mariadb/sql/ha_partition.h
monty@mysql.com 74cc73d461 This changeset is largely a handler cleanup changeset (WL#3281), but includes fixes and cleanups that was found necessary while testing the handler changes
Changes that requires code changes in other code of other storage engines.
(Note that all changes are very straightforward and one should find all issues
by compiling a --debug build and fixing all compiler errors and all
asserts in field.cc while running the test suite),

- New optional handler function introduced: reset()
  This is called after every DML statement to make it easy for a handler to
  statement specific cleanups.
  (The only case it's not called is if force the file to be closed)

- handler::extra(HA_EXTRA_RESET) is removed. Code that was there before
  should be moved to handler::reset()

- table->read_set contains a bitmap over all columns that are needed
  in the query.  read_row() and similar functions only needs to read these
  columns
- table->write_set contains a bitmap over all columns that will be updated
  in the query. write_row() and update_row() only needs to update these
  columns.
  The above bitmaps should now be up to date in all context
  (including ALTER TABLE, filesort()).

  The handler is informed of any changes to the bitmap after
  fix_fields() by calling the virtual function
  handler::column_bitmaps_signal(). If the handler does caching of
  these bitmaps (instead of using table->read_set, table->write_set),
  it should redo the caching in this code. as the signal() may be sent
  several times, it's probably best to set a variable in the signal
  and redo the caching on read_row() / write_row() if the variable was
  set.

- Removed the read_set and write_set bitmap objects from the handler class

- Removed all column bit handling functions from the handler class.
  (Now one instead uses the normal bitmap functions in my_bitmap.c instead
  of handler dedicated bitmap functions)

- field->query_id is removed. One should instead instead check
  table->read_set and table->write_set if a field is used in the query.

- handler::extra(HA_EXTRA_RETRIVE_ALL_COLS) and
  handler::extra(HA_EXTRA_RETRIEVE_PRIMARY_KEY) are removed. One should now
  instead use table->read_set to check for which columns to retrieve.

- If a handler needs to call Field->val() or Field->store() on columns
  that are not used in the query, one should install a temporary
  all-columns-used map while doing so. For this, we provide the following
  functions:

  my_bitmap_map *old_map= dbug_tmp_use_all_columns(table, table->read_set);
  field->val();
  dbug_tmp_restore_column_map(table->read_set, old_map);

  and similar for the write map:

  my_bitmap_map *old_map= dbug_tmp_use_all_columns(table, table->write_set);
  field->val();
  dbug_tmp_restore_column_map(table->write_set, old_map);

  If this is not done, you will sooner or later hit a DBUG_ASSERT
  in the field store() / val() functions.
  (For not DBUG binaries, the dbug_tmp_restore_column_map() and
  dbug_tmp_restore_column_map() are inline dummy functions and should
  be optimized away be the compiler).

- If one needs to temporary set the column map for all binaries (and not
  just to avoid the DBUG_ASSERT() in the Field::store() / Field::val()
  methods) one should use the functions tmp_use_all_columns() and
  tmp_restore_column_map() instead of the above dbug_ variants.

- All 'status' fields in the handler base class (like records,
  data_file_length etc) are now stored in a 'stats' struct. This makes
  it easier to know what status variables are provided by the base
  handler.  This requires some trivial variable names in the extra()
  function.

- New virtual function handler::records().  This is called to optimize
  COUNT(*) if (handler::table_flags() & HA_HAS_RECORDS()) is true.
  (stats.records is not supposed to be an exact value. It's only has to
  be 'reasonable enough' for the optimizer to be able to choose a good
  optimization path).

- Non virtual handler::init() function added for caching of virtual
  constants from engine.

- Removed has_transactions() virtual method. Now one should instead return
  HA_NO_TRANSACTIONS in table_flags() if the table handler DOES NOT support
  transactions.

- The 'xxxx_create_handler()' function now has a MEM_ROOT_root argument
  that is to be used with 'new handler_name()' to allocate the handler
  in the right area.  The xxxx_create_handler() function is also
  responsible for any initialization of the object before returning.

  For example, one should change:

  static handler *myisam_create_handler(TABLE_SHARE *table)
  {
    return new ha_myisam(table);
  }

  ->

  static handler *myisam_create_handler(TABLE_SHARE *table, MEM_ROOT *mem_root)
  {
    return new (mem_root) ha_myisam(table);
  }

- New optional virtual function: use_hidden_primary_key().
  This is called in case of an update/delete when
  (table_flags() and HA_PRIMARY_KEY_REQUIRED_FOR_DELETE) is defined
  but we don't have a primary key. This allows the handler to take precisions
  in remembering any hidden primary key to able to update/delete any
  found row. The default handler marks all columns to be read.

- handler::table_flags() now returns a ulonglong (to allow for more flags).

- New/changed table_flags()
  - HA_HAS_RECORDS	    Set if ::records() is supported
  - HA_NO_TRANSACTIONS	    Set if engine doesn't support transactions
  - HA_PRIMARY_KEY_REQUIRED_FOR_DELETE
                            Set if we should mark all primary key columns for
			    read when reading rows as part of a DELETE
			    statement. If there is no primary key,
			    all columns are marked for read.
  - HA_PARTIAL_COLUMN_READ  Set if engine will not read all columns in some
			    cases (based on table->read_set)
 - HA_PRIMARY_KEY_ALLOW_RANDOM_ACCESS
   			    Renamed to HA_PRIMARY_KEY_REQUIRED_FOR_POSITION.
 - HA_DUPP_POS              Renamed to HA_DUPLICATE_POS
 - HA_REQUIRES_KEY_COLUMNS_FOR_DELETE
			    Set this if we should mark ALL key columns for
			    read when when reading rows as part of a DELETE
			    statement. In case of an update we will mark
			    all keys for read for which key part changed
			    value.
  - HA_STATS_RECORDS_IS_EXACT
			     Set this if stats.records is exact.
			     (This saves us some extra records() calls
			     when optimizing COUNT(*))
			    

- Removed table_flags()
  - HA_NOT_EXACT_COUNT     Now one should instead use HA_HAS_RECORDS if
			   handler::records() gives an exact count() and
			   HA_STATS_RECORDS_IS_EXACT if stats.records is exact.
  - HA_READ_RND_SAME	   Removed (no one supported this one)

- Removed not needed functions ha_retrieve_all_cols() and ha_retrieve_all_pk()

- Renamed handler::dupp_pos to handler::dup_pos

- Removed not used variable handler::sortkey


Upper level handler changes:

- ha_reset() now does some overall checks and calls ::reset()
- ha_table_flags() added. This is a cached version of table_flags(). The
  cache is updated on engine creation time and updated on open.


MySQL level changes (not obvious from the above):

- DBUG_ASSERT() added to check that column usage matches what is set
  in the column usage bit maps. (This found a LOT of bugs in current
  column marking code).

- In 5.1 before, all used columns was marked in read_set and only updated
  columns was marked in write_set. Now we only mark columns for which we
  need a value in read_set.

- Column bitmaps are created in open_binary_frm() and open_table_from_share().
  (Before this was in table.cc)

- handler::table_flags() calls are replaced with handler::ha_table_flags()

- For calling field->val() you must have the corresponding bit set in
  table->read_set. For calling field->store() you must have the
  corresponding bit set in table->write_set. (There are asserts in
  all store()/val() functions to catch wrong usage)

- thd->set_query_id is renamed to thd->mark_used_columns and instead
  of setting this to an integer value, this has now the values:
  MARK_COLUMNS_NONE, MARK_COLUMNS_READ, MARK_COLUMNS_WRITE
  Changed also all variables named 'set_query_id' to mark_used_columns.

- In filesort() we now inform the handler of exactly which columns are needed
  doing the sort and choosing the rows.

- The TABLE_SHARE object has a 'all_set' column bitmap one can use
  when one needs a column bitmap with all columns set.
  (This is used for table->use_all_columns() and other places)

- The TABLE object has 3 column bitmaps:
  - def_read_set     Default bitmap for columns to be read
  - def_write_set    Default bitmap for columns to be written
  - tmp_set          Can be used as a temporary bitmap when needed.
  The table object has also two pointer to bitmaps read_set and write_set
  that the handler should use to find out which columns are used in which way.

- count() optimization now calls handler::records() instead of using
  handler->stats.records (if (table_flags() & HA_HAS_RECORDS) is true).

- Added extra argument to Item::walk() to indicate if we should also
  traverse sub queries.

- Added TABLE parameter to cp_buffer_from_ref()

- Don't close tables created with CREATE ... SELECT but keep them in
  the table cache. (Faster usage of newly created tables).


New interfaces:

- table->clear_column_bitmaps() to initialize the bitmaps for tables
  at start of new statements.

- table->column_bitmaps_set() to set up new column bitmaps and signal
  the handler about this.

- table->column_bitmaps_set_no_signal() for some few cases where we need
  to setup new column bitmaps but don't signal the handler (as the handler
  has already been signaled about these before). Used for the momement
  only in opt_range.cc when doing ROR scans.

- table->use_all_columns() to install a bitmap where all columns are marked
  as use in the read and the write set.

- table->default_column_bitmaps() to install the normal read and write
  column bitmaps, but not signaling the handler about this.
  This is mainly used when creating TABLE instances.

- table->mark_columns_needed_for_delete(),
  table->mark_columns_needed_for_delete() and
  table->mark_columns_needed_for_insert() to allow us to put additional
  columns in column usage maps if handler so requires.
  (The handler indicates what it neads in handler->table_flags())

- table->prepare_for_position() to allow us to tell handler that it
  needs to read primary key parts to be able to store them in
  future table->position() calls.
  (This replaces the table->file->ha_retrieve_all_pk function)

- table->mark_auto_increment_column() to tell handler are going to update
  columns part of any auto_increment key.

- table->mark_columns_used_by_index() to mark all columns that is part of
  an index.  It will also send extra(HA_EXTRA_KEYREAD) to handler to allow
  it to quickly know that it only needs to read colums that are part
  of the key.  (The handler can also use the column map for detecting this,
  but simpler/faster handler can just monitor the extra() call).

- table->mark_columns_used_by_index_no_reset() to in addition to other columns,
  also mark all columns that is used by the given key.

- table->restore_column_maps_after_mark_index() to restore to default
  column maps after a call to table->mark_columns_used_by_index().

- New item function register_field_in_read_map(), for marking used columns
  in table->read_map. Used by filesort() to mark all used columns

- Maintain in TABLE->merge_keys set of all keys that are used in query.
  (Simplices some optimization loops)

- Maintain Field->part_of_key_not_clustered which is like Field->part_of_key
  but the field in the clustered key is not assumed to be part of all index.
  (used in opt_range.cc for faster loops)

-  dbug_tmp_use_all_columns(), dbug_tmp_restore_column_map()
   tmp_use_all_columns() and tmp_restore_column_map() functions to temporally
   mark all columns as usable.  The 'dbug_' version is primarily intended
   inside a handler when it wants to just call Field:store() & Field::val()
   functions, but don't need the column maps set for any other usage.
   (ie:: bitmap_is_set() is never called)

- We can't use compare_records() to skip updates for handlers that returns
  a partial column set and the read_set doesn't cover all columns in the
  write set. The reason for this is that if we have a column marked only for
  write we can't in the MySQL level know if the value changed or not.
  The reason this worked before was that MySQL marked all to be written
  columns as also to be read. The new 'optimal' bitmaps exposed this 'hidden
  bug'.

- open_table_from_share() does not anymore setup temporary MEM_ROOT
  object as a thread specific variable for the handler. Instead we
  send the to-be-used MEMROOT to get_new_handler().
  (Simpler, faster code)



Bugs fixed:

- Column marking was not done correctly in a lot of cases.
  (ALTER TABLE, when using triggers, auto_increment fields etc)
  (Could potentially result in wrong values inserted in table handlers
  relying on that the old column maps or field->set_query_id was correct)
  Especially when it comes to triggers, there may be cases where the
  old code would cause lost/wrong values for NDB and/or InnoDB tables.

- Split thd->options flag OPTION_STATUS_NO_TRANS_UPDATE to two flags:
  OPTION_STATUS_NO_TRANS_UPDATE and OPTION_KEEP_LOG.
  This allowed me to remove some wrong warnings about:
  "Some non-transactional changed tables couldn't be rolled back"

- Fixed handling of INSERT .. SELECT and CREATE ... SELECT that wrongly reset
  (thd->options & OPTION_STATUS_NO_TRANS_UPDATE) which caused us to loose
  some warnings about
  "Some non-transactional changed tables couldn't be rolled back")

- Fixed use of uninitialized memory in ha_ndbcluster.cc::delete_table()
  which could cause delete_table to report random failures.

- Fixed core dumps for some tests when running with --debug

- Added missing FN_LIBCHAR in mysql_rm_tmp_tables()
  (This has probably caused us to not properly remove temporary files after
  crash)

- slow_logs was not properly initialized, which could maybe cause
  extra/lost entries in slow log.

- If we get an duplicate row on insert, change column map to read and
  write all columns while retrying the operation. This is required by
  the definition of REPLACE and also ensures that fields that are only
  part of UPDATE are properly handled.  This fixed a bug in NDB and
  REPLACE where REPLACE wrongly copied some column values from the replaced
  row.

- For table handler that doesn't support NULL in keys, we would give an error
  when creating a primary key with NULL fields, even after the fields has been
  automaticly converted to NOT NULL.

- Creating a primary key on a SPATIAL key, would fail if field was not
  declared as NOT NULL.


Cleanups:

- Removed not used condition argument to setup_tables

- Removed not needed item function reset_query_id_processor().

- Field->add_index is removed. Now this is instead maintained in
  (field->flags & FIELD_IN_ADD_INDEX)

- Field->fieldnr is removed (use field->field_index instead)

- New argument to filesort() to indicate that it should return a set of
  row pointers (not used columns). This allowed me to remove some references
  to sql_command in filesort and should also enable us to return column
  results in some cases where we couldn't before.

- Changed column bitmap handling in opt_range.cc to be aligned with TABLE
  bitmap, which allowed me to use bitmap functions instead of looping over
  all fields to create some needed bitmaps. (Faster and smaller code)

- Broke up found too long lines

- Moved some variable declaration at start of function for better code
  readability.

- Removed some not used arguments from functions.
  (setup_fields(), mysql_prepare_insert_check_table())

- setup_fields() now takes an enum instead of an int for marking columns
   usage.

- For internal temporary tables, use handler::write_row(),
  handler::delete_row() and handler::update_row() instead of
  handler::ha_xxxx() for faster execution.

- Changed some constants to enum's and define's.

- Using separate column read and write sets allows for easier checking
  of timestamp field was set by statement.

- Remove calls to free_io_cache() as this is now done automaticly in ha_reset()

- Don't build table->normalized_path as this is now identical to table->path
  (after bar's fixes to convert filenames)

- Fixed some missed DBUG_PRINT(.."%lx") to use "0x%lx" to make it easier to
  do comparision with the 'convert-dbug-for-diff' tool.


Things left to do in 5.1:

- We wrongly log failed CREATE TABLE ... SELECT in some cases when using
  row based logging (as shown by testcase binlog_row_mix_innodb_myisam.result)
  Mats has promised to look into this.

- Test that my fix for CREATE TABLE ... SELECT is indeed correct.
  (I added several test cases for this, but in this case it's better that
  someone else also tests this throughly).
  Lars has promosed to do this.
2006-06-04 18:52:22 +03:00

953 lines
39 KiB
C++

/* Copyright (C) 2005 MySQL AB
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */
#ifdef __GNUC__
#pragma interface /* gcc class implementation */
#endif
enum partition_keywords
{
PKW_HASH= 0, PKW_RANGE, PKW_LIST, PKW_KEY, PKW_MAXVALUE, PKW_LINEAR
};
/*
PARTITION_SHARE is a structure that will be shared amoung all open handlers
The partition implements the minimum of what you will probably need.
*/
typedef struct st_partition_share
{
char *table_name;
uint table_name_length, use_count;
pthread_mutex_t mutex;
THR_LOCK lock;
} PARTITION_SHARE;
#define PARTITION_BYTES_IN_POS 2
class ha_partition :public handler
{
private:
enum partition_index_scan_type
{
partition_index_read= 0,
partition_index_first= 1,
partition_index_last= 2,
partition_no_index_scan= 3
};
/* Data for the partition handler */
int m_mode; // Open mode
uint m_open_test_lock; // Open test_if_locked
char *m_file_buffer; // Buffer with names
char *m_name_buffer_ptr; // Pointer to first partition name
handlerton **m_engine_array; // Array of types of the handlers
handler **m_file; // Array of references to handler inst.
handler **m_new_file; // Array of references to new handlers
handler **m_reorged_file; // Reorganised partitions
handler **m_added_file; // Added parts kept for errors
partition_info *m_part_info; // local reference to partition
byte *m_start_key_ref; // Reference of start key in current
// index scan info
Field **m_part_field_array; // Part field array locally to save acc
byte *m_ordered_rec_buffer; // Row and key buffer for ord. idx scan
KEY *m_curr_key_info; // Current index
byte *m_rec0; // table->record[0]
QUEUE m_queue; // Prio queue used by sorted read
/*
Since the partition handler is a handler on top of other handlers, it
is necessary to keep information about what the underlying handler
characteristics is. It is not possible to keep any handler instances
for this since the MySQL Server sometimes allocating the handler object
without freeing them.
*/
u_long m_table_flags;
u_long m_low_byte_first;
uint m_reorged_parts; // Number of reorganised parts
uint m_tot_parts; // Total number of partitions;
uint m_no_locks; // For engines like ha_blackhole, which needs no locks
uint m_last_part; // Last file that we update,write
int m_lock_type; // Remembers type of last
// external_lock
part_id_range m_part_spec; // Which parts to scan
uint m_scan_value; // Value passed in rnd_init
// call
uint m_ref_length; // Length of position in this
// handler object
key_range m_start_key; // index read key range
enum partition_index_scan_type m_index_scan_type;// What type of index
// scan
uint m_top_entry; // Which partition is to
// deliver next result
uint m_rec_length; // Local copy of record length
bool m_ordered; // Ordered/Unordered index scan
bool m_pkey_is_clustered; // Is primary key clustered
bool m_create_handler; // Handler used to create table
bool m_is_sub_partitioned; // Is subpartitioned
bool m_ordered_scan_ongoing;
/*
We keep track if all underlying handlers are MyISAM since MyISAM has a
great number of extra flags not needed by other handlers.
*/
bool m_myisam; // Are all underlying handlers
// MyISAM
/*
We keep track of InnoDB handlers below since it requires proper setting
of query_id in fields at index_init and index_read calls.
*/
bool m_innodb; // Are all underlying handlers
// InnoDB
/*
When calling extra(HA_EXTRA_CACHE) we do not pass this to the underlying
handlers immediately. Instead we cache it and call the underlying
immediately before starting the scan on the partition. This is to
prevent allocating a READ CACHE for each partition in parallel when
performing a full table scan on MyISAM partitioned table.
This state is cleared by extra(HA_EXTRA_NO_CACHE).
*/
bool m_extra_cache;
uint m_extra_cache_size;
void init_handler_variables();
/*
Variables for lock structures.
*/
THR_LOCK_DATA lock; /* MySQL lock */
PARTITION_SHARE *share; /* Shared lock info */
public:
virtual void set_part_info(partition_info *part_info)
{
m_part_info= part_info;
m_is_sub_partitioned= part_info->is_sub_partitioned();
}
/*
-------------------------------------------------------------------------
MODULE create/delete handler object
-------------------------------------------------------------------------
Object create/delete methode. The normal called when a table object
exists. There is also a method to create the handler object with only
partition information. This is used from mysql_create_table when the
table is to be created and the engine type is deduced to be the
partition handler.
-------------------------------------------------------------------------
*/
ha_partition(TABLE_SHARE * table);
ha_partition(partition_info * part_info);
~ha_partition();
/*
A partition handler has no characteristics in itself. It only inherits
those from the underlying handlers. Here we set-up those constants to
enable later calls of the methods to retrieve constants from the under-
lying handlers. Returns false if not successful.
*/
bool initialise_partition(MEM_ROOT *mem_root);
/*
-------------------------------------------------------------------------
MODULE meta data changes
-------------------------------------------------------------------------
Meta data routines to CREATE, DROP, RENAME table and often used at
ALTER TABLE (update_create_info used from ALTER TABLE and SHOW ..).
update_table_comment is used in SHOW TABLE commands to provide a
chance for the handler to add any interesting comments to the table
comments not provided by the users comment.
create_handler_files is called before opening a new handler object
with openfrm to call create. It is used to create any local handler
object needed in opening the object in openfrm
-------------------------------------------------------------------------
*/
virtual int delete_table(const char *from);
virtual int rename_table(const char *from, const char *to);
virtual int create(const char *name, TABLE *form,
HA_CREATE_INFO *create_info);
virtual int create_handler_files(const char *name,
const char *old_name, int action_flag,
HA_CREATE_INFO *create_info);
virtual void update_create_info(HA_CREATE_INFO *create_info);
virtual char *update_table_comment(const char *comment);
virtual int change_partitions(HA_CREATE_INFO *create_info,
const char *path,
ulonglong *copied,
ulonglong *deleted,
const void *pack_frm_data,
uint pack_frm_len);
virtual int drop_partitions(const char *path);
virtual int rename_partitions(const char *path);
bool get_no_parts(const char *name, uint *no_parts)
{
DBUG_ENTER("ha_partition::get_no_parts");
*no_parts= m_tot_parts;
DBUG_RETURN(0);
}
private:
int prepare_for_delete();
int copy_partitions(ulonglong *copied, ulonglong *deleted);
void cleanup_new_partition(uint part_count);
int prepare_new_partition(TABLE *table, HA_CREATE_INFO *create_info,
handler *file, const char *part_name);
/*
delete_table, rename_table and create uses very similar logic which
is packed into this routine.
*/
uint del_ren_cre_table(const char *from, const char *to,
TABLE *table_arg, HA_CREATE_INFO *create_info);
/*
One method to create the table_name.par file containing the names of the
underlying partitions, their engine and the number of partitions.
And one method to read it in.
*/
bool create_handler_file(const char *name);
bool get_from_handler_file(const char *name, MEM_ROOT *mem_root);
bool new_handlers_from_part_info(MEM_ROOT *mem_root);
bool create_handlers(MEM_ROOT *mem_root);
void clear_handler_file();
void set_up_table_before_create(TABLE *table_arg,
const char *partition_name_with_path,
HA_CREATE_INFO *info,
uint part_id);
partition_element *find_partition_element(uint part_id);
public:
/*
-------------------------------------------------------------------------
MODULE open/close object
-------------------------------------------------------------------------
Open and close handler object to ensure all underlying files and
objects allocated and deallocated for query handling is handled
properly.
-------------------------------------------------------------------------
A handler object is opened as part of its initialisation and before
being used for normal queries (not before meta-data changes always.
If the object was opened it will also be closed before being deleted.
*/
virtual int open(const char *name, int mode, uint test_if_locked);
virtual int close(void);
/*
-------------------------------------------------------------------------
MODULE start/end statement
-------------------------------------------------------------------------
This module contains methods that are used to understand start/end of
statements, transaction boundaries, and aid for proper concurrency
control.
The partition handler need not implement abort and commit since this
will be handled by any underlying handlers implementing transactions.
There is only one call to each handler type involved per transaction
and these go directly to the handlers supporting transactions
currently InnoDB, BDB and NDB).
-------------------------------------------------------------------------
*/
virtual THR_LOCK_DATA **store_lock(THD * thd, THR_LOCK_DATA ** to,
enum thr_lock_type lock_type);
virtual int external_lock(THD * thd, int lock_type);
/*
When table is locked a statement is started by calling start_stmt
instead of external_lock
*/
virtual int start_stmt(THD * thd, thr_lock_type lock_type);
/*
Lock count is number of locked underlying handlers (I assume)
*/
virtual uint lock_count(void) const;
/*
Call to unlock rows not to be updated in transaction
*/
virtual void unlock_row();
/*
-------------------------------------------------------------------------
MODULE change record
-------------------------------------------------------------------------
This part of the handler interface is used to change the records
after INSERT, DELETE, UPDATE, REPLACE method calls but also other
special meta-data operations as ALTER TABLE, LOAD DATA, TRUNCATE.
-------------------------------------------------------------------------
These methods are used for insert (write_row), update (update_row)
and delete (delete_row). All methods to change data always work on
one row at a time. update_row and delete_row also contains the old
row.
delete_all_rows will delete all rows in the table in one call as a
special optimisation for DELETE from table;
Bulk inserts are supported if all underlying handlers support it.
start_bulk_insert and end_bulk_insert is called before and after a
number of calls to write_row.
Not yet though.
*/
virtual int write_row(byte * buf);
virtual int update_row(const byte * old_data, byte * new_data);
virtual int delete_row(const byte * buf);
virtual int delete_all_rows(void);
virtual void start_bulk_insert(ha_rows rows);
virtual int end_bulk_insert();
/*
-------------------------------------------------------------------------
MODULE full table scan
-------------------------------------------------------------------------
This module is used for the most basic access method for any table
handler. This is to fetch all data through a full table scan. No
indexes are needed to implement this part.
It contains one method to start the scan (rnd_init) that can also be
called multiple times (typical in a nested loop join). Then proceeding
to the next record (rnd_next) and closing the scan (rnd_end).
To remember a record for later access there is a method (position)
and there is a method used to retrieve the record based on the stored
position.
The position can be a file position, a primary key, a ROWID dependent
on the handler below.
-------------------------------------------------------------------------
*/
/*
unlike index_init(), rnd_init() can be called two times
without rnd_end() in between (it only makes sense if scan=1).
then the second call should prepare for the new table scan
(e.g if rnd_init allocates the cursor, second call should
position it to the start of the table, no need to deallocate
and allocate it again
*/
virtual int rnd_init(bool scan);
virtual int rnd_end();
virtual int rnd_next(byte * buf);
virtual int rnd_pos(byte * buf, byte * pos);
virtual void position(const byte * record);
/*
-------------------------------------------------------------------------
MODULE index scan
-------------------------------------------------------------------------
This part of the handler interface is used to perform access through
indexes. The interface is defined as a scan interface but the handler
can also use key lookup if the index is a unique index or a primary
key index.
Index scans are mostly useful for SELECT queries but are an important
part also of UPDATE, DELETE, REPLACE and CREATE TABLE table AS SELECT
and so forth.
Naturally an index is needed for an index scan and indexes can either
be ordered, hash based. Some ordered indexes can return data in order
but not necessarily all of them.
There are many flags that define the behavior of indexes in the
various handlers. These methods are found in the optimizer module.
-------------------------------------------------------------------------
index_read is called to start a scan of an index. The find_flag defines
the semantics of the scan. These flags are defined in
include/my_base.h
index_read_idx is the same but also initializes index before calling doing
the same thing as index_read. Thus it is similar to index_init followed
by index_read. This is also how we implement it.
index_read/index_read_idx does also return the first row. Thus for
key lookups, the index_read will be the only call to the handler in
the index scan.
index_init initializes an index before using it and index_end does
any end processing needed.
*/
virtual int index_read(byte * buf, const byte * key,
uint key_len, enum ha_rkey_function find_flag);
virtual int index_read_idx(byte * buf, uint idx, const byte * key,
uint key_len, enum ha_rkey_function find_flag);
virtual int index_init(uint idx, bool sorted);
virtual int index_end();
/*
These methods are used to jump to next or previous entry in the index
scan. There are also methods to jump to first and last entry.
*/
virtual int index_next(byte * buf);
virtual int index_prev(byte * buf);
virtual int index_first(byte * buf);
virtual int index_last(byte * buf);
virtual int index_next_same(byte * buf, const byte * key, uint keylen);
virtual int index_read_last(byte * buf, const byte * key, uint keylen);
/*
read_first_row is virtual method but is only implemented by
handler.cc, no storage engine has implemented it so neither
will the partition handler.
virtual int read_first_row(byte *buf, uint primary_key);
*/
/*
We don't implement multi read range yet, will do later.
virtual int read_multi_range_first(KEY_MULTI_RANGE **found_range_p,
KEY_MULTI_RANGE *ranges, uint range_count,
bool sorted, HANDLER_BUFFER *buffer);
virtual int read_multi_range_next(KEY_MULTI_RANGE **found_range_p);
*/
virtual int read_range_first(const key_range * start_key,
const key_range * end_key,
bool eq_range, bool sorted);
virtual int read_range_next();
private:
int common_index_read(byte * buf, const byte * key,
uint key_len, enum ha_rkey_function find_flag);
int common_first_last(byte * buf);
int partition_scan_set_up(byte * buf, bool idx_read_flag);
int handle_unordered_next(byte * buf, bool next_same);
int handle_unordered_scan_next_partition(byte * buf);
byte *queue_buf(uint part_id)
{
return (m_ordered_rec_buffer +
(part_id * (m_rec_length + PARTITION_BYTES_IN_POS)));
}
byte *rec_buf(uint part_id)
{
return (queue_buf(part_id) +
PARTITION_BYTES_IN_POS);
}
int handle_ordered_index_scan(byte * buf);
int handle_ordered_next(byte * buf, bool next_same);
int handle_ordered_prev(byte * buf);
void return_top_record(byte * buf);
void include_partition_fields_in_used_fields();
public:
/*
-------------------------------------------------------------------------
MODULE information calls
-------------------------------------------------------------------------
This calls are used to inform the handler of specifics of the ongoing
scans and other actions. Most of these are used for optimisation
purposes.
-------------------------------------------------------------------------
*/
virtual void info(uint);
void get_dynamic_partition_info(PARTITION_INFO *stat_info,
uint part_id);
virtual int extra(enum ha_extra_function operation);
virtual int extra_opt(enum ha_extra_function operation, ulong cachesize);
virtual int reset(void);
private:
static const uint NO_CURRENT_PART_ID= 0xFFFFFFFF;
int loop_extra(enum ha_extra_function operation);
void late_extra_cache(uint partition_id);
void late_extra_no_cache(uint partition_id);
void prepare_extra_cache(uint cachesize);
public:
/*
-------------------------------------------------------------------------
MODULE optimiser support
-------------------------------------------------------------------------
-------------------------------------------------------------------------
*/
/*
NOTE !!!!!!
-------------------------------------------------------------------------
-------------------------------------------------------------------------
One important part of the public handler interface that is not depicted in
the methods is the attribute records
which is defined in the base class. This is looked upon directly and is
set by calling info(HA_STATUS_INFO) ?
-------------------------------------------------------------------------
*/
/*
keys_to_use_for_scanning can probably be implemented as the
intersection of all underlying handlers if mixed handlers are used.
This method is used to derive whether an index can be used for
index-only scanning when performing an ORDER BY query.
Only called from one place in sql_select.cc
*/
virtual const key_map *keys_to_use_for_scanning();
/*
Called in test_quick_select to determine if indexes should be used.
*/
virtual double scan_time();
/*
The next method will never be called if you do not implement indexes.
*/
virtual double read_time(uint index, uint ranges, ha_rows rows);
/*
For the given range how many records are estimated to be in this range.
Used by optimiser to calculate cost of using a particular index.
*/
virtual ha_rows records_in_range(uint inx, key_range * min_key,
key_range * max_key);
/*
Upper bound of number records returned in scan is sum of all
underlying handlers.
*/
virtual ha_rows estimate_rows_upper_bound();
/*
table_cache_type is implemented by the underlying handler but all
underlying handlers must have the same implementation for it to work.
*/
virtual uint8 table_cache_type();
/*
-------------------------------------------------------------------------
MODULE print messages
-------------------------------------------------------------------------
This module contains various methods that returns text messages for
table types, index type and error messages.
-------------------------------------------------------------------------
*/
/*
The name of the index type that will be used for display
Here we must ensure that all handlers use the same index type
for each index created.
*/
virtual const char *index_type(uint inx);
/* The name of the table type that will be used for display purposes */
virtual const char *table_type() const;
/* The name of the row type used for the underlying tables. */
virtual enum row_type get_row_type() const;
/*
Handler specific error messages
*/
virtual void print_error(int error, myf errflag);
virtual bool get_error_message(int error, String * buf);
/*
-------------------------------------------------------------------------
MODULE handler characteristics
-------------------------------------------------------------------------
This module contains a number of methods defining limitations and
characteristics of the handler. The partition handler will calculate
this characteristics based on underlying handler characteristics.
-------------------------------------------------------------------------
This is a list of flags that says what the storage engine
implements. The current table flags are documented in handler.h
The partition handler will support whatever the underlying handlers
support except when specifically mentioned below about exceptions
to this rule.
HA_READ_RND_SAME:
Not currently used. (Means that the handler supports the rnd_same() call)
(MyISAM, HEAP)
HA_TABLE_SCAN_ON_INDEX:
Used to avoid scanning full tables on an index. If this flag is set then
the handler always has a primary key (hidden if not defined) and this
index is used for scanning rather than a full table scan in all
situations.
(InnoDB, BDB, Federated)
HA_REC_NOT_IN_SEQ:
This flag is set for handlers that cannot guarantee that the rows are
returned accroding to incremental positions (0, 1, 2, 3...).
This also means that rnd_next() should return HA_ERR_RECORD_DELETED
if it finds a deleted row.
(MyISAM (not fixed length row), BDB, HEAP, NDB, InooDB)
HA_CAN_GEOMETRY:
Can the storage engine handle spatial data.
Used to check that no spatial attributes are declared unless
the storage engine is capable of handling it.
(MyISAM)
HA_FAST_KEY_READ:
Setting this flag indicates that the handler is equally fast in
finding a row by key as by position.
This flag is used in a very special situation in conjunction with
filesort's. For further explanation see intro to init_read_record.
(BDB, HEAP, InnoDB)
HA_NULL_IN_KEY:
Is NULL values allowed in indexes.
If this is not allowed then it is not possible to use an index on a
NULLable field.
(BDB, HEAP, MyISAM, NDB, InnoDB)
HA_DUPLICATE_POS:
Tells that we can the position for the conflicting duplicate key
record is stored in table->file->dupp_ref. (insert uses rnd_pos() on
this to find the duplicated row)
(MyISAM)
HA_CAN_INDEX_BLOBS:
Is the storage engine capable of defining an index of a prefix on
a BLOB attribute.
(BDB, Federated, MyISAM, InnoDB)
HA_AUTO_PART_KEY:
Auto increment fields can be part of a multi-part key. For second part
auto-increment keys, the auto_incrementing is done in handler.cc
(BDB, Federated, MyISAM, NDB)
HA_REQUIRE_PRIMARY_KEY:
Can't define a table without primary key (and cannot handle a table
with hidden primary key)
(No handler has this limitation currently)
HA_STATS_RECORDS_IS_EXACT:
Does the counter of records after the info call specify an exact
value or not. If it does this flag is set.
Only MyISAM and HEAP uses exact count.
HA_CAN_INSERT_DELAYED:
Can the storage engine support delayed inserts.
To start with the partition handler will not support delayed inserts.
Further investigation needed.
(HEAP, MyISAM)
HA_PRIMARY_KEY_IN_READ_INDEX:
This parameter is set when the handler will also return the primary key
when doing read-only-key on another index.
HA_NOT_DELETE_WITH_CACHE:
Seems to be an old MyISAM feature that is no longer used. No handler
has it defined but it is checked in init_read_record.
Further investigation needed.
(No handler defines it)
HA_NO_PREFIX_CHAR_KEYS:
Indexes on prefixes of character fields is not allowed.
(NDB)
HA_CAN_FULLTEXT:
Does the storage engine support fulltext indexes
The partition handler will start by not supporting fulltext indexes.
(MyISAM)
HA_CAN_SQL_HANDLER:
Can the HANDLER interface in the MySQL API be used towards this
storage engine.
(MyISAM, InnoDB)
HA_NO_AUTO_INCREMENT:
Set if the storage engine does not support auto increment fields.
(Currently not set by any handler)
HA_HAS_CHECKSUM:
Special MyISAM feature. Has special SQL support in CREATE TABLE.
No special handling needed by partition handler.
(MyISAM)
HA_FILE_BASED:
Should file names always be in lower case (used by engines
that map table names to file names.
Since partition handler has a local file this flag is set.
(BDB, Federated, MyISAM)
HA_CAN_BIT_FIELD:
Is the storage engine capable of handling bit fields?
(MyISAM, NDB)
HA_NEED_READ_RANGE_BUFFER:
Is Read Multi-Range supported => need multi read range buffer
This parameter specifies whether a buffer for read multi range
is needed by the handler. Whether the handler supports this
feature or not is dependent of whether the handler implements
read_multi_range* calls or not. The only handler currently
supporting this feature is NDB so the partition handler need
not handle this call. There are methods in handler.cc that will
transfer those calls into index_read and other calls in the
index scan module.
(NDB)
*/
virtual ulonglong table_flags() const
{ return m_table_flags; }
/*
This is a bitmap of flags that says how the storage engine
implements indexes. The current index flags are documented in
handler.h. If you do not implement indexes, just return zero
here.
part is the key part to check. First key part is 0
If all_parts it's set, MySQL want to know the flags for the combined
index up to and including 'part'.
HA_READ_NEXT:
Does the index support read next, this is assumed in the server
code and never checked so all indexes must support this.
Note that the handler can be used even if it doesn't have any index.
(BDB, HEAP, MyISAM, Federated, NDB, InnoDB)
HA_READ_PREV:
Can the index be used to scan backwards.
(BDB, HEAP, MyISAM, NDB, InnoDB)
HA_READ_ORDER:
Can the index deliver its record in index order. Typically true for
all ordered indexes and not true for hash indexes.
In first step this is not true for partition handler until a merge
sort has been implemented in partition handler.
Used to set keymap part_of_sortkey
This keymap is only used to find indexes usable for resolving an ORDER BY
in the query. Thus in most cases index_read will work just fine without
order in result production. When this flag is set it is however safe to
order all output started by index_read since most engines do this. With
read_multi_range calls there is a specific flag setting order or not
order so in those cases ordering of index output can be avoided.
(BDB, InnoDB, HEAP, MyISAM, NDB)
HA_READ_RANGE:
Specify whether index can handle ranges, typically true for all
ordered indexes and not true for hash indexes.
Used by optimiser to check if ranges (as key >= 5) can be optimised
by index.
(BDB, InnoDB, NDB, MyISAM, HEAP)
HA_ONLY_WHOLE_INDEX:
Can't use part key searches. This is typically true for hash indexes
and typically not true for ordered indexes.
(Federated, NDB, HEAP)
HA_KEYREAD_ONLY:
Does the storage engine support index-only scans on this index.
Enables use of HA_EXTRA_KEYREAD and HA_EXTRA_NO_KEYREAD
Used to set key_map keys_for_keyread and to check in optimiser for
index-only scans. When doing a read under HA_EXTRA_KEYREAD the handler
only have to fill in the columns the key covers. If
HA_PRIMARY_KEY_IN_READ_INDEX is set then also the PRIMARY KEY columns
must be updated in the row.
(BDB, InnoDB, MyISAM)
*/
virtual ulong index_flags(uint inx, uint part, bool all_parts) const
{
return m_file[0]->index_flags(inx, part, all_parts);
}
/*
extensions of table handler files
*/
virtual const char **bas_ext() const;
/*
unireg.cc will call the following to make sure that the storage engine
can handle the data it is about to send.
The maximum supported values is the minimum of all handlers in the table
*/
uint min_of_the_max_uint(uint (handler::*operator_func)(void) const) const;
virtual uint max_supported_record_length() const;
virtual uint max_supported_keys() const;
virtual uint max_supported_key_parts() const;
virtual uint max_supported_key_length() const;
virtual uint max_supported_key_part_length() const;
/*
All handlers in a partitioned table must have the same low_byte_first
*/
virtual bool low_byte_first() const
{ return m_low_byte_first; }
/*
The extra record buffer length is the maximum needed by all handlers.
The minimum record length is the maximum of all involved handlers.
*/
virtual uint extra_rec_buf_length() const;
virtual uint min_record_length(uint options) const;
/*
Primary key is clustered can only be true if all underlying handlers have
this feature.
*/
virtual bool primary_key_is_clustered()
{ return m_pkey_is_clustered; }
/*
-------------------------------------------------------------------------
MODULE compare records
-------------------------------------------------------------------------
cmp_ref checks if two references are the same. For most handlers this is
a simple memcmp of the reference. However some handlers use primary key
as reference and this can be the same even if memcmp says they are
different. This is due to character sets and end spaces and so forth.
For the partition handler the reference is first two bytes providing the
partition identity of the referred record and then the reference of the
underlying handler.
Thus cmp_ref for the partition handler always returns FALSE for records
not in the same partition and uses cmp_ref on the underlying handler
to check whether the rest of the reference part is also the same.
-------------------------------------------------------------------------
*/
virtual int cmp_ref(const byte * ref1, const byte * ref2);
/*
-------------------------------------------------------------------------
MODULE auto increment
-------------------------------------------------------------------------
This module is used to handle the support of auto increments.
This variable in the handler is used as part of the handler interface
It is maintained by the parent handler object and should not be
touched by child handler objects (see handler.cc for its use).
auto_increment_column_changed
-------------------------------------------------------------------------
*/
virtual void restore_auto_increment();
virtual ulonglong get_auto_increment();
/*
-------------------------------------------------------------------------
MODULE initialise handler for HANDLER call
-------------------------------------------------------------------------
This method is a special InnoDB method called before a HANDLER query.
-------------------------------------------------------------------------
*/
virtual void init_table_handle_for_HANDLER();
/*
The remainder of this file defines the handler methods not implemented
by the partition handler
*/
/*
-------------------------------------------------------------------------
MODULE foreign key support
-------------------------------------------------------------------------
The following methods are used to implement foreign keys as supported by
InnoDB. Implement this ??
get_foreign_key_create_info is used by SHOW CREATE TABLE to get a textual
description of how the CREATE TABLE part to define FOREIGN KEY's is done.
free_foreign_key_create_info is used to free the memory area that provided
this description.
can_switch_engines checks if it is ok to switch to a new engine based on
the foreign key info in the table.
-------------------------------------------------------------------------
virtual char* get_foreign_key_create_info()
virtual void free_foreign_key_create_info(char* str)
virtual int get_foreign_key_list(THD *thd,
List<FOREIGN_KEY_INFO> *f_key_list)
virtual uint referenced_by_foreign_key()
*/
virtual bool can_switch_engines();
/*
-------------------------------------------------------------------------
MODULE fulltext index
-------------------------------------------------------------------------
Fulltext stuff not yet.
-------------------------------------------------------------------------
virtual int ft_init() { return HA_ERR_WRONG_COMMAND; }
virtual FT_INFO *ft_init_ext(uint flags,uint inx,const byte *key,
uint keylen)
{ return NULL; }
virtual int ft_read(byte *buf) { return HA_ERR_WRONG_COMMAND; }
*/
/*
-------------------------------------------------------------------------
MODULE restart full table scan at position (MyISAM)
-------------------------------------------------------------------------
The following method is only used by MyISAM when used as
temporary tables in a join.
virtual int restart_rnd_next(byte *buf, byte *pos);
*/
/*
-------------------------------------------------------------------------
MODULE on-line ALTER TABLE
-------------------------------------------------------------------------
These methods are in the handler interface but never used (yet)
They are to be used by on-line alter table add/drop index:
-------------------------------------------------------------------------
virtual ulong index_ddl_flags(KEY *wanted_index) const
virtual int add_index(TABLE *table_arg,KEY *key_info,uint num_of_keys);
virtual int drop_index(TABLE *table_arg,uint *key_num,uint num_of_keys);
*/
/*
-------------------------------------------------------------------------
MODULE tablespace support
-------------------------------------------------------------------------
Admin of table spaces is not applicable to the partition handler (InnoDB)
This means that the following method is not implemented:
-------------------------------------------------------------------------
virtual int discard_or_import_tablespace(my_bool discard)
*/
/*
-------------------------------------------------------------------------
MODULE admin MyISAM
-------------------------------------------------------------------------
-------------------------------------------------------------------------
OPTIMIZE TABLE, CHECK TABLE, ANALYZE TABLE and REPAIR TABLE are
mapped to a routine that handles looping over a given set of
partitions and those routines send a flag indicating to execute on
all partitions.
-------------------------------------------------------------------------
*/
virtual int optimize(THD* thd, HA_CHECK_OPT *check_opt);
virtual int analyze(THD* thd, HA_CHECK_OPT *check_opt);
virtual int check(THD* thd, HA_CHECK_OPT *check_opt);
virtual int repair(THD* thd, HA_CHECK_OPT *check_opt);
virtual int optimize_partitions(THD *thd);
virtual int analyze_partitions(THD *thd);
virtual int check_partitions(THD *thd);
virtual int repair_partitions(THD *thd);
private:
int handle_opt_partitions(THD *thd, HA_CHECK_OPT *check_opt,
uint flags, bool all_parts);
public:
/*
-------------------------------------------------------------------------
Admin commands not supported currently (almost purely MyISAM routines)
This means that the following methods are not implemented:
-------------------------------------------------------------------------
virtual int backup(TD* thd, HA_CHECK_OPT *check_opt);
virtual int restore(THD* thd, HA_CHECK_OPT *check_opt);
virtual int assign_to_keycache(THD* thd, HA_CHECK_OPT *check_opt);
virtual int preload_keys(THD *thd, HA_CHECK_OPT *check_opt);
virtual bool check_and_repair(THD *thd);
virtual int dump(THD* thd, int fd = -1);
virtual int net_read_dump(NET* net);
virtual uint checksum() const;
virtual bool is_crashed() const;
virtual bool auto_repair() const;
-------------------------------------------------------------------------
MODULE enable/disable indexes
-------------------------------------------------------------------------
Enable/Disable Indexes are not supported currently (Heap, MyISAM)
This means that the following methods are not implemented:
-------------------------------------------------------------------------
virtual int disable_indexes(uint mode);
virtual int enable_indexes(uint mode);
virtual int indexes_are_disabled(void);
*/
/*
-------------------------------------------------------------------------
MODULE append_create_info
-------------------------------------------------------------------------
append_create_info is only used by MyISAM MERGE tables and the partition
handler will not support this handler as underlying handler.
Implement this??
-------------------------------------------------------------------------
virtual void append_create_info(String *packet)
*/
};