2005-05-06 10:39:30 +02:00
|
|
|
/* Copyright (C) 2000-2003 MySQL AB
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
|
|
it under the terms of the GNU General Public License as published by
|
|
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
|
|
(at your option) any later version.
|
|
|
|
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
GNU General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
along with this program; if not, write to the Free Software
|
2005-02-17 22:52:40 +01:00
|
|
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
This file defines the NDB Cluster handler: the interface between MySQL and
|
|
|
|
NDB Cluster
|
|
|
|
*/
|
|
|
|
|
2005-05-26 12:09:14 +02:00
|
|
|
#ifdef USE_PRAGMA_IMPLEMENTATION
|
2005-06-03 22:46:03 +02:00
|
|
|
#pragma implementation // gcc: Class implementation
|
2004-04-15 09:14:14 +02:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#include "mysql_priv.h"
|
|
|
|
|
|
|
|
#include <my_dir.h>
|
|
|
|
#include "ha_ndbcluster.h"
|
|
|
|
#include <ndbapi/NdbApi.hpp>
|
|
|
|
#include <ndbapi/NdbScanFilter.hpp>
|
2005-11-07 12:19:28 +01:00
|
|
|
#include <../util/Bitmask.hpp>
|
2005-09-15 02:33:28 +02:00
|
|
|
#include <ndbapi/NdbIndexStat.hpp>
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-12-20 12:36:14 +01:00
|
|
|
// options from from mysqld.cc
|
|
|
|
extern my_bool opt_ndb_optimized_node_selection;
|
|
|
|
extern const char *opt_ndbcluster_connectstring;
|
|
|
|
|
2005-11-07 16:25:06 +01:00
|
|
|
const char *ndb_distribution_names[]= {"KEYHASH", "LINHASH", NullS};
|
|
|
|
TYPELIB ndb_distribution_typelib= { array_elements(ndb_distribution_names)-1,
|
|
|
|
"", ndb_distribution_names, NULL };
|
|
|
|
const char *opt_ndb_distribution= ndb_distribution_names[ND_KEYHASH];
|
|
|
|
enum ndb_distribution opt_ndb_distribution_id= ND_KEYHASH;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Default value for parallelism
|
2004-12-06 09:05:30 +01:00
|
|
|
static const int parallelism= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
// Default value for max number of transactions
|
|
|
|
// createable against NDB from this handler
|
2005-06-17 16:07:54 +02:00
|
|
|
static const int max_transactions= 3; // should really be 2 but there is a transaction to much allocated when loch table is used
|
2004-07-23 15:46:56 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
static const char *ha_ndb_ext=".ndb";
|
2005-11-06 00:20:37 +01:00
|
|
|
static const char share_prefix[]= "./";
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
static int ndbcluster_close_connection(THD *thd);
|
|
|
|
static int ndbcluster_commit(THD *thd, bool all);
|
|
|
|
static int ndbcluster_rollback(THD *thd, bool all);
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
static handler* ndbcluster_create_handler(TABLE_SHARE *table);
|
2005-02-17 22:52:40 +01:00
|
|
|
|
2005-09-19 21:06:23 +02:00
|
|
|
handlerton ndbcluster_hton = {
|
2005-12-21 19:18:40 +01:00
|
|
|
MYSQL_HANDLERTON_INTERFACE_VERSION,
|
2005-03-13 21:58:09 +01:00
|
|
|
"ndbcluster",
|
2005-10-03 04:44:28 +02:00
|
|
|
SHOW_OPTION_YES,
|
|
|
|
"Clustered, fault-tolerant, memory-based tables",
|
|
|
|
DB_TYPE_NDBCLUSTER,
|
|
|
|
ndbcluster_init,
|
2005-02-17 22:52:40 +01:00
|
|
|
0, /* slot */
|
|
|
|
0, /* savepoint size */
|
|
|
|
ndbcluster_close_connection,
|
|
|
|
NULL, /* savepoint_set */
|
|
|
|
NULL, /* savepoint_rollback */
|
|
|
|
NULL, /* savepoint_release */
|
|
|
|
ndbcluster_commit,
|
|
|
|
ndbcluster_rollback,
|
|
|
|
NULL, /* prepare */
|
2005-02-18 12:53:43 +01:00
|
|
|
NULL, /* recover */
|
|
|
|
NULL, /* commit_by_xid */
|
A fix and a test case for Bug#10760 and complementary cleanups.
The idea of the patch
is that every cursor gets its own lock id for table level locking.
Thus cursors are protected from updates performed within the same
connection. Additionally a list of transient (must be closed at
commit) cursors is maintained and all transient cursors are closed
when necessary. Lastly, this patch adds support for deadlock
timeouts to TLL locking when using cursors.
+ post-review fixes.
include/thr_lock.h:
- add a notion of lock owner to table level locking. When using
cursors, lock owner can not be identified by a thread id any more,
as we must protect cursors from updates issued within the same
connection (thread). So, each cursor has its own lock identifier to
use with table level locking.
- extend return values of thr_lock and thr_multi_lock with
THR_LOCK_TIMEOUT and THR_LOCK_DEADLOCK, since these conditions
are now possible (see comments to thr_lock.c)
mysys/thr_lock.c:
Better support for cursors:
- use THR_LOCK_OWNER * as lock identifier, not pthread_t.
- check and return an error for a trivial deadlock case, when an
update statement is issued to a table locked by a cursor which has
been previously opened in the same connection.
- add support for locking timeouts: with use of cursors, trivial
deadlocks can occur. For now the only remedy is the lock wait timeout,
which is initialized from a new global variable 'table_lock_wait_timeout'
Example of a deadlock (assuming the storage engine does not downgrade
locks):
con1: open cursor for select * from t1;
con2: open cursor for select * from t2;
con1: update t2 set id=id*2; -- blocked
con2: update t1 set id=id*2; -- deadlock
Lock timeouts are active only if a connection is using cursors.
- the check in the wait_for_lock loop has been changed from
data->cond != cond to data->cond != 0. data->cond is zeroed
in every place it's changed.
- added comments
sql/examples/ha_archive.cc:
- extend the handlerton with the info about cursor behaviour at commit.
sql/examples/ha_archive.h:
- ctor moved to .cc to make use of archive handlerton
sql/examples/ha_example.cc:
- add handlerton instance, init handler::ht with it
sql/examples/ha_example.h:
- ctor moved to .cc to make use of ha_example handlerton
sql/examples/ha_tina.cc:
- add handlerton instance, init handler::ht with it
sql/examples/ha_tina.h:
- ctor moved to .cc to make use of CSV handlerton
sql/ha_berkeley.cc:
- init handlerton::flags and handler::ht
sql/ha_berkeley.h:
- ctor moved to .cc to make use of BerkeleyDB handlerton
sql/ha_blackhole.cc:
- add handlerton instance, init handler::ht with it
sql/ha_blackhole.h:
- ctor moved to .cc to make use of blackhole handlerton
sql/ha_federated.cc:
- add handlerton instance, init handler::ht with it
sql/ha_federated.h:
- ctor moved to .cc to make use of federated handlerton
sql/ha_heap.cc:
- add handlerton instance, init handler::ht with it
sql/ha_heap.h:
- ctor moved to .cc to make use of ha_heap handlerton
sql/ha_innodb.cc:
- init handlerton::flags and handler::ht of innobase storage engine
sql/ha_innodb.h:
- ctor moved to .cc to make use of archive handlerton
sql/ha_myisam.cc:
- add handlerton instance, init handler::ht with it
sql/ha_myisam.h:
- ctor moved to .cc to make use of MyISAM handlerton
sql/ha_myisammrg.cc:
- init handler::ht in the ctor
sql/ha_myisammrg.h:
- ctor moved to .cc to make use of MyISAM MERGE handlerton
sql/ha_ndbcluster.cc:
- init handlerton::flags and handler::ht
sql/handler.cc:
- drop support for ISAM storage engine, which was removed from 5.0
- close all "transient" cursors at COMMIT/ROLLBACK. A "transient"
SQL level cursor is a cursor that uses tables that have a transaction-
specific state.
sql/handler.h:
- extend struct handlerton with flags, add handlerton *ht to every
handler instance.
sql/lock.cc:
- extend mysql_lock_tables to send error to the client if
thr_multi_lock returns a timeout or a deadlock error.
sql/mysqld.cc:
- add server option --table_lock_wait_timeout (in seconds)
sql/set_var.cc:
- add new global variable 'table_lock_wait_timeout' to specify
a wait timeout for table-level locks of MySQL (in seconds). The default
timeout is 50 seconds. The timeout is active only if the connection
has open cursors.
sql/sql_class.cc:
- implement Statement_map::close_transient_cursors
- safety suggests that we need an assert ensuring
llock_info->n_cursors is functioning properly, adjust destruction of
the Statement_map to allow such assert in THD::~THD
sql/sql_class.h:
- add support for Cursors registry to Statement map.
sql/sql_prepare.cc:
- maintain a list of cursors that must be closed at commit/rollback.
sql/sql_select.cc:
- extend class Cursor to support specific at-COMMIT/ROLLBACK behavior.
If a cursor uses tables of a storage engine that
invalidates all open tables at COMMIT/ROLLBACK, it must be closed
before COMMIT/ROLLBACK is executed.
sql/sql_select.h:
- add an own lock_id and commit/rollback status flag to class Cursor
tests/mysql_client_test.c:
A test case for Bug#10760 and complementary issues: test a simple
deadlock case too.
mysql-test/var:
New BitKeeper file ``mysql-test/var''
2005-07-19 20:21:12 +02:00
|
|
|
NULL, /* rollback_by_xid */
|
2005-07-20 18:02:36 +02:00
|
|
|
NULL, /* create_cursor_read_view */
|
|
|
|
NULL, /* set_cursor_read_view */
|
|
|
|
NULL, /* close_cursor_read_view */
|
2005-11-07 16:25:06 +01:00
|
|
|
ndbcluster_create_handler, /* Create a new handler */
|
|
|
|
ndbcluster_drop_database, /* Drop a database */
|
|
|
|
ndbcluster_end, /* Panic call */
|
|
|
|
NULL, /* Start Consistent Snapshot */
|
|
|
|
NULL, /* Flush logs */
|
|
|
|
ndbcluster_show_status, /* Show status */
|
A fix and a test case for Bug#10760 and complementary cleanups.
The idea of the patch
is that every cursor gets its own lock id for table level locking.
Thus cursors are protected from updates performed within the same
connection. Additionally a list of transient (must be closed at
commit) cursors is maintained and all transient cursors are closed
when necessary. Lastly, this patch adds support for deadlock
timeouts to TLL locking when using cursors.
+ post-review fixes.
include/thr_lock.h:
- add a notion of lock owner to table level locking. When using
cursors, lock owner can not be identified by a thread id any more,
as we must protect cursors from updates issued within the same
connection (thread). So, each cursor has its own lock identifier to
use with table level locking.
- extend return values of thr_lock and thr_multi_lock with
THR_LOCK_TIMEOUT and THR_LOCK_DEADLOCK, since these conditions
are now possible (see comments to thr_lock.c)
mysys/thr_lock.c:
Better support for cursors:
- use THR_LOCK_OWNER * as lock identifier, not pthread_t.
- check and return an error for a trivial deadlock case, when an
update statement is issued to a table locked by a cursor which has
been previously opened in the same connection.
- add support for locking timeouts: with use of cursors, trivial
deadlocks can occur. For now the only remedy is the lock wait timeout,
which is initialized from a new global variable 'table_lock_wait_timeout'
Example of a deadlock (assuming the storage engine does not downgrade
locks):
con1: open cursor for select * from t1;
con2: open cursor for select * from t2;
con1: update t2 set id=id*2; -- blocked
con2: update t1 set id=id*2; -- deadlock
Lock timeouts are active only if a connection is using cursors.
- the check in the wait_for_lock loop has been changed from
data->cond != cond to data->cond != 0. data->cond is zeroed
in every place it's changed.
- added comments
sql/examples/ha_archive.cc:
- extend the handlerton with the info about cursor behaviour at commit.
sql/examples/ha_archive.h:
- ctor moved to .cc to make use of archive handlerton
sql/examples/ha_example.cc:
- add handlerton instance, init handler::ht with it
sql/examples/ha_example.h:
- ctor moved to .cc to make use of ha_example handlerton
sql/examples/ha_tina.cc:
- add handlerton instance, init handler::ht with it
sql/examples/ha_tina.h:
- ctor moved to .cc to make use of CSV handlerton
sql/ha_berkeley.cc:
- init handlerton::flags and handler::ht
sql/ha_berkeley.h:
- ctor moved to .cc to make use of BerkeleyDB handlerton
sql/ha_blackhole.cc:
- add handlerton instance, init handler::ht with it
sql/ha_blackhole.h:
- ctor moved to .cc to make use of blackhole handlerton
sql/ha_federated.cc:
- add handlerton instance, init handler::ht with it
sql/ha_federated.h:
- ctor moved to .cc to make use of federated handlerton
sql/ha_heap.cc:
- add handlerton instance, init handler::ht with it
sql/ha_heap.h:
- ctor moved to .cc to make use of ha_heap handlerton
sql/ha_innodb.cc:
- init handlerton::flags and handler::ht of innobase storage engine
sql/ha_innodb.h:
- ctor moved to .cc to make use of archive handlerton
sql/ha_myisam.cc:
- add handlerton instance, init handler::ht with it
sql/ha_myisam.h:
- ctor moved to .cc to make use of MyISAM handlerton
sql/ha_myisammrg.cc:
- init handler::ht in the ctor
sql/ha_myisammrg.h:
- ctor moved to .cc to make use of MyISAM MERGE handlerton
sql/ha_ndbcluster.cc:
- init handlerton::flags and handler::ht
sql/handler.cc:
- drop support for ISAM storage engine, which was removed from 5.0
- close all "transient" cursors at COMMIT/ROLLBACK. A "transient"
SQL level cursor is a cursor that uses tables that have a transaction-
specific state.
sql/handler.h:
- extend struct handlerton with flags, add handlerton *ht to every
handler instance.
sql/lock.cc:
- extend mysql_lock_tables to send error to the client if
thr_multi_lock returns a timeout or a deadlock error.
sql/mysqld.cc:
- add server option --table_lock_wait_timeout (in seconds)
sql/set_var.cc:
- add new global variable 'table_lock_wait_timeout' to specify
a wait timeout for table-level locks of MySQL (in seconds). The default
timeout is 50 seconds. The timeout is active only if the connection
has open cursors.
sql/sql_class.cc:
- implement Statement_map::close_transient_cursors
- safety suggests that we need an assert ensuring
llock_info->n_cursors is functioning properly, adjust destruction of
the Statement_map to allow such assert in THD::~THD
sql/sql_class.h:
- add support for Cursors registry to Statement map.
sql/sql_prepare.cc:
- maintain a list of cursors that must be closed at commit/rollback.
sql/sql_select.cc:
- extend class Cursor to support specific at-COMMIT/ROLLBACK behavior.
If a cursor uses tables of a storage engine that
invalidates all open tables at COMMIT/ROLLBACK, it must be closed
before COMMIT/ROLLBACK is executed.
sql/sql_select.h:
- add an own lock_id and commit/rollback status flag to class Cursor
tests/mysql_client_test.c:
A test case for Bug#10760 and complementary issues: test a simple
deadlock case too.
mysql-test/var:
New BitKeeper file ``mysql-test/var''
2005-07-19 20:21:12 +02:00
|
|
|
HTON_NO_FLAGS
|
2005-02-17 22:52:40 +01:00
|
|
|
};
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
static handler *ndbcluster_create_handler(TABLE_SHARE *table)
|
2005-11-07 16:25:06 +01:00
|
|
|
{
|
|
|
|
return new ha_ndbcluster(table);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
#define NDB_HIDDEN_PRIMARY_KEY_LENGTH 8
|
2004-05-13 11:56:45 +02:00
|
|
|
|
2005-04-15 15:52:07 +02:00
|
|
|
#define NDB_FAILED_AUTO_INCREMENT ~(Uint64)0
|
2005-04-15 16:14:08 +02:00
|
|
|
#define NDB_AUTO_INCREMENT_RETRIES 10
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-17 12:45:23 +01:00
|
|
|
#define NDB_INVALID_SCHEMA_OBJECT 241
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
#define ERR_PRINT(err) \
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_PRINT("error", ("%d message: %s", err.code, err.message))
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
#define ERR_RETURN(err) \
|
|
|
|
{ \
|
2004-11-17 10:07:52 +01:00
|
|
|
const NdbError& tmp= err; \
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_PRINT(tmp); \
|
2004-11-17 10:07:52 +01:00
|
|
|
DBUG_RETURN(ndb_to_mysql_error(&tmp)); \
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-09-15 02:33:28 +02:00
|
|
|
#define ERR_BREAK(err, code) \
|
|
|
|
{ \
|
|
|
|
const NdbError& tmp= err; \
|
|
|
|
ERR_PRINT(tmp); \
|
|
|
|
code= ndb_to_mysql_error(&tmp); \
|
|
|
|
break; \
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Typedefs for long names
|
2005-11-06 00:20:37 +01:00
|
|
|
typedef NdbDictionary::Object NDBOBJ;
|
2004-04-15 09:14:14 +02:00
|
|
|
typedef NdbDictionary::Column NDBCOL;
|
2004-09-07 21:53:59 +02:00
|
|
|
typedef NdbDictionary::Table NDBTAB;
|
2004-04-15 09:14:14 +02:00
|
|
|
typedef NdbDictionary::Index NDBINDEX;
|
|
|
|
typedef NdbDictionary::Dictionary NDBDICT;
|
2005-11-06 00:20:37 +01:00
|
|
|
typedef NdbDictionary::Event NDBEVENT;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
static int ndbcluster_inited= 0;
|
|
|
|
static int ndbcluster_util_inited= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-11 13:59:22 +02:00
|
|
|
static Ndb* g_ndb= NULL;
|
2004-09-06 20:30:57 +02:00
|
|
|
static Ndb_cluster_connection* g_ndb_cluster_connection= NULL;
|
2004-05-11 13:59:22 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Handler synchronization
|
|
|
|
pthread_mutex_t ndbcluster_mutex;
|
|
|
|
|
|
|
|
// Table lock handling
|
|
|
|
static HASH ndbcluster_open_tables;
|
|
|
|
|
|
|
|
static byte *ndbcluster_get_key(NDB_SHARE *share,uint *length,
|
|
|
|
my_bool not_used __attribute__((unused)));
|
2005-11-06 00:20:37 +01:00
|
|
|
static NDB_SHARE *get_share(const char *key,
|
|
|
|
bool create_if_not_exists= TRUE,
|
|
|
|
bool have_lock= FALSE);
|
|
|
|
static void free_share(NDB_SHARE **share, bool have_lock= FALSE);
|
|
|
|
static void real_free_share(NDB_SHARE **share);
|
|
|
|
static void ndb_set_fragmentation(NDBTAB &tab, TABLE *table, uint pk_len);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
static int packfrm(const void *data, uint len, const void **pack_data, uint *pack_len);
|
|
|
|
static int unpackfrm(const void **data, uint *len,
|
2005-02-16 14:18:32 +01:00
|
|
|
const void* pack_data);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
static int ndb_get_table_statistics(Ndb*, const char *,
|
2005-02-16 14:18:32 +01:00
|
|
|
struct Ndb_statistics *);
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
#ifndef DBUG_OFF
|
|
|
|
void print_records(TABLE *table, const char *record)
|
|
|
|
{
|
|
|
|
if (_db_on_)
|
|
|
|
{
|
|
|
|
for (uint j= 0; j < table->s->fields; j++)
|
|
|
|
{
|
|
|
|
char buf[40];
|
|
|
|
int pos= 0;
|
|
|
|
Field *field= table->field[j];
|
|
|
|
const byte* field_ptr= field->ptr - table->record[0] + record;
|
|
|
|
int pack_len= field->pack_length();
|
|
|
|
int n= pack_len < 10 ? pack_len : 10;
|
|
|
|
|
|
|
|
for (int i= 0; i < n && pos < 20; i++)
|
|
|
|
{
|
|
|
|
pos+= sprintf(&buf[pos]," %x", (int) (unsigned char) field_ptr[i]);
|
|
|
|
}
|
|
|
|
buf[pos]= 0;
|
|
|
|
DBUG_PRINT("info",("[%u]field_ptr[0->%d]: %s", j, n, buf));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define print_records(a,b)
|
|
|
|
#endif
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
// Util thread variables
|
|
|
|
static pthread_t ndb_util_thread;
|
|
|
|
pthread_mutex_t LOCK_ndb_util_thread;
|
|
|
|
pthread_cond_t COND_ndb_util_thread;
|
2005-10-08 16:39:55 +02:00
|
|
|
pthread_handler_t ndb_util_thread_func(void *arg);
|
2005-02-11 22:33:52 +01:00
|
|
|
ulong ndb_cache_check_time;
|
2004-09-03 13:55:40 +02:00
|
|
|
|
2004-10-20 19:22:58 +02:00
|
|
|
/*
|
|
|
|
Dummy buffer to read zero pack_length fields
|
|
|
|
which are mapped to 1 char
|
|
|
|
*/
|
2004-12-10 16:55:04 +01:00
|
|
|
static uint32 dummy_buf;
|
2004-10-20 19:22:58 +02:00
|
|
|
|
2005-01-13 18:24:19 +01:00
|
|
|
/*
|
|
|
|
Stats that can be retrieved from ndb
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct Ndb_statistics {
|
|
|
|
Uint64 row_count;
|
|
|
|
Uint64 commit_count;
|
|
|
|
Uint64 row_size;
|
|
|
|
Uint64 fragment_memory;
|
|
|
|
};
|
|
|
|
|
2005-01-14 12:32:33 +01:00
|
|
|
/* Status variables shown with 'show status like 'Ndb%' */
|
|
|
|
|
|
|
|
static long ndb_cluster_node_id= 0;
|
|
|
|
static const char * ndb_connected_host= 0;
|
|
|
|
static long ndb_connected_port= 0;
|
|
|
|
static long ndb_number_of_replicas= 0;
|
|
|
|
static long ndb_number_of_storage_nodes= 0;
|
|
|
|
|
|
|
|
static int update_status_variables(Ndb_cluster_connection *c)
|
|
|
|
{
|
|
|
|
ndb_cluster_node_id= c->node_id();
|
|
|
|
ndb_connected_port= c->get_connected_port();
|
|
|
|
ndb_connected_host= c->get_connected_host();
|
|
|
|
ndb_number_of_replicas= 0;
|
|
|
|
ndb_number_of_storage_nodes= c->no_db_nodes();
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct show_var_st ndb_status_variables[]= {
|
|
|
|
{"cluster_node_id", (char*) &ndb_cluster_node_id, SHOW_LONG},
|
|
|
|
{"connected_host", (char*) &ndb_connected_host, SHOW_CHAR_PTR},
|
|
|
|
{"connected_port", (char*) &ndb_connected_port, SHOW_LONG},
|
|
|
|
// {"number_of_replicas", (char*) &ndb_number_of_replicas, SHOW_LONG},
|
|
|
|
{"number_of_storage_nodes",(char*) &ndb_number_of_storage_nodes, SHOW_LONG},
|
|
|
|
{NullS, NullS, SHOW_LONG}
|
|
|
|
};
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/* instantiated in storage/ndb/src/ndbapi/Ndbif.cpp */
|
|
|
|
extern Uint64 g_latest_trans_gci;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Error handling functions
|
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/* Note for merge: old mapping table, moved to storage/ndb/ndberror.c */
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
static int ndb_to_mysql_error(const NdbError *ndberr)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
/* read the mysql mapped error code */
|
|
|
|
int error= ndberr->mysql_code;
|
2004-11-17 09:15:53 +01:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
switch (error)
|
|
|
|
{
|
|
|
|
/* errors for which we do not add warnings, just return mapped error code
|
|
|
|
*/
|
|
|
|
case HA_ERR_NO_SUCH_TABLE:
|
|
|
|
case HA_ERR_KEY_NOT_FOUND:
|
|
|
|
case HA_ERR_FOUND_DUPP_KEY:
|
|
|
|
return error;
|
|
|
|
|
|
|
|
/* Mapping missing, go with the ndb error code*/
|
|
|
|
case -1:
|
|
|
|
error= ndberr->code;
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/* Mapping exists, go with the mapped code */
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/*
|
|
|
|
Push the NDB error message as warning
|
|
|
|
- Used to be able to use SHOW WARNINGS toget more info on what the error is
|
|
|
|
- Used by replication to see if the error was temporary
|
|
|
|
*/
|
|
|
|
if (ndberr->status == NdbError::TemporaryError)
|
2004-11-17 09:15:53 +01:00
|
|
|
push_warning_printf(current_thd, MYSQL_ERROR::WARN_LEVEL_ERROR,
|
2005-11-06 00:20:37 +01:00
|
|
|
ER_GET_TEMPORARY_ERRMSG, ER(ER_GET_TEMPORARY_ERRMSG),
|
|
|
|
ndberr->code, ndberr->message, "NDB");
|
|
|
|
else
|
|
|
|
push_warning_printf(current_thd, MYSQL_ERROR::WARN_LEVEL_ERROR,
|
|
|
|
ER_GET_ERRMSG, ER(ER_GET_ERRMSG),
|
|
|
|
ndberr->code, ndberr->message, "NDB");
|
|
|
|
return error;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
int execute_no_commit_ignore_no_key(ha_ndbcluster *h, NdbTransaction *trans)
|
|
|
|
{
|
|
|
|
int res= trans->execute(NdbTransaction::NoCommit,
|
|
|
|
NdbTransaction::AO_IgnoreError,
|
|
|
|
h->m_force_send);
|
|
|
|
if (res == 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
const NdbError &err= trans->getNdbError();
|
|
|
|
if (err.classification != NdbError::ConstraintViolation &&
|
|
|
|
err.classification != NdbError::NoDataFound)
|
|
|
|
return res;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
return 0;
|
|
|
|
}
|
2004-09-17 16:58:08 +02:00
|
|
|
|
|
|
|
inline
|
2005-01-04 13:47:16 +01:00
|
|
|
int execute_no_commit(ha_ndbcluster *h, NdbTransaction *trans)
|
2004-09-17 16:58:08 +02:00
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#ifdef NOT_USED
|
2005-01-11 17:00:31 +01:00
|
|
|
int m_batch_execute= 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
if (m_batch_execute)
|
2004-09-17 16:58:08 +02:00
|
|
|
return 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#endif
|
2005-11-06 00:20:37 +01:00
|
|
|
return h->m_ignore_no_key ?
|
|
|
|
execute_no_commit_ignore_no_key(h,trans) :
|
|
|
|
trans->execute(NdbTransaction::NoCommit,
|
|
|
|
NdbTransaction::AbortOnError,
|
|
|
|
h->m_force_send);
|
2004-09-28 21:11:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
inline
|
2005-01-04 13:47:16 +01:00
|
|
|
int execute_commit(ha_ndbcluster *h, NdbTransaction *trans)
|
2004-09-28 21:11:50 +02:00
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#ifdef NOT_USED
|
2005-01-11 17:00:31 +01:00
|
|
|
int m_batch_execute= 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
if (m_batch_execute)
|
2004-09-28 21:11:50 +02:00
|
|
|
return 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#endif
|
2005-01-05 17:59:24 +01:00
|
|
|
return trans->execute(NdbTransaction::Commit,
|
2005-02-16 14:18:32 +01:00
|
|
|
NdbTransaction::AbortOnError,
|
|
|
|
h->m_force_send);
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
inline
|
2005-01-04 13:47:16 +01:00
|
|
|
int execute_commit(THD *thd, NdbTransaction *trans)
|
2004-11-17 09:15:53 +01:00
|
|
|
{
|
|
|
|
#ifdef NOT_USED
|
2005-01-11 17:00:31 +01:00
|
|
|
int m_batch_execute= 0;
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_batch_execute)
|
|
|
|
return 0;
|
|
|
|
#endif
|
2005-01-05 17:59:24 +01:00
|
|
|
return trans->execute(NdbTransaction::Commit,
|
2005-02-16 14:18:32 +01:00
|
|
|
NdbTransaction::AbortOnError,
|
|
|
|
thd->variables.ndb_force_send);
|
2004-09-28 21:11:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
inline
|
2005-01-04 13:47:16 +01:00
|
|
|
int execute_no_commit_ie(ha_ndbcluster *h, NdbTransaction *trans)
|
2004-09-28 21:11:50 +02:00
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#ifdef NOT_USED
|
2005-01-11 17:00:31 +01:00
|
|
|
int m_batch_execute= 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
if (m_batch_execute)
|
2004-09-28 21:11:50 +02:00
|
|
|
return 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#endif
|
2005-01-05 17:59:24 +01:00
|
|
|
return trans->execute(NdbTransaction::NoCommit,
|
2005-02-16 14:18:32 +01:00
|
|
|
NdbTransaction::AO_IgnoreError,
|
|
|
|
h->m_force_send);
|
2004-09-17 16:58:08 +02:00
|
|
|
}
|
|
|
|
|
2004-12-17 21:13:22 +01:00
|
|
|
/*
|
|
|
|
Place holder for ha_ndbcluster thread specific data
|
|
|
|
*/
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Thd_ndb::Thd_ndb()
|
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
ndb= new Ndb(g_ndb_cluster_connection, "");
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
lock_count= 0;
|
|
|
|
count= 0;
|
2005-02-17 22:52:40 +01:00
|
|
|
all= NULL;
|
|
|
|
stmt= NULL;
|
2004-09-14 18:17:01 +02:00
|
|
|
error= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
Thd_ndb::~Thd_ndb()
|
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
if (ndb)
|
2005-11-17 12:45:23 +01:00
|
|
|
{
|
|
|
|
#ifndef DBUG_OFF
|
2006-01-05 23:47:49 +01:00
|
|
|
Ndb::Free_list_usage tmp;
|
|
|
|
tmp.m_name= 0;
|
2005-11-17 12:45:23 +01:00
|
|
|
while (ndb->get_free_list_usage(&tmp))
|
|
|
|
{
|
|
|
|
uint leaked= (uint) tmp.m_created - tmp.m_free;
|
|
|
|
if (leaked)
|
|
|
|
fprintf(stderr, "NDB: Found %u %s%s that %s not been released\n",
|
|
|
|
leaked, tmp.m_name,
|
|
|
|
(leaked == 1)?"":"'s",
|
|
|
|
(leaked == 1)?"has":"have");
|
|
|
|
}
|
|
|
|
#endif
|
2004-09-14 14:47:34 +02:00
|
|
|
delete ndb;
|
2006-01-05 23:47:49 +01:00
|
|
|
ndb= NULL;
|
2005-11-17 12:45:23 +01:00
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
changed_tables.empty();
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
inline
|
|
|
|
Thd_ndb *
|
|
|
|
get_thd_ndb(THD *thd) { return (Thd_ndb *) thd->ha_data[ndbcluster_hton.slot]; }
|
|
|
|
|
|
|
|
inline
|
|
|
|
void
|
|
|
|
set_thd_ndb(THD *thd, Thd_ndb *thd_ndb) { thd->ha_data[ndbcluster_hton.slot]= thd_ndb; }
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
inline
|
|
|
|
Ndb *ha_ndbcluster::get_ndb()
|
|
|
|
{
|
2005-02-18 12:53:43 +01:00
|
|
|
return get_thd_ndb(current_thd)->ndb;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* manage uncommitted insert/deletes during transactio to get records correct
|
|
|
|
*/
|
|
|
|
|
2005-03-22 17:42:08 +01:00
|
|
|
struct Ndb_local_table_statistics {
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
int no_uncommitted_rows_count;
|
2004-09-14 14:47:34 +02:00
|
|
|
ulong last_count;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
ha_rows records;
|
|
|
|
};
|
|
|
|
|
2004-10-04 01:20:05 +02:00
|
|
|
void ha_ndbcluster::set_rec_per_key()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::get_status_const");
|
2005-01-06 12:00:13 +01:00
|
|
|
for (uint i=0 ; i < table->s->keys ; i++)
|
2004-10-04 01:20:05 +02:00
|
|
|
{
|
|
|
|
table->key_info[i].rec_per_key[table->key_info[i].key_parts-1]= 1;
|
|
|
|
}
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
void ha_ndbcluster::records_update()
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::records_update");
|
2005-03-22 17:42:08 +01:00
|
|
|
struct Ndb_local_table_statistics *info=
|
|
|
|
(struct Ndb_local_table_statistics *)m_table_info;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_PRINT("info", ("id=%d, no_uncommitted_rows_count=%d",
|
2005-02-16 14:18:32 +01:00
|
|
|
((const NDBTAB *)m_table)->getTableId(),
|
|
|
|
info->no_uncommitted_rows_count));
|
2004-09-20 17:52:19 +02:00
|
|
|
// if (info->records == ~(ha_rows)0)
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2005-01-13 18:24:19 +01:00
|
|
|
struct Ndb_statistics stat;
|
2005-07-04 02:42:33 +02:00
|
|
|
if (ndb_get_table_statistics(ndb, m_tabname, &stat) == 0){
|
2005-01-13 18:24:19 +01:00
|
|
|
mean_rec_length= stat.row_size;
|
|
|
|
data_file_length= stat.fragment_memory;
|
|
|
|
info->records= stat.row_count;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
}
|
2004-09-14 18:17:01 +02:00
|
|
|
{
|
|
|
|
THD *thd= current_thd;
|
2005-02-17 22:52:40 +01:00
|
|
|
if (get_thd_ndb(thd)->error)
|
2004-09-14 18:17:01 +02:00
|
|
|
info->no_uncommitted_rows_count= 0;
|
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
records= info->records+ info->no_uncommitted_rows_count;
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2004-09-14 17:51:57 +02:00
|
|
|
void ha_ndbcluster::no_uncommitted_rows_execute_failure()
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
2004-09-14 17:51:57 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_execute_failure");
|
2005-02-18 12:53:43 +01:00
|
|
|
get_thd_ndb(current_thd)->error= 1;
|
2004-09-14 17:51:57 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
void ha_ndbcluster::no_uncommitted_rows_init(THD *thd)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_init");
|
2005-03-22 17:42:08 +01:00
|
|
|
struct Ndb_local_table_statistics *info=
|
|
|
|
(struct Ndb_local_table_statistics *)m_table_info;
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
2004-09-14 14:47:34 +02:00
|
|
|
if (info->last_count != thd_ndb->count)
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
{
|
2004-12-07 15:15:49 +01:00
|
|
|
info->last_count= thd_ndb->count;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
info->no_uncommitted_rows_count= 0;
|
|
|
|
info->records= ~(ha_rows)0;
|
|
|
|
DBUG_PRINT("info", ("id=%d, no_uncommitted_rows_count=%d",
|
2005-02-16 14:18:32 +01:00
|
|
|
((const NDBTAB *)m_table)->getTableId(),
|
|
|
|
info->no_uncommitted_rows_count));
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ha_ndbcluster::no_uncommitted_rows_update(int c)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_update");
|
2005-03-22 17:42:08 +01:00
|
|
|
struct Ndb_local_table_statistics *info=
|
|
|
|
(struct Ndb_local_table_statistics *)m_table_info;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
info->no_uncommitted_rows_count+= c;
|
|
|
|
DBUG_PRINT("info", ("id=%d, no_uncommitted_rows_count=%d",
|
2005-02-16 14:18:32 +01:00
|
|
|
((const NDBTAB *)m_table)->getTableId(),
|
|
|
|
info->no_uncommitted_rows_count));
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ha_ndbcluster::no_uncommitted_rows_reset(THD *thd)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_reset");
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
|
|
|
thd_ndb->count++;
|
|
|
|
thd_ndb->error= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Take care of the error that occured in NDB
|
2005-02-17 22:52:40 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
RETURN
|
2005-02-16 14:18:32 +01:00
|
|
|
0 No error
|
2004-04-15 09:14:14 +02:00
|
|
|
# The mapped error code
|
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
void
|
|
|
|
ha_ndbcluster::invalidate_dictionary_cache(TABLE *table, Ndb *ndb,
|
|
|
|
const char *tabname, bool global)
|
2005-03-07 10:21:38 +01:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2005-04-28 10:46:39 +02:00
|
|
|
DBUG_ENTER("invalidate_dictionary_cache");
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("invalidating %s", tabname));
|
2005-04-28 10:46:39 +02:00
|
|
|
|
2005-04-27 18:17:41 +02:00
|
|
|
if (global)
|
2005-04-28 09:39:29 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
const NDBTAB *tab= dict->getTable(tabname);
|
2005-04-28 10:46:39 +02:00
|
|
|
if (!tab)
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
if (tab->getObjectStatus() == NdbDictionary::Object::Invalid)
|
2005-04-28 09:39:29 +02:00
|
|
|
{
|
|
|
|
// Global cache has already been invalidated
|
2005-11-06 00:20:37 +01:00
|
|
|
dict->removeCachedTable(tabname);
|
2005-04-28 09:39:29 +02:00
|
|
|
global= FALSE;
|
|
|
|
}
|
|
|
|
else
|
2005-11-06 00:20:37 +01:00
|
|
|
dict->invalidateTable(tabname);
|
2005-04-28 09:39:29 +02:00
|
|
|
}
|
2005-04-27 18:17:41 +02:00
|
|
|
else
|
2005-11-06 00:20:37 +01:00
|
|
|
dict->removeCachedTable(tabname);
|
2005-03-09 23:38:11 +01:00
|
|
|
table->s->version=0L; /* Free when thread is ready */
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ha_ndbcluster::invalidate_dictionary_cache(bool global)
|
|
|
|
{
|
|
|
|
NDBDICT *dict= get_ndb()->getDictionary();
|
|
|
|
invalidate_dictionary_cache(table, get_ndb(), m_tabname, global);
|
2005-03-07 10:21:38 +01:00
|
|
|
/* Invalidate indexes */
|
2005-03-09 23:38:11 +01:00
|
|
|
for (uint i= 0; i < table->s->keys; i++)
|
2005-03-07 10:21:38 +01:00
|
|
|
{
|
|
|
|
NDBINDEX *index = (NDBINDEX *) m_index[i].index;
|
|
|
|
NDBINDEX *unique_index = (NDBINDEX *) m_index[i].unique_index;
|
|
|
|
NDB_INDEX_TYPE idx_type= m_index[i].type;
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (idx_type) {
|
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
|
|
|
case ORDERED_INDEX:
|
2005-04-27 18:17:41 +02:00
|
|
|
if (global)
|
|
|
|
dict->invalidateIndex(index->getName(), m_tabname);
|
|
|
|
else
|
|
|
|
dict->removeCachedIndex(index->getName(), m_tabname);
|
2005-03-10 00:10:10 +01:00
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case UNIQUE_ORDERED_INDEX:
|
2005-04-27 18:17:41 +02:00
|
|
|
if (global)
|
|
|
|
dict->invalidateIndex(index->getName(), m_tabname);
|
|
|
|
else
|
|
|
|
dict->removeCachedIndex(index->getName(), m_tabname);
|
2005-07-04 02:42:33 +02:00
|
|
|
case UNIQUE_INDEX:
|
2005-04-27 18:17:41 +02:00
|
|
|
if (global)
|
|
|
|
dict->invalidateIndex(unique_index->getName(), m_tabname);
|
|
|
|
else
|
|
|
|
dict->removeCachedIndex(unique_index->getName(), m_tabname);
|
2005-03-07 10:21:38 +01:00
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case PRIMARY_KEY_INDEX:
|
|
|
|
case UNDEFINED_INDEX:
|
2005-03-07 10:21:38 +01:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2005-01-04 13:47:16 +01:00
|
|
|
int ha_ndbcluster::ndb_err(NdbTransaction *trans)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-08-31 12:07:48 +02:00
|
|
|
int res;
|
2005-04-04 12:26:05 +02:00
|
|
|
NdbError err= trans->getNdbError();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndb_err");
|
|
|
|
|
|
|
|
ERR_PRINT(err);
|
|
|
|
switch (err.classification) {
|
|
|
|
case NdbError::SchemaError:
|
2005-04-27 18:17:41 +02:00
|
|
|
invalidate_dictionary_cache(TRUE);
|
2005-04-04 12:26:05 +02:00
|
|
|
|
|
|
|
if (err.code==284)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Check if the table is _really_ gone or if the table has
|
|
|
|
been alterend and thus changed table id
|
|
|
|
*/
|
|
|
|
NDBDICT *dict= get_ndb()->getDictionary();
|
|
|
|
DBUG_PRINT("info", ("Check if table %s is really gone", m_tabname));
|
|
|
|
if (!(dict->getTable(m_tabname)))
|
|
|
|
{
|
|
|
|
err= dict->getNdbError();
|
|
|
|
DBUG_PRINT("info", ("Table not found, error: %d", err.code));
|
2005-11-07 12:19:28 +01:00
|
|
|
if (err.code != 709 && err.code != 723)
|
2005-04-04 12:26:05 +02:00
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
2005-04-07 20:17:37 +02:00
|
|
|
DBUG_PRINT("info", ("Table exists but must have changed"));
|
2005-04-04 12:26:05 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-08-31 12:07:48 +02:00
|
|
|
res= ndb_to_mysql_error(&err);
|
|
|
|
DBUG_PRINT("info", ("transformed ndbcluster error %d to mysql error %d",
|
2005-02-16 14:18:32 +01:00
|
|
|
err.code, res));
|
2004-08-31 12:07:48 +02:00
|
|
|
if (res == HA_ERR_FOUND_DUPP_KEY)
|
2005-06-17 14:45:54 +02:00
|
|
|
{
|
|
|
|
if (m_rows_to_insert == 1)
|
2005-06-20 18:02:47 +02:00
|
|
|
m_dupkey= table->s->primary_key;
|
2005-06-17 14:45:54 +02:00
|
|
|
else
|
2005-06-27 15:46:41 +02:00
|
|
|
{
|
|
|
|
/* We are batching inserts, offending key is not available */
|
2005-06-17 14:45:54 +02:00
|
|
|
m_dupkey= (uint) -1;
|
2005-06-27 15:46:41 +02:00
|
|
|
}
|
2005-06-17 14:45:54 +02:00
|
|
|
}
|
2004-08-31 12:07:48 +02:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-05-11 13:59:22 +02:00
|
|
|
/*
|
2004-05-13 11:56:45 +02:00
|
|
|
Override the default get_error_message in order to add the
|
2004-05-11 13:59:22 +02:00
|
|
|
error message of NDB
|
|
|
|
*/
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
bool ha_ndbcluster::get_error_message(int error,
|
2005-02-16 14:18:32 +01:00
|
|
|
String *buf)
|
2004-05-11 13:59:22 +02:00
|
|
|
{
|
2004-05-13 11:56:45 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::get_error_message");
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("enter", ("error: %d", error));
|
2004-05-11 13:59:22 +02:00
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2004-09-15 14:44:21 +02:00
|
|
|
if (!ndb)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(FALSE);
|
2004-05-13 11:56:45 +02:00
|
|
|
|
2004-09-15 14:44:21 +02:00
|
|
|
const NdbError err= ndb->getNdbError(error);
|
2004-05-24 12:35:39 +02:00
|
|
|
bool temporary= err.status==NdbError::TemporaryError;
|
|
|
|
buf->set(err.message, strlen(err.message), &my_charset_bin);
|
|
|
|
DBUG_PRINT("exit", ("message: %s, temporary: %d", buf->ptr(), temporary));
|
|
|
|
DBUG_RETURN(temporary);
|
2004-05-11 13:59:22 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-06-20 07:53:53 +02:00
|
|
|
#ifndef DBUG_OFF
|
2004-07-22 12:38:09 +02:00
|
|
|
/*
|
|
|
|
Check if type is supported by NDB.
|
|
|
|
*/
|
|
|
|
|
2005-06-20 07:53:53 +02:00
|
|
|
static bool ndb_supported_type(enum_field_types type)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
|
|
|
switch (type) {
|
2004-07-23 13:21:39 +02:00
|
|
|
case MYSQL_TYPE_TINY:
|
|
|
|
case MYSQL_TYPE_SHORT:
|
|
|
|
case MYSQL_TYPE_LONG:
|
|
|
|
case MYSQL_TYPE_INT24:
|
|
|
|
case MYSQL_TYPE_LONGLONG:
|
|
|
|
case MYSQL_TYPE_FLOAT:
|
|
|
|
case MYSQL_TYPE_DOUBLE:
|
2005-02-16 21:19:42 +01:00
|
|
|
case MYSQL_TYPE_DECIMAL:
|
|
|
|
case MYSQL_TYPE_NEWDECIMAL:
|
2004-07-23 13:21:39 +02:00
|
|
|
case MYSQL_TYPE_TIMESTAMP:
|
|
|
|
case MYSQL_TYPE_DATETIME:
|
|
|
|
case MYSQL_TYPE_DATE:
|
|
|
|
case MYSQL_TYPE_NEWDATE:
|
|
|
|
case MYSQL_TYPE_TIME:
|
|
|
|
case MYSQL_TYPE_YEAR:
|
|
|
|
case MYSQL_TYPE_STRING:
|
|
|
|
case MYSQL_TYPE_VAR_STRING:
|
2005-01-07 11:55:20 +01:00
|
|
|
case MYSQL_TYPE_VARCHAR:
|
2004-07-23 13:21:39 +02:00
|
|
|
case MYSQL_TYPE_TINY_BLOB:
|
|
|
|
case MYSQL_TYPE_BLOB:
|
|
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
|
|
|
case MYSQL_TYPE_LONG_BLOB:
|
|
|
|
case MYSQL_TYPE_ENUM:
|
|
|
|
case MYSQL_TYPE_SET:
|
2004-12-23 15:28:41 +01:00
|
|
|
case MYSQL_TYPE_BIT:
|
2005-10-25 08:20:39 +02:00
|
|
|
case MYSQL_TYPE_GEOMETRY:
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
case MYSQL_TYPE_NULL:
|
2004-07-23 13:21:39 +02:00
|
|
|
break;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
2005-06-20 07:53:53 +02:00
|
|
|
#endif /* !DBUG_OFF */
|
2004-07-22 12:38:09 +02:00
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Instruct NDB to set the value of the hidden primary key
|
|
|
|
*/
|
|
|
|
|
|
|
|
bool ha_ndbcluster::set_hidden_key(NdbOperation *ndb_op,
|
2005-02-16 14:18:32 +01:00
|
|
|
uint fieldnr, const byte *field_ptr)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("set_hidden_key");
|
2005-11-07 12:19:28 +01:00
|
|
|
DBUG_RETURN(ndb_op->equal(fieldnr, (char*)field_ptr) != 0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Instruct NDB to set the value of one primary key attribute
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::set_ndb_key(NdbOperation *ndb_op, Field *field,
|
|
|
|
uint fieldnr, const byte *field_ptr)
|
|
|
|
{
|
|
|
|
uint32 pack_len= field->pack_length();
|
|
|
|
DBUG_ENTER("set_ndb_key");
|
|
|
|
DBUG_PRINT("enter", ("%d: %s, ndb_type: %u, len=%d",
|
|
|
|
fieldnr, field->field_name, field->type(),
|
|
|
|
pack_len));
|
|
|
|
DBUG_DUMP("key", (char*)field_ptr, pack_len);
|
|
|
|
|
2005-06-20 07:53:53 +02:00
|
|
|
DBUG_ASSERT(ndb_supported_type(field->type()));
|
|
|
|
DBUG_ASSERT(! (field->flags & BLOB_FLAG));
|
|
|
|
// Common implementation for most field types
|
|
|
|
DBUG_RETURN(ndb_op->equal(fieldnr, (char*) field_ptr, pack_len) != 0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Instruct NDB to set the value of one attribute
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::set_ndb_value(NdbOperation *ndb_op, Field *field,
|
2005-11-06 00:20:37 +01:00
|
|
|
uint fieldnr, int row_offset,
|
|
|
|
bool *set_blob_value)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
const byte* field_ptr= field->ptr + row_offset;
|
|
|
|
uint32 pack_len= field->pack_length();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("set_ndb_value");
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("enter", ("%d: %s type: %u len=%d is_null=%s",
|
2004-04-15 09:14:14 +02:00
|
|
|
fieldnr, field->field_name, field->type(),
|
2005-11-06 00:20:37 +01:00
|
|
|
pack_len, field->is_null(row_offset) ? "Y" : "N"));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_DUMP("value", (char*) field_ptr, pack_len);
|
2004-07-22 12:38:09 +02:00
|
|
|
|
2005-06-20 07:53:53 +02:00
|
|
|
DBUG_ASSERT(ndb_supported_type(field->type()));
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-10-20 19:22:58 +02:00
|
|
|
// ndb currently does not support size 0
|
2004-12-10 16:55:04 +01:00
|
|
|
uint32 empty_field;
|
2004-10-20 19:22:58 +02:00
|
|
|
if (pack_len == 0)
|
|
|
|
{
|
2004-12-10 16:55:04 +01:00
|
|
|
pack_len= sizeof(empty_field);
|
|
|
|
field_ptr= (byte *)&empty_field;
|
2005-11-06 00:20:37 +01:00
|
|
|
if (field->is_null(row_offset))
|
2005-02-16 14:18:32 +01:00
|
|
|
empty_field= 0;
|
2004-12-10 16:55:04 +01:00
|
|
|
else
|
2005-02-16 14:18:32 +01:00
|
|
|
empty_field= 1;
|
2004-10-20 19:22:58 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
if (! (field->flags & BLOB_FLAG))
|
|
|
|
{
|
2004-12-23 15:28:41 +01:00
|
|
|
if (field->type() != MYSQL_TYPE_BIT)
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (field->is_null(row_offset))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("field is NULL"));
|
2005-02-16 14:18:32 +01:00
|
|
|
// Set value to NULL
|
2005-11-07 12:19:28 +01:00
|
|
|
DBUG_RETURN((ndb_op->setValue(fieldnr, (char*)NULL) != 0));
|
2005-11-06 00:20:37 +01:00
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
// Common implementation for most field types
|
2005-11-07 12:19:28 +01:00
|
|
|
DBUG_RETURN(ndb_op->setValue(fieldnr, (char*)field_ptr) != 0);
|
2004-12-23 15:28:41 +01:00
|
|
|
}
|
|
|
|
else // if (field->type() == MYSQL_TYPE_BIT)
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
longlong bits= field->val_int();
|
2004-12-23 15:28:41 +01:00
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
// Round up bit field length to nearest word boundry
|
|
|
|
pack_len= ((pack_len + 3) >> 2) << 2;
|
2004-12-23 15:28:41 +01:00
|
|
|
DBUG_ASSERT(pack_len <= 8);
|
2005-11-06 00:20:37 +01:00
|
|
|
if (field->is_null(row_offset))
|
2004-12-23 15:28:41 +01:00
|
|
|
// Set value to NULL
|
2005-11-07 12:19:28 +01:00
|
|
|
DBUG_RETURN((ndb_op->setValue(fieldnr, (char*)NULL) != 0));
|
2004-12-23 15:28:41 +01:00
|
|
|
DBUG_PRINT("info", ("bit field"));
|
2005-01-03 14:29:27 +01:00
|
|
|
DBUG_DUMP("value", (char*)&bits, pack_len);
|
2005-01-20 15:50:56 +01:00
|
|
|
#ifdef WORDS_BIGENDIAN
|
2005-02-16 14:18:32 +01:00
|
|
|
if (pack_len < 5)
|
|
|
|
{
|
2005-11-07 12:19:28 +01:00
|
|
|
DBUG_RETURN(ndb_op->setValue(fieldnr, ((char*)&bits)+4) != 0);
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-01-20 15:50:56 +01:00
|
|
|
#endif
|
2005-11-07 12:19:28 +01:00
|
|
|
DBUG_RETURN(ndb_op->setValue(fieldnr, (char*)&bits) != 0);
|
2004-12-23 15:28:41 +01:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
// Blob type
|
2004-07-23 15:46:56 +02:00
|
|
|
NdbBlob *ndb_blob= ndb_op->getBlobHandle(fieldnr);
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_blob != NULL)
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (field->is_null(row_offset))
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(ndb_blob->setNull() != 0);
|
|
|
|
|
|
|
|
Field_blob *field_blob= (Field_blob*)field;
|
|
|
|
|
|
|
|
// Get length and pointer to data
|
|
|
|
uint32 blob_len= field_blob->get_length(field_ptr);
|
|
|
|
char* blob_ptr= NULL;
|
|
|
|
field_blob->get_ptr(&blob_ptr);
|
|
|
|
|
2004-09-02 12:42:03 +02:00
|
|
|
// Looks like NULL ptr signals length 0 blob
|
|
|
|
if (blob_ptr == NULL) {
|
|
|
|
DBUG_ASSERT(blob_len == 0);
|
2004-09-15 17:31:15 +02:00
|
|
|
blob_ptr= (char*)"";
|
2004-09-02 12:42:03 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
|
2005-07-19 22:04:43 +02:00
|
|
|
DBUG_PRINT("value", ("set blob ptr=%p len=%u",
|
|
|
|
blob_ptr, blob_len));
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_DUMP("value", (char*)blob_ptr, min(blob_len, 26));
|
|
|
|
|
2004-10-01 13:16:49 +02:00
|
|
|
if (set_blob_value)
|
2005-02-16 14:18:32 +01:00
|
|
|
*set_blob_value= TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
// No callback needed to write value
|
|
|
|
DBUG_RETURN(ndb_blob->setValue(blob_ptr, blob_len) != 0);
|
|
|
|
}
|
|
|
|
DBUG_RETURN(1);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Callback to read all blob values.
|
|
|
|
- not done in unpack_record because unpack_record is valid
|
|
|
|
after execute(Commit) but reading blobs is not
|
|
|
|
- may only generate read operations; they have to be executed
|
|
|
|
somewhere before the data is available
|
|
|
|
- due to single buffer for all blobs, we let the last blob
|
|
|
|
process all blobs (last so that all are active)
|
|
|
|
- null bit is still set in unpack_record
|
|
|
|
- TODO allocate blob part aligned buffers
|
|
|
|
*/
|
|
|
|
|
2004-07-29 10:44:53 +02:00
|
|
|
NdbBlob::ActiveHook g_get_ndb_blobs_value;
|
2004-07-22 12:38:09 +02:00
|
|
|
|
2004-07-29 10:44:53 +02:00
|
|
|
int g_get_ndb_blobs_value(NdbBlob *ndb_blob, void *arg)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2004-07-29 10:44:53 +02:00
|
|
|
DBUG_ENTER("g_get_ndb_blobs_value");
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_blob->blobsNextBlob() != NULL)
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
ha_ndbcluster *ha= (ha_ndbcluster *)arg;
|
|
|
|
DBUG_RETURN(ha->get_ndb_blobs_value(ndb_blob));
|
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::get_ndb_blobs_value(NdbBlob *last_ndb_blob)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("get_ndb_blobs_value");
|
|
|
|
|
|
|
|
// Field has no field number so cannot use TABLE blob_field
|
|
|
|
// Loop twice, first only counting total buffer size
|
|
|
|
for (int loop= 0; loop <= 1; loop++)
|
|
|
|
{
|
|
|
|
uint32 offset= 0;
|
2005-01-06 12:00:13 +01:00
|
|
|
for (uint i= 0; i < table->s->fields; i++)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
NdbValue value= m_value[i];
|
|
|
|
if (value.ptr != NULL && (field->flags & BLOB_FLAG))
|
|
|
|
{
|
|
|
|
Field_blob *field_blob= (Field_blob *)field;
|
|
|
|
NdbBlob *ndb_blob= value.blob;
|
|
|
|
Uint64 blob_len= 0;
|
|
|
|
if (ndb_blob->getLength(blob_len) != 0)
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
// Align to Uint64
|
|
|
|
uint32 blob_size= blob_len;
|
|
|
|
if (blob_size % 8 != 0)
|
|
|
|
blob_size+= 8 - blob_size % 8;
|
|
|
|
if (loop == 1)
|
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
char *buf= m_blobs_buffer + offset;
|
2004-07-22 12:38:09 +02:00
|
|
|
uint32 len= 0xffffffff; // Max uint32
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("value", ("read blob ptr=%lx len=%u",
|
|
|
|
buf, (uint) blob_len));
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_blob->readData(buf, len) != 0)
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
DBUG_ASSERT(len == blob_len);
|
|
|
|
field_blob->set_ptr(len, buf);
|
|
|
|
}
|
|
|
|
offset+= blob_size;
|
|
|
|
}
|
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
if (loop == 0 && offset > m_blobs_buffer_size)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
my_free(m_blobs_buffer, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
m_blobs_buffer_size= 0;
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("value", ("allocate blobs buffer size %u", offset));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_buffer= my_malloc(offset, MYF(MY_WME));
|
|
|
|
if (m_blobs_buffer == NULL)
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(-1);
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_buffer_size= offset;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Instruct NDB to fetch one field
|
2004-07-22 12:38:09 +02:00
|
|
|
- data is read directly into buffer provided by field
|
|
|
|
if field is NULL, data is read into memory provided by NDBAPI
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
int ha_ndbcluster::get_ndb_value(NdbOperation *ndb_op, Field *field,
|
2004-09-20 12:40:53 +02:00
|
|
|
uint fieldnr, byte* buf)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("get_ndb_value");
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("enter", ("fieldnr: %d flags: %o", fieldnr,
|
|
|
|
(int)(field != NULL ? field->flags : 0)));
|
|
|
|
|
|
|
|
if (field != NULL)
|
|
|
|
{
|
2005-06-20 07:53:53 +02:00
|
|
|
DBUG_ASSERT(buf);
|
|
|
|
DBUG_ASSERT(ndb_supported_type(field->type()));
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_ASSERT(field->ptr != NULL);
|
|
|
|
if (! (field->flags & BLOB_FLAG))
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2004-12-23 15:28:41 +01:00
|
|
|
if (field->type() != MYSQL_TYPE_BIT)
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
byte *field_buf;
|
|
|
|
if (field->pack_length() != 0)
|
|
|
|
field_buf= buf + (field->ptr - table->record[0]);
|
|
|
|
else
|
|
|
|
field_buf= (byte *)&dummy_buf;
|
|
|
|
m_value[fieldnr].rec= ndb_op->getValue(fieldnr,
|
|
|
|
field_buf);
|
|
|
|
}
|
2004-12-23 15:28:41 +01:00
|
|
|
else // if (field->type() == MYSQL_TYPE_BIT)
|
|
|
|
{
|
|
|
|
m_value[fieldnr].rec= ndb_op->getValue(fieldnr);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(m_value[fieldnr].rec == NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Blob type
|
|
|
|
NdbBlob *ndb_blob= ndb_op->getBlobHandle(fieldnr);
|
|
|
|
m_value[fieldnr].blob= ndb_blob;
|
|
|
|
if (ndb_blob != NULL)
|
|
|
|
{
|
|
|
|
// Set callback
|
|
|
|
void *arg= (void *)this;
|
2004-07-29 10:44:53 +02:00
|
|
|
DBUG_RETURN(ndb_blob->setActiveHook(g_get_ndb_blobs_value, arg) != 0);
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Used for hidden key only
|
|
|
|
m_value[fieldnr].rec= ndb_op->getValue(fieldnr, NULL);
|
|
|
|
DBUG_RETURN(m_value[fieldnr].rec == NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Check if any set or get of blob value in current query.
|
|
|
|
*/
|
2005-04-28 14:45:27 +02:00
|
|
|
bool ha_ndbcluster::uses_blob_value()
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->blob_fields == 0)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
uint no_fields= table->s->fields;
|
2004-07-22 12:38:09 +02:00
|
|
|
int i;
|
|
|
|
// They always put blobs at the end..
|
|
|
|
for (i= no_fields - 1; i >= 0; i--)
|
|
|
|
{
|
2005-04-28 14:45:27 +02:00
|
|
|
if ((m_write_op && ha_get_bit_in_write_set(i+1)) ||
|
|
|
|
(!m_write_op && ha_get_bit_in_read_set(i+1)))
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get metadata for this table from NDB
|
|
|
|
|
|
|
|
IMPLEMENTATION
|
|
|
|
- check that frm-file on disk is equal to frm-file
|
|
|
|
of table accessed in NDB
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
|
|
|
|
RETURN
|
|
|
|
0 ok
|
|
|
|
-2 Meta data has changed; Re-read data and try again
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
static int cmp_frm(const NDBTAB *ndbtab, const void *pack_data,
|
|
|
|
uint pack_length)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("cmp_frm");
|
|
|
|
/*
|
|
|
|
Compare FrmData in NDB with frm file from disk.
|
|
|
|
*/
|
|
|
|
if ((pack_length != ndbtab->getFrmLength()) ||
|
|
|
|
(memcmp(pack_data, ndbtab->getFrmData(), pack_length)))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::get_metadata(const char *path)
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
const NDBTAB *tab;
|
|
|
|
int error;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool invalidating_ndb_table= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("get_metadata");
|
|
|
|
DBUG_PRINT("enter", ("m_tabname: %s, path: %s", m_tabname, path));
|
|
|
|
|
2004-09-24 18:58:25 +02:00
|
|
|
do {
|
|
|
|
const void *data, *pack_data;
|
|
|
|
uint length, pack_length;
|
|
|
|
|
|
|
|
if (!(tab= dict->getTable(m_tabname)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
2005-04-28 09:39:29 +02:00
|
|
|
// Check if thread has stale local cache
|
2005-04-27 18:17:41 +02:00
|
|
|
if (tab->getObjectStatus() == NdbDictionary::Object::Invalid)
|
|
|
|
{
|
|
|
|
invalidate_dictionary_cache(FALSE);
|
|
|
|
if (!(tab= dict->getTable(m_tabname)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
DBUG_PRINT("info", ("Table schema version: %d", tab->getObjectVersion()));
|
|
|
|
}
|
2004-09-24 18:58:25 +02:00
|
|
|
/*
|
|
|
|
Compare FrmData in NDB with frm file from disk.
|
|
|
|
*/
|
|
|
|
error= 0;
|
|
|
|
if (readfrm(path, &data, &length) ||
|
2005-02-16 14:18:32 +01:00
|
|
|
packfrm(data, length, &pack_data, &pack_length))
|
2004-09-24 18:58:25 +02:00
|
|
|
{
|
|
|
|
my_free((char*)data, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
my_free((char*)pack_data, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
if (cmp_frm(tab, pack_data, pack_length))
|
2004-09-24 18:58:25 +02:00
|
|
|
{
|
|
|
|
if (!invalidating_ndb_table)
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Invalidating table"));
|
2005-04-27 18:17:41 +02:00
|
|
|
invalidate_dictionary_cache(TRUE);
|
2005-02-16 14:18:32 +01:00
|
|
|
invalidating_ndb_table= TRUE;
|
2004-09-24 18:58:25 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("error",
|
2005-11-06 00:20:37 +01:00
|
|
|
("metadata, pack_length: %d getFrmLength: %d memcmp: %d",
|
2005-02-16 14:18:32 +01:00
|
|
|
pack_length, tab->getFrmLength(),
|
2005-11-06 00:20:37 +01:00
|
|
|
memcmp(pack_data, tab->getFrmData(), pack_length)));
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_DUMP("pack_data", (char*)pack_data, pack_length);
|
|
|
|
DBUG_DUMP("frm", (char*)tab->getFrmData(), tab->getFrmLength());
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
error= HA_ERR_TABLE_DEF_CHANGED;
|
2005-02-16 14:18:32 +01:00
|
|
|
invalidating_ndb_table= FALSE;
|
2004-09-24 18:58:25 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
invalidating_ndb_table= FALSE;
|
2004-09-24 18:58:25 +02:00
|
|
|
}
|
|
|
|
my_free((char*)data, MYF(0));
|
|
|
|
my_free((char*)pack_data, MYF(0));
|
|
|
|
} while (invalidating_ndb_table);
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (error)
|
|
|
|
DBUG_RETURN(error);
|
2004-09-03 13:55:40 +02:00
|
|
|
|
2005-04-27 18:17:41 +02:00
|
|
|
m_table_version= tab->getObjectVersion();
|
2005-03-22 17:42:08 +01:00
|
|
|
m_table= (void *)tab;
|
|
|
|
m_table_info= NULL; // Set in external lock
|
|
|
|
|
|
|
|
DBUG_RETURN(build_index_list(ndb, table, ILBP_OPEN));
|
2004-05-24 12:35:39 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-01-26 11:31:46 +01:00
|
|
|
static int fix_unique_index_attr_order(NDB_INDEX_DATA &data,
|
2005-02-16 14:18:32 +01:00
|
|
|
const NDBINDEX *index,
|
|
|
|
KEY *key_info)
|
2005-01-26 11:31:46 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("fix_unique_index_attr_order");
|
|
|
|
unsigned sz= index->getNoOfIndexColumns();
|
|
|
|
|
|
|
|
if (data.unique_index_attrid_map)
|
|
|
|
my_free((char*)data.unique_index_attrid_map, MYF(0));
|
|
|
|
data.unique_index_attrid_map= (unsigned char*)my_malloc(sz,MYF(MY_WME));
|
|
|
|
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
DBUG_ASSERT(key_info->key_parts == sz);
|
|
|
|
for (unsigned i= 0; key_part != end; key_part++, i++)
|
|
|
|
{
|
|
|
|
const char *field_name= key_part->field->field_name;
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
data.unique_index_attrid_map[i]= 255;
|
|
|
|
#endif
|
|
|
|
for (unsigned j= 0; j < sz; j++)
|
|
|
|
{
|
2005-03-22 17:42:08 +01:00
|
|
|
const NDBCOL *c= index->getColumn(j);
|
BUG#9626
- Fix valgrind warnings
- Remove static db, schema and table name buffers from Ndb.cpp
mysys/thr_alarm.c:
Initialise sact to zero
ndb/include/kernel/ndb_limits.h:
Set MAX_ATTR_NAME_SIZE to NAME_LEN which is tha maximum in MySQL
ndb/include/kernel/signaldata/GetTabInfo.hpp:
Clean up
ndb/include/ndbapi/Ndb.hpp:
Remove NDB_MAX_INTERNAL_NAME_LENGTH and all char buffers for schema, db and tablename.
Made them dynamic and moved to NdbImpl.hpp
ndb/include/ndbapi/ndbapi_limits.h:
Remove the static length's of attr, db, schema and table name.
ndb/src/common/transporter/Packer.cpp:
Set theSignalId to ~0 when unpacking signal
ndb/src/ndbapi/Ndb.cpp:
Moved schema, database and tablename to NdbImpl.hpp
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Add NdbIndexImpl::init and NdbEventImpl::init
Init all vars in NdbIndexImpl, NdbEventImpl, NdbTableImpl and NdbIndexImpl
Delete the pseudo column NDB$RANGE_NO
Copy tablename to internal buff in NdbDictInterface to get proper alignment.
Convert length of table name from bytes words, when setting sz of LinearSectionPtr
Set LinearSectionPtr array size to number of LinearSections used - save some stack.
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
Add NdbEventImpl::init and NdbIndexImpl::init
Remove clearNewProperties and copyNewProperties, it's easier to check if all vars are initied if it's done in the same func.
Add buffer for tabname ti NdbDictInterface, memset it to 0 in initializer.
ndb/src/ndbapi/NdbImpl.hpp:
Use BaseString for table, schema and db names.
ndb/src/ndbapi/Ndbinit.cpp:
Move schema and db name to NdbImpl and use BaseString
ndb/src/ndbapi/ndb_cluster_connection.cpp:
Destroy ndb_global_event_buffer_mutex and ndb_print_state_mutex
sql/ha_ndbcluster.cc:
Check if pTrans is not null before calling closeTransaction
Remove NDB_MAX_ATTR_NAME_SIZE
Remove truncation of attr names. When attr name length is same in NDB as in MySQL this will be checked in functin check_column_name
2005-05-18 20:50:29 +02:00
|
|
|
if (strcmp(field_name, c->getName()) == 0)
|
2005-01-26 11:31:46 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
data.unique_index_attrid_map[i]= j;
|
|
|
|
break;
|
2005-01-26 11:31:46 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
DBUG_ASSERT(data.unique_index_attrid_map[i] != 255);
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-08-19 11:10:35 +02:00
|
|
|
|
2005-03-22 17:42:08 +01:00
|
|
|
int ha_ndbcluster::build_index_list(Ndb *ndb, TABLE *tab, enum ILBP phase)
|
2004-05-24 12:35:39 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
uint i;
|
2004-08-19 11:10:35 +02:00
|
|
|
int error= 0;
|
2005-01-11 17:00:31 +01:00
|
|
|
const char *index_name;
|
2004-09-25 14:35:27 +02:00
|
|
|
char unique_index_name[FN_LEN];
|
2004-05-24 12:35:39 +02:00
|
|
|
static const char* unique_suffix= "$unique";
|
2004-08-19 11:10:35 +02:00
|
|
|
KEY* key_info= tab->key_info;
|
2005-01-06 12:00:13 +01:00
|
|
|
const char **key_name= tab->s->keynames.type_names;
|
2005-03-22 17:42:08 +01:00
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::build_index_list");
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Save information about all known indexes
|
2005-01-06 12:00:13 +01:00
|
|
|
for (i= 0; i < tab->s->keys; i++, key_info++, key_name++)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
2004-08-19 11:10:35 +02:00
|
|
|
index_name= *key_name;
|
2004-05-24 12:35:39 +02:00
|
|
|
NDB_INDEX_TYPE idx_type= get_index_type_from_table(i);
|
2004-08-18 19:13:39 +02:00
|
|
|
m_index[i].type= idx_type;
|
2004-05-24 12:35:39 +02:00
|
|
|
if (idx_type == UNIQUE_ORDERED_INDEX || idx_type == UNIQUE_INDEX)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
strxnmov(unique_index_name, FN_LEN-1, index_name, unique_suffix, NullS);
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Created unique index name \'%s\' for index %d",
|
2005-02-16 14:18:32 +01:00
|
|
|
unique_index_name, i));
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
2004-08-19 11:10:35 +02:00
|
|
|
// Create secondary indexes if in create phase
|
|
|
|
if (phase == ILBP_CREATE)
|
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Creating index %u: %s", i, index_name));
|
|
|
|
switch (idx_type){
|
2005-02-16 14:18:32 +01:00
|
|
|
|
2004-08-19 11:10:35 +02:00
|
|
|
case PRIMARY_KEY_INDEX:
|
2005-02-16 14:18:32 +01:00
|
|
|
// Do nothing, already created
|
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
2005-02-16 14:18:32 +01:00
|
|
|
error= create_ordered_index(index_name, key_info);
|
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
case UNIQUE_ORDERED_INDEX:
|
2005-02-16 14:18:32 +01:00
|
|
|
if (!(error= create_ordered_index(index_name, key_info)))
|
|
|
|
error= create_unique_index(unique_index_name, key_info);
|
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
case UNIQUE_INDEX:
|
2005-02-16 14:18:32 +01:00
|
|
|
if (!(error= check_index_fields_not_null(i)))
|
|
|
|
error= create_unique_index(unique_index_name, key_info);
|
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
case ORDERED_INDEX:
|
2005-02-16 14:18:32 +01:00
|
|
|
error= create_ordered_index(index_name, key_info);
|
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
default:
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_ASSERT(FALSE);
|
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
}
|
|
|
|
if (error)
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("error", ("Failed to create index %u", i));
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
intern_drop_table();
|
2005-02-16 14:18:32 +01:00
|
|
|
break;
|
2004-08-19 11:10:35 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
// Add handles to index objects
|
2004-09-25 14:35:27 +02:00
|
|
|
if (idx_type != PRIMARY_KEY_INDEX && idx_type != UNIQUE_INDEX)
|
2004-08-18 19:13:39 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Get handle to index %s", index_name));
|
2004-08-19 11:10:35 +02:00
|
|
|
const NDBINDEX *index= dict->getIndex(index_name, m_tabname);
|
2004-08-18 19:13:39 +02:00
|
|
|
if (!index) DBUG_RETURN(1);
|
2004-08-19 07:33:35 +02:00
|
|
|
m_index[i].index= (void *) index;
|
2005-09-15 02:33:28 +02:00
|
|
|
// ordered index - add stats
|
|
|
|
NDB_INDEX_DATA& d=m_index[i];
|
|
|
|
delete d.index_stat;
|
|
|
|
d.index_stat=NULL;
|
|
|
|
THD *thd=current_thd;
|
|
|
|
if (thd->variables.ndb_index_stat_enable)
|
|
|
|
{
|
|
|
|
d.index_stat=new NdbIndexStat(index);
|
|
|
|
d.index_stat_cache_entries=thd->variables.ndb_index_stat_cache_entries;
|
|
|
|
d.index_stat_update_freq=thd->variables.ndb_index_stat_update_freq;
|
|
|
|
d.index_stat_query_count=0;
|
|
|
|
d.index_stat->alloc_cache(d.index_stat_cache_entries);
|
|
|
|
DBUG_PRINT("info", ("index %s stat=on cache_entries=%u update_freq=%u",
|
|
|
|
index->getName(),
|
|
|
|
d.index_stat_cache_entries,
|
|
|
|
d.index_stat_update_freq));
|
|
|
|
} else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("index %s stat=off", index->getName()));
|
|
|
|
}
|
2004-08-18 19:13:39 +02:00
|
|
|
}
|
2004-09-25 14:35:27 +02:00
|
|
|
if (idx_type == UNIQUE_ORDERED_INDEX || idx_type == UNIQUE_INDEX)
|
2004-08-18 19:13:39 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Get handle to unique_index %s", unique_index_name));
|
|
|
|
const NDBINDEX *index= dict->getIndex(unique_index_name, m_tabname);
|
2004-08-18 19:13:39 +02:00
|
|
|
if (!index) DBUG_RETURN(1);
|
2004-08-19 07:33:35 +02:00
|
|
|
m_index[i].unique_index= (void *) index;
|
2005-01-26 11:31:46 +01:00
|
|
|
error= fix_unique_index_attr_order(m_index[i], index, key_info);
|
|
|
|
}
|
2004-08-18 19:13:39 +02:00
|
|
|
}
|
2004-08-19 11:10:35 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(error);
|
2004-08-18 19:13:39 +02:00
|
|
|
}
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Decode the type of an index from information
|
|
|
|
provided in table object
|
|
|
|
*/
|
2004-05-24 12:35:39 +02:00
|
|
|
NDB_INDEX_TYPE ha_ndbcluster::get_index_type_from_table(uint inx) const
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
bool is_hash_index= (table_share->key_info[inx].algorithm ==
|
|
|
|
HA_KEY_ALG_HASH);
|
|
|
|
if (inx == table_share->primary_key)
|
2004-05-24 12:35:39 +02:00
|
|
|
return is_hash_index ? PRIMARY_KEY_INDEX : PRIMARY_KEY_ORDERED_INDEX;
|
2005-01-06 12:00:13 +01:00
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
return ((table_share->key_info[inx].flags & HA_NOSAME) ?
|
2005-01-06 12:00:13 +01:00
|
|
|
(is_hash_index ? UNIQUE_INDEX : UNIQUE_ORDERED_INDEX) :
|
|
|
|
ORDERED_INDEX);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-11-22 10:35:03 +01:00
|
|
|
int ha_ndbcluster::check_index_fields_not_null(uint inx)
|
|
|
|
{
|
|
|
|
KEY* key_info= table->key_info + inx;
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::check_index_fields_not_null");
|
2004-11-22 10:35:03 +01:00
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
|
|
|
if (field->maybe_null())
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
my_printf_error(ER_NULL_COLUMN_IN_INDEX,ER(ER_NULL_COLUMN_IN_INDEX),
|
|
|
|
MYF(0),field->field_name);
|
|
|
|
DBUG_RETURN(ER_NULL_COLUMN_IN_INDEX);
|
2004-11-22 10:35:03 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
void ha_ndbcluster::release_metadata()
|
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
uint i;
|
2004-04-30 12:25:31 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("release_metadata");
|
|
|
|
DBUG_PRINT("enter", ("m_tabname: %s", m_tabname));
|
|
|
|
|
|
|
|
m_table= NULL;
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table_info= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
// Release index list
|
2004-04-30 12:25:31 +02:00
|
|
|
for (i= 0; i < MAX_KEY; i++)
|
|
|
|
{
|
2004-08-18 19:13:39 +02:00
|
|
|
m_index[i].unique_index= NULL;
|
|
|
|
m_index[i].index= NULL;
|
2005-01-26 11:31:46 +01:00
|
|
|
if (m_index[i].unique_index_attrid_map)
|
|
|
|
{
|
|
|
|
my_free((char *)m_index[i].unique_index_attrid_map, MYF(0));
|
|
|
|
m_index[i].unique_index_attrid_map= NULL;
|
|
|
|
}
|
2005-09-15 02:33:28 +02:00
|
|
|
delete m_index[i].index_stat;
|
|
|
|
m_index[i].index_stat=NULL;
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
int ha_ndbcluster::get_ndb_lock_type(enum thr_lock_type type)
|
2004-05-10 14:12:28 +02:00
|
|
|
{
|
2004-10-13 10:08:18 +02:00
|
|
|
if (type >= TL_WRITE_ALLOW_WRITE)
|
2004-09-29 01:42:34 +02:00
|
|
|
return NdbOperation::LM_Exclusive;
|
2005-04-28 14:45:27 +02:00
|
|
|
else if (uses_blob_value())
|
2004-10-05 13:51:32 +02:00
|
|
|
return NdbOperation::LM_Read;
|
2004-07-22 12:38:09 +02:00
|
|
|
else
|
2004-09-29 01:42:34 +02:00
|
|
|
return NdbOperation::LM_CommittedRead;
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
static const ulong index_type_flags[]=
|
|
|
|
{
|
|
|
|
/* UNDEFINED_INDEX */
|
|
|
|
0,
|
|
|
|
|
|
|
|
/* PRIMARY_KEY_INDEX */
|
2004-06-24 12:04:22 +02:00
|
|
|
HA_ONLY_WHOLE_INDEX,
|
2004-05-24 12:35:39 +02:00
|
|
|
|
|
|
|
/* PRIMARY_KEY_ORDERED_INDEX */
|
2004-05-10 14:46:06 +02:00
|
|
|
/*
|
2004-06-24 14:49:34 +02:00
|
|
|
Enable HA_KEYREAD_ONLY when "sorted" indexes are supported,
|
2004-05-10 14:46:06 +02:00
|
|
|
thus ORDERD BY clauses can be optimized by reading directly
|
|
|
|
through the index.
|
|
|
|
*/
|
2004-06-24 14:49:34 +02:00
|
|
|
// HA_KEYREAD_ONLY |
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_NEXT |
|
2004-12-20 15:12:42 +01:00
|
|
|
HA_READ_PREV |
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_RANGE |
|
|
|
|
HA_READ_ORDER,
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/* UNIQUE_INDEX */
|
2004-06-24 12:04:22 +02:00
|
|
|
HA_ONLY_WHOLE_INDEX,
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
/* UNIQUE_ORDERED_INDEX */
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_NEXT |
|
2004-12-20 15:12:42 +01:00
|
|
|
HA_READ_PREV |
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_RANGE |
|
|
|
|
HA_READ_ORDER,
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/* ORDERED_INDEX */
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_NEXT |
|
2004-12-20 15:12:42 +01:00
|
|
|
HA_READ_PREV |
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_RANGE |
|
|
|
|
HA_READ_ORDER
|
2004-04-15 09:14:14 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
static const int index_flags_size= sizeof(index_type_flags)/sizeof(ulong);
|
|
|
|
|
|
|
|
inline NDB_INDEX_TYPE ha_ndbcluster::get_index_type(uint idx_no) const
|
|
|
|
{
|
|
|
|
DBUG_ASSERT(idx_no < MAX_KEY);
|
2004-08-18 19:13:39 +02:00
|
|
|
return m_index[idx_no].type;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get the flags for an index
|
|
|
|
|
|
|
|
RETURN
|
|
|
|
flags depending on the type of the index.
|
|
|
|
*/
|
|
|
|
|
2004-07-08 14:45:25 +02:00
|
|
|
inline ulong ha_ndbcluster::index_flags(uint idx_no, uint part,
|
|
|
|
bool all_parts) const
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_flags");
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
DBUG_PRINT("enter", ("idx_no: %u", idx_no));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(get_index_type_from_table(idx_no) < index_flags_size);
|
2005-09-30 12:20:15 +02:00
|
|
|
DBUG_RETURN(index_type_flags[get_index_type_from_table(idx_no)] |
|
|
|
|
HA_KEY_SCAN_NOT_ROR);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-01-07 11:55:20 +01:00
|
|
|
static void shrink_varchar(Field* field, const byte* & ptr, char* buf)
|
|
|
|
{
|
2005-09-01 14:12:48 +02:00
|
|
|
if (field->type() == MYSQL_TYPE_VARCHAR && ptr != NULL) {
|
2005-01-07 11:55:20 +01:00
|
|
|
Field_varstring* f= (Field_varstring*)field;
|
2005-04-21 12:53:40 +02:00
|
|
|
if (f->length_bytes == 1) {
|
2005-01-07 11:55:20 +01:00
|
|
|
uint pack_len= field->pack_length();
|
|
|
|
DBUG_ASSERT(1 <= pack_len && pack_len <= 256);
|
|
|
|
if (ptr[1] == 0) {
|
|
|
|
buf[0]= ptr[0];
|
|
|
|
} else {
|
2005-05-06 10:39:30 +02:00
|
|
|
DBUG_ASSERT(FALSE);
|
2005-01-07 11:55:20 +01:00
|
|
|
buf[0]= 255;
|
|
|
|
}
|
|
|
|
memmove(buf + 1, ptr + 2, pack_len - 1);
|
|
|
|
ptr= buf;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::set_primary_key(NdbOperation *op, const byte *key)
|
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
KEY* key_info= table->key_info + table->s->primary_key;
|
2004-04-15 09:14:14 +02:00
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
DBUG_ENTER("set_primary_key");
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
2005-01-07 11:55:20 +01:00
|
|
|
const byte* ptr= key;
|
|
|
|
char buf[256];
|
|
|
|
shrink_varchar(field, ptr, buf);
|
2004-04-15 09:14:14 +02:00
|
|
|
if (set_ndb_key(op, field,
|
2005-02-16 14:18:32 +01:00
|
|
|
key_part->fieldnr-1, ptr))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2005-01-07 11:55:20 +01:00
|
|
|
key += key_part->store_length;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-07-06 11:23:36 +02:00
|
|
|
int ha_ndbcluster::set_primary_key_from_record(NdbOperation *op, const byte *record)
|
2004-05-17 09:51:02 +02:00
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
KEY* key_info= table->key_info + table->s->primary_key;
|
2004-05-17 09:51:02 +02:00
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
2005-06-29 16:13:50 +02:00
|
|
|
DBUG_ENTER("set_primary_key_from_record");
|
2004-05-17 09:51:02 +02:00
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
|
|
|
if (set_ndb_key(op, field,
|
2005-07-06 11:23:36 +02:00
|
|
|
key_part->fieldnr-1, record+key_part->offset))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-11-30 13:37:03 +01:00
|
|
|
int
|
|
|
|
ha_ndbcluster::set_index_key(NdbOperation *op,
|
2005-02-16 14:18:32 +01:00
|
|
|
const KEY *key_info,
|
|
|
|
const byte * key_ptr)
|
2004-11-30 13:37:03 +01:00
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::set_index_key");
|
2004-11-30 13:37:03 +01:00
|
|
|
uint i;
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
|
|
|
|
for (i= 0; key_part != end; key_part++, i++)
|
|
|
|
{
|
2005-01-07 11:55:20 +01:00
|
|
|
Field* field= key_part->field;
|
|
|
|
const byte* ptr= key_part->null_bit ? key_ptr + 1 : key_ptr;
|
|
|
|
char buf[256];
|
|
|
|
shrink_varchar(field, ptr, buf);
|
2005-01-26 13:22:07 +01:00
|
|
|
if (set_ndb_key(op, field, m_index[active_index].unique_index_attrid_map[i], ptr))
|
2004-11-30 13:37:03 +01:00
|
|
|
ERR_RETURN(m_active_trans->getNdbError());
|
|
|
|
key_ptr+= key_part->store_length;
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-01-18 16:31:05 +01:00
|
|
|
inline
|
|
|
|
int ha_ndbcluster::define_read_attrs(byte* buf, NdbOperation* op)
|
|
|
|
{
|
|
|
|
uint i;
|
|
|
|
DBUG_ENTER("define_read_attrs");
|
|
|
|
|
|
|
|
// Define attributes to read
|
|
|
|
for (i= 0; i < table->s->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
2005-04-28 14:45:27 +02:00
|
|
|
if (ha_get_bit_in_read_set(i+1) ||
|
|
|
|
((field->flags & PRI_KEY_FLAG)))
|
2005-01-18 16:31:05 +01:00
|
|
|
{
|
|
|
|
if (get_ndb_value(op, field, i, buf))
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2005-01-18 16:31:05 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
m_value[i].ptr= NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (table->s->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Getting hidden key"));
|
|
|
|
// Scanning table with no primary key
|
|
|
|
int hidden_no= table->s->fields;
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
|
|
|
if (!tab->getColumn(hidden_no))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
#endif
|
|
|
|
if (get_ndb_value(op, NULL, hidden_no, NULL))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2005-07-12 20:01:22 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Read one record from NDB using primary key
|
|
|
|
*/
|
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
int ha_ndbcluster::pk_read(const byte *key, uint key_len, byte *buf,
|
|
|
|
uint32 part_id)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-01-11 17:00:31 +01:00
|
|
|
uint no_fields= table->s->fields;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbOperation *op;
|
2005-01-11 17:00:31 +01:00
|
|
|
|
2004-11-17 10:07:52 +01:00
|
|
|
int res;
|
|
|
|
DBUG_ENTER("pk_read");
|
|
|
|
DBUG_PRINT("enter", ("key_len: %u", key_len));
|
|
|
|
DBUG_DUMP("key", (char*)key, key_len);
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2004-09-28 19:35:07 +02:00
|
|
|
op->readTuple(lm) != 0)
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-17 10:07:52 +01:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
op->setPartitionId(part_id);
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key == MAX_KEY)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
// This table has no primary key, use "hidden" primary key
|
|
|
|
DBUG_PRINT("info", ("Using hidden key"));
|
|
|
|
DBUG_DUMP("key", (char*)key, 8);
|
|
|
|
if (set_hidden_key(op, no_fields, key))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-17 10:07:52 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Read key at the same time, for future reference
|
2004-09-20 12:40:53 +02:00
|
|
|
if (get_ndb_value(op, NULL, no_fields, NULL))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if ((res= set_primary_key(op, key)))
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if ((res= define_read_attrs(buf, op)))
|
2004-11-17 10:07:52 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-28 21:11:50 +02:00
|
|
|
if (execute_no_commit_ie(this,trans) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
|
|
|
|
// The value have now been fetched from NDB
|
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
/*
|
|
|
|
Read one complementing record from NDB using primary key from old_data
|
|
|
|
*/
|
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
int ha_ndbcluster::complemented_pk_read(const byte *old_data, byte *new_data,
|
|
|
|
uint32 old_part_id)
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
uint no_fields= table->s->fields, i;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-07-22 18:35:51 +02:00
|
|
|
NdbOperation *op;
|
|
|
|
DBUG_ENTER("complemented_pk_read");
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= FALSE;
|
2004-07-22 18:35:51 +02:00
|
|
|
|
2005-04-28 14:45:27 +02:00
|
|
|
if (ha_get_all_bit_in_read_set())
|
|
|
|
{
|
2004-07-22 18:35:51 +02:00
|
|
|
// We have allready retrieved all fields, nothing to complement
|
|
|
|
DBUG_RETURN(0);
|
2005-04-28 14:45:27 +02:00
|
|
|
}
|
2004-07-22 18:35:51 +02:00
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2004-09-28 19:35:07 +02:00
|
|
|
op->readTuple(lm) != 0)
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-12-01 12:43:30 +01:00
|
|
|
int res;
|
2005-07-04 20:36:38 +02:00
|
|
|
if ((res= set_primary_key_from_record(op, old_data)))
|
2004-12-01 12:43:30 +01:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2005-07-18 13:31:02 +02:00
|
|
|
|
|
|
|
if (m_use_partition_function)
|
|
|
|
op->setPartitionId(old_part_id);
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
// Read all unreferenced non-key field(s)
|
|
|
|
for (i= 0; i < no_fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
2004-12-01 12:43:30 +01:00
|
|
|
if (!((field->flags & PRI_KEY_FLAG) ||
|
2005-04-28 14:45:27 +02:00
|
|
|
(ha_get_bit_in_read_set(i+1))))
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
2004-09-20 12:40:53 +02:00
|
|
|
if (get_ndb_value(op, field, i, new_data))
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-07-22 18:35:51 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0)
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
|
|
|
|
// The value have now been fetched from NDB
|
|
|
|
unpack_record(new_data);
|
|
|
|
table->status= 0;
|
2004-12-01 12:43:30 +01:00
|
|
|
|
|
|
|
/**
|
|
|
|
* restore m_value
|
|
|
|
*/
|
|
|
|
for (i= 0; i < no_fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if (!((field->flags & PRI_KEY_FLAG) ||
|
2005-04-28 14:45:27 +02:00
|
|
|
(ha_get_bit_in_read_set(i+1))))
|
2004-12-01 12:43:30 +01:00
|
|
|
{
|
|
|
|
m_value[i].ptr= NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-11-18 12:11:56 +01:00
|
|
|
/*
|
|
|
|
Peek to check if a particular row already exists
|
|
|
|
*/
|
|
|
|
|
2005-06-29 16:13:50 +02:00
|
|
|
int ha_ndbcluster::peek_row(const byte *record)
|
2004-11-18 12:11:56 +01:00
|
|
|
{
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-11-18 12:11:56 +01:00
|
|
|
NdbOperation *op;
|
|
|
|
DBUG_ENTER("peek_row");
|
2004-12-08 00:36:40 +01:00
|
|
|
|
2004-11-18 12:11:56 +01:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
|
|
|
op->readTuple(lm) != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-12-08 00:36:40 +01:00
|
|
|
|
2004-11-18 12:11:56 +01:00
|
|
|
int res;
|
2005-06-29 16:13:50 +02:00
|
|
|
if ((res= set_primary_key_from_record(op, record)))
|
2004-11-18 12:11:56 +01:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-12-08 00:36:40 +01:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
{
|
|
|
|
uint32 part_id;
|
|
|
|
int error;
|
|
|
|
if ((error= m_part_info->get_partition_id(m_part_info, &part_id)))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
op->setPartitionId(part_id);
|
|
|
|
}
|
|
|
|
|
2004-11-18 12:11:56 +01:00
|
|
|
if (execute_no_commit_ie(this,trans) != 0)
|
2004-12-08 00:36:40 +01:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
2004-11-18 12:11:56 +01:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-07-22 18:35:51 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Read one record from NDB using unique secondary index
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::unique_index_read(const byte *key,
|
2005-02-16 14:18:32 +01:00
|
|
|
uint key_len, byte *buf)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-11-17 10:07:52 +01:00
|
|
|
int res;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbIndexOperation *op;
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::unique_index_read");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("key_len: %u, index: %u", key_len, active_index));
|
|
|
|
DBUG_DUMP("key", (char*)key, key_len);
|
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
2004-08-18 19:13:39 +02:00
|
|
|
if (!(op= trans->getNdbIndexOperation((NDBINDEX *)
|
2005-02-16 14:18:32 +01:00
|
|
|
m_index[active_index].unique_index,
|
2004-09-07 21:53:59 +02:00
|
|
|
(const NDBTAB *) m_table)) ||
|
2004-09-28 19:35:07 +02:00
|
|
|
op->readTuple(lm) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
// Set secondary index key(s)
|
2005-07-04 02:42:33 +02:00
|
|
|
if ((res= set_index_key(op, table->key_info + active_index, key)))
|
2004-11-30 13:37:03 +01:00
|
|
|
DBUG_RETURN(res);
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if ((res= define_read_attrs(buf, op)))
|
2004-11-17 10:07:52 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-28 21:11:50 +02:00
|
|
|
if (execute_no_commit_ie(this,trans) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
// The value have now been fetched from NDB
|
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-12-08 00:36:40 +01:00
|
|
|
inline int ha_ndbcluster::fetch_next(NdbScanOperation* cursor)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("fetch_next");
|
2004-05-10 14:12:28 +02:00
|
|
|
int check;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2004-10-13 10:08:18 +02:00
|
|
|
bool contact_ndb= m_lock.type < TL_WRITE_ALLOW_WRITE;
|
2004-05-10 14:12:28 +02:00
|
|
|
do {
|
|
|
|
DBUG_PRINT("info", ("Call nextResult, contact_ndb: %d", contact_ndb));
|
2004-07-22 12:38:09 +02:00
|
|
|
/*
|
|
|
|
We can only handle one tuple with blobs at a time.
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_ops_pending && m_blobs_pending)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0)
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
|
|
|
m_blobs_pending= FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
2004-12-07 15:15:49 +01:00
|
|
|
|
|
|
|
if ((check= cursor->nextResult(contact_ndb, m_force_send)) == 0)
|
2004-05-10 14:12:28 +02:00
|
|
|
{
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
else if (check == 1 || check == 2)
|
|
|
|
{
|
|
|
|
// 1: No more records
|
|
|
|
// 2: No more cached records
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
/*
|
2005-02-16 14:18:32 +01:00
|
|
|
Before fetching more rows and releasing lock(s),
|
|
|
|
all pending update or delete operations should
|
|
|
|
be sent to NDB
|
2004-05-10 14:12:28 +02:00
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
DBUG_PRINT("info", ("ops_pending: %d", m_ops_pending));
|
|
|
|
if (m_ops_pending)
|
2004-09-03 17:11:09 +02:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
if (m_transaction_on)
|
|
|
|
{
|
|
|
|
if (execute_no_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (execute_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(-1);
|
2005-07-04 02:42:33 +02:00
|
|
|
if (trans->restart() != 0)
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
|
|
|
DBUG_ASSERT(0);
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
m_ops_pending= 0;
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
contact_ndb= (check == 2);
|
|
|
|
}
|
2004-12-08 00:36:40 +01:00
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
} while (check == 2);
|
2004-12-08 00:36:40 +01:00
|
|
|
|
2004-12-07 15:15:49 +01:00
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get the next record of a started scan. Try to fetch
|
|
|
|
it locally from NdbApi cached records if possible,
|
|
|
|
otherwise ask NDB for more.
|
|
|
|
|
|
|
|
NOTE
|
|
|
|
If this is a update/delete make sure to not contact
|
|
|
|
NDB before any pending ops have been sent to NDB.
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-12-07 15:15:49 +01:00
|
|
|
*/
|
|
|
|
|
|
|
|
inline int ha_ndbcluster::next_result(byte *buf)
|
|
|
|
{
|
|
|
|
int res;
|
|
|
|
DBUG_ENTER("next_result");
|
|
|
|
|
2004-12-15 15:13:11 +01:00
|
|
|
if (!m_active_cursor)
|
|
|
|
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if ((res= fetch_next(m_active_cursor)) == 0)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("One more record found"));
|
|
|
|
|
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
else if (res == 1)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
|
|
|
// No more records
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
|
|
|
|
DBUG_PRINT("info", ("No more records"));
|
|
|
|
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_RETURN(ndb_err(m_active_trans));
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-04-30 13:38:41 +02:00
|
|
|
/*
|
2004-10-08 10:50:50 +02:00
|
|
|
Set bounds for ordered index scan.
|
2004-04-30 13:38:41 +02:00
|
|
|
*/
|
|
|
|
|
2004-06-11 13:49:22 +02:00
|
|
|
int ha_ndbcluster::set_bounds(NdbIndexScanOperation *op,
|
2005-09-15 02:33:28 +02:00
|
|
|
uint inx,
|
|
|
|
bool rir,
|
2005-02-16 14:18:32 +01:00
|
|
|
const key_range *keys[2],
|
|
|
|
uint range_no)
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
2005-09-15 02:33:28 +02:00
|
|
|
const KEY *const key_info= table->key_info + inx;
|
2004-10-08 10:50:50 +02:00
|
|
|
const uint key_parts= key_info->key_parts;
|
|
|
|
uint key_tot_len[2];
|
|
|
|
uint tot_len;
|
2004-10-25 22:35:46 +02:00
|
|
|
uint i, j;
|
2004-04-30 13:38:41 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("set_bounds");
|
2004-10-08 10:50:50 +02:00
|
|
|
DBUG_PRINT("info", ("key_parts=%d", key_parts));
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
for (j= 0; j <= 1; j++)
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
2004-10-08 10:50:50 +02:00
|
|
|
const key_range *key= keys[j];
|
|
|
|
if (key != NULL)
|
|
|
|
{
|
|
|
|
// for key->flag see ha_rkey_function
|
|
|
|
DBUG_PRINT("info", ("key %d length=%d flag=%d",
|
|
|
|
j, key->length, key->flag));
|
|
|
|
key_tot_len[j]= key->length;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("key %d not present", j));
|
|
|
|
key_tot_len[j]= 0;
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
}
|
|
|
|
tot_len= 0;
|
2004-08-31 10:19:10 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
for (i= 0; i < key_parts; i++)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO *key_part= &key_info->key_part[i];
|
|
|
|
Field *field= key_part->field;
|
2005-01-11 17:00:31 +01:00
|
|
|
#ifndef DBUG_OFF
|
2004-10-08 10:50:50 +02:00
|
|
|
uint part_len= key_part->length;
|
2005-01-11 17:00:31 +01:00
|
|
|
#endif
|
2004-10-08 10:50:50 +02:00
|
|
|
uint part_store_len= key_part->store_length;
|
|
|
|
// Info about each key part
|
|
|
|
struct part_st {
|
|
|
|
bool part_last;
|
|
|
|
const key_range *key;
|
|
|
|
const byte *part_ptr;
|
|
|
|
bool part_null;
|
|
|
|
int bound_type;
|
|
|
|
const char* bound_ptr;
|
|
|
|
};
|
|
|
|
struct part_st part[2];
|
|
|
|
|
|
|
|
for (j= 0; j <= 1; j++)
|
|
|
|
{
|
2004-12-07 15:15:49 +01:00
|
|
|
struct part_st &p= part[j];
|
2004-10-08 10:50:50 +02:00
|
|
|
p.key= NULL;
|
|
|
|
p.bound_type= -1;
|
|
|
|
if (tot_len < key_tot_len[j])
|
|
|
|
{
|
|
|
|
p.part_last= (tot_len + part_store_len >= key_tot_len[j]);
|
|
|
|
p.key= keys[j];
|
|
|
|
p.part_ptr= &p.key->key[tot_len];
|
2004-11-08 11:11:49 +01:00
|
|
|
p.part_null= key_part->null_bit && *p.part_ptr;
|
2004-10-08 10:50:50 +02:00
|
|
|
p.bound_ptr= (const char *)
|
2004-11-08 11:11:49 +01:00
|
|
|
p.part_null ? 0 : key_part->null_bit ? p.part_ptr + 1 : p.part_ptr;
|
2004-10-08 10:50:50 +02:00
|
|
|
|
|
|
|
if (j == 0)
|
|
|
|
{
|
|
|
|
switch (p.key->flag)
|
|
|
|
{
|
|
|
|
case HA_READ_KEY_EXACT:
|
2005-09-15 02:33:28 +02:00
|
|
|
if (! rir)
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundEQ;
|
|
|
|
else // differs for records_in_range
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLE;
|
2004-10-08 10:50:50 +02:00
|
|
|
break;
|
2004-12-20 15:12:42 +01:00
|
|
|
// ascending
|
2004-10-08 10:50:50 +02:00
|
|
|
case HA_READ_KEY_OR_NEXT:
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLE;
|
|
|
|
break;
|
|
|
|
case HA_READ_AFTER_KEY:
|
|
|
|
if (! p.part_last)
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLE;
|
|
|
|
else
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLT;
|
|
|
|
break;
|
2004-12-20 15:12:42 +01:00
|
|
|
// descending
|
|
|
|
case HA_READ_PREFIX_LAST: // weird
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundEQ;
|
|
|
|
break;
|
|
|
|
case HA_READ_PREFIX_LAST_OR_PREV: // weird
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGE;
|
|
|
|
break;
|
|
|
|
case HA_READ_BEFORE_KEY:
|
|
|
|
if (! p.part_last)
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGE;
|
|
|
|
else
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGT;
|
|
|
|
break;
|
2004-10-08 10:50:50 +02:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (j == 1) {
|
|
|
|
switch (p.key->flag)
|
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
// ascending
|
2004-10-08 10:50:50 +02:00
|
|
|
case HA_READ_BEFORE_KEY:
|
|
|
|
if (! p.part_last)
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGE;
|
|
|
|
else
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGT;
|
|
|
|
break;
|
|
|
|
case HA_READ_AFTER_KEY: // weird
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGE;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
2004-12-20 15:12:42 +01:00
|
|
|
// descending strangely sets no end key
|
2004-10-08 10:50:50 +02:00
|
|
|
}
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
if (p.bound_type == -1)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("key %d unknown flag %d", j, p.key->flag));
|
2005-05-06 10:39:30 +02:00
|
|
|
DBUG_ASSERT(FALSE);
|
2004-10-08 10:50:50 +02:00
|
|
|
// Stop setting bounds but continue with what we have
|
2005-02-16 14:18:32 +01:00
|
|
|
op->end_of_bound(range_no);
|
2004-10-08 10:50:50 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
// Seen with e.g. b = 1 and c > 1
|
|
|
|
if (part[0].bound_type == NdbIndexScanOperation::BoundLE &&
|
|
|
|
part[1].bound_type == NdbIndexScanOperation::BoundGE &&
|
|
|
|
memcmp(part[0].part_ptr, part[1].part_ptr, part_store_len) == 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("replace LE/GE pair by EQ"));
|
|
|
|
part[0].bound_type= NdbIndexScanOperation::BoundEQ;
|
|
|
|
part[1].bound_type= -1;
|
|
|
|
}
|
|
|
|
// Not seen but was in previous version
|
|
|
|
if (part[0].bound_type == NdbIndexScanOperation::BoundEQ &&
|
|
|
|
part[1].bound_type == NdbIndexScanOperation::BoundGE &&
|
|
|
|
memcmp(part[0].part_ptr, part[1].part_ptr, part_store_len) == 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("remove GE from EQ/GE pair"));
|
|
|
|
part[1].bound_type= -1;
|
|
|
|
}
|
2004-09-06 10:33:12 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
for (j= 0; j <= 1; j++)
|
|
|
|
{
|
2004-12-07 15:15:49 +01:00
|
|
|
struct part_st &p= part[j];
|
2004-10-08 10:50:50 +02:00
|
|
|
// Set bound if not done with this key
|
|
|
|
if (p.key != NULL)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("key %d:%d offset=%d length=%d last=%d bound=%d",
|
|
|
|
j, i, tot_len, part_len, p.part_last, p.bound_type));
|
|
|
|
DBUG_DUMP("info", (const char*)p.part_ptr, part_store_len);
|
|
|
|
|
|
|
|
// Set bound if not cancelled via type -1
|
|
|
|
if (p.bound_type != -1)
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-01-07 11:55:20 +01:00
|
|
|
const char* ptr= p.bound_ptr;
|
|
|
|
char buf[256];
|
|
|
|
shrink_varchar(field, ptr, buf);
|
2005-01-26 13:22:07 +01:00
|
|
|
if (op->setBound(i, p.bound_type, ptr))
|
2004-10-08 10:50:50 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2004-10-08 10:50:50 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
tot_len+= part_store_len;
|
2004-09-06 10:33:12 +02:00
|
|
|
}
|
2004-12-06 09:05:30 +01:00
|
|
|
op->end_of_bound(range_no);
|
2004-04-30 13:38:41 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
2004-05-10 14:12:28 +02:00
|
|
|
Start ordered index scan in NDB
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2004-04-30 13:38:41 +02:00
|
|
|
int ha_ndbcluster::ordered_index_scan(const key_range *start_key,
|
2005-02-16 14:18:32 +01:00
|
|
|
const key_range *end_key,
|
2005-07-18 13:31:02 +02:00
|
|
|
bool sorted, bool descending,
|
|
|
|
byte* buf, part_id_range *part_spec)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-11-17 10:07:52 +01:00
|
|
|
int res;
|
2004-10-05 12:17:42 +02:00
|
|
|
bool restart;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-06-11 13:49:22 +02:00
|
|
|
NdbIndexScanOperation *op;
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::ordered_index_scan");
|
|
|
|
DBUG_PRINT("enter", ("index: %u, sorted: %d, descending: %d",
|
|
|
|
active_index, sorted, descending));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("Starting new ordered scan on %s", m_tabname));
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
|
2004-10-07 11:57:48 +02:00
|
|
|
// Check that sorted seems to be initialised
|
|
|
|
DBUG_ASSERT(sorted == 0 || sorted == 1);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
if (m_active_cursor == 0)
|
2004-10-05 12:17:42 +02:00
|
|
|
{
|
2005-05-06 10:39:30 +02:00
|
|
|
restart= FALSE;
|
2004-10-05 12:17:42 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op= trans->getNdbIndexScanOperation((NDBINDEX *)
|
2005-02-16 14:18:32 +01:00
|
|
|
m_index[active_index].index,
|
|
|
|
(const NDBTAB *) m_table)) ||
|
|
|
|
op->readTuples(lm, 0, parallelism, sorted, descending))
|
2004-10-05 12:17:42 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function && part_spec != NULL &&
|
|
|
|
part_spec->start_part == part_spec->end_part)
|
|
|
|
op->setPartitionId(part_spec->start_part);
|
2004-12-07 15:15:49 +01:00
|
|
|
m_active_cursor= op;
|
2004-10-05 12:17:42 +02:00
|
|
|
} else {
|
2005-05-06 10:39:30 +02:00
|
|
|
restart= TRUE;
|
2004-12-07 15:15:49 +01:00
|
|
|
op= (NdbIndexScanOperation*)m_active_cursor;
|
2004-10-05 12:17:42 +02:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function && part_spec != NULL &&
|
|
|
|
part_spec->start_part == part_spec->end_part)
|
|
|
|
op->setPartitionId(part_spec->start_part);
|
2004-10-05 12:17:42 +02:00
|
|
|
DBUG_ASSERT(op->getSorted() == sorted);
|
|
|
|
DBUG_ASSERT(op->getLockMode() ==
|
2005-02-16 14:18:32 +01:00
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type));
|
2005-07-04 02:42:33 +02:00
|
|
|
if (op->reset_bounds(m_force_send))
|
2004-10-05 12:17:42 +02:00
|
|
|
DBUG_RETURN(ndb_err(m_active_trans));
|
|
|
|
}
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2004-05-10 14:46:06 +02:00
|
|
|
{
|
2004-10-08 10:50:50 +02:00
|
|
|
const key_range *keys[2]= { start_key, end_key };
|
2005-09-15 02:33:28 +02:00
|
|
|
res= set_bounds(op, active_index, false, keys);
|
2004-11-17 10:07:52 +01:00
|
|
|
if (res)
|
|
|
|
DBUG_RETURN(res);
|
2004-05-10 14:46:06 +02:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
|
|
|
|
if (!restart && generate_scan_filter(m_cond_stack, op))
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2004-11-17 10:07:52 +01:00
|
|
|
if (!restart && (res= define_read_attrs(buf, op)))
|
2004-05-10 14:46:06 +02:00
|
|
|
{
|
2004-11-17 10:07:52 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-10-05 12:17:42 +02:00
|
|
|
}
|
2004-11-17 10:07:52 +01:00
|
|
|
|
|
|
|
if (execute_no_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
|
|
|
|
DBUG_RETURN(next_result(buf));
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
2004-05-10 14:12:28 +02:00
|
|
|
Start full table scan in NDB
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::full_table_scan(byte *buf)
|
|
|
|
{
|
2004-11-17 10:07:52 +01:00
|
|
|
int res;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbScanOperation *op;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("full_table_scan");
|
|
|
|
DBUG_PRINT("enter", ("Starting new scan on %s", m_tabname));
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op=trans->getNdbScanOperation((const NDBTAB *) m_table)) ||
|
2004-12-07 15:15:49 +01:00
|
|
|
op->readTuples(lm, 0, parallelism))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-12-07 15:15:49 +01:00
|
|
|
m_active_cursor= op;
|
2005-01-07 15:33:24 +01:00
|
|
|
if (generate_scan_filter(m_cond_stack, op))
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2005-07-04 02:42:33 +02:00
|
|
|
if ((res= define_read_attrs(buf, op)))
|
2004-11-17 10:07:52 +01:00
|
|
|
DBUG_RETURN(res);
|
|
|
|
|
|
|
|
if (execute_no_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
DBUG_PRINT("exit", ("Scan started successfully"));
|
|
|
|
DBUG_RETURN(next_result(buf));
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Insert one record into NDB
|
|
|
|
*/
|
|
|
|
int ha_ndbcluster::write_row(byte *record)
|
|
|
|
{
|
2004-08-11 15:06:17 +02:00
|
|
|
bool has_auto_increment;
|
2004-04-15 09:14:14 +02:00
|
|
|
uint i;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbOperation *op;
|
|
|
|
int res;
|
2004-09-14 21:02:23 +02:00
|
|
|
THD *thd= current_thd;
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= TRUE;
|
2004-09-14 21:02:23 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("write_row");
|
2004-09-15 14:44:21 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
if (!m_use_write && m_ignore_dup_key && table->s->primary_key != MAX_KEY)
|
2004-09-15 14:44:21 +02:00
|
|
|
{
|
2005-06-29 16:13:50 +02:00
|
|
|
int peek_res= peek_row(record);
|
2004-11-18 12:11:56 +01:00
|
|
|
|
|
|
|
if (!peek_res)
|
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
m_dupkey= table->s->primary_key;
|
2004-11-18 12:11:56 +01:00
|
|
|
DBUG_RETURN(HA_ERR_FOUND_DUPP_KEY);
|
|
|
|
}
|
|
|
|
if (peek_res != HA_ERR_KEY_NOT_FOUND)
|
|
|
|
DBUG_RETURN(peek_res);
|
2004-09-15 14:44:21 +02:00
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(thd->status_var.ha_write_count, &LOCK_status);
|
2004-10-04 00:27:04 +02:00
|
|
|
if (table->timestamp_field_type & TIMESTAMP_AUTO_SET_ON_INSERT)
|
|
|
|
table->timestamp_field->set_time();
|
2004-07-29 11:35:38 +02:00
|
|
|
has_auto_increment= (table->next_number_field && record == table->record[0]);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
res= (m_use_write) ? op->writeTuple() :op->insertTuple();
|
|
|
|
if (res != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
{
|
|
|
|
uint32 part_id;
|
|
|
|
int error;
|
|
|
|
if ((error= m_part_info->get_partition_id(m_part_info, &part_id)))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
op->setPartitionId(part_id);
|
|
|
|
}
|
|
|
|
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key == MAX_KEY)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
// Table has hidden primary key
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2005-04-15 16:14:08 +02:00
|
|
|
Uint64 auto_value= NDB_FAILED_AUTO_INCREMENT;
|
|
|
|
uint retries= NDB_AUTO_INCREMENT_RETRIES;
|
|
|
|
do {
|
|
|
|
auto_value= ndb->getAutoIncrementValue((const NDBTAB *) m_table);
|
|
|
|
} while (auto_value == NDB_FAILED_AUTO_INCREMENT &&
|
|
|
|
--retries &&
|
|
|
|
ndb->getNdbError().status == NdbError::TemporaryError);
|
2005-04-15 15:52:07 +02:00
|
|
|
if (auto_value == NDB_FAILED_AUTO_INCREMENT)
|
|
|
|
ERR_RETURN(ndb->getNdbError());
|
2005-01-06 12:00:13 +01:00
|
|
|
if (set_hidden_key(op, table->s->fields, (const byte*)&auto_value))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
2004-08-13 13:51:18 +02:00
|
|
|
|
2004-09-16 04:55:22 +02:00
|
|
|
if (has_auto_increment)
|
|
|
|
{
|
2005-04-08 12:34:52 +02:00
|
|
|
THD *thd= table->in_use;
|
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= FALSE;
|
2004-08-13 13:51:18 +02:00
|
|
|
update_auto_increment();
|
2005-04-08 12:34:52 +02:00
|
|
|
/* Ensure that handler is always called for auto_increment values */
|
|
|
|
thd->next_insert_id= 0;
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= !auto_increment_column_changed;
|
2004-09-16 04:55:22 +02:00
|
|
|
}
|
2004-08-13 13:51:18 +02:00
|
|
|
|
2005-06-29 16:13:50 +02:00
|
|
|
if ((res= set_primary_key_from_record(op, record)))
|
2005-04-19 11:21:26 +02:00
|
|
|
return res;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Set non-key attribute(s)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool set_blob_value= FALSE;
|
2005-01-06 12:00:13 +01:00
|
|
|
for (i= 0; i < table->s->fields; i++)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if (!(field->flags & PRI_KEY_FLAG) &&
|
2005-11-06 00:20:37 +01:00
|
|
|
(ha_get_bit_in_write_set(i + 1) || !m_use_write) &&
|
|
|
|
set_ndb_value(op, field, i, record-table->record[0], &set_blob_value))
|
2004-08-13 13:51:18 +02:00
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2004-08-13 13:51:18 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
m_rows_changed++;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Execute write operation
|
|
|
|
NOTE When doing inserts with many values in
|
|
|
|
each INSERT statement it should not be necessary
|
|
|
|
to NoCommit the transaction between each row.
|
|
|
|
Find out how this is detected!
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_inserted++;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_update(1);
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_not_flushed= TRUE;
|
2005-04-13 12:42:23 +02:00
|
|
|
if ((m_rows_to_insert == (ha_rows) 1) ||
|
2004-11-03 15:53:26 +01:00
|
|
|
((m_rows_inserted % m_bulk_insert_rows) == 0) ||
|
2005-04-21 13:09:19 +02:00
|
|
|
m_primary_key_update ||
|
2004-10-01 13:16:49 +02:00
|
|
|
set_blob_value)
|
2004-04-29 14:38:35 +02:00
|
|
|
{
|
|
|
|
// Send rows to NDB
|
|
|
|
DBUG_PRINT("info", ("Sending inserts to NDB, "\
|
2005-02-16 14:18:32 +01:00
|
|
|
"rows_inserted:%d, bulk_insert_rows: %d",
|
|
|
|
(int)m_rows_inserted, (int)m_bulk_insert_rows));
|
2004-09-03 17:11:09 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_not_flushed= FALSE;
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_transaction_on)
|
2004-08-13 13:51:18 +02:00
|
|
|
{
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0)
|
2004-09-07 14:30:11 +02:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
|
|
|
no_uncommitted_rows_execute_failure();
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-07 14:30:11 +02:00
|
|
|
}
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
|
|
|
else
|
2004-08-13 13:51:18 +02:00
|
|
|
{
|
2004-09-28 21:11:50 +02:00
|
|
|
if (execute_commit(this,trans) != 0)
|
2004-09-03 17:11:09 +02:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
|
|
|
no_uncommitted_rows_execute_failure();
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
if (trans->restart() != 0)
|
2005-01-11 17:00:31 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_ASSERT(0);
|
|
|
|
DBUG_RETURN(-1);
|
2005-01-11 17:00:31 +01:00
|
|
|
}
|
2004-08-13 13:51:18 +02:00
|
|
|
}
|
2004-04-29 14:38:35 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
if ((has_auto_increment) && (m_skip_auto_increment))
|
2004-08-10 14:38:24 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2004-07-29 11:35:38 +02:00
|
|
|
Uint64 next_val= (Uint64) table->next_number_field->val_int() + 1;
|
2004-08-10 14:38:24 +02:00
|
|
|
DBUG_PRINT("info",
|
2005-02-16 14:18:32 +01:00
|
|
|
("Trying to set next auto increment value to %lu",
|
2004-08-24 17:00:45 +02:00
|
|
|
(ulong) next_val));
|
2004-12-30 19:56:09 +01:00
|
|
|
if (ndb->setAutoIncrementValue((const NDBTAB *) m_table, next_val, TRUE))
|
2004-08-10 14:38:24 +02:00
|
|
|
DBUG_PRINT("info",
|
2005-02-16 14:18:32 +01:00
|
|
|
("Setting next auto increment value to %u", next_val));
|
2004-07-29 11:35:38 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
2004-07-29 11:35:38 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/* Compare if a key in a row has changed */
|
|
|
|
|
|
|
|
int ha_ndbcluster::key_cmp(uint keynr, const byte * old_row,
|
2005-02-16 14:18:32 +01:00
|
|
|
const byte * new_row)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
KEY_PART_INFO *key_part=table->key_info[keynr].key_part;
|
|
|
|
KEY_PART_INFO *end=key_part+table->key_info[keynr].key_parts;
|
|
|
|
|
|
|
|
for (; key_part != end ; key_part++)
|
|
|
|
{
|
|
|
|
if (key_part->null_bit)
|
|
|
|
{
|
|
|
|
if ((old_row[key_part->null_offset] & key_part->null_bit) !=
|
2005-02-16 14:18:32 +01:00
|
|
|
(new_row[key_part->null_offset] & key_part->null_bit))
|
|
|
|
return 1;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-12-06 01:00:37 +01:00
|
|
|
if (key_part->key_part_flag & (HA_BLOB_PART | HA_VAR_LENGTH_PART))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
|
|
|
|
if (key_part->field->cmp_binary((char*) (old_row + key_part->offset),
|
2005-02-16 14:18:32 +01:00
|
|
|
(char*) (new_row + key_part->offset),
|
|
|
|
(ulong) key_part->length))
|
|
|
|
return 1;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (memcmp(old_row+key_part->offset, new_row+key_part->offset,
|
2005-02-16 14:18:32 +01:00
|
|
|
key_part->length))
|
|
|
|
return 1;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Update one record in NDB using primary key
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
|
|
|
|
{
|
|
|
|
THD *thd= current_thd;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-12-07 15:15:49 +01:00
|
|
|
NdbScanOperation* cursor= m_active_cursor;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbOperation *op;
|
|
|
|
uint i;
|
2005-07-18 13:31:02 +02:00
|
|
|
uint32 old_part_id= 0, new_part_id= 0;
|
|
|
|
int error;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("update_row");
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= TRUE;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(thd->status_var.ha_update_count, &LOCK_status);
|
2004-10-04 00:27:04 +02:00
|
|
|
if (table->timestamp_field_type & TIMESTAMP_AUTO_SET_ON_UPDATE)
|
2005-04-01 17:59:46 +02:00
|
|
|
{
|
2004-10-04 00:27:04 +02:00
|
|
|
table->timestamp_field->set_time();
|
2005-04-28 14:45:27 +02:00
|
|
|
ha_set_bit_in_write_set(table->timestamp_field->fieldnr);
|
2005-04-01 17:59:46 +02:00
|
|
|
}
|
2004-10-04 00:27:04 +02:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function &&
|
|
|
|
(error= get_parts_for_update(old_data, new_data, table->record[0],
|
|
|
|
m_part_info, &old_part_id, &new_part_id)))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
/* Check for update of primary key for special handling */
|
2005-01-06 12:00:13 +01:00
|
|
|
if ((table->s->primary_key != MAX_KEY) &&
|
2005-07-18 13:31:02 +02:00
|
|
|
(key_cmp(table->s->primary_key, old_data, new_data)) ||
|
|
|
|
(old_part_id != new_part_id))
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
2005-07-06 11:23:36 +02:00
|
|
|
int read_res, insert_res, delete_res, undo_res;
|
2004-07-22 18:35:51 +02:00
|
|
|
|
2005-04-19 11:21:26 +02:00
|
|
|
DBUG_PRINT("info", ("primary key update, doing pk read+delete+insert"));
|
2004-07-22 18:35:51 +02:00
|
|
|
// Get all old fields, since we optimize away fields not in query
|
2005-07-18 13:31:02 +02:00
|
|
|
read_res= complemented_pk_read(old_data, new_data, old_part_id);
|
2004-07-22 18:35:51 +02:00
|
|
|
if (read_res)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("pk read failed"));
|
|
|
|
DBUG_RETURN(read_res);
|
|
|
|
}
|
2004-07-23 15:46:56 +02:00
|
|
|
// Delete old row
|
2004-11-03 15:53:26 +01:00
|
|
|
m_primary_key_update= TRUE;
|
2004-07-23 15:46:56 +02:00
|
|
|
delete_res= delete_row(old_data);
|
2004-11-03 15:53:26 +01:00
|
|
|
m_primary_key_update= FALSE;
|
2004-07-23 15:46:56 +02:00
|
|
|
if (delete_res)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("delete failed"));
|
2005-04-19 11:21:26 +02:00
|
|
|
DBUG_RETURN(delete_res);
|
2004-07-23 15:46:56 +02:00
|
|
|
}
|
2005-04-19 11:21:26 +02:00
|
|
|
// Insert new row
|
|
|
|
DBUG_PRINT("info", ("delete succeded"));
|
2005-04-21 13:09:19 +02:00
|
|
|
m_primary_key_update= TRUE;
|
2005-04-19 11:21:26 +02:00
|
|
|
insert_res= write_row(new_data);
|
2005-04-21 13:09:19 +02:00
|
|
|
m_primary_key_update= FALSE;
|
2005-04-19 11:21:26 +02:00
|
|
|
if (insert_res)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("insert failed"));
|
|
|
|
if (trans->commitStatus() == NdbConnection::Started)
|
|
|
|
{
|
2005-07-06 11:23:36 +02:00
|
|
|
// Undo delete_row(old_data)
|
2005-04-19 11:21:26 +02:00
|
|
|
m_primary_key_update= TRUE;
|
2005-07-06 11:23:36 +02:00
|
|
|
undo_res= write_row((byte *)old_data);
|
|
|
|
if (undo_res)
|
|
|
|
push_warning(current_thd,
|
|
|
|
MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
|
|
undo_res,
|
|
|
|
"NDB failed undoing delete at primary key update");
|
2005-04-19 11:21:26 +02:00
|
|
|
m_primary_key_update= FALSE;
|
|
|
|
}
|
|
|
|
DBUG_RETURN(insert_res);
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("delete+insert succeeded"));
|
2004-07-23 15:46:56 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-07-22 18:35:51 +02:00
|
|
|
}
|
2004-07-09 13:18:56 +02:00
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
if (cursor)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
/*
|
|
|
|
We are scanning records and want to update the record
|
|
|
|
that was just found, call updateTuple on the cursor
|
|
|
|
to take over the lock to a new update operation
|
|
|
|
And thus setting the primary key of the record from
|
|
|
|
the active record in cursor
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("info", ("Calling updateTuple on cursor"));
|
2004-12-07 15:15:49 +01:00
|
|
|
if (!(op= cursor->updateCurrentTuple()))
|
2004-05-10 14:12:28 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending++;
|
2005-04-28 14:45:27 +02:00
|
|
|
if (uses_blob_value())
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_pending= TRUE;
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
cursor->setPartitionId(new_part_id);
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2005-02-16 14:18:32 +01:00
|
|
|
op->updateTuple() != 0)
|
2004-05-10 14:12:28 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
op->setPartitionId(new_part_id);
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key == MAX_KEY)
|
2004-05-10 14:12:28 +02:00
|
|
|
{
|
|
|
|
// This table has no primary key, use "hidden" primary key
|
|
|
|
DBUG_PRINT("info", ("Using hidden key"));
|
|
|
|
|
|
|
|
// Require that the PK for this record has previously been
|
|
|
|
// read into m_value
|
2005-01-06 12:00:13 +01:00
|
|
|
uint no_fields= table->s->fields;
|
2004-12-01 12:43:30 +01:00
|
|
|
const NdbRecAttr* rec= m_value[no_fields].rec;
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_ASSERT(rec);
|
|
|
|
DBUG_DUMP("key", (char*)rec->aRef(), NDB_HIDDEN_PRIMARY_KEY_LENGTH);
|
|
|
|
|
|
|
|
if (set_hidden_key(op, no_fields, rec->aRef()))
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
2005-06-29 16:13:50 +02:00
|
|
|
if ((res= set_primary_key_from_record(op, old_data)))
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
m_rows_changed++;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Set non-key attribute(s)
|
2005-01-06 12:00:13 +01:00
|
|
|
for (i= 0; i < table->s->fields; i++)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
2005-04-28 14:45:27 +02:00
|
|
|
if (ha_get_bit_in_write_set(i+1) &&
|
2004-04-15 09:14:14 +02:00
|
|
|
(!(field->flags & PRI_KEY_FLAG)) &&
|
2005-11-06 00:20:37 +01:00
|
|
|
set_ndb_value(op, field, i, new_data - table->record[0]))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Execute update operation
|
2004-09-17 16:58:08 +02:00
|
|
|
if (!cursor && execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Delete one record from NDB, using primary key
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::delete_row(const byte *record)
|
|
|
|
{
|
2004-09-14 21:02:23 +02:00
|
|
|
THD *thd= current_thd;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-12-07 15:15:49 +01:00
|
|
|
NdbScanOperation* cursor= m_active_cursor;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbOperation *op;
|
2005-07-18 13:31:02 +02:00
|
|
|
uint32 part_id;
|
|
|
|
int error;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("delete_row");
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= TRUE;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(thd->status_var.ha_delete_count,&LOCK_status);
|
2005-03-31 14:33:50 +02:00
|
|
|
m_rows_changed++;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function &&
|
|
|
|
(error= get_part_for_delete(record, table->record[0], m_part_info,
|
|
|
|
&part_id)))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
if (cursor)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
/*
|
2004-07-23 10:33:51 +02:00
|
|
|
We are scanning records and want to delete the record
|
2004-05-10 14:12:28 +02:00
|
|
|
that was just found, call deleteTuple on the cursor
|
2004-07-23 13:13:50 +02:00
|
|
|
to take over the lock to a new delete operation
|
2004-05-10 14:12:28 +02:00
|
|
|
And thus setting the primary key of the record from
|
|
|
|
the active record in cursor
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("info", ("Calling deleteTuple on cursor"));
|
2004-12-07 15:15:49 +01:00
|
|
|
if (cursor->deleteCurrentTuple() != 0)
|
2004-05-10 14:12:28 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending++;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
cursor->setPartitionId(part_id);
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_update(-1);
|
|
|
|
|
2005-04-19 11:21:26 +02:00
|
|
|
if (!m_primary_key_update)
|
|
|
|
// If deleting from cursor, NoCommit will be handled in next_result
|
|
|
|
DBUG_RETURN(0);
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
else
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op=trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2005-02-16 14:18:32 +01:00
|
|
|
op->deleteTuple() != 0)
|
2004-05-10 14:12:28 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
op->setPartitionId(part_id);
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_update(-1);
|
|
|
|
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key == MAX_KEY)
|
2004-05-10 14:12:28 +02:00
|
|
|
{
|
|
|
|
// This table has no primary key, use "hidden" primary key
|
|
|
|
DBUG_PRINT("info", ("Using hidden key"));
|
2005-01-06 12:00:13 +01:00
|
|
|
uint no_fields= table->s->fields;
|
2004-12-01 12:43:30 +01:00
|
|
|
const NdbRecAttr* rec= m_value[no_fields].rec;
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_ASSERT(rec != NULL);
|
|
|
|
|
|
|
|
if (set_hidden_key(op, no_fields, rec->aRef()))
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
2005-06-29 16:13:50 +02:00
|
|
|
if ((res= set_primary_key_from_record(op, record)))
|
|
|
|
return res;
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Execute delete operation
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Unpack a record read from NDB
|
|
|
|
|
|
|
|
SYNOPSIS
|
|
|
|
unpack_record()
|
2005-02-16 14:18:32 +01:00
|
|
|
buf Buffer to store read row
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
NOTE
|
|
|
|
The data for each row is read directly into the
|
|
|
|
destination buffer. This function is primarily
|
|
|
|
called in order to check if any fields should be
|
|
|
|
set to null.
|
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
static void ndb_unpack_record(TABLE *table, NdbValue *value,
|
|
|
|
MY_BITMAP *defined, byte *buf)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
Field **p_field= table->field, *field= *p_field;
|
2004-04-15 09:14:14 +02:00
|
|
|
uint row_offset= (uint) (buf - table->record[0]);
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_ENTER("ndb_unpack_record");
|
2004-12-01 12:43:30 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Set null flag(s)
|
2005-01-06 12:00:13 +01:00
|
|
|
bzero(buf, table->s->null_bytes);
|
2005-11-06 00:20:37 +01:00
|
|
|
for ( ; field;
|
|
|
|
p_field++, value++, field= *p_field)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
if ((*value).ptr)
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (!(field->flags & BLOB_FLAG))
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
int is_null= (*value).rec->isNULL();
|
|
|
|
if (is_null)
|
2004-12-23 15:28:41 +01:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (is_null > 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info",("[%u] NULL",
|
|
|
|
(*value).rec->getColumn()->getColumnNo()));
|
|
|
|
field->set_null(row_offset);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info",("[%u] UNDEFINED",
|
|
|
|
(*value).rec->getColumn()->getColumnNo()));
|
|
|
|
bitmap_clear_bit(defined,
|
|
|
|
(*value).rec->getColumn()->getColumnNo());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (field->type() == MYSQL_TYPE_BIT)
|
|
|
|
{
|
|
|
|
byte *save_field_ptr= field->ptr;
|
|
|
|
field->ptr= save_field_ptr + row_offset;
|
|
|
|
if (field->pack_length() < 5)
|
2004-12-23 15:28:41 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("bit field H'%.8X",
|
2005-02-16 14:18:32 +01:00
|
|
|
(*value).rec->u_32_value()));
|
2005-11-06 00:20:37 +01:00
|
|
|
((Field_bit*) field)->store((longlong)
|
|
|
|
(*value).rec->u_32_value(), FALSE);
|
2004-12-23 15:28:41 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("bit field H'%.8X%.8X",
|
2005-11-06 00:20:37 +01:00
|
|
|
*(Uint32*) (*value).rec->aRef(),
|
|
|
|
*((Uint32*) (*value).rec->aRef()+1)));
|
|
|
|
((Field_bit*) field)->store((longlong)
|
|
|
|
(*value).rec->u_64_value(),TRUE);
|
Added option --valgrind-mysqltest to mysql-test-run
Added flag to Field::store(longlong) to specify if value is unsigned.
This fixes bug #12750: Incorrect storage of 9999999999999999999 in DECIMAL(19, 0)
Fixed warning from valgrind in CREATE ... SELECT
Fixed double free of mysql.options if reconnect failed
mysql-test/mysql-test-run.sh:
Added option --valgrind-mysqltest to allow one to run mysqltest with valgrind
mysql-test/r/bigint.result:
Update results after fix for Field::store(longlong)
mysql-test/r/range.result:
Update results after fix for Field::store(longlong)
mysql-test/r/strict.result:
Update results after fix for Field::store(longlong)
(This fixes some wrong results when storing things into bigint columns)
mysql-test/r/type_ranges.result:
Update results after fix for Field::store(longlong)
mysql-test/t/bigint.test:
Added testing for #12750: Incorrect storage of 9999999999999999999 in DECIMAL(19, 0)
mysql-test/t/innodb.test:
Removed comments affected by this bug fix
mysql-test/t/mysqldump.test:
Fixed result to not depend on existing config files
mysql-test/t/range.test:
0xff numbers are now unsigned
mysql-test/t/strict.test:
Added errors for things that previously (wrongly) succeeded
sql-common/client.c:
Fixed double free of mysql.options if reconnect failed
sql/field.cc:
Added flag to Field::store(longlong) to specify if value is unsigned
sql/field.h:
Added flag to Field::store(longlong) to specify if value is unsigned
sql/field_conv.cc:
Fixed calls to Field::store(longlong,flag)
sql/ha_ndbcluster.cc:
Fixed calls to Field::store(longlong,flag)
sql/handler.cc:
Fixed calls to Field::store(longlong,flag)
sql/item.cc:
Fixed calls to Field::store(longlong,flag)
sql/item_sum.cc:
Fixed calls to Field::store(longlong,flag)
sql/sp.cc:
Fixed calls to Field::store(longlong,flag)
sql/sql_acl.cc:
Fixed calls to Field::store(longlong,flag)
sql/sql_help.cc:
Fixed calls to Field::store(longlong,flag)
sql/sql_show.cc:
Fixed calls to Field::store(longlong,flag)
sql/sql_table.cc:
Fixed varning from valgrind
sql/sql_udf.cc:
Fixed calls to Field::store(longlong,flag)
sql/tztime.cc:
Fixed calls to Field::store(longlong,flag)
sql/unireg.cc:
Fixed calls to Field::store(longlong,flag)
2005-09-14 00:41:44 +02:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
field->ptr= save_field_ptr;
|
|
|
|
DBUG_PRINT("info",("[%u] SET",
|
|
|
|
(*value).rec->getColumn()->getColumnNo()));
|
|
|
|
DBUG_DUMP("info", (const char*) field->ptr, field->field_length);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info",("[%u] SET",
|
|
|
|
(*value).rec->getColumn()->getColumnNo()));
|
|
|
|
DBUG_DUMP("info", (const char*) field->ptr, field->field_length);
|
2004-12-23 15:28:41 +01:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
NdbBlob *ndb_blob= (*value).blob;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool isNull= TRUE;
|
2005-01-11 17:00:31 +01:00
|
|
|
#ifndef DBUG_OFF
|
|
|
|
int ret=
|
|
|
|
#endif
|
2005-02-16 14:18:32 +01:00
|
|
|
ndb_blob->getNull(isNull);
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_ASSERT(ret == 0);
|
|
|
|
if (isNull)
|
2005-11-06 00:20:37 +01:00
|
|
|
field->set_null(row_offset);
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ha_ndbcluster::unpack_record(byte *buf)
|
|
|
|
{
|
|
|
|
ndb_unpack_record(table, m_value, 0, buf);
|
2004-04-15 09:14:14 +02:00
|
|
|
#ifndef DBUG_OFF
|
|
|
|
// Read and print all values that was fetched
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key == MAX_KEY)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
// Table with hidden primary key
|
2005-01-06 12:00:13 +01:00
|
|
|
int hidden_no= table->s->fields;
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
2004-04-15 09:14:14 +02:00
|
|
|
const NDBCOL *hidden_col= tab->getColumn(hidden_no);
|
2004-12-01 12:43:30 +01:00
|
|
|
const NdbRecAttr* rec= m_value[hidden_no].rec;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(rec);
|
|
|
|
DBUG_PRINT("hidden", ("%d: %s \"%llu\"", hidden_no,
|
|
|
|
hidden_col->getName(), rec->u_64_value()));
|
|
|
|
}
|
2005-01-17 08:08:51 +01:00
|
|
|
//print_results();
|
2004-04-15 09:14:14 +02:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Utility function to print/dump the fetched field
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::print_results()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("print_results");
|
|
|
|
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
if (!_db_on_)
|
|
|
|
DBUG_VOID_RETURN;
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-01-11 17:00:31 +01:00
|
|
|
char buf_type[MAX_FIELD_WIDTH], buf_val[MAX_FIELD_WIDTH];
|
2005-02-11 22:33:52 +01:00
|
|
|
String type(buf_type, sizeof(buf_type), &my_charset_bin);
|
2005-01-11 17:00:31 +01:00
|
|
|
String val(buf_val, sizeof(buf_val), &my_charset_bin);
|
2005-02-17 14:22:44 +01:00
|
|
|
for (uint f= 0; f < table->s->fields; f++)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-02-11 22:33:52 +01:00
|
|
|
/* Use DBUG_PRINT since DBUG_FILE cannot be filtered out */
|
2004-12-17 09:01:54 +01:00
|
|
|
char buf[2000];
|
2004-04-15 09:14:14 +02:00
|
|
|
Field *field;
|
2004-12-20 15:12:42 +01:00
|
|
|
void* ptr;
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbValue value;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-01-11 17:00:31 +01:00
|
|
|
buf[0]= 0;
|
2005-02-11 22:33:52 +01:00
|
|
|
field= table->field[f];
|
2004-07-22 12:38:09 +02:00
|
|
|
if (!(value= m_value[f]).ptr)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
strmov(buf, "not read");
|
2004-12-17 09:01:54 +01:00
|
|
|
goto print_value;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-01-11 17:00:31 +01:00
|
|
|
|
2004-12-20 15:12:42 +01:00
|
|
|
ptr= field->ptr;
|
2004-07-22 12:38:09 +02:00
|
|
|
|
|
|
|
if (! (field->flags & BLOB_FLAG))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
if (value.rec->isNULL())
|
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
strmov(buf, "NULL");
|
2004-12-17 09:01:54 +01:00
|
|
|
goto print_value;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
2005-01-11 17:00:31 +01:00
|
|
|
type.length(0);
|
|
|
|
val.length(0);
|
|
|
|
field->sql_type(type);
|
|
|
|
field->val_str(&val);
|
|
|
|
my_snprintf(buf, sizeof(buf), "%s %s", type.c_ptr(), val.c_ptr());
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-02-17 14:22:44 +01:00
|
|
|
NdbBlob *ndb_blob= value.blob;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool isNull= TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
ndb_blob->getNull(isNull);
|
2005-07-04 02:42:33 +02:00
|
|
|
if (isNull)
|
|
|
|
strmov(buf, "NULL");
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2004-12-17 09:01:54 +01:00
|
|
|
print_value:
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_PRINT("value", ("%u,%s: %s", f, field->field_name, buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
int ha_ndbcluster::index_init(uint index, bool sorted)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_init");
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_PRINT("enter", ("index: %u sorted: %d", index, sorted));
|
|
|
|
active_index= index;
|
|
|
|
m_sorted= sorted;
|
|
|
|
DBUG_RETURN(0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_end()
|
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_end");
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_RETURN(close_scan());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-11-30 07:41:26 +01:00
|
|
|
/**
|
|
|
|
* Check if key contains null
|
|
|
|
*/
|
|
|
|
static
|
|
|
|
int
|
|
|
|
check_null_in_key(const KEY* key_info, const byte *key, uint key_len)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO *curr_part, *end_part;
|
2004-12-07 15:15:49 +01:00
|
|
|
const byte* end_ptr= key + key_len;
|
2004-11-30 07:41:26 +01:00
|
|
|
curr_part= key_info->key_part;
|
|
|
|
end_part= curr_part + key_info->key_parts;
|
|
|
|
|
|
|
|
|
|
|
|
for (; curr_part != end_part && key < end_ptr; curr_part++)
|
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
if (curr_part->null_bit && *key)
|
2004-11-30 07:41:26 +01:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
key += curr_part->store_length;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::index_read(byte *buf,
|
2005-02-16 14:18:32 +01:00
|
|
|
const byte *key, uint key_len,
|
|
|
|
enum ha_rkey_function find_flag)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
key_range start_key;
|
|
|
|
bool descending= FALSE;
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_read");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("active_index: %u, key_len: %u, find_flag: %d",
|
|
|
|
active_index, key_len, find_flag));
|
|
|
|
|
2004-12-07 15:15:49 +01:00
|
|
|
start_key.key= key;
|
|
|
|
start_key.length= key_len;
|
|
|
|
start_key.flag= find_flag;
|
2005-07-18 13:31:02 +02:00
|
|
|
descending= FALSE;
|
2004-12-20 15:12:42 +01:00
|
|
|
switch (find_flag) {
|
|
|
|
case HA_READ_KEY_OR_PREV:
|
|
|
|
case HA_READ_BEFORE_KEY:
|
|
|
|
case HA_READ_PREFIX_LAST:
|
|
|
|
case HA_READ_PREFIX_LAST_OR_PREV:
|
|
|
|
descending= TRUE;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(read_range_first_to_buf(&start_key, 0, descending,
|
|
|
|
m_sorted, buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_read_idx(byte *buf, uint index_no,
|
2005-02-16 14:18:32 +01:00
|
|
|
const byte *key, uint key_len,
|
|
|
|
enum ha_rkey_function find_flag)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_key_count, &LOCK_status);
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_read_idx");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("index_no: %u, key_len: %u", index_no, key_len));
|
2005-07-18 13:31:02 +02:00
|
|
|
index_init(index_no, 0);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(index_read(buf, key, key_len, find_flag));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_next(byte *buf)
|
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_next");
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_next_count,
|
2005-02-16 14:18:32 +01:00
|
|
|
&LOCK_status);
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(next_result(buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_prev(byte *buf)
|
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_prev");
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_prev_count,
|
2005-02-16 14:18:32 +01:00
|
|
|
&LOCK_status);
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_RETURN(next_result(buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_first(byte *buf)
|
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_first");
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_first_count,
|
2005-02-16 14:18:32 +01:00
|
|
|
&LOCK_status);
|
2004-09-03 13:55:40 +02:00
|
|
|
// Start the ordered index scan and fetch the first row
|
|
|
|
|
|
|
|
// Only HA_READ_ORDER indexes get called by index_first
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(ordered_index_scan(0, 0, TRUE, FALSE, buf, NULL));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_last(byte *buf)
|
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::index_last");
|
2004-09-13 15:48:01 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_last_count,&LOCK_status);
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(ordered_index_scan(0, 0, TRUE, TRUE, buf, NULL));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-12-20 15:12:42 +01:00
|
|
|
int ha_ndbcluster::index_read_last(byte * buf, const byte * key, uint key_len)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::index_read_last");
|
|
|
|
DBUG_RETURN(index_read(buf, key, key_len, HA_READ_PREFIX_LAST));
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-07 16:22:42 +02:00
|
|
|
int ha_ndbcluster::read_range_first_to_buf(const key_range *start_key,
|
2005-02-16 14:18:32 +01:00
|
|
|
const key_range *end_key,
|
2005-07-18 13:31:02 +02:00
|
|
|
bool desc, bool sorted,
|
2005-02-16 14:18:32 +01:00
|
|
|
byte* buf)
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
part_id_range part_spec;
|
|
|
|
ndb_index_type type= get_index_type(active_index);
|
|
|
|
const KEY* key_info= table->key_info+active_index;
|
|
|
|
int error;
|
2004-09-07 16:22:42 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::read_range_first_to_buf");
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_PRINT("info", ("desc: %d, sorted: %d", desc, sorted));
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
if (m_use_partition_function)
|
|
|
|
{
|
|
|
|
get_partition_set(table, buf, active_index, start_key, &part_spec);
|
|
|
|
if (part_spec.start_part > part_spec.end_part)
|
|
|
|
{
|
|
|
|
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
|
|
|
}
|
|
|
|
else if (part_spec.start_part == part_spec.end_part)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Only one partition is required to scan, if sorted is required we
|
|
|
|
don't need it any more since output from one ordered partitioned
|
|
|
|
index is always sorted.
|
|
|
|
*/
|
|
|
|
sorted= FALSE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
m_write_op= FALSE;
|
|
|
|
switch (type){
|
2004-05-24 12:35:39 +02:00
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
2004-04-30 13:38:41 +02:00
|
|
|
case PRIMARY_KEY_INDEX:
|
2004-05-10 14:46:06 +02:00
|
|
|
if (start_key &&
|
2005-02-16 14:18:32 +01:00
|
|
|
start_key->length == key_info->key_length &&
|
|
|
|
start_key->flag == HA_READ_KEY_EXACT)
|
2004-05-17 09:51:02 +02:00
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
if (m_active_cursor && (error= close_scan()))
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(error);
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(pk_read(start_key->key, start_key->length, buf,
|
|
|
|
part_spec.start_part));
|
2004-05-17 09:51:02 +02:00
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
break;
|
2004-05-24 12:35:39 +02:00
|
|
|
case UNIQUE_ORDERED_INDEX:
|
2004-04-30 13:38:41 +02:00
|
|
|
case UNIQUE_INDEX:
|
2004-11-30 07:41:26 +01:00
|
|
|
if (start_key && start_key->length == key_info->key_length &&
|
2005-02-16 14:18:32 +01:00
|
|
|
start_key->flag == HA_READ_KEY_EXACT &&
|
|
|
|
!check_null_in_key(key_info, start_key->key, start_key->length))
|
2004-05-17 09:51:02 +02:00
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
if (m_active_cursor && (error= close_scan()))
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(error);
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(unique_index_read(start_key->key, start_key->length, buf));
|
2004-05-17 09:51:02 +02:00
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-05-10 14:46:06 +02:00
|
|
|
// Start the ordered index scan and fetch the first row
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(ordered_index_scan(start_key, end_key, sorted, desc, buf,
|
|
|
|
&part_spec));
|
2004-04-30 13:38:41 +02:00
|
|
|
}
|
|
|
|
|
2004-10-21 12:39:08 +02:00
|
|
|
int ha_ndbcluster::read_range_first(const key_range *start_key,
|
2005-02-16 14:18:32 +01:00
|
|
|
const key_range *end_key,
|
|
|
|
bool eq_r, bool sorted)
|
2004-10-21 12:39:08 +02:00
|
|
|
{
|
|
|
|
byte* buf= table->record[0];
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_range_first");
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_RETURN(read_range_first_to_buf(start_key, end_key, FALSE,
|
|
|
|
sorted, buf));
|
2004-10-21 12:39:08 +02:00
|
|
|
}
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
int ha_ndbcluster::read_range_next()
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_range_next");
|
|
|
|
DBUG_RETURN(next_result(table->record[0]));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::rnd_init(bool scan)
|
|
|
|
{
|
2004-12-07 15:15:49 +01:00
|
|
|
NdbScanOperation *cursor= m_active_cursor;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("rnd_init");
|
|
|
|
DBUG_PRINT("enter", ("scan: %d", scan));
|
2004-07-09 13:18:56 +02:00
|
|
|
// Check if scan is to be restarted
|
2004-08-24 23:13:32 +02:00
|
|
|
if (cursor)
|
|
|
|
{
|
|
|
|
if (!scan)
|
|
|
|
DBUG_RETURN(1);
|
2005-07-04 02:42:33 +02:00
|
|
|
if (cursor->restart(m_force_send) != 0)
|
2005-01-11 17:00:31 +01:00
|
|
|
{
|
|
|
|
DBUG_ASSERT(0);
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
2004-08-24 23:13:32 +02:00
|
|
|
}
|
2005-07-18 13:31:02 +02:00
|
|
|
index_init(table->s->primary_key, 0);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-05-11 13:59:22 +02:00
|
|
|
int ha_ndbcluster::close_scan()
|
|
|
|
{
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_ENTER("close_scan");
|
|
|
|
|
2004-12-08 00:36:40 +01:00
|
|
|
m_multi_cursor= 0;
|
|
|
|
if (!m_active_cursor && !m_multi_cursor)
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_RETURN(1);
|
|
|
|
|
2004-12-08 00:36:40 +01:00
|
|
|
NdbScanOperation *cursor= m_active_cursor ? m_active_cursor : m_multi_cursor;
|
2004-08-04 10:54:42 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_ops_pending)
|
2004-08-04 10:54:42 +02:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
Take over any pending transactions to the
|
|
|
|
deleteing/updating transaction before closing the scan
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
DBUG_PRINT("info", ("ops_pending: %d", m_ops_pending));
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-08-04 10:54:42 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-08-04 10:54:42 +02:00
|
|
|
}
|
|
|
|
|
2005-05-06 10:39:30 +02:00
|
|
|
cursor->close(m_force_send, TRUE);
|
2004-12-08 00:36:40 +01:00
|
|
|
m_active_cursor= m_multi_cursor= NULL;
|
2004-05-28 11:23:44 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-05-11 13:59:22 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::rnd_end()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("rnd_end");
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_RETURN(close_scan());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::rnd_next(byte *buf)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("rnd_next");
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_rnd_next_count,
|
2005-02-16 14:18:32 +01:00
|
|
|
&LOCK_status);
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (!m_active_cursor)
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(full_table_scan(buf));
|
|
|
|
DBUG_RETURN(next_result(buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
An "interesting" record has been found and it's pk
|
|
|
|
retrieved by calling position
|
|
|
|
Now it's time to read the record from db once
|
|
|
|
again
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::rnd_pos(byte *buf, byte *pos)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("rnd_pos");
|
2004-09-14 21:02:23 +02:00
|
|
|
statistic_increment(current_thd->status_var.ha_read_rnd_count,
|
2005-02-16 14:18:32 +01:00
|
|
|
&LOCK_status);
|
2004-04-15 09:14:14 +02:00
|
|
|
// The primary key for the record is stored in pos
|
|
|
|
// Perform a pk_read using primary key "index"
|
2005-07-18 13:31:02 +02:00
|
|
|
{
|
|
|
|
part_id_range part_spec;
|
|
|
|
if (m_use_partition_function)
|
|
|
|
{
|
|
|
|
key_range key_spec;
|
|
|
|
KEY *key_info= table->key_info + active_index;
|
|
|
|
key_spec.key= pos;
|
|
|
|
key_spec.length= ref_length;
|
|
|
|
key_spec.flag= HA_READ_KEY_EXACT;
|
|
|
|
get_full_part_id_from_key(table, buf, key_info, &key_spec, &part_spec);
|
|
|
|
DBUG_ASSERT(part_spec.start_part == part_spec.end_part);
|
|
|
|
}
|
|
|
|
DBUG_RETURN(pk_read(pos, ref_length, buf, part_spec.start_part));
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Store the primary key of this record in ref
|
|
|
|
variable, so that the row can be retrieved again later
|
|
|
|
using "reference" in rnd_pos
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::position(const byte *record)
|
|
|
|
{
|
|
|
|
KEY *key_info;
|
|
|
|
KEY_PART_INFO *key_part;
|
|
|
|
KEY_PART_INFO *end;
|
|
|
|
byte *buff;
|
|
|
|
DBUG_ENTER("position");
|
|
|
|
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key != MAX_KEY)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
key_info= table->key_info + table->s->primary_key;
|
2004-04-15 09:14:14 +02:00
|
|
|
key_part= key_info->key_part;
|
|
|
|
end= key_part + key_info->key_parts;
|
|
|
|
buff= ref;
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
if (key_part->null_bit) {
|
|
|
|
/* Store 0 if the key part is a NULL part */
|
|
|
|
if (record[key_part->null_offset]
|
|
|
|
& key_part->null_bit) {
|
|
|
|
*buff++= 1;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
*buff++= 0;
|
|
|
|
}
|
2005-12-20 20:53:29 +01:00
|
|
|
|
|
|
|
size_t len = key_part->length;
|
|
|
|
const byte * ptr = record + key_part->offset;
|
|
|
|
Field *field = key_part->field;
|
|
|
|
if ((field->type() == MYSQL_TYPE_VARCHAR) &&
|
|
|
|
((Field_varstring*)field)->length_bytes == 1)
|
|
|
|
{
|
|
|
|
/**
|
|
|
|
* Keys always use 2 bytes length
|
|
|
|
*/
|
|
|
|
buff[0] = ptr[0];
|
|
|
|
buff[1] = 0;
|
|
|
|
memcpy(buff+2, ptr + 1, len);
|
|
|
|
len += 2;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
memcpy(buff, ptr, len);
|
|
|
|
}
|
|
|
|
buff += len;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// No primary key, get hidden key
|
|
|
|
DBUG_PRINT("info", ("Getting hidden key"));
|
2005-01-06 12:00:13 +01:00
|
|
|
int hidden_no= table->s->fields;
|
2004-12-01 12:43:30 +01:00
|
|
|
const NdbRecAttr* rec= m_value[hidden_no].rec;
|
2005-01-11 17:00:31 +01:00
|
|
|
memcpy(ref, (const void*)rec->aRef(), ref_length);
|
|
|
|
#ifndef DBUG_OFF
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
2004-04-15 09:14:14 +02:00
|
|
|
const NDBCOL *hidden_col= tab->getColumn(hidden_no);
|
|
|
|
DBUG_ASSERT(hidden_col->getPrimaryKey() &&
|
|
|
|
hidden_col->getAutoIncrement() &&
|
|
|
|
rec != NULL &&
|
|
|
|
ref_length == NDB_HIDDEN_PRIMARY_KEY_LENGTH);
|
2005-01-11 17:00:31 +01:00
|
|
|
#endif
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_DUMP("ref", (char*)ref, ref_length);
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ha_ndbcluster::info(uint flag)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("info");
|
|
|
|
DBUG_PRINT("enter", ("flag: %d", flag));
|
|
|
|
|
|
|
|
if (flag & HA_STATUS_POS)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_POS"));
|
|
|
|
if (flag & HA_STATUS_NO_LOCK)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_NO_LOCK"));
|
|
|
|
if (flag & HA_STATUS_TIME)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_TIME"));
|
|
|
|
if (flag & HA_STATUS_VARIABLE)
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("info", ("HA_STATUS_VARIABLE"));
|
2004-09-20 17:52:19 +02:00
|
|
|
if (m_table_info)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
2005-02-16 14:18:32 +01:00
|
|
|
records= 100;
|
2004-11-17 09:15:53 +01:00
|
|
|
else
|
2005-02-16 14:18:32 +01:00
|
|
|
records_update();
|
2004-09-20 17:52:19 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
if ((my_errno= check_ndb_connection()))
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
Ndb *ndb= get_ndb();
|
2005-01-13 18:24:19 +01:00
|
|
|
struct Ndb_statistics stat;
|
|
|
|
if (current_thd->variables.ndb_use_exact_count &&
|
2005-02-16 14:18:32 +01:00
|
|
|
ndb_get_table_statistics(ndb, m_tabname, &stat) == 0)
|
2005-01-13 18:24:19 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
mean_rec_length= stat.row_size;
|
|
|
|
data_file_length= stat.fragment_memory;
|
|
|
|
records= stat.row_count;
|
2005-01-13 18:24:19 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
mean_rec_length= 0;
|
|
|
|
records= 100;
|
2005-01-13 18:24:19 +01:00
|
|
|
}
|
2004-09-20 17:52:19 +02:00
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
2004-10-04 01:20:05 +02:00
|
|
|
if (flag & HA_STATUS_CONST)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_CONST"));
|
|
|
|
set_rec_per_key();
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
if (flag & HA_STATUS_ERRKEY)
|
2004-08-31 12:07:48 +02:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("info", ("HA_STATUS_ERRKEY"));
|
2004-11-03 15:53:26 +01:00
|
|
|
errkey= m_dupkey;
|
2004-08-31 12:07:48 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
if (flag & HA_STATUS_AUTO)
|
2005-08-10 09:38:36 +02:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("info", ("HA_STATUS_AUTO"));
|
2005-08-10 09:38:36 +02:00
|
|
|
if (m_table)
|
|
|
|
{
|
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
|
|
|
|
auto_increment_value=
|
|
|
|
ndb->readAutoIncrementValue((const NDBTAB *) m_table);
|
|
|
|
}
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::extra(enum ha_extra_function operation)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("extra");
|
|
|
|
switch (operation) {
|
|
|
|
case HA_EXTRA_RESET: /* Reset database to after open */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_RESET"));
|
2004-12-17 21:13:22 +01:00
|
|
|
DBUG_PRINT("info", ("Clearing condition stack"));
|
|
|
|
cond_clear();
|
2005-12-22 10:29:00 +01:00
|
|
|
if (m_part_info)
|
|
|
|
bitmap_clear_all(&m_part_info->used_partitions);
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case HA_EXTRA_IGNORE_DUP_KEY: /* Dup keys don't rollback everything*/
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_IGNORE_DUP_KEY"));
|
2004-09-15 14:44:21 +02:00
|
|
|
if (current_thd->lex->sql_command == SQLCOM_REPLACE)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Turning ON use of write instead of insert"));
|
|
|
|
m_use_write= TRUE;
|
|
|
|
} else
|
|
|
|
{
|
2004-11-18 12:11:56 +01:00
|
|
|
DBUG_PRINT("info", ("Ignoring duplicate key"));
|
|
|
|
m_ignore_dup_key= TRUE;
|
2004-09-15 14:44:21 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_IGNORE_DUP_KEY:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_IGNORE_DUP_KEY"));
|
|
|
|
DBUG_PRINT("info", ("Turning OFF use of write instead of insert"));
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
m_use_write= FALSE;
|
2004-11-18 12:11:56 +01:00
|
|
|
m_ignore_dup_key= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
2005-07-18 13:31:02 +02:00
|
|
|
default:
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-04-29 14:38:35 +02:00
|
|
|
/*
|
|
|
|
Start of an insert, remember number of rows to be inserted, it will
|
|
|
|
be used in write_row and get_autoincrement to send an optimal number
|
|
|
|
of rows in each roundtrip to the server
|
|
|
|
|
|
|
|
SYNOPSIS
|
|
|
|
rows number of rows to insert, 0 if unknown
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::start_bulk_insert(ha_rows rows)
|
|
|
|
{
|
|
|
|
int bytes, batch;
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
2004-04-29 14:38:35 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("start_bulk_insert");
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("enter", ("rows: %d", (int)rows));
|
2004-04-29 14:38:35 +02:00
|
|
|
|
2005-04-13 12:42:23 +02:00
|
|
|
m_rows_inserted= (ha_rows) 0;
|
|
|
|
if (rows == (ha_rows) 0)
|
2005-05-07 22:04:52 +02:00
|
|
|
{
|
2005-04-07 17:44:09 +02:00
|
|
|
/* We don't know how many will be inserted, guess */
|
|
|
|
m_rows_to_insert= m_autoincrement_prefetch;
|
2005-05-07 22:04:52 +02:00
|
|
|
}
|
2005-04-07 17:44:09 +02:00
|
|
|
else
|
|
|
|
m_rows_to_insert= rows;
|
2004-04-29 14:38:35 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
Calculate how many rows that should be inserted
|
|
|
|
per roundtrip to NDB. This is done in order to minimize the
|
|
|
|
number of roundtrips as much as possible. However performance will
|
|
|
|
degrade if too many bytes are inserted, thus it's limited by this
|
|
|
|
calculation.
|
|
|
|
*/
|
2004-07-23 15:46:56 +02:00
|
|
|
const int bytesperbatch= 8192;
|
2004-04-29 14:38:35 +02:00
|
|
|
bytes= 12 + tab->getRowSizeInBytes() + 4 * tab->getNoOfColumns();
|
2004-05-10 14:46:06 +02:00
|
|
|
batch= bytesperbatch/bytes;
|
2004-04-29 14:38:35 +02:00
|
|
|
batch= batch == 0 ? 1 : batch;
|
|
|
|
DBUG_PRINT("info", ("batch: %d, bytes: %d", batch, bytes));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_rows= batch;
|
2004-04-29 14:38:35 +02:00
|
|
|
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
End of an insert
|
|
|
|
*/
|
|
|
|
int ha_ndbcluster::end_bulk_insert()
|
|
|
|
{
|
2004-07-23 15:46:56 +02:00
|
|
|
int error= 0;
|
|
|
|
|
2004-04-29 14:38:35 +02:00
|
|
|
DBUG_ENTER("end_bulk_insert");
|
2004-07-23 15:46:56 +02:00
|
|
|
// Check if last inserts need to be flushed
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_bulk_insert_not_flushed)
|
2004-07-23 15:46:56 +02:00
|
|
|
{
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction *trans= m_active_trans;
|
2004-07-23 15:46:56 +02:00
|
|
|
// Send rows to NDB
|
|
|
|
DBUG_PRINT("info", ("Sending inserts to NDB, "\
|
|
|
|
"rows_inserted:%d, bulk_insert_rows: %d",
|
2005-04-13 12:42:23 +02:00
|
|
|
(int) m_rows_inserted, (int) m_bulk_insert_rows));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_not_flushed= FALSE;
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-09-07 12:45:19 +02:00
|
|
|
my_errno= error= ndb_err(trans);
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-07-23 15:46:56 +02:00
|
|
|
}
|
|
|
|
|
2005-04-13 12:42:23 +02:00
|
|
|
m_rows_inserted= (ha_rows) 0;
|
|
|
|
m_rows_to_insert= (ha_rows) 1;
|
2004-07-23 15:46:56 +02:00
|
|
|
DBUG_RETURN(error);
|
2004-04-29 14:38:35 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::extra_opt(enum ha_extra_function operation, ulong cache_size)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("extra_opt");
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("enter", ("cache_size: %lu", cache_size));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(extra(operation));
|
|
|
|
}
|
|
|
|
|
2005-04-27 11:25:08 +02:00
|
|
|
static const char *ha_ndbcluster_exts[] = {
|
|
|
|
ha_ndb_ext,
|
|
|
|
NullS
|
|
|
|
};
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-03-21 12:29:08 +01:00
|
|
|
const char** ha_ndbcluster::bas_ext() const
|
2005-04-27 11:25:08 +02:00
|
|
|
{
|
|
|
|
return ha_ndbcluster_exts;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
How many seeks it will take to read through the table
|
|
|
|
This is to be comparable to the number returned by records_in_range so
|
|
|
|
that we can decide if we should scan the table or use keys.
|
|
|
|
*/
|
|
|
|
|
|
|
|
double ha_ndbcluster::scan_time()
|
|
|
|
{
|
2004-09-03 13:55:40 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::scan_time()");
|
|
|
|
double res= rows2double(records*1000);
|
|
|
|
DBUG_PRINT("exit", ("table: %s value: %f",
|
2005-02-16 14:18:32 +01:00
|
|
|
m_tabname, res));
|
2004-09-03 13:55:40 +02:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-06-20 17:16:22 +02:00
|
|
|
/*
|
|
|
|
Convert MySQL table locks into locks supported by Ndb Cluster.
|
|
|
|
Note that MySQL Cluster does currently not support distributed
|
|
|
|
table locks, so to be safe one should set cluster in Single
|
|
|
|
User Mode, before relying on table locks when updating tables
|
|
|
|
from several MySQL servers
|
|
|
|
*/
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
THR_LOCK_DATA **ha_ndbcluster::store_lock(THD *thd,
|
|
|
|
THR_LOCK_DATA **to,
|
|
|
|
enum thr_lock_type lock_type)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("store_lock");
|
|
|
|
if (lock_type != TL_IGNORE && m_lock.type == TL_UNLOCK)
|
|
|
|
{
|
2004-09-28 21:11:50 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/* If we are not doing a LOCK TABLE, then allow multiple
|
|
|
|
writers */
|
|
|
|
|
2004-10-13 10:08:18 +02:00
|
|
|
/* Since NDB does not currently have table locks
|
|
|
|
this is treated as a ordinary lock */
|
|
|
|
|
2005-06-09 13:28:29 +02:00
|
|
|
if ((lock_type >= TL_WRITE_CONCURRENT_INSERT &&
|
2004-04-15 09:14:14 +02:00
|
|
|
lock_type <= TL_WRITE) && !thd->in_lock_tables)
|
|
|
|
lock_type= TL_WRITE_ALLOW_WRITE;
|
|
|
|
|
|
|
|
/* In queries of type INSERT INTO t1 SELECT ... FROM t2 ...
|
|
|
|
MySQL would use the lock TL_READ_NO_INSERT on t2, and that
|
|
|
|
would conflict with TL_WRITE_ALLOW_WRITE, blocking all inserts
|
|
|
|
to t2. Convert the lock to a normal read lock to allow
|
|
|
|
concurrent inserts to t2. */
|
|
|
|
|
|
|
|
if (lock_type == TL_READ_NO_INSERT && !thd->in_lock_tables)
|
|
|
|
lock_type= TL_READ;
|
|
|
|
|
|
|
|
m_lock.type=lock_type;
|
|
|
|
}
|
|
|
|
*to++= &m_lock;
|
2004-05-10 14:12:28 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("exit", ("lock_type: %d", lock_type));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(to);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
#define PRINT_OPTION_FLAGS(t) { \
|
|
|
|
if (t->options & OPTION_NOT_AUTOCOMMIT) \
|
|
|
|
DBUG_PRINT("thd->options", ("OPTION_NOT_AUTOCOMMIT")); \
|
|
|
|
if (t->options & OPTION_BEGIN) \
|
|
|
|
DBUG_PRINT("thd->options", ("OPTION_BEGIN")); \
|
|
|
|
if (t->options & OPTION_TABLE_LOCK) \
|
|
|
|
DBUG_PRINT("thd->options", ("OPTION_TABLE_LOCK")); \
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define PRINT_OPTION_FLAGS(t)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
As MySQL will execute an external lock for every new table it uses
|
|
|
|
we can use this to start the transactions.
|
|
|
|
If we are in auto_commit mode we just need to start a transaction
|
2005-02-17 22:52:40 +01:00
|
|
|
for the statement, this will be stored in thd_ndb.stmt.
|
2004-04-15 09:14:14 +02:00
|
|
|
If not, we have to start a master transaction if there doesn't exist
|
2005-02-17 22:52:40 +01:00
|
|
|
one from before, this will be stored in thd_ndb.all
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
When a table lock is held one transaction will be started which holds
|
|
|
|
the table lock and for each statement a hupp transaction will be started
|
2005-03-22 17:48:28 +01:00
|
|
|
If we are locking the table then:
|
2005-03-22 17:42:08 +01:00
|
|
|
- save the NdbDictionary::Table for easy access
|
2005-03-22 17:48:28 +01:00
|
|
|
- save reference to table statistics
|
|
|
|
- refresh list of the indexes for the table if needed (if altered)
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::external_lock(THD *thd, int lock_type)
|
|
|
|
{
|
|
|
|
int error=0;
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction* trans= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("external_lock");
|
|
|
|
/*
|
|
|
|
Check that this handler instance has a connection
|
|
|
|
set up to the Ndb object of thd
|
|
|
|
*/
|
2005-03-15 15:03:25 +01:00
|
|
|
if (check_ndb_connection(thd))
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(1);
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= thd_ndb->ndb;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("enter", ("this: %x thd: %lx thd_ndb: %lx "
|
|
|
|
"thd_ndb->lock_count: %d",
|
|
|
|
this, thd, thd_ndb, thd_ndb->lock_count));
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (lock_type != F_UNLCK)
|
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("info", ("lock_type != F_UNLCK"));
|
2005-12-10 00:00:33 +01:00
|
|
|
if (!thd->transaction.on)
|
|
|
|
m_transaction_on= FALSE;
|
|
|
|
else
|
|
|
|
m_transaction_on= thd->variables.ndb_use_transactions;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
if (!thd_ndb->lock_count++)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
PRINT_OPTION_FLAGS(thd);
|
2005-10-14 11:23:02 +02:00
|
|
|
if (!(thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN)))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
// Autocommit transaction
|
2005-02-17 22:52:40 +01:00
|
|
|
DBUG_ASSERT(!thd_ndb->stmt);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("trans",("Starting transaction stmt"));
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
trans= ndb->startTransaction();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (trans == NULL)
|
2004-12-30 19:56:09 +01:00
|
|
|
ERR_RETURN(ndb->getNdbError());
|
2005-02-16 14:18:32 +01:00
|
|
|
no_uncommitted_rows_reset(thd);
|
2005-02-17 22:52:40 +01:00
|
|
|
thd_ndb->stmt= trans;
|
2005-12-10 00:00:33 +01:00
|
|
|
if (m_transaction_on)
|
|
|
|
trans_register_ha(thd, FALSE, &ndbcluster_hton);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-02-17 22:52:40 +01:00
|
|
|
if (!thd_ndb->all)
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
// Not autocommit transaction
|
|
|
|
// A "master" transaction ha not been started yet
|
|
|
|
DBUG_PRINT("trans",("starting transaction, all"));
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
trans= ndb->startTransaction();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (trans == NULL)
|
2004-12-30 19:56:09 +01:00
|
|
|
ERR_RETURN(ndb->getNdbError());
|
2005-02-16 14:18:32 +01:00
|
|
|
no_uncommitted_rows_reset(thd);
|
2005-02-17 22:52:40 +01:00
|
|
|
thd_ndb->all= trans;
|
2005-12-10 00:00:33 +01:00
|
|
|
if (m_transaction_on)
|
|
|
|
trans_register_ha(thd, TRUE, &ndbcluster_hton);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
If this is the start of a LOCK TABLE, a table look
|
|
|
|
should be taken on the table in NDB
|
|
|
|
|
|
|
|
Check if it should be read or write lock
|
|
|
|
*/
|
|
|
|
if (thd->options & (OPTION_TABLE_LOCK))
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
//lockThisTable();
|
|
|
|
DBUG_PRINT("info", ("Locking the table..." ));
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
This is the place to make sure this handler instance
|
|
|
|
has a started transaction.
|
|
|
|
|
|
|
|
The transaction is started by the first handler on which
|
|
|
|
MySQL Server calls external lock
|
|
|
|
|
|
|
|
Other handlers in the same stmt or transaction should use
|
|
|
|
the same NDB transaction. This is done by setting up the m_active_trans
|
|
|
|
pointer to point to the NDB transaction.
|
|
|
|
*/
|
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
// store thread specific data first to set the right context
|
|
|
|
m_force_send= thd->variables.ndb_force_send;
|
|
|
|
m_ha_not_exact_count= !thd->variables.ndb_use_exact_count;
|
2005-04-13 12:42:23 +02:00
|
|
|
m_autoincrement_prefetch=
|
|
|
|
(ha_rows) thd->variables.ndb_autoincrement_prefetch_sz;
|
2004-11-17 09:15:53 +01:00
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
m_active_trans= thd_ndb->all ? thd_ndb->all : thd_ndb->stmt;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(m_active_trans);
|
2004-05-10 14:12:28 +02:00
|
|
|
// Start of transaction
|
2005-03-15 15:03:25 +01:00
|
|
|
m_rows_changed= 0;
|
|
|
|
m_ops_pending= 0;
|
2004-09-20 17:52:19 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-09-20 17:52:19 +02:00
|
|
|
const NDBTAB *tab;
|
|
|
|
void *tab_info;
|
|
|
|
if (!(tab= dict->getTable(m_tabname, &tab_info)))
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(dict->getNdbError());
|
2005-04-27 18:17:41 +02:00
|
|
|
DBUG_PRINT("info", ("Table schema version: %d",
|
|
|
|
tab->getObjectVersion()));
|
|
|
|
// Check if thread has stale local cache
|
2005-08-18 14:02:25 +02:00
|
|
|
// New transaction must not use old tables... (trans != 0)
|
|
|
|
// Running might...
|
|
|
|
if ((trans && tab->getObjectStatus() != NdbDictionary::Object::Retrieved)
|
|
|
|
|| tab->getObjectStatus() == NdbDictionary::Object::Invalid)
|
2005-04-27 18:17:41 +02:00
|
|
|
{
|
|
|
|
invalidate_dictionary_cache(FALSE);
|
2005-04-29 11:37:47 +02:00
|
|
|
if (!(tab= dict->getTable(m_tabname, &tab_info)))
|
2005-04-27 18:17:41 +02:00
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
DBUG_PRINT("info", ("Table schema version: %d",
|
|
|
|
tab->getObjectVersion()));
|
|
|
|
}
|
2005-11-24 16:21:07 +01:00
|
|
|
if (m_table_version < tab->getObjectVersion())
|
2005-11-17 12:45:23 +01:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
The table has been altered, caller has to retry
|
|
|
|
*/
|
|
|
|
NdbError err= ndb->getNdbError(NDB_INVALID_SCHEMA_OBJECT);
|
|
|
|
DBUG_RETURN(ndb_to_mysql_error(&err));
|
|
|
|
}
|
2005-11-24 16:21:07 +01:00
|
|
|
if (m_table != (void *)tab)
|
|
|
|
{
|
|
|
|
m_table= (void *)tab;
|
|
|
|
m_table_version = tab->getObjectVersion();
|
|
|
|
if (!(my_errno= build_index_list(ndb, table, ILBP_OPEN)))
|
|
|
|
DBUG_RETURN(my_errno);
|
2005-11-17 12:45:23 +01:00
|
|
|
}
|
2004-09-20 17:52:19 +02:00
|
|
|
m_table_info= tab_info;
|
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_init(thd);
|
2005-03-15 15:03:25 +01:00
|
|
|
}
|
|
|
|
else
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("info", ("lock_type == F_UNLCK"));
|
2005-03-15 15:03:25 +01:00
|
|
|
|
|
|
|
if (ndb_cache_check_time && m_rows_changed)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Rows has changed and util thread is running"));
|
|
|
|
if (thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Add share to list of tables to be invalidated"));
|
|
|
|
/* NOTE push_back allocates memory using transactions mem_root! */
|
|
|
|
thd_ndb->changed_tables.push_back(m_share, &thd->transaction.mem_root);
|
|
|
|
}
|
|
|
|
|
|
|
|
pthread_mutex_lock(&m_share->mutex);
|
|
|
|
DBUG_PRINT("info", ("Invalidating commit_count"));
|
|
|
|
m_share->commit_count= 0;
|
|
|
|
m_share->commit_count_lock++;
|
|
|
|
pthread_mutex_unlock(&m_share->mutex);
|
|
|
|
}
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
if (!--thd_ndb->lock_count)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("trans", ("Last external_lock"));
|
|
|
|
PRINT_OPTION_FLAGS(thd);
|
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
if (thd_ndb->stmt)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
Unlock is done without a transaction commit / rollback.
|
|
|
|
This happens if the thread didn't update any rows
|
|
|
|
We must in this case close the transaction to release resources
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("trans",("ending non-updating transaction"));
|
2004-12-30 19:56:09 +01:00
|
|
|
ndb->closeTransaction(m_active_trans);
|
2005-02-17 22:52:40 +01:00
|
|
|
thd_ndb->stmt= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
}
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table_info= NULL;
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
/*
|
|
|
|
This is the place to make sure this handler instance
|
|
|
|
no longer are connected to the active transaction.
|
|
|
|
|
|
|
|
And since the handler is no longer part of the transaction
|
|
|
|
it can't have open cursors, ops or blobs pending.
|
|
|
|
*/
|
|
|
|
m_active_trans= NULL;
|
|
|
|
|
2004-09-07 12:54:31 +02:00
|
|
|
if (m_active_cursor)
|
|
|
|
DBUG_PRINT("warning", ("m_active_cursor != NULL"));
|
2004-09-13 14:46:38 +02:00
|
|
|
m_active_cursor= NULL;
|
|
|
|
|
2004-12-08 00:36:40 +01:00
|
|
|
if (m_multi_cursor)
|
|
|
|
DBUG_PRINT("warning", ("m_multi_cursor != NULL"));
|
|
|
|
m_multi_cursor= NULL;
|
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_blobs_pending)
|
2004-09-07 12:54:31 +02:00
|
|
|
DBUG_PRINT("warning", ("blobs_pending != 0"));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_pending= 0;
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_ops_pending)
|
2004-09-07 12:54:31 +02:00
|
|
|
DBUG_PRINT("warning", ("ops_pending != 0L"));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2005-10-14 11:23:02 +02:00
|
|
|
Start a transaction for running a statement if one is not
|
|
|
|
already running in a transaction. This will be the case in
|
|
|
|
a BEGIN; COMMIT; block
|
|
|
|
When using LOCK TABLE's external_lock will start a transaction
|
|
|
|
since ndb does not currently does not support table locking
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2005-09-30 20:20:10 +02:00
|
|
|
int ha_ndbcluster::start_stmt(THD *thd, thr_lock_type lock_type)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
int error=0;
|
|
|
|
DBUG_ENTER("start_stmt");
|
|
|
|
PRINT_OPTION_FLAGS(thd);
|
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
2005-10-14 11:28:01 +02:00
|
|
|
NdbTransaction *trans= (thd_ndb->stmt)?thd_ndb->stmt:thd_ndb->all;
|
2004-04-15 09:14:14 +02:00
|
|
|
if (!trans){
|
2005-02-17 22:52:40 +01:00
|
|
|
Ndb *ndb= thd_ndb->ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("trans",("Starting transaction stmt"));
|
2004-12-30 19:56:09 +01:00
|
|
|
trans= ndb->startTransaction();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (trans == NULL)
|
2004-12-30 19:56:09 +01:00
|
|
|
ERR_RETURN(ndb->getNdbError());
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_reset(thd);
|
2005-02-17 22:52:40 +01:00
|
|
|
thd_ndb->stmt= trans;
|
|
|
|
trans_register_ha(thd, FALSE, &ndbcluster_hton);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
m_active_trans= trans;
|
2004-05-05 16:04:23 +02:00
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
// Start of statement
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2005-03-15 15:03:25 +01:00
|
|
|
Commit a transaction started in NDB
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
int ndbcluster_commit(THD *thd, bool all)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
int res= 0;
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
|
|
|
Ndb *ndb= thd_ndb->ndb;
|
|
|
|
NdbTransaction *trans= all ? thd_ndb->all : thd_ndb->stmt;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("ndbcluster_commit");
|
|
|
|
DBUG_PRINT("transaction",("%s",
|
2005-03-15 15:03:25 +01:00
|
|
|
trans == thd_ndb->stmt ?
|
2004-04-15 09:14:14 +02:00
|
|
|
"stmt" : "all"));
|
|
|
|
DBUG_ASSERT(ndb && trans);
|
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
if (execute_commit(thd,trans) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
const NdbError err= trans->getNdbError();
|
2004-07-02 16:14:08 +02:00
|
|
|
const NdbOperation *error_op= trans->getNdbErrorOperation();
|
2005-03-15 15:03:25 +01:00
|
|
|
ERR_PRINT(err);
|
2004-04-15 09:14:14 +02:00
|
|
|
res= ndb_to_mysql_error(&err);
|
2005-03-15 15:03:25 +01:00
|
|
|
if (res != -1)
|
2004-07-02 16:14:08 +02:00
|
|
|
ndbcluster_print_error(res, error_op);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
ndb->closeTransaction(trans);
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if (all)
|
2005-02-21 10:53:19 +01:00
|
|
|
thd_ndb->all= NULL;
|
|
|
|
else
|
|
|
|
thd_ndb->stmt= NULL;
|
2005-03-15 15:03:25 +01:00
|
|
|
|
|
|
|
/* Clear commit_count for tables changed by transaction */
|
|
|
|
NDB_SHARE* share;
|
|
|
|
List_iterator_fast<NDB_SHARE> it(thd_ndb->changed_tables);
|
|
|
|
while ((share= it++))
|
|
|
|
{
|
|
|
|
pthread_mutex_lock(&share->mutex);
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("Invalidate commit_count for %s, share->commit_count: %d ",
|
|
|
|
share->key, share->commit_count));
|
2005-03-15 15:03:25 +01:00
|
|
|
share->commit_count= 0;
|
|
|
|
share->commit_count_lock++;
|
|
|
|
pthread_mutex_unlock(&share->mutex);
|
|
|
|
}
|
|
|
|
thd_ndb->changed_tables.empty();
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(res);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Rollback a transaction started in NDB
|
|
|
|
*/
|
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
int ndbcluster_rollback(THD *thd, bool all)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
int res= 0;
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
|
|
|
Ndb *ndb= thd_ndb->ndb;
|
|
|
|
NdbTransaction *trans= all ? thd_ndb->all : thd_ndb->stmt;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("ndbcluster_rollback");
|
|
|
|
DBUG_PRINT("transaction",("%s",
|
2005-02-17 22:52:40 +01:00
|
|
|
trans == thd_ndb->stmt ?
|
2004-04-15 09:14:14 +02:00
|
|
|
"stmt" : "all"));
|
|
|
|
DBUG_ASSERT(ndb && trans);
|
|
|
|
|
2005-01-05 17:59:24 +01:00
|
|
|
if (trans->execute(NdbTransaction::Rollback) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
const NdbError err= trans->getNdbError();
|
2004-07-02 16:14:08 +02:00
|
|
|
const NdbOperation *error_op= trans->getNdbErrorOperation();
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_PRINT(err);
|
|
|
|
res= ndb_to_mysql_error(&err);
|
2004-07-02 16:14:08 +02:00
|
|
|
if (res != -1)
|
|
|
|
ndbcluster_print_error(res, error_op);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
ndb->closeTransaction(trans);
|
2005-02-21 10:53:19 +01:00
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if (all)
|
2005-02-21 10:53:19 +01:00
|
|
|
thd_ndb->all= NULL;
|
|
|
|
else
|
|
|
|
thd_ndb->stmt= NULL;
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
/* Clear list of tables changed by transaction */
|
|
|
|
thd_ndb->changed_tables.empty();
|
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2004-07-22 12:38:09 +02:00
|
|
|
Define NDB column based on Field.
|
|
|
|
Returns 0 or mysql error code.
|
|
|
|
Not member of ha_ndbcluster because NDBCOL cannot be declared.
|
2005-01-08 00:03:19 +01:00
|
|
|
|
|
|
|
MySQL text types with character set "binary" are mapped to true
|
|
|
|
NDB binary types without a character set. This may change.
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
static int create_ndb_column(NDBCOL &col,
|
|
|
|
Field *field,
|
|
|
|
HA_CREATE_INFO *info)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
// Set name
|
BUG#9626
- Fix valgrind warnings
- Remove static db, schema and table name buffers from Ndb.cpp
mysys/thr_alarm.c:
Initialise sact to zero
ndb/include/kernel/ndb_limits.h:
Set MAX_ATTR_NAME_SIZE to NAME_LEN which is tha maximum in MySQL
ndb/include/kernel/signaldata/GetTabInfo.hpp:
Clean up
ndb/include/ndbapi/Ndb.hpp:
Remove NDB_MAX_INTERNAL_NAME_LENGTH and all char buffers for schema, db and tablename.
Made them dynamic and moved to NdbImpl.hpp
ndb/include/ndbapi/ndbapi_limits.h:
Remove the static length's of attr, db, schema and table name.
ndb/src/common/transporter/Packer.cpp:
Set theSignalId to ~0 when unpacking signal
ndb/src/ndbapi/Ndb.cpp:
Moved schema, database and tablename to NdbImpl.hpp
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Add NdbIndexImpl::init and NdbEventImpl::init
Init all vars in NdbIndexImpl, NdbEventImpl, NdbTableImpl and NdbIndexImpl
Delete the pseudo column NDB$RANGE_NO
Copy tablename to internal buff in NdbDictInterface to get proper alignment.
Convert length of table name from bytes words, when setting sz of LinearSectionPtr
Set LinearSectionPtr array size to number of LinearSections used - save some stack.
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
Add NdbEventImpl::init and NdbIndexImpl::init
Remove clearNewProperties and copyNewProperties, it's easier to check if all vars are initied if it's done in the same func.
Add buffer for tabname ti NdbDictInterface, memset it to 0 in initializer.
ndb/src/ndbapi/NdbImpl.hpp:
Use BaseString for table, schema and db names.
ndb/src/ndbapi/Ndbinit.cpp:
Move schema and db name to NdbImpl and use BaseString
ndb/src/ndbapi/ndb_cluster_connection.cpp:
Destroy ndb_global_event_buffer_mutex and ndb_print_state_mutex
sql/ha_ndbcluster.cc:
Check if pTrans is not null before calling closeTransaction
Remove NDB_MAX_ATTR_NAME_SIZE
Remove truncation of attr names. When attr name length is same in NDB as in MySQL this will be checked in functin check_column_name
2005-05-18 20:50:29 +02:00
|
|
|
col.setName(field->field_name);
|
2004-09-15 17:44:13 +02:00
|
|
|
// Get char set
|
|
|
|
CHARSET_INFO *cs= field->charset();
|
2004-07-22 12:38:09 +02:00
|
|
|
// Set type and sizes
|
|
|
|
const enum enum_field_types mysql_type= field->real_type();
|
|
|
|
switch (mysql_type) {
|
|
|
|
// Numeric types
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_TINY:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Tinyunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Tinyint);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_SHORT:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Smallunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Smallint);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_LONG:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Unsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Int);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_INT24:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Mediumunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Mediumint);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_LONGLONG:
|
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Bigunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Bigint);
|
|
|
|
col.setLength(1);
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_FLOAT:
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Float);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_DOUBLE:
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Double);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2005-02-16 12:45:59 +01:00
|
|
|
case MYSQL_TYPE_DECIMAL:
|
|
|
|
{
|
|
|
|
Field_decimal *f= (Field_decimal*)field;
|
|
|
|
uint precision= f->pack_length();
|
|
|
|
uint scale= f->decimals();
|
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
{
|
|
|
|
col.setType(NDBCOL::Olddecimalunsigned);
|
|
|
|
precision-= (scale > 0);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
col.setType(NDBCOL::Olddecimal);
|
|
|
|
precision-= 1 + (scale > 0);
|
|
|
|
}
|
|
|
|
col.setPrecision(precision);
|
|
|
|
col.setScale(scale);
|
|
|
|
col.setLength(1);
|
|
|
|
}
|
|
|
|
break;
|
2005-02-16 21:19:42 +01:00
|
|
|
case MYSQL_TYPE_NEWDECIMAL:
|
|
|
|
{
|
|
|
|
Field_new_decimal *f= (Field_new_decimal*)field;
|
2005-05-05 17:06:49 +02:00
|
|
|
uint precision= f->precision;
|
2005-02-16 21:19:42 +01:00
|
|
|
uint scale= f->decimals();
|
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
{
|
|
|
|
col.setType(NDBCOL::Decimalunsigned);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
col.setType(NDBCOL::Decimal);
|
|
|
|
}
|
|
|
|
col.setPrecision(precision);
|
|
|
|
col.setScale(scale);
|
|
|
|
col.setLength(1);
|
|
|
|
}
|
|
|
|
break;
|
2004-07-22 12:38:09 +02:00
|
|
|
// Date types
|
|
|
|
case MYSQL_TYPE_DATETIME:
|
|
|
|
col.setType(NDBCOL::Datetime);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2005-01-27 17:33:35 +01:00
|
|
|
case MYSQL_TYPE_DATE: // ?
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
2004-07-22 12:38:09 +02:00
|
|
|
case MYSQL_TYPE_NEWDATE:
|
2005-01-08 16:57:51 +01:00
|
|
|
col.setType(NDBCOL::Date);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-07-22 12:38:09 +02:00
|
|
|
case MYSQL_TYPE_TIME:
|
2005-01-08 19:28:44 +01:00
|
|
|
col.setType(NDBCOL::Time);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2005-01-27 17:33:35 +01:00
|
|
|
case MYSQL_TYPE_YEAR:
|
|
|
|
col.setType(NDBCOL::Year);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_TIMESTAMP:
|
|
|
|
col.setType(NDBCOL::Timestamp);
|
|
|
|
col.setLength(1);
|
2004-07-22 12:38:09 +02:00
|
|
|
break;
|
|
|
|
// Char types
|
|
|
|
case MYSQL_TYPE_STRING:
|
2004-10-20 19:22:58 +02:00
|
|
|
if (field->pack_length() == 0)
|
2004-12-09 14:17:58 +01:00
|
|
|
{
|
|
|
|
col.setType(NDBCOL::Bit);
|
|
|
|
col.setLength(1);
|
|
|
|
}
|
2005-01-08 00:03:19 +01:00
|
|
|
else if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2004-12-10 16:55:04 +01:00
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Binary);
|
2004-12-10 16:55:04 +01:00
|
|
|
col.setLength(field->pack_length());
|
2004-09-15 17:44:13 +02:00
|
|
|
}
|
2004-10-20 19:22:58 +02:00
|
|
|
else
|
2004-12-10 16:55:04 +01:00
|
|
|
{
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setCharset(cs);
|
2004-10-20 19:22:58 +02:00
|
|
|
col.setLength(field->pack_length());
|
2004-12-10 16:55:04 +01:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
break;
|
2005-01-07 11:55:20 +01:00
|
|
|
case MYSQL_TYPE_VAR_STRING: // ?
|
|
|
|
case MYSQL_TYPE_VARCHAR:
|
|
|
|
{
|
|
|
|
Field_varstring* f= (Field_varstring*)field;
|
|
|
|
if (f->length_bytes == 1)
|
|
|
|
{
|
2005-01-08 00:03:19 +01:00
|
|
|
if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2005-01-07 11:55:20 +01:00
|
|
|
col.setType(NDBCOL::Varbinary);
|
|
|
|
else {
|
|
|
|
col.setType(NDBCOL::Varchar);
|
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (f->length_bytes == 2)
|
|
|
|
{
|
2005-01-08 00:03:19 +01:00
|
|
|
if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2005-01-07 11:55:20 +01:00
|
|
|
col.setType(NDBCOL::Longvarbinary);
|
|
|
|
else {
|
|
|
|
col.setType(NDBCOL::Longvarchar);
|
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
return HA_ERR_UNSUPPORTED;
|
|
|
|
}
|
|
|
|
col.setLength(field->field_length);
|
2004-09-15 17:44:13 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
break;
|
|
|
|
// Blob types (all come in as MYSQL_TYPE_BLOB)
|
|
|
|
mysql_type_tiny_blob:
|
|
|
|
case MYSQL_TYPE_TINY_BLOB:
|
2005-01-08 00:03:19 +01:00
|
|
|
if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setInlineSize(256);
|
|
|
|
// No parts
|
|
|
|
col.setPartSize(0);
|
|
|
|
col.setStripeSize(0);
|
|
|
|
break;
|
2004-11-29 13:50:13 +01:00
|
|
|
//mysql_type_blob:
|
2005-10-25 08:20:39 +02:00
|
|
|
case MYSQL_TYPE_GEOMETRY:
|
2004-07-22 12:38:09 +02:00
|
|
|
case MYSQL_TYPE_BLOB:
|
2005-01-08 00:03:19 +01:00
|
|
|
if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
// Use "<=" even if "<" is the exact condition
|
|
|
|
if (field->max_length() <= (1 << 8))
|
|
|
|
goto mysql_type_tiny_blob;
|
|
|
|
else if (field->max_length() <= (1 << 16))
|
|
|
|
{
|
|
|
|
col.setInlineSize(256);
|
|
|
|
col.setPartSize(2000);
|
|
|
|
col.setStripeSize(16);
|
|
|
|
}
|
|
|
|
else if (field->max_length() <= (1 << 24))
|
|
|
|
goto mysql_type_medium_blob;
|
|
|
|
else
|
|
|
|
goto mysql_type_long_blob;
|
|
|
|
break;
|
|
|
|
mysql_type_medium_blob:
|
|
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
2005-01-08 00:03:19 +01:00
|
|
|
if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setInlineSize(256);
|
|
|
|
col.setPartSize(4000);
|
|
|
|
col.setStripeSize(8);
|
|
|
|
break;
|
|
|
|
mysql_type_long_blob:
|
|
|
|
case MYSQL_TYPE_LONG_BLOB:
|
2005-01-08 00:03:19 +01:00
|
|
|
if ((field->flags & BINARY_FLAG) && cs == &my_charset_bin)
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setInlineSize(256);
|
|
|
|
col.setPartSize(8000);
|
|
|
|
col.setStripeSize(4);
|
|
|
|
break;
|
|
|
|
// Other types
|
|
|
|
case MYSQL_TYPE_ENUM:
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_SET:
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case MYSQL_TYPE_BIT:
|
|
|
|
{
|
2004-12-23 15:28:41 +01:00
|
|
|
int no_of_bits= field->field_length*8 + ((Field_bit *) field)->bit_len;
|
|
|
|
col.setType(NDBCOL::Bit);
|
|
|
|
if (!no_of_bits)
|
|
|
|
col.setLength(1);
|
|
|
|
else
|
|
|
|
col.setLength(no_of_bits);
|
|
|
|
break;
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
case MYSQL_TYPE_NULL:
|
|
|
|
goto mysql_type_unsupported;
|
|
|
|
mysql_type_unsupported:
|
|
|
|
default:
|
|
|
|
return HA_ERR_UNSUPPORTED;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
// Set nullable and pk
|
|
|
|
col.setNullable(field->maybe_null());
|
|
|
|
col.setPrimaryKey(field->flags & PRI_KEY_FLAG);
|
|
|
|
// Set autoincrement
|
|
|
|
if (field->flags & AUTO_INCREMENT_FLAG)
|
|
|
|
{
|
|
|
|
col.setAutoIncrement(TRUE);
|
|
|
|
ulonglong value= info->auto_increment_value ?
|
2004-09-02 17:51:44 +02:00
|
|
|
info->auto_increment_value : (ulonglong) 1;
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("info", ("Autoincrement key, initial: %llu", value));
|
|
|
|
col.setAutoIncrementInitialValue(value);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
else
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
col.setAutoIncrement(FALSE);
|
2004-07-22 12:38:09 +02:00
|
|
|
return 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/*
|
|
|
|
Create a table in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::create(const char *name,
|
2005-02-16 14:18:32 +01:00
|
|
|
TABLE *form,
|
|
|
|
HA_CREATE_INFO *info)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
NDBTAB tab;
|
|
|
|
NDBCOL col;
|
2004-11-08 11:06:36 +01:00
|
|
|
uint pack_length, length, i, pk_length= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
const void *data, *pack_data;
|
|
|
|
char name2[FN_HEADLEN];
|
2005-11-21 12:27:58 +01:00
|
|
|
bool create_from_engine= (info->table_options & HA_OPTION_CREATE_FROM_ENGINE);
|
2005-11-25 11:57:13 +01:00
|
|
|
|
2005-01-07 11:55:20 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::create");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("name: %s", name));
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
|
2005-12-31 06:01:26 +01:00
|
|
|
strcpy(name2, name);
|
|
|
|
DBUG_ASSERT(*fn_rext((char*)name2) == 0);
|
2004-04-15 09:14:14 +02:00
|
|
|
set_dbname(name2);
|
2004-09-13 14:46:38 +02:00
|
|
|
set_tabname(name2);
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
table= form;
|
2004-09-13 14:46:38 +02:00
|
|
|
if (create_from_engine)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Table alreay exists in NDB and frm file has been created by
|
|
|
|
caller.
|
|
|
|
Do Ndb specific stuff, such as create a .ndb file
|
|
|
|
*/
|
2005-11-06 00:20:37 +01:00
|
|
|
if ((my_errno= write_ndb_file()))
|
|
|
|
DBUG_RETURN(my_errno);
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("table", ("name: %s", m_tabname));
|
|
|
|
tab.setName(m_tabname);
|
|
|
|
tab.setLogging(!(info->options & HA_LEX_CREATE_TMP_TABLE));
|
|
|
|
|
|
|
|
// Save frm data for this table
|
|
|
|
if (readfrm(name, &data, &length))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
if (packfrm(data, length, &pack_data, &pack_length))
|
|
|
|
DBUG_RETURN(2);
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("setFrm data=%lx len=%d", pack_data, pack_length));
|
2004-04-15 09:14:14 +02:00
|
|
|
tab.setFrm(pack_data, pack_length);
|
|
|
|
my_free((char*)data, MYF(0));
|
|
|
|
my_free((char*)pack_data, MYF(0));
|
|
|
|
|
2005-01-06 12:00:13 +01:00
|
|
|
for (i= 0; i < form->s->fields; i++)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
Field *field= form->field[i];
|
|
|
|
DBUG_PRINT("info", ("name: %s, type: %u, pack_length: %d",
|
|
|
|
field->field_name, field->real_type(),
|
2005-02-16 14:18:32 +01:00
|
|
|
field->pack_length()));
|
2004-07-23 15:46:56 +02:00
|
|
|
if ((my_errno= create_ndb_column(col, field, info)))
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
2005-11-07 12:19:28 +01:00
|
|
|
|
|
|
|
if (
|
|
|
|
#ifdef NDB_DISKDATA
|
|
|
|
info->store_on_disk ||
|
|
|
|
#else
|
|
|
|
getenv("NDB_DEFAULT_DISK"))
|
|
|
|
#endif
|
|
|
|
col.setStorageType(NdbDictionary::Column::StorageTypeDisk);
|
|
|
|
else
|
|
|
|
col.setStorageType(NdbDictionary::Column::StorageTypeMemory);
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
tab.addColumn(col);
|
2005-07-04 02:42:33 +02:00
|
|
|
if (col.getPrimaryKey())
|
2004-11-08 11:06:36 +01:00
|
|
|
pk_length += (field->pack_length() + 3) / 4;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-11-07 12:19:28 +01:00
|
|
|
|
|
|
|
KEY* key_info;
|
|
|
|
for (i= 0, key_info= form->key_info; i < form->s->keys; i++, key_info++)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO *key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO *end= key_part + key_info->key_parts;
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
tab.getColumn(key_part->fieldnr-1)->setStorageType(
|
|
|
|
NdbDictionary::Column::StorageTypeMemory);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef NDB_DISKDATA
|
|
|
|
if (info->store_on_disk)
|
|
|
|
if (info->tablespace)
|
|
|
|
tab.setTablespace(info->tablespace);
|
|
|
|
else
|
|
|
|
tab.setTablespace("DEFAULT-TS");
|
|
|
|
#endif
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// No primary key, create shadow key as 64 bit, auto increment
|
2005-01-06 12:00:13 +01:00
|
|
|
if (form->s->primary_key == MAX_KEY)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Generating shadow key"));
|
|
|
|
col.setName("$PK");
|
|
|
|
col.setType(NdbDictionary::Column::Bigunsigned);
|
|
|
|
col.setLength(1);
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
col.setNullable(FALSE);
|
2004-04-15 09:14:14 +02:00
|
|
|
col.setPrimaryKey(TRUE);
|
|
|
|
col.setAutoIncrement(TRUE);
|
|
|
|
tab.addColumn(col);
|
2004-11-08 11:06:36 +01:00
|
|
|
pk_length += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure that blob tables don't have to big part size
|
2005-01-06 12:00:13 +01:00
|
|
|
for (i= 0; i < form->s->fields; i++)
|
2004-11-08 11:06:36 +01:00
|
|
|
{
|
|
|
|
/**
|
|
|
|
* The extra +7 concists
|
|
|
|
* 2 - words from pk in blob table
|
|
|
|
* 5 - from extra words added by tup/dict??
|
|
|
|
*/
|
|
|
|
switch (form->field[i]->real_type()) {
|
2005-10-25 08:20:39 +02:00
|
|
|
case MYSQL_TYPE_GEOMETRY:
|
2004-11-08 11:06:36 +01:00
|
|
|
case MYSQL_TYPE_BLOB:
|
|
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
|
|
|
case MYSQL_TYPE_LONG_BLOB:
|
|
|
|
{
|
2004-12-07 15:15:49 +01:00
|
|
|
NdbDictionary::Column * col= tab.getColumn(i);
|
|
|
|
int size= pk_length + (col->getPartSize()+3)/4 + 7;
|
2005-07-04 02:42:33 +02:00
|
|
|
if (size > NDB_MAX_TUPLE_SIZE_IN_WORDS &&
|
2005-02-16 14:18:32 +01:00
|
|
|
(pk_length+7) < NDB_MAX_TUPLE_SIZE_IN_WORDS)
|
2004-11-08 11:06:36 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
size= NDB_MAX_TUPLE_SIZE_IN_WORDS - pk_length - 7;
|
|
|
|
col->setPartSize(4*size);
|
2004-11-08 11:06:36 +01:00
|
|
|
}
|
|
|
|
/**
|
|
|
|
* If size > NDB_MAX and pk_length+7 >= NDB_MAX
|
|
|
|
* then the table can't be created anyway, so skip
|
|
|
|
* changing part size, and have error later
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
// Check partition info
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
partition_info *part_info= form->part_info;
|
2005-07-18 13:31:02 +02:00
|
|
|
if (part_info)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
if ((error= set_up_partition_info(part_info, form, (void*)&tab)))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ndb_set_fragmentation(tab, form, pk_length);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if ((my_errno= check_ndb_connection()))
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
|
|
|
|
// Create the table in NDB
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-09-13 14:46:38 +02:00
|
|
|
if (dict->createTable(tab) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
|
|
|
ERR_PRINT(err);
|
|
|
|
my_errno= ndb_to_mysql_error(&err);
|
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("Table %s/%s created successfully",
|
|
|
|
m_dbname, m_tabname));
|
2004-04-30 12:25:31 +02:00
|
|
|
|
2004-08-19 11:10:35 +02:00
|
|
|
// Create secondary indexes
|
2005-03-22 17:42:08 +01:00
|
|
|
my_errno= build_index_list(ndb, form, ILBP_CREATE);
|
2004-08-18 19:13:39 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if (!my_errno)
|
|
|
|
my_errno= write_ndb_file();
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-30 12:25:31 +02:00
|
|
|
int ha_ndbcluster::create_ordered_index(const char *name,
|
2005-02-16 14:18:32 +01:00
|
|
|
KEY *key_info)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::create_ordered_index");
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(create_index(name, key_info, FALSE));
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::create_unique_index(const char *name,
|
2005-02-16 14:18:32 +01:00
|
|
|
KEY *key_info)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
|
|
|
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::create_unique_index");
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(create_index(name, key_info, TRUE));
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Create an index in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::create_index(const char *name,
|
2005-02-16 14:18:32 +01:00
|
|
|
KEY *key_info,
|
|
|
|
bool unique)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NdbDictionary::Dictionary *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
KEY_PART_INFO *key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO *end= key_part + key_info->key_parts;
|
|
|
|
|
2004-12-20 15:12:42 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::create_index");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("name: %s ", name));
|
2004-04-30 12:25:31 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbDictionary::Index ndb_index(name);
|
2004-04-30 12:25:31 +02:00
|
|
|
if (unique)
|
2004-04-15 09:14:14 +02:00
|
|
|
ndb_index.setType(NdbDictionary::Index::UniqueHashIndex);
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ndb_index.setType(NdbDictionary::Index::OrderedIndex);
|
|
|
|
// TODO Only temporary ordered indexes supported
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
ndb_index.setLogging(FALSE);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
ndb_index.setTable(m_tabname);
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field *field= key_part->field;
|
|
|
|
DBUG_PRINT("info", ("attr: %s", field->field_name));
|
BUG#9626
- Fix valgrind warnings
- Remove static db, schema and table name buffers from Ndb.cpp
mysys/thr_alarm.c:
Initialise sact to zero
ndb/include/kernel/ndb_limits.h:
Set MAX_ATTR_NAME_SIZE to NAME_LEN which is tha maximum in MySQL
ndb/include/kernel/signaldata/GetTabInfo.hpp:
Clean up
ndb/include/ndbapi/Ndb.hpp:
Remove NDB_MAX_INTERNAL_NAME_LENGTH and all char buffers for schema, db and tablename.
Made them dynamic and moved to NdbImpl.hpp
ndb/include/ndbapi/ndbapi_limits.h:
Remove the static length's of attr, db, schema and table name.
ndb/src/common/transporter/Packer.cpp:
Set theSignalId to ~0 when unpacking signal
ndb/src/ndbapi/Ndb.cpp:
Moved schema, database and tablename to NdbImpl.hpp
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Add NdbIndexImpl::init and NdbEventImpl::init
Init all vars in NdbIndexImpl, NdbEventImpl, NdbTableImpl and NdbIndexImpl
Delete the pseudo column NDB$RANGE_NO
Copy tablename to internal buff in NdbDictInterface to get proper alignment.
Convert length of table name from bytes words, when setting sz of LinearSectionPtr
Set LinearSectionPtr array size to number of LinearSections used - save some stack.
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
Add NdbEventImpl::init and NdbIndexImpl::init
Remove clearNewProperties and copyNewProperties, it's easier to check if all vars are initied if it's done in the same func.
Add buffer for tabname ti NdbDictInterface, memset it to 0 in initializer.
ndb/src/ndbapi/NdbImpl.hpp:
Use BaseString for table, schema and db names.
ndb/src/ndbapi/Ndbinit.cpp:
Move schema and db name to NdbImpl and use BaseString
ndb/src/ndbapi/ndb_cluster_connection.cpp:
Destroy ndb_global_event_buffer_mutex and ndb_print_state_mutex
sql/ha_ndbcluster.cc:
Check if pTrans is not null before calling closeTransaction
Remove NDB_MAX_ATTR_NAME_SIZE
Remove truncation of attr names. When attr name length is same in NDB as in MySQL this will be checked in functin check_column_name
2005-05-18 20:50:29 +02:00
|
|
|
ndb_index.addColumnName(field->field_name);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
if (dict->createIndex(ndb_index))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
|
|
|
|
// Success
|
|
|
|
DBUG_PRINT("info", ("Created index %s", name));
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Rename a table in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::rename_table(const char *from, const char *to)
|
|
|
|
{
|
2004-12-06 14:51:10 +01:00
|
|
|
NDBDICT *dict;
|
2004-04-15 09:14:14 +02:00
|
|
|
char new_tabname[FN_HEADLEN];
|
2004-12-06 14:51:10 +01:00
|
|
|
const NDBTAB *orig_tab;
|
|
|
|
int result;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("ha_ndbcluster::rename_table");
|
2004-12-06 14:51:10 +01:00
|
|
|
DBUG_PRINT("info", ("Renaming %s to %s", from, to));
|
2004-04-15 09:14:14 +02:00
|
|
|
set_dbname(from);
|
|
|
|
set_tabname(from);
|
|
|
|
set_tabname(to, new_tabname);
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if (check_ndb_connection())
|
|
|
|
DBUG_RETURN(my_errno= HA_ERR_NO_CONNECTION);
|
|
|
|
|
2005-01-03 11:56:51 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
dict= ndb->getDictionary();
|
2004-12-06 14:51:10 +01:00
|
|
|
if (!(orig_tab= dict->getTable(m_tabname)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
2005-04-28 09:39:29 +02:00
|
|
|
// Check if thread has stale local cache
|
|
|
|
if (orig_tab->getObjectStatus() == NdbDictionary::Object::Invalid)
|
|
|
|
{
|
|
|
|
dict->removeCachedTable(m_tabname);
|
|
|
|
if (!(orig_tab= dict->getTable(m_tabname)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
}
|
2004-12-06 14:51:10 +01:00
|
|
|
m_table= (void *)orig_tab;
|
|
|
|
// Change current database to that of target table
|
|
|
|
set_dbname(to);
|
2005-01-03 11:56:51 +01:00
|
|
|
ndb->setDatabaseName(m_dbname);
|
2005-11-06 00:20:37 +01:00
|
|
|
|
|
|
|
if ((result= alter_table_name(new_tabname)))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(result);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Rename .ndb file
|
|
|
|
if ((result= handler::rename_table(from, to)))
|
2004-09-13 14:46:38 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
// ToDo in 4.1 should rollback alter table...
|
|
|
|
DBUG_RETURN(result);
|
2004-09-13 14:46:38 +02:00
|
|
|
}
|
2004-12-06 14:51:10 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(result);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Rename a table in NDB Cluster using alter table
|
|
|
|
*/
|
|
|
|
|
2004-12-06 14:51:10 +01:00
|
|
|
int ha_ndbcluster::alter_table_name(const char *to)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-12-06 14:51:10 +01:00
|
|
|
const NDBTAB *orig_tab= (const NDBTAB *) m_table;
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_ENTER("alter_table_name");
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-12-07 10:27:22 +01:00
|
|
|
NdbDictionary::Table new_tab= *orig_tab;
|
2004-12-06 14:51:10 +01:00
|
|
|
new_tab.setName(to);
|
|
|
|
if (dict->alterTable(new_tab) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
|
|
|
|
m_table= NULL;
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table_info= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2005-02-17 14:22:44 +01:00
|
|
|
Delete table from NDB Cluster
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/* static version which does not need a handler */
|
|
|
|
|
|
|
|
int
|
|
|
|
ha_ndbcluster::delete_table(ha_ndbcluster *h, Ndb *ndb,
|
|
|
|
const char *path,
|
|
|
|
const char *db,
|
|
|
|
const char *table_name)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::ndbcluster_delete_table");
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
|
|
|
|
|
|
|
/* Drop the table from NDB */
|
|
|
|
|
|
|
|
int res;
|
|
|
|
if (h)
|
|
|
|
{
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
res= h->intern_drop_table();
|
2005-11-06 00:20:37 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ndb->setDatabaseName(db);
|
|
|
|
res= dict->dropTable(table_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (res)
|
|
|
|
{
|
|
|
|
DBUG_RETURN(res);
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::delete_table(const char *name)
|
|
|
|
{
|
2005-02-17 14:22:44 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::delete_table");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("name: %s", name));
|
|
|
|
set_dbname(name);
|
|
|
|
set_tabname(name);
|
2005-02-17 14:22:44 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (check_ndb_connection())
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2005-02-17 14:22:44 +01:00
|
|
|
|
|
|
|
/* Call ancestor function to delete .ndb file */
|
2004-09-13 14:46:38 +02:00
|
|
|
handler::delete_table(name);
|
2005-11-06 00:20:37 +01:00
|
|
|
|
|
|
|
DBUG_RETURN(delete_table(this, get_ndb(),name, m_dbname, m_tabname));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2005-02-17 14:22:44 +01:00
|
|
|
Drop table in NDB Cluster
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
int ha_ndbcluster::intern_drop_table()
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NdbDictionary::Dictionary *dict= ndb->getDictionary();
|
2005-02-17 14:22:44 +01:00
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
DBUG_ENTER("intern_drop_table");
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("Deleting %s", m_tabname));
|
|
|
|
release_metadata();
|
2005-02-17 14:22:44 +01:00
|
|
|
if (dict->dropTable(m_tabname))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-15 21:10:31 +02:00
|
|
|
ulonglong ha_ndbcluster::get_auto_increment()
|
2004-04-29 14:38:35 +02:00
|
|
|
{
|
2004-09-15 21:10:31 +02:00
|
|
|
int cache_size;
|
|
|
|
Uint64 auto_value;
|
2004-07-02 11:50:28 +02:00
|
|
|
DBUG_ENTER("get_auto_increment");
|
|
|
|
DBUG_PRINT("enter", ("m_tabname: %s", m_tabname));
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2005-04-08 16:49:14 +02:00
|
|
|
|
2005-04-07 20:08:01 +02:00
|
|
|
if (m_rows_inserted > m_rows_to_insert)
|
2005-05-07 22:04:52 +02:00
|
|
|
{
|
2005-04-07 20:08:01 +02:00
|
|
|
/* We guessed too low */
|
|
|
|
m_rows_to_insert+= m_autoincrement_prefetch;
|
2005-05-07 22:04:52 +02:00
|
|
|
}
|
2004-11-20 18:36:41 +01:00
|
|
|
cache_size=
|
2005-07-04 02:42:33 +02:00
|
|
|
(int) ((m_rows_to_insert - m_rows_inserted < m_autoincrement_prefetch) ?
|
|
|
|
m_rows_to_insert - m_rows_inserted :
|
|
|
|
((m_rows_to_insert > m_autoincrement_prefetch) ?
|
|
|
|
m_rows_to_insert : m_autoincrement_prefetch));
|
2005-04-16 03:35:54 +02:00
|
|
|
auto_value= NDB_FAILED_AUTO_INCREMENT;
|
2005-04-15 16:14:08 +02:00
|
|
|
uint retries= NDB_AUTO_INCREMENT_RETRIES;
|
|
|
|
do {
|
|
|
|
auto_value=
|
|
|
|
(m_skip_auto_increment) ?
|
|
|
|
ndb->readAutoIncrementValue((const NDBTAB *) m_table)
|
|
|
|
: ndb->getAutoIncrementValue((const NDBTAB *) m_table, cache_size);
|
|
|
|
} while (auto_value == NDB_FAILED_AUTO_INCREMENT &&
|
|
|
|
--retries &&
|
|
|
|
ndb->getNdbError().status == NdbError::TemporaryError);
|
2005-04-15 15:52:07 +02:00
|
|
|
if (auto_value == NDB_FAILED_AUTO_INCREMENT)
|
2005-10-14 11:23:02 +02:00
|
|
|
{
|
|
|
|
const NdbError err= ndb->getNdbError();
|
|
|
|
sql_print_error("Error %lu in ::get_auto_increment(): %s",
|
|
|
|
(ulong) err.code, err.message);
|
|
|
|
DBUG_RETURN(~(ulonglong) 0);
|
|
|
|
}
|
2004-07-02 11:50:28 +02:00
|
|
|
DBUG_RETURN((longlong)auto_value);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Constructor for the NDB Cluster table handler
|
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
#define HA_NDBCLUSTER_TABLE_FLAGS \
|
|
|
|
HA_REC_NOT_IN_SEQ | \
|
|
|
|
HA_NULL_IN_KEY | \
|
|
|
|
HA_AUTO_PART_KEY | \
|
|
|
|
HA_NO_PREFIX_CHAR_KEYS | \
|
|
|
|
HA_NEED_READ_RANGE_BUFFER | \
|
|
|
|
HA_CAN_GEOMETRY | \
|
|
|
|
HA_CAN_BIT_FIELD
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
ha_ndbcluster::ha_ndbcluster(TABLE_SHARE *table_arg):
|
A fix and a test case for Bug#10760 and complementary cleanups.
The idea of the patch
is that every cursor gets its own lock id for table level locking.
Thus cursors are protected from updates performed within the same
connection. Additionally a list of transient (must be closed at
commit) cursors is maintained and all transient cursors are closed
when necessary. Lastly, this patch adds support for deadlock
timeouts to TLL locking when using cursors.
+ post-review fixes.
include/thr_lock.h:
- add a notion of lock owner to table level locking. When using
cursors, lock owner can not be identified by a thread id any more,
as we must protect cursors from updates issued within the same
connection (thread). So, each cursor has its own lock identifier to
use with table level locking.
- extend return values of thr_lock and thr_multi_lock with
THR_LOCK_TIMEOUT and THR_LOCK_DEADLOCK, since these conditions
are now possible (see comments to thr_lock.c)
mysys/thr_lock.c:
Better support for cursors:
- use THR_LOCK_OWNER * as lock identifier, not pthread_t.
- check and return an error for a trivial deadlock case, when an
update statement is issued to a table locked by a cursor which has
been previously opened in the same connection.
- add support for locking timeouts: with use of cursors, trivial
deadlocks can occur. For now the only remedy is the lock wait timeout,
which is initialized from a new global variable 'table_lock_wait_timeout'
Example of a deadlock (assuming the storage engine does not downgrade
locks):
con1: open cursor for select * from t1;
con2: open cursor for select * from t2;
con1: update t2 set id=id*2; -- blocked
con2: update t1 set id=id*2; -- deadlock
Lock timeouts are active only if a connection is using cursors.
- the check in the wait_for_lock loop has been changed from
data->cond != cond to data->cond != 0. data->cond is zeroed
in every place it's changed.
- added comments
sql/examples/ha_archive.cc:
- extend the handlerton with the info about cursor behaviour at commit.
sql/examples/ha_archive.h:
- ctor moved to .cc to make use of archive handlerton
sql/examples/ha_example.cc:
- add handlerton instance, init handler::ht with it
sql/examples/ha_example.h:
- ctor moved to .cc to make use of ha_example handlerton
sql/examples/ha_tina.cc:
- add handlerton instance, init handler::ht with it
sql/examples/ha_tina.h:
- ctor moved to .cc to make use of CSV handlerton
sql/ha_berkeley.cc:
- init handlerton::flags and handler::ht
sql/ha_berkeley.h:
- ctor moved to .cc to make use of BerkeleyDB handlerton
sql/ha_blackhole.cc:
- add handlerton instance, init handler::ht with it
sql/ha_blackhole.h:
- ctor moved to .cc to make use of blackhole handlerton
sql/ha_federated.cc:
- add handlerton instance, init handler::ht with it
sql/ha_federated.h:
- ctor moved to .cc to make use of federated handlerton
sql/ha_heap.cc:
- add handlerton instance, init handler::ht with it
sql/ha_heap.h:
- ctor moved to .cc to make use of ha_heap handlerton
sql/ha_innodb.cc:
- init handlerton::flags and handler::ht of innobase storage engine
sql/ha_innodb.h:
- ctor moved to .cc to make use of archive handlerton
sql/ha_myisam.cc:
- add handlerton instance, init handler::ht with it
sql/ha_myisam.h:
- ctor moved to .cc to make use of MyISAM handlerton
sql/ha_myisammrg.cc:
- init handler::ht in the ctor
sql/ha_myisammrg.h:
- ctor moved to .cc to make use of MyISAM MERGE handlerton
sql/ha_ndbcluster.cc:
- init handlerton::flags and handler::ht
sql/handler.cc:
- drop support for ISAM storage engine, which was removed from 5.0
- close all "transient" cursors at COMMIT/ROLLBACK. A "transient"
SQL level cursor is a cursor that uses tables that have a transaction-
specific state.
sql/handler.h:
- extend struct handlerton with flags, add handlerton *ht to every
handler instance.
sql/lock.cc:
- extend mysql_lock_tables to send error to the client if
thr_multi_lock returns a timeout or a deadlock error.
sql/mysqld.cc:
- add server option --table_lock_wait_timeout (in seconds)
sql/set_var.cc:
- add new global variable 'table_lock_wait_timeout' to specify
a wait timeout for table-level locks of MySQL (in seconds). The default
timeout is 50 seconds. The timeout is active only if the connection
has open cursors.
sql/sql_class.cc:
- implement Statement_map::close_transient_cursors
- safety suggests that we need an assert ensuring
llock_info->n_cursors is functioning properly, adjust destruction of
the Statement_map to allow such assert in THD::~THD
sql/sql_class.h:
- add support for Cursors registry to Statement map.
sql/sql_prepare.cc:
- maintain a list of cursors that must be closed at commit/rollback.
sql/sql_select.cc:
- extend class Cursor to support specific at-COMMIT/ROLLBACK behavior.
If a cursor uses tables of a storage engine that
invalidates all open tables at COMMIT/ROLLBACK, it must be closed
before COMMIT/ROLLBACK is executed.
sql/sql_select.h:
- add an own lock_id and commit/rollback status flag to class Cursor
tests/mysql_client_test.c:
A test case for Bug#10760 and complementary issues: test a simple
deadlock case too.
mysql-test/var:
New BitKeeper file ``mysql-test/var''
2005-07-19 20:21:12 +02:00
|
|
|
handler(&ndbcluster_hton, table_arg),
|
2004-04-15 09:14:14 +02:00
|
|
|
m_active_trans(NULL),
|
|
|
|
m_active_cursor(NULL),
|
|
|
|
m_table(NULL),
|
2005-04-27 18:17:41 +02:00
|
|
|
m_table_version(-1),
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
m_table_info(NULL),
|
2005-11-06 00:20:37 +01:00
|
|
|
m_table_flags(HA_NDBCLUSTER_TABLE_FLAGS),
|
2004-09-06 20:30:57 +02:00
|
|
|
m_share(0),
|
2005-07-18 13:31:02 +02:00
|
|
|
m_part_info(NULL),
|
|
|
|
m_use_partition_function(FALSE),
|
|
|
|
m_sorted(FALSE),
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
m_use_write(FALSE),
|
2004-11-18 12:11:56 +01:00
|
|
|
m_ignore_dup_key(FALSE),
|
2004-11-03 15:53:26 +01:00
|
|
|
m_primary_key_update(FALSE),
|
2005-11-06 00:20:37 +01:00
|
|
|
m_ignore_no_key(FALSE),
|
2005-04-13 12:42:23 +02:00
|
|
|
m_rows_to_insert((ha_rows) 1),
|
|
|
|
m_rows_inserted((ha_rows) 0),
|
|
|
|
m_bulk_insert_rows((ha_rows) 1024),
|
2005-04-13 16:24:17 +02:00
|
|
|
m_rows_changed((ha_rows) 0),
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_not_flushed(FALSE),
|
|
|
|
m_ops_pending(0),
|
|
|
|
m_skip_auto_increment(TRUE),
|
|
|
|
m_blobs_pending(0),
|
|
|
|
m_blobs_buffer(0),
|
|
|
|
m_blobs_buffer_size(0),
|
2004-11-17 09:15:53 +01:00
|
|
|
m_dupkey((uint) -1),
|
|
|
|
m_ha_not_exact_count(FALSE),
|
|
|
|
m_force_send(TRUE),
|
2005-04-13 12:42:23 +02:00
|
|
|
m_autoincrement_prefetch((ha_rows) 32),
|
2004-11-22 14:41:46 +01:00
|
|
|
m_transaction_on(TRUE),
|
2005-02-11 22:05:24 +01:00
|
|
|
m_cond_stack(NULL),
|
|
|
|
m_multi_cursor(NULL)
|
2005-03-15 15:03:25 +01:00
|
|
|
{
|
2004-04-30 12:25:31 +02:00
|
|
|
int i;
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster");
|
|
|
|
|
|
|
|
m_tabname[0]= '\0';
|
|
|
|
m_dbname[0]= '\0';
|
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
records= ~(ha_rows)0; // uninitialized
|
2004-04-15 09:14:14 +02:00
|
|
|
block_size= 1024;
|
|
|
|
|
2004-04-30 12:25:31 +02:00
|
|
|
for (i= 0; i < MAX_KEY; i++)
|
|
|
|
{
|
2005-01-26 11:31:46 +01:00
|
|
|
m_index[i].type= UNDEFINED_INDEX;
|
|
|
|
m_index[i].unique_index= NULL;
|
|
|
|
m_index[i].index= NULL;
|
|
|
|
m_index[i].unique_index_attrid_map= NULL;
|
2005-09-15 02:33:28 +02:00
|
|
|
m_index[i].index_stat=NULL;
|
|
|
|
m_index[i].index_stat_cache_entries=0;
|
|
|
|
m_index[i].index_stat_update_freq=0;
|
|
|
|
m_index[i].index_stat_query_count=0;
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Destructor for NDB Cluster table handler
|
|
|
|
*/
|
|
|
|
|
|
|
|
ha_ndbcluster::~ha_ndbcluster()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("~ha_ndbcluster");
|
|
|
|
|
2004-09-03 17:11:09 +02:00
|
|
|
if (m_share)
|
2005-11-06 00:20:37 +01:00
|
|
|
{
|
|
|
|
free_share(&m_share);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
release_metadata();
|
2004-11-03 15:53:26 +01:00
|
|
|
my_free(m_blobs_buffer, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
m_blobs_buffer= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
// Check for open cursor/transaction
|
2004-09-07 12:54:31 +02:00
|
|
|
if (m_active_cursor) {
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(m_active_cursor == NULL);
|
2004-09-07 12:54:31 +02:00
|
|
|
if (m_active_trans) {
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(m_active_trans == NULL);
|
|
|
|
|
2004-12-17 21:13:22 +01:00
|
|
|
// Discard the condition stack
|
|
|
|
DBUG_PRINT("info", ("Clearing condition stack"));
|
|
|
|
cond_clear();
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Open a table for further use
|
|
|
|
- fetch metadata for this table from NDB
|
|
|
|
- check that table exists
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
|
|
|
|
RETURN
|
|
|
|
0 ok
|
|
|
|
< 0 Table has changed
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::open(const char *name, int mode, uint test_if_locked)
|
|
|
|
{
|
2004-10-04 01:20:05 +02:00
|
|
|
int res;
|
2004-04-15 09:14:14 +02:00
|
|
|
KEY *key;
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::open");
|
|
|
|
DBUG_PRINT("enter", ("name: %s mode: %d test_if_locked: %d",
|
|
|
|
name, mode, test_if_locked));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
/*
|
|
|
|
Setup ref_length to make room for the whole
|
|
|
|
primary key to be written in the ref variable
|
|
|
|
*/
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-01-06 12:00:13 +01:00
|
|
|
if (table->s->primary_key != MAX_KEY)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-01-06 12:00:13 +01:00
|
|
|
key= table->key_info+table->s->primary_key;
|
2004-04-15 09:14:14 +02:00
|
|
|
ref_length= key->key_length;
|
|
|
|
DBUG_PRINT("info", (" ref_length: %d", ref_length));
|
|
|
|
}
|
|
|
|
// Init table lock structure
|
|
|
|
if (!(m_share=get_share(name)))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
thr_lock_data_init(&m_share->lock,&m_lock,(void*) 0);
|
|
|
|
|
|
|
|
set_dbname(name);
|
|
|
|
set_tabname(name);
|
|
|
|
|
2004-09-03 17:11:09 +02:00
|
|
|
if (check_ndb_connection()) {
|
2005-11-06 00:20:37 +01:00
|
|
|
free_share(&m_share);
|
|
|
|
m_share= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2004-10-04 01:20:05 +02:00
|
|
|
res= get_metadata(name);
|
|
|
|
if (!res)
|
|
|
|
info(HA_STATUS_VARIABLE | HA_STATUS_CONST);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-07-18 13:31:02 +02:00
|
|
|
|
2004-10-04 01:20:05 +02:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2005-11-25 09:11:52 +01:00
|
|
|
void ha_ndbcluster::set_part_info(partition_info *part_info)
|
|
|
|
{
|
|
|
|
m_part_info= part_info;
|
|
|
|
if (!(m_part_info->part_type == HASH_PARTITION &&
|
|
|
|
m_part_info->list_of_part_fields &&
|
|
|
|
!is_sub_partitioned(m_part_info)))
|
|
|
|
m_use_partition_function= TRUE;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
Close the table
|
|
|
|
- release resources setup by open()
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::close(void)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("close");
|
2005-11-06 00:20:37 +01:00
|
|
|
free_share(&m_share);
|
|
|
|
m_share= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
release_metadata();
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
Thd_ndb* ha_ndbcluster::seize_thd_ndb()
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
Thd_ndb *thd_ndb;
|
|
|
|
DBUG_ENTER("seize_thd_ndb");
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
thd_ndb= new Thd_ndb();
|
2005-03-22 17:42:08 +01:00
|
|
|
thd_ndb->ndb->getDictionary()->set_local_table_data_size(
|
|
|
|
sizeof(Ndb_local_table_statistics)
|
|
|
|
);
|
2004-09-14 14:47:34 +02:00
|
|
|
if (thd_ndb->ndb->init(max_transactions) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
ERR_PRINT(thd_ndb->ndb->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
TODO
|
|
|
|
Alt.1 If init fails because to many allocated Ndb
|
|
|
|
wait on condition for a Ndb object to be released.
|
|
|
|
Alt.2 Seize/release from pool, wait until next release
|
|
|
|
*/
|
2004-09-14 14:47:34 +02:00
|
|
|
delete thd_ndb;
|
|
|
|
thd_ndb= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-09-14 14:47:34 +02:00
|
|
|
DBUG_RETURN(thd_ndb);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
void ha_ndbcluster::release_thd_ndb(Thd_ndb* thd_ndb)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
DBUG_ENTER("release_thd_ndb");
|
|
|
|
delete thd_ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2004-09-20 17:34:32 +02:00
|
|
|
If this thread already has a Thd_ndb object allocated
|
2004-04-15 09:14:14 +02:00
|
|
|
in current THD, reuse it. Otherwise
|
2004-09-20 17:34:32 +02:00
|
|
|
seize a Thd_ndb object, assign it to current THD and use it.
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
*/
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
Ndb* check_ndb_in_thd(THD* thd)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
2004-09-14 14:47:34 +02:00
|
|
|
if (!thd_ndb)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-20 17:34:32 +02:00
|
|
|
if (!(thd_ndb= ha_ndbcluster::seize_thd_ndb()))
|
2005-02-11 22:33:52 +01:00
|
|
|
return NULL;
|
2005-02-17 22:52:40 +01:00
|
|
|
set_thd_ndb(thd, thd_ndb);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
return thd_ndb->ndb;
|
2004-09-13 14:46:38 +02:00
|
|
|
}
|
|
|
|
|
2004-09-20 17:34:32 +02:00
|
|
|
|
2004-12-06 14:51:10 +01:00
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
int ha_ndbcluster::check_ndb_connection(THD* thd)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("check_ndb_connection");
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2004-12-30 19:56:09 +01:00
|
|
|
ndb->setDatabaseName(m_dbname);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-09-20 17:34:32 +02:00
|
|
|
|
2005-02-17 22:52:40 +01:00
|
|
|
int ndbcluster_close_connection(THD *thd)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-02-17 22:52:40 +01:00
|
|
|
Thd_ndb *thd_ndb= get_thd_ndb(thd);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndbcluster_close_connection");
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
if (thd_ndb)
|
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
ha_ndbcluster::release_thd_ndb(thd_ndb);
|
2005-02-17 22:52:40 +01:00
|
|
|
set_thd_ndb(thd, NULL); // not strictly required but does not hurt either
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
2005-02-17 22:52:40 +01:00
|
|
|
DBUG_RETURN(0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Try to discover one table from NDB
|
|
|
|
*/
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
int ndbcluster_discover(THD* thd, const char *db, const char *name,
|
2005-02-16 14:18:32 +01:00
|
|
|
const void** frmblob, uint* frmlen)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
uint len;
|
|
|
|
const void* data;
|
|
|
|
const NDBTAB* tab;
|
2004-09-13 14:46:38 +02:00
|
|
|
Ndb* ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndbcluster_discover");
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_PRINT("enter", ("db: %s, name: %s", db, name));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
|
|
|
ndb->setDatabaseName(db);
|
2004-09-06 20:30:57 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
NDBDICT* dict= ndb->getDictionary();
|
2005-03-22 17:42:08 +01:00
|
|
|
dict->set_local_table_data_size(sizeof(Ndb_local_table_statistics));
|
2004-09-13 14:46:38 +02:00
|
|
|
dict->invalidateTable(name);
|
|
|
|
if (!(tab= dict->getTable(name)))
|
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
2005-11-07 12:19:28 +01:00
|
|
|
if (err.code == 709 || err.code == 723)
|
2005-06-16 15:17:47 +02:00
|
|
|
DBUG_RETURN(-1);
|
2004-09-13 14:46:38 +02:00
|
|
|
ERR_RETURN(err);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("Found table %s", tab->getName()));
|
|
|
|
|
|
|
|
len= tab->getFrmLength();
|
|
|
|
if (len == 0 || tab->getFrmData() == NULL)
|
|
|
|
{
|
BUG#10365 Cluster handler uses non-standard error codes
- Added better error messages when trying to open a table that can't be discovered or unpacked. The most likely cause of this is that it does not have any frm data, probably since it has been created from NdbApi or is a NDB system table.
- Separated functionality that was in ha_create_table_from_engine into two functions. One that checks if the table exists and another one that tries to create the table from the engine.
mysql-test/r/ndb_autodiscover.result:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
mysql-test/t/ndb_autodiscover.test:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
ndb/test/ndbapi/create_tab.cpp:
Set connectstring before creating Ndb object.
sql/ha_ndbcluster.cc:
Rename and use the function ndbcluster_table_exists_in_engine.
Correct return valu from ndbcluster_discover
Remove old code "ndb_discover_tables"
sql/ha_ndbcluster.h:
Rename function ndbcluster_table_exists to ndb ndbcluster_table_exists_in_engine
sql/handler.cc:
Update comment of ha_create_table_from_engine
Remove parameter create_if_found from ha_create_table_from_engine, the function ha_table_exists_in_engine is now used toi check if table is found in engine.
Cleanup return codes from ha_create_table_from_engine.
Change name of ha_table_exists to ha_table_exists_in_engine, update comment and returne codes.
sql/handler.h:
Remove paramter create_if_cound from ha_create_table_from_engine
Rename ha_table_exists to ha_table_exists_in_engine
sql/sql_base.cc:
Use the function ha_table_exists_in_engine to detect if table exists in enegine.
If it exists, call function ha_create_table_from_engine to try and create it.
If create of table fails, set correct error message.
sql/sql_table.cc:
Add comments, remove parameter create_if_found to ha_create_table_from_engine.
When dropping a table, try to discover it from engine. If discover fails, use same error message as if the table didn't exists.
Maybe another message should be displayed here, ex: "Table could not be dropped, unpack failed"
When creating a new table, use ha_table_exists_in_engine to check if a table with the given name already exists.
2005-06-08 13:31:59 +02:00
|
|
|
DBUG_PRINT("error", ("No frm data found."));
|
|
|
|
DBUG_RETURN(1);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
if (unpackfrm(&data, &len, tab->getFrmData()))
|
BUG#10365 Cluster handler uses non-standard error codes
- Added better error messages when trying to open a table that can't be discovered or unpacked. The most likely cause of this is that it does not have any frm data, probably since it has been created from NdbApi or is a NDB system table.
- Separated functionality that was in ha_create_table_from_engine into two functions. One that checks if the table exists and another one that tries to create the table from the engine.
mysql-test/r/ndb_autodiscover.result:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
mysql-test/t/ndb_autodiscover.test:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
ndb/test/ndbapi/create_tab.cpp:
Set connectstring before creating Ndb object.
sql/ha_ndbcluster.cc:
Rename and use the function ndbcluster_table_exists_in_engine.
Correct return valu from ndbcluster_discover
Remove old code "ndb_discover_tables"
sql/ha_ndbcluster.h:
Rename function ndbcluster_table_exists to ndb ndbcluster_table_exists_in_engine
sql/handler.cc:
Update comment of ha_create_table_from_engine
Remove parameter create_if_found from ha_create_table_from_engine, the function ha_table_exists_in_engine is now used toi check if table is found in engine.
Cleanup return codes from ha_create_table_from_engine.
Change name of ha_table_exists to ha_table_exists_in_engine, update comment and returne codes.
sql/handler.h:
Remove paramter create_if_cound from ha_create_table_from_engine
Rename ha_table_exists to ha_table_exists_in_engine
sql/sql_base.cc:
Use the function ha_table_exists_in_engine to detect if table exists in enegine.
If it exists, call function ha_create_table_from_engine to try and create it.
If create of table fails, set correct error message.
sql/sql_table.cc:
Add comments, remove parameter create_if_found to ha_create_table_from_engine.
When dropping a table, try to discover it from engine. If discover fails, use same error message as if the table didn't exists.
Maybe another message should be displayed here, ex: "Table could not be dropped, unpack failed"
When creating a new table, use ha_table_exists_in_engine to check if a table with the given name already exists.
2005-06-08 13:31:59 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Could not unpack table"));
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
*frmlen= len;
|
|
|
|
*frmblob= data;
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2004-09-13 14:46:38 +02:00
|
|
|
Check if a table exists in NDB
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
*/
|
2004-04-15 09:14:14 +02:00
|
|
|
|
BUG#10365 Cluster handler uses non-standard error codes
- Added better error messages when trying to open a table that can't be discovered or unpacked. The most likely cause of this is that it does not have any frm data, probably since it has been created from NdbApi or is a NDB system table.
- Separated functionality that was in ha_create_table_from_engine into two functions. One that checks if the table exists and another one that tries to create the table from the engine.
mysql-test/r/ndb_autodiscover.result:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
mysql-test/t/ndb_autodiscover.test:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
ndb/test/ndbapi/create_tab.cpp:
Set connectstring before creating Ndb object.
sql/ha_ndbcluster.cc:
Rename and use the function ndbcluster_table_exists_in_engine.
Correct return valu from ndbcluster_discover
Remove old code "ndb_discover_tables"
sql/ha_ndbcluster.h:
Rename function ndbcluster_table_exists to ndb ndbcluster_table_exists_in_engine
sql/handler.cc:
Update comment of ha_create_table_from_engine
Remove parameter create_if_found from ha_create_table_from_engine, the function ha_table_exists_in_engine is now used toi check if table is found in engine.
Cleanup return codes from ha_create_table_from_engine.
Change name of ha_table_exists to ha_table_exists_in_engine, update comment and returne codes.
sql/handler.h:
Remove paramter create_if_cound from ha_create_table_from_engine
Rename ha_table_exists to ha_table_exists_in_engine
sql/sql_base.cc:
Use the function ha_table_exists_in_engine to detect if table exists in enegine.
If it exists, call function ha_create_table_from_engine to try and create it.
If create of table fails, set correct error message.
sql/sql_table.cc:
Add comments, remove parameter create_if_found to ha_create_table_from_engine.
When dropping a table, try to discover it from engine. If discover fails, use same error message as if the table didn't exists.
Maybe another message should be displayed here, ex: "Table could not be dropped, unpack failed"
When creating a new table, use ha_table_exists_in_engine to check if a table with the given name already exists.
2005-06-08 13:31:59 +02:00
|
|
|
int ndbcluster_table_exists_in_engine(THD* thd, const char *db, const char *name)
|
2004-09-13 14:46:38 +02:00
|
|
|
{
|
|
|
|
const NDBTAB* tab;
|
|
|
|
Ndb* ndb;
|
BUG#10365 Cluster handler uses non-standard error codes
- Added better error messages when trying to open a table that can't be discovered or unpacked. The most likely cause of this is that it does not have any frm data, probably since it has been created from NdbApi or is a NDB system table.
- Separated functionality that was in ha_create_table_from_engine into two functions. One that checks if the table exists and another one that tries to create the table from the engine.
mysql-test/r/ndb_autodiscover.result:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
mysql-test/t/ndb_autodiscover.test:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
ndb/test/ndbapi/create_tab.cpp:
Set connectstring before creating Ndb object.
sql/ha_ndbcluster.cc:
Rename and use the function ndbcluster_table_exists_in_engine.
Correct return valu from ndbcluster_discover
Remove old code "ndb_discover_tables"
sql/ha_ndbcluster.h:
Rename function ndbcluster_table_exists to ndb ndbcluster_table_exists_in_engine
sql/handler.cc:
Update comment of ha_create_table_from_engine
Remove parameter create_if_found from ha_create_table_from_engine, the function ha_table_exists_in_engine is now used toi check if table is found in engine.
Cleanup return codes from ha_create_table_from_engine.
Change name of ha_table_exists to ha_table_exists_in_engine, update comment and returne codes.
sql/handler.h:
Remove paramter create_if_cound from ha_create_table_from_engine
Rename ha_table_exists to ha_table_exists_in_engine
sql/sql_base.cc:
Use the function ha_table_exists_in_engine to detect if table exists in enegine.
If it exists, call function ha_create_table_from_engine to try and create it.
If create of table fails, set correct error message.
sql/sql_table.cc:
Add comments, remove parameter create_if_found to ha_create_table_from_engine.
When dropping a table, try to discover it from engine. If discover fails, use same error message as if the table didn't exists.
Maybe another message should be displayed here, ex: "Table could not be dropped, unpack failed"
When creating a new table, use ha_table_exists_in_engine to check if a table with the given name already exists.
2005-06-08 13:31:59 +02:00
|
|
|
DBUG_ENTER("ndbcluster_table_exists_in_engine");
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("enter", ("db: %s, name: %s", db, name));
|
2004-09-13 14:46:38 +02:00
|
|
|
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2004-09-13 14:46:38 +02:00
|
|
|
ndb->setDatabaseName(db);
|
|
|
|
|
|
|
|
NDBDICT* dict= ndb->getDictionary();
|
2005-03-22 17:42:08 +01:00
|
|
|
dict->set_local_table_data_size(sizeof(Ndb_local_table_statistics));
|
2004-09-13 14:46:38 +02:00
|
|
|
dict->invalidateTable(name);
|
|
|
|
if (!(tab= dict->getTable(name)))
|
2005-03-15 15:03:25 +01:00
|
|
|
{
|
2004-09-13 14:46:38 +02:00
|
|
|
const NdbError err= dict->getNdbError();
|
2005-11-07 12:19:28 +01:00
|
|
|
if (err.code == 709 || err.code == 723)
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
ERR_RETURN(err);
|
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_PRINT("info", ("Found table %s", tab->getName()));
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
extern "C" byte* tables_get_key(const char *entry, uint *length,
|
2005-02-16 14:18:32 +01:00
|
|
|
my_bool not_used __attribute__((unused)))
|
2004-09-21 12:13:58 +02:00
|
|
|
{
|
|
|
|
*length= strlen(entry);
|
|
|
|
return (byte*) entry;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2005-04-22 15:28:23 +02:00
|
|
|
/*
|
|
|
|
Drop a database in NDB Cluster
|
2005-11-07 16:25:06 +01:00
|
|
|
NOTE add a dummy void function, since stupid handlerton is returning void instead of int...
|
|
|
|
*/
|
2005-04-22 15:28:23 +02:00
|
|
|
|
2005-11-07 16:25:06 +01:00
|
|
|
int ndbcluster_drop_database_impl(const char *path)
|
2005-04-22 15:28:23 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_drop_database");
|
|
|
|
THD *thd= current_thd;
|
|
|
|
char dbname[FN_HEADLEN];
|
|
|
|
Ndb* ndb;
|
|
|
|
NdbDictionary::Dictionary::List list;
|
|
|
|
uint i;
|
|
|
|
char *tabname;
|
|
|
|
List<char> drop_list;
|
2005-04-25 11:31:16 +02:00
|
|
|
int ret= 0;
|
2005-04-22 15:28:23 +02:00
|
|
|
ha_ndbcluster::set_dbname(path, (char *)&dbname);
|
|
|
|
DBUG_PRINT("enter", ("db: %s", dbname));
|
|
|
|
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
2005-11-07 16:25:06 +01:00
|
|
|
DBUG_RETURN(-1);
|
2005-04-22 15:28:23 +02:00
|
|
|
|
|
|
|
// List tables in NDB
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
|
|
|
if (dict->listObjects(list,
|
|
|
|
NdbDictionary::Object::UserTable) != 0)
|
2005-11-07 16:25:06 +01:00
|
|
|
DBUG_RETURN(-1);
|
2005-04-22 15:28:23 +02:00
|
|
|
for (i= 0 ; i < list.count ; i++)
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
NdbDictionary::Dictionary::List::Element& elmt= list.elements[i];
|
|
|
|
DBUG_PRINT("info", ("Found %s/%s in NDB", elmt.database, elmt.name));
|
2005-04-22 15:28:23 +02:00
|
|
|
|
|
|
|
// Add only tables that belongs to db
|
2005-11-06 00:20:37 +01:00
|
|
|
if (my_strcasecmp(system_charset_info, elmt.database, dbname))
|
2005-04-22 15:28:23 +02:00
|
|
|
continue;
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("%s must be dropped", elmt.name));
|
|
|
|
drop_list.push_back(thd->strdup(elmt.name));
|
2005-04-22 15:28:23 +02:00
|
|
|
}
|
|
|
|
// Drop any tables belonging to database
|
2005-11-06 00:20:37 +01:00
|
|
|
char full_path[FN_REFLEN];
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
char *tmp= strxnmov(full_path, FN_REFLEN-1, share_prefix, dbname, "/",
|
|
|
|
NullS);
|
2005-04-22 15:28:23 +02:00
|
|
|
ndb->setDatabaseName(dbname);
|
|
|
|
List_iterator_fast<char> it(drop_list);
|
|
|
|
while ((tabname=it++))
|
2005-04-25 11:31:16 +02:00
|
|
|
{
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
strxnmov(tmp, FN_REFLEN - (tmp - full_path)-1, tabname, NullS);
|
2005-11-06 00:20:37 +01:00
|
|
|
if (ha_ndbcluster::delete_table(0, ndb, full_path, dbname, tabname))
|
2005-04-25 11:31:16 +02:00
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
2005-11-07 12:19:28 +01:00
|
|
|
if (err.code != 709 && err.code != 723)
|
2005-04-25 11:34:47 +02:00
|
|
|
{
|
|
|
|
ERR_PRINT(err);
|
2005-04-25 11:31:16 +02:00
|
|
|
ret= ndb_to_mysql_error(&err);
|
2005-04-25 11:34:47 +02:00
|
|
|
}
|
2005-04-25 11:31:16 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
DBUG_RETURN(ret);
|
2005-04-22 15:28:23 +02:00
|
|
|
}
|
|
|
|
|
2005-11-07 16:25:06 +01:00
|
|
|
void ndbcluster_drop_database(char *path)
|
|
|
|
{
|
|
|
|
ndbcluster_drop_database_impl(path);
|
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
/*
|
|
|
|
find all tables in ndb and discover those needed
|
|
|
|
*/
|
|
|
|
static int ndbcluster_find_all_files(THD *thd)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_find_all_files");
|
|
|
|
Ndb* ndb;
|
|
|
|
char key[FN_REFLEN];
|
|
|
|
NdbDictionary::Dictionary::List list;
|
|
|
|
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
|
|
|
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
|
|
|
|
|
|
|
int unhandled, retries= 5;
|
|
|
|
do
|
|
|
|
{
|
|
|
|
if (dict->listObjects(list, NdbDictionary::Object::UserTable) != 0)
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
unhandled= 0;
|
|
|
|
for (uint i= 0 ; i < list.count ; i++)
|
|
|
|
{
|
|
|
|
NDBDICT::List::Element& elmt= list.elements[i];
|
|
|
|
DBUG_PRINT("info", ("Found %s.%s in NDB", elmt.database, elmt.name));
|
|
|
|
if (!(elmt.state == NDBOBJ::StateBuilding ||
|
|
|
|
elmt.state == NDBOBJ::StateOnline))
|
|
|
|
{
|
|
|
|
sql_print_information("NDB: skipping setup table %s.%s, in state %d",
|
|
|
|
elmt.database, elmt.name, elmt.state);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
ndb->setDatabaseName(elmt.database);
|
|
|
|
const NDBTAB *ndbtab;
|
|
|
|
|
|
|
|
if (!(ndbtab= dict->getTable(elmt.name)))
|
|
|
|
{
|
|
|
|
sql_print_error("NDB: failed to setup table %s.%s, error: %d, %s",
|
|
|
|
elmt.database, elmt.name,
|
|
|
|
dict->getNdbError().code,
|
|
|
|
dict->getNdbError().message);
|
|
|
|
unhandled++;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ndbtab->getFrmLength() == 0)
|
|
|
|
continue;
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
strxnmov(key, FN_LEN-1, mysql_data_home, "/",
|
2005-11-06 00:20:37 +01:00
|
|
|
elmt.database, "/", elmt.name, NullS);
|
|
|
|
const void *data= 0, *pack_data= 0;
|
|
|
|
uint length, pack_length;
|
|
|
|
int discover= 0;
|
|
|
|
if (readfrm(key, &data, &length) ||
|
|
|
|
packfrm(data, length, &pack_data, &pack_length))
|
|
|
|
{
|
|
|
|
discover= 1;
|
|
|
|
sql_print_information("NDB: missing frm for %s.%s, discovering...",
|
|
|
|
elmt.database, elmt.name);
|
|
|
|
}
|
|
|
|
else if (cmp_frm(ndbtab, pack_data, pack_length))
|
|
|
|
{
|
|
|
|
discover= 1;
|
|
|
|
sql_print_information("NDB: mismatch in frm for %s.%s, discovering...",
|
|
|
|
elmt.database, elmt.name);
|
|
|
|
}
|
|
|
|
my_free((char*) data, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
my_free((char*) pack_data, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
|
|
|
|
if (discover)
|
|
|
|
{
|
|
|
|
/* ToDo 4.1 database needs to be created if missing */
|
|
|
|
pthread_mutex_lock(&LOCK_open);
|
|
|
|
if (ha_create_table_from_engine(thd, elmt.database, elmt.name))
|
|
|
|
{
|
|
|
|
/* ToDo 4.1 handle error */
|
|
|
|
}
|
|
|
|
pthread_mutex_unlock(&LOCK_open);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
while (unhandled && retries--);
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2005-04-22 15:28:23 +02:00
|
|
|
|
2004-09-21 12:13:58 +02:00
|
|
|
int ndbcluster_find_files(THD *thd,const char *db,const char *path,
|
2005-02-16 14:18:32 +01:00
|
|
|
const char *wild, bool dir, List<char> *files)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-10-25 22:35:46 +02:00
|
|
|
DBUG_ENTER("ndbcluster_find_files");
|
|
|
|
DBUG_PRINT("enter", ("db: %s", db));
|
|
|
|
{ // extra bracket to avoid gcc 2.95.3 warning
|
2004-04-15 09:14:14 +02:00
|
|
|
uint i;
|
2004-09-13 14:46:38 +02:00
|
|
|
Ndb* ndb;
|
2004-09-21 12:13:58 +02:00
|
|
|
char name[FN_REFLEN];
|
2004-09-26 16:11:24 +02:00
|
|
|
HASH ndb_tables, ok_tables;
|
2005-11-06 00:20:37 +01:00
|
|
|
NDBDICT::List list;
|
2004-09-13 14:46:38 +02:00
|
|
|
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
|
|
|
|
2004-09-21 12:13:58 +02:00
|
|
|
if (dir)
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_RETURN(0); // Discover of databases not yet supported
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// List tables in NDB
|
2004-09-13 14:46:38 +02:00
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (dict->listObjects(list,
|
2005-02-16 14:18:32 +01:00
|
|
|
NdbDictionary::Object::UserTable) != 0)
|
2004-09-13 06:40:25 +02:00
|
|
|
ERR_RETURN(dict->getNdbError());
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
if (hash_init(&ndb_tables, system_charset_info,list.count,0,0,
|
2005-02-16 14:18:32 +01:00
|
|
|
(hash_get_key)tables_get_key,0,0))
|
2004-09-26 16:11:24 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Failed to init HASH ndb_tables"));
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (hash_init(&ok_tables, system_charset_info,32,0,0,
|
2005-02-16 14:18:32 +01:00
|
|
|
(hash_get_key)tables_get_key,0,0))
|
2004-09-26 16:11:24 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Failed to init HASH ok_tables"));
|
|
|
|
hash_free(&ndb_tables);
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
for (i= 0 ; i < list.count ; i++)
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
NDBDICT::List::Element& elmt= list.elements[i];
|
|
|
|
DBUG_PRINT("info", ("Found %s/%s in NDB", elmt.database, elmt.name));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-10-08 13:37:13 +02:00
|
|
|
// Add only tables that belongs to db
|
2005-11-06 00:20:37 +01:00
|
|
|
if (my_strcasecmp(system_charset_info, elmt.database, db))
|
2004-10-08 13:37:13 +02:00
|
|
|
continue;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Apply wildcard to list of tables in NDB
|
2004-09-21 12:13:58 +02:00
|
|
|
if (wild)
|
2004-09-13 14:46:38 +02:00
|
|
|
{
|
2004-09-21 12:13:58 +02:00
|
|
|
if (lower_case_table_names)
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (wild_case_compare(files_charset_info, elmt.name, wild))
|
2005-02-16 14:18:32 +01:00
|
|
|
continue;
|
2004-09-21 12:13:58 +02:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
else if (wild_compare(elmt.name,wild,0))
|
2005-02-16 14:18:32 +01:00
|
|
|
continue;
|
2004-09-21 12:13:58 +02:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("Inserting %s into ndb_tables hash", elmt.name));
|
|
|
|
my_hash_insert(&ndb_tables, (byte*)thd->strdup(elmt.name));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
char *file_name;
|
|
|
|
List_iterator<char> it(*files);
|
|
|
|
List<char> delete_list;
|
|
|
|
while ((file_name=it++))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s", file_name));
|
|
|
|
if (hash_search(&ndb_tables, file_name, strlen(file_name)))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s existed in NDB _and_ on disk ", file_name));
|
|
|
|
// File existed in NDB and as frm file, put in ok_tables list
|
|
|
|
my_hash_insert(&ok_tables, (byte*)file_name);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// File is not in NDB, check for .ndb file with this name
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
(void)strxnmov(name, FN_REFLEN-1,
|
2005-02-16 14:18:32 +01:00
|
|
|
mysql_data_home,"/",db,"/",file_name,ha_ndb_ext,NullS);
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_PRINT("info", ("Check access for %s", name));
|
2004-09-21 12:13:58 +02:00
|
|
|
if (access(name, F_OK))
|
2004-09-26 16:11:24 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s did not exist on disk", name));
|
|
|
|
// .ndb file did not exist on disk, another table type
|
|
|
|
continue;
|
2004-09-13 14:46:38 +02:00
|
|
|
}
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_PRINT("info", ("%s existed on disk", name));
|
|
|
|
// The .ndb file exists on disk, but it's not in list of tables in ndb
|
|
|
|
// Verify that handler agrees table is gone.
|
BUG#10365 Cluster handler uses non-standard error codes
- Added better error messages when trying to open a table that can't be discovered or unpacked. The most likely cause of this is that it does not have any frm data, probably since it has been created from NdbApi or is a NDB system table.
- Separated functionality that was in ha_create_table_from_engine into two functions. One that checks if the table exists and another one that tries to create the table from the engine.
mysql-test/r/ndb_autodiscover.result:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
mysql-test/t/ndb_autodiscover.test:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
ndb/test/ndbapi/create_tab.cpp:
Set connectstring before creating Ndb object.
sql/ha_ndbcluster.cc:
Rename and use the function ndbcluster_table_exists_in_engine.
Correct return valu from ndbcluster_discover
Remove old code "ndb_discover_tables"
sql/ha_ndbcluster.h:
Rename function ndbcluster_table_exists to ndb ndbcluster_table_exists_in_engine
sql/handler.cc:
Update comment of ha_create_table_from_engine
Remove parameter create_if_found from ha_create_table_from_engine, the function ha_table_exists_in_engine is now used toi check if table is found in engine.
Cleanup return codes from ha_create_table_from_engine.
Change name of ha_table_exists to ha_table_exists_in_engine, update comment and returne codes.
sql/handler.h:
Remove paramter create_if_cound from ha_create_table_from_engine
Rename ha_table_exists to ha_table_exists_in_engine
sql/sql_base.cc:
Use the function ha_table_exists_in_engine to detect if table exists in enegine.
If it exists, call function ha_create_table_from_engine to try and create it.
If create of table fails, set correct error message.
sql/sql_table.cc:
Add comments, remove parameter create_if_found to ha_create_table_from_engine.
When dropping a table, try to discover it from engine. If discover fails, use same error message as if the table didn't exists.
Maybe another message should be displayed here, ex: "Table could not be dropped, unpack failed"
When creating a new table, use ha_table_exists_in_engine to check if a table with the given name already exists.
2005-06-08 13:31:59 +02:00
|
|
|
if (ndbcluster_table_exists_in_engine(thd, db, file_name) == 0)
|
2004-09-26 16:11:24 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("NDB says %s does not exists", file_name));
|
|
|
|
it.remove();
|
|
|
|
// Put in list of tables to remove from disk
|
|
|
|
delete_list.push_back(thd->strdup(file_name));
|
|
|
|
}
|
|
|
|
}
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Check for new files to discover
|
|
|
|
DBUG_PRINT("info", ("Checking for new files to discover"));
|
|
|
|
List<char> create_list;
|
|
|
|
for (i= 0 ; i < ndb_tables.records ; i++)
|
2004-09-21 12:13:58 +02:00
|
|
|
{
|
2004-09-26 16:11:24 +02:00
|
|
|
file_name= hash_element(&ndb_tables, i);
|
|
|
|
if (!hash_search(&ok_tables, file_name, strlen(file_name)))
|
2004-09-21 12:13:58 +02:00
|
|
|
{
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
strxnmov(name, sizeof(name)-1,
|
2005-11-06 00:20:37 +01:00
|
|
|
mysql_data_home, "/", db, "/", file_name, reg_ext, NullS);
|
|
|
|
if (access(name, F_OK))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s must be discovered", file_name));
|
|
|
|
// File is in list of ndb tables and not in ok_tables
|
|
|
|
// This table need to be created
|
|
|
|
create_list.push_back(thd->strdup(file_name));
|
|
|
|
}
|
2004-09-26 16:11:24 +02:00
|
|
|
}
|
|
|
|
}
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Lock mutex before deleting and creating frm files
|
|
|
|
pthread_mutex_lock(&LOCK_open);
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
if (!global_read_lock)
|
|
|
|
{
|
|
|
|
// Delete old files
|
|
|
|
List_iterator_fast<char> it3(delete_list);
|
|
|
|
while ((file_name=it3++))
|
2005-02-17 14:22:44 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Remove table %s/%s", db, file_name));
|
2004-09-26 16:11:24 +02:00
|
|
|
// Delete the table and all related files
|
|
|
|
TABLE_LIST table_list;
|
|
|
|
bzero((char*) &table_list,sizeof(table_list));
|
|
|
|
table_list.db= (char*) db;
|
2005-01-06 12:00:13 +01:00
|
|
|
table_list.alias= table_list.table_name= (char*)file_name;
|
2005-02-17 14:22:44 +01:00
|
|
|
(void)mysql_rm_table_part2(thd, &table_list,
|
2005-02-21 16:13:29 +01:00
|
|
|
/* if_exists */ FALSE,
|
|
|
|
/* drop_temporary */ FALSE,
|
|
|
|
/* drop_view */ FALSE,
|
|
|
|
/* dont_log_query*/ TRUE);
|
2005-02-17 14:22:44 +01:00
|
|
|
/* Clear error message that is returned when table is deleted */
|
|
|
|
thd->clear_error();
|
2004-09-21 12:13:58 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Create new files
|
|
|
|
List_iterator_fast<char> it2(create_list);
|
|
|
|
while ((file_name=it2++))
|
|
|
|
{
|
BUG#9626 Several serious errors reported by Valgrind in latest 5.0 bk tree
- Fix several valgrind warnings.
ndb/include/kernel/AttributeList.hpp:
Include ndb_limits.h
ndb/include/ndbapi/Ndb.hpp:
Add new function internalize_table_name
Proper formatting
ndb/include/util/SimpleProperties.hpp:
Add function "add" for adding strings/data to word buffers without reading after the string/data
Fix comment for UtilBufferWriter
ndb/src/common/util/SimpleProperties.cpp:
Add function "add" for adding strings/data to word buffers without reading after the string/data
ndb/src/ndbapi/DictCache.cpp:
Add DBUG printouts to all functions in GlobalDictCache
Change ndbout_c to DBUG_PRINT
Add function GlobalDictCache::printCache
ndb/src/ndbapi/DictCache.hpp:
Add function GlobalDictCache::printCache
ndb/src/ndbapi/Ndb.cpp:
Change DEBUG_TRACE to DBUG_PRINT, DBUG_ENTER
Use new function internalize_table_name, and create internal tabname on stack
ndb/src/ndbapi/NdbDictionary.cpp:
Add DBUG_ENTER calls
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Pass BaseString as reference when passing internal table name to internal functions.
Remove check for exceeding MAX_SECTION_SIZE.
Remove pekkas fix that saves internalName in save_me variable, not needed when tablename is stack variable.
NdbDictInterface::gettable(int tableid, ...) - Not used
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
Remove include of ndb_limits.h
Use BaseString& for internal functions
Remove m_namebuf, not needed, uses m_buffer
Formatting
NdbDictInterface::getTable(int tableId, ...) - Not used
ndb/src/ndbapi/NdbImpl.hpp:
Remove buffer for m_internalname, placed on stack of caller.
Moved impl of internalize_table_name and internalize_index_anme to Ndb::internalize_table_name and Ndb::internalize_index_name
ndb/src/ndbapi/NdbLinHash.hpp:
Save also the terminating null character in chain->str to get nicer printouts.
ndb/src/ndbapi/NdbTransaction.cpp:
Remove hardcoded buffer sizes.
Remove include of ndb_limits.h
sql/ha_ndbcluster.cc:
Fix "typo", use file_name var since that is the current files name,
2005-06-02 14:32:06 +02:00
|
|
|
DBUG_PRINT("info", ("Table %s need discovery", file_name));
|
BUG#10365 Cluster handler uses non-standard error codes
- Added better error messages when trying to open a table that can't be discovered or unpacked. The most likely cause of this is that it does not have any frm data, probably since it has been created from NdbApi or is a NDB system table.
- Separated functionality that was in ha_create_table_from_engine into two functions. One that checks if the table exists and another one that tries to create the table from the engine.
mysql-test/r/ndb_autodiscover.result:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
mysql-test/t/ndb_autodiscover.test:
Add tests for reading from a table that can't be discovered(SYSTAB_0)
Discovery is not performed during create table anymore.
ndb/test/ndbapi/create_tab.cpp:
Set connectstring before creating Ndb object.
sql/ha_ndbcluster.cc:
Rename and use the function ndbcluster_table_exists_in_engine.
Correct return valu from ndbcluster_discover
Remove old code "ndb_discover_tables"
sql/ha_ndbcluster.h:
Rename function ndbcluster_table_exists to ndb ndbcluster_table_exists_in_engine
sql/handler.cc:
Update comment of ha_create_table_from_engine
Remove parameter create_if_found from ha_create_table_from_engine, the function ha_table_exists_in_engine is now used toi check if table is found in engine.
Cleanup return codes from ha_create_table_from_engine.
Change name of ha_table_exists to ha_table_exists_in_engine, update comment and returne codes.
sql/handler.h:
Remove paramter create_if_cound from ha_create_table_from_engine
Rename ha_table_exists to ha_table_exists_in_engine
sql/sql_base.cc:
Use the function ha_table_exists_in_engine to detect if table exists in enegine.
If it exists, call function ha_create_table_from_engine to try and create it.
If create of table fails, set correct error message.
sql/sql_table.cc:
Add comments, remove parameter create_if_found to ha_create_table_from_engine.
When dropping a table, try to discover it from engine. If discover fails, use same error message as if the table didn't exists.
Maybe another message should be displayed here, ex: "Table could not be dropped, unpack failed"
When creating a new table, use ha_table_exists_in_engine to check if a table with the given name already exists.
2005-06-08 13:31:59 +02:00
|
|
|
if (ha_create_table_from_engine(thd, db, file_name) == 0)
|
2004-10-07 09:16:46 +02:00
|
|
|
files->push_back(thd->strdup(file_name));
|
2004-09-26 16:11:24 +02:00
|
|
|
}
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
pthread_mutex_unlock(&LOCK_open);
|
2004-09-26 16:11:24 +02:00
|
|
|
|
|
|
|
hash_free(&ok_tables);
|
2004-09-21 12:13:58 +02:00
|
|
|
hash_free(&ndb_tables);
|
2004-10-25 22:35:46 +02:00
|
|
|
} // extra bracket to avoid gcc 2.95.3 warning
|
2004-09-21 12:13:58 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Initialise all gloal variables before creating
|
|
|
|
a NDB Cluster table handler
|
|
|
|
*/
|
|
|
|
|
2005-01-14 12:32:33 +01:00
|
|
|
/* Call back after cluster connect */
|
|
|
|
static int connect_callback()
|
|
|
|
{
|
|
|
|
update_status_variables(g_ndb_cluster_connection);
|
2005-11-06 00:20:37 +01:00
|
|
|
pthread_cond_signal(&COND_ndb_util_thread);
|
2005-01-14 12:32:33 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-10-03 04:44:28 +02:00
|
|
|
bool ndbcluster_init()
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-13 06:40:25 +02:00
|
|
|
int res;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndbcluster_init");
|
2005-10-03 04:44:28 +02:00
|
|
|
|
|
|
|
if (have_ndbcluster != SHOW_OPTION_YES)
|
|
|
|
goto ndbcluster_init_error;
|
|
|
|
|
2004-08-20 18:10:47 +02:00
|
|
|
// Set connectstring if specified
|
2004-12-20 12:36:14 +01:00
|
|
|
if (opt_ndbcluster_connectstring != 0)
|
|
|
|
DBUG_PRINT("connectstring", ("%s", opt_ndbcluster_connectstring));
|
2004-09-06 20:30:57 +02:00
|
|
|
if ((g_ndb_cluster_connection=
|
2004-12-20 12:36:14 +01:00
|
|
|
new Ndb_cluster_connection(opt_ndbcluster_connectstring)) == 0)
|
2004-09-06 20:30:57 +02:00
|
|
|
{
|
2004-12-20 12:36:14 +01:00
|
|
|
DBUG_PRINT("error",("Ndb_cluster_connection(%s)",
|
2005-02-16 14:18:32 +01:00
|
|
|
opt_ndbcluster_connectstring));
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-08-20 18:10:47 +02:00
|
|
|
}
|
2004-09-13 06:40:25 +02:00
|
|
|
|
2004-12-20 12:36:14 +01:00
|
|
|
g_ndb_cluster_connection->set_optimized_node_selection
|
|
|
|
(opt_ndb_optimized_node_selection);
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Create a Ndb object to open the connection to NDB
|
2005-02-21 23:29:07 +01:00
|
|
|
if ( (g_ndb= new Ndb(g_ndb_cluster_connection, "sys")) == 0 )
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("failed to create global ndb object"));
|
|
|
|
goto ndbcluster_init_error;
|
|
|
|
}
|
2005-03-22 17:42:08 +01:00
|
|
|
g_ndb->getDictionary()->set_local_table_data_size(sizeof(Ndb_local_table_statistics));
|
2004-04-15 09:14:14 +02:00
|
|
|
if (g_ndb->init() != 0)
|
|
|
|
{
|
|
|
|
ERR_PRINT (g_ndb->getNdbError());
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-09-13 06:40:25 +02:00
|
|
|
|
2004-11-30 18:47:47 +01:00
|
|
|
if ((res= g_ndb_cluster_connection->connect(0,0,0)) == 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-01-14 12:32:33 +01:00
|
|
|
connect_callback();
|
2004-11-30 18:47:47 +01:00
|
|
|
DBUG_PRINT("info",("NDBCLUSTER storage engine at %s on port %d",
|
2005-02-16 14:18:32 +01:00
|
|
|
g_ndb_cluster_connection->get_connected_host(),
|
|
|
|
g_ndb_cluster_connection->get_connected_port()));
|
2005-02-24 18:58:56 +01:00
|
|
|
g_ndb_cluster_connection->wait_until_ready(10,3);
|
2004-09-13 06:40:25 +02:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
else if (res == 1)
|
2004-09-13 06:40:25 +02:00
|
|
|
{
|
2005-01-14 12:32:33 +01:00
|
|
|
if (g_ndb_cluster_connection->start_connect_thread(connect_callback))
|
2005-01-11 17:00:31 +01:00
|
|
|
{
|
2004-09-27 02:24:23 +02:00
|
|
|
DBUG_PRINT("error", ("g_ndb_cluster_connection->start_connect_thread()"));
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
|
|
|
}
|
2005-01-11 17:00:31 +01:00
|
|
|
#ifndef DBUG_OFF
|
2004-11-30 18:47:47 +01:00
|
|
|
{
|
|
|
|
char buf[1024];
|
2005-01-11 17:00:31 +01:00
|
|
|
DBUG_PRINT("info",
|
2005-02-16 14:18:32 +01:00
|
|
|
("NDBCLUSTER storage engine not started, "
|
|
|
|
"will connect using %s",
|
|
|
|
g_ndb_cluster_connection->
|
|
|
|
get_connectstring(buf,sizeof(buf))));
|
2004-09-27 02:24:23 +02:00
|
|
|
}
|
2005-01-11 17:00:31 +01:00
|
|
|
#endif
|
2004-09-13 06:40:25 +02:00
|
|
|
}
|
2004-09-27 02:24:23 +02:00
|
|
|
else
|
2004-09-13 06:40:25 +02:00
|
|
|
{
|
|
|
|
DBUG_ASSERT(res == -1);
|
|
|
|
DBUG_PRINT("error", ("permanent error"));
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-09-13 06:40:25 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
(void) hash_init(&ndbcluster_open_tables,system_charset_info,32,0,0,
|
|
|
|
(hash_get_key) ndbcluster_get_key,0,0);
|
|
|
|
pthread_mutex_init(&ndbcluster_mutex,MY_MUTEX_INIT_FAST);
|
2005-02-11 22:33:52 +01:00
|
|
|
pthread_mutex_init(&LOCK_ndb_util_thread, MY_MUTEX_INIT_FAST);
|
|
|
|
pthread_cond_init(&COND_ndb_util_thread, NULL);
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
// Create utility thread
|
|
|
|
pthread_t tmp;
|
|
|
|
if (pthread_create(&tmp, &connection_attrib, ndb_util_thread_func, 0))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Could not create ndb utility thread"));
|
2005-02-21 23:29:07 +01:00
|
|
|
hash_free(&ndbcluster_open_tables);
|
|
|
|
pthread_mutex_destroy(&ndbcluster_mutex);
|
|
|
|
pthread_mutex_destroy(&LOCK_ndb_util_thread);
|
|
|
|
pthread_cond_destroy(&COND_ndb_util_thread);
|
2005-02-11 22:33:52 +01:00
|
|
|
goto ndbcluster_init_error;
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
ndbcluster_inited= 1;
|
2005-10-03 04:44:28 +02:00
|
|
|
DBUG_RETURN(FALSE);
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-10-03 04:44:28 +02:00
|
|
|
ndbcluster_init_error:
|
2005-07-04 02:42:33 +02:00
|
|
|
if (g_ndb)
|
2005-02-21 23:29:07 +01:00
|
|
|
delete g_ndb;
|
|
|
|
g_ndb= NULL;
|
|
|
|
if (g_ndb_cluster_connection)
|
|
|
|
delete g_ndb_cluster_connection;
|
|
|
|
g_ndb_cluster_connection= NULL;
|
2005-10-03 04:44:28 +02:00
|
|
|
have_ndbcluster= SHOW_OPTION_DISABLED; // If we couldn't use handler
|
|
|
|
DBUG_RETURN(TRUE);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
End use of the NDB Cluster table handler
|
|
|
|
- free all global variables allocated by
|
2005-02-11 22:33:52 +01:00
|
|
|
ndbcluster_init()
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2005-11-07 16:25:06 +01:00
|
|
|
int ndbcluster_end(ha_panic_function type)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_end");
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-02-21 23:29:07 +01:00
|
|
|
if (!ndbcluster_inited)
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
// Kill ndb utility thread
|
|
|
|
(void) pthread_mutex_lock(&LOCK_ndb_util_thread);
|
|
|
|
DBUG_PRINT("exit",("killing ndb util thread: %lx", ndb_util_thread));
|
|
|
|
(void) pthread_cond_signal(&COND_ndb_util_thread);
|
|
|
|
(void) pthread_mutex_unlock(&LOCK_ndb_util_thread);
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if (g_ndb)
|
2005-11-17 12:45:23 +01:00
|
|
|
{
|
|
|
|
#ifndef DBUG_OFF
|
2006-01-05 23:47:49 +01:00
|
|
|
Ndb::Free_list_usage tmp;
|
|
|
|
tmp.m_name= 0;
|
2005-11-17 12:45:23 +01:00
|
|
|
while (g_ndb->get_free_list_usage(&tmp))
|
|
|
|
{
|
|
|
|
uint leaked= (uint) tmp.m_created - tmp.m_free;
|
|
|
|
if (leaked)
|
|
|
|
fprintf(stderr, "NDB: Found %u %s%s that %s not been released\n",
|
|
|
|
leaked, tmp.m_name,
|
|
|
|
(leaked == 1)?"":"'s",
|
|
|
|
(leaked == 1)?"has":"have");
|
|
|
|
}
|
|
|
|
#endif
|
2004-09-13 06:40:25 +02:00
|
|
|
delete g_ndb;
|
2006-01-05 23:47:49 +01:00
|
|
|
g_ndb= NULL;
|
2005-11-17 12:45:23 +01:00
|
|
|
}
|
2006-01-05 23:47:49 +01:00
|
|
|
delete g_ndb_cluster_connection;
|
2004-09-06 20:30:57 +02:00
|
|
|
g_ndb_cluster_connection= NULL;
|
2005-02-21 23:29:07 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
hash_free(&ndbcluster_open_tables);
|
|
|
|
pthread_mutex_destroy(&ndbcluster_mutex);
|
2005-02-11 22:33:52 +01:00
|
|
|
pthread_mutex_destroy(&LOCK_ndb_util_thread);
|
|
|
|
pthread_cond_destroy(&COND_ndb_util_thread);
|
2004-04-15 09:14:14 +02:00
|
|
|
ndbcluster_inited= 0;
|
2005-11-06 00:20:37 +01:00
|
|
|
ndbcluster_util_inited= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-06-23 15:34:45 +02:00
|
|
|
/*
|
|
|
|
Static error print function called from
|
|
|
|
static handler method ndbcluster_commit
|
|
|
|
and ndbcluster_rollback
|
|
|
|
*/
|
2004-07-02 16:14:08 +02:00
|
|
|
|
|
|
|
void ndbcluster_print_error(int error, const NdbOperation *error_op)
|
2004-06-22 09:54:00 +02:00
|
|
|
{
|
2004-06-23 15:34:45 +02:00
|
|
|
DBUG_ENTER("ndbcluster_print_error");
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
TABLE_SHARE share;
|
2004-07-02 16:14:08 +02:00
|
|
|
const char *tab_name= (error_op) ? error_op->getTableName() : "";
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
share.db.str= (char*) "";
|
|
|
|
share.db.length= 0;
|
|
|
|
share.table_name.str= (char *) tab_name;
|
|
|
|
share.table_name.length= strlen(tab_name);
|
|
|
|
ha_ndbcluster error_handler(&share);
|
2004-06-23 15:34:45 +02:00
|
|
|
error_handler.print_error(error, MYF(0));
|
2004-07-06 08:43:57 +02:00
|
|
|
DBUG_VOID_RETURN;
|
2004-06-22 09:54:00 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-04-22 15:28:23 +02:00
|
|
|
/**
|
|
|
|
* Set a given location from full pathname to database name
|
|
|
|
*
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
2005-04-22 15:28:23 +02:00
|
|
|
void ha_ndbcluster::set_dbname(const char *path_name, char *dbname)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
char *end, *ptr;
|
|
|
|
|
|
|
|
/* Scan name from the end */
|
2005-04-22 15:28:23 +02:00
|
|
|
ptr= strend(path_name)-1;
|
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
|
|
|
ptr--;
|
|
|
|
end= ptr;
|
2004-04-15 09:14:14 +02:00
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
|
|
|
uint name_len= end - ptr;
|
2005-04-22 15:28:23 +02:00
|
|
|
memcpy(dbname, ptr + 1, name_len);
|
|
|
|
dbname[name_len]= '\0';
|
2004-04-15 09:14:14 +02:00
|
|
|
#ifdef __WIN__
|
|
|
|
/* Put to lower case */
|
2005-04-22 15:28:23 +02:00
|
|
|
|
|
|
|
ptr= dbname;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
while (*ptr != '\0') {
|
2004-07-23 15:46:56 +02:00
|
|
|
*ptr= tolower(*ptr);
|
2004-04-15 09:14:14 +02:00
|
|
|
ptr++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2005-04-22 15:28:23 +02:00
|
|
|
/*
|
|
|
|
Set m_dbname from full pathname to table file
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::set_dbname(const char *path_name)
|
|
|
|
{
|
|
|
|
set_dbname(path_name, m_dbname);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/**
|
|
|
|
* Set a given location from full pathname to table file
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
ha_ndbcluster::set_tabname(const char *path_name, char * tabname)
|
|
|
|
{
|
|
|
|
char *end, *ptr;
|
|
|
|
|
|
|
|
/* Scan name from the end */
|
2004-07-23 15:46:56 +02:00
|
|
|
end= strend(path_name)-1;
|
|
|
|
ptr= end;
|
2004-04-15 09:14:14 +02:00
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
2004-07-23 15:46:56 +02:00
|
|
|
uint name_len= end - ptr;
|
2004-04-15 09:14:14 +02:00
|
|
|
memcpy(tabname, ptr + 1, end - ptr);
|
2004-07-23 15:46:56 +02:00
|
|
|
tabname[name_len]= '\0';
|
2004-04-15 09:14:14 +02:00
|
|
|
#ifdef __WIN__
|
|
|
|
/* Put to lower case */
|
2004-07-23 15:46:56 +02:00
|
|
|
ptr= tabname;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
while (*ptr != '\0') {
|
|
|
|
*ptr= tolower(*ptr);
|
|
|
|
ptr++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2005-04-22 15:28:23 +02:00
|
|
|
Set m_tabname from full pathname to table file
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2005-04-22 15:28:23 +02:00
|
|
|
void ha_ndbcluster::set_tabname(const char *path_name)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-04-22 15:28:23 +02:00
|
|
|
set_tabname(path_name, m_tabname);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
ha_rows
|
2004-05-16 13:48:32 +02:00
|
|
|
ha_ndbcluster::records_in_range(uint inx, key_range *min_key,
|
|
|
|
key_range *max_key)
|
|
|
|
{
|
|
|
|
KEY *key_info= table->key_info + inx;
|
2004-04-15 09:14:14 +02:00
|
|
|
uint key_length= key_info->key_length;
|
2004-05-24 12:35:39 +02:00
|
|
|
NDB_INDEX_TYPE idx_type= get_index_type(inx);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("records_in_range");
|
2004-05-24 12:35:39 +02:00
|
|
|
// Prevent partial read of hash indexes by returning HA_POS_ERROR
|
|
|
|
if ((idx_type == UNIQUE_INDEX || idx_type == PRIMARY_KEY_INDEX) &&
|
|
|
|
((min_key && min_key->length < key_length) ||
|
|
|
|
(max_key && max_key->length < key_length)))
|
|
|
|
DBUG_RETURN(HA_POS_ERROR);
|
|
|
|
|
|
|
|
// Read from hash index with full key
|
|
|
|
// This is a "const" table which returns only one record!
|
|
|
|
if ((idx_type != ORDERED_INDEX) &&
|
|
|
|
((min_key && min_key->length == key_length) ||
|
|
|
|
(max_key && max_key->length == key_length)))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
|
2005-09-15 02:33:28 +02:00
|
|
|
if ((idx_type == PRIMARY_KEY_ORDERED_INDEX ||
|
|
|
|
idx_type == UNIQUE_ORDERED_INDEX ||
|
|
|
|
idx_type == ORDERED_INDEX) &&
|
|
|
|
m_index[inx].index_stat != NULL)
|
|
|
|
{
|
|
|
|
NDB_INDEX_DATA& d=m_index[inx];
|
|
|
|
NDBINDEX* index=(NDBINDEX*)d.index;
|
|
|
|
Ndb* ndb=get_ndb();
|
|
|
|
NdbTransaction* trans=NULL;
|
|
|
|
NdbIndexScanOperation* op=NULL;
|
|
|
|
int res=0;
|
|
|
|
Uint64 rows;
|
|
|
|
|
|
|
|
do
|
|
|
|
{
|
|
|
|
// We must provide approx table rows
|
|
|
|
Uint64 table_rows=0;
|
|
|
|
Ndb_local_table_statistics *info=
|
|
|
|
(Ndb_local_table_statistics *)m_table_info;
|
|
|
|
if (info->records != ~(ha_rows)0 && info->records != 0)
|
|
|
|
{
|
|
|
|
table_rows = info->records;
|
|
|
|
DBUG_PRINT("info", ("use info->records: %llu", table_rows));
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
Ndb_statistics stat;
|
|
|
|
if ((res=ndb_get_table_statistics(ndb, m_tabname, &stat)) != 0)
|
|
|
|
break;
|
|
|
|
table_rows=stat.row_count;
|
|
|
|
DBUG_PRINT("info", ("use db row_count: %llu", table_rows));
|
|
|
|
if (table_rows == 0) {
|
|
|
|
// Problem if autocommit=0
|
|
|
|
#ifdef ndb_get_table_statistics_uses_active_trans
|
|
|
|
rows=0;
|
|
|
|
break;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Define scan op for the range
|
|
|
|
if ((trans=m_active_trans) == NULL)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("no active trans"));
|
|
|
|
if (! (trans=ndb->startTransaction()))
|
|
|
|
ERR_BREAK(ndb->getNdbError(), res);
|
|
|
|
}
|
|
|
|
if (! (op=trans->getNdbIndexScanOperation(index, (NDBTAB*)m_table)))
|
|
|
|
ERR_BREAK(trans->getNdbError(), res);
|
|
|
|
if ((op->readTuples(NdbOperation::LM_CommittedRead)) == -1)
|
|
|
|
ERR_BREAK(op->getNdbError(), res);
|
|
|
|
const key_range *keys[2]={ min_key, max_key };
|
|
|
|
if ((res=set_bounds(op, inx, true, keys)) != 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
// Decide if db should be contacted
|
|
|
|
int flags=0;
|
|
|
|
if (d.index_stat_query_count < d.index_stat_cache_entries ||
|
|
|
|
(d.index_stat_update_freq != 0 &&
|
|
|
|
d.index_stat_query_count % d.index_stat_update_freq == 0))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("force stat from db"));
|
|
|
|
flags|=NdbIndexStat::RR_UseDb;
|
|
|
|
}
|
|
|
|
if (d.index_stat->records_in_range(index, op, table_rows, &rows, flags) == -1)
|
|
|
|
ERR_BREAK(d.index_stat->getNdbError(), res);
|
|
|
|
d.index_stat_query_count++;
|
|
|
|
} while (0);
|
|
|
|
|
|
|
|
if (trans != m_active_trans && rows == 0)
|
|
|
|
rows = 1;
|
|
|
|
if (trans != m_active_trans && trans != NULL)
|
|
|
|
ndb->closeTransaction(trans);
|
|
|
|
if (res != 0)
|
|
|
|
DBUG_RETURN(HA_POS_ERROR);
|
|
|
|
DBUG_RETURN(rows);
|
|
|
|
}
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_RETURN(10); /* Good guess when you don't know anything */
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
ulong ha_ndbcluster::table_flags(void) const
|
|
|
|
{
|
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return m_table_flags | HA_NOT_EXACT_COUNT;
|
|
|
|
else
|
|
|
|
return m_table_flags;
|
|
|
|
}
|
|
|
|
const char * ha_ndbcluster::table_type() const
|
|
|
|
{
|
2005-09-21 04:16:24 +02:00
|
|
|
return("NDBCLUSTER");
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_record_length() const
|
|
|
|
{
|
|
|
|
return NDB_MAX_TUPLE_SIZE;
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_keys() const
|
|
|
|
{
|
|
|
|
return MAX_KEY;
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_key_parts() const
|
|
|
|
{
|
|
|
|
return NDB_MAX_NO_OF_ATTRIBUTES_IN_KEY;
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_key_length() const
|
|
|
|
{
|
|
|
|
return NDB_MAX_KEY_SIZE;
|
|
|
|
}
|
|
|
|
bool ha_ndbcluster::low_byte_first() const
|
|
|
|
{
|
|
|
|
#ifdef WORDS_BIGENDIAN
|
|
|
|
return FALSE;
|
|
|
|
#else
|
|
|
|
return TRUE;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
bool ha_ndbcluster::has_transactions()
|
|
|
|
{
|
2004-11-22 14:41:46 +01:00
|
|
|
return m_transaction_on;
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
|
|
|
const char* ha_ndbcluster::index_type(uint key_number)
|
|
|
|
{
|
|
|
|
switch (get_index_type(key_number)) {
|
|
|
|
case ORDERED_INDEX:
|
|
|
|
case UNIQUE_ORDERED_INDEX:
|
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
|
|
|
return "BTREE";
|
|
|
|
case UNIQUE_INDEX:
|
|
|
|
case PRIMARY_KEY_INDEX:
|
|
|
|
default:
|
|
|
|
return "HASH";
|
|
|
|
}
|
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
uint8 ha_ndbcluster::table_cache_type()
|
|
|
|
{
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::table_cache_type=HA_CACHE_TBL_ASKTRANSACT");
|
|
|
|
DBUG_RETURN(HA_CACHE_TBL_ASKTRANSACT);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
uint ndb_get_commitcount(THD *thd, char *dbname, char *tabname,
|
2005-02-16 14:18:32 +01:00
|
|
|
Uint64 *commit_count)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
2005-03-15 15:03:25 +01:00
|
|
|
char name[FN_REFLEN];
|
|
|
|
NDB_SHARE *share;
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
DBUG_ENTER("ndb_get_commitcount");
|
|
|
|
|
|
|
|
(void)strxnmov(name, FN_REFLEN-1, share_prefix, dbname, "/", tabname, NullS);
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("enter", ("name: %s", name));
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
|
|
|
if (!(share=(NDB_SHARE*) hash_search(&ndbcluster_open_tables,
|
|
|
|
(byte*) name,
|
|
|
|
strlen(name))))
|
|
|
|
{
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("Table %s not found in ndbcluster_open_tables", name));
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
share->use_count++;
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
|
|
|
|
pthread_mutex_lock(&share->mutex);
|
2005-02-11 22:33:52 +01:00
|
|
|
if (ndb_cache_check_time > 0)
|
|
|
|
{
|
2005-03-15 15:03:25 +01:00
|
|
|
if (share->commit_count != 0)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
2005-03-15 15:03:25 +01:00
|
|
|
*commit_count= share->commit_count;
|
|
|
|
DBUG_PRINT("info", ("Getting commit_count: %llu from share",
|
|
|
|
share->commit_count));
|
|
|
|
pthread_mutex_unlock(&share->mutex);
|
2005-11-06 00:20:37 +01:00
|
|
|
free_share(&share);
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_RETURN(0);
|
2005-02-11 22:33:52 +01:00
|
|
|
}
|
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("info", ("Get commit_count from NDB"));
|
2005-02-11 22:33:52 +01:00
|
|
|
Ndb *ndb;
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
ndb->setDatabaseName(dbname);
|
2005-03-15 15:03:25 +01:00
|
|
|
uint lock= share->commit_count_lock;
|
|
|
|
pthread_mutex_unlock(&share->mutex);
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
struct Ndb_statistics stat;
|
|
|
|
if (ndb_get_table_statistics(ndb, tabname, &stat))
|
2005-03-15 15:03:25 +01:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
free_share(&share);
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(1);
|
2005-03-15 15:03:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
pthread_mutex_lock(&share->mutex);
|
2005-07-04 02:42:33 +02:00
|
|
|
if (share->commit_count_lock == lock)
|
2005-03-15 15:03:25 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Setting commit_count to %llu", stat.commit_count));
|
|
|
|
share->commit_count= stat.commit_count;
|
|
|
|
*commit_count= stat.commit_count;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Discarding commit_count, comit_count_lock changed"));
|
|
|
|
*commit_count= 0;
|
|
|
|
}
|
|
|
|
pthread_mutex_unlock(&share->mutex);
|
2005-11-06 00:20:37 +01:00
|
|
|
free_share(&share);
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Check if a cached query can be used.
|
|
|
|
This is done by comparing the supplied engine_data to commit_count of
|
|
|
|
the table.
|
|
|
|
The commit_count is either retrieved from the share for the table, where
|
|
|
|
it has been cached by the util thread. If the util thread is not started,
|
|
|
|
NDB has to be contacetd to retrieve the commit_count, this will introduce
|
|
|
|
a small delay while waiting for NDB to answer.
|
|
|
|
|
|
|
|
|
|
|
|
SYNOPSIS
|
|
|
|
ndbcluster_cache_retrieval_allowed
|
|
|
|
thd thread handle
|
|
|
|
full_name concatenation of database name,
|
|
|
|
the null character '\0', and the table
|
|
|
|
name
|
|
|
|
full_name_len length of the full name,
|
|
|
|
i.e. len(dbname) + len(tablename) + 1
|
|
|
|
|
|
|
|
engine_data parameter retrieved when query was first inserted into
|
|
|
|
the cache. If the value of engine_data is changed,
|
|
|
|
all queries for this table should be invalidated.
|
|
|
|
|
|
|
|
RETURN VALUE
|
|
|
|
TRUE Yes, use the query from cache
|
|
|
|
FALSE No, don't use the cached query, and if engine_data
|
|
|
|
has changed, all queries for this table should be invalidated
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
static my_bool
|
|
|
|
ndbcluster_cache_retrieval_allowed(THD *thd,
|
2005-02-16 14:18:32 +01:00
|
|
|
char *full_name, uint full_name_len,
|
|
|
|
ulonglong *engine_data)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_cache_retrieval_allowed");
|
|
|
|
|
|
|
|
Uint64 commit_count;
|
|
|
|
bool is_autocommit= !(thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN));
|
|
|
|
char *dbname= full_name;
|
|
|
|
char *tabname= dbname+strlen(dbname)+1;
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("enter", ("dbname: %s, tabname: %s, is_autocommit: %d",
|
|
|
|
dbname, tabname, is_autocommit));
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
if (!is_autocommit)
|
2005-03-15 15:03:25 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("exit", ("No, don't use cache in transaction"));
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(FALSE);
|
2005-03-15 15:03:25 +01:00
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
if (ndb_get_commitcount(thd, dbname, tabname, &commit_count))
|
|
|
|
{
|
2005-03-15 15:03:25 +01:00
|
|
|
*engine_data= 0; /* invalidate */
|
|
|
|
DBUG_PRINT("exit", ("No, could not retrieve commit_count"));
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("info", ("*engine_data: %llu, commit_count: %llu",
|
2005-02-16 14:18:32 +01:00
|
|
|
*engine_data, commit_count));
|
2005-03-15 15:03:25 +01:00
|
|
|
if (commit_count == 0)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
2005-03-15 15:03:25 +01:00
|
|
|
*engine_data= 0; /* invalidate */
|
|
|
|
DBUG_PRINT("exit", ("No, local commit has been performed"));
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
else if (*engine_data != commit_count)
|
|
|
|
{
|
|
|
|
*engine_data= commit_count; /* invalidate */
|
|
|
|
DBUG_PRINT("exit", ("No, commit_count has changed"));
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("exit", ("OK to use cache, engine_data: %llu", *engine_data));
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(TRUE);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
Register a table for use in the query cache. Fetch the commit_count
|
|
|
|
for the table and return it in engine_data, this will later be used
|
|
|
|
to check if the table has changed, before the cached query is reused.
|
|
|
|
|
|
|
|
SYNOPSIS
|
|
|
|
ha_ndbcluster::can_query_cache_table
|
|
|
|
thd thread handle
|
|
|
|
full_name concatenation of database name,
|
|
|
|
the null character '\0', and the table
|
|
|
|
name
|
|
|
|
full_name_len length of the full name,
|
|
|
|
i.e. len(dbname) + len(tablename) + 1
|
|
|
|
qc_engine_callback function to be called before using cache on this table
|
|
|
|
engine_data out, commit_count for this table
|
|
|
|
|
|
|
|
RETURN VALUE
|
|
|
|
TRUE Yes, it's ok to cahce this query
|
|
|
|
FALSE No, don't cach the query
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
my_bool
|
|
|
|
ha_ndbcluster::register_query_cache_table(THD *thd,
|
2005-02-16 14:18:32 +01:00
|
|
|
char *full_name, uint full_name_len,
|
|
|
|
qc_engine_callback *engine_callback,
|
|
|
|
ulonglong *engine_data)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::register_query_cache_table");
|
|
|
|
|
|
|
|
bool is_autocommit= !(thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN));
|
2005-03-15 15:03:25 +01:00
|
|
|
|
|
|
|
DBUG_PRINT("enter",("dbname: %s, tabname: %s, is_autocommit: %d",
|
|
|
|
m_dbname, m_tabname, is_autocommit));
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
if (!is_autocommit)
|
2005-03-15 15:03:25 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("exit", ("Can't register table during transaction"))
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(FALSE);
|
2005-03-15 15:03:25 +01:00
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
Uint64 commit_count;
|
|
|
|
if (ndb_get_commitcount(thd, m_dbname, m_tabname, &commit_count))
|
|
|
|
{
|
|
|
|
*engine_data= 0;
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("exit", ("Error, could not get commitcount"))
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
*engine_data= commit_count;
|
|
|
|
*engine_callback= ndbcluster_cache_retrieval_allowed;
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("exit", ("commit_count: %llu", commit_count));
|
|
|
|
DBUG_RETURN(commit_count > 0);
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
#ifndef DBUG_OFF
|
|
|
|
static void dbug_print_table(const char *info, TABLE *table)
|
|
|
|
{
|
|
|
|
if (table == 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info",("%s: (null)", info));
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info",
|
|
|
|
("%s: %s.%s s->fields: %d "
|
|
|
|
"reclength: %d rec_buff_length: %d record[0]: %lx "
|
|
|
|
"record[1]: %lx",
|
|
|
|
info,
|
|
|
|
table->s->db,
|
|
|
|
table->s->table_name,
|
|
|
|
table->s->fields,
|
|
|
|
table->s->reclength,
|
|
|
|
table->s->rec_buff_length,
|
|
|
|
table->record[0],
|
|
|
|
table->record[1]));
|
|
|
|
|
|
|
|
for (unsigned int i= 0; i < table->s->fields; i++)
|
|
|
|
{
|
|
|
|
Field *f= table->field[i];
|
|
|
|
DBUG_PRINT("info",
|
|
|
|
("[%d] \"%s\"(0x%lx:%s%s%s%s%s%s) type: %d pack_length: %d "
|
|
|
|
"ptr: 0x%lx[+%d] null_bit: %u null_ptr: 0x%lx[+%d]",
|
|
|
|
i,
|
|
|
|
f->field_name,
|
|
|
|
f->flags,
|
|
|
|
(f->flags & PRI_KEY_FLAG) ? "pri" : "attr",
|
|
|
|
(f->flags & NOT_NULL_FLAG) ? "" : ",nullable",
|
|
|
|
(f->flags & UNSIGNED_FLAG) ? ",unsigned" : ",signed",
|
|
|
|
(f->flags & ZEROFILL_FLAG) ? ",zerofill" : "",
|
|
|
|
(f->flags & BLOB_FLAG) ? ",blob" : "",
|
|
|
|
(f->flags & BINARY_FLAG) ? ",binary" : "",
|
|
|
|
f->real_type(),
|
|
|
|
f->pack_length(),
|
|
|
|
f->ptr, f->ptr - table->record[0],
|
|
|
|
f->null_bit,
|
|
|
|
f->null_ptr, (byte*) f->null_ptr - table->record[0]));
|
|
|
|
if (f->type() == MYSQL_TYPE_BIT)
|
|
|
|
{
|
|
|
|
Field_bit *g= (Field_bit*) f;
|
|
|
|
DBUG_PRINT("MYSQL_TYPE_BIT",("field_length: %d bit_ptr: 0x%lx[+%d] "
|
|
|
|
"bit_ofs: %u bit_len: %u",
|
|
|
|
g->field_length, g->bit_ptr,
|
|
|
|
(byte*) g->bit_ptr-table->record[0],
|
|
|
|
g->bit_ofs, g->bit_len));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
2005-02-11 22:33:52 +01:00
|
|
|
Handling the shared NDB_SHARE structure that is needed to
|
2004-04-15 09:14:14 +02:00
|
|
|
provide table locking.
|
|
|
|
It's also used for sharing data with other NDB handlers
|
|
|
|
in the same MySQL Server. There is currently not much
|
|
|
|
data we want to or can share.
|
|
|
|
*/
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
static byte *ndbcluster_get_key(NDB_SHARE *share,uint *length,
|
2005-02-16 14:18:32 +01:00
|
|
|
my_bool not_used __attribute__((unused)))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
*length= share->key_length;
|
|
|
|
return (byte*) share->key;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
static void dbug_print_open_tables()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("dbug_print_open_tables");
|
|
|
|
for (uint i= 0; i < ndbcluster_open_tables.records; i++)
|
|
|
|
{
|
|
|
|
NDB_SHARE *share= (NDB_SHARE*) hash_element(&ndbcluster_open_tables, i);
|
|
|
|
DBUG_PRINT("share",
|
|
|
|
("[%d] 0x%lx key: %s key_length: %d",
|
|
|
|
i, share, share->key, share->key_length));
|
|
|
|
DBUG_PRINT("share",
|
|
|
|
("db.tablename: %s.%s use_count: %d commit_count: %d",
|
|
|
|
share->db, share->table_name,
|
|
|
|
share->use_count, share->commit_count));
|
|
|
|
}
|
|
|
|
DBUG_VOID_RETURN;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
#else
|
|
|
|
#define dbug_print_open_tables()
|
|
|
|
#endif
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/*
|
|
|
|
Increase refcount on existing share.
|
|
|
|
Always returns share and cannot fail.
|
|
|
|
*/
|
|
|
|
static NDB_SHARE *get_share(NDB_SHARE *share)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
2005-11-06 00:20:37 +01:00
|
|
|
share->use_count++;
|
|
|
|
|
|
|
|
dbug_print_open_tables();
|
|
|
|
|
|
|
|
DBUG_PRINT("get_share",
|
|
|
|
("0x%lx key: %s key_length: %d",
|
|
|
|
share, share->key, share->key_length));
|
|
|
|
DBUG_PRINT("get_share",
|
|
|
|
("db.tablename: %s.%s use_count: %d commit_count: %d",
|
|
|
|
share->db, share->table_name,
|
|
|
|
share->use_count, share->commit_count));
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
return share;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get a share object for key
|
|
|
|
|
|
|
|
Returns share for key, and increases the refcount on the share.
|
|
|
|
|
|
|
|
create_if_not_exists == TRUE:
|
|
|
|
creates share if it does not alreade exist
|
|
|
|
returns 0 only due to out of memory, and then sets my_error
|
|
|
|
|
|
|
|
create_if_not_exists == FALSE:
|
|
|
|
returns 0 if share does not exist
|
|
|
|
|
|
|
|
have_lock == TRUE, pthread_mutex_lock(&ndbcluster_mutex) already taken
|
|
|
|
*/
|
|
|
|
static NDB_SHARE *get_share(const char *key, bool create_if_not_exists,
|
|
|
|
bool have_lock)
|
|
|
|
{
|
|
|
|
NDB_SHARE *share;
|
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
|
|
|
uint length= (uint) strlen(key);
|
|
|
|
if (!(share= (NDB_SHARE*) hash_search(&ndbcluster_open_tables,
|
|
|
|
(byte*) key,
|
|
|
|
length)))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (!create_if_not_exists)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("get_share: %s does not exist", key));
|
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if ((share= (NDB_SHARE*) my_malloc(sizeof(*share),
|
2004-04-15 09:14:14 +02:00
|
|
|
MYF(MY_WME | MY_ZEROFILL))))
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
MEM_ROOT **root_ptr=
|
|
|
|
my_pthread_getspecific_ptr(MEM_ROOT**, THR_MALLOC);
|
|
|
|
MEM_ROOT *old_root= *root_ptr;
|
|
|
|
init_sql_alloc(&share->mem_root, 1024, 0);
|
|
|
|
*root_ptr= &share->mem_root; // remember to reset before return
|
|
|
|
|
|
|
|
/* enough space for key, db, and table_name */
|
|
|
|
share->key= alloc_root(*root_ptr, 2 * (length + 1));
|
|
|
|
share->key_length= length;
|
|
|
|
strmov(share->key, key);
|
2004-04-15 09:14:14 +02:00
|
|
|
if (my_hash_insert(&ndbcluster_open_tables, (byte*) share))
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
free_root(&share->mem_root, MYF(0));
|
|
|
|
my_free((gptr) share, 0);
|
|
|
|
*root_ptr= old_root;
|
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
2004-04-15 09:14:14 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
thr_lock_init(&share->lock);
|
2005-11-06 00:20:37 +01:00
|
|
|
pthread_mutex_init(&share->mutex, MY_MUTEX_INIT_FAST);
|
2005-02-11 22:33:52 +01:00
|
|
|
share->commit_count= 0;
|
2005-03-15 15:03:25 +01:00
|
|
|
share->commit_count_lock= 0;
|
2005-11-06 00:20:37 +01:00
|
|
|
share->db= share->key + length + 1;
|
|
|
|
ha_ndbcluster::set_dbname(key, share->db);
|
|
|
|
share->table_name= share->db + strlen(share->db) + 1;
|
|
|
|
ha_ndbcluster::set_tabname(key, share->table_name);
|
|
|
|
*root_ptr= old_root;
|
2005-03-15 15:03:25 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("error", ("get_share: failed to alloc share"));
|
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
my_error(ER_OUTOFMEMORY, MYF(0), sizeof(*share));
|
2005-03-15 15:03:25 +01:00
|
|
|
return 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
share->use_count++;
|
2005-03-15 15:03:25 +01:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
dbug_print_open_tables();
|
|
|
|
|
|
|
|
DBUG_PRINT("get_share",
|
|
|
|
("0x%lx key: %s key_length: %d key: %s",
|
|
|
|
share, share->key, share->key_length, key));
|
|
|
|
DBUG_PRINT("get_share",
|
|
|
|
("db.tablename: %s.%s use_count: %d commit_count: %d",
|
|
|
|
share->db, share->table_name,
|
|
|
|
share->use_count, share->commit_count));
|
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
2004-04-15 09:14:14 +02:00
|
|
|
return share;
|
|
|
|
}
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
static void real_free_share(NDB_SHARE **share)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("real_free_share",
|
|
|
|
("0x%lx key: %s key_length: %d",
|
|
|
|
(*share), (*share)->key, (*share)->key_length));
|
|
|
|
DBUG_PRINT("real_free_share",
|
|
|
|
("db.tablename: %s.%s use_count: %d commit_count: %d",
|
|
|
|
(*share)->db, (*share)->table_name,
|
|
|
|
(*share)->use_count, (*share)->commit_count));
|
|
|
|
|
|
|
|
hash_delete(&ndbcluster_open_tables, (byte*) *share);
|
|
|
|
thr_lock_delete(&(*share)->lock);
|
|
|
|
pthread_mutex_destroy(&(*share)->mutex);
|
|
|
|
free_root(&(*share)->mem_root, MYF(0));
|
|
|
|
|
|
|
|
my_free((gptr) *share, MYF(0));
|
|
|
|
*share= 0;
|
|
|
|
|
|
|
|
dbug_print_open_tables();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
decrease refcount of share
|
|
|
|
calls real_free_share when refcount reaches 0
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
have_lock == TRUE, pthread_mutex_lock(&ndbcluster_mutex) already taken
|
|
|
|
*/
|
|
|
|
static void free_share(NDB_SHARE **share, bool have_lock)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
|
|
|
if ((*share)->util_lock == current_thd)
|
|
|
|
(*share)->util_lock= 0;
|
|
|
|
if (!--(*share)->use_count)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2005-11-06 00:20:37 +01:00
|
|
|
real_free_share(share);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
else
|
|
|
|
{
|
|
|
|
dbug_print_open_tables();
|
|
|
|
DBUG_PRINT("free_share",
|
|
|
|
("0x%lx key: %s key_length: %d",
|
|
|
|
*share, (*share)->key, (*share)->key_length));
|
|
|
|
DBUG_PRINT("free_share",
|
|
|
|
("db.tablename: %s.%s use_count: %d commit_count: %d",
|
|
|
|
(*share)->db, (*share)->table_name,
|
|
|
|
(*share)->use_count, (*share)->commit_count));
|
|
|
|
}
|
|
|
|
if (!have_lock)
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Internal representation of the frm blob
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct frm_blob_struct
|
|
|
|
{
|
|
|
|
struct frm_blob_header
|
|
|
|
{
|
|
|
|
uint ver; // Version of header
|
|
|
|
uint orglen; // Original length of compressed data
|
|
|
|
uint complen; // Compressed length of data, 0=uncompressed
|
|
|
|
} head;
|
|
|
|
char data[1];
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int packfrm(const void *data, uint len,
|
2005-02-16 14:18:32 +01:00
|
|
|
const void **pack_data, uint *pack_len)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
int error;
|
|
|
|
ulong org_len, comp_len;
|
|
|
|
uint blob_len;
|
|
|
|
frm_blob_struct* blob;
|
|
|
|
DBUG_ENTER("packfrm");
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("enter", ("data: 0x%lx len: %d", data, len));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
error= 1;
|
2004-07-23 15:46:56 +02:00
|
|
|
org_len= len;
|
2004-04-15 09:14:14 +02:00
|
|
|
if (my_compress((byte*)data, &org_len, &comp_len))
|
|
|
|
goto err;
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("info", ("org_len: %d comp_len: %d", org_len, comp_len));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_DUMP("compressed", (char*)data, org_len);
|
|
|
|
|
|
|
|
error= 2;
|
|
|
|
blob_len= sizeof(frm_blob_struct::frm_blob_header)+org_len;
|
|
|
|
if (!(blob= (frm_blob_struct*) my_malloc(blob_len,MYF(MY_WME))))
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
// Store compressed blob in machine independent format
|
|
|
|
int4store((char*)(&blob->head.ver), 1);
|
|
|
|
int4store((char*)(&blob->head.orglen), comp_len);
|
|
|
|
int4store((char*)(&blob->head.complen), org_len);
|
|
|
|
|
|
|
|
// Copy frm data into blob, already in machine independent format
|
|
|
|
memcpy(blob->data, data, org_len);
|
|
|
|
|
2004-07-23 15:46:56 +02:00
|
|
|
*pack_data= blob;
|
|
|
|
*pack_len= blob_len;
|
|
|
|
error= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("exit",
|
|
|
|
("pack_data: 0x%lx pack_len: %d", *pack_data, *pack_len));
|
2004-04-15 09:14:14 +02:00
|
|
|
err:
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int unpackfrm(const void **unpack_data, uint *unpack_len,
|
2005-02-16 14:18:32 +01:00
|
|
|
const void *pack_data)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-07-23 15:46:56 +02:00
|
|
|
const frm_blob_struct *blob= (frm_blob_struct*)pack_data;
|
2004-04-15 09:14:14 +02:00
|
|
|
byte *data;
|
|
|
|
ulong complen, orglen, ver;
|
|
|
|
DBUG_ENTER("unpackfrm");
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("enter", ("pack_data: 0x%lx", pack_data));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
complen= uint4korr((char*)&blob->head.complen);
|
|
|
|
orglen= uint4korr((char*)&blob->head.orglen);
|
|
|
|
ver= uint4korr((char*)&blob->head.ver);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("blob",("ver: %d complen: %d orglen: %d",
|
2005-02-16 14:18:32 +01:00
|
|
|
ver,complen,orglen));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_DUMP("blob->data", (char*) blob->data, complen);
|
|
|
|
|
|
|
|
if (ver != 1)
|
|
|
|
DBUG_RETURN(1);
|
2004-07-23 15:46:56 +02:00
|
|
|
if (!(data= my_malloc(max(orglen, complen), MYF(MY_WME))))
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(2);
|
|
|
|
memcpy(data, blob->data, complen);
|
|
|
|
|
|
|
|
if (my_uncompress(data, &complen, &orglen))
|
|
|
|
{
|
|
|
|
my_free((char*)data, MYF(0));
|
|
|
|
DBUG_RETURN(3);
|
|
|
|
}
|
|
|
|
|
2004-07-23 15:46:56 +02:00
|
|
|
*unpack_data= data;
|
|
|
|
*unpack_len= complen;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
DBUG_PRINT("exit", ("frmdata: 0x%lx, len: %d", *unpack_data, *unpack_len));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-09-03 13:55:40 +02:00
|
|
|
|
|
|
|
static
|
|
|
|
int
|
2005-01-13 18:24:19 +01:00
|
|
|
ndb_get_table_statistics(Ndb* ndb, const char * table,
|
2005-02-16 14:18:32 +01:00
|
|
|
struct Ndb_statistics * ndbstat)
|
2004-09-03 13:55:40 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ndb_get_table_statistics");
|
|
|
|
DBUG_PRINT("enter", ("table: %s", table));
|
2005-01-04 13:47:16 +01:00
|
|
|
NdbTransaction* pTrans= ndb->startTransaction();
|
2005-11-06 00:20:37 +01:00
|
|
|
if (pTrans == NULL)
|
|
|
|
ERR_RETURN(ndb->getNdbError());
|
2004-09-03 13:55:40 +02:00
|
|
|
do
|
|
|
|
{
|
|
|
|
NdbScanOperation* pOp= pTrans->getNdbScanOperation(table);
|
|
|
|
if (pOp == NULL)
|
|
|
|
break;
|
|
|
|
|
2004-12-07 15:15:49 +01:00
|
|
|
if (pOp->readTuples(NdbOperation::LM_CommittedRead))
|
2004-09-03 13:55:40 +02:00
|
|
|
break;
|
|
|
|
|
|
|
|
int check= pOp->interpret_exit_last_row();
|
|
|
|
if (check == -1)
|
|
|
|
break;
|
|
|
|
|
2005-01-20 22:28:08 +01:00
|
|
|
Uint64 rows, commits, mem;
|
|
|
|
Uint32 size;
|
2004-09-03 13:55:40 +02:00
|
|
|
pOp->getValue(NdbDictionary::Column::ROW_COUNT, (char*)&rows);
|
|
|
|
pOp->getValue(NdbDictionary::Column::COMMIT_COUNT, (char*)&commits);
|
2005-01-13 18:24:19 +01:00
|
|
|
pOp->getValue(NdbDictionary::Column::ROW_SIZE, (char*)&size);
|
|
|
|
pOp->getValue(NdbDictionary::Column::FRAGMENT_MEMORY, (char*)&mem);
|
2004-09-03 13:55:40 +02:00
|
|
|
|
2005-01-05 17:59:24 +01:00
|
|
|
check= pTrans->execute(NdbTransaction::NoCommit,
|
2005-02-16 14:18:32 +01:00
|
|
|
NdbTransaction::AbortOnError,
|
|
|
|
TRUE);
|
2004-09-03 13:55:40 +02:00
|
|
|
if (check == -1)
|
|
|
|
break;
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
Uint32 count= 0;
|
2004-09-03 13:55:40 +02:00
|
|
|
Uint64 sum_rows= 0;
|
|
|
|
Uint64 sum_commits= 0;
|
2005-01-13 18:24:19 +01:00
|
|
|
Uint64 sum_row_size= 0;
|
|
|
|
Uint64 sum_mem= 0;
|
2005-07-18 14:33:18 +02:00
|
|
|
while ((check= pOp->nextResult(TRUE, TRUE)) == 0)
|
2004-09-03 13:55:40 +02:00
|
|
|
{
|
|
|
|
sum_rows+= rows;
|
|
|
|
sum_commits+= commits;
|
2005-01-13 18:24:19 +01:00
|
|
|
if (sum_row_size < size)
|
2005-02-16 14:18:32 +01:00
|
|
|
sum_row_size= size;
|
2005-01-13 18:24:19 +01:00
|
|
|
sum_mem+= mem;
|
2005-03-15 15:03:25 +01:00
|
|
|
count++;
|
2004-09-03 13:55:40 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
if (check == -1)
|
|
|
|
break;
|
|
|
|
|
2004-12-07 15:15:49 +01:00
|
|
|
pOp->close(TRUE);
|
2004-11-22 14:41:46 +01:00
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
ndb->closeTransaction(pTrans);
|
2005-01-13 18:24:19 +01:00
|
|
|
|
|
|
|
ndbstat->row_count= sum_rows;
|
|
|
|
ndbstat->commit_count= sum_commits;
|
|
|
|
ndbstat->row_size= sum_row_size;
|
|
|
|
ndbstat->fragment_memory= sum_mem;
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
DBUG_PRINT("exit", ("records: %llu commits: %llu "
|
|
|
|
"row_size: %llu mem: %llu count: %u",
|
|
|
|
sum_rows, sum_commits, sum_row_size,
|
|
|
|
sum_mem, count));
|
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
DBUG_RETURN(0);
|
2005-07-18 14:33:18 +02:00
|
|
|
} while (0);
|
2004-09-03 13:55:40 +02:00
|
|
|
|
BUG#9626
- Fix valgrind warnings
- Remove static db, schema and table name buffers from Ndb.cpp
mysys/thr_alarm.c:
Initialise sact to zero
ndb/include/kernel/ndb_limits.h:
Set MAX_ATTR_NAME_SIZE to NAME_LEN which is tha maximum in MySQL
ndb/include/kernel/signaldata/GetTabInfo.hpp:
Clean up
ndb/include/ndbapi/Ndb.hpp:
Remove NDB_MAX_INTERNAL_NAME_LENGTH and all char buffers for schema, db and tablename.
Made them dynamic and moved to NdbImpl.hpp
ndb/include/ndbapi/ndbapi_limits.h:
Remove the static length's of attr, db, schema and table name.
ndb/src/common/transporter/Packer.cpp:
Set theSignalId to ~0 when unpacking signal
ndb/src/ndbapi/Ndb.cpp:
Moved schema, database and tablename to NdbImpl.hpp
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Add NdbIndexImpl::init and NdbEventImpl::init
Init all vars in NdbIndexImpl, NdbEventImpl, NdbTableImpl and NdbIndexImpl
Delete the pseudo column NDB$RANGE_NO
Copy tablename to internal buff in NdbDictInterface to get proper alignment.
Convert length of table name from bytes words, when setting sz of LinearSectionPtr
Set LinearSectionPtr array size to number of LinearSections used - save some stack.
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
Add NdbEventImpl::init and NdbIndexImpl::init
Remove clearNewProperties and copyNewProperties, it's easier to check if all vars are initied if it's done in the same func.
Add buffer for tabname ti NdbDictInterface, memset it to 0 in initializer.
ndb/src/ndbapi/NdbImpl.hpp:
Use BaseString for table, schema and db names.
ndb/src/ndbapi/Ndbinit.cpp:
Move schema and db name to NdbImpl and use BaseString
ndb/src/ndbapi/ndb_cluster_connection.cpp:
Destroy ndb_global_event_buffer_mutex and ndb_print_state_mutex
sql/ha_ndbcluster.cc:
Check if pTrans is not null before calling closeTransaction
Remove NDB_MAX_ATTR_NAME_SIZE
Remove truncation of attr names. When attr name length is same in NDB as in MySQL this will be checked in functin check_column_name
2005-05-18 20:50:29 +02:00
|
|
|
if (pTrans)
|
|
|
|
ndb->closeTransaction(pTrans);
|
2004-09-03 13:55:40 +02:00
|
|
|
DBUG_PRINT("exit", ("failed"));
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
/*
|
|
|
|
Create a .ndb file to serve as a placeholder indicating
|
|
|
|
that the table with this name is a ndb table
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::write_ndb_file()
|
|
|
|
{
|
|
|
|
File file;
|
|
|
|
bool error=1;
|
|
|
|
char path[FN_REFLEN];
|
|
|
|
|
|
|
|
DBUG_ENTER("write_ndb_file");
|
|
|
|
DBUG_PRINT("enter", ("db: %s, name: %s", m_dbname, m_tabname));
|
|
|
|
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
(void)strxnmov(path, FN_REFLEN-1,
|
2005-02-16 14:18:32 +01:00
|
|
|
mysql_data_home,"/",m_dbname,"/",m_tabname,ha_ndb_ext,NullS);
|
2004-09-13 14:46:38 +02:00
|
|
|
|
|
|
|
if ((file=my_create(path, CREATE_MODE,O_RDWR | O_TRUNC,MYF(MY_WME))) >= 0)
|
|
|
|
{
|
|
|
|
// It's an empty file
|
|
|
|
error=0;
|
|
|
|
my_close(file,MYF(0));
|
|
|
|
}
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
2004-11-17 10:07:52 +01:00
|
|
|
int
|
2004-12-28 17:01:07 +01:00
|
|
|
ha_ndbcluster::read_multi_range_first(KEY_MULTI_RANGE **found_range_p,
|
2005-02-16 14:18:32 +01:00
|
|
|
KEY_MULTI_RANGE *ranges,
|
|
|
|
uint range_count,
|
|
|
|
bool sorted,
|
|
|
|
HANDLER_BUFFER *buffer)
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_multi_range_first");
|
2005-04-28 14:45:27 +02:00
|
|
|
m_write_op= FALSE;
|
2004-12-08 00:36:40 +01:00
|
|
|
|
2004-11-17 10:07:52 +01:00
|
|
|
int res;
|
|
|
|
KEY* key_info= table->key_info + active_index;
|
2004-12-07 15:15:49 +01:00
|
|
|
NDB_INDEX_TYPE index_type= get_index_type(active_index);
|
2005-01-07 07:15:08 +01:00
|
|
|
ulong reclength= table->s->reclength;
|
2004-11-17 10:07:52 +01:00
|
|
|
NdbOperation* op;
|
|
|
|
|
2005-04-28 14:45:27 +02:00
|
|
|
if (uses_blob_value())
|
2004-11-17 14:21:56 +01:00
|
|
|
{
|
|
|
|
/**
|
|
|
|
* blobs can't be batched currently
|
|
|
|
*/
|
2005-05-06 10:39:30 +02:00
|
|
|
m_disable_multi_read= TRUE;
|
2004-11-25 19:12:05 +01:00
|
|
|
DBUG_RETURN(handler::read_multi_range_first(found_range_p,
|
2005-02-16 14:18:32 +01:00
|
|
|
ranges,
|
|
|
|
range_count,
|
|
|
|
sorted,
|
|
|
|
buffer));
|
2004-11-17 14:21:56 +01:00
|
|
|
}
|
|
|
|
|
2005-05-06 10:39:30 +02:00
|
|
|
m_disable_multi_read= FALSE;
|
2004-11-30 07:58:55 +01:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Copy arguments into member variables
|
|
|
|
*/
|
2004-12-28 17:01:07 +01:00
|
|
|
m_multi_ranges= ranges;
|
|
|
|
multi_range_curr= ranges;
|
|
|
|
multi_range_end= ranges+range_count;
|
2004-11-17 10:07:52 +01:00
|
|
|
multi_range_sorted= sorted;
|
|
|
|
multi_range_buffer= buffer;
|
|
|
|
|
2004-11-30 07:58:55 +01:00
|
|
|
/**
|
|
|
|
* read multi range will read ranges as follows (if not ordered)
|
|
|
|
*
|
|
|
|
* input read order
|
|
|
|
* ====== ==========
|
|
|
|
* pk-op 1 pk-op 1
|
|
|
|
* pk-op 2 pk-op 2
|
|
|
|
* range 3 range (3,5) NOTE result rows will be intermixed
|
|
|
|
* pk-op 4 pk-op 4
|
|
|
|
* range 5
|
|
|
|
* pk-op 6 pk-ok 6
|
2005-01-11 17:00:31 +01:00
|
|
|
*/
|
|
|
|
|
2005-01-11 14:06:44 +01:00
|
|
|
/**
|
2004-11-30 07:58:55 +01:00
|
|
|
* Variables for loop
|
|
|
|
*/
|
2004-12-08 00:36:40 +01:00
|
|
|
byte *curr= (byte*)buffer->buffer;
|
|
|
|
byte *end_of_buffer= (byte*)buffer->buffer_end;
|
2004-11-17 10:07:52 +01:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
|
|
|
const NDBINDEX *unique_idx= (NDBINDEX *) m_index[active_index].unique_index;
|
2004-11-30 07:58:55 +01:00
|
|
|
const NDBINDEX *idx= (NDBINDEX *) m_index[active_index].index;
|
2004-11-26 15:15:22 +01:00
|
|
|
const NdbOperation* lastOp= m_active_trans->getLastDefinedOperation();
|
|
|
|
NdbIndexScanOperation* scanOp= 0;
|
2005-06-01 15:35:09 +02:00
|
|
|
for (; multi_range_curr<multi_range_end && curr+reclength <= end_of_buffer;
|
|
|
|
multi_range_curr++)
|
2004-11-26 15:15:22 +01:00
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
part_id_range part_spec;
|
|
|
|
if (m_use_partition_function)
|
|
|
|
{
|
|
|
|
get_partition_set(table, curr, active_index,
|
|
|
|
&multi_range_curr->start_key,
|
|
|
|
&part_spec);
|
|
|
|
if (part_spec.start_part > part_spec.end_part)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
We can skip this partition since the key won't fit into any
|
|
|
|
partition
|
|
|
|
*/
|
|
|
|
curr += reclength;
|
|
|
|
multi_range_curr->range_flag |= SKIP_RANGE;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
switch(index_type){
|
2005-07-04 02:42:33 +02:00
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
|
|
|
if (!(multi_range_curr->start_key.length == key_info->key_length &&
|
2005-07-18 13:31:02 +02:00
|
|
|
multi_range_curr->start_key.flag == HA_READ_KEY_EXACT))
|
|
|
|
goto range;
|
|
|
|
// else fall through
|
2004-11-26 15:15:22 +01:00
|
|
|
case PRIMARY_KEY_INDEX:
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
2004-12-28 17:01:07 +01:00
|
|
|
multi_range_curr->range_flag |= UNIQUE_RANGE;
|
2004-11-17 10:07:52 +01:00
|
|
|
if ((op= m_active_trans->getNdbOperation(tab)) &&
|
2005-02-16 14:18:32 +01:00
|
|
|
!op->readTuple(lm) &&
|
|
|
|
!set_primary_key(op, multi_range_curr->start_key.key) &&
|
|
|
|
!define_read_attrs(curr, op) &&
|
2005-07-18 13:31:02 +02:00
|
|
|
(op->setAbortOption(AO_IgnoreError), TRUE) &&
|
|
|
|
(!m_use_partition_function ||
|
|
|
|
(op->setPartitionId(part_spec.start_part), true)))
|
2005-02-16 14:18:32 +01:00
|
|
|
curr += reclength;
|
2004-11-17 10:07:52 +01:00
|
|
|
else
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(op ? op->getNdbError() : m_active_trans->getNdbError());
|
2004-11-26 15:15:22 +01:00
|
|
|
break;
|
2004-11-17 10:07:52 +01:00
|
|
|
}
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case UNIQUE_ORDERED_INDEX:
|
|
|
|
if (!(multi_range_curr->start_key.length == key_info->key_length &&
|
2005-07-18 13:31:02 +02:00
|
|
|
multi_range_curr->start_key.flag == HA_READ_KEY_EXACT &&
|
|
|
|
!check_null_in_key(key_info, multi_range_curr->start_key.key,
|
|
|
|
multi_range_curr->start_key.length)))
|
|
|
|
goto range;
|
|
|
|
// else fall through
|
2004-11-26 15:15:22 +01:00
|
|
|
case UNIQUE_INDEX:
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
2004-12-28 17:01:07 +01:00
|
|
|
multi_range_curr->range_flag |= UNIQUE_RANGE;
|
2004-11-17 10:07:52 +01:00
|
|
|
if ((op= m_active_trans->getNdbIndexOperation(unique_idx, tab)) &&
|
2005-02-16 14:18:32 +01:00
|
|
|
!op->readTuple(lm) &&
|
|
|
|
!set_index_key(op, key_info, multi_range_curr->start_key.key) &&
|
|
|
|
!define_read_attrs(curr, op) &&
|
2005-05-06 10:39:30 +02:00
|
|
|
(op->setAbortOption(AO_IgnoreError), TRUE))
|
2005-02-16 14:18:32 +01:00
|
|
|
curr += reclength;
|
2004-11-17 10:07:52 +01:00
|
|
|
else
|
2005-02-16 14:18:32 +01:00
|
|
|
ERR_RETURN(op ? op->getNdbError() : m_active_trans->getNdbError());
|
2004-11-26 15:15:22 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-07-18 13:31:02 +02:00
|
|
|
case ORDERED_INDEX: {
|
2004-11-26 15:15:22 +01:00
|
|
|
range:
|
2004-12-28 17:01:07 +01:00
|
|
|
multi_range_curr->range_flag &= ~(uint)UNIQUE_RANGE;
|
2004-11-26 15:15:22 +01:00
|
|
|
if (scanOp == 0)
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
if (m_multi_cursor)
|
|
|
|
{
|
|
|
|
scanOp= m_multi_cursor;
|
|
|
|
DBUG_ASSERT(scanOp->getSorted() == sorted);
|
|
|
|
DBUG_ASSERT(scanOp->getLockMode() ==
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type));
|
2005-07-04 02:42:33 +02:00
|
|
|
if (scanOp->reset_bounds(m_force_send))
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(ndb_err(m_active_trans));
|
|
|
|
|
|
|
|
end_of_buffer -= reclength;
|
|
|
|
}
|
|
|
|
else if ((scanOp= m_active_trans->getNdbIndexScanOperation(idx, tab))
|
2005-05-06 10:39:30 +02:00
|
|
|
&&!scanOp->readTuples(lm, 0, parallelism, sorted, FALSE, TRUE)
|
2005-02-16 14:18:32 +01:00
|
|
|
&&!generate_scan_filter(m_cond_stack, scanOp)
|
|
|
|
&&!define_read_attrs(end_of_buffer-reclength, scanOp))
|
|
|
|
{
|
|
|
|
m_multi_cursor= scanOp;
|
|
|
|
m_multi_range_cursor_result_ptr= end_of_buffer-reclength;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ERR_RETURN(scanOp ? scanOp->getNdbError() :
|
|
|
|
m_active_trans->getNdbError());
|
|
|
|
}
|
2004-11-26 15:15:22 +01:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
|
2004-12-28 17:01:07 +01:00
|
|
|
const key_range *keys[2]= { &multi_range_curr->start_key,
|
2005-02-16 14:18:32 +01:00
|
|
|
&multi_range_curr->end_key };
|
2005-09-15 02:33:28 +02:00
|
|
|
if ((res= set_bounds(scanOp, active_index, false, keys,
|
|
|
|
multi_range_curr-ranges)))
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-11-26 15:15:22 +01:00
|
|
|
break;
|
2004-11-17 10:07:52 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case UNDEFINED_INDEX:
|
2005-01-11 14:06:44 +01:00
|
|
|
DBUG_ASSERT(FALSE);
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
break;
|
|
|
|
}
|
2004-11-17 10:07:52 +01:00
|
|
|
}
|
|
|
|
|
2004-12-28 17:01:07 +01:00
|
|
|
if (multi_range_curr != multi_range_end)
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
2004-11-30 07:58:55 +01:00
|
|
|
/**
|
|
|
|
* Mark that we're using entire buffer (even if might not) as
|
|
|
|
* we haven't read all ranges for some reason
|
|
|
|
* This as we don't want mysqld to reuse the buffer when we read
|
|
|
|
* the remaining ranges
|
|
|
|
*/
|
2004-11-26 15:15:22 +01:00
|
|
|
buffer->end_of_used_area= (byte*)buffer->buffer_end;
|
2004-11-17 10:07:52 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
buffer->end_of_used_area= curr;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Set first operation in multi range
|
|
|
|
*/
|
|
|
|
m_current_multi_operation=
|
|
|
|
lastOp ? lastOp->next() : m_active_trans->getFirstDefinedOperation();
|
2004-11-26 15:15:22 +01:00
|
|
|
if (!(res= execute_no_commit_ie(this, m_active_trans)))
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
2004-12-28 17:01:07 +01:00
|
|
|
m_multi_range_defined= multi_range_curr;
|
|
|
|
multi_range_curr= ranges;
|
2004-11-26 15:15:22 +01:00
|
|
|
m_multi_range_result_ptr= (byte*)buffer->buffer;
|
|
|
|
DBUG_RETURN(read_multi_range_next(found_range_p));
|
2004-11-17 10:07:52 +01:00
|
|
|
}
|
|
|
|
ERR_RETURN(m_active_trans->getNdbError());
|
|
|
|
}
|
|
|
|
|
2004-12-08 00:36:40 +01:00
|
|
|
#if 0
|
2005-07-18 13:31:02 +02:00
|
|
|
#define DBUG_MULTI_RANGE(x) DBUG_PRINT("info", ("read_multi_range_next: case %d\n", x));
|
2004-12-08 00:36:40 +01:00
|
|
|
#else
|
|
|
|
#define DBUG_MULTI_RANGE(x)
|
|
|
|
#endif
|
|
|
|
|
2004-11-17 10:07:52 +01:00
|
|
|
int
|
2004-12-28 17:01:07 +01:00
|
|
|
ha_ndbcluster::read_multi_range_next(KEY_MULTI_RANGE ** multi_range_found_p)
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_multi_range_next");
|
2004-11-26 15:15:22 +01:00
|
|
|
if (m_disable_multi_read)
|
2004-11-30 07:58:55 +01:00
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_MULTI_RANGE(11);
|
2004-11-26 15:15:22 +01:00
|
|
|
DBUG_RETURN(handler::read_multi_range_next(multi_range_found_p));
|
2004-11-30 07:58:55 +01:00
|
|
|
}
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2004-11-26 15:15:22 +01:00
|
|
|
int res;
|
2004-12-07 15:15:49 +01:00
|
|
|
int range_no;
|
2005-01-07 07:15:08 +01:00
|
|
|
ulong reclength= table->s->reclength;
|
2004-11-26 15:15:22 +01:00
|
|
|
const NdbOperation* op= m_current_multi_operation;
|
2005-06-01 15:35:09 +02:00
|
|
|
for (;multi_range_curr < m_multi_range_defined; multi_range_curr++)
|
2004-11-17 10:07:52 +01:00
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_MULTI_RANGE(12);
|
|
|
|
if (multi_range_curr->range_flag & SKIP_RANGE)
|
|
|
|
continue;
|
2004-12-28 17:01:07 +01:00
|
|
|
if (multi_range_curr->range_flag & UNIQUE_RANGE)
|
2004-11-26 15:15:22 +01:00
|
|
|
{
|
2004-12-07 15:15:49 +01:00
|
|
|
if (op->getNdbError().code == 0)
|
2005-07-18 13:31:02 +02:00
|
|
|
{
|
|
|
|
DBUG_MULTI_RANGE(13);
|
2005-02-16 14:18:32 +01:00
|
|
|
goto found_next;
|
2005-07-18 13:31:02 +02:00
|
|
|
}
|
2004-11-26 15:15:22 +01:00
|
|
|
|
|
|
|
op= m_active_trans->getNextCompletedOperation(op);
|
|
|
|
m_multi_range_result_ptr += reclength;
|
2004-12-07 15:15:49 +01:00
|
|
|
continue;
|
2004-11-26 15:15:22 +01:00
|
|
|
}
|
2004-12-07 15:15:49 +01:00
|
|
|
else if (m_multi_cursor && !multi_range_sorted)
|
2004-11-26 15:15:22 +01:00
|
|
|
{
|
2004-12-08 00:36:40 +01:00
|
|
|
DBUG_MULTI_RANGE(1);
|
|
|
|
if ((res= fetch_next(m_multi_cursor)) == 0)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_MULTI_RANGE(2);
|
|
|
|
range_no= m_multi_cursor->get_range_no();
|
|
|
|
goto found;
|
2004-12-07 15:15:49 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_MULTI_RANGE(14);
|
2005-02-16 14:18:32 +01:00
|
|
|
goto close_scan;
|
2004-12-07 15:15:49 +01:00
|
|
|
}
|
|
|
|
}
|
2004-12-08 00:36:40 +01:00
|
|
|
else if (m_multi_cursor && multi_range_sorted)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
2004-12-08 00:36:40 +01:00
|
|
|
if (m_active_cursor && (res= fetch_next(m_multi_cursor)))
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_MULTI_RANGE(3);
|
|
|
|
goto close_scan;
|
2004-12-08 00:36:40 +01:00
|
|
|
}
|
2004-11-29 13:50:13 +01:00
|
|
|
|
2004-12-07 15:15:49 +01:00
|
|
|
range_no= m_multi_cursor->get_range_no();
|
2004-12-28 17:01:07 +01:00
|
|
|
uint current_range_no= multi_range_curr - m_multi_ranges;
|
2005-01-11 14:06:44 +01:00
|
|
|
if ((uint) range_no == current_range_no)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_MULTI_RANGE(4);
|
2004-12-07 15:15:49 +01:00
|
|
|
// return current row
|
2005-02-16 14:18:32 +01:00
|
|
|
goto found;
|
2004-12-07 15:15:49 +01:00
|
|
|
}
|
2004-12-28 17:01:07 +01:00
|
|
|
else if (range_no > (int)current_range_no)
|
2004-12-07 15:15:49 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_MULTI_RANGE(5);
|
|
|
|
// wait with current row
|
|
|
|
m_active_cursor= 0;
|
|
|
|
continue;
|
2004-12-07 15:15:49 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_MULTI_RANGE(6);
|
|
|
|
// First fetch from cursor
|
|
|
|
DBUG_ASSERT(range_no == -1);
|
2005-07-04 02:42:33 +02:00
|
|
|
if ((res= m_multi_cursor->nextResult(true)))
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_MULTI_RANGE(15);
|
2005-02-16 14:18:32 +01:00
|
|
|
goto close_scan;
|
|
|
|
}
|
|
|
|
multi_range_curr--; // Will be increased in for-loop
|
|
|
|
continue;
|
2004-12-07 15:15:49 +01:00
|
|
|
}
|
2004-11-26 15:15:22 +01:00
|
|
|
}
|
2004-12-08 00:36:40 +01:00
|
|
|
else /** m_multi_cursor == 0 */
|
2004-11-26 15:15:22 +01:00
|
|
|
{
|
2004-12-08 00:36:40 +01:00
|
|
|
DBUG_MULTI_RANGE(7);
|
2004-11-30 07:58:55 +01:00
|
|
|
/**
|
|
|
|
* Corresponds to range 5 in example in read_multi_range_first
|
|
|
|
*/
|
|
|
|
(void)1;
|
2004-12-07 15:15:49 +01:00
|
|
|
continue;
|
2004-11-26 15:15:22 +01:00
|
|
|
}
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2005-05-06 10:39:30 +02:00
|
|
|
DBUG_ASSERT(FALSE); // Should only get here via goto's
|
2004-12-07 15:15:49 +01:00
|
|
|
close_scan:
|
|
|
|
if (res == 1)
|
|
|
|
{
|
2005-05-06 10:39:30 +02:00
|
|
|
m_multi_cursor->close(FALSE, TRUE);
|
2004-12-07 15:15:49 +01:00
|
|
|
m_active_cursor= m_multi_cursor= 0;
|
2004-12-08 00:36:40 +01:00
|
|
|
DBUG_MULTI_RANGE(8);
|
2004-12-07 15:15:49 +01:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-07-18 13:31:02 +02:00
|
|
|
DBUG_MULTI_RANGE(9);
|
2004-12-07 15:15:49 +01:00
|
|
|
DBUG_RETURN(ndb_err(m_active_trans));
|
|
|
|
}
|
|
|
|
}
|
2004-11-26 15:15:22 +01:00
|
|
|
|
2004-12-28 17:01:07 +01:00
|
|
|
if (multi_range_curr == multi_range_end)
|
2005-07-18 13:31:02 +02:00
|
|
|
{
|
|
|
|
DBUG_MULTI_RANGE(16);
|
2004-11-17 10:07:52 +01:00
|
|
|
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
2005-07-18 13:31:02 +02:00
|
|
|
}
|
2004-11-26 15:15:22 +01:00
|
|
|
|
2004-11-17 10:07:52 +01:00
|
|
|
/**
|
|
|
|
* Read remaining ranges
|
|
|
|
*/
|
|
|
|
DBUG_RETURN(read_multi_range_first(multi_range_found_p,
|
2005-02-16 14:18:32 +01:00
|
|
|
multi_range_curr,
|
|
|
|
multi_range_end - multi_range_curr,
|
|
|
|
multi_range_sorted,
|
|
|
|
multi_range_buffer));
|
2004-11-26 15:15:22 +01:00
|
|
|
|
|
|
|
found:
|
2004-11-30 07:58:55 +01:00
|
|
|
/**
|
|
|
|
* Found a record belonging to a scan
|
|
|
|
*/
|
2004-12-07 15:15:49 +01:00
|
|
|
m_active_cursor= m_multi_cursor;
|
2004-12-28 17:01:07 +01:00
|
|
|
* multi_range_found_p= m_multi_ranges + range_no;
|
2004-12-07 15:15:49 +01:00
|
|
|
memcpy(table->record[0], m_multi_range_cursor_result_ptr, reclength);
|
|
|
|
setup_recattr(m_active_cursor->getFirstRecAttr());
|
2004-11-26 15:15:22 +01:00
|
|
|
unpack_record(table->record[0]);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-07 15:15:49 +01:00
|
|
|
|
2004-11-29 13:50:13 +01:00
|
|
|
found_next:
|
2004-11-30 07:58:55 +01:00
|
|
|
/**
|
|
|
|
* Found a record belonging to a pk/index op,
|
|
|
|
* copy result and move to next to prepare for next call
|
|
|
|
*/
|
2004-12-28 17:01:07 +01:00
|
|
|
* multi_range_found_p= multi_range_curr;
|
2004-11-29 13:50:13 +01:00
|
|
|
memcpy(table->record[0], m_multi_range_result_ptr, reclength);
|
2004-12-01 12:43:30 +01:00
|
|
|
setup_recattr(op->getFirstRecAttr());
|
2004-11-29 13:50:13 +01:00
|
|
|
unpack_record(table->record[0]);
|
2004-12-01 12:43:30 +01:00
|
|
|
table->status= 0;
|
|
|
|
|
2004-11-29 13:50:13 +01:00
|
|
|
multi_range_curr++;
|
2004-12-01 12:43:30 +01:00
|
|
|
m_current_multi_operation= m_active_trans->getNextCompletedOperation(op);
|
2004-11-29 13:50:13 +01:00
|
|
|
m_multi_range_result_ptr += reclength;
|
|
|
|
DBUG_RETURN(0);
|
2004-11-17 10:07:52 +01:00
|
|
|
}
|
|
|
|
|
2004-12-01 12:43:30 +01:00
|
|
|
int
|
|
|
|
ha_ndbcluster::setup_recattr(const NdbRecAttr* curr)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("setup_recattr");
|
|
|
|
|
|
|
|
Field **field, **end;
|
|
|
|
NdbValue *value= m_value;
|
|
|
|
|
2005-01-07 07:15:08 +01:00
|
|
|
end= table->field + table->s->fields;
|
2004-12-01 12:43:30 +01:00
|
|
|
|
|
|
|
for (field= table->field; field < end; field++, value++)
|
|
|
|
{
|
|
|
|
if ((* value).ptr)
|
|
|
|
{
|
|
|
|
DBUG_ASSERT(curr != 0);
|
2004-12-07 15:15:49 +01:00
|
|
|
(* value).rec= curr;
|
|
|
|
curr= curr->next();
|
2004-12-01 12:43:30 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-12-08 00:36:40 +01:00
|
|
|
DBUG_RETURN(0);
|
2004-12-01 12:43:30 +01:00
|
|
|
}
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
char*
|
|
|
|
ha_ndbcluster::update_table_comment(
|
2005-02-16 14:18:32 +01:00
|
|
|
/* out: table comment + additional */
|
|
|
|
const char* comment)/* in: table comment defined by user */
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
uint length= strlen(comment);
|
2005-07-04 02:42:33 +02:00
|
|
|
if (length > 64000 - 3)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
return((char*)comment); /* string too long */
|
|
|
|
}
|
|
|
|
|
|
|
|
Ndb* ndb;
|
|
|
|
if (!(ndb= get_ndb()))
|
|
|
|
{
|
|
|
|
return((char*)comment);
|
|
|
|
}
|
|
|
|
|
|
|
|
ndb->setDatabaseName(m_dbname);
|
|
|
|
NDBDICT* dict= ndb->getDictionary();
|
|
|
|
const NDBTAB* tab;
|
|
|
|
if (!(tab= dict->getTable(m_tabname)))
|
|
|
|
{
|
|
|
|
return((char*)comment);
|
|
|
|
}
|
|
|
|
|
|
|
|
char *str;
|
|
|
|
const char *fmt="%s%snumber_of_replicas: %d";
|
|
|
|
const unsigned fmt_len_plus_extra= length + strlen(fmt);
|
|
|
|
if ((str= my_malloc(fmt_len_plus_extra, MYF(0))) == NULL)
|
|
|
|
{
|
|
|
|
return (char*)comment;
|
|
|
|
}
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
my_snprintf(str,fmt_len_plus_extra,fmt,comment,
|
|
|
|
length > 0 ? " ":"",
|
|
|
|
tab->getReplicaCount());
|
2005-02-11 22:33:52 +01:00
|
|
|
return str;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Utility thread main loop
|
2005-10-08 16:39:55 +02:00
|
|
|
pthread_handler_t ndb_util_thread_func(void *arg __attribute__((unused)))
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
THD *thd; /* needs to be first for thread_stack */
|
2005-03-15 15:03:25 +01:00
|
|
|
Ndb* ndb;
|
2005-02-11 22:33:52 +01:00
|
|
|
struct timespec abstime;
|
2005-11-06 00:20:37 +01:00
|
|
|
List<NDB_SHARE> util_open_tables;
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
my_thread_init();
|
|
|
|
DBUG_ENTER("ndb_util_thread");
|
|
|
|
DBUG_PRINT("enter", ("ndb_cache_check_time: %d", ndb_cache_check_time));
|
|
|
|
|
|
|
|
thd= new THD; /* note that contructor of THD uses DBUG_ */
|
|
|
|
THD_CHECK_SENTRY(thd);
|
2005-03-15 15:03:25 +01:00
|
|
|
ndb= new Ndb(g_ndb_cluster_connection, "");
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
pthread_detach_this_thread();
|
|
|
|
ndb_util_thread= pthread_self();
|
|
|
|
|
|
|
|
thd->thread_stack= (char*)&thd; /* remember where our stack is */
|
2005-03-15 15:03:25 +01:00
|
|
|
if (thd->store_globals() && (ndb->init() != -1))
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
thd->cleanup();
|
|
|
|
delete thd;
|
2005-04-06 12:27:43 +02:00
|
|
|
delete ndb;
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_RETURN(NULL);
|
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
thd->init_for_queries();
|
|
|
|
thd->version=refresh_version;
|
|
|
|
thd->set_time();
|
|
|
|
thd->main_security_ctx.host_or_ip= "";
|
|
|
|
thd->client_capabilities = 0;
|
|
|
|
my_net_init(&thd->net, 0);
|
|
|
|
thd->main_security_ctx.master_access= ~0;
|
|
|
|
thd->main_security_ctx.priv_user = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
wait for mysql server to start
|
|
|
|
*/
|
|
|
|
pthread_mutex_lock(&LOCK_server_started);
|
|
|
|
while (!mysqld_server_started)
|
|
|
|
pthread_cond_wait(&COND_server_started, &LOCK_server_started);
|
|
|
|
pthread_mutex_unlock(&LOCK_server_started);
|
|
|
|
|
|
|
|
/*
|
|
|
|
Wait for cluster to start
|
|
|
|
*/
|
|
|
|
pthread_mutex_lock(&LOCK_ndb_util_thread);
|
|
|
|
while (!ndb_cluster_node_id)
|
|
|
|
{
|
|
|
|
/* ndb not connected yet */
|
|
|
|
set_timespec(abstime, 1);
|
|
|
|
pthread_cond_timedwait(&COND_ndb_util_thread,
|
|
|
|
&LOCK_ndb_util_thread,
|
|
|
|
&abstime);
|
|
|
|
if (abort_loop)
|
|
|
|
{
|
|
|
|
pthread_mutex_unlock(&LOCK_ndb_util_thread);
|
|
|
|
goto ndb_util_thread_end;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
pthread_mutex_unlock(&LOCK_ndb_util_thread);
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get all table definitions from the storage node
|
|
|
|
*/
|
|
|
|
ndbcluster_find_all_files(thd);
|
|
|
|
|
|
|
|
ndbcluster_util_inited= 1;
|
2005-02-11 22:33:52 +01:00
|
|
|
|
2005-02-21 14:04:54 +01:00
|
|
|
set_timespec(abstime, 0);
|
2005-11-06 00:20:37 +01:00
|
|
|
for (;!abort_loop;)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
|
|
|
|
pthread_mutex_lock(&LOCK_ndb_util_thread);
|
2005-10-11 23:58:22 +02:00
|
|
|
pthread_cond_timedwait(&COND_ndb_util_thread,
|
|
|
|
&LOCK_ndb_util_thread,
|
|
|
|
&abstime);
|
2005-02-11 22:33:52 +01:00
|
|
|
pthread_mutex_unlock(&LOCK_ndb_util_thread);
|
2005-11-06 00:20:37 +01:00
|
|
|
#ifdef NDB_EXTRA_DEBUG_UTIL_THREAD
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_PRINT("ndb_util_thread", ("Started, ndb_cache_check_time: %d",
|
|
|
|
ndb_cache_check_time));
|
2005-11-06 00:20:37 +01:00
|
|
|
#endif
|
2005-02-11 22:33:52 +01:00
|
|
|
if (abort_loop)
|
|
|
|
break; /* Shutting down server */
|
|
|
|
|
|
|
|
if (ndb_cache_check_time == 0)
|
|
|
|
{
|
2005-03-15 15:03:25 +01:00
|
|
|
/* Wake up in 1 second to check if value has changed */
|
|
|
|
set_timespec(abstime, 1);
|
2005-02-11 22:33:52 +01:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Lock mutex and fill list with pointers to all open tables */
|
|
|
|
NDB_SHARE *share;
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
|
|
|
for (uint i= 0; i < ndbcluster_open_tables.records; i++)
|
|
|
|
{
|
|
|
|
share= (NDB_SHARE *)hash_element(&ndbcluster_open_tables, i);
|
|
|
|
share->use_count++; /* Make sure the table can't be closed */
|
|
|
|
DBUG_PRINT("ndb_util_thread",
|
|
|
|
("Found open table[%d]: %s, use_count: %d",
|
|
|
|
i, share->table_name, share->use_count));
|
|
|
|
|
|
|
|
/* Store pointer to table */
|
|
|
|
util_open_tables.push_back(share);
|
|
|
|
}
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
|
|
|
|
/* Iterate through the open files list */
|
|
|
|
List_iterator_fast<NDB_SHARE> it(util_open_tables);
|
2005-02-16 14:18:32 +01:00
|
|
|
while ((share= it++))
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("ndb_util_thread",
|
2005-03-15 15:03:25 +01:00
|
|
|
("Fetching commit count for: %s",
|
2005-11-06 00:20:37 +01:00
|
|
|
share->key));
|
2005-02-11 22:33:52 +01:00
|
|
|
|
|
|
|
/* Contact NDB to get commit count for table */
|
2005-11-06 00:20:37 +01:00
|
|
|
ndb->setDatabaseName(share->db);
|
2005-03-15 15:03:25 +01:00
|
|
|
struct Ndb_statistics stat;
|
|
|
|
|
|
|
|
uint lock;
|
|
|
|
pthread_mutex_lock(&share->mutex);
|
|
|
|
lock= share->commit_count_lock;
|
|
|
|
pthread_mutex_unlock(&share->mutex);
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
if (ndb_get_table_statistics(ndb, share->table_name, &stat) == 0)
|
2005-02-11 22:33:52 +01:00
|
|
|
{
|
|
|
|
DBUG_PRINT("ndb_util_thread",
|
2005-03-15 15:03:25 +01:00
|
|
|
("Table: %s, commit_count: %llu, rows: %llu",
|
2005-11-06 00:20:37 +01:00
|
|
|
share->key, stat.commit_count, stat.row_count));
|
2005-02-11 22:33:52 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("ndb_util_thread",
|
|
|
|
("Error: Could not get commit count for table %s",
|
2005-11-06 00:20:37 +01:00
|
|
|
share->key));
|
2005-03-15 15:03:25 +01:00
|
|
|
stat.commit_count= 0;
|
2005-02-11 22:33:52 +01:00
|
|
|
}
|
2005-03-15 15:03:25 +01:00
|
|
|
|
|
|
|
pthread_mutex_lock(&share->mutex);
|
|
|
|
if (share->commit_count_lock == lock)
|
|
|
|
share->commit_count= stat.commit_count;
|
|
|
|
pthread_mutex_unlock(&share->mutex);
|
|
|
|
|
2005-02-11 22:33:52 +01:00
|
|
|
/* Decrease the use count and possibly free share */
|
2005-11-06 00:20:37 +01:00
|
|
|
free_share(&share);
|
2005-02-11 22:33:52 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Clear the list of open tables */
|
|
|
|
util_open_tables.empty();
|
|
|
|
|
2005-03-15 15:03:25 +01:00
|
|
|
/* Calculate new time to wake up */
|
|
|
|
int secs= 0;
|
|
|
|
int msecs= ndb_cache_check_time;
|
|
|
|
|
|
|
|
struct timeval tick_time;
|
|
|
|
gettimeofday(&tick_time, 0);
|
|
|
|
abstime.tv_sec= tick_time.tv_sec;
|
|
|
|
abstime.tv_nsec= tick_time.tv_usec * 1000;
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
if (msecs >= 1000){
|
2005-03-15 15:03:25 +01:00
|
|
|
secs= msecs / 1000;
|
|
|
|
msecs= msecs % 1000;
|
|
|
|
}
|
|
|
|
|
|
|
|
abstime.tv_sec+= secs;
|
|
|
|
abstime.tv_nsec+= msecs * 1000000;
|
|
|
|
if (abstime.tv_nsec >= 1000000000) {
|
|
|
|
abstime.tv_sec+= 1;
|
|
|
|
abstime.tv_nsec-= 1000000000;
|
|
|
|
}
|
2005-02-11 22:33:52 +01:00
|
|
|
}
|
2005-11-06 00:20:37 +01:00
|
|
|
ndb_util_thread_end:
|
|
|
|
net_end(&thd->net);
|
2005-02-11 22:33:52 +01:00
|
|
|
thd->cleanup();
|
|
|
|
delete thd;
|
2005-04-06 12:27:43 +02:00
|
|
|
delete ndb;
|
2005-02-11 22:33:52 +01:00
|
|
|
DBUG_PRINT("exit", ("ndb_util_thread"));
|
|
|
|
my_thread_end();
|
|
|
|
pthread_exit(0);
|
|
|
|
DBUG_RETURN(NULL);
|
|
|
|
}
|
|
|
|
|
2004-12-17 21:13:22 +01:00
|
|
|
/*
|
|
|
|
Condition pushdown
|
|
|
|
*/
|
2005-02-16 14:18:32 +01:00
|
|
|
/*
|
|
|
|
Push a condition to ndbcluster storage engine for evaluation
|
|
|
|
during table and index scans. The conditions will be stored on a stack
|
|
|
|
for possibly storing several conditions. The stack can be popped
|
|
|
|
by calling cond_pop, handler::extra(HA_EXTRA_RESET) (handler::reset())
|
|
|
|
will clear the stack.
|
|
|
|
The current implementation supports arbitrary AND/OR nested conditions
|
|
|
|
with comparisons between columns and constants (including constant
|
|
|
|
expressions and function calls) and the following comparison operators:
|
|
|
|
=, !=, >, >=, <, <=, "is null", and "is not null".
|
|
|
|
|
|
|
|
RETURN
|
|
|
|
NULL The condition was supported and will be evaluated for each
|
|
|
|
row found during the scan
|
|
|
|
cond The condition was not supported and all rows will be returned from
|
|
|
|
the scan for evaluation (and thus not saved on stack)
|
|
|
|
*/
|
2004-12-17 21:13:22 +01:00
|
|
|
const
|
|
|
|
COND*
|
|
|
|
ha_ndbcluster::cond_push(const COND *cond)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("cond_push");
|
2005-05-08 22:42:40 +02:00
|
|
|
Ndb_cond_stack *ndb_cond = new Ndb_cond_stack();
|
|
|
|
DBUG_EXECUTE("where",print_where((COND *)cond, m_tabname););
|
|
|
|
if (m_cond_stack)
|
2005-05-07 22:41:30 +02:00
|
|
|
ndb_cond->next= m_cond_stack;
|
2005-05-08 22:42:40 +02:00
|
|
|
else
|
|
|
|
ndb_cond->next= NULL;
|
|
|
|
m_cond_stack= ndb_cond;
|
|
|
|
|
|
|
|
if (serialize_cond(cond, ndb_cond))
|
|
|
|
{
|
2005-02-11 22:05:24 +01:00
|
|
|
DBUG_RETURN(NULL);
|
2005-05-08 22:42:40 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
cond_pop();
|
2005-02-11 22:05:24 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
DBUG_RETURN(cond);
|
|
|
|
}
|
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
/*
|
|
|
|
Pop the top condition from the condition stack of the handler instance.
|
|
|
|
*/
|
2004-12-17 21:13:22 +01:00
|
|
|
void
|
|
|
|
ha_ndbcluster::cond_pop()
|
|
|
|
{
|
|
|
|
Ndb_cond_stack *ndb_cond_stack= m_cond_stack;
|
|
|
|
if (ndb_cond_stack)
|
|
|
|
{
|
|
|
|
m_cond_stack= ndb_cond_stack->next;
|
|
|
|
delete ndb_cond_stack;
|
|
|
|
}
|
2005-02-11 22:05:24 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
/*
|
|
|
|
Clear the condition stack
|
|
|
|
*/
|
2004-12-17 21:13:22 +01:00
|
|
|
void
|
|
|
|
ha_ndbcluster::cond_clear()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("cond_clear");
|
|
|
|
while (m_cond_stack)
|
|
|
|
cond_pop();
|
|
|
|
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
/*
|
|
|
|
Serialize the item tree into a linked list represented by Ndb_cond
|
|
|
|
for fast generation of NbdScanFilter. Adds information such as
|
|
|
|
position of fields that is not directly available in the Item tree.
|
|
|
|
Also checks if condition is supported.
|
|
|
|
*/
|
2004-12-17 21:13:22 +01:00
|
|
|
void ndb_serialize_cond(const Item *item, void *arg)
|
|
|
|
{
|
|
|
|
Ndb_cond_traverse_context *context= (Ndb_cond_traverse_context *) arg;
|
|
|
|
DBUG_ENTER("ndb_serialize_cond");
|
|
|
|
|
2005-02-11 22:05:24 +01:00
|
|
|
// Check if we are skipping arguments to a function to be evaluated
|
|
|
|
if (context->skip)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Skiping argument %d", context->skip));
|
|
|
|
context->skip--;
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (item->type()) {
|
|
|
|
case Item::FUNC_ITEM:
|
|
|
|
{
|
2005-02-11 22:05:24 +01:00
|
|
|
Item_func *func_item= (Item_func *) item;
|
|
|
|
context->skip+= func_item->argument_count();
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::INT_ITEM:
|
|
|
|
case Item::REAL_ITEM:
|
|
|
|
case Item::STRING_ITEM:
|
|
|
|
case Item::VARBIN_ITEM:
|
|
|
|
case Item::DECIMAL_ITEM:
|
2005-02-11 22:05:24 +01:00
|
|
|
break;
|
|
|
|
default:
|
2005-02-16 14:18:32 +01:00
|
|
|
context->supported= FALSE;
|
2005-02-11 22:05:24 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
|
2005-02-11 22:05:24 +01:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
if (context->supported)
|
2004-12-17 21:13:22 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Ndb_rewrite_context *rewrite_context= context->rewrite_stack;
|
|
|
|
const Item_func *func_item;
|
|
|
|
// Check if we are rewriting some unsupported function call
|
|
|
|
if (rewrite_context &&
|
|
|
|
(func_item= rewrite_context->func_item) &&
|
|
|
|
rewrite_context->count++ == 0)
|
2005-02-11 22:05:24 +01:00
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (func_item->functype()) {
|
|
|
|
case Item_func::BETWEEN:
|
2005-02-16 14:18:32 +01:00
|
|
|
/*
|
2005-05-18 14:32:05 +02:00
|
|
|
Rewrite
|
|
|
|
<field>|<const> BETWEEN <const1>|<field1> AND <const2>|<field2>
|
|
|
|
to <field>|<const> > <const1>|<field1> AND
|
|
|
|
<field>|<const> < <const2>|<field2>
|
|
|
|
or actually in prefix format
|
|
|
|
BEGIN(AND) GT(<field>|<const>, <const1>|<field1>),
|
|
|
|
LT(<field>|<const>, <const2>|<field2>), END()
|
2005-02-16 14:18:32 +01:00
|
|
|
*/
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::IN_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
/*
|
|
|
|
Rewrite <field>|<const> IN(<const1>|<field1>, <const2>|<field2>,..)
|
|
|
|
to <field>|<const> = <const1>|<field1> OR
|
|
|
|
<field> = <const2>|<field2> ...
|
|
|
|
or actually in prefix format
|
|
|
|
BEGIN(OR) EQ(<field>|<const>, <const1><field1>),
|
|
|
|
EQ(<field>|<const>, <const2>|<field2>), ... END()
|
|
|
|
Each part of the disjunction is added for each call
|
|
|
|
to ndb_serialize_cond and end of rewrite statement
|
|
|
|
is wrapped in end of ndb_serialize_cond
|
|
|
|
*/
|
|
|
|
if (context->expecting(item->type()))
|
|
|
|
{
|
|
|
|
// This is the <field>|<const> item, save it in the rewrite context
|
|
|
|
rewrite_context->left_hand_item= item;
|
|
|
|
if (item->type() == Item::FUNC_ITEM)
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_func *func_item= (Item_func *) item;
|
|
|
|
if (func_item->functype() == Item_func::UNKNOWN_FUNC &&
|
|
|
|
func_item->const_item())
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
// Skip any arguments since we will evaluate function instead
|
|
|
|
DBUG_PRINT("info", ("Skip until end of arguments marker"));
|
|
|
|
context->skip= func_item->argument_count();
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("Found unsupported functional expression in BETWEEN|IN"));
|
|
|
|
context->supported= FALSE;
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2005-02-23 15:51:26 +01:00
|
|
|
else
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
// Non-supported BETWEEN|IN expression
|
|
|
|
DBUG_PRINT("info", ("Found unexpected item of type %u in BETWEEN|IN",
|
|
|
|
item->type()));
|
2005-02-23 15:51:26 +01:00
|
|
|
context->supported= FALSE;
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_VOID_RETURN;
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
Ndb_cond_stack *ndb_stack= context->stack_ptr;
|
|
|
|
Ndb_cond *prev_cond= context->cond_ptr;
|
|
|
|
Ndb_cond *curr_cond= context->cond_ptr= new Ndb_cond();
|
|
|
|
if (!ndb_stack->ndb_cond)
|
|
|
|
ndb_stack->ndb_cond= curr_cond;
|
|
|
|
curr_cond->prev= prev_cond;
|
|
|
|
if (prev_cond) prev_cond->next= curr_cond;
|
|
|
|
// Check if we are rewriting some unsupported function call
|
|
|
|
if (context->rewrite_stack)
|
|
|
|
{
|
|
|
|
Ndb_rewrite_context *rewrite_context= context->rewrite_stack;
|
|
|
|
const Item_func *func_item= rewrite_context->func_item;
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (func_item->functype()) {
|
|
|
|
case Item_func::BETWEEN:
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Rewrite
|
|
|
|
<field>|<const> BETWEEN <const1>|<field1> AND <const2>|<field2>
|
|
|
|
to <field>|<const> > <const1>|<field1> AND
|
|
|
|
<field>|<const> < <const2>|<field2>
|
|
|
|
or actually in prefix format
|
|
|
|
BEGIN(AND) GT(<field>|<const>, <const1>|<field1>),
|
|
|
|
LT(<field>|<const>, <const2>|<field2>), END()
|
|
|
|
*/
|
2005-05-18 14:32:05 +02:00
|
|
|
if (rewrite_context->count == 2)
|
|
|
|
{
|
|
|
|
// Lower limit of BETWEEN
|
|
|
|
DBUG_PRINT("info", ("GE_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(Item_func::GE_FUNC, 2);
|
|
|
|
}
|
|
|
|
else if (rewrite_context->count == 3)
|
|
|
|
{
|
|
|
|
// Upper limit of BETWEEN
|
|
|
|
DBUG_PRINT("info", ("LE_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(Item_func::LE_FUNC, 2);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Illegal BETWEEN expression
|
|
|
|
DBUG_PRINT("info", ("Illegal BETWEEN expression"));
|
|
|
|
context->supported= FALSE;
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::IN_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
/*
|
|
|
|
Rewrite <field>|<const> IN(<const1>|<field1>, <const2>|<field2>,..)
|
|
|
|
to <field>|<const> = <const1>|<field1> OR
|
|
|
|
<field> = <const2>|<field2> ...
|
|
|
|
or actually in prefix format
|
|
|
|
BEGIN(OR) EQ(<field>|<const>, <const1><field1>),
|
|
|
|
EQ(<field>|<const>, <const2>|<field2>), ... END()
|
|
|
|
Each part of the disjunction is added for each call
|
|
|
|
to ndb_serialize_cond and end of rewrite statement
|
|
|
|
is wrapped in end of ndb_serialize_cond
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("info", ("EQ_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(Item_func::EQ_FUNC, 2);
|
2005-02-16 14:18:32 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
default:
|
|
|
|
context->supported= FALSE;
|
2005-02-18 21:43:51 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
// Handle left hand <field>|<const>
|
|
|
|
context->rewrite_stack= NULL; // Disable rewrite mode
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FUNC_ITEM);
|
|
|
|
ndb_serialize_cond(rewrite_context->left_hand_item, arg);
|
|
|
|
context->skip= 0; // Any FUNC_ITEM expression has already been parsed
|
|
|
|
context->rewrite_stack= rewrite_context; // Enable rewrite mode
|
|
|
|
if (!context->supported)
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
|
|
|
|
prev_cond= context->cond_ptr;
|
|
|
|
curr_cond= context->cond_ptr= new Ndb_cond();
|
|
|
|
prev_cond->next= curr_cond;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check for end of AND/OR expression
|
|
|
|
if (!item)
|
|
|
|
{
|
|
|
|
// End marker for condition group
|
|
|
|
DBUG_PRINT("info", ("End of condition group"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_END_COND);
|
|
|
|
}
|
|
|
|
else
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
|
|
|
switch (item->type()) {
|
|
|
|
case Item::FIELD_ITEM:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_field *field_item= (Item_field *) item;
|
|
|
|
Field *field= field_item->field;
|
|
|
|
enum_field_types type= field->type();
|
|
|
|
/*
|
|
|
|
Check that the field is part of the table of the handler
|
|
|
|
instance and that we expect a field with of this result type.
|
|
|
|
*/
|
|
|
|
if (context->table == field->table)
|
|
|
|
{
|
|
|
|
const NDBTAB *tab= (const NDBTAB *) context->ndb_table;
|
|
|
|
DBUG_PRINT("info", ("FIELD_ITEM"));
|
|
|
|
DBUG_PRINT("info", ("table %s", tab->getName()));
|
|
|
|
DBUG_PRINT("info", ("column %s", field->field_name));
|
|
|
|
DBUG_PRINT("info", ("result type %d", field->result_type()));
|
|
|
|
|
|
|
|
// Check that we are expecting a field and with the correct
|
|
|
|
// result type
|
|
|
|
if (context->expecting(Item::FIELD_ITEM) &&
|
|
|
|
(context->expecting_field_result(field->result_type()) ||
|
2005-07-31 21:33:46 +02:00
|
|
|
// Date and year can be written as string or int
|
2005-05-18 14:32:05 +02:00
|
|
|
((type == MYSQL_TYPE_TIME ||
|
|
|
|
type == MYSQL_TYPE_DATE ||
|
|
|
|
type == MYSQL_TYPE_YEAR ||
|
|
|
|
type == MYSQL_TYPE_DATETIME)
|
2005-07-31 21:33:46 +02:00
|
|
|
? (context->expecting_field_result(STRING_RESULT) ||
|
|
|
|
context->expecting_field_result(INT_RESULT))
|
|
|
|
: true)) &&
|
2005-05-18 14:32:05 +02:00
|
|
|
// Bit fields no yet supported in scan filter
|
|
|
|
type != MYSQL_TYPE_BIT)
|
|
|
|
{
|
|
|
|
const NDBCOL *col= tab->getColumn(field->field_name);
|
|
|
|
DBUG_ASSERT(col);
|
|
|
|
curr_cond->ndb_item= new Ndb_item(field, col->getColumnNo());
|
|
|
|
context->dont_expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_no_field_result();
|
|
|
|
if (context->expect_mask)
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
// We have not seen second argument yet
|
|
|
|
if (type == MYSQL_TYPE_TIME ||
|
|
|
|
type == MYSQL_TYPE_DATE ||
|
|
|
|
type == MYSQL_TYPE_YEAR ||
|
|
|
|
type == MYSQL_TYPE_DATETIME)
|
|
|
|
{
|
|
|
|
context->expect_only(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
}
|
|
|
|
else
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (field->result_type()) {
|
|
|
|
case STRING_RESULT:
|
2005-05-18 14:32:05 +02:00
|
|
|
// Expect char string or binary string
|
|
|
|
context->expect_only(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect_collation(field_item->collation.collation);
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case REAL_RESULT:
|
2005-05-18 14:32:05 +02:00
|
|
|
context->expect_only(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
2005-06-10 10:34:05 +02:00
|
|
|
context->expect(Item::INT_ITEM);
|
2005-05-18 14:32:05 +02:00
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case INT_RESULT:
|
2005-05-18 14:32:05 +02:00
|
|
|
context->expect_only(Item::INT_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case DECIMAL_RESULT:
|
2005-05-18 14:32:05 +02:00
|
|
|
context->expect_only(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
2005-06-10 10:34:05 +02:00
|
|
|
context->expect(Item::INT_ITEM);
|
2005-05-18 14:32:05 +02:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
|
|
|
else
|
2005-02-23 15:51:26 +01:00
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
2005-05-18 14:32:05 +02:00
|
|
|
// Check that field and string constant collations are the same
|
|
|
|
if ((field->result_type() == STRING_RESULT) &&
|
|
|
|
!context->expecting_collation(item->collation.collation)
|
|
|
|
&& type != MYSQL_TYPE_TIME
|
|
|
|
&& type != MYSQL_TYPE_DATE
|
|
|
|
&& type != MYSQL_TYPE_YEAR
|
|
|
|
&& type != MYSQL_TYPE_DATETIME)
|
2005-02-23 15:51:26 +01:00
|
|
|
{
|
2005-02-23 17:18:06 +01:00
|
|
|
DBUG_PRINT("info", ("Found non-matching collation %s",
|
2005-05-18 14:32:05 +02:00
|
|
|
item->collation.collation->name));
|
|
|
|
context->supported= FALSE;
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
else
|
|
|
|
{
|
2005-07-31 21:33:46 +02:00
|
|
|
DBUG_PRINT("info", ("Was not expecting field of type %u(%u)",
|
|
|
|
field->result_type(), type));
|
2005-05-18 14:32:05 +02:00
|
|
|
context->supported= FALSE;
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
else
|
2005-05-18 14:32:05 +02:00
|
|
|
{
|
Table definition cache, part 2
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
Other noteworthy changes:
- In TABLE_SHARE the most common strings are now LEX_STRING's
- Better error message when table is not found
- Variable table_cache is now renamed 'table_open_cache'
- New variable 'table_definition_cache' that is the number of table defintions that will be cached
- strxnmov() calls are now fixed to avoid overflows
- strxnmov() will now always add one end \0 to result
- engine objects are now created with a TABLE_SHARE object instead of a TABLE object.
- After creating a field object one must call field->init(table) before using it
- For a busy system this change will give you:
- Less memory usage for table object
- Faster opening of tables (if it's has been in use or is in table definition cache)
- Allow you to cache many table definitions objects
- Faster drop of table
mysql-test/mysql-test-run.sh:
Fixed some problems with --gdb option
Test both with socket and tcp/ip port that all old servers are killed
mysql-test/r/flush_table.result:
More tests with lock table with 2 threads + flush table
mysql-test/r/information_schema.result:
Removed old (now wrong) result
mysql-test/r/innodb.result:
Better error messages (thanks to TDC patch)
mysql-test/r/merge.result:
Extra flush table test
mysql-test/r/ndb_bitfield.result:
Better error messages (thanks to TDC patch)
mysql-test/r/ndb_partition_error.result:
Better error messages (thanks to TDC patch)
mysql-test/r/query_cache.result:
Remove tables left from old tests
mysql-test/r/temp_table.result:
Test truncate with temporary tables
mysql-test/r/variables.result:
Table_cache -> Table_open_cache
mysql-test/t/flush_table.test:
More tests with lock table with 2 threads + flush table
mysql-test/t/merge.test:
Extra flush table test
mysql-test/t/multi_update.test:
Added 'sleep' to make test predictable
mysql-test/t/query_cache.test:
Remove tables left from old tests
mysql-test/t/temp_table.test:
Test truncate with temporary tables
mysql-test/t/variables.test:
Table_cache -> Table_open_cache
mysql-test/valgrind.supp:
Remove warning that may happens becasue threads dies in different order
mysys/hash.c:
Fixed wrong DBUG_PRINT
mysys/mf_dirname.c:
More DBUG
mysys/mf_pack.c:
Better comment
mysys/mf_tempdir.c:
More DBUG
Ensure that we call cleanup_dirname() on all temporary directory paths.
If we don't do this, we will get a failure when comparing temporary table
names as in some cases the temporary table name is run through convert_dirname())
mysys/my_alloc.c:
Indentation fix
sql/examples/ha_example.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_example.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/examples/ha_tina.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/field.cc:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Use s->db instead of s->table_cache_key
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/field.h:
Update for table definition cache:
- Field creation now takes TABLE_SHARE instead of TABLE as argument
(This is becasue field definitions are now cached in TABLE_SHARE)
When a field is created, one now must call field->init(TABLE) before using it
- Added Field::clone() to create a field in TABLE from a field in TABLE_SHARE
- make_field() takes TABLE_SHARE as argument instead of TABLE
- move_field() -> move_field_offset()
sql/ha_archive.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_archive.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_berkeley.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Changed name of argument create() to not hide internal 'table' variable.
table->s -> table_share
sql/ha_berkeley.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_blackhole.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_federated.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed comments
Remove index variable and replace with pointers (simple optimization)
move_field() -> move_field_offset()
Removed some strlen() calls
sql/ha_federated.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_heap.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Simplify delete_table() and create() as the given file names are now without extension
sql/ha_heap.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_innodb.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisam.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Remove not needed fn_format()
Fixed for new table->s structure
sql/ha_myisam.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_myisammrg.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Don't set 'is_view' for MERGE tables
Use new interface to find_temporary_table()
sql/ha_myisammrg.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Added flag HA_NO_COPY_ON_ALTER
sql/ha_ndbcluster.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Fixed wrong calls to strxnmov()
Give error HA_ERR_TABLE_DEF_CHANGED if table definition has changed
drop_table -> intern_drop_table()
table->s -> table_share
Move part_info to TABLE
Fixed comments & DBUG print's
New arguments to print_error()
sql/ha_ndbcluster.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
sql/ha_partition.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
We can't set up or use part_info when creating handler as there is not yet any table object
New ha_intialise() to work with TDC (Done by Mikael)
sql/ha_partition.h:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
Got set_part_info() from Mikael
sql/handler.cc:
We new use TABLE_SHARE instead of TABLE when creating engine handlers
ha_delete_table() now also takes database as an argument
handler::ha_open() now takes TABLE as argument
ha_open() now calls ha_allocate_read_write_set()
Simplify ha_allocate_read_write_set()
Remove ha_deallocate_read_write_set()
Use table_share (Cached by table definition cache)
sql/handler.h:
New table flag: HA_NO_COPY_ON_ALTER (used by merge tables)
Remove ha_deallocate_read_write_set()
get_new_handler() now takes TABLE_SHARE as argument
ha_delete_table() now gets database as argument
sql/item.cc:
table_name and db are now LEX_STRING objects
When creating fields, we have now have to call field->init(table)
move_field -> move_field_offset()
sql/item.h:
tmp_table_field_from_field_type() now takes an extra paramenter 'fixed_length' to allow one to force usage of CHAR
instead of BLOB
sql/item_cmpfunc.cc:
Fixed call to tmp_table_field_from_field_type()
sql/item_create.cc:
Assert if new not handled cast type
sql/item_func.cc:
When creating fields, we have now have to call field->init(table)
dummy_table used by 'sp' now needs a TABLE_SHARE object
sql/item_subselect.cc:
Trivial code cleanups
sql/item_sum.cc:
When creating fields, we have now have to call field->init(table)
sql/item_timefunc.cc:
Item_func_str_to_date::tmp_table_field() now replaced by call to
tmp_table_field_from_field_type() (see item_timefunc.h)
sql/item_timefunc.h:
Simply tmp_table_field()
sql/item_uniq.cc:
When creating fields, we have now have to call field->init(table)
sql/key.cc:
Added 'KEY' argument to 'find_ref_key' to simplify code
sql/lock.cc:
More debugging
Use create_table_def_key() to create key for table cache
Allocate TABLE_SHARE properly when creating name lock
Fix that locked_table_name doesn't test same table twice
sql/mysql_priv.h:
New functions for table definition cache
New interfaces to a lot of functions.
New faster interface to find_temporary_table() and close_temporary_table()
sql/mysqld.cc:
Added support for table definition cache of size 'table_def_size'
Fixed som calls to strnmov()
Changed name of 'table_cache' to 'table_open_cache'
sql/opt_range.cc:
Use new interfaces
Fixed warnings from valgrind
sql/parse_file.cc:
Safer calls to strxnmov()
Fixed typo
sql/set_var.cc:
Added variable 'table_definition_cache'
Variable table_cache renamed to 'table_open_cache'
sql/slave.cc:
Use new interface
sql/sp.cc:
Proper use of TABLE_SHARE
sql/sp_head.cc:
Remove compiler warnings
We have now to call field->init(table)
sql/sp_head.h:
Pointers to parsed strings are now const
sql/sql_acl.cc:
table_name is now a LEX_STRING
sql/sql_base.cc:
Main implementation of table definition cache
(The #ifdef's are there for the future when table definition cache will replace open table cache)
Now table definitions are cached indepndent of open tables, which will speed up things when a table is in use at once from several places
Views are not yet cached; For the moment we only cache if a table is a view or not.
Faster implementation of find_temorary_table()
Replace 'wait_for_refresh()' with the more general function 'wait_for_condition()'
Drop table is slightly faster as we can use the table definition cache to know the type of the table
sql/sql_cache.cc:
table_cache_key and table_name are now LEX_STRING
'sDBUG print fixes
sql/sql_class.cc:
table_cache_key is now a LEX_STRING
safer strxnmov()
sql/sql_class.h:
Added number of open table shares (table definitions)
sql/sql_db.cc:
safer strxnmov()
sql/sql_delete.cc:
Use new interface to find_temporary_table()
sql/sql_derived.cc:
table_name is now a LEX_STRING
sql/sql_handler.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_insert.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
sql/sql_lex.cc:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_lex.h:
Make parsed string a const (to quickly find out if anything is trying to change the query string)
sql/sql_load.cc:
Safer strxnmov()
sql/sql_parse.cc:
Better error if wrong DB name
sql/sql_partition.cc:
part_info moved to TABLE from TABLE_SHARE
Indentation changes
sql/sql_select.cc:
Indentation fixes
Call field->init(TABLE) for new created fields
Update create_tmp_table() to use TABLE_SHARE properly
sql/sql_select.h:
Call field->init(TABLE) for new created fields
sql/sql_show.cc:
table_name is now a LEX_STRING
part_info moved to TABLE
sql/sql_table.cc:
Use table definition cache to speed up delete of tables
Fixed calls to functions with new interfaces
Don't use 'share_not_to_be_used'
Instead of doing openfrm() when doing repair, we now have to call
get_table_share() followed by open_table_from_share().
Replace some fn_format() with faster unpack_filename().
Safer strxnmov()
part_info is now in TABLE
Added Mikaels patch for partition and ALTER TABLE
Instead of using 'TABLE_SHARE->is_view' use 'table_flags() & HA_NO_COPY_ON_ALTER
sql/sql_test.cc:
table_name and table_cache_key are now LEX_STRING's
sql/sql_trigger.cc:
TABLE_SHARE->db and TABLE_SHARE->table_name are now LEX_STRING's
safer strxnmov()
Removed compiler warnings
sql/sql_update.cc:
Call field->init(TABLE) after field is created
sql/sql_view.cc:
safer strxnmov()
Create common TABLE_SHARE object for views to allow us to cache if table is a view
sql/structs.h:
Added SHOW_TABLE_DEFINITIONS
sql/table.cc:
Creation and destruct of TABLE_SHARE objects that are common for many TABLE objects
The table opening process now works the following way:
- Create common TABLE_SHARE object
- Read the .frm file and unpack it into the TABLE_SHARE object
- Create a TABLE object based on the information in the TABLE_SHARE
object and open a handler to the table object
open_table_def() is written in such a way that it should be trival to add parsing of the .frm files in new formats
sql/table.h:
TABLE objects for the same database table now share a common TABLE_SHARE object
In TABLE_SHARE the most common strings are now LEX_STRING's
sql/unireg.cc:
Changed arguments to rea_create_table() to have same order as other functions
Call field->init(table) for new created fields
sql/unireg.h:
Added OPEN_VIEW
strings/strxnmov.c:
Change strxnmov() to always add end \0
This makes usage of strxnmov() safer as most of MySQL code assumes that strxnmov() will create a null terminated string
2005-11-23 21:45:02 +01:00
|
|
|
DBUG_PRINT("info", ("Was not expecting field from table %s (%s)",
|
|
|
|
context->table->s->table_name.str,
|
|
|
|
field->table->s->table_name.str));
|
2005-02-16 14:18:32 +01:00
|
|
|
context->supported= FALSE;
|
2005-05-18 14:32:05 +02:00
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::FUNC_ITEM:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_func *func_item= (Item_func *) item;
|
|
|
|
// Check that we expect a function or functional expression here
|
|
|
|
if (context->expecting(Item::FUNC_ITEM) ||
|
|
|
|
func_item->functype() == Item_func::UNKNOWN_FUNC)
|
|
|
|
context->expect_nothing();
|
|
|
|
else
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
// Did not expect function here
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (func_item->functype()) {
|
|
|
|
case Item_func::EQ_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("EQ_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::NE_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("NE_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::LT_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("LT_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::LE_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("LE_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::GE_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("GE_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::GT_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("GT_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::REAL_ITEM);
|
|
|
|
context->expect(Item::DECIMAL_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::LIKE_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("LIKE_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect(Item::FUNC_ITEM);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::ISNULL_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("ISNULL_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::ISNOTNULL_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("ISNOTNULL_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::FIELD_ITEM);
|
|
|
|
context->expect_field_result(STRING_RESULT);
|
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(INT_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::NOT_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("NOT_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(func_item->functype(),
|
|
|
|
func_item);
|
|
|
|
context->expect(Item::FUNC_ITEM);
|
2005-02-23 15:51:26 +01:00
|
|
|
context->expect(Item::COND_ITEM);
|
2005-05-18 14:32:05 +02:00
|
|
|
break;
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::BETWEEN:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("BETWEEN, rewriting using AND"));
|
2005-09-12 09:36:07 +02:00
|
|
|
Item_func_between *between_func= (Item_func_between *) func_item;
|
2005-05-18 14:32:05 +02:00
|
|
|
Ndb_rewrite_context *rewrite_context=
|
|
|
|
new Ndb_rewrite_context(func_item);
|
|
|
|
rewrite_context->next= context->rewrite_stack;
|
|
|
|
context->rewrite_stack= rewrite_context;
|
2005-09-12 09:36:07 +02:00
|
|
|
if (between_func->negated)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("NOT_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(Item_func::NOT_FUNC, 1);
|
|
|
|
prev_cond= curr_cond;
|
|
|
|
curr_cond= context->cond_ptr= new Ndb_cond();
|
|
|
|
curr_cond->prev= prev_cond;
|
|
|
|
prev_cond->next= curr_cond;
|
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("COND_AND_FUNC"));
|
2005-09-12 09:36:07 +02:00
|
|
|
curr_cond->ndb_item=
|
|
|
|
new Ndb_item(Item_func::COND_AND_FUNC,
|
|
|
|
func_item->argument_count() - 1);
|
2005-02-16 14:18:32 +01:00
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
2005-05-18 14:32:05 +02:00
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FUNC_ITEM);
|
|
|
|
break;
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::IN_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("IN_FUNC, rewriting using OR"));
|
2005-09-12 09:36:07 +02:00
|
|
|
Item_func_in *in_func= (Item_func_in *) func_item;
|
2005-05-18 14:32:05 +02:00
|
|
|
Ndb_rewrite_context *rewrite_context=
|
|
|
|
new Ndb_rewrite_context(func_item);
|
|
|
|
rewrite_context->next= context->rewrite_stack;
|
|
|
|
context->rewrite_stack= rewrite_context;
|
2005-09-12 09:36:07 +02:00
|
|
|
if (in_func->negated)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("NOT_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(Item_func::NOT_FUNC, 1);
|
|
|
|
prev_cond= curr_cond;
|
|
|
|
curr_cond= context->cond_ptr= new Ndb_cond();
|
|
|
|
curr_cond->prev= prev_cond;
|
|
|
|
prev_cond->next= curr_cond;
|
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("COND_OR_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(Item_func::COND_OR_FUNC,
|
|
|
|
func_item->argument_count() - 1);
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect(Item::INT_ITEM);
|
|
|
|
context->expect(Item::STRING_ITEM);
|
|
|
|
context->expect(Item::VARBIN_ITEM);
|
|
|
|
context->expect(Item::FUNC_ITEM);
|
|
|
|
break;
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::UNKNOWN_FUNC:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("UNKNOWN_FUNC %s",
|
|
|
|
func_item->const_item()?"const":""));
|
|
|
|
DBUG_PRINT("info", ("result type %d", func_item->result_type()));
|
|
|
|
if (func_item->const_item())
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
|
|
|
switch (func_item->result_type()) {
|
|
|
|
case STRING_RESULT:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::STRING_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(STRING_RESULT);
|
|
|
|
context->expect_collation(func_item->collation.collation);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
// Check that string result have correct collation
|
|
|
|
if (!context->expecting_collation(item->collation.collation))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Found non-matching collation %s",
|
|
|
|
item->collation.collation->name));
|
|
|
|
context->supported= FALSE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Skip any arguments since we will evaluate function instead
|
|
|
|
DBUG_PRINT("info", ("Skip until end of arguments marker"));
|
|
|
|
context->skip= func_item->argument_count();
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case REAL_RESULT:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::REAL_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(REAL_RESULT);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Skip any arguments since we will evaluate function instead
|
|
|
|
DBUG_PRINT("info", ("Skip until end of arguments marker"));
|
|
|
|
context->skip= func_item->argument_count();
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case INT_RESULT:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::INT_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(INT_RESULT);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Skip any arguments since we will evaluate function instead
|
|
|
|
DBUG_PRINT("info", ("Skip until end of arguments marker"));
|
|
|
|
context->skip= func_item->argument_count();
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case DECIMAL_RESULT:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::DECIMAL_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(DECIMAL_RESULT);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
|
|
|
// Skip any arguments since we will evaluate function instead
|
|
|
|
DBUG_PRINT("info", ("Skip until end of arguments marker"));
|
|
|
|
context->skip= func_item->argument_count();
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
else
|
|
|
|
// Function does not return constant expression
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
default:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("Found func_item of type %d",
|
|
|
|
func_item->functype()));
|
|
|
|
context->supported= FALSE;
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
}
|
|
|
|
break;
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::STRING_ITEM:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("STRING_ITEM"));
|
|
|
|
if (context->expecting(Item::STRING_ITEM))
|
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
#ifndef DBUG_OFF
|
2005-05-18 14:32:05 +02:00
|
|
|
char buff[256];
|
|
|
|
String str(buff,(uint32) sizeof(buff), system_charset_info);
|
|
|
|
str.length(0);
|
|
|
|
Item_string *string_item= (Item_string *) item;
|
|
|
|
DBUG_PRINT("info", ("value \"%s\"",
|
|
|
|
string_item->val_str(&str)->ptr()));
|
2005-02-16 14:18:32 +01:00
|
|
|
#endif
|
2005-05-18 14:32:05 +02:00
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::STRING_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(STRING_RESULT);
|
|
|
|
context->expect_collation(item->collation.collation);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
// Check that we are comparing with a field with same collation
|
|
|
|
if (!context->expecting_collation(item->collation.collation))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Found non-matching collation %s",
|
|
|
|
item->collation.collation->name));
|
|
|
|
context->supported= FALSE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::INT_ITEM:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("INT_ITEM"));
|
|
|
|
if (context->expecting(Item::INT_ITEM))
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_int *int_item= (Item_int *) item;
|
|
|
|
DBUG_PRINT("info", ("value %d", int_item->value));
|
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::INT_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(INT_RESULT);
|
2005-06-10 10:34:05 +02:00
|
|
|
context->expect_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
2005-05-18 14:32:05 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
|
|
|
else
|
2005-05-18 14:32:05 +02:00
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::REAL_ITEM:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("REAL_ITEM %s"));
|
|
|
|
if (context->expecting(Item::REAL_ITEM))
|
2005-02-23 15:51:26 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_float *float_item= (Item_float *) item;
|
|
|
|
DBUG_PRINT("info", ("value %f", float_item->value));
|
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::REAL_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(REAL_RESULT);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
else
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::VARBIN_ITEM:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("VARBIN_ITEM"));
|
|
|
|
if (context->expecting(Item::VARBIN_ITEM))
|
2005-02-16 14:18:32 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::VARBIN_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(STRING_RESULT);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
|
|
|
else
|
2005-05-18 14:32:05 +02:00
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::DECIMAL_ITEM:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("DECIMAL_ITEM %s"));
|
|
|
|
if (context->expecting(Item::DECIMAL_ITEM))
|
2005-02-23 15:51:26 +01:00
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_decimal *decimal_item= (Item_decimal *) item;
|
|
|
|
DBUG_PRINT("info", ("value %f", decimal_item->val_real()));
|
|
|
|
NDB_ITEM_QUALIFICATION q;
|
|
|
|
q.value_type= Item::DECIMAL_ITEM;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_VALUE, q, item);
|
|
|
|
if (context->expect_field_result_mask)
|
|
|
|
{
|
|
|
|
// We have not seen the field argument yet
|
|
|
|
context->expect_only(Item::FIELD_ITEM);
|
|
|
|
context->expect_only_field_result(REAL_RESULT);
|
|
|
|
context->expect_field_result(DECIMAL_RESULT);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Expect another logical expression
|
|
|
|
context->expect_only(Item::FUNC_ITEM);
|
|
|
|
context->expect(Item::COND_ITEM);
|
|
|
|
}
|
2005-02-23 15:51:26 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
else
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item::COND_ITEM:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
Item_cond *cond_item= (Item_cond *) item;
|
|
|
|
|
|
|
|
if (context->expecting(Item::COND_ITEM))
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
|
|
|
switch (cond_item->functype()) {
|
|
|
|
case Item_func::COND_AND_FUNC:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("COND_AND_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(cond_item->functype(),
|
|
|
|
cond_item);
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case Item_func::COND_OR_FUNC:
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("COND_OR_FUNC"));
|
|
|
|
curr_cond->ndb_item= new Ndb_item(cond_item->functype(),
|
|
|
|
cond_item);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
DBUG_PRINT("info", ("COND_ITEM %d", cond_item->functype()));
|
|
|
|
context->supported= FALSE;
|
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
else
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
|
|
|
/* Did not expect condition */
|
2005-05-18 14:32:05 +02:00
|
|
|
context->supported= FALSE;
|
2005-07-04 02:42:33 +02:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
break;
|
2005-02-16 14:18:32 +01:00
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
default:
|
|
|
|
{
|
2005-05-18 14:32:05 +02:00
|
|
|
DBUG_PRINT("info", ("Found item of type %d", item->type()));
|
2005-02-16 14:18:32 +01:00
|
|
|
context->supported= FALSE;
|
2005-05-18 14:32:05 +02:00
|
|
|
}
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
if (context->supported && context->rewrite_stack)
|
|
|
|
{
|
|
|
|
Ndb_rewrite_context *rewrite_context= context->rewrite_stack;
|
|
|
|
if (rewrite_context->count ==
|
|
|
|
rewrite_context->func_item->argument_count())
|
|
|
|
{
|
|
|
|
// Rewrite is done, wrap an END() at the en
|
|
|
|
DBUG_PRINT("info", ("End of condition group"));
|
|
|
|
prev_cond= curr_cond;
|
|
|
|
curr_cond= context->cond_ptr= new Ndb_cond();
|
2005-09-12 09:36:07 +02:00
|
|
|
curr_cond->prev= prev_cond;
|
2005-05-18 14:32:05 +02:00
|
|
|
prev_cond->next= curr_cond;
|
|
|
|
curr_cond->ndb_item= new Ndb_item(NDB_END_COND);
|
|
|
|
// Pop rewrite stack
|
|
|
|
context->rewrite_stack= context->rewrite_stack->next;
|
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-05-18 14:32:05 +02:00
|
|
|
|
2004-12-17 21:13:22 +01:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
|
|
|
ha_ndbcluster::serialize_cond(const COND *cond, Ndb_cond_stack *ndb_cond)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("serialize_cond");
|
|
|
|
Item *item= (Item *) cond;
|
2005-02-16 14:18:32 +01:00
|
|
|
Ndb_cond_traverse_context context(table, (void *)m_table, ndb_cond);
|
2005-02-23 15:51:26 +01:00
|
|
|
// Expect a logical expression
|
|
|
|
context.expect(Item::FUNC_ITEM);
|
|
|
|
context.expect(Item::COND_ITEM);
|
2004-12-17 21:13:22 +01:00
|
|
|
item->traverse_cond(&ndb_serialize_cond, (void *) &context, Item::PREFIX);
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("The pushed condition is %ssupported", (context.supported)?"":"not "));
|
2004-12-17 21:13:22 +01:00
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(context.supported);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
int
|
|
|
|
ha_ndbcluster::build_scan_filter_predicate(Ndb_cond * &cond,
|
2005-02-18 21:43:51 +01:00
|
|
|
NdbScanFilter *filter,
|
|
|
|
bool negated)
|
2004-12-17 21:13:22 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("build_scan_filter_predicate");
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (cond->ndb_item->type) {
|
|
|
|
case NDB_FUNCTION:
|
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!cond->next)
|
|
|
|
break;
|
|
|
|
Ndb_item *a= cond->next->ndb_item;
|
2005-02-23 15:51:26 +01:00
|
|
|
Ndb_item *b, *field, *value= NULL;
|
2005-08-12 12:54:42 +02:00
|
|
|
LINT_INIT(field);
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (cond->ndb_item->argument_count()) {
|
|
|
|
case 1:
|
2005-02-23 15:51:26 +01:00
|
|
|
field=
|
|
|
|
(a->type == NDB_FIELD)? a : NULL;
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
case 2:
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!cond->next->next)
|
2005-02-16 14:18:32 +01:00
|
|
|
break;
|
2005-02-23 15:51:26 +01:00
|
|
|
b= cond->next->next->ndb_item;
|
|
|
|
value=
|
2005-02-16 14:18:32 +01:00
|
|
|
(a->type == NDB_VALUE)? a
|
|
|
|
: (b->type == NDB_VALUE)? b
|
|
|
|
: NULL;
|
2005-02-23 15:51:26 +01:00
|
|
|
field=
|
2005-02-16 14:18:32 +01:00
|
|
|
(a->type == NDB_FIELD)? a
|
|
|
|
: (b->type == NDB_FIELD)? b
|
|
|
|
: NULL;
|
2005-02-23 15:51:26 +01:00
|
|
|
break;
|
2005-03-09 16:39:23 +01:00
|
|
|
default:
|
2005-07-18 13:31:02 +02:00
|
|
|
field= NULL; //Keep compiler happy
|
|
|
|
DBUG_ASSERT(0);
|
2005-02-23 15:51:26 +01:00
|
|
|
break;
|
|
|
|
}
|
2005-07-04 02:42:33 +02:00
|
|
|
switch ((negated) ?
|
|
|
|
Ndb_item::negate(cond->ndb_item->qualification.function_type)
|
|
|
|
: cond->ndb_item->qualification.function_type) {
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_EQ_FUNC:
|
2005-02-23 15:51:26 +01:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
2005-01-04 20:56:04 +01:00
|
|
|
DBUG_PRINT("info", ("Generating EQ filter"));
|
2005-01-07 15:33:24 +01:00
|
|
|
if (filter->cmp(NdbScanFilter::COND_EQ,
|
2005-02-16 14:18:32 +01:00
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_NE_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
2005-01-04 20:56:04 +01:00
|
|
|
DBUG_PRINT("info", ("Generating NE filter"));
|
2005-01-07 15:33:24 +01:00
|
|
|
if (filter->cmp(NdbScanFilter::COND_NE,
|
2005-02-16 14:18:32 +01:00
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_LT_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
2004-12-17 21:13:22 +01:00
|
|
|
if (a == field)
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating LT filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_LT,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
else
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating GT filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_GT,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_LE_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
2004-12-17 21:13:22 +01:00
|
|
|
if (a == field)
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating LE filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_LE,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
else
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating GE filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_GE,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_GE_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
2004-12-17 21:13:22 +01:00
|
|
|
if (a == field)
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating GE filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_GE,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
else
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating LE filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_LE,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_GT_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
2004-12-17 21:13:22 +01:00
|
|
|
if (a == field)
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating GT filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_GT,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
else
|
2005-01-04 14:20:18 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_PRINT("info", ("Generating LT filter"));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_LT,
|
|
|
|
field->get_field_no(),
|
|
|
|
field->get_val(),
|
|
|
|
field->pack_length()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-04 14:20:18 +01:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_LIKE_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-24 11:09:14 +01:00
|
|
|
if ((value->qualification.value_type != Item::STRING_ITEM) &&
|
|
|
|
(value->qualification.value_type != Item::VARBIN_ITEM))
|
|
|
|
break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
|
|
|
DBUG_PRINT("info", ("Generating LIKE filter: like(%d,%s,%d)",
|
2005-02-18 21:43:51 +01:00
|
|
|
field->get_field_no(), value->get_val(),
|
|
|
|
value->pack_length()));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_LIKE,
|
|
|
|
field->get_field_no(),
|
2005-02-23 15:51:26 +01:00
|
|
|
value->get_val(),
|
|
|
|
value->pack_length()) == -1)
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(1);
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_NOTLIKE_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2004-12-17 21:13:22 +01:00
|
|
|
if (!value || !field) break;
|
2005-02-24 11:09:14 +01:00
|
|
|
if ((value->qualification.value_type != Item::STRING_ITEM) &&
|
|
|
|
(value->qualification.value_type != Item::VARBIN_ITEM))
|
2005-08-01 11:50:43 +02:00
|
|
|
break;
|
2005-02-11 22:05:24 +01:00
|
|
|
// Save value in right format for the field type
|
|
|
|
value->save_in_field(field);
|
|
|
|
DBUG_PRINT("info", ("Generating NOTLIKE filter: notlike(%d,%s,%d)",
|
2005-02-18 21:43:51 +01:00
|
|
|
field->get_field_no(), value->get_val(),
|
|
|
|
value->pack_length()));
|
|
|
|
if (filter->cmp(NdbScanFilter::COND_NOT_LIKE,
|
|
|
|
field->get_field_no(),
|
2005-02-23 15:51:26 +01:00
|
|
|
value->get_val(),
|
|
|
|
value->pack_length()) == -1)
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(1);
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_ISNULL_FUNC:
|
2005-02-23 15:51:26 +01:00
|
|
|
if (!field)
|
|
|
|
break;
|
|
|
|
DBUG_PRINT("info", ("Generating ISNULL filter"));
|
|
|
|
if (filter->isnull(field->get_field_no()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_ISNOTNULL_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2005-02-23 15:51:26 +01:00
|
|
|
if (!field)
|
|
|
|
break;
|
|
|
|
DBUG_PRINT("info", ("Generating ISNOTNULL filter"));
|
|
|
|
if (filter->isnotnull(field->get_field_no()) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-01-07 15:33:24 +01:00
|
|
|
cond= cond->next->next;
|
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("Found illegal condition"));
|
2005-01-07 15:33:24 +01:00
|
|
|
DBUG_RETURN(1);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
int
|
2005-02-23 15:51:26 +01:00
|
|
|
ha_ndbcluster::build_scan_filter_group(Ndb_cond* &cond, NdbScanFilter *filter)
|
2004-12-17 21:13:22 +01:00
|
|
|
{
|
2005-02-23 15:51:26 +01:00
|
|
|
uint level=0;
|
2005-05-06 10:39:30 +02:00
|
|
|
bool negated= FALSE;
|
2004-12-17 21:13:22 +01:00
|
|
|
DBUG_ENTER("build_scan_filter_group");
|
2005-07-04 02:42:33 +02:00
|
|
|
|
2005-02-23 15:51:26 +01:00
|
|
|
do
|
|
|
|
{
|
2005-07-04 02:42:33 +02:00
|
|
|
if (!cond)
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
switch (cond->ndb_item->type) {
|
|
|
|
case NDB_FUNCTION:
|
|
|
|
{
|
|
|
|
switch (cond->ndb_item->qualification.function_type) {
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_COND_AND_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2005-02-23 15:51:26 +01:00
|
|
|
level++;
|
|
|
|
DBUG_PRINT("info", ("Generating %s group %u", (negated)?"NAND":"AND",
|
|
|
|
level));
|
|
|
|
if ((negated) ? filter->begin(NdbScanFilter::NAND)
|
|
|
|
: filter->begin(NdbScanFilter::AND) == -1)
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(1);
|
2005-05-06 10:39:30 +02:00
|
|
|
negated= FALSE;
|
2005-02-23 15:51:26 +01:00
|
|
|
cond= cond->next;
|
|
|
|
break;
|
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_COND_OR_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2005-02-23 15:51:26 +01:00
|
|
|
level++;
|
|
|
|
DBUG_PRINT("info", ("Generating %s group %u", (negated)?"NOR":"OR",
|
|
|
|
level));
|
|
|
|
if ((negated) ? filter->begin(NdbScanFilter::NOR)
|
|
|
|
: filter->begin(NdbScanFilter::OR) == -1)
|
|
|
|
DBUG_RETURN(1);
|
2005-05-06 10:39:30 +02:00
|
|
|
negated= FALSE;
|
2005-02-23 15:51:26 +01:00
|
|
|
cond= cond->next;
|
|
|
|
break;
|
|
|
|
}
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_NOT_FUNC:
|
2005-07-04 02:42:33 +02:00
|
|
|
{
|
2005-06-01 11:56:47 +02:00
|
|
|
DBUG_PRINT("info", ("Generating negated query"));
|
2005-02-23 15:51:26 +01:00
|
|
|
cond= cond->next;
|
2005-05-06 10:39:30 +02:00
|
|
|
negated= TRUE;
|
2005-02-23 15:51:26 +01:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
if (build_scan_filter_predicate(cond, filter, negated))
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(1);
|
2005-05-06 10:39:30 +02:00
|
|
|
negated= FALSE;
|
2005-02-23 15:51:26 +01:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
break;
|
2005-07-04 02:42:33 +02:00
|
|
|
}
|
|
|
|
case NDB_END_COND:
|
2005-02-23 15:51:26 +01:00
|
|
|
DBUG_PRINT("info", ("End of group %u", level));
|
|
|
|
level--;
|
2005-01-07 15:33:24 +01:00
|
|
|
if (cond) cond= cond->next;
|
|
|
|
if (filter->end() == -1)
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(1);
|
2005-06-01 11:56:47 +02:00
|
|
|
if (!negated)
|
|
|
|
break;
|
|
|
|
// else fall through (NOT END is an illegal condition)
|
2005-07-04 02:42:33 +02:00
|
|
|
default:
|
|
|
|
{
|
2005-02-23 15:51:26 +01:00
|
|
|
DBUG_PRINT("info", ("Illegal scan filter"));
|
2005-02-18 21:43:51 +01:00
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-06-01 11:56:47 +02:00
|
|
|
} while (level > 0 || negated);
|
2004-12-17 21:13:22 +01:00
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
int
|
|
|
|
ha_ndbcluster::build_scan_filter(Ndb_cond * &cond, NdbScanFilter *filter)
|
2004-12-17 21:13:22 +01:00
|
|
|
{
|
|
|
|
bool simple_cond= TRUE;
|
|
|
|
DBUG_ENTER("build_scan_filter");
|
|
|
|
|
2005-07-04 02:42:33 +02:00
|
|
|
switch (cond->ndb_item->type) {
|
|
|
|
case NDB_FUNCTION:
|
|
|
|
switch (cond->ndb_item->qualification.function_type) {
|
2005-08-01 11:50:43 +02:00
|
|
|
case NDB_COND_AND_FUNC:
|
|
|
|
case NDB_COND_OR_FUNC:
|
2005-05-18 14:32:05 +02:00
|
|
|
simple_cond= FALSE;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
if (simple_cond && filter->begin() == -1)
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
if (build_scan_filter_group(cond, filter))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
if (simple_cond && filter->end() == -1)
|
|
|
|
DBUG_RETURN(1);
|
2004-12-17 21:13:22 +01:00
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
int
|
2004-12-17 21:13:22 +01:00
|
|
|
ha_ndbcluster::generate_scan_filter(Ndb_cond_stack *ndb_cond_stack,
|
2005-02-16 14:18:32 +01:00
|
|
|
NdbScanOperation *op)
|
2004-12-17 21:13:22 +01:00
|
|
|
{
|
|
|
|
DBUG_ENTER("generate_scan_filter");
|
|
|
|
if (ndb_cond_stack)
|
|
|
|
{
|
2005-06-01 11:56:47 +02:00
|
|
|
DBUG_PRINT("info", ("Generating scan filter"));
|
2004-12-17 21:13:22 +01:00
|
|
|
NdbScanFilter filter(op);
|
|
|
|
bool multiple_cond= FALSE;
|
|
|
|
// Wrap an AND group around multiple conditions
|
|
|
|
if (ndb_cond_stack->next) {
|
|
|
|
multiple_cond= TRUE;
|
2005-01-07 15:33:24 +01:00
|
|
|
if (filter.begin() == -1)
|
2005-02-16 14:18:32 +01:00
|
|
|
DBUG_RETURN(1);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
for (Ndb_cond_stack *stack= ndb_cond_stack;
|
2005-02-16 14:18:32 +01:00
|
|
|
(stack);
|
|
|
|
stack= stack->next)
|
2004-12-17 21:13:22 +01:00
|
|
|
{
|
2005-02-16 14:18:32 +01:00
|
|
|
Ndb_cond *cond= stack->ndb_cond;
|
2005-01-07 15:33:24 +01:00
|
|
|
|
2005-02-16 14:18:32 +01:00
|
|
|
if (build_scan_filter(cond, &filter))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("build_scan_filter failed"));
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
2005-01-07 15:33:24 +01:00
|
|
|
if (multiple_cond && filter.end() == -1)
|
|
|
|
DBUG_RETURN(1);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Empty stack"));
|
|
|
|
}
|
|
|
|
|
2005-01-07 15:33:24 +01:00
|
|
|
DBUG_RETURN(0);
|
2004-12-17 21:13:22 +01:00
|
|
|
}
|
|
|
|
|
2005-11-06 00:20:37 +01:00
|
|
|
/*
|
|
|
|
Implements the SHOW NDB STATUS command.
|
|
|
|
*/
|
2005-11-07 16:25:06 +01:00
|
|
|
bool
|
|
|
|
ndbcluster_show_status(THD* thd, stat_print_fn *stat_print,
|
|
|
|
enum ha_stat_type stat_type)
|
2005-09-19 15:35:07 +02:00
|
|
|
{
|
2005-11-07 16:25:06 +01:00
|
|
|
char buf[IO_SIZE];
|
2005-09-19 15:35:07 +02:00
|
|
|
DBUG_ENTER("ndbcluster_show_status");
|
|
|
|
|
|
|
|
if (have_ndbcluster != SHOW_OPTION_YES)
|
|
|
|
{
|
2005-11-07 16:25:06 +01:00
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
if (stat_type != HA_ENGINE_STATUS)
|
|
|
|
{
|
|
|
|
DBUG_RETURN(FALSE);
|
2005-09-19 15:35:07 +02:00
|
|
|
}
|
|
|
|
|
2005-09-20 12:07:18 +02:00
|
|
|
if (get_thd_ndb(thd) && get_thd_ndb(thd)->ndb)
|
2005-09-19 15:35:07 +02:00
|
|
|
{
|
2005-09-20 12:07:18 +02:00
|
|
|
Ndb* ndb= (get_thd_ndb(thd))->ndb;
|
2006-01-05 23:47:49 +01:00
|
|
|
Ndb::Free_list_usage tmp;
|
|
|
|
tmp.m_name= 0;
|
2005-09-19 15:35:07 +02:00
|
|
|
while (ndb->get_free_list_usage(&tmp))
|
|
|
|
{
|
2005-12-21 19:18:40 +01:00
|
|
|
uint buflen=
|
|
|
|
my_snprintf(buf, sizeof(buf),
|
2005-11-07 16:25:06 +01:00
|
|
|
"created=%u, free=%u, sizeof=%u",
|
|
|
|
tmp.m_created, tmp.m_free, tmp.m_sizeof);
|
2005-12-21 19:18:40 +01:00
|
|
|
if (stat_print(thd, ndbcluster_hton.name, strlen(ndbcluster_hton.name),
|
|
|
|
tmp.m_name, strlen(tmp.m_name), buf, buflen))
|
2005-11-07 16:25:06 +01:00
|
|
|
DBUG_RETURN(TRUE);
|
2005-09-19 15:35:07 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
send_eof(thd);
|
|
|
|
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
2005-07-18 13:31:02 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
Create a table in NDB Cluster
|
|
|
|
*/
|
|
|
|
static uint get_no_fragments(ulonglong max_rows)
|
|
|
|
{
|
|
|
|
#if MYSQL_VERSION_ID >= 50000
|
|
|
|
uint acc_row_size= 25 + /*safety margin*/ 2;
|
|
|
|
#else
|
|
|
|
uint acc_row_size= pk_length*4;
|
|
|
|
/* add acc overhead */
|
|
|
|
if (pk_length <= 8) /* main page will set the limit */
|
|
|
|
acc_row_size+= 25 + /*safety margin*/ 2;
|
|
|
|
else /* overflow page will set the limit */
|
|
|
|
acc_row_size+= 4 + /*safety margin*/ 4;
|
|
|
|
#endif
|
|
|
|
ulonglong acc_fragment_size= 512*1024*1024;
|
|
|
|
#if MYSQL_VERSION_ID >= 50100
|
|
|
|
return (max_rows*acc_row_size)/acc_fragment_size+1;
|
|
|
|
#else
|
|
|
|
return ((max_rows*acc_row_size)/acc_fragment_size+1
|
|
|
|
+1/*correct rounding*/)/2;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Routine to adjust default number of partitions to always be a multiple
|
|
|
|
of number of nodes and never more than 4 times the number of nodes.
|
|
|
|
|
|
|
|
*/
|
|
|
|
static bool adjusted_frag_count(uint no_fragments, uint no_nodes,
|
|
|
|
uint &reported_frags)
|
|
|
|
{
|
|
|
|
uint i= 0;
|
|
|
|
reported_frags= no_nodes;
|
|
|
|
while (reported_frags < no_fragments && ++i < 4 &&
|
|
|
|
(reported_frags + no_nodes) < MAX_PARTITIONS)
|
|
|
|
reported_frags+= no_nodes;
|
|
|
|
return (reported_frags < no_fragments);
|
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::get_default_no_partitions(ulonglong max_rows)
|
|
|
|
{
|
|
|
|
uint reported_frags;
|
|
|
|
uint no_fragments= get_no_fragments(max_rows);
|
|
|
|
uint no_nodes= g_ndb_cluster_connection->no_db_nodes();
|
|
|
|
adjusted_frag_count(no_fragments, no_nodes, reported_frags);
|
|
|
|
return (int)reported_frags;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
User defined partitioning set-up. We need to check how many fragments the
|
|
|
|
user wants defined and which node groups to put those into. Later we also
|
|
|
|
want to attach those partitions to a tablespace.
|
|
|
|
|
|
|
|
All the functionality of the partition function, partition limits and so
|
|
|
|
forth are entirely handled by the MySQL Server. There is one exception to
|
|
|
|
this rule for PARTITION BY KEY where NDB handles the hash function and
|
|
|
|
this type can thus be handled transparently also by NDB API program.
|
|
|
|
For RANGE, HASH and LIST and subpartitioning the NDB API programs must
|
|
|
|
implement the function to map to a partition.
|
|
|
|
*/
|
|
|
|
|
|
|
|
uint ha_ndbcluster::set_up_partition_info(partition_info *part_info,
|
|
|
|
TABLE *table,
|
|
|
|
void *tab_par)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::set_up_partition_info");
|
|
|
|
ushort node_group[MAX_PARTITIONS];
|
|
|
|
ulong ng_index= 0, i, j;
|
|
|
|
NDBTAB *tab= (NDBTAB*)tab_par;
|
|
|
|
NDBTAB::FragmentType ftype= NDBTAB::UserDefined;
|
|
|
|
partition_element *part_elem;
|
|
|
|
|
|
|
|
if (part_info->part_type == HASH_PARTITION &&
|
|
|
|
part_info->list_of_part_fields == TRUE)
|
|
|
|
{
|
|
|
|
Field **fields= part_info->part_field_array;
|
|
|
|
|
|
|
|
if (part_info->linear_hash_ind)
|
|
|
|
ftype= NDBTAB::DistrKeyLin;
|
|
|
|
else
|
|
|
|
ftype= NDBTAB::DistrKeyHash;
|
|
|
|
|
|
|
|
for (i= 0; i < part_info->part_field_list.elements; i++)
|
|
|
|
{
|
|
|
|
NDBCOL *col= tab->getColumn(fields[i]->fieldnr - 1);
|
|
|
|
DBUG_PRINT("info",("setting dist key on %s", col->getName()));
|
|
|
|
col->setPartitionKey(TRUE);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
List_iterator<partition_element> part_it(part_info->partitions);
|
|
|
|
for (i= 0; i < part_info->no_parts; i++)
|
|
|
|
{
|
|
|
|
part_elem= part_it++;
|
|
|
|
if (!is_sub_partitioned(part_info))
|
|
|
|
{
|
|
|
|
node_group[ng_index++]= part_elem->nodegroup_id;
|
|
|
|
//Here we should insert tablespace id based on tablespace name
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
List_iterator<partition_element> sub_it(part_elem->subpartitions);
|
|
|
|
for (j= 0; j < part_info->no_subparts; j++)
|
|
|
|
{
|
|
|
|
part_elem= sub_it++;
|
|
|
|
node_group[ng_index++]= part_elem->nodegroup_id;
|
|
|
|
//Here we should insert tablespace id based on tablespace name
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
{
|
|
|
|
uint no_nodes= g_ndb_cluster_connection->no_db_nodes();
|
|
|
|
if (ng_index > 4 * no_nodes)
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1300);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
tab->setNodeGroupIds(&node_group, ng_index);
|
|
|
|
tab->setFragmentType(ftype);
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
This routine is used to set-up fragmentation when the user has only specified
|
|
|
|
ENGINE = NDB and no user defined partitioning what so ever. Thus all values
|
|
|
|
will be based on default values. We will choose Linear Hash or Hash with
|
|
|
|
perfect spread dependent on a session variable defined in MySQL.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static void ndb_set_fragmentation(NDBTAB &tab, TABLE *form, uint pk_length)
|
|
|
|
{
|
2005-08-22 11:41:37 +02:00
|
|
|
NDBTAB::FragmentType ftype= NDBTAB::DistrKeyHash;
|
2005-07-18 13:31:02 +02:00
|
|
|
ushort node_group[MAX_PARTITIONS];
|
|
|
|
uint no_nodes= g_ndb_cluster_connection->no_db_nodes(), no_fragments, i;
|
|
|
|
DBUG_ENTER("ndb_set_fragmentation");
|
|
|
|
|
|
|
|
if (form->s->max_rows == (ha_rows) 0)
|
|
|
|
{
|
|
|
|
no_fragments= no_nodes;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Ensure that we get enough fragments to handle all rows and ensure that
|
|
|
|
the table is fully distributed by keeping the number of fragments a
|
|
|
|
multiple of the number of nodes.
|
|
|
|
*/
|
|
|
|
uint fragments= get_no_fragments(form->s->max_rows);
|
|
|
|
if (adjusted_frag_count(fragments, no_nodes, no_fragments))
|
|
|
|
{
|
|
|
|
push_warning(current_thd,
|
|
|
|
MYSQL_ERROR::WARN_LEVEL_WARN, ER_UNKNOWN_ERROR,
|
|
|
|
"Ndb might have problems storing the max amount of rows specified");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
Always start with node group 0 and continue with next node group from
|
|
|
|
there
|
|
|
|
*/
|
|
|
|
node_group[0]= 0;
|
|
|
|
for (i= 1; i < no_fragments; i++)
|
|
|
|
node_group[i]= UNDEF_NODEGROUP;
|
2005-08-22 11:41:37 +02:00
|
|
|
switch (opt_ndb_distribution_id)
|
|
|
|
{
|
|
|
|
case ND_KEYHASH:
|
2005-08-22 13:36:55 +02:00
|
|
|
ftype= NDBTAB::DistrKeyHash;
|
2005-08-22 11:41:37 +02:00
|
|
|
break;
|
|
|
|
case ND_LINHASH:
|
2005-07-18 13:31:02 +02:00
|
|
|
ftype= NDBTAB::DistrKeyLin;
|
2005-08-22 11:41:37 +02:00
|
|
|
break;
|
|
|
|
}
|
2005-07-18 13:31:02 +02:00
|
|
|
tab.setFragmentType(ftype);
|
|
|
|
tab.setNodeGroupIds(&node_group, no_fragments);
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
2005-07-22 23:15:53 +02:00
|
|
|
|
2005-07-22 22:43:59 +02:00
|
|
|
bool ha_ndbcluster::check_if_incompatible_data(HA_CREATE_INFO *info,
|
|
|
|
uint table_changes)
|
|
|
|
{
|
2005-07-23 10:50:31 +02:00
|
|
|
/*
|
|
|
|
TODO: Remove the dummy return below, when cluster gets
|
|
|
|
signal from alter table when only .frm is changed. Cluster
|
|
|
|
needs it to manage the copies.
|
|
|
|
*/
|
2005-07-22 22:43:59 +02:00
|
|
|
return COMPATIBLE_DATA_NO;
|
|
|
|
|
|
|
|
if (table_changes != IS_EQUAL_YES)
|
|
|
|
return COMPATIBLE_DATA_NO;
|
|
|
|
|
|
|
|
/* Check that auto_increment value was not changed */
|
|
|
|
if ((info->used_fields & HA_CREATE_USED_AUTO) &&
|
|
|
|
info->auto_increment_value != 0)
|
|
|
|
return COMPATIBLE_DATA_NO;
|
|
|
|
|
|
|
|
/* Check that row format didn't change */
|
|
|
|
if ((info->used_fields & HA_CREATE_USED_AUTO) &&
|
|
|
|
get_row_type() != info->row_type)
|
|
|
|
return COMPATIBLE_DATA_NO;
|
|
|
|
|
|
|
|
return COMPATIBLE_DATA_YES;
|
|
|
|
}
|
|
|
|
|
2005-11-07 12:19:28 +01:00
|
|
|
#ifdef NDB_DISKDATA
|
|
|
|
bool set_up_tablespace(st_alter_tablespace *info,
|
|
|
|
NdbDictionary::Tablespace *ndb_ts)
|
|
|
|
{
|
|
|
|
ndb_ts->setName(info->tablespace_name);
|
|
|
|
ndb_ts->setExtentSize(info->extent_size);
|
|
|
|
ndb_ts->setDefaultLogfileGroup(info->logfile_group_name);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool set_up_datafile(st_alter_tablespace *info,
|
|
|
|
NdbDictionary::Datafile *ndb_df)
|
|
|
|
{
|
|
|
|
if (info->max_size > 0)
|
|
|
|
{
|
|
|
|
my_error(ER_TABLESPACE_AUTO_EXTEND_ERROR, MYF(0));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
ndb_df->setPath(info->data_file_name);
|
|
|
|
ndb_df->setSize(info->initial_size);
|
|
|
|
ndb_df->setTablespace(info->tablespace_name);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool set_up_logfile_group(st_alter_tablespace *info,
|
|
|
|
NdbDictionary::LogfileGroup *ndb_lg)
|
|
|
|
{
|
|
|
|
ndb_lg->setName(info->logfile_group_name);
|
|
|
|
ndb_lg->setUndoBufferSize(info->undo_buffer_size);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool set_up_undofile(st_alter_tablespace *info,
|
|
|
|
NdbDictionary::Undofile *ndb_uf)
|
|
|
|
{
|
|
|
|
ndb_uf->setPath(info->undo_file_name);
|
|
|
|
ndb_uf->setSize(info->initial_size);
|
|
|
|
ndb_uf->setLogfileGroup(info->logfile_group_name);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::alter_tablespace(st_alter_tablespace *info)
|
|
|
|
{
|
|
|
|
Ndb *ndb;
|
|
|
|
NDBDICT *dict;
|
|
|
|
int error;
|
|
|
|
DBUG_ENTER("ha_ndbcluster::alter_tablespace");
|
|
|
|
if (check_ndb_connection())
|
|
|
|
{
|
|
|
|
DBUG_RETURN(my_errno= HA_ERR_NO_CONNECTION);
|
|
|
|
}
|
|
|
|
ndb= get_ndb();
|
|
|
|
dict= ndb->getDictionary();
|
|
|
|
switch (info->ts_cmd_type){
|
|
|
|
case (CREATE_TABLESPACE):
|
|
|
|
{
|
|
|
|
NdbDictionary::Tablespace ndb_ts;
|
|
|
|
NdbDictionary::Datafile ndb_df;
|
|
|
|
if (set_up_tablespace(info, &ndb_ts))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
if (set_up_datafile(info, &ndb_df))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
if (error= dict->createTablespace(ndb_ts))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createTablespace returned %d", error));
|
|
|
|
my_error(ER_CREATE_TABLESPACE_FAILED, MYF(0), "TABLESPACE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("Successfully created Tablespace"));
|
|
|
|
if (error= dict->createDatafile(ndb_df))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createDatafile returned %d", error));
|
|
|
|
my_error(ER_CREATE_TABLESPACE_FAILED, MYF(0), "DATAFILE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case (ALTER_TABLESPACE):
|
|
|
|
{
|
|
|
|
if (info->ts_alter_tablespace_type == ALTER_TABLESPACE_ADD_FILE)
|
|
|
|
{
|
|
|
|
NdbDictionary::Datafile ndb_df;
|
|
|
|
if (set_up_datafile(info, &ndb_df))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
if (error= dict->createDatafile(ndb_df))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createDatafile returned %d", error));
|
|
|
|
my_error(ER_ALTER_TABLESPACE_FAILED, MYF(0), "CREATE DATAFILE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if(info->ts_alter_tablespace_type == ALTER_TABLESPACE_DROP_FILE)
|
|
|
|
{
|
|
|
|
NdbDictionary::Datafile df = dict->getDatafile(0,
|
|
|
|
info->data_file_name);
|
|
|
|
if (strcmp(df.getPath(), info->data_file_name) == 0)
|
|
|
|
{
|
|
|
|
if (error= dict->dropDatafile(df))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createDatafile returned %d", error));
|
|
|
|
my_error(ER_ALTER_TABLESPACE_FAILED, MYF(0), " DROP DATAFILE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("No such datafile"));
|
|
|
|
my_error(ER_ALTER_TABLESPACE_FAILED, MYF(0), " NO SUCH FILE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Unsupported alter tablespace: %d",
|
|
|
|
info->ts_alter_tablespace_type));
|
|
|
|
DBUG_RETURN(HA_ADMIN_NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case (CREATE_LOGFILE_GROUP):
|
|
|
|
{
|
|
|
|
NdbDictionary::LogfileGroup ndb_lg;
|
|
|
|
NdbDictionary::Undofile ndb_uf;
|
|
|
|
if (info->undo_file_name == NULL)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
REDO files in LOGFILE GROUP not supported yet
|
|
|
|
*/
|
|
|
|
DBUG_RETURN(HA_ADMIN_NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
if (set_up_logfile_group(info, &ndb_lg))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
if (error= dict->createLogfileGroup(ndb_lg))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createLogfileGroup returned %d", error));
|
|
|
|
my_error(ER_CREATE_TABLESPACE_FAILED, MYF(0), "LOGFILE GROUP");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("Successfully created Logfile Group"));
|
|
|
|
if (set_up_undofile(info, &ndb_uf))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
if (error= dict->createUndofile(ndb_uf))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createUndofile returned %d", error));
|
|
|
|
my_error(ER_CREATE_TABLESPACE_FAILED, MYF(0), "UNDOFILE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case (ALTER_LOGFILE_GROUP):
|
|
|
|
{
|
|
|
|
if (info->undo_file_name == NULL)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
REDO files in LOGFILE GROUP not supported yet
|
|
|
|
*/
|
|
|
|
DBUG_RETURN(HA_ADMIN_NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
NdbDictionary::Undofile ndb_uf;
|
|
|
|
if (set_up_undofile(info, &ndb_uf))
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
if (error= dict->createUndofile(ndb_uf))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("createUndofile returned %d", error));
|
|
|
|
my_error(ER_ALTER_TABLESPACE_FAILED, MYF(0), "CREATE UNDOFILE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case (DROP_TABLESPACE):
|
|
|
|
{
|
|
|
|
if (error= dict->dropTablespace(
|
|
|
|
dict->getTablespace(info->tablespace_name)))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("dropTablespace returned %d", error));
|
|
|
|
my_error(ER_DROP_TABLESPACE_FAILED, MYF(0), "TABLESPACE");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case (DROP_LOGFILE_GROUP):
|
|
|
|
{
|
|
|
|
if (error= dict->dropLogfileGroup(dict->getLogfileGroup(info->logfile_group_name)))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("dropLogfileGroup returned %d", error));
|
|
|
|
my_error(ER_DROP_TABLESPACE_FAILED, MYF(0), "LOGFILE GROUP");
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case (CHANGE_FILE_TABLESPACE):
|
|
|
|
{
|
|
|
|
DBUG_RETURN(HA_ADMIN_NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
case (ALTER_ACCESS_MODE_TABLESPACE):
|
|
|
|
{
|
|
|
|
DBUG_RETURN(HA_ADMIN_NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
{
|
|
|
|
DBUG_RETURN(HA_ADMIN_NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* NDB_DISKDATA */
|