2004-04-15 09:14:14 +02:00
|
|
|
/* Copyright (C) 2000-2003 MySQL AB
|
|
|
|
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
|
|
it under the terms of the GNU General Public License as published by
|
|
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
|
|
(at your option) any later version.
|
|
|
|
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
GNU General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
along with this program; if not, write to the Free Software
|
|
|
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
This file defines the NDB Cluster handler: the interface between MySQL and
|
|
|
|
NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifdef __GNUC__
|
|
|
|
#pragma implementation // gcc: Class implementation
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#include "mysql_priv.h"
|
|
|
|
|
|
|
|
#ifdef HAVE_NDBCLUSTER_DB
|
|
|
|
#include <my_dir.h>
|
|
|
|
#include "ha_ndbcluster.h"
|
|
|
|
#include <ndbapi/NdbApi.hpp>
|
|
|
|
#include <ndbapi/NdbScanFilter.hpp>
|
|
|
|
|
2004-12-20 12:36:14 +01:00
|
|
|
// options from from mysqld.cc
|
|
|
|
extern my_bool opt_ndb_optimized_node_selection;
|
|
|
|
extern const char *opt_ndbcluster_connectstring;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Default value for parallelism
|
|
|
|
static const int parallelism= 240;
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
// Default value for max number of transactions
|
|
|
|
// createable against NDB from this handler
|
2004-07-23 15:46:56 +02:00
|
|
|
static const int max_transactions= 256;
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
static const char *ha_ndb_ext=".ndb";
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
#define NDB_HIDDEN_PRIMARY_KEY_LENGTH 8
|
2004-05-13 11:56:45 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
#define ERR_PRINT(err) \
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_PRINT("error", ("%d message: %s", err.code, err.message))
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
#define ERR_RETURN(err) \
|
|
|
|
{ \
|
|
|
|
ERR_PRINT(err); \
|
|
|
|
DBUG_RETURN(ndb_to_mysql_error(&err)); \
|
|
|
|
}
|
|
|
|
|
|
|
|
// Typedefs for long names
|
|
|
|
typedef NdbDictionary::Column NDBCOL;
|
2004-09-07 21:53:59 +02:00
|
|
|
typedef NdbDictionary::Table NDBTAB;
|
2004-04-15 09:14:14 +02:00
|
|
|
typedef NdbDictionary::Index NDBINDEX;
|
|
|
|
typedef NdbDictionary::Dictionary NDBDICT;
|
|
|
|
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool ndbcluster_inited= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-11 13:59:22 +02:00
|
|
|
static Ndb* g_ndb= NULL;
|
2004-09-06 20:30:57 +02:00
|
|
|
static Ndb_cluster_connection* g_ndb_cluster_connection= NULL;
|
2004-05-11 13:59:22 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Handler synchronization
|
|
|
|
pthread_mutex_t ndbcluster_mutex;
|
|
|
|
|
|
|
|
// Table lock handling
|
|
|
|
static HASH ndbcluster_open_tables;
|
|
|
|
|
|
|
|
static byte *ndbcluster_get_key(NDB_SHARE *share,uint *length,
|
|
|
|
my_bool not_used __attribute__((unused)));
|
|
|
|
static NDB_SHARE *get_share(const char *table_name);
|
|
|
|
static void free_share(NDB_SHARE *share);
|
|
|
|
|
|
|
|
static int packfrm(const void *data, uint len, const void **pack_data, uint *pack_len);
|
|
|
|
static int unpackfrm(const void **data, uint *len,
|
|
|
|
const void* pack_data);
|
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
static int ndb_get_table_statistics(Ndb*, const char *,
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Uint64* rows, Uint64* commits);
|
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
|
2004-10-20 19:22:58 +02:00
|
|
|
/*
|
|
|
|
Dummy buffer to read zero pack_length fields
|
|
|
|
which are mapped to 1 char
|
|
|
|
*/
|
|
|
|
static byte dummy_buf[1];
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Error handling functions
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct err_code_mapping
|
|
|
|
{
|
|
|
|
int ndb_err;
|
|
|
|
int my_err;
|
2004-11-17 09:15:53 +01:00
|
|
|
int show_warning;
|
2004-04-15 09:14:14 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
static const err_code_mapping err_map[]=
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
{ 626, HA_ERR_KEY_NOT_FOUND, 0 },
|
|
|
|
{ 630, HA_ERR_FOUND_DUPP_KEY, 0 },
|
2004-11-19 10:27:16 +01:00
|
|
|
{ 893, HA_ERR_FOUND_DUPP_KEY, 0 },
|
2004-11-17 09:15:53 +01:00
|
|
|
{ 721, HA_ERR_TABLE_EXIST, 1 },
|
|
|
|
{ 4244, HA_ERR_TABLE_EXIST, 1 },
|
|
|
|
|
|
|
|
{ 709, HA_ERR_NO_SUCH_TABLE, 1 },
|
|
|
|
{ 284, HA_ERR_NO_SUCH_TABLE, 1 },
|
|
|
|
|
|
|
|
{ 266, HA_ERR_LOCK_WAIT_TIMEOUT, 1 },
|
|
|
|
{ 274, HA_ERR_LOCK_WAIT_TIMEOUT, 1 },
|
|
|
|
{ 296, HA_ERR_LOCK_WAIT_TIMEOUT, 1 },
|
|
|
|
{ 297, HA_ERR_LOCK_WAIT_TIMEOUT, 1 },
|
|
|
|
{ 237, HA_ERR_LOCK_WAIT_TIMEOUT, 1 },
|
|
|
|
|
|
|
|
{ 623, HA_ERR_RECORD_FILE_FULL, 1 },
|
|
|
|
{ 624, HA_ERR_RECORD_FILE_FULL, 1 },
|
|
|
|
{ 625, HA_ERR_RECORD_FILE_FULL, 1 },
|
|
|
|
{ 826, HA_ERR_RECORD_FILE_FULL, 1 },
|
|
|
|
{ 827, HA_ERR_RECORD_FILE_FULL, 1 },
|
|
|
|
{ 832, HA_ERR_RECORD_FILE_FULL, 1 },
|
|
|
|
|
|
|
|
{ 0, 1, 0 },
|
|
|
|
|
|
|
|
{ -1, -1, 1 }
|
2004-04-15 09:14:14 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
static int ndb_to_mysql_error(const NdbError *err)
|
|
|
|
{
|
|
|
|
uint i;
|
2004-11-17 09:15:53 +01:00
|
|
|
for (i=0; err_map[i].ndb_err != err->code && err_map[i].my_err != -1; i++);
|
|
|
|
if (err_map[i].show_warning)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
// Push the NDB error message as warning
|
|
|
|
push_warning_printf(current_thd, MYSQL_ERROR::WARN_LEVEL_ERROR,
|
|
|
|
ER_GET_ERRMSG, ER(ER_GET_ERRMSG),
|
|
|
|
err->code, err->message, "NDB");
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-11-17 09:15:53 +01:00
|
|
|
if (err_map[i].my_err == -1)
|
|
|
|
return err->code;
|
2004-04-15 09:14:14 +02:00
|
|
|
return err_map[i].my_err;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-17 16:58:08 +02:00
|
|
|
|
|
|
|
inline
|
|
|
|
int execute_no_commit(ha_ndbcluster *h, NdbConnection *trans)
|
|
|
|
{
|
|
|
|
int m_batch_execute= 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#ifdef NOT_USED
|
|
|
|
if (m_batch_execute)
|
2004-09-17 16:58:08 +02:00
|
|
|
return 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#endif
|
2004-11-17 09:15:53 +01:00
|
|
|
return trans->execute(NoCommit,AbortOnError,h->m_force_send);
|
2004-09-28 21:11:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
inline
|
|
|
|
int execute_commit(ha_ndbcluster *h, NdbConnection *trans)
|
|
|
|
{
|
|
|
|
int m_batch_execute= 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#ifdef NOT_USED
|
|
|
|
if (m_batch_execute)
|
2004-09-28 21:11:50 +02:00
|
|
|
return 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#endif
|
2004-11-17 09:15:53 +01:00
|
|
|
return trans->execute(Commit,AbortOnError,h->m_force_send);
|
|
|
|
}
|
|
|
|
|
|
|
|
inline
|
|
|
|
int execute_commit(THD *thd, NdbConnection *trans)
|
|
|
|
{
|
|
|
|
int m_batch_execute= 0;
|
|
|
|
#ifdef NOT_USED
|
|
|
|
if (m_batch_execute)
|
|
|
|
return 0;
|
|
|
|
#endif
|
|
|
|
return trans->execute(Commit,AbortOnError,thd->variables.ndb_force_send);
|
2004-09-28 21:11:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
inline
|
|
|
|
int execute_no_commit_ie(ha_ndbcluster *h, NdbConnection *trans)
|
|
|
|
{
|
|
|
|
int m_batch_execute= 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#ifdef NOT_USED
|
|
|
|
if (m_batch_execute)
|
2004-09-28 21:11:50 +02:00
|
|
|
return 0;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
#endif
|
2004-12-17 10:36:14 +01:00
|
|
|
return trans->execute(NoCommit, AO_IgnoreError,h->m_force_send);
|
2004-09-17 16:58:08 +02:00
|
|
|
}
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
/*
|
|
|
|
Place holder for ha_ndbcluster thread specific data
|
|
|
|
*/
|
|
|
|
|
|
|
|
Thd_ndb::Thd_ndb()
|
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
ndb= new Ndb(g_ndb_cluster_connection, "");
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
lock_count= 0;
|
|
|
|
count= 0;
|
2004-09-14 18:17:01 +02:00
|
|
|
error= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
Thd_ndb::~Thd_ndb()
|
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
if (ndb)
|
|
|
|
delete ndb;
|
2005-01-01 22:47:50 +01:00
|
|
|
ndb= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
inline
|
|
|
|
Ndb *ha_ndbcluster::get_ndb()
|
|
|
|
{
|
|
|
|
return ((Thd_ndb*)current_thd->transaction.thd_ndb)->ndb;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* manage uncommitted insert/deletes during transactio to get records correct
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct Ndb_table_local_info {
|
|
|
|
int no_uncommitted_rows_count;
|
2004-09-14 14:47:34 +02:00
|
|
|
ulong last_count;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
ha_rows records;
|
|
|
|
};
|
|
|
|
|
2004-10-04 01:20:05 +02:00
|
|
|
void ha_ndbcluster::set_rec_per_key()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::get_status_const");
|
|
|
|
for (uint i=0 ; i < table->keys ; i++)
|
|
|
|
{
|
|
|
|
table->key_info[i].rec_per_key[table->key_info[i].key_parts-1]= 1;
|
|
|
|
}
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
void ha_ndbcluster::records_update()
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::records_update");
|
|
|
|
struct Ndb_table_local_info *info= (struct Ndb_table_local_info *)m_table_info;
|
|
|
|
DBUG_PRINT("info", ("id=%d, no_uncommitted_rows_count=%d",
|
|
|
|
((const NDBTAB *)m_table)->getTableId(),
|
|
|
|
info->no_uncommitted_rows_count));
|
2004-09-20 17:52:19 +02:00
|
|
|
// if (info->records == ~(ha_rows)0)
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Uint64 rows;
|
2004-12-30 19:56:09 +01:00
|
|
|
if(ndb_get_table_statistics(ndb, m_tabname, &rows, 0) == 0){
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
info->records= rows;
|
|
|
|
}
|
|
|
|
}
|
2004-09-14 18:17:01 +02:00
|
|
|
{
|
|
|
|
THD *thd= current_thd;
|
|
|
|
if (((Thd_ndb*)(thd->transaction.thd_ndb))->error)
|
|
|
|
info->no_uncommitted_rows_count= 0;
|
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
records= info->records+ info->no_uncommitted_rows_count;
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2004-09-14 17:51:57 +02:00
|
|
|
void ha_ndbcluster::no_uncommitted_rows_execute_failure()
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
2004-09-14 17:51:57 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_execute_failure");
|
2004-09-14 18:17:01 +02:00
|
|
|
THD *thd= current_thd;
|
|
|
|
((Thd_ndb*)(thd->transaction.thd_ndb))->error= 1;
|
2004-09-14 17:51:57 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
void ha_ndbcluster::no_uncommitted_rows_init(THD *thd)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_init");
|
|
|
|
struct Ndb_table_local_info *info= (struct Ndb_table_local_info *)m_table_info;
|
|
|
|
Thd_ndb *thd_ndb= (Thd_ndb *)thd->transaction.thd_ndb;
|
2004-09-14 14:47:34 +02:00
|
|
|
if (info->last_count != thd_ndb->count)
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
info->last_count = thd_ndb->count;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
info->no_uncommitted_rows_count= 0;
|
|
|
|
info->records= ~(ha_rows)0;
|
|
|
|
DBUG_PRINT("info", ("id=%d, no_uncommitted_rows_count=%d",
|
|
|
|
((const NDBTAB *)m_table)->getTableId(),
|
|
|
|
info->no_uncommitted_rows_count));
|
|
|
|
}
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ha_ndbcluster::no_uncommitted_rows_update(int c)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_update");
|
2004-09-17 16:58:08 +02:00
|
|
|
struct Ndb_table_local_info *info=
|
|
|
|
(struct Ndb_table_local_info *)m_table_info;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
info->no_uncommitted_rows_count+= c;
|
|
|
|
DBUG_PRINT("info", ("id=%d, no_uncommitted_rows_count=%d",
|
|
|
|
((const NDBTAB *)m_table)->getTableId(),
|
|
|
|
info->no_uncommitted_rows_count));
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ha_ndbcluster::no_uncommitted_rows_reset(THD *thd)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::no_uncommitted_rows_reset");
|
|
|
|
((Thd_ndb*)(thd->transaction.thd_ndb))->count++;
|
2004-09-14 18:17:01 +02:00
|
|
|
((Thd_ndb*)(thd->transaction.thd_ndb))->error= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Take care of the error that occured in NDB
|
|
|
|
|
|
|
|
RETURN
|
|
|
|
0 No error
|
|
|
|
# The mapped error code
|
|
|
|
*/
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::ndb_err(NdbConnection *trans)
|
|
|
|
{
|
2004-08-31 12:07:48 +02:00
|
|
|
int res;
|
2004-04-15 09:14:14 +02:00
|
|
|
const NdbError err= trans->getNdbError();
|
|
|
|
DBUG_ENTER("ndb_err");
|
|
|
|
|
|
|
|
ERR_PRINT(err);
|
|
|
|
switch (err.classification) {
|
|
|
|
case NdbError::SchemaError:
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("info", ("invalidateTable %s", m_tabname));
|
|
|
|
dict->invalidateTable(m_tabname);
|
2004-09-13 14:46:38 +02:00
|
|
|
table->version=0L; /* Free when thread is ready */
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-08-31 12:07:48 +02:00
|
|
|
res= ndb_to_mysql_error(&err);
|
|
|
|
DBUG_PRINT("info", ("transformed ndbcluster error %d to mysql error %d",
|
|
|
|
err.code, res));
|
|
|
|
if (res == HA_ERR_FOUND_DUPP_KEY)
|
2004-11-03 15:53:26 +01:00
|
|
|
m_dupkey= table->primary_key;
|
2004-08-31 12:07:48 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-05-11 13:59:22 +02:00
|
|
|
/*
|
2004-05-13 11:56:45 +02:00
|
|
|
Override the default get_error_message in order to add the
|
2004-05-11 13:59:22 +02:00
|
|
|
error message of NDB
|
|
|
|
*/
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
bool ha_ndbcluster::get_error_message(int error,
|
|
|
|
String *buf)
|
2004-05-11 13:59:22 +02:00
|
|
|
{
|
2004-05-13 11:56:45 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::get_error_message");
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("enter", ("error: %d", error));
|
2004-09-20 12:40:53 +02:00
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2004-09-15 14:44:21 +02:00
|
|
|
if (!ndb)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(FALSE);
|
2004-05-13 11:56:45 +02:00
|
|
|
|
2004-09-15 14:44:21 +02:00
|
|
|
const NdbError err= ndb->getNdbError(error);
|
2004-05-24 12:35:39 +02:00
|
|
|
bool temporary= err.status==NdbError::TemporaryError;
|
|
|
|
buf->set(err.message, strlen(err.message), &my_charset_bin);
|
|
|
|
DBUG_PRINT("exit", ("message: %s, temporary: %d", buf->ptr(), temporary));
|
|
|
|
DBUG_RETURN(temporary);
|
2004-05-11 13:59:22 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
/*
|
|
|
|
Check if type is supported by NDB.
|
2004-09-13 14:46:38 +02:00
|
|
|
TODO Use this once in open(), not in every operation
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
static inline bool ndb_supported_type(enum_field_types type)
|
|
|
|
{
|
|
|
|
switch (type) {
|
2004-07-23 13:21:39 +02:00
|
|
|
case MYSQL_TYPE_DECIMAL:
|
|
|
|
case MYSQL_TYPE_TINY:
|
|
|
|
case MYSQL_TYPE_SHORT:
|
|
|
|
case MYSQL_TYPE_LONG:
|
|
|
|
case MYSQL_TYPE_INT24:
|
|
|
|
case MYSQL_TYPE_LONGLONG:
|
|
|
|
case MYSQL_TYPE_FLOAT:
|
|
|
|
case MYSQL_TYPE_DOUBLE:
|
|
|
|
case MYSQL_TYPE_TIMESTAMP:
|
|
|
|
case MYSQL_TYPE_DATETIME:
|
|
|
|
case MYSQL_TYPE_DATE:
|
|
|
|
case MYSQL_TYPE_NEWDATE:
|
|
|
|
case MYSQL_TYPE_TIME:
|
|
|
|
case MYSQL_TYPE_YEAR:
|
|
|
|
case MYSQL_TYPE_STRING:
|
|
|
|
case MYSQL_TYPE_VAR_STRING:
|
|
|
|
case MYSQL_TYPE_TINY_BLOB:
|
|
|
|
case MYSQL_TYPE_BLOB:
|
|
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
|
|
|
case MYSQL_TYPE_LONG_BLOB:
|
|
|
|
case MYSQL_TYPE_ENUM:
|
|
|
|
case MYSQL_TYPE_SET:
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
case MYSQL_TYPE_NULL:
|
|
|
|
case MYSQL_TYPE_GEOMETRY:
|
2004-07-23 13:21:39 +02:00
|
|
|
break;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Instruct NDB to set the value of the hidden primary key
|
|
|
|
*/
|
|
|
|
|
|
|
|
bool ha_ndbcluster::set_hidden_key(NdbOperation *ndb_op,
|
|
|
|
uint fieldnr, const byte *field_ptr)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("set_hidden_key");
|
|
|
|
DBUG_RETURN(ndb_op->equal(fieldnr, (char*)field_ptr,
|
|
|
|
NDB_HIDDEN_PRIMARY_KEY_LENGTH) != 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Instruct NDB to set the value of one primary key attribute
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::set_ndb_key(NdbOperation *ndb_op, Field *field,
|
|
|
|
uint fieldnr, const byte *field_ptr)
|
|
|
|
{
|
|
|
|
uint32 pack_len= field->pack_length();
|
|
|
|
DBUG_ENTER("set_ndb_key");
|
|
|
|
DBUG_PRINT("enter", ("%d: %s, ndb_type: %u, len=%d",
|
|
|
|
fieldnr, field->field_name, field->type(),
|
|
|
|
pack_len));
|
|
|
|
DBUG_DUMP("key", (char*)field_ptr, pack_len);
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_supported_type(field->type()))
|
|
|
|
{
|
|
|
|
if (! (field->flags & BLOB_FLAG))
|
|
|
|
// Common implementation for most field types
|
|
|
|
DBUG_RETURN(ndb_op->equal(fieldnr, (char*) field_ptr, pack_len) != 0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
// Unhandled field types
|
|
|
|
DBUG_PRINT("error", ("Field type %d not supported", field->type()));
|
|
|
|
DBUG_RETURN(2);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Instruct NDB to set the value of one attribute
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::set_ndb_value(NdbOperation *ndb_op, Field *field,
|
2004-10-01 13:16:49 +02:00
|
|
|
uint fieldnr, bool *set_blob_value)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
const byte* field_ptr= field->ptr;
|
|
|
|
uint32 pack_len= field->pack_length();
|
|
|
|
DBUG_ENTER("set_ndb_value");
|
|
|
|
DBUG_PRINT("enter", ("%d: %s, type: %u, len=%d, is_null=%s",
|
|
|
|
fieldnr, field->field_name, field->type(),
|
|
|
|
pack_len, field->is_null()?"Y":"N"));
|
|
|
|
DBUG_DUMP("value", (char*) field_ptr, pack_len);
|
2004-07-22 12:38:09 +02:00
|
|
|
|
|
|
|
if (ndb_supported_type(field->type()))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-10-20 19:22:58 +02:00
|
|
|
// ndb currently does not support size 0
|
|
|
|
const byte *empty_field= "";
|
|
|
|
if (pack_len == 0)
|
|
|
|
{
|
|
|
|
pack_len= 1;
|
|
|
|
field_ptr= empty_field;
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
if (! (field->flags & BLOB_FLAG))
|
|
|
|
{
|
|
|
|
if (field->is_null())
|
|
|
|
// Set value to NULL
|
|
|
|
DBUG_RETURN((ndb_op->setValue(fieldnr, (char*)NULL, pack_len) != 0));
|
|
|
|
// Common implementation for most field types
|
|
|
|
DBUG_RETURN(ndb_op->setValue(fieldnr, (char*)field_ptr, pack_len) != 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Blob type
|
2004-07-23 15:46:56 +02:00
|
|
|
NdbBlob *ndb_blob= ndb_op->getBlobHandle(fieldnr);
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_blob != NULL)
|
|
|
|
{
|
|
|
|
if (field->is_null())
|
|
|
|
DBUG_RETURN(ndb_blob->setNull() != 0);
|
|
|
|
|
|
|
|
Field_blob *field_blob= (Field_blob*)field;
|
|
|
|
|
|
|
|
// Get length and pointer to data
|
|
|
|
uint32 blob_len= field_blob->get_length(field_ptr);
|
|
|
|
char* blob_ptr= NULL;
|
|
|
|
field_blob->get_ptr(&blob_ptr);
|
|
|
|
|
2004-09-02 12:42:03 +02:00
|
|
|
// Looks like NULL ptr signals length 0 blob
|
|
|
|
if (blob_ptr == NULL) {
|
|
|
|
DBUG_ASSERT(blob_len == 0);
|
2004-09-15 17:31:15 +02:00
|
|
|
blob_ptr= (char*)"";
|
2004-09-02 12:42:03 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("value", ("set blob ptr=%x len=%u",
|
|
|
|
(unsigned)blob_ptr, blob_len));
|
|
|
|
DBUG_DUMP("value", (char*)blob_ptr, min(blob_len, 26));
|
|
|
|
|
2004-10-01 13:16:49 +02:00
|
|
|
if (set_blob_value)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
*set_blob_value= TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
// No callback needed to write value
|
|
|
|
DBUG_RETURN(ndb_blob->setValue(blob_ptr, blob_len) != 0);
|
|
|
|
}
|
|
|
|
DBUG_RETURN(1);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
// Unhandled field types
|
|
|
|
DBUG_PRINT("error", ("Field type %d not supported", field->type()));
|
|
|
|
DBUG_RETURN(2);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Callback to read all blob values.
|
|
|
|
- not done in unpack_record because unpack_record is valid
|
|
|
|
after execute(Commit) but reading blobs is not
|
|
|
|
- may only generate read operations; they have to be executed
|
|
|
|
somewhere before the data is available
|
|
|
|
- due to single buffer for all blobs, we let the last blob
|
|
|
|
process all blobs (last so that all are active)
|
|
|
|
- null bit is still set in unpack_record
|
|
|
|
- TODO allocate blob part aligned buffers
|
|
|
|
*/
|
|
|
|
|
2004-07-29 10:44:53 +02:00
|
|
|
NdbBlob::ActiveHook g_get_ndb_blobs_value;
|
2004-07-22 12:38:09 +02:00
|
|
|
|
2004-07-29 10:44:53 +02:00
|
|
|
int g_get_ndb_blobs_value(NdbBlob *ndb_blob, void *arg)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2004-07-29 10:44:53 +02:00
|
|
|
DBUG_ENTER("g_get_ndb_blobs_value");
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_blob->blobsNextBlob() != NULL)
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
ha_ndbcluster *ha= (ha_ndbcluster *)arg;
|
|
|
|
DBUG_RETURN(ha->get_ndb_blobs_value(ndb_blob));
|
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::get_ndb_blobs_value(NdbBlob *last_ndb_blob)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("get_ndb_blobs_value");
|
|
|
|
|
|
|
|
// Field has no field number so cannot use TABLE blob_field
|
|
|
|
// Loop twice, first only counting total buffer size
|
|
|
|
for (int loop= 0; loop <= 1; loop++)
|
|
|
|
{
|
|
|
|
uint32 offset= 0;
|
|
|
|
for (uint i= 0; i < table->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
NdbValue value= m_value[i];
|
|
|
|
if (value.ptr != NULL && (field->flags & BLOB_FLAG))
|
|
|
|
{
|
|
|
|
Field_blob *field_blob= (Field_blob *)field;
|
|
|
|
NdbBlob *ndb_blob= value.blob;
|
|
|
|
Uint64 blob_len= 0;
|
|
|
|
if (ndb_blob->getLength(blob_len) != 0)
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
// Align to Uint64
|
|
|
|
uint32 blob_size= blob_len;
|
|
|
|
if (blob_size % 8 != 0)
|
|
|
|
blob_size+= 8 - blob_size % 8;
|
|
|
|
if (loop == 1)
|
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
char *buf= m_blobs_buffer + offset;
|
2004-07-22 12:38:09 +02:00
|
|
|
uint32 len= 0xffffffff; // Max uint32
|
|
|
|
DBUG_PRINT("value", ("read blob ptr=%x len=%u",
|
|
|
|
(uint)buf, (uint)blob_len));
|
|
|
|
if (ndb_blob->readData(buf, len) != 0)
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
DBUG_ASSERT(len == blob_len);
|
|
|
|
field_blob->set_ptr(len, buf);
|
|
|
|
}
|
|
|
|
offset+= blob_size;
|
|
|
|
}
|
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
if (loop == 0 && offset > m_blobs_buffer_size)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
my_free(m_blobs_buffer, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
m_blobs_buffer_size= 0;
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("value", ("allocate blobs buffer size %u", offset));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_buffer= my_malloc(offset, MYF(MY_WME));
|
|
|
|
if (m_blobs_buffer == NULL)
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(-1);
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_buffer_size= offset;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Instruct NDB to fetch one field
|
2004-07-22 12:38:09 +02:00
|
|
|
- data is read directly into buffer provided by field
|
|
|
|
if field is NULL, data is read into memory provided by NDBAPI
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
int ha_ndbcluster::get_ndb_value(NdbOperation *ndb_op, Field *field,
|
2004-09-20 12:40:53 +02:00
|
|
|
uint fieldnr, byte* buf)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("get_ndb_value");
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("enter", ("fieldnr: %d flags: %o", fieldnr,
|
|
|
|
(int)(field != NULL ? field->flags : 0)));
|
|
|
|
|
|
|
|
if (field != NULL)
|
|
|
|
{
|
2004-09-20 12:40:53 +02:00
|
|
|
DBUG_ASSERT(buf);
|
2004-07-22 12:38:09 +02:00
|
|
|
if (ndb_supported_type(field->type()))
|
|
|
|
{
|
|
|
|
DBUG_ASSERT(field->ptr != NULL);
|
|
|
|
if (! (field->flags & BLOB_FLAG))
|
2004-09-20 12:40:53 +02:00
|
|
|
{
|
2004-10-20 19:22:58 +02:00
|
|
|
byte *field_buf;
|
|
|
|
if (field->pack_length() != 0)
|
|
|
|
field_buf= buf + (field->ptr - table->record[0]);
|
|
|
|
else
|
|
|
|
field_buf= dummy_buf;
|
2004-09-20 12:40:53 +02:00
|
|
|
m_value[fieldnr].rec= ndb_op->getValue(fieldnr,
|
|
|
|
field_buf);
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(m_value[fieldnr].rec == NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Blob type
|
|
|
|
NdbBlob *ndb_blob= ndb_op->getBlobHandle(fieldnr);
|
|
|
|
m_value[fieldnr].blob= ndb_blob;
|
|
|
|
if (ndb_blob != NULL)
|
|
|
|
{
|
|
|
|
// Set callback
|
|
|
|
void *arg= (void *)this;
|
2004-07-29 10:44:53 +02:00
|
|
|
DBUG_RETURN(ndb_blob->setActiveHook(g_get_ndb_blobs_value, arg) != 0);
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
// Unhandled field types
|
|
|
|
DBUG_PRINT("error", ("Field type %d not supported", field->type()));
|
|
|
|
DBUG_RETURN(2);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Used for hidden key only
|
|
|
|
m_value[fieldnr].rec= ndb_op->getValue(fieldnr, NULL);
|
|
|
|
DBUG_RETURN(m_value[fieldnr].rec == NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Check if any set or get of blob value in current query.
|
|
|
|
*/
|
|
|
|
bool ha_ndbcluster::uses_blob_value(bool all_fields)
|
|
|
|
{
|
|
|
|
if (table->blob_fields == 0)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
if (all_fields)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
|
|
|
uint no_fields= table->fields;
|
|
|
|
int i;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
THD *thd= table->in_use;
|
2004-07-22 12:38:09 +02:00
|
|
|
// They always put blobs at the end..
|
|
|
|
for (i= no_fields - 1; i >= 0; i--)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if (thd->query_id == field->query_id)
|
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
return FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get metadata for this table from NDB
|
|
|
|
|
|
|
|
IMPLEMENTATION
|
|
|
|
- save the NdbDictionary::Table for easy access
|
|
|
|
- check that frm-file on disk is equal to frm-file
|
|
|
|
of table accessed in NDB
|
|
|
|
- build a list of the indexes for the table
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::get_metadata(const char *path)
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
const NDBTAB *tab;
|
|
|
|
int error;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool invalidating_ndb_table= FALSE;
|
2004-09-24 18:58:25 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("get_metadata");
|
|
|
|
DBUG_PRINT("enter", ("m_tabname: %s, path: %s", m_tabname, path));
|
|
|
|
|
2004-09-24 18:58:25 +02:00
|
|
|
do {
|
|
|
|
const void *data, *pack_data;
|
|
|
|
uint length, pack_length;
|
|
|
|
|
|
|
|
if (!(tab= dict->getTable(m_tabname)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
DBUG_PRINT("info", ("Table schema version: %d", tab->getObjectVersion()));
|
|
|
|
/*
|
|
|
|
Compare FrmData in NDB with frm file from disk.
|
|
|
|
*/
|
|
|
|
error= 0;
|
|
|
|
if (readfrm(path, &data, &length) ||
|
|
|
|
packfrm(data, length, &pack_data, &pack_length))
|
|
|
|
{
|
|
|
|
my_free((char*)data, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
my_free((char*)pack_data, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-24 18:58:25 +02:00
|
|
|
if ((pack_length != tab->getFrmLength()) ||
|
|
|
|
(memcmp(pack_data, tab->getFrmData(), pack_length)))
|
|
|
|
{
|
|
|
|
if (!invalidating_ndb_table)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Invalidating table"));
|
|
|
|
dict->invalidateTable(m_tabname);
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
invalidating_ndb_table= TRUE;
|
2004-09-24 18:58:25 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error",
|
|
|
|
("metadata, pack_length: %d getFrmLength: %d memcmp: %d",
|
|
|
|
pack_length, tab->getFrmLength(),
|
|
|
|
memcmp(pack_data, tab->getFrmData(), pack_length)));
|
|
|
|
DBUG_DUMP("pack_data", (char*)pack_data, pack_length);
|
|
|
|
DBUG_DUMP("frm", (char*)tab->getFrmData(), tab->getFrmLength());
|
2004-09-28 08:53:21 +02:00
|
|
|
error= 3;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
invalidating_ndb_table= FALSE;
|
2004-09-24 18:58:25 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
invalidating_ndb_table= FALSE;
|
2004-09-24 18:58:25 +02:00
|
|
|
}
|
|
|
|
my_free((char*)data, MYF(0));
|
|
|
|
my_free((char*)pack_data, MYF(0));
|
|
|
|
} while (invalidating_ndb_table);
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (error)
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table= NULL;
|
|
|
|
m_table_info= NULL;
|
2004-09-03 13:55:40 +02:00
|
|
|
|
2004-08-19 11:10:35 +02:00
|
|
|
DBUG_RETURN(build_index_list(table, ILBP_OPEN));
|
2004-05-24 12:35:39 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-08-19 11:10:35 +02:00
|
|
|
|
2004-08-19 11:51:06 +02:00
|
|
|
int ha_ndbcluster::build_index_list(TABLE *tab, enum ILBP phase)
|
2004-05-24 12:35:39 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
uint i;
|
2004-08-19 11:10:35 +02:00
|
|
|
int error= 0;
|
2004-09-25 14:35:27 +02:00
|
|
|
const char *name, *index_name;
|
|
|
|
char unique_index_name[FN_LEN];
|
2004-05-24 12:35:39 +02:00
|
|
|
static const char* unique_suffix= "$unique";
|
2004-08-19 11:10:35 +02:00
|
|
|
KEY* key_info= tab->key_info;
|
|
|
|
const char **key_name= tab->keynames.type_names;
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NdbDictionary::Dictionary *dict= ndb->getDictionary();
|
2004-08-19 11:10:35 +02:00
|
|
|
DBUG_ENTER("build_index_list");
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Save information about all known indexes
|
2004-08-19 11:10:35 +02:00
|
|
|
for (i= 0; i < tab->keys; i++, key_info++, key_name++)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
2004-08-19 11:10:35 +02:00
|
|
|
index_name= *key_name;
|
2004-05-24 12:35:39 +02:00
|
|
|
NDB_INDEX_TYPE idx_type= get_index_type_from_table(i);
|
2004-08-18 19:13:39 +02:00
|
|
|
m_index[i].type= idx_type;
|
2004-05-24 12:35:39 +02:00
|
|
|
if (idx_type == UNIQUE_ORDERED_INDEX || idx_type == UNIQUE_INDEX)
|
2004-04-30 12:25:31 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
strxnmov(unique_index_name, FN_LEN, index_name, unique_suffix, NullS);
|
|
|
|
DBUG_PRINT("info", ("Created unique index name \'%s\' for index %d",
|
|
|
|
unique_index_name, i));
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
2004-08-19 11:10:35 +02:00
|
|
|
// Create secondary indexes if in create phase
|
|
|
|
if (phase == ILBP_CREATE)
|
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Creating index %u: %s", i, index_name));
|
|
|
|
switch (idx_type){
|
2004-08-19 11:10:35 +02:00
|
|
|
|
|
|
|
case PRIMARY_KEY_INDEX:
|
|
|
|
// Do nothing, already created
|
|
|
|
break;
|
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
|
|
|
error= create_ordered_index(index_name, key_info);
|
|
|
|
break;
|
|
|
|
case UNIQUE_ORDERED_INDEX:
|
|
|
|
if (!(error= create_ordered_index(index_name, key_info)))
|
2004-09-25 14:35:27 +02:00
|
|
|
error= create_unique_index(unique_index_name, key_info);
|
2004-08-19 11:10:35 +02:00
|
|
|
break;
|
|
|
|
case UNIQUE_INDEX:
|
2004-11-22 10:35:03 +01:00
|
|
|
if (!(error= check_index_fields_not_null(i)))
|
|
|
|
error= create_unique_index(unique_index_name, key_info);
|
2004-08-19 11:10:35 +02:00
|
|
|
break;
|
|
|
|
case ORDERED_INDEX:
|
|
|
|
error= create_ordered_index(index_name, key_info);
|
|
|
|
break;
|
|
|
|
default:
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_ASSERT(FALSE);
|
2004-08-19 11:10:35 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (error)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Failed to create index %u", i));
|
|
|
|
drop_table();
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Add handles to index objects
|
2004-09-25 14:35:27 +02:00
|
|
|
if (idx_type != PRIMARY_KEY_INDEX && idx_type != UNIQUE_INDEX)
|
2004-08-18 19:13:39 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Get handle to index %s", index_name));
|
2004-08-19 11:10:35 +02:00
|
|
|
const NDBINDEX *index= dict->getIndex(index_name, m_tabname);
|
2004-08-18 19:13:39 +02:00
|
|
|
if (!index) DBUG_RETURN(1);
|
2004-08-19 07:33:35 +02:00
|
|
|
m_index[i].index= (void *) index;
|
2004-08-18 19:13:39 +02:00
|
|
|
}
|
2004-09-25 14:35:27 +02:00
|
|
|
if (idx_type == UNIQUE_ORDERED_INDEX || idx_type == UNIQUE_INDEX)
|
2004-08-18 19:13:39 +02:00
|
|
|
{
|
2004-09-25 14:35:27 +02:00
|
|
|
DBUG_PRINT("info", ("Get handle to unique_index %s", unique_index_name));
|
|
|
|
const NDBINDEX *index= dict->getIndex(unique_index_name, m_tabname);
|
2004-08-18 19:13:39 +02:00
|
|
|
if (!index) DBUG_RETURN(1);
|
2004-08-19 07:33:35 +02:00
|
|
|
m_index[i].unique_index= (void *) index;
|
2004-08-18 19:13:39 +02:00
|
|
|
}
|
|
|
|
}
|
2004-08-19 11:10:35 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(error);
|
2004-08-18 19:13:39 +02:00
|
|
|
}
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Decode the type of an index from information
|
|
|
|
provided in table object
|
|
|
|
*/
|
2004-05-24 12:35:39 +02:00
|
|
|
NDB_INDEX_TYPE ha_ndbcluster::get_index_type_from_table(uint inx) const
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
bool is_hash_index= (table->key_info[inx].algorithm == HA_KEY_ALG_HASH);
|
|
|
|
if (inx == table->primary_key)
|
|
|
|
return is_hash_index ? PRIMARY_KEY_INDEX : PRIMARY_KEY_ORDERED_INDEX;
|
2004-04-15 09:14:14 +02:00
|
|
|
else
|
2004-05-24 12:35:39 +02:00
|
|
|
return ((table->key_info[inx].flags & HA_NOSAME) ?
|
|
|
|
(is_hash_index ? UNIQUE_INDEX : UNIQUE_ORDERED_INDEX) :
|
2004-04-15 09:14:14 +02:00
|
|
|
ORDERED_INDEX);
|
|
|
|
}
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-11-22 10:35:03 +01:00
|
|
|
int ha_ndbcluster::check_index_fields_not_null(uint inx)
|
|
|
|
{
|
|
|
|
KEY* key_info= table->key_info + inx;
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
DBUG_ENTER("check_index_fields_not_null");
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
|
|
|
if (field->maybe_null())
|
|
|
|
{
|
|
|
|
my_printf_error(ER_NULL_COLUMN_IN_INDEX,ER(ER_NULL_COLUMN_IN_INDEX),
|
|
|
|
MYF(0),field->field_name);
|
|
|
|
DBUG_RETURN(ER_NULL_COLUMN_IN_INDEX);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
void ha_ndbcluster::release_metadata()
|
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
uint i;
|
2004-04-30 12:25:31 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("release_metadata");
|
|
|
|
DBUG_PRINT("enter", ("m_tabname: %s", m_tabname));
|
|
|
|
|
|
|
|
m_table= NULL;
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table_info= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
// Release index list
|
2004-04-30 12:25:31 +02:00
|
|
|
for (i= 0; i < MAX_KEY; i++)
|
|
|
|
{
|
2004-08-18 19:13:39 +02:00
|
|
|
m_index[i].unique_index= NULL;
|
|
|
|
m_index[i].index= NULL;
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
int ha_ndbcluster::get_ndb_lock_type(enum thr_lock_type type)
|
2004-05-10 14:12:28 +02:00
|
|
|
{
|
2004-10-13 10:08:18 +02:00
|
|
|
if (type >= TL_WRITE_ALLOW_WRITE)
|
2004-09-29 01:42:34 +02:00
|
|
|
return NdbOperation::LM_Exclusive;
|
2004-11-03 15:53:26 +01:00
|
|
|
else if (uses_blob_value(m_retrieve_all_fields))
|
2004-10-05 13:51:32 +02:00
|
|
|
return NdbOperation::LM_Read;
|
2004-07-22 12:38:09 +02:00
|
|
|
else
|
2004-09-29 01:42:34 +02:00
|
|
|
return NdbOperation::LM_CommittedRead;
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
static const ulong index_type_flags[]=
|
|
|
|
{
|
|
|
|
/* UNDEFINED_INDEX */
|
|
|
|
0,
|
|
|
|
|
|
|
|
/* PRIMARY_KEY_INDEX */
|
2004-06-24 12:04:22 +02:00
|
|
|
HA_ONLY_WHOLE_INDEX,
|
2004-05-24 12:35:39 +02:00
|
|
|
|
|
|
|
/* PRIMARY_KEY_ORDERED_INDEX */
|
2004-05-10 14:46:06 +02:00
|
|
|
/*
|
2004-06-24 14:49:34 +02:00
|
|
|
Enable HA_KEYREAD_ONLY when "sorted" indexes are supported,
|
2004-05-10 14:46:06 +02:00
|
|
|
thus ORDERD BY clauses can be optimized by reading directly
|
|
|
|
through the index.
|
|
|
|
*/
|
2004-06-24 14:49:34 +02:00
|
|
|
// HA_KEYREAD_ONLY |
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_NEXT |
|
|
|
|
HA_READ_RANGE |
|
|
|
|
HA_READ_ORDER,
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/* UNIQUE_INDEX */
|
2004-06-24 12:04:22 +02:00
|
|
|
HA_ONLY_WHOLE_INDEX,
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
/* UNIQUE_ORDERED_INDEX */
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_NEXT |
|
|
|
|
HA_READ_RANGE |
|
|
|
|
HA_READ_ORDER,
|
2004-05-24 12:35:39 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/* ORDERED_INDEX */
|
2004-09-03 13:55:40 +02:00
|
|
|
HA_READ_NEXT |
|
|
|
|
HA_READ_RANGE |
|
|
|
|
HA_READ_ORDER
|
2004-04-15 09:14:14 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
static const int index_flags_size= sizeof(index_type_flags)/sizeof(ulong);
|
|
|
|
|
|
|
|
inline NDB_INDEX_TYPE ha_ndbcluster::get_index_type(uint idx_no) const
|
|
|
|
{
|
|
|
|
DBUG_ASSERT(idx_no < MAX_KEY);
|
2004-08-18 19:13:39 +02:00
|
|
|
return m_index[idx_no].type;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Get the flags for an index
|
|
|
|
|
|
|
|
RETURN
|
|
|
|
flags depending on the type of the index.
|
|
|
|
*/
|
|
|
|
|
2004-07-08 14:45:25 +02:00
|
|
|
inline ulong ha_ndbcluster::index_flags(uint idx_no, uint part,
|
|
|
|
bool all_parts) const
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("index_flags");
|
2004-05-10 14:46:06 +02:00
|
|
|
DBUG_PRINT("info", ("idx_no: %d", idx_no));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(get_index_type_from_table(idx_no) < index_flags_size);
|
|
|
|
DBUG_RETURN(index_type_flags[get_index_type_from_table(idx_no)]);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::set_primary_key(NdbOperation *op, const byte *key)
|
|
|
|
{
|
|
|
|
KEY* key_info= table->key_info + table->primary_key;
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
DBUG_ENTER("set_primary_key");
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
|
|
|
if (set_ndb_key(op, field,
|
|
|
|
key_part->fieldnr-1, key))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
key += key_part->length;
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-05-17 09:51:02 +02:00
|
|
|
int ha_ndbcluster::set_primary_key_from_old_data(NdbOperation *op, const byte *old_data)
|
|
|
|
{
|
|
|
|
KEY* key_info= table->key_info + table->primary_key;
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
DBUG_ENTER("set_primary_key_from_old_data");
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
|
|
|
if (set_ndb_key(op, field,
|
|
|
|
key_part->fieldnr-1, old_data+key_part->offset))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::set_primary_key(NdbOperation *op)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("set_primary_key");
|
|
|
|
KEY* key_info= table->key_info + table->primary_key;
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO* end= key_part+key_info->key_parts;
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field* field= key_part->field;
|
|
|
|
if (set_ndb_key(op, field,
|
|
|
|
key_part->fieldnr-1, field->ptr))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Read one record from NDB using primary key
|
|
|
|
*/
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
int ha_ndbcluster::pk_read(const byte *key, uint key_len, byte *buf)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
uint no_fields= table->fields, i;
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbOperation *op;
|
|
|
|
THD *thd= current_thd;
|
|
|
|
DBUG_ENTER("pk_read");
|
|
|
|
DBUG_PRINT("enter", ("key_len: %u", key_len));
|
|
|
|
DBUG_DUMP("key", (char*)key, key_len);
|
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2004-09-28 19:35:07 +02:00
|
|
|
op->readTuple(lm) != 0)
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
if (table->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
// This table has no primary key, use "hidden" primary key
|
|
|
|
DBUG_PRINT("info", ("Using hidden key"));
|
|
|
|
DBUG_DUMP("key", (char*)key, 8);
|
|
|
|
if (set_hidden_key(op, no_fields, key))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Read key at the same time, for future reference
|
2004-09-20 12:40:53 +02:00
|
|
|
if (get_ndb_value(op, NULL, no_fields, NULL))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
|
|
|
if ((res= set_primary_key(op, key)))
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
// Read all wanted non-key field(s) unless HA_EXTRA_RETRIEVE_ALL_COLS
|
2004-04-15 09:14:14 +02:00
|
|
|
for (i= 0; i < no_fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
2004-07-22 18:35:51 +02:00
|
|
|
if ((thd->query_id == field->query_id) ||
|
2004-11-18 12:11:56 +01:00
|
|
|
m_retrieve_all_fields ||
|
|
|
|
(field->flags & PRI_KEY_FLAG) && m_retrieve_primary_key)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-20 12:40:53 +02:00
|
|
|
if (get_ndb_value(op, field, i, buf))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Attribute was not to be read
|
2004-07-22 12:38:09 +02:00
|
|
|
m_value[i].ptr= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-09-28 21:11:50 +02:00
|
|
|
if (execute_no_commit_ie(this,trans) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
|
|
|
|
// The value have now been fetched from NDB
|
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
/*
|
|
|
|
Read one complementing record from NDB using primary key from old_data
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::complemented_pk_read(const byte *old_data, byte *new_data)
|
|
|
|
{
|
|
|
|
uint no_fields= table->fields, i;
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbOperation *op;
|
|
|
|
THD *thd= current_thd;
|
|
|
|
DBUG_ENTER("complemented_pk_read");
|
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_retrieve_all_fields)
|
2004-07-22 18:35:51 +02:00
|
|
|
// We have allready retrieved all fields, nothing to complement
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2004-09-28 19:35:07 +02:00
|
|
|
op->readTuple(lm) != 0)
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-07-22 18:35:51 +02:00
|
|
|
|
|
|
|
int res;
|
2004-07-23 15:46:56 +02:00
|
|
|
if ((res= set_primary_key_from_old_data(op, old_data)))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-07-22 18:35:51 +02:00
|
|
|
|
|
|
|
// Read all unreferenced non-key field(s)
|
|
|
|
for (i= 0; i < no_fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if (!(field->flags & PRI_KEY_FLAG) &&
|
|
|
|
(thd->query_id != field->query_id))
|
|
|
|
{
|
2004-09-20 12:40:53 +02:00
|
|
|
if (get_ndb_value(op, field, i, new_data))
|
2004-07-23 10:33:51 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-07-22 18:35:51 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0)
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
|
|
|
|
// The value have now been fetched from NDB
|
|
|
|
unpack_record(new_data);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-11-18 12:11:56 +01:00
|
|
|
/*
|
|
|
|
Peek to check if a particular row already exists
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::peek_row()
|
|
|
|
{
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbOperation *op;
|
|
|
|
THD *thd= current_thd;
|
|
|
|
DBUG_ENTER("peek_row");
|
|
|
|
|
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
|
|
|
op->readTuple(lm) != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
int res;
|
|
|
|
if ((res= set_primary_key(op)))
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
if (execute_no_commit_ie(this,trans) != 0)
|
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-07-22 18:35:51 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Read one record from NDB using unique secondary index
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::unique_index_read(const byte *key,
|
|
|
|
uint key_len, byte *buf)
|
|
|
|
{
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbIndexOperation *op;
|
|
|
|
THD *thd= current_thd;
|
|
|
|
byte *key_ptr;
|
|
|
|
KEY* key_info;
|
|
|
|
KEY_PART_INFO *key_part, *end;
|
|
|
|
uint i;
|
|
|
|
DBUG_ENTER("unique_index_read");
|
|
|
|
DBUG_PRINT("enter", ("key_len: %u, index: %u", key_len, active_index));
|
|
|
|
DBUG_DUMP("key", (char*)key, key_len);
|
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
2004-08-18 19:13:39 +02:00
|
|
|
if (!(op= trans->getNdbIndexOperation((NDBINDEX *)
|
|
|
|
m_index[active_index].unique_index,
|
2004-09-07 21:53:59 +02:00
|
|
|
(const NDBTAB *) m_table)) ||
|
2004-09-28 19:35:07 +02:00
|
|
|
op->readTuple(lm) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
// Set secondary index key(s)
|
|
|
|
key_ptr= (byte *) key;
|
|
|
|
key_info= table->key_info + active_index;
|
|
|
|
DBUG_ASSERT(key_info->key_length == key_len);
|
|
|
|
end= (key_part= key_info->key_part) + key_info->key_parts;
|
|
|
|
|
|
|
|
for (i= 0; key_part != end; key_part++, i++)
|
|
|
|
{
|
2004-11-30 07:41:26 +01:00
|
|
|
if (set_ndb_key(op, key_part->field, i,
|
|
|
|
key_part->null_bit ? key_ptr + 1 : key_ptr))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-30 07:41:26 +01:00
|
|
|
key_ptr+= key_part->store_length;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Get non-index attribute(s)
|
|
|
|
for (i= 0; i < table->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if ((thd->query_id == field->query_id) ||
|
2004-11-18 12:11:56 +01:00
|
|
|
(field->flags & PRI_KEY_FLAG)) // && m_retrieve_primary_key ??
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-20 12:40:53 +02:00
|
|
|
if (get_ndb_value(op, field, i, buf))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Attribute was not to be read
|
2004-07-22 12:38:09 +02:00
|
|
|
m_value[i].ptr= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-09-28 21:11:50 +02:00
|
|
|
if (execute_no_commit_ie(this,trans) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
table->status= STATUS_NOT_FOUND;
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
// The value have now been fetched from NDB
|
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2004-05-10 14:12:28 +02:00
|
|
|
Get the next record of a started scan. Try to fetch
|
|
|
|
it locally from NdbApi cached records if possible,
|
|
|
|
otherwise ask NDB for more.
|
|
|
|
|
|
|
|
NOTE
|
|
|
|
If this is a update/delete make sure to not contact
|
|
|
|
NDB before any pending ops have been sent to NDB.
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
inline int ha_ndbcluster::next_result(byte *buf)
|
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
int check;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbResultSet *cursor= m_active_cursor;
|
|
|
|
DBUG_ENTER("next_result");
|
2004-05-10 14:12:28 +02:00
|
|
|
|
|
|
|
if (!cursor)
|
|
|
|
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
|
|
|
|
|
|
|
/*
|
|
|
|
If this an update or delete, call nextResult with false
|
|
|
|
to process any records already cached in NdbApi
|
|
|
|
*/
|
2004-10-13 10:08:18 +02:00
|
|
|
bool contact_ndb= m_lock.type < TL_WRITE_ALLOW_WRITE;
|
2004-05-10 14:12:28 +02:00
|
|
|
do {
|
|
|
|
DBUG_PRINT("info", ("Call nextResult, contact_ndb: %d", contact_ndb));
|
2004-07-22 12:38:09 +02:00
|
|
|
/*
|
|
|
|
We can only handle one tuple with blobs at a time.
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_ops_pending && m_blobs_pending)
|
2004-07-22 12:38:09 +02:00
|
|
|
{
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0)
|
2004-09-03 17:11:09 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
|
|
|
m_blobs_pending= FALSE;
|
2004-07-22 12:38:09 +02:00
|
|
|
}
|
2004-11-22 14:41:46 +01:00
|
|
|
check= cursor->nextResult(contact_ndb, m_force_send);
|
2004-05-10 14:12:28 +02:00
|
|
|
if (check == 0)
|
|
|
|
{
|
|
|
|
// One more record found
|
|
|
|
DBUG_PRINT("info", ("One more record found"));
|
2004-07-22 12:38:09 +02:00
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
else if (check == 1 || check == 2)
|
|
|
|
{
|
|
|
|
// 1: No more records
|
|
|
|
// 2: No more cached records
|
|
|
|
|
|
|
|
/*
|
|
|
|
Before fetching more rows and releasing lock(s),
|
|
|
|
all pending update or delete operations should
|
|
|
|
be sent to NDB
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
DBUG_PRINT("info", ("ops_pending: %d", m_ops_pending));
|
|
|
|
if (m_ops_pending)
|
2004-09-03 17:11:09 +02:00
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
// if (current_thd->transaction.on)
|
|
|
|
if (m_transaction_on)
|
2004-10-04 14:26:26 +02:00
|
|
|
{
|
|
|
|
if (execute_no_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (execute_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-10-05 12:04:09 +02:00
|
|
|
int res= trans->restart();
|
|
|
|
DBUG_ASSERT(res == 0);
|
2004-10-04 14:26:26 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
|
|
|
|
contact_ndb= (check == 2);
|
|
|
|
}
|
|
|
|
} while (check == 2);
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
table->status= STATUS_NOT_FOUND;
|
2004-05-13 11:56:45 +02:00
|
|
|
if (check == -1)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
// No more records
|
|
|
|
DBUG_PRINT("info", ("No more records"));
|
|
|
|
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
|
|
|
}
|
|
|
|
|
2004-04-30 13:38:41 +02:00
|
|
|
/*
|
2004-10-08 10:50:50 +02:00
|
|
|
Set bounds for ordered index scan.
|
2004-04-30 13:38:41 +02:00
|
|
|
*/
|
|
|
|
|
2004-06-11 13:49:22 +02:00
|
|
|
int ha_ndbcluster::set_bounds(NdbIndexScanOperation *op,
|
2004-10-08 10:50:50 +02:00
|
|
|
const key_range *keys[2])
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
2004-10-08 10:50:50 +02:00
|
|
|
const KEY *const key_info= table->key_info + active_index;
|
|
|
|
const uint key_parts= key_info->key_parts;
|
|
|
|
uint key_tot_len[2];
|
|
|
|
uint tot_len;
|
2004-10-25 22:35:46 +02:00
|
|
|
uint i, j;
|
2004-04-30 13:38:41 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("set_bounds");
|
2004-10-08 10:50:50 +02:00
|
|
|
DBUG_PRINT("info", ("key_parts=%d", key_parts));
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
for (j= 0; j <= 1; j++)
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
2004-10-08 10:50:50 +02:00
|
|
|
const key_range *key= keys[j];
|
|
|
|
if (key != NULL)
|
|
|
|
{
|
|
|
|
// for key->flag see ha_rkey_function
|
|
|
|
DBUG_PRINT("info", ("key %d length=%d flag=%d",
|
|
|
|
j, key->length, key->flag));
|
|
|
|
key_tot_len[j]= key->length;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("key %d not present", j));
|
|
|
|
key_tot_len[j]= 0;
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
}
|
2004-10-08 10:50:50 +02:00
|
|
|
tot_len= 0;
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
for (i= 0; i < key_parts; i++)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO *key_part= &key_info->key_part[i];
|
|
|
|
Field *field= key_part->field;
|
|
|
|
uint part_len= key_part->length;
|
|
|
|
uint part_store_len= key_part->store_length;
|
|
|
|
// Info about each key part
|
|
|
|
struct part_st {
|
|
|
|
bool part_last;
|
|
|
|
const key_range *key;
|
|
|
|
const byte *part_ptr;
|
|
|
|
bool part_null;
|
|
|
|
int bound_type;
|
|
|
|
const char* bound_ptr;
|
|
|
|
};
|
|
|
|
struct part_st part[2];
|
|
|
|
|
|
|
|
for (j= 0; j <= 1; j++)
|
|
|
|
{
|
|
|
|
struct part_st &p = part[j];
|
|
|
|
p.key= NULL;
|
|
|
|
p.bound_type= -1;
|
|
|
|
if (tot_len < key_tot_len[j])
|
|
|
|
{
|
|
|
|
p.part_last= (tot_len + part_store_len >= key_tot_len[j]);
|
|
|
|
p.key= keys[j];
|
|
|
|
p.part_ptr= &p.key->key[tot_len];
|
2004-11-08 11:11:49 +01:00
|
|
|
p.part_null= key_part->null_bit && *p.part_ptr;
|
2004-10-08 10:50:50 +02:00
|
|
|
p.bound_ptr= (const char *)
|
2004-11-08 11:11:49 +01:00
|
|
|
p.part_null ? 0 : key_part->null_bit ? p.part_ptr + 1 : p.part_ptr;
|
2004-10-08 10:50:50 +02:00
|
|
|
|
|
|
|
if (j == 0)
|
|
|
|
{
|
|
|
|
switch (p.key->flag)
|
|
|
|
{
|
|
|
|
case HA_READ_KEY_EXACT:
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundEQ;
|
|
|
|
break;
|
|
|
|
case HA_READ_KEY_OR_NEXT:
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLE;
|
|
|
|
break;
|
|
|
|
case HA_READ_AFTER_KEY:
|
|
|
|
if (! p.part_last)
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLE;
|
|
|
|
else
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundLT;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (j == 1) {
|
|
|
|
switch (p.key->flag)
|
|
|
|
{
|
|
|
|
case HA_READ_BEFORE_KEY:
|
|
|
|
if (! p.part_last)
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGE;
|
|
|
|
else
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGT;
|
|
|
|
break;
|
|
|
|
case HA_READ_AFTER_KEY: // weird
|
|
|
|
p.bound_type= NdbIndexScanOperation::BoundGE;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
if (p.bound_type == -1)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("key %d unknown flag %d", j, p.key->flag));
|
|
|
|
DBUG_ASSERT(false);
|
|
|
|
// Stop setting bounds but continue with what we have
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2004-09-06 10:33:12 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
// Seen with e.g. b = 1 and c > 1
|
|
|
|
if (part[0].bound_type == NdbIndexScanOperation::BoundLE &&
|
|
|
|
part[1].bound_type == NdbIndexScanOperation::BoundGE &&
|
|
|
|
memcmp(part[0].part_ptr, part[1].part_ptr, part_store_len) == 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("replace LE/GE pair by EQ"));
|
|
|
|
part[0].bound_type= NdbIndexScanOperation::BoundEQ;
|
|
|
|
part[1].bound_type= -1;
|
|
|
|
}
|
|
|
|
// Not seen but was in previous version
|
|
|
|
if (part[0].bound_type == NdbIndexScanOperation::BoundEQ &&
|
|
|
|
part[1].bound_type == NdbIndexScanOperation::BoundGE &&
|
|
|
|
memcmp(part[0].part_ptr, part[1].part_ptr, part_store_len) == 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("remove GE from EQ/GE pair"));
|
|
|
|
part[1].bound_type= -1;
|
|
|
|
}
|
2004-09-06 10:33:12 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
for (j= 0; j <= 1; j++)
|
|
|
|
{
|
|
|
|
struct part_st &p = part[j];
|
|
|
|
// Set bound if not done with this key
|
|
|
|
if (p.key != NULL)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("key %d:%d offset=%d length=%d last=%d bound=%d",
|
|
|
|
j, i, tot_len, part_len, p.part_last, p.bound_type));
|
|
|
|
DBUG_DUMP("info", (const char*)p.part_ptr, part_store_len);
|
|
|
|
|
|
|
|
// Set bound if not cancelled via type -1
|
|
|
|
if (p.bound_type != -1)
|
2004-10-28 17:35:06 +02:00
|
|
|
{
|
|
|
|
char truncated_field_name[NDB_MAX_ATTR_NAME_SIZE];
|
|
|
|
strnmov(truncated_field_name,field->field_name,sizeof(truncated_field_name));
|
|
|
|
truncated_field_name[sizeof(truncated_field_name)-1]= '\0';
|
|
|
|
if (op->setBound(truncated_field_name, p.bound_type, p.bound_ptr))
|
2004-10-08 10:50:50 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2004-10-28 17:35:06 +02:00
|
|
|
}
|
2004-10-08 10:50:50 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
tot_len+= part_store_len;
|
2004-09-06 10:33:12 +02:00
|
|
|
}
|
2004-10-08 10:50:50 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-09-06 10:33:12 +02:00
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-10-21 12:39:08 +02:00
|
|
|
inline
|
|
|
|
int ha_ndbcluster::define_read_attrs(byte* buf, NdbOperation* op)
|
|
|
|
{
|
|
|
|
uint i;
|
|
|
|
THD *thd= current_thd;
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
|
|
|
|
DBUG_ENTER("define_read_attrs");
|
|
|
|
|
|
|
|
// Define attributes to read
|
|
|
|
for (i= 0; i < table->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if ((thd->query_id == field->query_id) ||
|
|
|
|
(field->flags & PRI_KEY_FLAG) ||
|
2004-11-03 15:53:26 +01:00
|
|
|
m_retrieve_all_fields)
|
2004-10-21 12:39:08 +02:00
|
|
|
{
|
|
|
|
if (get_ndb_value(op, field, i, buf))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
m_value[i].ptr= NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (table->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Getting hidden key"));
|
|
|
|
// Scanning table with no primary key
|
|
|
|
int hidden_no= table->fields;
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
|
|
|
if (!tab->getColumn(hidden_no))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
#endif
|
|
|
|
if (get_ndb_value(op, NULL, hidden_no, NULL))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
|
|
|
|
if (execute_no_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
DBUG_PRINT("exit", ("Scan started successfully"));
|
|
|
|
DBUG_RETURN(next_result(buf));
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
2004-05-10 14:12:28 +02:00
|
|
|
Start ordered index scan in NDB
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2004-04-30 13:38:41 +02:00
|
|
|
int ha_ndbcluster::ordered_index_scan(const key_range *start_key,
|
|
|
|
const key_range *end_key,
|
|
|
|
bool sorted, byte* buf)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-10-05 12:17:42 +02:00
|
|
|
bool restart;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbConnection *trans= m_active_trans;
|
2004-05-10 14:12:28 +02:00
|
|
|
NdbResultSet *cursor;
|
2004-06-11 13:49:22 +02:00
|
|
|
NdbIndexScanOperation *op;
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ordered_index_scan");
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_PRINT("enter", ("index: %u, sorted: %d", active_index, sorted));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("enter", ("Starting new ordered scan on %s", m_tabname));
|
2004-09-06 10:33:12 +02:00
|
|
|
|
2004-10-07 11:57:48 +02:00
|
|
|
// Check that sorted seems to be initialised
|
|
|
|
DBUG_ASSERT(sorted == 0 || sorted == 1);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-10-08 10:50:50 +02:00
|
|
|
if (m_active_cursor == 0)
|
2004-10-05 12:17:42 +02:00
|
|
|
{
|
|
|
|
restart= false;
|
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op= trans->getNdbIndexScanOperation((NDBINDEX *)
|
|
|
|
m_index[active_index].index,
|
|
|
|
(const NDBTAB *) m_table)) ||
|
|
|
|
!(cursor= op->readTuples(lm, 0, parallelism, sorted)))
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
m_active_cursor= cursor;
|
|
|
|
} else {
|
|
|
|
restart= true;
|
|
|
|
op= (NdbIndexScanOperation*)m_active_cursor->getOperation();
|
|
|
|
|
|
|
|
DBUG_ASSERT(op->getSorted() == sorted);
|
|
|
|
DBUG_ASSERT(op->getLockMode() ==
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type));
|
2004-11-22 14:41:46 +01:00
|
|
|
if(op->reset_bounds(m_force_send))
|
2004-10-05 12:17:42 +02:00
|
|
|
DBUG_RETURN(ndb_err(m_active_trans));
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-05-10 14:46:06 +02:00
|
|
|
{
|
2004-10-08 10:50:50 +02:00
|
|
|
const key_range *keys[2]= { start_key, end_key };
|
|
|
|
int ret= set_bounds(op, keys);
|
|
|
|
if (ret)
|
|
|
|
DBUG_RETURN(ret);
|
2004-05-10 14:46:06 +02:00
|
|
|
}
|
2004-10-08 10:50:50 +02:00
|
|
|
|
|
|
|
if (!restart)
|
2004-10-05 12:17:42 +02:00
|
|
|
{
|
|
|
|
DBUG_RETURN(define_read_attrs(buf, op));
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (execute_no_commit(this,trans) != 0)
|
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
|
|
|
|
DBUG_RETURN(next_result(buf));
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2004-05-10 14:12:28 +02:00
|
|
|
Start a filtered scan in NDB.
|
|
|
|
|
|
|
|
NOTE
|
|
|
|
This function is here as an example of how to start a
|
|
|
|
filtered scan. It should be possible to replace full_table_scan
|
|
|
|
with this function and make a best effort attempt
|
|
|
|
at filtering out the irrelevant data by converting the "items"
|
|
|
|
into interpreted instructions.
|
|
|
|
This would speed up table scans where there is a limiting WHERE clause
|
|
|
|
that doesn't match any index in the table.
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::filtered_scan(const byte *key, uint key_len,
|
|
|
|
byte *buf,
|
|
|
|
enum ha_rkey_function find_flag)
|
|
|
|
{
|
|
|
|
NdbConnection *trans= m_active_trans;
|
2004-05-10 14:12:28 +02:00
|
|
|
NdbResultSet *cursor;
|
|
|
|
NdbScanOperation *op;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("filtered_scan");
|
|
|
|
DBUG_PRINT("enter", ("key_len: %u, index: %u",
|
|
|
|
key_len, active_index));
|
|
|
|
DBUG_DUMP("key", (char*)key, key_len);
|
|
|
|
DBUG_PRINT("info", ("Starting a new filtered scan on %s",
|
|
|
|
m_tabname));
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op= trans->getNdbScanOperation((const NDBTAB *) m_table)) ||
|
|
|
|
!(cursor= op->readTuples(lm, 0, parallelism)))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
m_active_cursor= cursor;
|
2004-06-11 13:49:22 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
// Start scan filter
|
|
|
|
NdbScanFilter sf(op);
|
|
|
|
sf.begin();
|
|
|
|
|
|
|
|
// Set filter using the supplied key data
|
|
|
|
byte *key_ptr= (byte *) key;
|
|
|
|
uint tot_len= 0;
|
|
|
|
KEY* key_info= table->key_info + active_index;
|
|
|
|
for (uint k= 0; k < key_info->key_parts; k++)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO* key_part= key_info->key_part+k;
|
|
|
|
Field* field= key_part->field;
|
|
|
|
uint ndb_fieldnr= key_part->fieldnr-1;
|
|
|
|
DBUG_PRINT("key_part", ("fieldnr: %d", ndb_fieldnr));
|
2004-11-18 12:11:56 +01:00
|
|
|
//const NDBCOL *col= ((const NDBTAB *) m_table)->getColumn(ndb_fieldnr);
|
2004-04-15 09:14:14 +02:00
|
|
|
uint32 field_len= field->pack_length();
|
|
|
|
DBUG_DUMP("key", (char*)key, field_len);
|
|
|
|
|
|
|
|
DBUG_PRINT("info", ("Column %s, type: %d, len: %d",
|
|
|
|
field->field_name, field->real_type(), field_len));
|
|
|
|
|
|
|
|
// Define scan filter
|
|
|
|
if (field->real_type() == MYSQL_TYPE_STRING)
|
|
|
|
sf.eq(ndb_fieldnr, key_ptr, field_len);
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (field_len == 8)
|
|
|
|
sf.eq(ndb_fieldnr, (Uint64)*key_ptr);
|
|
|
|
else if (field_len <= 4)
|
|
|
|
sf.eq(ndb_fieldnr, (Uint32)*key_ptr);
|
|
|
|
else
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
key_ptr += field_len;
|
|
|
|
tot_len += field_len;
|
|
|
|
|
|
|
|
if (tot_len >= key_len)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
// End scan filter
|
|
|
|
sf.end();
|
|
|
|
}
|
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(define_read_attrs(buf, op));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2004-05-10 14:12:28 +02:00
|
|
|
Start full table scan in NDB
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::full_table_scan(byte *buf)
|
|
|
|
{
|
|
|
|
uint i;
|
|
|
|
NdbResultSet *cursor;
|
|
|
|
NdbScanOperation *op;
|
2004-05-10 14:12:28 +02:00
|
|
|
NdbConnection *trans= m_active_trans;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("full_table_scan");
|
|
|
|
DBUG_PRINT("enter", ("Starting new scan on %s", m_tabname));
|
|
|
|
|
2004-09-28 19:35:07 +02:00
|
|
|
NdbOperation::LockMode lm=
|
|
|
|
(NdbOperation::LockMode)get_ndb_lock_type(m_lock.type);
|
|
|
|
if (!(op=trans->getNdbScanOperation((const NDBTAB *) m_table)) ||
|
|
|
|
!(cursor= op->readTuples(lm, 0, parallelism)))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
m_active_cursor= cursor;
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(define_read_attrs(buf, op));
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Insert one record into NDB
|
|
|
|
*/
|
|
|
|
int ha_ndbcluster::write_row(byte *record)
|
|
|
|
{
|
2004-08-11 15:06:17 +02:00
|
|
|
bool has_auto_increment;
|
2004-04-15 09:14:14 +02:00
|
|
|
uint i;
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
NdbOperation *op;
|
|
|
|
int res;
|
|
|
|
DBUG_ENTER("write_row");
|
2004-09-15 14:44:21 +02:00
|
|
|
|
2004-11-18 12:11:56 +01:00
|
|
|
if(m_ignore_dup_key && table->primary_key != MAX_KEY)
|
2004-09-15 14:44:21 +02:00
|
|
|
{
|
2004-11-18 12:11:56 +01:00
|
|
|
int peek_res= peek_row();
|
|
|
|
|
|
|
|
if (!peek_res)
|
|
|
|
{
|
|
|
|
m_dupkey= table->primary_key;
|
|
|
|
DBUG_RETURN(HA_ERR_FOUND_DUPP_KEY);
|
|
|
|
}
|
|
|
|
if (peek_res != HA_ERR_KEY_NOT_FOUND)
|
|
|
|
DBUG_RETURN(peek_res);
|
2004-09-15 14:44:21 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
statistic_increment(ha_write_count,&LOCK_status);
|
2004-10-04 00:27:04 +02:00
|
|
|
if (table->timestamp_field_type & TIMESTAMP_AUTO_SET_ON_INSERT)
|
|
|
|
table->timestamp_field->set_time();
|
2004-07-29 11:35:38 +02:00
|
|
|
has_auto_increment= (table->next_number_field && record == table->record[0]);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)))
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
res= (m_use_write) ? op->writeTuple() :op->insertTuple();
|
|
|
|
if (res != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
if (table->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
// Table has hidden primary key
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
Uint64 auto_value= ndb->getAutoIncrementValue((const NDBTAB *) m_table);
|
2004-04-15 09:14:14 +02:00
|
|
|
if (set_hidden_key(op, table->fields, (const byte*)&auto_value))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
2004-08-13 13:51:18 +02:00
|
|
|
|
2004-09-16 04:55:22 +02:00
|
|
|
if (has_auto_increment)
|
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= FALSE;
|
2004-08-13 13:51:18 +02:00
|
|
|
update_auto_increment();
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= !auto_increment_column_changed;
|
2004-09-16 04:55:22 +02:00
|
|
|
}
|
2004-08-13 13:51:18 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if ((res= set_primary_key(op)))
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set non-key attribute(s)
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool set_blob_value= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
for (i= 0; i < table->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
|
|
|
if (!(field->flags & PRI_KEY_FLAG) &&
|
2004-10-01 13:16:49 +02:00
|
|
|
set_ndb_value(op, field, i, &set_blob_value))
|
2004-08-13 13:51:18 +02:00
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(op->getNdbError());
|
2004-08-13 13:51:18 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Execute write operation
|
|
|
|
NOTE When doing inserts with many values in
|
|
|
|
each INSERT statement it should not be necessary
|
|
|
|
to NoCommit the transaction between each row.
|
|
|
|
Find out how this is detected!
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_inserted++;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_update(1);
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_not_flushed= TRUE;
|
|
|
|
if ((m_rows_to_insert == 1) ||
|
|
|
|
((m_rows_inserted % m_bulk_insert_rows) == 0) ||
|
2004-10-01 13:16:49 +02:00
|
|
|
set_blob_value)
|
2004-04-29 14:38:35 +02:00
|
|
|
{
|
2004-09-03 17:11:09 +02:00
|
|
|
THD *thd= current_thd;
|
2004-04-29 14:38:35 +02:00
|
|
|
// Send rows to NDB
|
|
|
|
DBUG_PRINT("info", ("Sending inserts to NDB, "\
|
|
|
|
"rows_inserted:%d, bulk_insert_rows: %d",
|
2004-11-03 15:53:26 +01:00
|
|
|
(int)m_rows_inserted, (int)m_bulk_insert_rows));
|
2004-09-03 17:11:09 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_not_flushed= FALSE;
|
2004-11-17 09:15:53 +01:00
|
|
|
// if (thd->transaction.on)
|
|
|
|
if (m_transaction_on)
|
2004-09-07 14:30:11 +02:00
|
|
|
{
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0)
|
2004-09-07 14:30:11 +02:00
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-09-07 14:30:11 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
|
|
|
}
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
|
|
|
else
|
2004-08-13 13:51:18 +02:00
|
|
|
{
|
2004-09-28 21:11:50 +02:00
|
|
|
if (execute_commit(this,trans) != 0)
|
2004-09-03 17:11:09 +02:00
|
|
|
{
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-09-07 14:30:11 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
2004-10-05 12:04:09 +02:00
|
|
|
int res= trans->restart();
|
|
|
|
DBUG_ASSERT(res == 0);
|
2004-08-13 13:51:18 +02:00
|
|
|
}
|
2004-04-29 14:38:35 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
if ((has_auto_increment) && (m_skip_auto_increment))
|
2004-08-10 14:38:24 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2004-07-29 11:35:38 +02:00
|
|
|
Uint64 next_val= (Uint64) table->next_number_field->val_int() + 1;
|
2004-08-10 14:38:24 +02:00
|
|
|
DBUG_PRINT("info",
|
2004-08-24 17:00:45 +02:00
|
|
|
("Trying to set next auto increment value to %lu",
|
|
|
|
(ulong) next_val));
|
2004-12-30 19:56:09 +01:00
|
|
|
if (ndb->setAutoIncrementValue((const NDBTAB *) m_table, next_val, TRUE))
|
2004-08-10 14:38:24 +02:00
|
|
|
DBUG_PRINT("info",
|
|
|
|
("Setting next auto increment value to %u", next_val));
|
2004-07-29 11:35:38 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
m_skip_auto_increment= TRUE;
|
2004-07-29 11:35:38 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/* Compare if a key in a row has changed */
|
|
|
|
|
|
|
|
int ha_ndbcluster::key_cmp(uint keynr, const byte * old_row,
|
|
|
|
const byte * new_row)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO *key_part=table->key_info[keynr].key_part;
|
|
|
|
KEY_PART_INFO *end=key_part+table->key_info[keynr].key_parts;
|
|
|
|
|
|
|
|
for (; key_part != end ; key_part++)
|
|
|
|
{
|
|
|
|
if (key_part->null_bit)
|
|
|
|
{
|
|
|
|
if ((old_row[key_part->null_offset] & key_part->null_bit) !=
|
|
|
|
(new_row[key_part->null_offset] & key_part->null_bit))
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
if (key_part->key_part_flag & (HA_BLOB_PART | HA_VAR_LENGTH))
|
|
|
|
{
|
|
|
|
|
|
|
|
if (key_part->field->cmp_binary((char*) (old_row + key_part->offset),
|
|
|
|
(char*) (new_row + key_part->offset),
|
|
|
|
(ulong) key_part->length))
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (memcmp(old_row+key_part->offset, new_row+key_part->offset,
|
|
|
|
key_part->length))
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Update one record in NDB using primary key
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::update_row(const byte *old_data, byte *new_data)
|
|
|
|
{
|
|
|
|
THD *thd= current_thd;
|
|
|
|
NdbConnection *trans= m_active_trans;
|
2004-05-10 14:12:28 +02:00
|
|
|
NdbResultSet* cursor= m_active_cursor;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbOperation *op;
|
|
|
|
uint i;
|
|
|
|
DBUG_ENTER("update_row");
|
|
|
|
|
|
|
|
statistic_increment(ha_update_count,&LOCK_status);
|
2004-10-04 00:27:04 +02:00
|
|
|
if (table->timestamp_field_type & TIMESTAMP_AUTO_SET_ON_UPDATE)
|
|
|
|
table->timestamp_field->set_time();
|
|
|
|
|
2004-07-22 18:35:51 +02:00
|
|
|
/* Check for update of primary key for special handling */
|
2004-05-10 14:12:28 +02:00
|
|
|
if ((table->primary_key != MAX_KEY) &&
|
|
|
|
(key_cmp(table->primary_key, old_data, new_data)))
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
2004-07-23 15:46:56 +02:00
|
|
|
int read_res, insert_res, delete_res;
|
2004-07-22 18:35:51 +02:00
|
|
|
|
2004-07-23 15:46:56 +02:00
|
|
|
DBUG_PRINT("info", ("primary key update, doing pk read+insert+delete"));
|
2004-07-22 18:35:51 +02:00
|
|
|
// Get all old fields, since we optimize away fields not in query
|
2004-07-23 15:46:56 +02:00
|
|
|
read_res= complemented_pk_read(old_data, new_data);
|
2004-07-22 18:35:51 +02:00
|
|
|
if (read_res)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("pk read failed"));
|
|
|
|
DBUG_RETURN(read_res);
|
|
|
|
}
|
|
|
|
// Insert new row
|
2004-07-23 15:46:56 +02:00
|
|
|
insert_res= write_row(new_data);
|
|
|
|
if (insert_res)
|
2004-07-22 18:35:51 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("insert failed"));
|
|
|
|
DBUG_RETURN(insert_res);
|
|
|
|
}
|
2004-07-23 15:46:56 +02:00
|
|
|
// Delete old row
|
|
|
|
DBUG_PRINT("info", ("insert succeded"));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_primary_key_update= TRUE;
|
2004-07-23 15:46:56 +02:00
|
|
|
delete_res= delete_row(old_data);
|
2004-11-03 15:53:26 +01:00
|
|
|
m_primary_key_update= FALSE;
|
2004-07-23 15:46:56 +02:00
|
|
|
if (delete_res)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("delete failed"));
|
|
|
|
// Undo write_row(new_data)
|
|
|
|
DBUG_RETURN(delete_row(new_data));
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("insert+delete succeeded"));
|
|
|
|
DBUG_RETURN(0);
|
2004-07-22 18:35:51 +02:00
|
|
|
}
|
2004-07-09 13:18:56 +02:00
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
if (cursor)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
/*
|
|
|
|
We are scanning records and want to update the record
|
|
|
|
that was just found, call updateTuple on the cursor
|
|
|
|
to take over the lock to a new update operation
|
|
|
|
And thus setting the primary key of the record from
|
|
|
|
the active record in cursor
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("info", ("Calling updateTuple on cursor"));
|
|
|
|
if (!(op= cursor->updateTuple()))
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending++;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
if (uses_blob_value(FALSE))
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_pending= TRUE;
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op= trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2004-05-10 14:12:28 +02:00
|
|
|
op->updateTuple() != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
|
|
|
if (table->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
// This table has no primary key, use "hidden" primary key
|
|
|
|
DBUG_PRINT("info", ("Using hidden key"));
|
|
|
|
|
|
|
|
// Require that the PK for this record has previously been
|
|
|
|
// read into m_value
|
|
|
|
uint no_fields= table->fields;
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbRecAttr* rec= m_value[no_fields].rec;
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_ASSERT(rec);
|
|
|
|
DBUG_DUMP("key", (char*)rec->aRef(), NDB_HIDDEN_PRIMARY_KEY_LENGTH);
|
|
|
|
|
|
|
|
if (set_hidden_key(op, no_fields, rec->aRef()))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
2004-05-17 09:51:02 +02:00
|
|
|
if ((res= set_primary_key_from_old_data(op, old_data)))
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(res);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Set non-key attribute(s)
|
|
|
|
for (i= 0; i < table->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= table->field[i];
|
2004-11-17 09:15:53 +01:00
|
|
|
if (((thd->query_id == field->query_id) || m_retrieve_all_fields) &&
|
2004-04-15 09:14:14 +02:00
|
|
|
(!(field->flags & PRI_KEY_FLAG)) &&
|
|
|
|
set_ndb_value(op, field, i))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Execute update operation
|
2004-09-17 16:58:08 +02:00
|
|
|
if (!cursor && execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Delete one record from NDB, using primary key
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::delete_row(const byte *record)
|
|
|
|
{
|
|
|
|
NdbConnection *trans= m_active_trans;
|
2004-05-10 14:12:28 +02:00
|
|
|
NdbResultSet* cursor= m_active_cursor;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbOperation *op;
|
|
|
|
DBUG_ENTER("delete_row");
|
|
|
|
|
|
|
|
statistic_increment(ha_delete_count,&LOCK_status);
|
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
if (cursor)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
/*
|
2004-07-23 10:33:51 +02:00
|
|
|
We are scanning records and want to delete the record
|
2004-05-10 14:12:28 +02:00
|
|
|
that was just found, call deleteTuple on the cursor
|
2004-07-23 13:13:50 +02:00
|
|
|
to take over the lock to a new delete operation
|
2004-05-10 14:12:28 +02:00
|
|
|
And thus setting the primary key of the record from
|
|
|
|
the active record in cursor
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("info", ("Calling deleteTuple on cursor"));
|
|
|
|
if (cursor->deleteTuple() != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending++;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_update(-1);
|
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
// If deleting from cursor, NoCommit will be handled in next_result
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
else
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-09-07 21:53:59 +02:00
|
|
|
if (!(op=trans->getNdbOperation((const NDBTAB *) m_table)) ||
|
2004-05-10 14:12:28 +02:00
|
|
|
op->deleteTuple() != 0)
|
|
|
|
ERR_RETURN(trans->getNdbError());
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_update(-1);
|
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
if (table->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
// This table has no primary key, use "hidden" primary key
|
|
|
|
DBUG_PRINT("info", ("Using hidden key"));
|
|
|
|
uint no_fields= table->fields;
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbRecAttr* rec= m_value[no_fields].rec;
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_ASSERT(rec != NULL);
|
|
|
|
|
|
|
|
if (set_hidden_key(op, no_fields, rec->aRef()))
|
|
|
|
ERR_RETURN(op->getNdbError());
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
int res;
|
2004-11-03 15:53:26 +01:00
|
|
|
if ((res= (m_primary_key_update ?
|
|
|
|
set_primary_key_from_old_data(op, record)
|
|
|
|
: set_primary_key(op))))
|
|
|
|
return res;
|
2004-05-10 14:12:28 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
// Execute delete operation
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Unpack a record read from NDB
|
|
|
|
|
|
|
|
SYNOPSIS
|
|
|
|
unpack_record()
|
|
|
|
buf Buffer to store read row
|
|
|
|
|
|
|
|
NOTE
|
|
|
|
The data for each row is read directly into the
|
|
|
|
destination buffer. This function is primarily
|
|
|
|
called in order to check if any fields should be
|
|
|
|
set to null.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::unpack_record(byte* buf)
|
|
|
|
{
|
|
|
|
uint row_offset= (uint) (buf - table->record[0]);
|
|
|
|
Field **field, **end;
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbValue *value= m_value;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("unpack_record");
|
|
|
|
|
|
|
|
// Set null flag(s)
|
|
|
|
bzero(buf, table->null_bytes);
|
|
|
|
for (field= table->field, end= field+table->fields;
|
|
|
|
field < end;
|
|
|
|
field++, value++)
|
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
if ((*value).ptr)
|
|
|
|
{
|
|
|
|
if (! ((*field)->flags & BLOB_FLAG))
|
|
|
|
{
|
|
|
|
if ((*value).rec->isNULL())
|
|
|
|
(*field)->set_null(row_offset);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
NdbBlob* ndb_blob= (*value).blob;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool isNull= TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
int ret= ndb_blob->getNull(isNull);
|
|
|
|
DBUG_ASSERT(ret == 0);
|
|
|
|
if (isNull)
|
|
|
|
(*field)->set_null(row_offset);
|
|
|
|
}
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
// Read and print all values that was fetched
|
|
|
|
if (table->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
// Table with hidden primary key
|
|
|
|
int hidden_no= table->fields;
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
2004-04-15 09:14:14 +02:00
|
|
|
const NDBCOL *hidden_col= tab->getColumn(hidden_no);
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbRecAttr* rec= m_value[hidden_no].rec;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(rec);
|
|
|
|
DBUG_PRINT("hidden", ("%d: %s \"%llu\"", hidden_no,
|
|
|
|
hidden_col->getName(), rec->u_64_value()));
|
|
|
|
}
|
|
|
|
print_results();
|
|
|
|
#endif
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Utility function to print/dump the fetched field
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::print_results()
|
|
|
|
{
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB*) m_table;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("print_results");
|
|
|
|
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
if (!_db_on_)
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
|
|
|
|
for (uint f=0; f<table->fields;f++)
|
|
|
|
{
|
|
|
|
Field *field;
|
|
|
|
const NDBCOL *col;
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbValue value;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
if (!(value= m_value[f]).ptr)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
fprintf(DBUG_FILE, "Field %d was not read\n", f);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
field= table->field[f];
|
|
|
|
DBUG_DUMP("field->ptr", (char*)field->ptr, field->pack_length());
|
|
|
|
col= tab->getColumn(f);
|
|
|
|
fprintf(DBUG_FILE, "%d: %s\t", f, col->getName());
|
2004-07-22 12:38:09 +02:00
|
|
|
|
|
|
|
NdbBlob *ndb_blob= NULL;
|
|
|
|
if (! (field->flags & BLOB_FLAG))
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
if (value.rec->isNULL())
|
|
|
|
{
|
|
|
|
fprintf(DBUG_FILE, "NULL\n");
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ndb_blob= value.blob;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
bool isNull= TRUE;
|
2004-07-22 12:38:09 +02:00
|
|
|
ndb_blob->getNull(isNull);
|
|
|
|
if (isNull) {
|
|
|
|
fprintf(DBUG_FILE, "NULL\n");
|
|
|
|
continue;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
switch (col->getType()) {
|
|
|
|
case NdbDictionary::Column::Tinyint: {
|
|
|
|
char value= *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Tinyint\t%d", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Tinyunsigned: {
|
|
|
|
unsigned char value= *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Tinyunsigned\t%u", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Smallint: {
|
|
|
|
short value= *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Smallint\t%d", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Smallunsigned: {
|
|
|
|
unsigned short value= *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Smallunsigned\t%u", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Mediumint: {
|
|
|
|
byte value[3];
|
|
|
|
memcpy(value, field->ptr, 3);
|
|
|
|
fprintf(DBUG_FILE, "Mediumint\t%d,%d,%d", value[0], value[1], value[2]);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Mediumunsigned: {
|
|
|
|
byte value[3];
|
|
|
|
memcpy(value, field->ptr, 3);
|
|
|
|
fprintf(DBUG_FILE, "Mediumunsigned\t%u,%u,%u", value[0], value[1], value[2]);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Int: {
|
|
|
|
fprintf(DBUG_FILE, "Int\t%lld", field->val_int());
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Unsigned: {
|
|
|
|
Uint32 value= (Uint32) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Unsigned\t%u", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Bigint: {
|
|
|
|
Int64 value= (Int64) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Bigint\t%lld", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Bigunsigned: {
|
|
|
|
Uint64 value= (Uint64) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Bigunsigned\t%llu", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Float: {
|
|
|
|
float value= (float) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Float\t%f", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Double: {
|
|
|
|
double value= (double) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Double\t%f", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Decimal: {
|
|
|
|
char *value= field->ptr;
|
|
|
|
|
|
|
|
fprintf(DBUG_FILE, "Decimal\t'%-*s'", field->pack_length(), value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Char:{
|
2004-11-13 17:34:34 +01:00
|
|
|
const char *value= (char *) field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Char\t'%.*s'", field->pack_length(), value);
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Varchar:
|
|
|
|
case NdbDictionary::Column::Binary:
|
|
|
|
case NdbDictionary::Column::Varbinary: {
|
2004-11-13 17:34:34 +01:00
|
|
|
const char *value= (char *) field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Var\t'%.*s'", field->pack_length(), value);
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Datetime: {
|
|
|
|
Uint64 value= (Uint64) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Datetime\t%llu", value);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Timespec: {
|
|
|
|
Uint64 value= (Uint64) *field->ptr;
|
|
|
|
fprintf(DBUG_FILE, "Timespec\t%llu", value);
|
|
|
|
break;
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
case NdbDictionary::Column::Blob: {
|
|
|
|
Uint64 len= 0;
|
|
|
|
ndb_blob->getLength(len);
|
|
|
|
fprintf(DBUG_FILE, "Blob\t[len=%u]", (unsigned)len);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Text: {
|
|
|
|
Uint64 len= 0;
|
|
|
|
ndb_blob->getLength(len);
|
|
|
|
fprintf(DBUG_FILE, "Text\t[len=%u]", (unsigned)len);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case NdbDictionary::Column::Undefined:
|
|
|
|
fprintf(DBUG_FILE, "Unknown type: %d", col->getType());
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
fprintf(DBUG_FILE, "\n");
|
|
|
|
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_init(uint index)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("index_init");
|
|
|
|
DBUG_PRINT("enter", ("index: %u", index));
|
|
|
|
DBUG_RETURN(handler::index_init(index));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_end()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("index_end");
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_RETURN(close_scan());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-11-30 07:41:26 +01:00
|
|
|
/**
|
|
|
|
* Check if key contains null
|
|
|
|
*/
|
|
|
|
static
|
|
|
|
int
|
|
|
|
check_null_in_key(const KEY* key_info, const byte *key, uint key_len)
|
|
|
|
{
|
|
|
|
KEY_PART_INFO *curr_part, *end_part;
|
|
|
|
const byte* end_ptr = key + key_len;
|
|
|
|
curr_part= key_info->key_part;
|
|
|
|
end_part= curr_part + key_info->key_parts;
|
|
|
|
|
|
|
|
|
|
|
|
for (; curr_part != end_part && key < end_ptr; curr_part++)
|
|
|
|
{
|
|
|
|
if(curr_part->null_bit && *key)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
key += curr_part->store_length;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::index_read(byte *buf,
|
2004-09-07 16:22:42 +02:00
|
|
|
const byte *key, uint key_len,
|
|
|
|
enum ha_rkey_function find_flag)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("index_read");
|
|
|
|
DBUG_PRINT("enter", ("active_index: %u, key_len: %u, find_flag: %d",
|
|
|
|
active_index, key_len, find_flag));
|
|
|
|
|
2004-10-06 08:48:26 +02:00
|
|
|
int error;
|
|
|
|
ndb_index_type type = get_index_type(active_index);
|
|
|
|
const KEY* key_info = table->key_info+active_index;
|
|
|
|
switch (type){
|
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
|
|
|
case PRIMARY_KEY_INDEX:
|
|
|
|
if (find_flag == HA_READ_KEY_EXACT && key_info->key_length == key_len)
|
|
|
|
{
|
2004-11-30 07:41:26 +01:00
|
|
|
if(m_active_cursor && (error= close_scan()))
|
|
|
|
DBUG_RETURN(error);
|
2004-10-06 08:48:26 +02:00
|
|
|
DBUG_RETURN(pk_read(key, key_len, buf));
|
|
|
|
}
|
|
|
|
else if (type == PRIMARY_KEY_INDEX)
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case UNIQUE_ORDERED_INDEX:
|
|
|
|
case UNIQUE_INDEX:
|
2004-11-30 07:41:26 +01:00
|
|
|
if (find_flag == HA_READ_KEY_EXACT && key_info->key_length == key_len &&
|
|
|
|
!check_null_in_key(key_info, key, key_len))
|
2004-10-06 08:48:26 +02:00
|
|
|
{
|
2004-11-30 07:41:26 +01:00
|
|
|
if(m_active_cursor && (error= close_scan()))
|
|
|
|
DBUG_RETURN(error);
|
2004-10-06 08:48:26 +02:00
|
|
|
DBUG_RETURN(unique_index_read(key, key_len, buf));
|
|
|
|
}
|
|
|
|
else if (type == UNIQUE_INDEX)
|
|
|
|
{
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case ORDERED_INDEX:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
case UNDEFINED_INDEX:
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_ASSERT(FALSE);
|
2004-10-07 14:19:46 +02:00
|
|
|
DBUG_RETURN(1);
|
2004-10-06 08:48:26 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2004-05-17 09:51:02 +02:00
|
|
|
key_range start_key;
|
2004-10-06 08:48:26 +02:00
|
|
|
start_key.key = key;
|
|
|
|
start_key.length = key_len;
|
|
|
|
start_key.flag = find_flag;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
error= ordered_index_scan(&start_key, 0, TRUE, buf);
|
2004-10-06 08:48:26 +02:00
|
|
|
DBUG_RETURN(error == HA_ERR_END_OF_FILE ? HA_ERR_KEY_NOT_FOUND : error);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_read_idx(byte *buf, uint index_no,
|
|
|
|
const byte *key, uint key_len,
|
|
|
|
enum ha_rkey_function find_flag)
|
|
|
|
{
|
|
|
|
statistic_increment(ha_read_key_count,&LOCK_status);
|
|
|
|
DBUG_ENTER("index_read_idx");
|
|
|
|
DBUG_PRINT("enter", ("index_no: %u, key_len: %u", index_no, key_len));
|
|
|
|
index_init(index_no);
|
|
|
|
DBUG_RETURN(index_read(buf, key, key_len, find_flag));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_next(byte *buf)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("index_next");
|
|
|
|
|
2004-07-23 15:46:56 +02:00
|
|
|
int error= 1;
|
2004-04-15 09:14:14 +02:00
|
|
|
statistic_increment(ha_read_next_count,&LOCK_status);
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(next_result(buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_prev(byte *buf)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("index_prev");
|
|
|
|
statistic_increment(ha_read_prev_count,&LOCK_status);
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_first(byte *buf)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("index_first");
|
|
|
|
statistic_increment(ha_read_first_count,&LOCK_status);
|
2004-09-03 13:55:40 +02:00
|
|
|
// Start the ordered index scan and fetch the first row
|
|
|
|
|
|
|
|
// Only HA_READ_ORDER indexes get called by index_first
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(ordered_index_scan(0, 0, TRUE, buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::index_last(byte *buf)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("index_last");
|
|
|
|
statistic_increment(ha_read_last_count,&LOCK_status);
|
2004-09-03 13:55:40 +02:00
|
|
|
int res;
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
if((res= ordered_index_scan(0, 0, TRUE, buf)) == 0){
|
2004-09-03 13:55:40 +02:00
|
|
|
NdbResultSet *cursor= m_active_cursor;
|
2004-11-22 14:41:46 +01:00
|
|
|
while((res= cursor->nextResult(TRUE, m_force_send)) == 0);
|
2004-09-03 13:55:40 +02:00
|
|
|
if(res == 1){
|
|
|
|
unpack_record(buf);
|
|
|
|
table->status= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
}
|
2004-11-10 11:02:51 +01:00
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-07 16:22:42 +02:00
|
|
|
inline
|
|
|
|
int ha_ndbcluster::read_range_first_to_buf(const key_range *start_key,
|
|
|
|
const key_range *end_key,
|
|
|
|
bool eq_range, bool sorted,
|
|
|
|
byte* buf)
|
|
|
|
{
|
|
|
|
KEY* key_info;
|
|
|
|
int error= 1;
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_range_first_to_buf");
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("info", ("eq_range: %d, sorted: %d", eq_range, sorted));
|
2004-04-30 13:38:41 +02:00
|
|
|
|
2004-05-10 14:46:06 +02:00
|
|
|
switch (get_index_type(active_index)){
|
2004-05-24 12:35:39 +02:00
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
2004-04-30 13:38:41 +02:00
|
|
|
case PRIMARY_KEY_INDEX:
|
2004-05-10 14:46:06 +02:00
|
|
|
key_info= table->key_info + active_index;
|
|
|
|
if (start_key &&
|
|
|
|
start_key->length == key_info->key_length &&
|
|
|
|
start_key->flag == HA_READ_KEY_EXACT)
|
2004-05-17 09:51:02 +02:00
|
|
|
{
|
2004-11-30 07:41:26 +01:00
|
|
|
if(m_active_cursor && (error= close_scan()))
|
|
|
|
DBUG_RETURN(error);
|
2004-05-17 09:51:02 +02:00
|
|
|
error= pk_read(start_key->key, start_key->length, buf);
|
|
|
|
DBUG_RETURN(error == HA_ERR_KEY_NOT_FOUND ? HA_ERR_END_OF_FILE : error);
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
break;
|
2004-05-24 12:35:39 +02:00
|
|
|
case UNIQUE_ORDERED_INDEX:
|
2004-04-30 13:38:41 +02:00
|
|
|
case UNIQUE_INDEX:
|
2004-05-10 14:46:06 +02:00
|
|
|
key_info= table->key_info + active_index;
|
2004-11-30 07:41:26 +01:00
|
|
|
if (start_key && start_key->length == key_info->key_length &&
|
|
|
|
start_key->flag == HA_READ_KEY_EXACT &&
|
|
|
|
!check_null_in_key(key_info, start_key->key, start_key->length))
|
2004-05-17 09:51:02 +02:00
|
|
|
{
|
2004-11-30 07:41:26 +01:00
|
|
|
if(m_active_cursor && (error= close_scan()))
|
|
|
|
DBUG_RETURN(error);
|
2004-05-17 09:51:02 +02:00
|
|
|
error= unique_index_read(start_key->key, start_key->length, buf);
|
|
|
|
DBUG_RETURN(error == HA_ERR_KEY_NOT_FOUND ? HA_ERR_END_OF_FILE : error);
|
|
|
|
}
|
2004-04-30 13:38:41 +02:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-05-10 14:46:06 +02:00
|
|
|
|
|
|
|
// Start the ordered index scan and fetch the first row
|
2004-10-05 12:17:42 +02:00
|
|
|
error= ordered_index_scan(start_key, end_key, sorted, buf);
|
2004-04-30 13:38:41 +02:00
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-10-21 12:39:08 +02:00
|
|
|
int ha_ndbcluster::read_range_first(const key_range *start_key,
|
|
|
|
const key_range *end_key,
|
|
|
|
bool eq_range, bool sorted)
|
|
|
|
{
|
|
|
|
byte* buf= table->record[0];
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_range_first");
|
|
|
|
|
|
|
|
DBUG_RETURN(read_range_first_to_buf(start_key,
|
|
|
|
end_key,
|
|
|
|
eq_range,
|
|
|
|
sorted,
|
|
|
|
buf));
|
|
|
|
}
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
int ha_ndbcluster::read_range_next()
|
2004-04-30 13:38:41 +02:00
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::read_range_next");
|
|
|
|
DBUG_RETURN(next_result(table->record[0]));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
int ha_ndbcluster::rnd_init(bool scan)
|
|
|
|
{
|
|
|
|
NdbResultSet *cursor= m_active_cursor;
|
|
|
|
DBUG_ENTER("rnd_init");
|
|
|
|
DBUG_PRINT("enter", ("scan: %d", scan));
|
2004-07-09 13:18:56 +02:00
|
|
|
// Check if scan is to be restarted
|
2004-08-24 23:13:32 +02:00
|
|
|
if (cursor)
|
|
|
|
{
|
|
|
|
if (!scan)
|
|
|
|
DBUG_RETURN(1);
|
2004-11-22 14:41:46 +01:00
|
|
|
int res= cursor->restart(m_force_send);
|
2004-10-05 12:04:09 +02:00
|
|
|
DBUG_ASSERT(res == 0);
|
2004-08-24 23:13:32 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
index_init(table->primary_key);
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-05-11 13:59:22 +02:00
|
|
|
int ha_ndbcluster::close_scan()
|
|
|
|
{
|
|
|
|
NdbResultSet *cursor= m_active_cursor;
|
2004-08-04 10:54:42 +02:00
|
|
|
NdbConnection *trans= m_active_trans;
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_ENTER("close_scan");
|
|
|
|
|
|
|
|
if (!cursor)
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
|
2004-08-04 10:54:42 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_ops_pending)
|
2004-08-04 10:54:42 +02:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
Take over any pending transactions to the
|
|
|
|
deleteing/updating transaction before closing the scan
|
|
|
|
*/
|
2004-11-03 15:53:26 +01:00
|
|
|
DBUG_PRINT("info", ("ops_pending: %d", m_ops_pending));
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-08-04 10:54:42 +02:00
|
|
|
DBUG_RETURN(ndb_err(trans));
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-08-04 10:54:42 +02:00
|
|
|
}
|
|
|
|
|
2004-11-22 14:41:46 +01:00
|
|
|
cursor->close(m_force_send);
|
2004-05-11 13:59:22 +02:00
|
|
|
m_active_cursor= NULL;
|
2004-05-28 11:23:44 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-05-11 13:59:22 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::rnd_end()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("rnd_end");
|
2004-05-11 13:59:22 +02:00
|
|
|
DBUG_RETURN(close_scan());
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::rnd_next(byte *buf)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("rnd_next");
|
|
|
|
statistic_increment(ha_read_rnd_next_count, &LOCK_status);
|
2004-05-10 14:12:28 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (!m_active_cursor)
|
2004-05-10 14:12:28 +02:00
|
|
|
DBUG_RETURN(full_table_scan(buf));
|
|
|
|
DBUG_RETURN(next_result(buf));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
An "interesting" record has been found and it's pk
|
|
|
|
retrieved by calling position
|
|
|
|
Now it's time to read the record from db once
|
|
|
|
again
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::rnd_pos(byte *buf, byte *pos)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("rnd_pos");
|
|
|
|
statistic_increment(ha_read_rnd_count,&LOCK_status);
|
|
|
|
// The primary key for the record is stored in pos
|
|
|
|
// Perform a pk_read using primary key "index"
|
|
|
|
DBUG_RETURN(pk_read(pos, ref_length, buf));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Store the primary key of this record in ref
|
|
|
|
variable, so that the row can be retrieved again later
|
|
|
|
using "reference" in rnd_pos
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::position(const byte *record)
|
|
|
|
{
|
|
|
|
KEY *key_info;
|
|
|
|
KEY_PART_INFO *key_part;
|
|
|
|
KEY_PART_INFO *end;
|
|
|
|
byte *buff;
|
|
|
|
DBUG_ENTER("position");
|
|
|
|
|
|
|
|
if (table->primary_key != MAX_KEY)
|
|
|
|
{
|
|
|
|
key_info= table->key_info + table->primary_key;
|
|
|
|
key_part= key_info->key_part;
|
|
|
|
end= key_part + key_info->key_parts;
|
|
|
|
buff= ref;
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
if (key_part->null_bit) {
|
|
|
|
/* Store 0 if the key part is a NULL part */
|
|
|
|
if (record[key_part->null_offset]
|
|
|
|
& key_part->null_bit) {
|
|
|
|
*buff++= 1;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
*buff++= 0;
|
|
|
|
}
|
|
|
|
memcpy(buff, record + key_part->offset, key_part->length);
|
|
|
|
buff += key_part->length;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// No primary key, get hidden key
|
|
|
|
DBUG_PRINT("info", ("Getting hidden key"));
|
|
|
|
int hidden_no= table->fields;
|
2004-07-22 12:38:09 +02:00
|
|
|
NdbRecAttr* rec= m_value[hidden_no].rec;
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
2004-04-15 09:14:14 +02:00
|
|
|
const NDBCOL *hidden_col= tab->getColumn(hidden_no);
|
|
|
|
DBUG_ASSERT(hidden_col->getPrimaryKey() &&
|
|
|
|
hidden_col->getAutoIncrement() &&
|
|
|
|
rec != NULL &&
|
|
|
|
ref_length == NDB_HIDDEN_PRIMARY_KEY_LENGTH);
|
|
|
|
memcpy(ref, (const void*)rec->aRef(), ref_length);
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_DUMP("ref", (char*)ref, ref_length);
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void ha_ndbcluster::info(uint flag)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("info");
|
|
|
|
DBUG_PRINT("enter", ("flag: %d", flag));
|
|
|
|
|
|
|
|
if (flag & HA_STATUS_POS)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_POS"));
|
|
|
|
if (flag & HA_STATUS_NO_LOCK)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_NO_LOCK"));
|
|
|
|
if (flag & HA_STATUS_TIME)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_TIME"));
|
|
|
|
if (flag & HA_STATUS_VARIABLE)
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("info", ("HA_STATUS_VARIABLE"));
|
2004-09-20 17:52:19 +02:00
|
|
|
if (m_table_info)
|
|
|
|
{
|
2004-11-17 09:15:53 +01:00
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
records= 100;
|
|
|
|
else
|
|
|
|
records_update();
|
2004-09-20 17:52:19 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
if ((my_errno= check_ndb_connection()))
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
Ndb *ndb= get_ndb();
|
2004-11-17 09:15:53 +01:00
|
|
|
Uint64 rows= 100;
|
|
|
|
if (current_thd->variables.ndb_use_exact_count)
|
2004-12-30 21:36:16 +01:00
|
|
|
ndb_get_table_statistics(ndb, m_tabname, &rows, 0);
|
2004-11-17 09:15:53 +01:00
|
|
|
records= rows;
|
2004-09-20 17:52:19 +02:00
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
}
|
2004-10-04 01:20:05 +02:00
|
|
|
if (flag & HA_STATUS_CONST)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_CONST"));
|
|
|
|
set_rec_per_key();
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
if (flag & HA_STATUS_ERRKEY)
|
2004-08-31 12:07:48 +02:00
|
|
|
{
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("info", ("HA_STATUS_ERRKEY"));
|
2004-11-03 15:53:26 +01:00
|
|
|
errkey= m_dupkey;
|
2004-08-31 12:07:48 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
if (flag & HA_STATUS_AUTO)
|
|
|
|
DBUG_PRINT("info", ("HA_STATUS_AUTO"));
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::extra(enum ha_extra_function operation)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("extra");
|
|
|
|
switch (operation) {
|
|
|
|
case HA_EXTRA_NORMAL: /* Optimize for space (def) */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NORMAL"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_QUICK: /* Optimize for speed */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_QUICK"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_RESET: /* Reset database to after open */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_RESET"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_CACHE: /* Cash record in HA_rrnd() */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_CACHE: /* End cacheing of records (def) */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_READCHECK: /* No readcheck on update */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_READCHECK"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_READCHECK: /* Use readcheck (def) */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_READCHECK"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_KEYREAD: /* Read only key to database */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_KEYREAD"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_KEYREAD: /* Normal read of records (def) */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_KEYREAD"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_USER_CHANGE: /* No user is allowed to write */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_USER_CHANGE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_KEY_CACHE:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_KEY_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_KEY_CACHE:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_KEY_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_WAIT_LOCK: /* Wait until file is avalably (def) */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_WAIT_LOCK"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_WAIT_LOCK: /* If file is locked, return quickly */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_WAIT_LOCK"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_WRITE_CACHE: /* Use write cache in ha_write() */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_WRITE_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_FLUSH_CACHE: /* flush write_record_cache */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_FLUSH_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_KEYS: /* Remove all update of keys */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_KEYS"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_KEYREAD_CHANGE_POS: /* Keyread, but change pos */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_KEYREAD_CHANGE_POS")); /* xxxxchk -r must be used */
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_REMEMBER_POS: /* Remember pos for next/prev */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_REMEMBER_POS"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_RESTORE_POS:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_RESTORE_POS"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_REINIT_CACHE: /* init cache from current record */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_REINIT_CACHE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_FORCE_REOPEN: /* Datafile have changed on disk */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_FORCE_REOPEN"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_FLUSH: /* Flush tables to disk */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_FLUSH"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_ROWS: /* Don't write rows */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_ROWS"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_RESET_STATE: /* Reset positions */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_RESET_STATE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_IGNORE_DUP_KEY: /* Dup keys don't rollback everything*/
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_IGNORE_DUP_KEY"));
|
2004-09-15 14:44:21 +02:00
|
|
|
if (current_thd->lex->sql_command == SQLCOM_REPLACE)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Turning ON use of write instead of insert"));
|
|
|
|
m_use_write= TRUE;
|
|
|
|
} else
|
|
|
|
{
|
2004-11-18 12:11:56 +01:00
|
|
|
DBUG_PRINT("info", ("Ignoring duplicate key"));
|
|
|
|
m_ignore_dup_key= TRUE;
|
2004-09-15 14:44:21 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case HA_EXTRA_NO_IGNORE_DUP_KEY:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_NO_IGNORE_DUP_KEY"));
|
|
|
|
DBUG_PRINT("info", ("Turning OFF use of write instead of insert"));
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
m_use_write= FALSE;
|
2004-11-18 12:11:56 +01:00
|
|
|
m_ignore_dup_key= FALSE;
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case HA_EXTRA_RETRIEVE_ALL_COLS: /* Retrieve all columns, not just those
|
|
|
|
where field->query_id is the same as
|
|
|
|
the current query id */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_RETRIEVE_ALL_COLS"));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_retrieve_all_fields= TRUE;
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case HA_EXTRA_PREPARE_FOR_DELETE:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_PREPARE_FOR_DELETE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_PREPARE_FOR_UPDATE: /* Remove read cache if problems */
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_PREPARE_FOR_UPDATE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_PRELOAD_BUFFER_SIZE:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_PRELOAD_BUFFER_SIZE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_RETRIEVE_PRIMARY_KEY:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_RETRIEVE_PRIMARY_KEY"));
|
2004-11-18 12:11:56 +01:00
|
|
|
m_retrieve_primary_key= TRUE;
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case HA_EXTRA_CHANGE_KEY_TO_UNIQUE:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_CHANGE_KEY_TO_UNIQUE"));
|
|
|
|
break;
|
|
|
|
case HA_EXTRA_CHANGE_KEY_TO_DUP:
|
|
|
|
DBUG_PRINT("info", ("HA_EXTRA_CHANGE_KEY_TO_DUP"));
|
|
|
|
break;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-04-29 14:38:35 +02:00
|
|
|
/*
|
|
|
|
Start of an insert, remember number of rows to be inserted, it will
|
|
|
|
be used in write_row and get_autoincrement to send an optimal number
|
|
|
|
of rows in each roundtrip to the server
|
|
|
|
|
|
|
|
SYNOPSIS
|
|
|
|
rows number of rows to insert, 0 if unknown
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::start_bulk_insert(ha_rows rows)
|
|
|
|
{
|
|
|
|
int bytes, batch;
|
2004-09-07 21:53:59 +02:00
|
|
|
const NDBTAB *tab= (const NDBTAB *) m_table;
|
2004-04-29 14:38:35 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("start_bulk_insert");
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("enter", ("rows: %d", (int)rows));
|
2004-04-29 14:38:35 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_inserted= 0;
|
|
|
|
m_rows_to_insert= rows;
|
2004-04-29 14:38:35 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
Calculate how many rows that should be inserted
|
|
|
|
per roundtrip to NDB. This is done in order to minimize the
|
|
|
|
number of roundtrips as much as possible. However performance will
|
|
|
|
degrade if too many bytes are inserted, thus it's limited by this
|
|
|
|
calculation.
|
|
|
|
*/
|
2004-07-23 15:46:56 +02:00
|
|
|
const int bytesperbatch= 8192;
|
2004-04-29 14:38:35 +02:00
|
|
|
bytes= 12 + tab->getRowSizeInBytes() + 4 * tab->getNoOfColumns();
|
2004-05-10 14:46:06 +02:00
|
|
|
batch= bytesperbatch/bytes;
|
2004-04-29 14:38:35 +02:00
|
|
|
batch= batch == 0 ? 1 : batch;
|
|
|
|
DBUG_PRINT("info", ("batch: %d, bytes: %d", batch, bytes));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_bulk_insert_rows= batch;
|
2004-04-29 14:38:35 +02:00
|
|
|
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
End of an insert
|
|
|
|
*/
|
|
|
|
int ha_ndbcluster::end_bulk_insert()
|
|
|
|
{
|
2004-07-23 15:46:56 +02:00
|
|
|
int error= 0;
|
|
|
|
|
2004-04-29 14:38:35 +02:00
|
|
|
DBUG_ENTER("end_bulk_insert");
|
2004-07-23 15:46:56 +02:00
|
|
|
// Check if last inserts need to be flushed
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_bulk_insert_not_flushed)
|
2004-07-23 15:46:56 +02:00
|
|
|
{
|
|
|
|
NdbConnection *trans= m_active_trans;
|
|
|
|
// Send rows to NDB
|
|
|
|
DBUG_PRINT("info", ("Sending inserts to NDB, "\
|
|
|
|
"rows_inserted:%d, bulk_insert_rows: %d",
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_inserted, m_bulk_insert_rows));
|
|
|
|
m_bulk_insert_not_flushed= FALSE;
|
2004-09-17 16:58:08 +02:00
|
|
|
if (execute_no_commit(this,trans) != 0) {
|
2004-09-14 17:51:57 +02:00
|
|
|
no_uncommitted_rows_execute_failure();
|
2004-09-07 12:45:19 +02:00
|
|
|
my_errno= error= ndb_err(trans);
|
2004-09-14 17:51:57 +02:00
|
|
|
}
|
2004-07-23 15:46:56 +02:00
|
|
|
}
|
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_inserted= 0;
|
|
|
|
m_rows_to_insert= 1;
|
2004-07-23 15:46:56 +02:00
|
|
|
DBUG_RETURN(error);
|
2004-04-29 14:38:35 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
int ha_ndbcluster::extra_opt(enum ha_extra_function operation, ulong cache_size)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("extra_opt");
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("enter", ("cache_size: %lu", cache_size));
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(extra(operation));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ha_ndbcluster::reset()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("reset");
|
|
|
|
// Reset what?
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
2005-01-14 14:33:26 +01:00
|
|
|
static const char *ha_ndb_bas_ext[]= { ha_ndb_ext, NullS };
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2005-01-14 14:33:26 +01:00
|
|
|
const char**
|
|
|
|
ha_ndbcluster::bas_ext() const
|
|
|
|
{
|
|
|
|
return ha_ndb_bas_ext;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
How many seeks it will take to read through the table
|
|
|
|
This is to be comparable to the number returned by records_in_range so
|
|
|
|
that we can decide if we should scan the table or use keys.
|
|
|
|
*/
|
|
|
|
|
|
|
|
double ha_ndbcluster::scan_time()
|
|
|
|
{
|
2004-09-03 13:55:40 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster::scan_time()");
|
|
|
|
double res= rows2double(records*1000);
|
|
|
|
DBUG_PRINT("exit", ("table: %s value: %f",
|
|
|
|
m_tabname, res));
|
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
THR_LOCK_DATA **ha_ndbcluster::store_lock(THD *thd,
|
|
|
|
THR_LOCK_DATA **to,
|
|
|
|
enum thr_lock_type lock_type)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("store_lock");
|
|
|
|
if (lock_type != TL_IGNORE && m_lock.type == TL_UNLOCK)
|
|
|
|
{
|
2004-09-28 21:11:50 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/* If we are not doing a LOCK TABLE, then allow multiple
|
|
|
|
writers */
|
|
|
|
|
2004-10-13 10:08:18 +02:00
|
|
|
/* Since NDB does not currently have table locks
|
|
|
|
this is treated as a ordinary lock */
|
|
|
|
|
2004-09-28 21:11:50 +02:00
|
|
|
if ((lock_type >= TL_WRITE_ALLOW_WRITE &&
|
2004-04-15 09:14:14 +02:00
|
|
|
lock_type <= TL_WRITE) && !thd->in_lock_tables)
|
|
|
|
lock_type= TL_WRITE_ALLOW_WRITE;
|
|
|
|
|
|
|
|
/* In queries of type INSERT INTO t1 SELECT ... FROM t2 ...
|
|
|
|
MySQL would use the lock TL_READ_NO_INSERT on t2, and that
|
|
|
|
would conflict with TL_WRITE_ALLOW_WRITE, blocking all inserts
|
|
|
|
to t2. Convert the lock to a normal read lock to allow
|
|
|
|
concurrent inserts to t2. */
|
|
|
|
|
|
|
|
if (lock_type == TL_READ_NO_INSERT && !thd->in_lock_tables)
|
|
|
|
lock_type= TL_READ;
|
|
|
|
|
|
|
|
m_lock.type=lock_type;
|
|
|
|
}
|
|
|
|
*to++= &m_lock;
|
2004-05-10 14:12:28 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("exit", ("lock_type: %d", lock_type));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(to);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef DBUG_OFF
|
|
|
|
#define PRINT_OPTION_FLAGS(t) { \
|
|
|
|
if (t->options & OPTION_NOT_AUTOCOMMIT) \
|
|
|
|
DBUG_PRINT("thd->options", ("OPTION_NOT_AUTOCOMMIT")); \
|
|
|
|
if (t->options & OPTION_BEGIN) \
|
|
|
|
DBUG_PRINT("thd->options", ("OPTION_BEGIN")); \
|
|
|
|
if (t->options & OPTION_TABLE_LOCK) \
|
|
|
|
DBUG_PRINT("thd->options", ("OPTION_TABLE_LOCK")); \
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define PRINT_OPTION_FLAGS(t)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
As MySQL will execute an external lock for every new table it uses
|
|
|
|
we can use this to start the transactions.
|
|
|
|
If we are in auto_commit mode we just need to start a transaction
|
|
|
|
for the statement, this will be stored in transaction.stmt.
|
|
|
|
If not, we have to start a master transaction if there doesn't exist
|
|
|
|
one from before, this will be stored in transaction.all
|
|
|
|
|
|
|
|
When a table lock is held one transaction will be started which holds
|
|
|
|
the table lock and for each statement a hupp transaction will be started
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::external_lock(THD *thd, int lock_type)
|
|
|
|
{
|
|
|
|
int error=0;
|
|
|
|
NdbConnection* trans= NULL;
|
|
|
|
|
|
|
|
DBUG_ENTER("external_lock");
|
|
|
|
/*
|
|
|
|
Check that this handler instance has a connection
|
|
|
|
set up to the Ndb object of thd
|
|
|
|
*/
|
|
|
|
if (check_ndb_connection())
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Thd_ndb *thd_ndb= (Thd_ndb*)thd->transaction.thd_ndb;
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= thd_ndb->ndb;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("enter", ("transaction.thd_ndb->lock_count: %d",
|
|
|
|
thd_ndb->lock_count));
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
if (lock_type != F_UNLCK)
|
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("info", ("lock_type != F_UNLCK"));
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
if (!thd_ndb->lock_count++)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
PRINT_OPTION_FLAGS(thd);
|
|
|
|
|
|
|
|
if (!(thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN | OPTION_TABLE_LOCK)))
|
|
|
|
{
|
|
|
|
// Autocommit transaction
|
|
|
|
DBUG_ASSERT(!thd->transaction.stmt.ndb_tid);
|
|
|
|
DBUG_PRINT("trans",("Starting transaction stmt"));
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
trans= ndb->startTransaction();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (trans == NULL)
|
2004-12-30 19:56:09 +01:00
|
|
|
ERR_RETURN(ndb->getNdbError());
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_reset(thd);
|
2004-04-15 09:14:14 +02:00
|
|
|
thd->transaction.stmt.ndb_tid= trans;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (!thd->transaction.all.ndb_tid)
|
|
|
|
{
|
|
|
|
// Not autocommit transaction
|
|
|
|
// A "master" transaction ha not been started yet
|
|
|
|
DBUG_PRINT("trans",("starting transaction, all"));
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
trans= ndb->startTransaction();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (trans == NULL)
|
2004-12-30 19:56:09 +01:00
|
|
|
ERR_RETURN(ndb->getNdbError());
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_reset(thd);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
If this is the start of a LOCK TABLE, a table look
|
|
|
|
should be taken on the table in NDB
|
|
|
|
|
|
|
|
Check if it should be read or write lock
|
|
|
|
*/
|
|
|
|
if (thd->options & (OPTION_TABLE_LOCK))
|
|
|
|
{
|
|
|
|
//lockThisTable();
|
|
|
|
DBUG_PRINT("info", ("Locking the table..." ));
|
|
|
|
}
|
|
|
|
|
|
|
|
thd->transaction.all.ndb_tid= trans;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
This is the place to make sure this handler instance
|
|
|
|
has a started transaction.
|
|
|
|
|
|
|
|
The transaction is started by the first handler on which
|
|
|
|
MySQL Server calls external lock
|
|
|
|
|
|
|
|
Other handlers in the same stmt or transaction should use
|
|
|
|
the same NDB transaction. This is done by setting up the m_active_trans
|
|
|
|
pointer to point to the NDB transaction.
|
|
|
|
*/
|
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
// store thread specific data first to set the right context
|
|
|
|
m_force_send= thd->variables.ndb_force_send;
|
|
|
|
m_ha_not_exact_count= !thd->variables.ndb_use_exact_count;
|
|
|
|
m_autoincrement_prefetch= thd->variables.ndb_autoincrement_prefetch_sz;
|
|
|
|
if (!thd->transaction.on)
|
|
|
|
m_transaction_on= FALSE;
|
|
|
|
else
|
|
|
|
m_transaction_on= thd->variables.ndb_use_transactions;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
m_active_trans= thd->transaction.all.ndb_tid ?
|
|
|
|
(NdbConnection*)thd->transaction.all.ndb_tid:
|
|
|
|
(NdbConnection*)thd->transaction.stmt.ndb_tid;
|
|
|
|
DBUG_ASSERT(m_active_trans);
|
2004-05-10 14:12:28 +02:00
|
|
|
// Start of transaction
|
2004-11-03 15:53:26 +01:00
|
|
|
m_retrieve_all_fields= FALSE;
|
2004-11-18 12:11:56 +01:00
|
|
|
m_retrieve_primary_key= FALSE;
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-09-20 17:52:19 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-09-20 17:52:19 +02:00
|
|
|
const NDBTAB *tab;
|
|
|
|
void *tab_info;
|
|
|
|
if (!(tab= dict->getTable(m_tabname, &tab_info)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
DBUG_PRINT("info", ("Table schema version: %d", tab->getObjectVersion()));
|
|
|
|
m_table= (void *)tab;
|
|
|
|
m_table_info= tab_info;
|
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_init(thd);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_PRINT("info", ("lock_type == F_UNLCK"));
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
if (!--thd_ndb->lock_count)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("trans", ("Last external_lock"));
|
|
|
|
PRINT_OPTION_FLAGS(thd);
|
|
|
|
|
|
|
|
if (thd->transaction.stmt.ndb_tid)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Unlock is done without a transaction commit / rollback.
|
|
|
|
This happens if the thread didn't update any rows
|
|
|
|
We must in this case close the transaction to release resources
|
|
|
|
*/
|
|
|
|
DBUG_PRINT("trans",("ending non-updating transaction"));
|
2004-12-30 19:56:09 +01:00
|
|
|
ndb->closeTransaction(m_active_trans);
|
2004-04-15 09:14:14 +02:00
|
|
|
thd->transaction.stmt.ndb_tid= 0;
|
|
|
|
}
|
|
|
|
}
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table= NULL;
|
|
|
|
m_table_info= NULL;
|
2004-09-13 14:46:38 +02:00
|
|
|
/*
|
|
|
|
This is the place to make sure this handler instance
|
|
|
|
no longer are connected to the active transaction.
|
|
|
|
|
|
|
|
And since the handler is no longer part of the transaction
|
|
|
|
it can't have open cursors, ops or blobs pending.
|
|
|
|
*/
|
|
|
|
m_active_trans= NULL;
|
|
|
|
|
2004-09-07 12:54:31 +02:00
|
|
|
if (m_active_cursor)
|
|
|
|
DBUG_PRINT("warning", ("m_active_cursor != NULL"));
|
2004-09-13 14:46:38 +02:00
|
|
|
m_active_cursor= NULL;
|
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_blobs_pending)
|
2004-09-07 12:54:31 +02:00
|
|
|
DBUG_PRINT("warning", ("blobs_pending != 0"));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_blobs_pending= 0;
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2004-11-03 15:53:26 +01:00
|
|
|
if (m_ops_pending)
|
2004-09-07 12:54:31 +02:00
|
|
|
DBUG_PRINT("warning", ("ops_pending != 0L"));
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
When using LOCK TABLE's external_lock is only called when the actual
|
|
|
|
TABLE LOCK is done.
|
|
|
|
Under LOCK TABLES, each used tables will force a call to start_stmt.
|
2004-08-30 12:13:10 +02:00
|
|
|
Ndb doesn't currently support table locks, and will do ordinary
|
|
|
|
startTransaction for each transaction/statement.
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::start_stmt(THD *thd)
|
|
|
|
{
|
|
|
|
int error=0;
|
|
|
|
DBUG_ENTER("start_stmt");
|
|
|
|
PRINT_OPTION_FLAGS(thd);
|
|
|
|
|
|
|
|
NdbConnection *trans= (NdbConnection*)thd->transaction.stmt.ndb_tid;
|
|
|
|
if (!trans){
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= ((Thd_ndb*)thd->transaction.thd_ndb)->ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_PRINT("trans",("Starting transaction stmt"));
|
|
|
|
|
|
|
|
NdbConnection *tablock_trans=
|
|
|
|
(NdbConnection*)thd->transaction.all.ndb_tid;
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("info", ("tablock_trans: %x", (uint)tablock_trans));
|
2004-08-30 12:13:10 +02:00
|
|
|
DBUG_ASSERT(tablock_trans);
|
2004-12-30 19:56:09 +01:00
|
|
|
// trans= ndb->hupp(tablock_trans);
|
|
|
|
trans= ndb->startTransaction();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (trans == NULL)
|
2004-12-30 19:56:09 +01:00
|
|
|
ERR_RETURN(ndb->getNdbError());
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
no_uncommitted_rows_reset(thd);
|
2004-04-15 09:14:14 +02:00
|
|
|
thd->transaction.stmt.ndb_tid= trans;
|
|
|
|
}
|
|
|
|
m_active_trans= trans;
|
2004-05-05 16:04:23 +02:00
|
|
|
|
2004-05-10 14:12:28 +02:00
|
|
|
// Start of statement
|
2004-11-03 15:53:26 +01:00
|
|
|
m_retrieve_all_fields= FALSE;
|
2004-11-18 12:11:56 +01:00
|
|
|
m_retrieve_primary_key= FALSE;
|
2004-11-03 15:53:26 +01:00
|
|
|
m_ops_pending= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Commit a transaction started in NDB
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ndbcluster_commit(THD *thd, void *ndb_transaction)
|
|
|
|
{
|
|
|
|
int res= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Ndb *ndb= ((Thd_ndb*)thd->transaction.thd_ndb)->ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbConnection *trans= (NdbConnection*)ndb_transaction;
|
|
|
|
|
|
|
|
DBUG_ENTER("ndbcluster_commit");
|
|
|
|
DBUG_PRINT("transaction",("%s",
|
|
|
|
trans == thd->transaction.stmt.ndb_tid ?
|
|
|
|
"stmt" : "all"));
|
|
|
|
DBUG_ASSERT(ndb && trans);
|
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
if (execute_commit(thd,trans) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
const NdbError err= trans->getNdbError();
|
2004-07-02 16:14:08 +02:00
|
|
|
const NdbOperation *error_op= trans->getNdbErrorOperation();
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_PRINT(err);
|
|
|
|
res= ndb_to_mysql_error(&err);
|
2004-08-31 12:07:48 +02:00
|
|
|
if (res != -1)
|
2004-07-02 16:14:08 +02:00
|
|
|
ndbcluster_print_error(res, error_op);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
ndb->closeTransaction(trans);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(res);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Rollback a transaction started in NDB
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ndbcluster_rollback(THD *thd, void *ndb_transaction)
|
|
|
|
{
|
|
|
|
int res= 0;
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Ndb *ndb= ((Thd_ndb*)thd->transaction.thd_ndb)->ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbConnection *trans= (NdbConnection*)ndb_transaction;
|
|
|
|
|
|
|
|
DBUG_ENTER("ndbcluster_rollback");
|
|
|
|
DBUG_PRINT("transaction",("%s",
|
|
|
|
trans == thd->transaction.stmt.ndb_tid ?
|
|
|
|
"stmt" : "all"));
|
|
|
|
DBUG_ASSERT(ndb && trans);
|
|
|
|
|
|
|
|
if (trans->execute(Rollback) != 0)
|
|
|
|
{
|
|
|
|
const NdbError err= trans->getNdbError();
|
2004-07-02 16:14:08 +02:00
|
|
|
const NdbOperation *error_op= trans->getNdbErrorOperation();
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_PRINT(err);
|
|
|
|
res= ndb_to_mysql_error(&err);
|
2004-07-02 16:14:08 +02:00
|
|
|
if (res != -1)
|
|
|
|
ndbcluster_print_error(res, error_op);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
ndb->closeTransaction(trans);
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2004-07-22 12:38:09 +02:00
|
|
|
Define NDB column based on Field.
|
|
|
|
Returns 0 or mysql error code.
|
|
|
|
Not member of ha_ndbcluster because NDBCOL cannot be declared.
|
2004-04-15 09:14:14 +02:00
|
|
|
*/
|
|
|
|
|
2004-07-22 12:38:09 +02:00
|
|
|
static int create_ndb_column(NDBCOL &col,
|
|
|
|
Field *field,
|
|
|
|
HA_CREATE_INFO *info)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-07-22 12:38:09 +02:00
|
|
|
// Set name
|
2004-10-28 17:35:06 +02:00
|
|
|
{
|
|
|
|
char truncated_field_name[NDB_MAX_ATTR_NAME_SIZE];
|
|
|
|
strnmov(truncated_field_name,field->field_name,sizeof(truncated_field_name));
|
|
|
|
truncated_field_name[sizeof(truncated_field_name)-1]= '\0';
|
|
|
|
col.setName(truncated_field_name);
|
|
|
|
}
|
2004-09-15 17:44:13 +02:00
|
|
|
// Get char set
|
|
|
|
CHARSET_INFO *cs= field->charset();
|
2004-07-22 12:38:09 +02:00
|
|
|
// Set type and sizes
|
|
|
|
const enum enum_field_types mysql_type= field->real_type();
|
|
|
|
switch (mysql_type) {
|
|
|
|
// Numeric types
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_DECIMAL:
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_TINY:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Tinyunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Tinyint);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_SHORT:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Smallunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Smallint);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_LONG:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Unsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Int);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_INT24:
|
2004-07-22 12:38:09 +02:00
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Mediumunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Mediumint);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_LONGLONG:
|
|
|
|
if (field->flags & UNSIGNED_FLAG)
|
|
|
|
col.setType(NDBCOL::Bigunsigned);
|
|
|
|
else
|
|
|
|
col.setType(NDBCOL::Bigint);
|
|
|
|
col.setLength(1);
|
2004-04-15 09:14:14 +02:00
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_FLOAT:
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Float);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
2004-04-15 09:14:14 +02:00
|
|
|
case MYSQL_TYPE_DOUBLE:
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Double);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
|
|
|
// Date types
|
|
|
|
case MYSQL_TYPE_TIMESTAMP:
|
|
|
|
col.setType(NDBCOL::Unsigned);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_DATETIME:
|
|
|
|
col.setType(NDBCOL::Datetime);
|
|
|
|
col.setLength(1);
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_DATE:
|
|
|
|
case MYSQL_TYPE_NEWDATE:
|
|
|
|
case MYSQL_TYPE_TIME:
|
|
|
|
case MYSQL_TYPE_YEAR:
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
|
|
|
// Char types
|
|
|
|
case MYSQL_TYPE_STRING:
|
|
|
|
if (field->flags & BINARY_FLAG)
|
|
|
|
col.setType(NDBCOL::Binary);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Char);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-10-20 19:22:58 +02:00
|
|
|
if (field->pack_length() == 0)
|
|
|
|
col.setLength(1); // currently ndb does not support size 0
|
|
|
|
else
|
|
|
|
col.setLength(field->pack_length());
|
2004-07-22 12:38:09 +02:00
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_VAR_STRING:
|
|
|
|
if (field->flags & BINARY_FLAG)
|
|
|
|
col.setType(NDBCOL::Varbinary);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Varchar);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
|
|
|
// Blob types (all come in as MYSQL_TYPE_BLOB)
|
|
|
|
mysql_type_tiny_blob:
|
|
|
|
case MYSQL_TYPE_TINY_BLOB:
|
|
|
|
if (field->flags & BINARY_FLAG)
|
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setInlineSize(256);
|
|
|
|
// No parts
|
|
|
|
col.setPartSize(0);
|
|
|
|
col.setStripeSize(0);
|
|
|
|
break;
|
|
|
|
mysql_type_blob:
|
|
|
|
case MYSQL_TYPE_BLOB:
|
|
|
|
if (field->flags & BINARY_FLAG)
|
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
// Use "<=" even if "<" is the exact condition
|
|
|
|
if (field->max_length() <= (1 << 8))
|
|
|
|
goto mysql_type_tiny_blob;
|
|
|
|
else if (field->max_length() <= (1 << 16))
|
|
|
|
{
|
|
|
|
col.setInlineSize(256);
|
|
|
|
col.setPartSize(2000);
|
|
|
|
col.setStripeSize(16);
|
|
|
|
}
|
|
|
|
else if (field->max_length() <= (1 << 24))
|
|
|
|
goto mysql_type_medium_blob;
|
|
|
|
else
|
|
|
|
goto mysql_type_long_blob;
|
|
|
|
break;
|
|
|
|
mysql_type_medium_blob:
|
|
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
|
|
|
if (field->flags & BINARY_FLAG)
|
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setInlineSize(256);
|
|
|
|
col.setPartSize(4000);
|
|
|
|
col.setStripeSize(8);
|
|
|
|
break;
|
|
|
|
mysql_type_long_blob:
|
|
|
|
case MYSQL_TYPE_LONG_BLOB:
|
|
|
|
if (field->flags & BINARY_FLAG)
|
|
|
|
col.setType(NDBCOL::Blob);
|
2004-09-15 17:44:13 +02:00
|
|
|
else {
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setType(NDBCOL::Text);
|
2004-09-15 17:44:13 +02:00
|
|
|
col.setCharset(cs);
|
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
col.setInlineSize(256);
|
|
|
|
col.setPartSize(8000);
|
|
|
|
col.setStripeSize(4);
|
|
|
|
break;
|
|
|
|
// Other types
|
|
|
|
case MYSQL_TYPE_ENUM:
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_SET:
|
|
|
|
col.setType(NDBCOL::Char);
|
|
|
|
col.setLength(field->pack_length());
|
|
|
|
break;
|
|
|
|
case MYSQL_TYPE_NULL:
|
|
|
|
case MYSQL_TYPE_GEOMETRY:
|
|
|
|
goto mysql_type_unsupported;
|
|
|
|
mysql_type_unsupported:
|
|
|
|
default:
|
|
|
|
return HA_ERR_UNSUPPORTED;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
// Set nullable and pk
|
|
|
|
col.setNullable(field->maybe_null());
|
|
|
|
col.setPrimaryKey(field->flags & PRI_KEY_FLAG);
|
|
|
|
// Set autoincrement
|
|
|
|
if (field->flags & AUTO_INCREMENT_FLAG)
|
|
|
|
{
|
|
|
|
col.setAutoIncrement(TRUE);
|
|
|
|
ulonglong value= info->auto_increment_value ?
|
2004-09-02 17:51:44 +02:00
|
|
|
info->auto_increment_value : (ulonglong) 1;
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_PRINT("info", ("Autoincrement key, initial: %llu", value));
|
|
|
|
col.setAutoIncrementInitialValue(value);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-07-22 12:38:09 +02:00
|
|
|
else
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
col.setAutoIncrement(FALSE);
|
2004-07-22 12:38:09 +02:00
|
|
|
return 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Create a table in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::create(const char *name,
|
|
|
|
TABLE *form,
|
|
|
|
HA_CREATE_INFO *info)
|
|
|
|
{
|
|
|
|
NDBTAB tab;
|
|
|
|
NDBCOL col;
|
2004-11-08 11:06:36 +01:00
|
|
|
uint pack_length, length, i, pk_length= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
const void *data, *pack_data;
|
2004-04-30 12:25:31 +02:00
|
|
|
const char **key_names= form->keynames.type_names;
|
2004-04-15 09:14:14 +02:00
|
|
|
char name2[FN_HEADLEN];
|
2004-09-13 14:46:38 +02:00
|
|
|
bool create_from_engine= (info->table_options & HA_CREATE_FROM_ENGINE);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("create");
|
|
|
|
DBUG_PRINT("enter", ("name: %s", name));
|
|
|
|
fn_format(name2, name, "", "",2); // Remove the .frm extension
|
|
|
|
set_dbname(name2);
|
2004-09-13 14:46:38 +02:00
|
|
|
set_tabname(name2);
|
|
|
|
|
|
|
|
if (create_from_engine)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
Table alreay exists in NDB and frm file has been created by
|
|
|
|
caller.
|
|
|
|
Do Ndb specific stuff, such as create a .ndb file
|
|
|
|
*/
|
|
|
|
my_errno= write_ndb_file();
|
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("table", ("name: %s", m_tabname));
|
|
|
|
tab.setName(m_tabname);
|
|
|
|
tab.setLogging(!(info->options & HA_LEX_CREATE_TMP_TABLE));
|
|
|
|
|
|
|
|
// Save frm data for this table
|
|
|
|
if (readfrm(name, &data, &length))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
if (packfrm(data, length, &pack_data, &pack_length))
|
|
|
|
DBUG_RETURN(2);
|
|
|
|
|
|
|
|
DBUG_PRINT("info", ("setFrm data=%x, len=%d", pack_data, pack_length));
|
|
|
|
tab.setFrm(pack_data, pack_length);
|
|
|
|
my_free((char*)data, MYF(0));
|
|
|
|
my_free((char*)pack_data, MYF(0));
|
|
|
|
|
|
|
|
for (i= 0; i < form->fields; i++)
|
|
|
|
{
|
|
|
|
Field *field= form->field[i];
|
|
|
|
DBUG_PRINT("info", ("name: %s, type: %u, pack_length: %d",
|
|
|
|
field->field_name, field->real_type(),
|
|
|
|
field->pack_length()));
|
2004-07-23 15:46:56 +02:00
|
|
|
if ((my_errno= create_ndb_column(col, field, info)))
|
2004-07-22 12:38:09 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
2004-04-15 09:14:14 +02:00
|
|
|
tab.addColumn(col);
|
2004-11-08 11:06:36 +01:00
|
|
|
if(col.getPrimaryKey())
|
|
|
|
pk_length += (field->pack_length() + 3) / 4;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// No primary key, create shadow key as 64 bit, auto increment
|
|
|
|
if (form->primary_key == MAX_KEY)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Generating shadow key"));
|
|
|
|
col.setName("$PK");
|
|
|
|
col.setType(NdbDictionary::Column::Bigunsigned);
|
|
|
|
col.setLength(1);
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
col.setNullable(FALSE);
|
2004-04-15 09:14:14 +02:00
|
|
|
col.setPrimaryKey(TRUE);
|
|
|
|
col.setAutoIncrement(TRUE);
|
|
|
|
tab.addColumn(col);
|
2004-11-08 11:06:36 +01:00
|
|
|
pk_length += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure that blob tables don't have to big part size
|
|
|
|
for (i= 0; i < form->fields; i++)
|
|
|
|
{
|
|
|
|
/**
|
|
|
|
* The extra +7 concists
|
|
|
|
* 2 - words from pk in blob table
|
|
|
|
* 5 - from extra words added by tup/dict??
|
|
|
|
*/
|
|
|
|
switch (form->field[i]->real_type()) {
|
|
|
|
case MYSQL_TYPE_BLOB:
|
|
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
|
|
|
case MYSQL_TYPE_LONG_BLOB:
|
|
|
|
{
|
|
|
|
NdbDictionary::Column * col = tab.getColumn(i);
|
|
|
|
int size = pk_length + (col->getPartSize()+3)/4 + 7;
|
|
|
|
if(size > NDB_MAX_TUPLE_SIZE_IN_WORDS &&
|
|
|
|
(pk_length+7) < NDB_MAX_TUPLE_SIZE_IN_WORDS)
|
|
|
|
{
|
|
|
|
size = NDB_MAX_TUPLE_SIZE_IN_WORDS - pk_length - 7;
|
|
|
|
col->setPartSize(4*size);
|
|
|
|
}
|
|
|
|
/**
|
|
|
|
* If size > NDB_MAX and pk_length+7 >= NDB_MAX
|
|
|
|
* then the table can't be created anyway, so skip
|
|
|
|
* changing part size, and have error later
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if ((my_errno= check_ndb_connection()))
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
|
|
|
|
// Create the table in NDB
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-09-13 14:46:38 +02:00
|
|
|
if (dict->createTable(tab) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
|
|
|
ERR_PRINT(err);
|
|
|
|
my_errno= ndb_to_mysql_error(&err);
|
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
}
|
|
|
|
DBUG_PRINT("info", ("Table %s/%s created successfully",
|
|
|
|
m_dbname, m_tabname));
|
2004-04-30 12:25:31 +02:00
|
|
|
|
2004-08-19 11:10:35 +02:00
|
|
|
// Create secondary indexes
|
|
|
|
my_errno= build_index_list(form, ILBP_CREATE);
|
2004-08-18 19:13:39 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if (!my_errno)
|
|
|
|
my_errno= write_ndb_file();
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(my_errno);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-30 12:25:31 +02:00
|
|
|
int ha_ndbcluster::create_ordered_index(const char *name,
|
|
|
|
KEY *key_info)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("create_ordered_index");
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(create_index(name, key_info, FALSE));
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
int ha_ndbcluster::create_unique_index(const char *name,
|
|
|
|
KEY *key_info)
|
|
|
|
{
|
|
|
|
|
2004-05-24 12:35:39 +02:00
|
|
|
DBUG_ENTER("create_unique_index");
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(create_index(name, key_info, TRUE));
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
Create an index in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::create_index(const char *name,
|
2004-04-30 12:25:31 +02:00
|
|
|
KEY *key_info,
|
|
|
|
bool unique)
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NdbDictionary::Dictionary *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
KEY_PART_INFO *key_part= key_info->key_part;
|
|
|
|
KEY_PART_INFO *end= key_part + key_info->key_parts;
|
|
|
|
|
|
|
|
DBUG_ENTER("create_index");
|
|
|
|
DBUG_PRINT("enter", ("name: %s ", name));
|
2004-04-30 12:25:31 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbDictionary::Index ndb_index(name);
|
2004-04-30 12:25:31 +02:00
|
|
|
if (unique)
|
2004-04-15 09:14:14 +02:00
|
|
|
ndb_index.setType(NdbDictionary::Index::UniqueHashIndex);
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ndb_index.setType(NdbDictionary::Index::OrderedIndex);
|
|
|
|
// TODO Only temporary ordered indexes supported
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
ndb_index.setLogging(FALSE);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
ndb_index.setTable(m_tabname);
|
|
|
|
|
|
|
|
for (; key_part != end; key_part++)
|
|
|
|
{
|
|
|
|
Field *field= key_part->field;
|
|
|
|
DBUG_PRINT("info", ("attr: %s", field->field_name));
|
2004-10-28 17:35:06 +02:00
|
|
|
{
|
|
|
|
char truncated_field_name[NDB_MAX_ATTR_NAME_SIZE];
|
|
|
|
strnmov(truncated_field_name,field->field_name,sizeof(truncated_field_name));
|
|
|
|
truncated_field_name[sizeof(truncated_field_name)-1]= '\0';
|
|
|
|
ndb_index.addColumnName(truncated_field_name);
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
if (dict->createIndex(ndb_index))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
|
|
|
|
// Success
|
|
|
|
DBUG_PRINT("info", ("Created index %s", name));
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Rename a table in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::rename_table(const char *from, const char *to)
|
|
|
|
{
|
2004-12-06 14:51:10 +01:00
|
|
|
NDBDICT *dict;
|
2004-04-15 09:14:14 +02:00
|
|
|
char new_tabname[FN_HEADLEN];
|
2004-12-06 14:51:10 +01:00
|
|
|
const NDBTAB *orig_tab;
|
|
|
|
int result;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("ha_ndbcluster::rename_table");
|
2004-12-06 14:51:10 +01:00
|
|
|
DBUG_PRINT("info", ("Renaming %s to %s", from, to));
|
2004-04-15 09:14:14 +02:00
|
|
|
set_dbname(from);
|
|
|
|
set_tabname(from);
|
|
|
|
set_tabname(to, new_tabname);
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if (check_ndb_connection())
|
|
|
|
DBUG_RETURN(my_errno= HA_ERR_NO_CONNECTION);
|
2005-01-01 22:40:40 +01:00
|
|
|
|
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
dict= ndb->getDictionary();
|
2004-12-06 14:51:10 +01:00
|
|
|
if (!(orig_tab= dict->getTable(m_tabname)))
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-12-06 14:51:10 +01:00
|
|
|
m_table= (void *)orig_tab;
|
|
|
|
// Change current database to that of target table
|
|
|
|
set_dbname(to);
|
2005-01-01 22:40:40 +01:00
|
|
|
ndb->setDatabaseName(m_dbname);
|
2004-12-06 14:51:10 +01:00
|
|
|
if (!(result= alter_table_name(new_tabname)))
|
2004-09-13 14:46:38 +02:00
|
|
|
{
|
2004-12-06 14:51:10 +01:00
|
|
|
// Rename .ndb file
|
|
|
|
result= handler::rename_table(from, to);
|
2004-09-13 14:46:38 +02:00
|
|
|
}
|
2004-12-06 14:51:10 +01:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(result);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Rename a table in NDB Cluster using alter table
|
|
|
|
*/
|
|
|
|
|
2004-12-06 14:51:10 +01:00
|
|
|
int ha_ndbcluster::alter_table_name(const char *to)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-12-06 14:51:10 +01:00
|
|
|
const NDBTAB *orig_tab= (const NDBTAB *) m_table;
|
|
|
|
int ret;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("alter_table_name_table");
|
|
|
|
|
2004-12-07 10:27:22 +01:00
|
|
|
NdbDictionary::Table new_tab= *orig_tab;
|
2004-12-06 14:51:10 +01:00
|
|
|
new_tab.setName(to);
|
|
|
|
if (dict->alterTable(new_tab) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
|
|
|
|
m_table= NULL;
|
2004-09-21 14:52:56 +02:00
|
|
|
m_table_info= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Delete a table from NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::delete_table(const char *name)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("delete_table");
|
|
|
|
DBUG_PRINT("enter", ("name: %s", name));
|
|
|
|
set_dbname(name);
|
|
|
|
set_tabname(name);
|
|
|
|
|
|
|
|
if (check_ndb_connection())
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2004-12-06 14:51:10 +01:00
|
|
|
// Remove .ndb file
|
2004-09-13 14:46:38 +02:00
|
|
|
handler::delete_table(name);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(drop_table());
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Drop a table in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::drop_table()
|
|
|
|
{
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
NdbDictionary::Dictionary *dict= ndb->getDictionary();
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("drop_table");
|
|
|
|
DBUG_PRINT("enter", ("Deleting %s", m_tabname));
|
|
|
|
|
|
|
|
if (dict->dropTable(m_tabname))
|
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
|
|
|
if (err.code == 709)
|
|
|
|
; // 709: No such table existed
|
|
|
|
else
|
|
|
|
ERR_RETURN(dict->getNdbError());
|
|
|
|
}
|
|
|
|
release_metadata();
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Drop a database in NDB Cluster
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ndbcluster_drop_database(const char *path)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_drop_database");
|
|
|
|
// TODO drop all tables for this database
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
longlong ha_ndbcluster::get_auto_increment()
|
2004-04-29 14:38:35 +02:00
|
|
|
{
|
2004-07-02 11:50:28 +02:00
|
|
|
DBUG_ENTER("get_auto_increment");
|
|
|
|
DBUG_PRINT("enter", ("m_tabname: %s", m_tabname));
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
2004-07-23 15:46:56 +02:00
|
|
|
int cache_size=
|
2004-11-17 09:15:53 +01:00
|
|
|
(m_rows_to_insert - m_rows_inserted < m_autoincrement_prefetch) ?
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_to_insert - m_rows_inserted
|
2004-11-17 09:15:53 +01:00
|
|
|
: (m_rows_to_insert > m_autoincrement_prefetch) ?
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_to_insert
|
2004-11-17 09:15:53 +01:00
|
|
|
: m_autoincrement_prefetch;
|
2004-04-29 14:38:35 +02:00
|
|
|
Uint64 auto_value=
|
2004-11-03 15:53:26 +01:00
|
|
|
(m_skip_auto_increment) ?
|
2004-12-30 19:56:09 +01:00
|
|
|
ndb->readAutoIncrementValue((const NDBTAB *) m_table)
|
|
|
|
: ndb->getAutoIncrementValue((const NDBTAB *) m_table, cache_size);
|
2004-07-02 11:50:28 +02:00
|
|
|
DBUG_RETURN((longlong)auto_value);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Constructor for the NDB Cluster table handler
|
|
|
|
*/
|
|
|
|
|
|
|
|
ha_ndbcluster::ha_ndbcluster(TABLE *table_arg):
|
|
|
|
handler(table_arg),
|
|
|
|
m_active_trans(NULL),
|
|
|
|
m_active_cursor(NULL),
|
|
|
|
m_table(NULL),
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
m_table_info(NULL),
|
2004-04-15 09:14:14 +02:00
|
|
|
m_table_flags(HA_REC_NOT_IN_SEQ |
|
2004-08-31 10:19:10 +02:00
|
|
|
HA_NULL_IN_KEY |
|
2004-09-15 17:31:15 +02:00
|
|
|
HA_AUTO_PART_KEY |
|
|
|
|
HA_NO_PREFIX_CHAR_KEYS),
|
2004-09-06 20:30:57 +02:00
|
|
|
m_share(0),
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
m_use_write(FALSE),
|
2004-11-18 12:11:56 +01:00
|
|
|
m_ignore_dup_key(FALSE),
|
2004-11-03 15:53:26 +01:00
|
|
|
m_primary_key_update(FALSE),
|
|
|
|
m_retrieve_all_fields(FALSE),
|
2004-11-18 12:11:56 +01:00
|
|
|
m_retrieve_primary_key(FALSE),
|
2004-11-03 15:53:26 +01:00
|
|
|
m_rows_to_insert(1),
|
|
|
|
m_rows_inserted(0),
|
|
|
|
m_bulk_insert_rows(1024),
|
|
|
|
m_bulk_insert_not_flushed(FALSE),
|
|
|
|
m_ops_pending(0),
|
|
|
|
m_skip_auto_increment(TRUE),
|
|
|
|
m_blobs_pending(0),
|
|
|
|
m_blobs_buffer(0),
|
|
|
|
m_blobs_buffer_size(0),
|
2004-11-17 09:15:53 +01:00
|
|
|
m_dupkey((uint) -1),
|
|
|
|
m_ha_not_exact_count(FALSE),
|
|
|
|
m_force_send(TRUE),
|
|
|
|
m_autoincrement_prefetch(32),
|
2004-11-25 19:56:10 +01:00
|
|
|
m_transaction_on(TRUE)
|
2004-11-23 17:38:00 +01:00
|
|
|
{
|
2004-04-30 12:25:31 +02:00
|
|
|
int i;
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ha_ndbcluster");
|
|
|
|
|
|
|
|
m_tabname[0]= '\0';
|
|
|
|
m_dbname[0]= '\0';
|
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
records= ~(ha_rows)0; // uninitialized
|
2004-04-15 09:14:14 +02:00
|
|
|
block_size= 1024;
|
|
|
|
|
2004-04-30 12:25:31 +02:00
|
|
|
for (i= 0; i < MAX_KEY; i++)
|
|
|
|
{
|
2004-08-18 19:13:39 +02:00
|
|
|
m_index[i].type= UNDEFINED_INDEX;
|
|
|
|
m_index[i].unique_index= NULL;
|
|
|
|
m_index[i].index= NULL;
|
2004-04-30 12:25:31 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Destructor for NDB Cluster table handler
|
|
|
|
*/
|
|
|
|
|
|
|
|
ha_ndbcluster::~ha_ndbcluster()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("~ha_ndbcluster");
|
|
|
|
|
2004-09-03 17:11:09 +02:00
|
|
|
if (m_share)
|
|
|
|
free_share(m_share);
|
2004-04-15 09:14:14 +02:00
|
|
|
release_metadata();
|
2004-11-03 15:53:26 +01:00
|
|
|
my_free(m_blobs_buffer, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
m_blobs_buffer= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
// Check for open cursor/transaction
|
2004-09-07 12:54:31 +02:00
|
|
|
if (m_active_cursor) {
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(m_active_cursor == NULL);
|
2004-09-07 12:54:31 +02:00
|
|
|
if (m_active_trans) {
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ASSERT(m_active_trans == NULL);
|
|
|
|
|
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Open a table for further use
|
|
|
|
- fetch metadata for this table from NDB
|
|
|
|
- check that table exists
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::open(const char *name, int mode, uint test_if_locked)
|
|
|
|
{
|
2004-10-04 01:20:05 +02:00
|
|
|
int res;
|
2004-04-15 09:14:14 +02:00
|
|
|
KEY *key;
|
|
|
|
DBUG_ENTER("open");
|
|
|
|
DBUG_PRINT("enter", ("name: %s mode: %d test_if_locked: %d",
|
|
|
|
name, mode, test_if_locked));
|
|
|
|
|
|
|
|
// Setup ref_length to make room for the whole
|
|
|
|
// primary key to be written in the ref variable
|
|
|
|
|
|
|
|
if (table->primary_key != MAX_KEY)
|
|
|
|
{
|
|
|
|
key= table->key_info+table->primary_key;
|
|
|
|
ref_length= key->key_length;
|
|
|
|
DBUG_PRINT("info", (" ref_length: %d", ref_length));
|
|
|
|
}
|
|
|
|
// Init table lock structure
|
|
|
|
if (!(m_share=get_share(name)))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
thr_lock_data_init(&m_share->lock,&m_lock,(void*) 0);
|
|
|
|
|
|
|
|
set_dbname(name);
|
|
|
|
set_tabname(name);
|
|
|
|
|
2004-09-03 17:11:09 +02:00
|
|
|
if (check_ndb_connection()) {
|
|
|
|
free_share(m_share); m_share= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2004-09-03 17:11:09 +02:00
|
|
|
}
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2004-10-04 01:20:05 +02:00
|
|
|
res= get_metadata(name);
|
|
|
|
if (!res)
|
|
|
|
info(HA_STATUS_VARIABLE | HA_STATUS_CONST);
|
|
|
|
|
|
|
|
DBUG_RETURN(res);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Close the table
|
|
|
|
- release resources setup by open()
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::close(void)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("close");
|
2004-09-03 17:11:09 +02:00
|
|
|
free_share(m_share); m_share= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
release_metadata();
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
Thd_ndb* ha_ndbcluster::seize_thd_ndb()
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
Thd_ndb *thd_ndb;
|
|
|
|
DBUG_ENTER("seize_thd_ndb");
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
thd_ndb= new Thd_ndb();
|
|
|
|
thd_ndb->ndb->getDictionary()->set_local_table_data_size(sizeof(Ndb_table_local_info));
|
|
|
|
if (thd_ndb->ndb->init(max_transactions) != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
ERR_PRINT(thd_ndb->ndb->getNdbError());
|
2004-04-15 09:14:14 +02:00
|
|
|
/*
|
|
|
|
TODO
|
|
|
|
Alt.1 If init fails because to many allocated Ndb
|
|
|
|
wait on condition for a Ndb object to be released.
|
|
|
|
Alt.2 Seize/release from pool, wait until next release
|
|
|
|
*/
|
2004-09-14 14:47:34 +02:00
|
|
|
delete thd_ndb;
|
|
|
|
thd_ndb= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-09-14 14:47:34 +02:00
|
|
|
DBUG_RETURN(thd_ndb);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
void ha_ndbcluster::release_thd_ndb(Thd_ndb* thd_ndb)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
DBUG_ENTER("release_thd_ndb");
|
|
|
|
delete thd_ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2004-09-20 17:34:32 +02:00
|
|
|
If this thread already has a Thd_ndb object allocated
|
2004-04-15 09:14:14 +02:00
|
|
|
in current THD, reuse it. Otherwise
|
2004-09-20 17:34:32 +02:00
|
|
|
seize a Thd_ndb object, assign it to current THD and use it.
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
*/
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
Ndb* check_ndb_in_thd(THD* thd)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_ENTER("check_ndb_in_thd");
|
2004-09-14 14:47:34 +02:00
|
|
|
Thd_ndb *thd_ndb= (Thd_ndb*)thd->transaction.thd_ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-14 14:47:34 +02:00
|
|
|
if (!thd_ndb)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-09-20 17:34:32 +02:00
|
|
|
if (!(thd_ndb= ha_ndbcluster::seize_thd_ndb()))
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_RETURN(NULL);
|
2004-09-14 14:47:34 +02:00
|
|
|
thd->transaction.thd_ndb= thd_ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-09-20 17:34:32 +02:00
|
|
|
DBUG_RETURN(thd_ndb->ndb);
|
2004-09-13 14:46:38 +02:00
|
|
|
}
|
|
|
|
|
2004-09-20 17:34:32 +02:00
|
|
|
|
2004-12-06 14:51:10 +01:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
int ha_ndbcluster::check_ndb_connection()
|
|
|
|
{
|
|
|
|
THD* thd= current_thd;
|
2004-12-30 19:56:09 +01:00
|
|
|
Ndb *ndb;
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_ENTER("check_ndb_connection");
|
|
|
|
|
2004-12-30 19:56:09 +01:00
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
2004-12-30 19:56:09 +01:00
|
|
|
ndb->setDatabaseName(m_dbname);
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-09-20 17:34:32 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
void ndbcluster_close_connection(THD *thd)
|
|
|
|
{
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
Thd_ndb *thd_ndb= (Thd_ndb*)thd->transaction.thd_ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndbcluster_close_connection");
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
if (thd_ndb)
|
|
|
|
{
|
2004-09-14 14:47:34 +02:00
|
|
|
ha_ndbcluster::release_thd_ndb(thd_ndb);
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
thd->transaction.thd_ndb= NULL;
|
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_VOID_RETURN;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Try to discover one table from NDB
|
|
|
|
*/
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
int ndbcluster_discover(THD* thd, const char *db, const char *name,
|
2004-04-15 09:14:14 +02:00
|
|
|
const void** frmblob, uint* frmlen)
|
|
|
|
{
|
|
|
|
uint len;
|
|
|
|
const void* data;
|
|
|
|
const NDBTAB* tab;
|
2004-09-13 14:46:38 +02:00
|
|
|
Ndb* ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndbcluster_discover");
|
2004-09-13 14:46:38 +02:00
|
|
|
DBUG_PRINT("enter", ("db: %s, name: %s", db, name));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
|
|
|
ndb->setDatabaseName(db);
|
2004-09-06 20:30:57 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
NDBDICT* dict= ndb->getDictionary();
|
2004-09-20 17:34:32 +02:00
|
|
|
dict->set_local_table_data_size(sizeof(Ndb_table_local_info));
|
2004-09-13 14:46:38 +02:00
|
|
|
dict->invalidateTable(name);
|
|
|
|
if (!(tab= dict->getTable(name)))
|
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
|
|
|
if (err.code == 709)
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
ERR_RETURN(err);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_PRINT("info", ("Found table %s", tab->getName()));
|
|
|
|
|
|
|
|
len= tab->getFrmLength();
|
|
|
|
if (len == 0 || tab->getFrmData() == NULL)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("No frm data found",
|
|
|
|
("Table is probably created via NdbApi"));
|
|
|
|
DBUG_RETURN(2);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (unpackfrm(&data, &len, tab->getFrmData()))
|
|
|
|
DBUG_RETURN(3);
|
|
|
|
|
|
|
|
*frmlen= len;
|
|
|
|
*frmblob= data;
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2004-09-13 14:46:38 +02:00
|
|
|
Check if a table exists in NDB
|
|
|
|
|
|
|
|
*/
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
int ndbcluster_table_exists(THD* thd, const char *db, const char *name)
|
|
|
|
{
|
|
|
|
uint len;
|
|
|
|
const void* data;
|
|
|
|
const NDBTAB* tab;
|
|
|
|
Ndb* ndb;
|
|
|
|
DBUG_ENTER("ndbcluster_table_exists");
|
|
|
|
DBUG_PRINT("enter", ("db: %s, name: %s", db, name));
|
|
|
|
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
|
|
|
ndb->setDatabaseName(db);
|
|
|
|
|
|
|
|
NDBDICT* dict= ndb->getDictionary();
|
2004-09-20 17:34:32 +02:00
|
|
|
dict->set_local_table_data_size(sizeof(Ndb_table_local_info));
|
2004-09-13 14:46:38 +02:00
|
|
|
dict->invalidateTable(name);
|
|
|
|
if (!(tab= dict->getTable(name)))
|
|
|
|
{
|
|
|
|
const NdbError err= dict->getNdbError();
|
|
|
|
if (err.code == 709)
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
ERR_RETURN(err);
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_PRINT("info", ("Found table %s", tab->getName()));
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
extern "C" byte* tables_get_key(const char *entry, uint *length,
|
|
|
|
my_bool not_used __attribute__((unused)))
|
2004-09-21 12:13:58 +02:00
|
|
|
{
|
|
|
|
*length= strlen(entry);
|
|
|
|
return (byte*) entry;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int ndbcluster_find_files(THD *thd,const char *db,const char *path,
|
2004-09-26 16:11:24 +02:00
|
|
|
const char *wild, bool dir, List<char> *files)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
2004-10-25 22:35:46 +02:00
|
|
|
DBUG_ENTER("ndbcluster_find_files");
|
|
|
|
DBUG_PRINT("enter", ("db: %s", db));
|
|
|
|
{ // extra bracket to avoid gcc 2.95.3 warning
|
2004-04-15 09:14:14 +02:00
|
|
|
uint i;
|
2004-09-13 14:46:38 +02:00
|
|
|
Ndb* ndb;
|
2004-09-21 12:13:58 +02:00
|
|
|
char name[FN_REFLEN];
|
2004-09-26 16:11:24 +02:00
|
|
|
HASH ndb_tables, ok_tables;
|
2004-04-15 09:14:14 +02:00
|
|
|
NdbDictionary::Dictionary::List list;
|
2004-09-13 14:46:38 +02:00
|
|
|
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
|
|
|
|
2004-09-21 12:13:58 +02:00
|
|
|
if (dir)
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_RETURN(0); // Discover of databases not yet supported
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// List tables in NDB
|
2004-09-13 14:46:38 +02:00
|
|
|
NDBDICT *dict= ndb->getDictionary();
|
2004-04-15 09:14:14 +02:00
|
|
|
if (dict->listObjects(list,
|
|
|
|
NdbDictionary::Object::UserTable) != 0)
|
2004-09-13 06:40:25 +02:00
|
|
|
ERR_RETURN(dict->getNdbError());
|
2004-09-13 14:46:38 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
if (hash_init(&ndb_tables, system_charset_info,list.count,0,0,
|
|
|
|
(hash_get_key)tables_get_key,0,0))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Failed to init HASH ndb_tables"));
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (hash_init(&ok_tables, system_charset_info,32,0,0,
|
|
|
|
(hash_get_key)tables_get_key,0,0))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("error", ("Failed to init HASH ok_tables"));
|
|
|
|
hash_free(&ndb_tables);
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
for (i= 0 ; i < list.count ; i++)
|
|
|
|
{
|
|
|
|
NdbDictionary::Dictionary::List::Element& t= list.elements[i];
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_PRINT("info", ("Found %s/%s in NDB", t.database, t.name));
|
2004-04-15 09:14:14 +02:00
|
|
|
|
2004-10-08 13:37:13 +02:00
|
|
|
// Add only tables that belongs to db
|
|
|
|
if (my_strcasecmp(system_charset_info, t.database, db))
|
|
|
|
continue;
|
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Apply wildcard to list of tables in NDB
|
2004-09-21 12:13:58 +02:00
|
|
|
if (wild)
|
2004-09-13 14:46:38 +02:00
|
|
|
{
|
2004-09-21 12:13:58 +02:00
|
|
|
if (lower_case_table_names)
|
|
|
|
{
|
|
|
|
if (wild_case_compare(files_charset_info, t.name, wild))
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
else if (wild_compare(t.name,wild,0))
|
|
|
|
continue;
|
|
|
|
}
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_PRINT("info", ("Inserting %s into ndb_tables hash", t.name));
|
|
|
|
my_hash_insert(&ndb_tables, (byte*)thd->strdup(t.name));
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
char *file_name;
|
|
|
|
List_iterator<char> it(*files);
|
|
|
|
List<char> delete_list;
|
|
|
|
while ((file_name=it++))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s", file_name));
|
|
|
|
if (hash_search(&ndb_tables, file_name, strlen(file_name)))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s existed in NDB _and_ on disk ", file_name));
|
|
|
|
// File existed in NDB and as frm file, put in ok_tables list
|
|
|
|
my_hash_insert(&ok_tables, (byte*)file_name);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// File is not in NDB, check for .ndb file with this name
|
2004-09-21 12:13:58 +02:00
|
|
|
(void)strxnmov(name, FN_REFLEN,
|
2004-09-26 16:11:24 +02:00
|
|
|
mysql_data_home,"/",db,"/",file_name,ha_ndb_ext,NullS);
|
|
|
|
DBUG_PRINT("info", ("Check access for %s", name));
|
2004-09-21 12:13:58 +02:00
|
|
|
if (access(name, F_OK))
|
2004-09-26 16:11:24 +02:00
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("%s did not exist on disk", name));
|
|
|
|
// .ndb file did not exist on disk, another table type
|
|
|
|
continue;
|
2004-09-13 14:46:38 +02:00
|
|
|
}
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_PRINT("info", ("%s existed on disk", name));
|
|
|
|
// The .ndb file exists on disk, but it's not in list of tables in ndb
|
|
|
|
// Verify that handler agrees table is gone.
|
|
|
|
if (ndbcluster_table_exists(thd, db, file_name) == 0)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("NDB says %s does not exists", file_name));
|
|
|
|
it.remove();
|
|
|
|
// Put in list of tables to remove from disk
|
|
|
|
delete_list.push_back(thd->strdup(file_name));
|
|
|
|
}
|
|
|
|
}
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Check for new files to discover
|
|
|
|
DBUG_PRINT("info", ("Checking for new files to discover"));
|
|
|
|
List<char> create_list;
|
|
|
|
for (i= 0 ; i < ndb_tables.records ; i++)
|
2004-09-21 12:13:58 +02:00
|
|
|
{
|
2004-09-26 16:11:24 +02:00
|
|
|
file_name= hash_element(&ndb_tables, i);
|
|
|
|
if (!hash_search(&ok_tables, file_name, strlen(file_name)))
|
2004-09-21 12:13:58 +02:00
|
|
|
{
|
2004-09-26 16:11:24 +02:00
|
|
|
DBUG_PRINT("info", ("%s must be discovered", file_name));
|
|
|
|
// File is in list of ndb tables and not in ok_tables
|
|
|
|
// This table need to be created
|
|
|
|
create_list.push_back(thd->strdup(file_name));
|
|
|
|
}
|
|
|
|
}
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Lock mutex before deleting and creating frm files
|
|
|
|
pthread_mutex_lock(&LOCK_open);
|
2004-09-21 12:13:58 +02:00
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
if (!global_read_lock)
|
|
|
|
{
|
|
|
|
// Delete old files
|
|
|
|
List_iterator_fast<char> it3(delete_list);
|
|
|
|
while ((file_name=it3++))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Remove table %s/%s",db, file_name ));
|
|
|
|
// Delete the table and all related files
|
|
|
|
TABLE_LIST table_list;
|
|
|
|
bzero((char*) &table_list,sizeof(table_list));
|
|
|
|
table_list.db= (char*) db;
|
2004-10-20 01:55:02 +02:00
|
|
|
table_list.alias=table_list.real_name=(char*)file_name;
|
2004-09-26 16:11:24 +02:00
|
|
|
(void)mysql_rm_table_part2(thd, &table_list,
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
/* if_exists */ TRUE,
|
|
|
|
/* drop_temporary */ FALSE,
|
|
|
|
/* dont_log_query*/ TRUE);
|
2004-09-21 12:13:58 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2004-09-26 16:11:24 +02:00
|
|
|
// Create new files
|
|
|
|
List_iterator_fast<char> it2(create_list);
|
|
|
|
while ((file_name=it2++))
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info", ("Table %s need discovery", name));
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
if (ha_create_table_from_engine(thd, db, file_name, TRUE) == 0)
|
2004-10-07 09:16:46 +02:00
|
|
|
files->push_back(thd->strdup(file_name));
|
2004-09-26 16:11:24 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
pthread_mutex_unlock(&LOCK_open);
|
|
|
|
|
|
|
|
hash_free(&ok_tables);
|
2004-09-21 12:13:58 +02:00
|
|
|
hash_free(&ndb_tables);
|
2004-10-25 22:35:46 +02:00
|
|
|
} // extra bracket to avoid gcc 2.95.3 warning
|
2004-09-21 12:13:58 +02:00
|
|
|
DBUG_RETURN(0);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Initialise all gloal variables before creating
|
|
|
|
a NDB Cluster table handler
|
|
|
|
*/
|
|
|
|
|
|
|
|
bool ndbcluster_init()
|
|
|
|
{
|
2004-09-13 06:40:25 +02:00
|
|
|
int res;
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_ENTER("ndbcluster_init");
|
2004-08-20 18:10:47 +02:00
|
|
|
// Set connectstring if specified
|
2004-12-20 12:36:14 +01:00
|
|
|
if (opt_ndbcluster_connectstring != 0)
|
|
|
|
DBUG_PRINT("connectstring", ("%s", opt_ndbcluster_connectstring));
|
2004-09-06 20:30:57 +02:00
|
|
|
if ((g_ndb_cluster_connection=
|
2004-12-20 12:36:14 +01:00
|
|
|
new Ndb_cluster_connection(opt_ndbcluster_connectstring)) == 0)
|
2004-09-06 20:30:57 +02:00
|
|
|
{
|
2004-12-20 12:36:14 +01:00
|
|
|
DBUG_PRINT("error",("Ndb_cluster_connection(%s)",
|
|
|
|
opt_ndbcluster_connectstring));
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-08-20 18:10:47 +02:00
|
|
|
}
|
2004-09-13 06:40:25 +02:00
|
|
|
|
2004-12-20 12:36:14 +01:00
|
|
|
g_ndb_cluster_connection->set_optimized_node_selection
|
|
|
|
(opt_ndb_optimized_node_selection);
|
|
|
|
|
2004-09-06 20:30:57 +02:00
|
|
|
// Create a Ndb object to open the connection to NDB
|
|
|
|
g_ndb= new Ndb(g_ndb_cluster_connection, "sys");
|
2004-09-14 14:47:34 +02:00
|
|
|
g_ndb->getDictionary()->set_local_table_data_size(sizeof(Ndb_table_local_info));
|
2004-09-06 20:30:57 +02:00
|
|
|
if (g_ndb->init() != 0)
|
2004-04-15 09:14:14 +02:00
|
|
|
{
|
|
|
|
ERR_PRINT (g_ndb->getNdbError());
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
2004-09-13 06:40:25 +02:00
|
|
|
|
2004-11-30 18:47:47 +01:00
|
|
|
if ((res= g_ndb_cluster_connection->connect(0,0,0)) == 0)
|
2004-09-13 06:40:25 +02:00
|
|
|
{
|
2004-11-30 18:47:47 +01:00
|
|
|
DBUG_PRINT("info",("NDBCLUSTER storage engine at %s on port %d",
|
|
|
|
g_ndb_cluster_connection->get_connected_host(),
|
|
|
|
g_ndb_cluster_connection->get_connected_port()));
|
2004-12-20 12:36:14 +01:00
|
|
|
g_ndb_cluster_connection->wait_until_ready(10,0);
|
2004-09-13 06:40:25 +02:00
|
|
|
}
|
2004-09-27 02:24:23 +02:00
|
|
|
else if(res == 1)
|
2004-09-13 06:40:25 +02:00
|
|
|
{
|
2004-09-27 02:24:23 +02:00
|
|
|
if (g_ndb_cluster_connection->start_connect_thread()) {
|
|
|
|
DBUG_PRINT("error", ("g_ndb_cluster_connection->start_connect_thread()"));
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
|
|
|
}
|
|
|
|
{
|
|
|
|
char buf[1024];
|
|
|
|
DBUG_PRINT("info",("NDBCLUSTER storage engine not started, will connect using %s",
|
|
|
|
g_ndb_cluster_connection->get_connectstring(buf,sizeof(buf))));
|
2004-09-27 02:24:23 +02:00
|
|
|
}
|
2004-09-13 06:40:25 +02:00
|
|
|
}
|
2004-09-27 02:24:23 +02:00
|
|
|
else
|
2004-09-13 06:40:25 +02:00
|
|
|
{
|
|
|
|
DBUG_ASSERT(res == -1);
|
|
|
|
DBUG_PRINT("error", ("permanent error"));
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-09-13 06:40:25 +02:00
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
(void) hash_init(&ndbcluster_open_tables,system_charset_info,32,0,0,
|
|
|
|
(hash_get_key) ndbcluster_get_key,0,0);
|
|
|
|
pthread_mutex_init(&ndbcluster_mutex,MY_MUTEX_INIT_FAST);
|
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder
2004-09-14 10:52:21 +02:00
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
ndbcluster_inited= 1;
|
|
|
|
#ifdef USE_DISCOVER_ON_STARTUP
|
2004-09-30 16:12:28 +02:00
|
|
|
if (ndb_discover_tables() != 0)
|
2004-11-30 18:47:47 +01:00
|
|
|
goto ndbcluster_init_error;
|
2004-04-15 09:14:14 +02:00
|
|
|
#endif
|
true,false -> TRUE, FALSE
Simple fixes/optimization of things discovered during review of new pushed code
include/my_sys.h:
Ensure that clear_alloc_root() interacts correctly with alloc_root_inited()
mysys/hash.c:
More comments
Simple optimization (merge identical code)
mysys/my_bitmap.c:
Change inline -> static inline
sql/examples/ha_archive.cc:
Fixed compiler warning
sql/ha_ndbcluster.cc:
true,false -> TRUE, FALSE
Change if (false) -> #ifdef NOT_USED
sql/ha_ndbcluster.h:
true,false -> TRUE, FALSE
sql/handler.cc:
More comments
Remove not needed initializations.
#ifdef not used code
sql/item_cmpfunc.h:
true,false -> TRUE, FALSE
sql/item_strfunc.cc:
Move local variables to function beginning
Remove wrong comments
sql/log_event.h:
true,false -> TRUE, FALSE
sql/sql_base.cc:
true,false -> TRUE, FALSE
More comments
sql/sql_help.cc:
true,false -> TRUE, FALSE
sql/sql_lex.cc:
Simple optimization of new code
sql/sql_parse.cc:
true,false -> TRUE, FALSE
sql/sql_prepare.cc:
true,false -> TRUE, FALSE
sql/sql_table.cc:
true,false -> TRUE, FALSE
sql/sql_yacc.yy:
true,false -> TRUE, FALSE
2004-10-14 17:03:46 +02:00
|
|
|
DBUG_RETURN(FALSE);
|
2004-11-30 18:47:47 +01:00
|
|
|
ndbcluster_init_error:
|
|
|
|
ndbcluster_end();
|
|
|
|
DBUG_RETURN(TRUE);
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
End use of the NDB Cluster table handler
|
|
|
|
- free all global variables allocated by
|
|
|
|
ndcluster_init()
|
|
|
|
*/
|
|
|
|
|
|
|
|
bool ndbcluster_end()
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_end");
|
2004-09-13 06:40:25 +02:00
|
|
|
if(g_ndb)
|
|
|
|
delete g_ndb;
|
2004-04-15 09:14:14 +02:00
|
|
|
g_ndb= NULL;
|
2004-09-06 20:30:57 +02:00
|
|
|
if (g_ndb_cluster_connection)
|
|
|
|
delete g_ndb_cluster_connection;
|
|
|
|
g_ndb_cluster_connection= NULL;
|
2004-04-15 09:14:14 +02:00
|
|
|
if (!ndbcluster_inited)
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
hash_free(&ndbcluster_open_tables);
|
|
|
|
pthread_mutex_destroy(&ndbcluster_mutex);
|
|
|
|
ndbcluster_inited= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
|
|
|
|
2004-06-23 15:34:45 +02:00
|
|
|
/*
|
|
|
|
Static error print function called from
|
|
|
|
static handler method ndbcluster_commit
|
|
|
|
and ndbcluster_rollback
|
|
|
|
*/
|
2004-07-02 16:14:08 +02:00
|
|
|
|
|
|
|
void ndbcluster_print_error(int error, const NdbOperation *error_op)
|
2004-06-22 09:54:00 +02:00
|
|
|
{
|
2004-06-23 15:34:45 +02:00
|
|
|
DBUG_ENTER("ndbcluster_print_error");
|
|
|
|
TABLE tab;
|
2004-07-02 16:14:08 +02:00
|
|
|
const char *tab_name= (error_op) ? error_op->getTableName() : "";
|
|
|
|
tab.table_name= (char *) tab_name;
|
2004-06-23 15:34:45 +02:00
|
|
|
ha_ndbcluster error_handler(&tab);
|
2004-07-02 16:14:08 +02:00
|
|
|
tab.file= &error_handler;
|
2004-06-23 15:34:45 +02:00
|
|
|
error_handler.print_error(error, MYF(0));
|
2004-07-06 08:43:57 +02:00
|
|
|
DBUG_VOID_RETURN;
|
2004-06-22 09:54:00 +02:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
Set m_tabname from full pathname to table file
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::set_tabname(const char *path_name)
|
|
|
|
{
|
|
|
|
char *end, *ptr;
|
|
|
|
|
|
|
|
/* Scan name from the end */
|
|
|
|
end= strend(path_name)-1;
|
|
|
|
ptr= end;
|
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
|
|
|
uint name_len= end - ptr;
|
|
|
|
memcpy(m_tabname, ptr + 1, end - ptr);
|
|
|
|
m_tabname[name_len]= '\0';
|
|
|
|
#ifdef __WIN__
|
|
|
|
/* Put to lower case */
|
|
|
|
ptr= m_tabname;
|
|
|
|
|
|
|
|
while (*ptr != '\0') {
|
2004-07-23 15:46:56 +02:00
|
|
|
*ptr= tolower(*ptr);
|
2004-04-15 09:14:14 +02:00
|
|
|
ptr++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Set a given location from full pathname to table file
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
ha_ndbcluster::set_tabname(const char *path_name, char * tabname)
|
|
|
|
{
|
|
|
|
char *end, *ptr;
|
|
|
|
|
|
|
|
/* Scan name from the end */
|
2004-07-23 15:46:56 +02:00
|
|
|
end= strend(path_name)-1;
|
|
|
|
ptr= end;
|
2004-04-15 09:14:14 +02:00
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
2004-07-23 15:46:56 +02:00
|
|
|
uint name_len= end - ptr;
|
2004-04-15 09:14:14 +02:00
|
|
|
memcpy(tabname, ptr + 1, end - ptr);
|
2004-07-23 15:46:56 +02:00
|
|
|
tabname[name_len]= '\0';
|
2004-04-15 09:14:14 +02:00
|
|
|
#ifdef __WIN__
|
|
|
|
/* Put to lower case */
|
2004-07-23 15:46:56 +02:00
|
|
|
ptr= tabname;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
while (*ptr != '\0') {
|
|
|
|
*ptr= tolower(*ptr);
|
|
|
|
ptr++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Set m_dbname from full pathname to table file
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
void ha_ndbcluster::set_dbname(const char *path_name)
|
|
|
|
{
|
|
|
|
char *end, *ptr;
|
|
|
|
|
|
|
|
/* Scan name from the end */
|
|
|
|
ptr= strend(path_name)-1;
|
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
|
|
|
ptr--;
|
|
|
|
end= ptr;
|
|
|
|
while (ptr >= path_name && *ptr != '\\' && *ptr != '/') {
|
|
|
|
ptr--;
|
|
|
|
}
|
|
|
|
uint name_len= end - ptr;
|
|
|
|
memcpy(m_dbname, ptr + 1, name_len);
|
|
|
|
m_dbname[name_len]= '\0';
|
|
|
|
#ifdef __WIN__
|
|
|
|
/* Put to lower case */
|
|
|
|
|
|
|
|
ptr= m_dbname;
|
|
|
|
|
|
|
|
while (*ptr != '\0') {
|
|
|
|
*ptr= tolower(*ptr);
|
|
|
|
ptr++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
ha_rows
|
2004-05-16 13:48:32 +02:00
|
|
|
ha_ndbcluster::records_in_range(uint inx, key_range *min_key,
|
|
|
|
key_range *max_key)
|
|
|
|
{
|
|
|
|
KEY *key_info= table->key_info + inx;
|
2004-04-15 09:14:14 +02:00
|
|
|
uint key_length= key_info->key_length;
|
2004-05-24 12:35:39 +02:00
|
|
|
NDB_INDEX_TYPE idx_type= get_index_type(inx);
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_ENTER("records_in_range");
|
2004-05-24 12:35:39 +02:00
|
|
|
// Prevent partial read of hash indexes by returning HA_POS_ERROR
|
|
|
|
if ((idx_type == UNIQUE_INDEX || idx_type == PRIMARY_KEY_INDEX) &&
|
|
|
|
((min_key && min_key->length < key_length) ||
|
|
|
|
(max_key && max_key->length < key_length)))
|
|
|
|
DBUG_RETURN(HA_POS_ERROR);
|
|
|
|
|
|
|
|
// Read from hash index with full key
|
|
|
|
// This is a "const" table which returns only one record!
|
|
|
|
if ((idx_type != ORDERED_INDEX) &&
|
|
|
|
((min_key && min_key->length == key_length) ||
|
|
|
|
(max_key && max_key->length == key_length)))
|
|
|
|
DBUG_RETURN(1);
|
|
|
|
|
|
|
|
DBUG_RETURN(10); /* Good guess when you don't know anything */
|
2004-04-15 09:14:14 +02:00
|
|
|
}
|
|
|
|
|
2004-11-17 09:15:53 +01:00
|
|
|
ulong ha_ndbcluster::table_flags(void) const
|
|
|
|
{
|
|
|
|
if (m_ha_not_exact_count)
|
|
|
|
return m_table_flags | HA_NOT_EXACT_COUNT;
|
|
|
|
else
|
|
|
|
return m_table_flags;
|
|
|
|
}
|
|
|
|
const char * ha_ndbcluster::table_type() const
|
|
|
|
{
|
|
|
|
return("ndbcluster");
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_record_length() const
|
|
|
|
{
|
|
|
|
return NDB_MAX_TUPLE_SIZE;
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_keys() const
|
|
|
|
{
|
|
|
|
return MAX_KEY;
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_key_parts() const
|
|
|
|
{
|
|
|
|
return NDB_MAX_NO_OF_ATTRIBUTES_IN_KEY;
|
|
|
|
}
|
|
|
|
uint ha_ndbcluster::max_supported_key_length() const
|
|
|
|
{
|
|
|
|
return NDB_MAX_KEY_SIZE;
|
|
|
|
}
|
|
|
|
bool ha_ndbcluster::low_byte_first() const
|
|
|
|
{
|
|
|
|
#ifdef WORDS_BIGENDIAN
|
|
|
|
return FALSE;
|
|
|
|
#else
|
|
|
|
return TRUE;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
bool ha_ndbcluster::has_transactions()
|
|
|
|
{
|
2004-11-22 14:41:46 +01:00
|
|
|
return m_transaction_on;
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
|
|
|
const char* ha_ndbcluster::index_type(uint key_number)
|
|
|
|
{
|
|
|
|
switch (get_index_type(key_number)) {
|
|
|
|
case ORDERED_INDEX:
|
|
|
|
case UNIQUE_ORDERED_INDEX:
|
|
|
|
case PRIMARY_KEY_ORDERED_INDEX:
|
|
|
|
return "BTREE";
|
|
|
|
case UNIQUE_INDEX:
|
|
|
|
case PRIMARY_KEY_INDEX:
|
|
|
|
default:
|
|
|
|
return "HASH";
|
|
|
|
}
|
|
|
|
}
|
|
|
|
uint8 ha_ndbcluster::table_cache_type()
|
|
|
|
{
|
2004-11-25 19:56:10 +01:00
|
|
|
DBUG_ENTER("ha_ndbcluster::table_cache_type=HA_CACHE_TBL_ASKTRANSACT");
|
|
|
|
DBUG_RETURN(HA_CACHE_TBL_ASKTRANSACT);
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
sql/ha_innodb.cc
enabled query cache for ndb
modified engine interface somewhat
sql/ha_innodb.h
enabled query cache for ndb
modified engine interface somewhat
sql/ha_ndbcluster.cc
enabled query cache for ndb
modified engine interface somewhat
ndb will only allow caching and retrieval if running autocommit
- return false, but do not invalidate
commit count is used as engine data, i.e.
- store commit count before store of cache
- allow retrieval if commit count has not changed on a table
- invalidate if commit count has changed
sql/ha_ndbcluster.h
enabled query cache for ndb
modified engine interface somewhat
sql/handler.cc
enabled query cache for ndb
modified engine interface somewhat
sql/handler.h
enabled query cache for ndb
modified engine interface somewhat
new virtual handler method cached_table_registration called on each table before alowing store in query cache
- return TRUE - ok to cache, FALSE - not allowed to cache, invalidate queries if engine_data below has changed
- sets ulonglong (engine_data) that is stored in query cache for each table
- sets callback to be called for each table before usage of cached query, callback = 0 -> no check later
sql/mysql_priv.h
enabled query cache for ndb
modified engine interface somewhat
callcack prototype for callback to engine before query cache retrieval
sql/sql_cache.cc
enabled query cache for ndb
modified engine interface somewhat
if callback is set on table in cache, do callback to check if allowed to use cache
if not allowed to use cache, check if engine_data has changed, if so, invalidate all queries with that table
+ changes to store and pass callback and engine_data around
sql/sql_cache.h
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/table.h
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/ha_innodb.cc:
enabled query cache for ndb
modified engine interface somewhat
sql/ha_innodb.h:
enabled query cache for ndb
modified engine interface somewhat
sql/ha_ndbcluster.cc:
enabled query cache for ndb
modified engine interface somewhat
ndb will only allow caching and retrieval if running autocommit
- return false, but do not invalidate
commit count is used as engine data, i.e.
- store commit count before store of cache
- allow retrieval if commit count has not changed on a table
- invalidate if commit count has changed
sql/ha_ndbcluster.h:
enabled query cache for ndb
modified engine interface somewhat
sql/handler.cc:
enabled query cache for ndb
modified engine interface somewhat
sql/handler.h:
enabled query cache for ndb
modified engine interface somewhat
new virtual handler method cached_table_registration called on each table before alowing store in query cache
- return TRUE - ok to cache, FALSE - not allowed to cache, invalidate queries if engine_data below has changed
- sets ulonglong (engine_data) that is stored in query cache for each table
- sets callback to be called for each table before usage of cached query, callback = 0 -> no check later
sql/mysql_priv.h:
enabled query cache for ndb
modified engine interface somewhat
callcack prototype for callback to engine before query cache retrieval
sql/sql_cache.cc:
enabled query cache for ndb
modified engine interface somewhat
if callback is set on table in cache, do callback to check if allowed to use cache
if not allowed to use cache, check if engine_data has changed, if so, invalidate all queries with that table
+ changes to store and pass callback and engine_data around
sql/sql_cache.h:
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/table.h:
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
2004-11-24 12:56:51 +01:00
|
|
|
static
|
|
|
|
my_bool
|
|
|
|
ndbcluster_cache_retrieval_allowed(
|
|
|
|
/*======================================*/
|
|
|
|
/* out: TRUE if permitted, FALSE if not;
|
|
|
|
note that the value FALSE means invalidation
|
|
|
|
of query cache if *engine_data is changed */
|
|
|
|
THD* thd, /* in: thd of the user who is trying to
|
|
|
|
store a result to the query cache or
|
|
|
|
retrieve it */
|
|
|
|
char* full_name, /* in: concatenation of database name,
|
|
|
|
the null character '\0', and the table
|
|
|
|
name */
|
|
|
|
uint full_name_len, /* in: length of the full name, i.e.
|
|
|
|
len(dbname) + len(tablename) + 1 */
|
|
|
|
ulonglong *engine_data) /* in: value set in call to
|
|
|
|
ha_ndbcluster::cached_table_registration
|
|
|
|
out: if return FALSE this is used to invalidate
|
|
|
|
all cached queries with this table*/
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ndbcluster_cache_retrieval_allowed");
|
|
|
|
char tabname[128];
|
|
|
|
char *dbname= full_name;
|
|
|
|
my_bool is_autocommit;
|
|
|
|
{
|
|
|
|
int dbname_len= strlen(full_name);
|
|
|
|
int tabname_len= full_name_len-dbname_len-1;
|
|
|
|
memcpy(tabname, full_name+dbname_len+1, tabname_len);
|
|
|
|
tabname[tabname_len]= '\0';
|
|
|
|
}
|
|
|
|
if (thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN))
|
|
|
|
is_autocommit = FALSE;
|
|
|
|
else
|
|
|
|
is_autocommit = TRUE;
|
|
|
|
DBUG_PRINT("enter",("dbname=%s, tabname=%s, autocommit=%d",
|
|
|
|
dbname,tabname,is_autocommit));
|
|
|
|
if (!is_autocommit)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info",("OPTION_NOT_AUTOCOMMIT=%d OPTION_BEGIN=%d",
|
|
|
|
thd->options & OPTION_NOT_AUTOCOMMIT,
|
|
|
|
thd->options & OPTION_BEGIN));
|
|
|
|
// ToDo enable cache inside a transaction
|
|
|
|
// no need to invalidate though so leave *engine_data
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
{
|
|
|
|
Ndb *ndb;
|
|
|
|
Uint64 commit_count;
|
|
|
|
if (!(ndb= check_ndb_in_thd(thd)))
|
|
|
|
{
|
|
|
|
*engine_data= *engine_data+1; // invalidate
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
ndb->setDatabaseName(dbname);
|
|
|
|
if (ndb_get_table_statistics(ndb, tabname, 0, &commit_count))
|
|
|
|
{
|
|
|
|
*engine_data= *engine_data+1; // invalidate
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
if (*engine_data != commit_count)
|
|
|
|
{
|
|
|
|
*engine_data= commit_count; // invalidate
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
DBUG_PRINT("exit",("*engine_data=%d ok, use cache",*engine_data));
|
|
|
|
DBUG_RETURN(TRUE);
|
|
|
|
}
|
|
|
|
|
|
|
|
my_bool
|
|
|
|
ha_ndbcluster::cached_table_registration(
|
|
|
|
/*======================================*/
|
|
|
|
/* out: TRUE if permitted, FALSE if not;
|
|
|
|
note that the value FALSE means invalidation
|
|
|
|
of query cache if *engine_data is changed */
|
|
|
|
THD* thd, /* in: thd of the user who is trying to
|
|
|
|
store a result to the query cache or
|
|
|
|
retrieve it */
|
|
|
|
char* full_name, /* in: concatenation of database name,
|
|
|
|
the null character '\0', and the table
|
|
|
|
name */
|
|
|
|
uint full_name_len, /* in: length of the full name, i.e.
|
|
|
|
len(dbname) + len(tablename) + 1 */
|
|
|
|
qc_engine_callback
|
|
|
|
*engine_callback, /* out: function to be called before using
|
|
|
|
cache on this table */
|
|
|
|
ulonglong *engine_data) /* out: if return FALSE this is used to
|
|
|
|
invalidate all cached queries with this table*/
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ha_ndbcluster::cached_table_registration");
|
|
|
|
my_bool is_autocommit;
|
|
|
|
if (thd->options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN))
|
|
|
|
is_autocommit = FALSE;
|
2004-11-22 14:41:46 +01:00
|
|
|
else
|
sql/ha_innodb.cc
enabled query cache for ndb
modified engine interface somewhat
sql/ha_innodb.h
enabled query cache for ndb
modified engine interface somewhat
sql/ha_ndbcluster.cc
enabled query cache for ndb
modified engine interface somewhat
ndb will only allow caching and retrieval if running autocommit
- return false, but do not invalidate
commit count is used as engine data, i.e.
- store commit count before store of cache
- allow retrieval if commit count has not changed on a table
- invalidate if commit count has changed
sql/ha_ndbcluster.h
enabled query cache for ndb
modified engine interface somewhat
sql/handler.cc
enabled query cache for ndb
modified engine interface somewhat
sql/handler.h
enabled query cache for ndb
modified engine interface somewhat
new virtual handler method cached_table_registration called on each table before alowing store in query cache
- return TRUE - ok to cache, FALSE - not allowed to cache, invalidate queries if engine_data below has changed
- sets ulonglong (engine_data) that is stored in query cache for each table
- sets callback to be called for each table before usage of cached query, callback = 0 -> no check later
sql/mysql_priv.h
enabled query cache for ndb
modified engine interface somewhat
callcack prototype for callback to engine before query cache retrieval
sql/sql_cache.cc
enabled query cache for ndb
modified engine interface somewhat
if callback is set on table in cache, do callback to check if allowed to use cache
if not allowed to use cache, check if engine_data has changed, if so, invalidate all queries with that table
+ changes to store and pass callback and engine_data around
sql/sql_cache.h
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/table.h
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/ha_innodb.cc:
enabled query cache for ndb
modified engine interface somewhat
sql/ha_innodb.h:
enabled query cache for ndb
modified engine interface somewhat
sql/ha_ndbcluster.cc:
enabled query cache for ndb
modified engine interface somewhat
ndb will only allow caching and retrieval if running autocommit
- return false, but do not invalidate
commit count is used as engine data, i.e.
- store commit count before store of cache
- allow retrieval if commit count has not changed on a table
- invalidate if commit count has changed
sql/ha_ndbcluster.h:
enabled query cache for ndb
modified engine interface somewhat
sql/handler.cc:
enabled query cache for ndb
modified engine interface somewhat
sql/handler.h:
enabled query cache for ndb
modified engine interface somewhat
new virtual handler method cached_table_registration called on each table before alowing store in query cache
- return TRUE - ok to cache, FALSE - not allowed to cache, invalidate queries if engine_data below has changed
- sets ulonglong (engine_data) that is stored in query cache for each table
- sets callback to be called for each table before usage of cached query, callback = 0 -> no check later
sql/mysql_priv.h:
enabled query cache for ndb
modified engine interface somewhat
callcack prototype for callback to engine before query cache retrieval
sql/sql_cache.cc:
enabled query cache for ndb
modified engine interface somewhat
if callback is set on table in cache, do callback to check if allowed to use cache
if not allowed to use cache, check if engine_data has changed, if so, invalidate all queries with that table
+ changes to store and pass callback and engine_data around
sql/sql_cache.h:
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/table.h:
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
2004-11-24 12:56:51 +01:00
|
|
|
is_autocommit = TRUE;
|
|
|
|
DBUG_PRINT("enter",("dbname=%s, tabname=%s, is_autocommit=%d",
|
|
|
|
m_dbname,m_tabname,is_autocommit));
|
|
|
|
if (!is_autocommit)
|
|
|
|
{
|
|
|
|
DBUG_PRINT("info",("OPTION_NOT_AUTOCOMMIT=%d OPTION_BEGIN=%d",
|
|
|
|
thd->options & OPTION_NOT_AUTOCOMMIT,
|
|
|
|
thd->options & OPTION_BEGIN));
|
|
|
|
// ToDo enable cache inside a transaction
|
|
|
|
// no need to invalidate though so leave *engine_data
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
{
|
|
|
|
Uint64 commit_count;
|
2005-01-14 14:33:26 +01:00
|
|
|
Ndb *ndb= get_ndb();
|
|
|
|
ndb->setDatabaseName(m_dbname);
|
|
|
|
if (ndb_get_table_statistics(ndb, m_tabname, 0, &commit_count))
|
sql/ha_innodb.cc
enabled query cache for ndb
modified engine interface somewhat
sql/ha_innodb.h
enabled query cache for ndb
modified engine interface somewhat
sql/ha_ndbcluster.cc
enabled query cache for ndb
modified engine interface somewhat
ndb will only allow caching and retrieval if running autocommit
- return false, but do not invalidate
commit count is used as engine data, i.e.
- store commit count before store of cache
- allow retrieval if commit count has not changed on a table
- invalidate if commit count has changed
sql/ha_ndbcluster.h
enabled query cache for ndb
modified engine interface somewhat
sql/handler.cc
enabled query cache for ndb
modified engine interface somewhat
sql/handler.h
enabled query cache for ndb
modified engine interface somewhat
new virtual handler method cached_table_registration called on each table before alowing store in query cache
- return TRUE - ok to cache, FALSE - not allowed to cache, invalidate queries if engine_data below has changed
- sets ulonglong (engine_data) that is stored in query cache for each table
- sets callback to be called for each table before usage of cached query, callback = 0 -> no check later
sql/mysql_priv.h
enabled query cache for ndb
modified engine interface somewhat
callcack prototype for callback to engine before query cache retrieval
sql/sql_cache.cc
enabled query cache for ndb
modified engine interface somewhat
if callback is set on table in cache, do callback to check if allowed to use cache
if not allowed to use cache, check if engine_data has changed, if so, invalidate all queries with that table
+ changes to store and pass callback and engine_data around
sql/sql_cache.h
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/table.h
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/ha_innodb.cc:
enabled query cache for ndb
modified engine interface somewhat
sql/ha_innodb.h:
enabled query cache for ndb
modified engine interface somewhat
sql/ha_ndbcluster.cc:
enabled query cache for ndb
modified engine interface somewhat
ndb will only allow caching and retrieval if running autocommit
- return false, but do not invalidate
commit count is used as engine data, i.e.
- store commit count before store of cache
- allow retrieval if commit count has not changed on a table
- invalidate if commit count has changed
sql/ha_ndbcluster.h:
enabled query cache for ndb
modified engine interface somewhat
sql/handler.cc:
enabled query cache for ndb
modified engine interface somewhat
sql/handler.h:
enabled query cache for ndb
modified engine interface somewhat
new virtual handler method cached_table_registration called on each table before alowing store in query cache
- return TRUE - ok to cache, FALSE - not allowed to cache, invalidate queries if engine_data below has changed
- sets ulonglong (engine_data) that is stored in query cache for each table
- sets callback to be called for each table before usage of cached query, callback = 0 -> no check later
sql/mysql_priv.h:
enabled query cache for ndb
modified engine interface somewhat
callcack prototype for callback to engine before query cache retrieval
sql/sql_cache.cc:
enabled query cache for ndb
modified engine interface somewhat
if callback is set on table in cache, do callback to check if allowed to use cache
if not allowed to use cache, check if engine_data has changed, if so, invalidate all queries with that table
+ changes to store and pass callback and engine_data around
sql/sql_cache.h:
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
sql/table.h:
enabled query cache for ndb
modified engine interface somewhat
changes to store callback and engine_data
2004-11-24 12:56:51 +01:00
|
|
|
{
|
|
|
|
*engine_data= 0;
|
|
|
|
DBUG_RETURN(FALSE);
|
|
|
|
}
|
|
|
|
*engine_data= commit_count;
|
|
|
|
}
|
|
|
|
*engine_callback= ndbcluster_cache_retrieval_allowed;
|
|
|
|
DBUG_PRINT("exit",("*engine_data=%d", *engine_data));
|
|
|
|
DBUG_RETURN(TRUE);
|
2004-11-17 09:15:53 +01:00
|
|
|
}
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
Handling the shared NDB_SHARE structure that is needed to
|
|
|
|
provide table locking.
|
|
|
|
It's also used for sharing data with other NDB handlers
|
|
|
|
in the same MySQL Server. There is currently not much
|
|
|
|
data we want to or can share.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static byte* ndbcluster_get_key(NDB_SHARE *share,uint *length,
|
|
|
|
my_bool not_used __attribute__((unused)))
|
|
|
|
{
|
|
|
|
*length=share->table_name_length;
|
|
|
|
return (byte*) share->table_name;
|
|
|
|
}
|
|
|
|
|
|
|
|
static NDB_SHARE* get_share(const char *table_name)
|
|
|
|
{
|
|
|
|
NDB_SHARE *share;
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
|
|
|
uint length=(uint) strlen(table_name);
|
|
|
|
if (!(share=(NDB_SHARE*) hash_search(&ndbcluster_open_tables,
|
|
|
|
(byte*) table_name,
|
|
|
|
length)))
|
|
|
|
{
|
|
|
|
if ((share=(NDB_SHARE *) my_malloc(sizeof(*share)+length+1,
|
|
|
|
MYF(MY_WME | MY_ZEROFILL))))
|
|
|
|
{
|
|
|
|
share->table_name_length=length;
|
|
|
|
share->table_name=(char*) (share+1);
|
|
|
|
strmov(share->table_name,table_name);
|
|
|
|
if (my_hash_insert(&ndbcluster_open_tables, (byte*) share))
|
|
|
|
{
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
my_free((gptr) share,0);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
thr_lock_init(&share->lock);
|
|
|
|
pthread_mutex_init(&share->mutex,MY_MUTEX_INIT_FAST);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
share->use_count++;
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
return share;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void free_share(NDB_SHARE *share)
|
|
|
|
{
|
|
|
|
pthread_mutex_lock(&ndbcluster_mutex);
|
|
|
|
if (!--share->use_count)
|
|
|
|
{
|
|
|
|
hash_delete(&ndbcluster_open_tables, (byte*) share);
|
|
|
|
thr_lock_delete(&share->lock);
|
|
|
|
pthread_mutex_destroy(&share->mutex);
|
|
|
|
my_free((gptr) share, MYF(0));
|
|
|
|
}
|
|
|
|
pthread_mutex_unlock(&ndbcluster_mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
Internal representation of the frm blob
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct frm_blob_struct
|
|
|
|
{
|
|
|
|
struct frm_blob_header
|
|
|
|
{
|
|
|
|
uint ver; // Version of header
|
|
|
|
uint orglen; // Original length of compressed data
|
|
|
|
uint complen; // Compressed length of data, 0=uncompressed
|
|
|
|
} head;
|
|
|
|
char data[1];
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int packfrm(const void *data, uint len,
|
|
|
|
const void **pack_data, uint *pack_len)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
ulong org_len, comp_len;
|
|
|
|
uint blob_len;
|
|
|
|
frm_blob_struct* blob;
|
|
|
|
DBUG_ENTER("packfrm");
|
|
|
|
DBUG_PRINT("enter", ("data: %x, len: %d", data, len));
|
|
|
|
|
|
|
|
error= 1;
|
2004-07-23 15:46:56 +02:00
|
|
|
org_len= len;
|
2004-04-15 09:14:14 +02:00
|
|
|
if (my_compress((byte*)data, &org_len, &comp_len))
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
DBUG_PRINT("info", ("org_len: %d, comp_len: %d", org_len, comp_len));
|
|
|
|
DBUG_DUMP("compressed", (char*)data, org_len);
|
|
|
|
|
|
|
|
error= 2;
|
|
|
|
blob_len= sizeof(frm_blob_struct::frm_blob_header)+org_len;
|
|
|
|
if (!(blob= (frm_blob_struct*) my_malloc(blob_len,MYF(MY_WME))))
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
// Store compressed blob in machine independent format
|
|
|
|
int4store((char*)(&blob->head.ver), 1);
|
|
|
|
int4store((char*)(&blob->head.orglen), comp_len);
|
|
|
|
int4store((char*)(&blob->head.complen), org_len);
|
|
|
|
|
|
|
|
// Copy frm data into blob, already in machine independent format
|
|
|
|
memcpy(blob->data, data, org_len);
|
|
|
|
|
2004-07-23 15:46:56 +02:00
|
|
|
*pack_data= blob;
|
|
|
|
*pack_len= blob_len;
|
|
|
|
error= 0;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("exit", ("pack_data: %x, pack_len: %d", *pack_data, *pack_len));
|
|
|
|
err:
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int unpackfrm(const void **unpack_data, uint *unpack_len,
|
|
|
|
const void *pack_data)
|
|
|
|
{
|
2004-07-23 15:46:56 +02:00
|
|
|
const frm_blob_struct *blob= (frm_blob_struct*)pack_data;
|
2004-04-15 09:14:14 +02:00
|
|
|
byte *data;
|
|
|
|
ulong complen, orglen, ver;
|
|
|
|
DBUG_ENTER("unpackfrm");
|
|
|
|
DBUG_PRINT("enter", ("pack_data: %x", pack_data));
|
|
|
|
|
|
|
|
complen= uint4korr((char*)&blob->head.complen);
|
|
|
|
orglen= uint4korr((char*)&blob->head.orglen);
|
|
|
|
ver= uint4korr((char*)&blob->head.ver);
|
|
|
|
|
|
|
|
DBUG_PRINT("blob",("ver: %d complen: %d orglen: %d",
|
|
|
|
ver,complen,orglen));
|
|
|
|
DBUG_DUMP("blob->data", (char*) blob->data, complen);
|
|
|
|
|
|
|
|
if (ver != 1)
|
|
|
|
DBUG_RETURN(1);
|
2004-07-23 15:46:56 +02:00
|
|
|
if (!(data= my_malloc(max(orglen, complen), MYF(MY_WME))))
|
2004-04-15 09:14:14 +02:00
|
|
|
DBUG_RETURN(2);
|
|
|
|
memcpy(data, blob->data, complen);
|
|
|
|
|
|
|
|
if (my_uncompress(data, &complen, &orglen))
|
|
|
|
{
|
|
|
|
my_free((char*)data, MYF(0));
|
|
|
|
DBUG_RETURN(3);
|
|
|
|
}
|
|
|
|
|
2004-07-23 15:46:56 +02:00
|
|
|
*unpack_data= data;
|
|
|
|
*unpack_len= complen;
|
2004-04-15 09:14:14 +02:00
|
|
|
|
|
|
|
DBUG_PRINT("exit", ("frmdata: %x, len: %d", *unpack_data, *unpack_len));
|
|
|
|
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
}
|
2004-09-03 13:55:40 +02:00
|
|
|
|
|
|
|
static
|
|
|
|
int
|
|
|
|
ndb_get_table_statistics(Ndb* ndb, const char * table,
|
|
|
|
Uint64* row_count, Uint64* commit_count)
|
|
|
|
{
|
|
|
|
DBUG_ENTER("ndb_get_table_statistics");
|
|
|
|
DBUG_PRINT("enter", ("table: %s", table));
|
2004-11-22 14:41:46 +01:00
|
|
|
NdbConnection* pTrans= ndb->startTransaction();
|
2004-09-03 13:55:40 +02:00
|
|
|
do
|
|
|
|
{
|
|
|
|
if (pTrans == NULL)
|
|
|
|
break;
|
2004-11-22 14:41:46 +01:00
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
NdbScanOperation* pOp= pTrans->getNdbScanOperation(table);
|
|
|
|
if (pOp == NULL)
|
|
|
|
break;
|
|
|
|
|
2004-09-29 01:42:34 +02:00
|
|
|
NdbResultSet* rs= pOp->readTuples(NdbOperation::LM_CommittedRead);
|
2004-09-03 13:55:40 +02:00
|
|
|
if (rs == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
int check= pOp->interpret_exit_last_row();
|
|
|
|
if (check == -1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
Uint64 rows, commits;
|
|
|
|
pOp->getValue(NdbDictionary::Column::ROW_COUNT, (char*)&rows);
|
|
|
|
pOp->getValue(NdbDictionary::Column::COMMIT_COUNT, (char*)&commits);
|
|
|
|
|
2004-11-22 14:41:46 +01:00
|
|
|
check= pTrans->execute(NoCommit, AbortOnError, TRUE);
|
2004-09-03 13:55:40 +02:00
|
|
|
if (check == -1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
Uint64 sum_rows= 0;
|
|
|
|
Uint64 sum_commits= 0;
|
2004-11-22 14:41:46 +01:00
|
|
|
while((check= rs->nextResult(TRUE, TRUE)) == 0)
|
2004-09-03 13:55:40 +02:00
|
|
|
{
|
|
|
|
sum_rows+= rows;
|
|
|
|
sum_commits+= commits;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (check == -1)
|
|
|
|
break;
|
|
|
|
|
2004-11-22 14:41:46 +01:00
|
|
|
rs->close(TRUE);
|
|
|
|
|
2004-09-03 13:55:40 +02:00
|
|
|
ndb->closeTransaction(pTrans);
|
|
|
|
if(row_count)
|
|
|
|
* row_count= sum_rows;
|
|
|
|
if(commit_count)
|
|
|
|
* commit_count= sum_commits;
|
|
|
|
DBUG_PRINT("exit", ("records: %u commits: %u", sum_rows, sum_commits));
|
|
|
|
DBUG_RETURN(0);
|
|
|
|
} while(0);
|
|
|
|
|
2004-11-22 14:41:46 +01:00
|
|
|
ndb->closeTransaction(pTrans);
|
2004-09-03 13:55:40 +02:00
|
|
|
DBUG_PRINT("exit", ("failed"));
|
|
|
|
DBUG_RETURN(-1);
|
|
|
|
}
|
|
|
|
|
2004-09-13 14:46:38 +02:00
|
|
|
/*
|
|
|
|
Create a .ndb file to serve as a placeholder indicating
|
|
|
|
that the table with this name is a ndb table
|
|
|
|
*/
|
|
|
|
|
|
|
|
int ha_ndbcluster::write_ndb_file()
|
|
|
|
{
|
|
|
|
File file;
|
|
|
|
bool error=1;
|
|
|
|
char path[FN_REFLEN];
|
|
|
|
|
|
|
|
DBUG_ENTER("write_ndb_file");
|
|
|
|
DBUG_PRINT("enter", ("db: %s, name: %s", m_dbname, m_tabname));
|
|
|
|
|
|
|
|
(void)strxnmov(path, FN_REFLEN,
|
|
|
|
mysql_data_home,"/",m_dbname,"/",m_tabname,ha_ndb_ext,NullS);
|
|
|
|
|
|
|
|
if ((file=my_create(path, CREATE_MODE,O_RDWR | O_TRUNC,MYF(MY_WME))) >= 0)
|
|
|
|
{
|
|
|
|
// It's an empty file
|
|
|
|
error=0;
|
|
|
|
my_close(file,MYF(0));
|
|
|
|
}
|
|
|
|
DBUG_RETURN(error);
|
|
|
|
}
|
|
|
|
|
2004-04-15 09:14:14 +02:00
|
|
|
#endif /* HAVE_NDBCLUSTER_DB */
|