mirror of
https://github.com/MariaDB/server.git
synced 2025-01-21 22:34:18 +01:00
eff3780dd8
Backport of: ------------------------------------------------------------ revno: 2630.4.1 committer: Dmitry Lenev <dlenev@mysql.com> branch nick: mysql-6.0-3726-w timestamp: Fri 2008-05-23 17:54:03 +0400 message: WL#3726 "DDL locking for all metadata objects". After review fixes in progress. ------------------------------------------------------------ This is the first patch in series. It transforms the metadata locking subsystem to use a dedicated module (mdl.h,cc). No significant changes in the locking protocol. The import passes the test suite with the exception of deprecated/removed 6.0 features, and MERGE tables. The latter are subject to a fix by WL#4144. Unfortunately, the original changeset comments got lost in a merge, thus this import has its own (largely insufficient) comments. This patch fixes Bug#25144 "replication / binlog with view breaks". Warning: this patch introduces an incompatible change: Under LOCK TABLES, it's no longer possible to FLUSH a table that was not locked for WRITE. Under LOCK TABLES, it's no longer possible to DROP a table or VIEW that was not locked for WRITE. ****** Backport of: ------------------------------------------------------------ revno: 2630.4.2 committer: Dmitry Lenev <dlenev@mysql.com> branch nick: mysql-6.0-3726-w timestamp: Sat 2008-05-24 14:03:45 +0400 message: WL#3726 "DDL locking for all metadata objects". After review fixes in progress. ****** Backport of: ------------------------------------------------------------ revno: 2630.4.3 committer: Dmitry Lenev <dlenev@mysql.com> branch nick: mysql-6.0-3726-w timestamp: Sat 2008-05-24 14:08:51 +0400 message: WL#3726 "DDL locking for all metadata objects" Fixed failing Windows builds by adding mdl.cc to the lists of files needed to build server/libmysqld on Windows. ****** Backport of: ------------------------------------------------------------ revno: 2630.4.4 committer: Dmitry Lenev <dlenev@mysql.com> branch nick: mysql-6.0-3726-w timestamp: Sat 2008-05-24 21:57:58 +0400 message: WL#3726 "DDL locking for all metadata objects". Fix for assert failures in kill.test which occured when one tried to kill ALTER TABLE statement on merge table while it was waiting in wait_while_table_is_used() for other connections to close this table. These assert failures stemmed from the fact that cleanup code in this case assumed that temporary table representing new version of table was open with adding to THD::temporary_tables list while code which were opening this temporary table wasn't always fulfilling this. This patch changes code that opens new version of table to always do this linking in. It also streamlines cleanup process for cases when error occurs while we have new version of table open. ****** WL#3726 "DDL locking for all metadata objects" Add libmysqld/mdl.cc to .bzrignore. ****** Backport of: ------------------------------------------------------------ revno: 2630.4.6 committer: Dmitry Lenev <dlenev@mysql.com> branch nick: mysql-6.0-3726-w timestamp: Sun 2008-05-25 00:33:22 +0400 message: WL#3726 "DDL locking for all metadata objects". Addition to the fix of assert failures in kill.test caused by changes for this worklog. Make sure we close the new table only once. .bzrignore: Add libmysqld/mdl.cc libmysqld/CMakeLists.txt: Added mdl.cc to the list of files needed for building of libmysqld. libmysqld/Makefile.am: Added files implementing new meta-data locking subsystem to the server. mysql-test/include/handler.inc: Use separate connection for waiting while threads performing DDL operations conflicting with open HANDLER tables reach blocked state. This is required because now we check and close tables open by HANDLER statements in this connection conflicting with DDL in another each time open_tables() is called and thus select from I_S which is used for waiting will unblock DDL operations if issued from connection with open HANDLERs. mysql-test/r/create.result: Adjusted test case after change in implementation of CREATE TABLE ... SELECT. We no longer have special check in open_table() which catches the case when we select from the table created. Instead we rely on unique_table() call which happens after opening and locking all tables. mysql-test/r/flush.result: FLUSH TABLES WITH READ LOCK can no longer happen under LOCK TABLES. Updated test accordingly. mysql-test/r/flush_table.result: Under LOCK TABLES we no longer allow to do FLUSH TABLES for tables locked for read. Updated test accordingly. mysql-test/r/handler_innodb.result: Use separate connection for waiting while threads performing DDL operations conflicting with open HANDLER tables reach blocked state. This is required because now we check and close tables open by HANDLER statements in this connection conflicting with DDL in another each time open_tables() is called and thus select from I_S which is used for waiting will unblock DDL operations if issued from connection with open HANDLERs. mysql-test/r/handler_myisam.result: Use separate connection for waiting while threads performing DDL operations conflicting with open HANDLER tables reach blocked state. This is required because now we check and close tables open by HANDLER statements in this connection conflicting with DDL in another each time open_tables() is called and thus select from I_S which is used for waiting will unblock DDL operations if issued from connection with open HANDLERs. mysql-test/r/information_schema.result: Additional test for WL#3726 "DDL locking for all metadata objects". Check that we use high-priority metadata lock requests when filling I_S tables. Rearrange tests to match 6.0 better (fewer merge conflicts). mysql-test/r/kill.result: Added tests checking that DDL and DML statements waiting for metadata locks can be interrupted by KILL command. mysql-test/r/lock.result: One no longer is allowed to do DROP VIEW under LOCK TABLES even if this view is locked by LOCK TABLES. The problem is that in such situation write locks on view are not mutually exclusive so upgrading metadata lock which is required for dropping of view will lead to deadlock. mysql-test/r/partition_column_prune.result: Update results (same results in 6.0), WL#3726 mysql-test/r/partition_pruning.result: Update results (same results in 6.0), WL#3726 mysql-test/r/ps_ddl.result: We no longer invalidate prepared CREATE TABLE ... SELECT statement if target table changes. This is OK since it is not strictly necessary. The first change is wrong, is caused by FLUSH TABLE now flushing all unused tables. This is a regression that Dmitri fixed in 6.0 in a follow up patch. mysql-test/r/sp.result: Under LOCK TABLES we no longer allow accessing views which were not explicitly locked. To access view we need to obtain metadata lock on it and doing this under LOCK TABLES may lead to deadlocks. mysql-test/r/view.result: One no longer is allowed to do DROP VIEW under LOCK TABLES even if this view is locked by LOCK TABLES. The problem is that in such situation even "write locks" on view are not mutually exclusive so upgrading metadata lock which is required for dropping of view will lead to deadlock mysql-test/r/view_grant.result: ALTER VIEW implementation was changed to open a view only after checking that user which does alter has appropriate privileges on it. This means that in case when user's privileges are insufficient for this we won't check that new view definer is the same as original one or user performing alter has SUPER privilege. Adjusted test case accordingly. mysql-test/r/view_multi.result: Added test case for bug#25144 "replication / binlog with view breaks". mysql-test/suite/rpl/t/disabled.def: Disable test for deprecated features (they don't work with new MDL). mysql-test/t/create.test: Adjusted test case after change in implementation of CREATE TABLE ... SELECT. We no longer have special check in open_table() which catches the case when we select from the table created. Instead we rely on unique_table() call which happens after opening and locking all tables. mysql-test/t/disabled.def: Disable merge.test, subject of WL#4144 mysql-test/t/flush.test: FLUSH TABLES WITH READ LOCK can no longer happen under LOCK TABLES. Updated test accordingly. mysql-test/t/flush_table.test: Under LOCK TABLES we no longer allow to do FLUSH TABLES for tables locked for read. Updated test accordingly. mysql-test/t/information_schema.test: Additional test for WL#3726 "DDL locking for all metadata objects". Check that we use high-priority metadata lock requests when filling I_S tables. Rearrange the results for easier merges with 6.0. mysql-test/t/kill.test: Added tests checking that DDL and DML statements waiting for metadata locks can be interrupted by KILL command. mysql-test/t/lock.test: One no longer is allowed to do DROP VIEW under LOCK TABLES even if this view is locked by LOCK TABLES. The problem is that in such situation write locks on view are not mutually exclusive so upgrading metadata lock which is required for dropping of view will lead to deadlock. mysql-test/t/lock_multi.test: Adjusted test case to the changes of status in various places caused by change in implementation FLUSH TABLES WITH READ LOCK, which is now takes global metadata lock before flushing tables and therefore waits on at these places. mysql-test/t/ps_ddl.test: We no longer invalidate prepared CREATE TABLE ... SELECT statement if target table changes. This is OK since it is not strictly necessary. The first change is wrong, is caused by FLUSH TABLE now flushing all unused tables. This is a regression that Dmitri fixed in 6.0 in a follow up patch. mysql-test/t/sp.test: Under LOCK TABLES we no longer allow accessing views which were not explicitly locked. To access view we need to obtain metadata lock on it and doing this under LOCK TABLES may lead to deadlocks. mysql-test/t/trigger_notembedded.test: Adjusted test case to the changes of status in various places caused by change in implementation FLUSH TABLES WITH READ LOCK, which is now takes global metadata lock before flushing tables and therefore waits on at these places. mysql-test/t/view.test: One no longer is allowed to do DROP VIEW under LOCK TABLES even if this view is locked by LOCK TABLES. The problem is that in such situation even "write locks" on view are not mutually exclusive so upgrading metadata lock which is required for dropping of view will lead to deadlock. mysql-test/t/view_grant.test: ALTER VIEW implementation was changed to open a view only after checking that user which does alter has appropriate privileges on it. This means that in case when user's privileges are insufficient for this we won't check that new view definer is the same as original one or user performing alter has SUPER privilege. Adjusted test case accordingly. mysql-test/t/view_multi.test: Added test case for bug#25144 "replication / binlog with view breaks". sql/CMakeLists.txt: Added mdl.cc to the list of files needed for building of server. sql/Makefile.am: Added files implementing new meta-data locking subsystem to the server. sql/event_db_repository.cc: Allocate metadata lock requests objects (MDL_LOCK) on execution memory root in cases when TABLE_LIST objects is also allocated there or on stack. sql/ha_ndbcluster.cc: Adjusted code to work nicely with new metadata locking subsystem. close_cached_tables() no longer has wait_for_placeholder argument. Instead of relying on this parameter and related behavior FLUSH TABLES WITH READ LOCK now takes global shared metadata lock. sql/ha_ndbcluster_binlog.cc: Adjusted code to work with new metadata locking subsystem. close_cached_tables() no longer has wait_for_placeholder argument. Instead of relying on this parameter and related behavior FLUSH TABLES WITH READ LOCK now takes global shared metadata lock. sql/handler.cc: update_frm_version(): Directly update TABLE_SHARE::mysql_version member instead of going through all TABLE instances for this table (old code was a legacy from pre-table-definition-cache days). sql/lock.cc: Use new metadata locking subsystem. Threw away most of functions related to name locking as now one is supposed to use metadata locking API instead. In lock_global_read_lock() and unlock_global_read_lock() in order to avoid problems with global read lock sneaking in at the moment when we perform FLUSH TABLES or ALTER TABLE under LOCK TABLES and when tables being reopened are protected only by metadata locks we also have to take global shared meta data lock. sql/log_event.cc: Adjusted code to work with new metadata locking subsystem. For tables open by slave thread for applying RBR events allocate memory for lock request object in the same chunk of memory as TABLE_LIST objects for them. In order to ensure that we keep these objects around until tables are open always close tables before calling Relay_log_info::clear_tables_to_lock(). Use new auxiliary Relay_log_info::slave_close_thread_tables() method to enforce this. sql/log_event_old.cc: Adjusted code to work with new metadata locking subsystem. Since for tables open by slave thread for applying RBR events memory for lock request object is allocated in the same chunk of memory as TABLE_LIST objects for them we have to ensure that we keep these objects around until tables are open. To ensure this we always close tables before calling Relay_log_info::clear_tables_to_lock(). To enfore this we use new auxiliary Relay_log_info::slave_close_thread_tables() method. sql/mdl.cc: Implemented new metadata locking subsystem and API described in WL3726 "DDL locking for all metadata objects". sql/mdl.h: Implemented new metadata locking subsystem and API described in WL3726 "DDL locking for all metadata objects". sql/mysql_priv.h: - close_thread_tables()/close_tables_for_reopen() now has one more argument which indicates that metadata locks should be released but not removed from the context in order to be used later in mdl_wait_for_locks() and tdc_wait_for_old_version(). - close_cached_table() routine is no longer public. - Thread waiting in wait_while_table_is_used() can be now killed so this function returns boolean to make caller aware of such situation. - We no longer have table cache as separate entity instead used and unused TABLE instances are linked to TABLE_SHARE objects in table definition cache. - Now third argument of open_table() is also used for requesting table repair or auto-discovery of table's new definition. So its type was changed from bool to enum. - Added tdc_open_view() function for opening view by getting its definition from disk (and table cache in future). - reopen_name_locked_table() no longer needs "link_in" argument as now we have exclusive metadata locks instead of dummy TABLE instances when this function is called. - find_locked_table() now takes head of list of TABLE instances instead of always scanning through THD::open_tables list. Also added find_write_locked_table() auxiliary. - reopen_tables(), close_cached_tables() no longer have mark_share_as_old and wait_for_placeholder arguments. Instead of relying on this parameters and related behavior FLUSH TABLES WITH READ LOCK now takes global shared metadata lock. - We no longer need drop_locked_tables() and abort_locked_tables(). - mysql_ha_rm_tables() now always assume that LOCK_open is not acquired by caller. - Added notify_thread_having_shared_lock() callback invoked by metadata locking subsystem when acquiring an exclusive lock, for each thread that has a conflicting shared metadata lock. - Introduced expel_table_from_cache() as replacement for remove_table_from_cache() (the main difference is that this new function assumes that caller follows metadata locking protocol and never waits). - Threw away most of functions related to name locking. One should use new metadata locking subsystem and API instead. sql/mysqld.cc: Got rid of call initializing/deinitializing table cache since now it is embedded into table definition cache. Added calls for initializing/ deinitializing metadata locking subsystem. sql/rpl_rli.cc: Introduced auxiliary Relay_log_info::slave_close_thread_tables() method which is used for enforcing that we always close tables open for RBR before deallocating TABLE_LIST elements and MDL_LOCK objects for them. sql/rpl_rli.h: Introduced auxiliary Relay_log_info::slave_close_thread_tables() method which is used for enforcing that we always close tables open for RBR before deallocating TABLE_LIST elements and MDL_LOCK objects for them. sql/set_var.cc: close_cached_tables() no longer has wait_for_placeholder argument. Instead of relying on this parameter and related behavior FLUSH TABLES WITH READ LOCK now takes global shared metadata lock. sql/sp_head.cc: For tables added to the statement's table list by prelocking algorithm we allocate these objects either on the same memory as corresponding table list elements or on THD::locked_tables_root (if we are building table list for LOCK TABLES). sql/sql_acl.cc: Allocate metadata lock requests objects (MDL_LOCK) on execution memory root in cases when we use stack TABLE_LIST objects to open tables. Got rid of redundant code by using unlock_locked_tables() function. sql/sql_base.cc: Changed code to use new MDL subsystem. Got rid of separate table cache. Now used and unused TABLE instances are linked to the TABLE_SHAREs in table definition cache. check_unused(): Adjusted code to the fact that we no longer have separate table cache. Removed dead code. table_def_free(): Free TABLE instances referenced from TABLE_SHARE objects before destroying table definition cache. get_table_share(): Added assert which ensures that noone will be able to access table (and its share) without acquiring some kind of metadata lock first. close_handle_and_leave_table_as_lock(): Adjusted code to the fact that TABLE instances now are linked to list in TABLE_SHARE. list_open_tables(): Changed this function to use table definition cache instead of table cache. free_cache_entry(): Unlink freed TABLE elements from the list of all TABLE instances for the table in TABLE_SHARE. kill_delayed_thread_for_table(): Added auxiliary for killing delayed insert threads for particular table. close_cached_tables(): Got rid of wait_for_refresh argument as we now rely on global shared metadata lock to prevent FLUSH WITH READ LOCK sneaking in when we are reopening tables. Heavily reworked this function to use new MDL code and not to rely on separate table cache entity. close_open_tables(): We no longer have separate table cache. close_thread_tables(): Release metadata locks after closing all tables. Added skip_mdl argument which allows us not to remove metadata lock requests from the context in case when we are going to use this requests later in mdl_wait_for_locks() and tdc_wait_for_old_versions(). close_thread_table()/close_table_for_reopen(): Since we no longer have separate table cache and all TABLE instances are linked to TABLE_SHARE objects in table definition cache we have to link/unlink TABLE object to/from appropriate lists in the share. name_lock_locked_table(): Moved redundant code to find_write_locked_table() function and adjusted code to the fact that wait_while_table_is_used() can now return with an error if our thread is killed. reopen_table_entry(): We no longer need "link_in" argument as with MDL we no longer call this function with dummy TABLE object pre-allocated and added to the THD::open_tables. Also now we add newly-open TABLE instance to the list of share's used TABLE instances. table_cache_insert_placeholder(): Got rid of name-locking legacy. lock_table_name_if_not_cached(): Moved to sql_table.cc the only place where it is used. It was also reimplemented using new MDL API. open_table(): - Reworked this function to use new MDL subsystem. - Changed code to deal with table definition cache directly instead of going through separate table cache. - Now third argument is also used for requesting table repair or auto-discovery of table's new definition. So its type was changed from bool to enum. find_locked_table()/find_write_locked_table(): Accept head of list of TABLE objects as first argument and use this list instead of always searching in THD::open_tables list. Also added auxiliary for finding write-locked locked tables. reopen_table(): Adjusted function to work with new MDL subsystem and to properly manuipulate with lists of used/unused TABLE instaces in TABLE_SHARE. reopen_tables(): Removed mark_share_as_old parameter. Instead of relying on it and related behavior FLUSH TABLES WITH READ LOCK now takes global shared metadata lock. Changed code after removing separate table cache. drop_locked_tables()/abort_locked_tables(): Got rid of functions which are no longer needed. unlock_locked_tables(): Moved this function from sql_parse.cc and changed it to release memory which was used for allocating metadata lock requests for tables open and locked by LOCK TABLES. tdc_open_view(): Intoduced function for opening a view by getting its definition from disk (and table cache in future). reopen_table_entry(): Introduced function for opening table definitions while holding exclusive metatadata lock on it. open_unireg_entry(): Got rid of this function. Most of its functionality is relocated to open_table() and open_table_fini() functions, and some of it to reopen_table_entry() and tdc_open_view(). Also code resposible for auto-repair and auto-discovery of tables was moved to separate function. open_table_entry_fini(): Introduced function which contains common actions which finalize process of TABLE object creation. auto_repair_table(): Moved code responsible for auto-repair of table being opened here. handle_failed_open_table_attempt() Moved code responsible for handling failing attempt to open table to one place (retry due to lock conflict/old version, auto-discovery and repair). open_tables(): - Flush open HANDLER tables if they have old version of if there is conflicting metadata lock against them (before this moment we had this code in open_table()). - When we open view which should be processed via derived table on the second execution of prepared statement or stored routine we still should call open_table() for it in order to obtain metadata lock on it and prepare its security context. - In cases when we discover that some special handling of failure to open table is needed call handle_failed_open_table_attempt() which handles all such scenarios. open_ltable(): Handling of various special scenarios of failure to open a table was moved to separate handle_failed_open_table_attempt() function. remove_db_from_cache(): Removed this function as it is no longer used. notify_thread_having_shared_lock(): Added callback which is invoked by MDL subsystem when acquiring an exclusive lock, for each thread that has a conflicting shared metadata lock. expel_table_from_cache(): Introduced function for removing unused TABLE instances. Unlike remove_table_from_cache() it relies on caller following MDL protocol and having appropriate locks when calling it and thus does not do any waiting if table is still in use. tdc_wait_for_old_version(): Added function which allows open_tables() to wait in cases when we discover that we should back-off due to presence of old version of table. abort_and_upgrade_lock(): Use new MDL calls. mysql_wait_completed_table(): Got rid of unused function. open_system_tables_for_read/for_update()/performance_schema_table(): Allocate MDL_LOCK objects on execution memory root in cases when TABLE_LIST objects for corresponding tables is allocated on stack. close_performance_schema_table(): Release metadata locks after closing tables. ****** Use I_P_List for free/used tables list in the table share. sql/sql_binlog.cc: Use Relay_log_info::slave_close_thread_tables() method to enforce that we always close tables open for RBR before deallocating TABLE_LIST elements and MDL_LOCK objects for them. sql/sql_class.cc: Added meta-data locking contexts as part of Open_tables_state context. Also introduced THD::locked_tables_root memory root which is to be used for allocating MDL_LOCK objects for tables in LOCK TABLES statement (end of lifetime for such objects is UNLOCK TABLES so we can't use statement or execution root for them). sql/sql_class.h: Added meta-data locking contexts as part of Open_tables_state context. Also introduced THD::locked_tables_root memory root which is to be used for allocating MDL_LOCK objects for tables in LOCK TABLES statement (end of lifetime for such objects is UNLOCK TABLES so we can't use statement or execution root for them). Note: handler_mdl_context and locked_tables_root and mdl_el_root will be removed by subsequent patches. sql/sql_db.cc: mysql_rm_db() does not really need to call remove_db_from_cache() as it drops each table in the database using mysql_rm_table_part2(), which performs all necessary operations on table (definition) cache. sql/sql_delete.cc: Use the new metadata locking API for TRUNCATE. sql/sql_handler.cc: Changed HANDLER implementation to use new metadata locking subsystem. Note that MDL_LOCK objects for HANDLER tables are allocated in the same chunk of heap memory as TABLE_LIST object for those tables. sql/sql_insert.cc: mysql_insert(): find_locked_table() now takes head of list of TABLE object as its argument instead of always scanning through THD::open_tables list. handle_delayed_insert(): Allocate metadata lock request object for table open by delayed insert thread on execution memroot. create_table_from_items(): We no longer allocate dummy TABLE objects for tables being created if they don't exist. As consequence reopen_name_locked_table() no longer has link_in argument. open_table() now has one more argument which is not relevant for temporary tables. sql/sql_parse.cc: - Moved unlock_locked_tables() routine to sql_base.cc and made available it in other files. Got rid of some redundant code by using this function. - Replaced boolean TABLE_LIST::create member with enum open_table_type member. - Use special memory root for allocating MDL_LOCK objects for tables open and locked by LOCK TABLES (these object should live till UNLOCK TABLES so we can't allocate them on statement nor execution memory root). Also properly set metadata lock upgradability attribure for those tables. - Under LOCK TABLES it is no longer allowed to flush tables which are not write-locked as this breaks metadata locking protocol and thus potentially might lead to deadlock. - Added auxiliary adjust_mdl_locks_upgradability() function. sql/sql_partition.cc: Adjusted code to the fact that reopen_tables() no longer has "mark_share_as_old" argument. Got rid of comments which are no longer true. sql/sql_plist.h: Added I_P_List template class for parametrized intrusive doubly linked lists and I_P_List_iterator for corresponding iterator. Unlike for I_List<> list elements of such list can participate in several lists. Unlike List<> such lists are doubly-linked and intrusive. sql/sql_plugin.cc: Allocate metadata lock requests objects (MDL_LOCK) on execution memory root in cases when we use stack TABLE_LIST objects to open tables. sql/sql_prepare.cc: Replaced boolean TABLE_LIST::create member with enum open_table_type member. This allows easily handle situation in which instead of opening the table we want only to take exclusive metadata lock on it. sql/sql_rename.cc: Use new metadata locking subsystem in implementation of RENAME TABLE. sql/sql_servers.cc: Allocate metadata lock requests objects (MDL_LOCK) on execution memory root in cases when we use stack TABLE_LIST objects to open tables. Got rid of redundant code by using unlock_locked_tables() function. sql/sql_show.cc: Acquire shared metadata lock when we are getting information for I_S table directly from TABLE_SHARE without doing full-blown table open. We use high priority lock request in this situation in order to avoid deadlocks. Also allocate metadata lock requests objects (MDL_LOCK) on execution memory root in cases when TABLE_LIST objects are also allocated there sql/sql_table.cc: mysql_rm_table(): Removed comment which is no longer relevant. mysql_rm_table_part2(): Now caller of mysql_ha_rm_tables() should not own LOCK_open. Adjusted code to use new metadata locking subsystem instead of name-locks. lock_table_name_if_not_cached(): Moved this function from sql_base.cc to this file and reimplemented it using metadata locking API. mysql_create_table(): Adjusted code to use new MDL API. wait_while_table_is_used(): Changed function to use new MDL subsystem. Made thread waiting in it killable (this also led to introduction of return value so caller can distinguish successful executions from situations when waiting was aborted). close_cached_tables(): Thread waiting in this function is killable now. As result it has return value for distinguishing between succes and failure. Got rid of redundant boradcast_refresh() call. prepare_for_repair(): Use MDL subsystem instead of name-locks. mysql_admin_table(): mysql_ha_rm_tables() now always assumes that caller doesn't own LOCK_open. mysql_repair_table(): We should mark all elements of table list as requiring upgradable metadata locks. mysql_create_table_like(): Use new MDL subsystem instead of name-locks. create_temporary_tables(): We don't need to obtain metadata locks when creating temporary table. mysql_fast_or_online_alter_table(): Thread waiting in wait_while_table_is_used() is now killable. mysql_alter_table(): Adjusted code to work with new MDL subsystem and to the fact that threads waiting in what_while_table_is_used() and close_cached_table() are now killable. sql/sql_test.cc: We no longer have separate table cache. TABLE instances are now associated with/linked to TABLE_SHARE objects in table definition cache. sql/sql_trigger.cc: Adjusted code to work with new metadata locking subsystem. Also reopen_tables() no longer has mark_share_as_old argument (Instead of relying on this parameter and related behavior FLUSH TABLES WITH READ LOCK now takes global shared metadata lock). sql/sql_udf.cc: Allocate metadata lock requests objects (MDL_LOCK) on execution memory root in cases when we use stack TABLE_LIST objects to open tables. sql/sql_update.cc: Adjusted code to work with new meta-data locking subsystem. sql/sql_view.cc: Added proper meta-data locking to implementations of CREATE/ALTER/DROP VIEW statements. Now we obtain exclusive meta-data lock on a view before creating/ changing/dropping it. This ensures that all concurrent statements that use this view will finish before our statement will proceed and therefore we will get correct order of statements in the binary log. Also ensure that TABLE_LIST::mdl_upgradable attribute is properly propagated for underlying tables of view. sql/table.cc: Added auxiliary alloc_mdl_locks() function for allocating metadata lock request objects for all elements of table list. sql/table.h: TABLE_SHARE: Got rid of unused members. Introduced members for storing lists of used and unused TABLE objects for this share. TABLE: Added members for linking TABLE objects into per-share lists of used and unused TABLE instances. Added member for holding pointer to metadata lock for this table. TABLE_LIST: Replaced boolean TABLE_LIST::create member with enum open_table_type member. This allows easily handle situation in which instead of opening the table we want only to take exclusive meta-data lock on it (we need this in order to handle ALTER VIEW and CREATE VIEW statements). Introduced new mdl_upgradable member for marking elements of table list for which we need to take upgradable shared metadata lock instead of plain shared metadata lock. Added pointer for holding pointer to MDL_LOCK for the table. Added auxiliary alloc_mdl_locks() function for allocating metadata lock requests objects for all elements of table list. Added auxiliary set_all_mdl_upgradable() function for marking all elements in table list as requiring upgradable metadata locks. storage/myisammrg/ha_myisammrg.cc: Allocate MDL_LOCK objects for underlying tables of MERGE table. To be reworked once Ingo pushes his patch for WL4144.
4018 lines
112 KiB
C++
4018 lines
112 KiB
C++
/* Copyright (C) 2000-2008 MySQL AB, 2008-2009 Sun Microsystems, Inc.
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; version 2 of the License.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */
|
|
|
|
|
|
/*****************************************************************************
|
|
**
|
|
** This file implements classes defined in sql_class.h
|
|
** Especially the classes to handle a result from a select
|
|
**
|
|
*****************************************************************************/
|
|
|
|
#ifdef USE_PRAGMA_IMPLEMENTATION
|
|
#pragma implementation // gcc: Class implementation
|
|
#endif
|
|
|
|
#include "mysql_priv.h"
|
|
#include "rpl_rli.h"
|
|
#include "rpl_filter.h"
|
|
#include "rpl_record.h"
|
|
#include "slave.h"
|
|
#include <my_bitmap.h>
|
|
#include "log_event.h"
|
|
#include <m_ctype.h>
|
|
#include <sys/stat.h>
|
|
#include <thr_alarm.h>
|
|
#ifdef __WIN__
|
|
#include <io.h>
|
|
#endif
|
|
#include <mysys_err.h>
|
|
|
|
#include "sp_rcontext.h"
|
|
#include "sp_cache.h"
|
|
#include "debug_sync.h"
|
|
|
|
/*
|
|
The following is used to initialise Table_ident with a internal
|
|
table name
|
|
*/
|
|
char internal_table_name[2]= "*";
|
|
char empty_c_string[1]= {0}; /* used for not defined db */
|
|
|
|
const char * const THD::DEFAULT_WHERE= "field list";
|
|
|
|
|
|
/*****************************************************************************
|
|
** Instansiate templates
|
|
*****************************************************************************/
|
|
|
|
#ifdef HAVE_EXPLICIT_TEMPLATE_INSTANTIATION
|
|
/* Used templates */
|
|
template class List<Key>;
|
|
template class List_iterator<Key>;
|
|
template class List<Key_part_spec>;
|
|
template class List_iterator<Key_part_spec>;
|
|
template class List<Alter_drop>;
|
|
template class List_iterator<Alter_drop>;
|
|
template class List<Alter_column>;
|
|
template class List_iterator<Alter_column>;
|
|
#endif
|
|
|
|
/****************************************************************************
|
|
** User variables
|
|
****************************************************************************/
|
|
|
|
extern "C" uchar *get_var_key(user_var_entry *entry, size_t *length,
|
|
my_bool not_used __attribute__((unused)))
|
|
{
|
|
*length= entry->name.length;
|
|
return (uchar*) entry->name.str;
|
|
}
|
|
|
|
extern "C" void free_user_var(user_var_entry *entry)
|
|
{
|
|
char *pos= (char*) entry+ALIGN_SIZE(sizeof(*entry));
|
|
if (entry->value && entry->value != pos)
|
|
my_free(entry->value, MYF(0));
|
|
my_free((char*) entry,MYF(0));
|
|
}
|
|
|
|
bool Key_part_spec::operator==(const Key_part_spec& other) const
|
|
{
|
|
return length == other.length &&
|
|
!my_strcasecmp(system_charset_info, field_name.str,
|
|
other.field_name.str);
|
|
}
|
|
|
|
/**
|
|
Construct an (almost) deep copy of this key. Only those
|
|
elements that are known to never change are not copied.
|
|
If out of memory, a partial copy is returned and an error is set
|
|
in THD.
|
|
*/
|
|
|
|
Key::Key(const Key &rhs, MEM_ROOT *mem_root)
|
|
:type(rhs.type),
|
|
key_create_info(rhs.key_create_info),
|
|
columns(rhs.columns, mem_root),
|
|
name(rhs.name),
|
|
generated(rhs.generated)
|
|
{
|
|
list_copy_and_replace_each_value(columns, mem_root);
|
|
}
|
|
|
|
/**
|
|
Construct an (almost) deep copy of this foreign key. Only those
|
|
elements that are known to never change are not copied.
|
|
If out of memory, a partial copy is returned and an error is set
|
|
in THD.
|
|
*/
|
|
|
|
Foreign_key::Foreign_key(const Foreign_key &rhs, MEM_ROOT *mem_root)
|
|
:Key(rhs),
|
|
ref_table(rhs.ref_table),
|
|
ref_columns(rhs.ref_columns),
|
|
delete_opt(rhs.delete_opt),
|
|
update_opt(rhs.update_opt),
|
|
match_opt(rhs.match_opt)
|
|
{
|
|
list_copy_and_replace_each_value(ref_columns, mem_root);
|
|
}
|
|
|
|
/*
|
|
Test if a foreign key (= generated key) is a prefix of the given key
|
|
(ignoring key name, key type and order of columns)
|
|
|
|
NOTES:
|
|
This is only used to test if an index for a FOREIGN KEY exists
|
|
|
|
IMPLEMENTATION
|
|
We only compare field names
|
|
|
|
RETURN
|
|
0 Generated key is a prefix of other key
|
|
1 Not equal
|
|
*/
|
|
|
|
bool foreign_key_prefix(Key *a, Key *b)
|
|
{
|
|
/* Ensure that 'a' is the generated key */
|
|
if (a->generated)
|
|
{
|
|
if (b->generated && a->columns.elements > b->columns.elements)
|
|
swap_variables(Key*, a, b); // Put shorter key in 'a'
|
|
}
|
|
else
|
|
{
|
|
if (!b->generated)
|
|
return TRUE; // No foreign key
|
|
swap_variables(Key*, a, b); // Put generated key in 'a'
|
|
}
|
|
|
|
/* Test if 'a' is a prefix of 'b' */
|
|
if (a->columns.elements > b->columns.elements)
|
|
return TRUE; // Can't be prefix
|
|
|
|
List_iterator<Key_part_spec> col_it1(a->columns);
|
|
List_iterator<Key_part_spec> col_it2(b->columns);
|
|
const Key_part_spec *col1, *col2;
|
|
|
|
#ifdef ENABLE_WHEN_INNODB_CAN_HANDLE_SWAPED_FOREIGN_KEY_COLUMNS
|
|
while ((col1= col_it1++))
|
|
{
|
|
bool found= 0;
|
|
col_it2.rewind();
|
|
while ((col2= col_it2++))
|
|
{
|
|
if (*col1 == *col2)
|
|
{
|
|
found= TRUE;
|
|
break;
|
|
}
|
|
}
|
|
if (!found)
|
|
return TRUE; // Error
|
|
}
|
|
return FALSE; // Is prefix
|
|
#else
|
|
while ((col1= col_it1++))
|
|
{
|
|
col2= col_it2++;
|
|
if (!(*col1 == *col2))
|
|
return TRUE;
|
|
}
|
|
return FALSE; // Is prefix
|
|
#endif
|
|
}
|
|
|
|
|
|
/****************************************************************************
|
|
** Thread specific functions
|
|
****************************************************************************/
|
|
|
|
Open_tables_state::Open_tables_state(THD *thd, ulong version_arg)
|
|
:version(version_arg), state_flags(0U)
|
|
{
|
|
reset_open_tables_state(thd);
|
|
}
|
|
|
|
/*
|
|
The following functions form part of the C plugin API
|
|
*/
|
|
|
|
extern "C" int mysql_tmpfile(const char *prefix)
|
|
{
|
|
char filename[FN_REFLEN];
|
|
File fd = create_temp_file(filename, mysql_tmpdir, prefix,
|
|
#ifdef __WIN__
|
|
O_BINARY | O_TRUNC | O_SEQUENTIAL |
|
|
O_SHORT_LIVED |
|
|
#endif /* __WIN__ */
|
|
O_CREAT | O_EXCL | O_RDWR | O_TEMPORARY,
|
|
MYF(MY_WME));
|
|
if (fd >= 0) {
|
|
#ifndef __WIN__
|
|
/*
|
|
This can be removed once the following bug is fixed:
|
|
Bug #28903 create_temp_file() doesn't honor O_TEMPORARY option
|
|
(file not removed) (Unix)
|
|
*/
|
|
unlink(filename);
|
|
#endif /* !__WIN__ */
|
|
}
|
|
|
|
return fd;
|
|
}
|
|
|
|
|
|
extern "C"
|
|
int thd_in_lock_tables(const THD *thd)
|
|
{
|
|
return test(thd->in_lock_tables);
|
|
}
|
|
|
|
|
|
extern "C"
|
|
int thd_tablespace_op(const THD *thd)
|
|
{
|
|
return test(thd->tablespace_op);
|
|
}
|
|
|
|
|
|
extern "C"
|
|
const char *set_thd_proc_info(THD *thd, const char *info,
|
|
const char *calling_function,
|
|
const char *calling_file,
|
|
const unsigned int calling_line)
|
|
{
|
|
const char *old_info= thd->proc_info;
|
|
DBUG_PRINT("proc_info", ("%s:%d %s", calling_file, calling_line,
|
|
(info != NULL) ? info : "(null)"));
|
|
#if defined(ENABLED_PROFILING)
|
|
thd->profiling.status_change(info, calling_function, calling_file, calling_line);
|
|
#endif
|
|
thd->proc_info= info;
|
|
return old_info;
|
|
}
|
|
|
|
extern "C"
|
|
const char* thd_enter_cond(MYSQL_THD thd, pthread_cond_t *cond,
|
|
pthread_mutex_t *mutex, const char *msg)
|
|
{
|
|
if (!thd)
|
|
thd= current_thd;
|
|
|
|
return thd->enter_cond(cond, mutex, msg);
|
|
}
|
|
|
|
extern "C"
|
|
void thd_exit_cond(MYSQL_THD thd, const char *old_msg)
|
|
{
|
|
if (!thd)
|
|
thd= current_thd;
|
|
|
|
thd->exit_cond(old_msg);
|
|
return;
|
|
}
|
|
|
|
extern "C"
|
|
void **thd_ha_data(const THD *thd, const struct handlerton *hton)
|
|
{
|
|
return (void **) &thd->ha_data[hton->slot].ha_ptr;
|
|
}
|
|
|
|
extern "C"
|
|
long long thd_test_options(const THD *thd, long long test_options)
|
|
{
|
|
return thd->options & test_options;
|
|
}
|
|
|
|
extern "C"
|
|
int thd_sql_command(const THD *thd)
|
|
{
|
|
return (int) thd->lex->sql_command;
|
|
}
|
|
|
|
extern "C"
|
|
int thd_tx_isolation(const THD *thd)
|
|
{
|
|
return (int) thd->variables.tx_isolation;
|
|
}
|
|
|
|
extern "C"
|
|
void thd_inc_row_count(THD *thd)
|
|
{
|
|
thd->warning_info->inc_current_row_for_warning();
|
|
}
|
|
|
|
|
|
/**
|
|
Dumps a text description of a thread, its security context
|
|
(user, host) and the current query.
|
|
|
|
@param thd thread context
|
|
@param buffer pointer to preferred result buffer
|
|
@param length length of buffer
|
|
@param max_query_len how many chars of query to copy (0 for all)
|
|
|
|
@req LOCK_thread_count
|
|
|
|
@note LOCK_thread_count mutex is not necessary when the function is invoked on
|
|
the currently running thread (current_thd) or if the caller in some other
|
|
way guarantees that access to thd->query is serialized.
|
|
|
|
@return Pointer to string
|
|
*/
|
|
|
|
extern "C"
|
|
char *thd_security_context(THD *thd, char *buffer, unsigned int length,
|
|
unsigned int max_query_len)
|
|
{
|
|
String str(buffer, length, &my_charset_latin1);
|
|
const Security_context *sctx= &thd->main_security_ctx;
|
|
char header[64];
|
|
int len;
|
|
/*
|
|
The pointers thd->query and thd->proc_info might change since they are
|
|
being modified concurrently. This is acceptable for proc_info since its
|
|
values doesn't have to very accurate and the memory it points to is static,
|
|
but we need to attempt a snapshot on the pointer values to avoid using NULL
|
|
values. The pointer to thd->query however, doesn't point to static memory
|
|
and has to be protected by LOCK_thread_count or risk pointing to
|
|
uninitialized memory.
|
|
*/
|
|
const char *proc_info= thd->proc_info;
|
|
|
|
len= my_snprintf(header, sizeof(header),
|
|
"MySQL thread id %lu, query id %lu",
|
|
thd->thread_id, (ulong) thd->query_id);
|
|
str.length(0);
|
|
str.append(header, len);
|
|
|
|
if (sctx->host)
|
|
{
|
|
str.append(' ');
|
|
str.append(sctx->host);
|
|
}
|
|
|
|
if (sctx->ip)
|
|
{
|
|
str.append(' ');
|
|
str.append(sctx->ip);
|
|
}
|
|
|
|
if (sctx->user)
|
|
{
|
|
str.append(' ');
|
|
str.append(sctx->user);
|
|
}
|
|
|
|
if (proc_info)
|
|
{
|
|
str.append(' ');
|
|
str.append(proc_info);
|
|
}
|
|
|
|
if (thd->query())
|
|
{
|
|
if (max_query_len < 1)
|
|
len= thd->query_length();
|
|
else
|
|
len= min(thd->query_length(), max_query_len);
|
|
str.append('\n');
|
|
str.append(thd->query(), len);
|
|
}
|
|
if (str.c_ptr_safe() == buffer)
|
|
return buffer;
|
|
|
|
/*
|
|
We have to copy the new string to the destination buffer because the string
|
|
was reallocated to a larger buffer to be able to fit.
|
|
*/
|
|
DBUG_ASSERT(buffer != NULL);
|
|
length= min(str.length(), length-1);
|
|
memcpy(buffer, str.c_ptr_quick(), length);
|
|
/* Make sure that the new string is null terminated */
|
|
buffer[length]= '\0';
|
|
return buffer;
|
|
}
|
|
|
|
|
|
/**
|
|
Implementation of Drop_table_error_handler::handle_condition().
|
|
The reason in having this implementation is to silence technical low-level
|
|
warnings during DROP TABLE operation. Currently we don't want to expose
|
|
the following warnings during DROP TABLE:
|
|
- Some of table files are missed or invalid (the table is going to be
|
|
deleted anyway, so why bother that something was missed);
|
|
- A trigger associated with the table does not have DEFINER (One of the
|
|
MySQL specifics now is that triggers are loaded for the table being
|
|
dropped. So, we may have a warning that trigger does not have DEFINER
|
|
attribute during DROP TABLE operation).
|
|
|
|
@return TRUE if the condition is handled.
|
|
*/
|
|
bool Drop_table_error_handler::handle_condition(THD *thd,
|
|
uint sql_errno,
|
|
const char* sqlstate,
|
|
MYSQL_ERROR::enum_warning_level level,
|
|
const char* msg,
|
|
MYSQL_ERROR ** cond_hdl)
|
|
{
|
|
*cond_hdl= NULL;
|
|
return ((sql_errno == EE_DELETE && my_errno == ENOENT) ||
|
|
sql_errno == ER_TRG_NO_DEFINER);
|
|
}
|
|
|
|
|
|
THD::THD()
|
|
:Statement(&main_lex, &main_mem_root, CONVENTIONAL_EXECUTION,
|
|
/* statement id */ 0),
|
|
Open_tables_state(this, refresh_version), rli_fake(0),
|
|
lock_id(&main_lock_id),
|
|
user_time(0), in_sub_stmt(0),
|
|
sql_log_bin_toplevel(false),
|
|
binlog_table_maps(0), binlog_flags(0UL),
|
|
table_map_for_update(0),
|
|
arg_of_last_insert_id_function(FALSE),
|
|
first_successful_insert_id_in_prev_stmt(0),
|
|
first_successful_insert_id_in_prev_stmt_for_binlog(0),
|
|
first_successful_insert_id_in_cur_stmt(0),
|
|
stmt_depends_on_first_successful_insert_id_in_prev_stmt(FALSE),
|
|
examined_row_count(0),
|
|
warning_info(&main_warning_info),
|
|
stmt_da(&main_da),
|
|
global_read_lock(0),
|
|
is_fatal_error(0),
|
|
transaction_rollback_request(0),
|
|
is_fatal_sub_stmt_error(0),
|
|
rand_used(0),
|
|
time_zone_used(0),
|
|
in_lock_tables(0),
|
|
bootstrap(0),
|
|
derived_tables_processing(FALSE),
|
|
spcont(NULL),
|
|
m_parser_state(NULL),
|
|
#if defined(ENABLED_DEBUG_SYNC)
|
|
debug_sync_control(0),
|
|
#endif /* defined(ENABLED_DEBUG_SYNC) */
|
|
main_warning_info(0),
|
|
mdl_el_root(NULL)
|
|
{
|
|
ulong tmp;
|
|
|
|
/*
|
|
Pass nominal parameters to init_alloc_root only to ensure that
|
|
the destructor works OK in case of an error. The main_mem_root
|
|
will be re-initialized in init_for_queries().
|
|
*/
|
|
init_sql_alloc(&main_mem_root, ALLOC_ROOT_MIN_BLOCK_SIZE, 0);
|
|
stmt_arena= this;
|
|
thread_stack= 0;
|
|
catalog= (char*)"std"; // the only catalog we have for now
|
|
main_security_ctx.init();
|
|
security_ctx= &main_security_ctx;
|
|
some_tables_deleted=no_errors=password= 0;
|
|
query_start_used= 0;
|
|
count_cuted_fields= CHECK_FIELD_IGNORE;
|
|
killed= NOT_KILLED;
|
|
col_access=0;
|
|
is_slave_error= thread_specific_used= FALSE;
|
|
my_hash_clear(&handler_tables_hash);
|
|
tmp_table=0;
|
|
used_tables=0;
|
|
cuted_fields= 0L;
|
|
sent_row_count= 0L;
|
|
limit_found_rows= 0;
|
|
row_count_func= -1;
|
|
statement_id_counter= 0UL;
|
|
// Must be reset to handle error with THD's created for init of mysqld
|
|
lex->current_select= 0;
|
|
start_time=(time_t) 0;
|
|
start_utime= prior_thr_create_utime= 0L;
|
|
utime_after_lock= 0L;
|
|
current_linfo = 0;
|
|
slave_thread = 0;
|
|
bzero(&variables, sizeof(variables));
|
|
thread_id= 0;
|
|
one_shot_set= 0;
|
|
file_id = 0;
|
|
query_id= 0;
|
|
query_name_consts= 0;
|
|
db_charset= global_system_variables.collation_database;
|
|
bzero(ha_data, sizeof(ha_data));
|
|
mysys_var=0;
|
|
binlog_evt_union.do_union= FALSE;
|
|
enable_slow_log= 0;
|
|
#ifndef DBUG_OFF
|
|
dbug_sentry=THD_SENTRY_MAGIC;
|
|
#endif
|
|
#ifndef EMBEDDED_LIBRARY
|
|
net.vio=0;
|
|
#endif
|
|
client_capabilities= 0; // minimalistic client
|
|
ull=0;
|
|
system_thread= NON_SYSTEM_THREAD;
|
|
cleanup_done= abort_on_warning= no_warnings_for_error= 0;
|
|
peer_port= 0; // For SHOW PROCESSLIST
|
|
transaction.m_pending_rows_event= 0;
|
|
transaction.on= 1;
|
|
#ifdef SIGNAL_WITH_VIO_CLOSE
|
|
active_vio = 0;
|
|
#endif
|
|
pthread_mutex_init(&LOCK_thd_data, MY_MUTEX_INIT_FAST);
|
|
|
|
/* Variables with default values */
|
|
proc_info="login";
|
|
where= THD::DEFAULT_WHERE;
|
|
server_id = ::server_id;
|
|
slave_net = 0;
|
|
command=COM_CONNECT;
|
|
*scramble= '\0';
|
|
|
|
init();
|
|
#if defined(ENABLED_PROFILING)
|
|
profiling.set_thd(this);
|
|
#endif
|
|
user_connect=(USER_CONN *)0;
|
|
my_hash_init(&user_vars, system_charset_info, USER_VARS_HASH_SIZE, 0, 0,
|
|
(my_hash_get_key) get_var_key,
|
|
(my_hash_free_key) free_user_var, 0);
|
|
|
|
sp_proc_cache= NULL;
|
|
sp_func_cache= NULL;
|
|
|
|
/* For user vars replication*/
|
|
if (opt_bin_log)
|
|
my_init_dynamic_array(&user_var_events,
|
|
sizeof(BINLOG_USER_VAR_EVENT *), 16, 16);
|
|
else
|
|
bzero((char*) &user_var_events, sizeof(user_var_events));
|
|
|
|
/* Protocol */
|
|
protocol= &protocol_text; // Default protocol
|
|
protocol_text.init(this);
|
|
protocol_binary.init(this);
|
|
|
|
tablespace_op=FALSE;
|
|
tmp= sql_rnd_with_mutex();
|
|
randominit(&rand, tmp + (ulong) &rand, tmp + (ulong) ::global_query_id);
|
|
substitute_null_with_insert_id = FALSE;
|
|
thr_lock_info_init(&lock_info); /* safety: will be reset after start */
|
|
thr_lock_owner_init(&main_lock_id, &lock_info);
|
|
|
|
m_internal_handler= NULL;
|
|
|
|
init_sql_alloc(&locked_tables_root, ALLOC_ROOT_MIN_BLOCK_SIZE, 0);
|
|
}
|
|
|
|
|
|
void THD::push_internal_handler(Internal_error_handler *handler)
|
|
{
|
|
if (m_internal_handler)
|
|
{
|
|
handler->m_prev_internal_handler= m_internal_handler;
|
|
m_internal_handler= handler;
|
|
}
|
|
else
|
|
{
|
|
m_internal_handler= handler;
|
|
}
|
|
}
|
|
|
|
bool THD::handle_condition(uint sql_errno,
|
|
const char* sqlstate,
|
|
MYSQL_ERROR::enum_warning_level level,
|
|
const char* msg,
|
|
MYSQL_ERROR ** cond_hdl)
|
|
{
|
|
if (!m_internal_handler)
|
|
{
|
|
*cond_hdl= NULL;
|
|
return FALSE;
|
|
}
|
|
|
|
for (Internal_error_handler *error_handler= m_internal_handler;
|
|
error_handler;
|
|
error_handler= m_internal_handler->m_prev_internal_handler)
|
|
{
|
|
if (error_handler-> handle_condition(this, sql_errno, sqlstate, level, msg,
|
|
cond_hdl))
|
|
{
|
|
return TRUE;
|
|
}
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
|
|
void THD::pop_internal_handler()
|
|
{
|
|
DBUG_ASSERT(m_internal_handler != NULL);
|
|
m_internal_handler= m_internal_handler->m_prev_internal_handler;
|
|
}
|
|
|
|
|
|
void THD::raise_error(uint sql_errno)
|
|
{
|
|
const char* msg= ER(sql_errno);
|
|
(void) raise_condition(sql_errno,
|
|
NULL,
|
|
MYSQL_ERROR::WARN_LEVEL_ERROR,
|
|
msg);
|
|
}
|
|
|
|
void THD::raise_error_printf(uint sql_errno, ...)
|
|
{
|
|
va_list args;
|
|
char ebuff[MYSQL_ERRMSG_SIZE];
|
|
DBUG_ENTER("THD::raise_error_printf");
|
|
DBUG_PRINT("my", ("nr: %d errno: %d", sql_errno, errno));
|
|
const char* format= ER(sql_errno);
|
|
va_start(args, sql_errno);
|
|
my_vsnprintf(ebuff, sizeof(ebuff), format, args);
|
|
va_end(args);
|
|
(void) raise_condition(sql_errno,
|
|
NULL,
|
|
MYSQL_ERROR::WARN_LEVEL_ERROR,
|
|
ebuff);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
void THD::raise_warning(uint sql_errno)
|
|
{
|
|
const char* msg= ER(sql_errno);
|
|
(void) raise_condition(sql_errno,
|
|
NULL,
|
|
MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
msg);
|
|
}
|
|
|
|
void THD::raise_warning_printf(uint sql_errno, ...)
|
|
{
|
|
va_list args;
|
|
char ebuff[MYSQL_ERRMSG_SIZE];
|
|
DBUG_ENTER("THD::raise_warning_printf");
|
|
DBUG_PRINT("enter", ("warning: %u", sql_errno));
|
|
const char* format= ER(sql_errno);
|
|
va_start(args, sql_errno);
|
|
my_vsnprintf(ebuff, sizeof(ebuff), format, args);
|
|
va_end(args);
|
|
(void) raise_condition(sql_errno,
|
|
NULL,
|
|
MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
ebuff);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
void THD::raise_note(uint sql_errno)
|
|
{
|
|
DBUG_ENTER("THD::raise_note");
|
|
DBUG_PRINT("enter", ("code: %d", sql_errno));
|
|
if (!(this->options & OPTION_SQL_NOTES))
|
|
DBUG_VOID_RETURN;
|
|
const char* msg= ER(sql_errno);
|
|
(void) raise_condition(sql_errno,
|
|
NULL,
|
|
MYSQL_ERROR::WARN_LEVEL_NOTE,
|
|
msg);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
void THD::raise_note_printf(uint sql_errno, ...)
|
|
{
|
|
va_list args;
|
|
char ebuff[MYSQL_ERRMSG_SIZE];
|
|
DBUG_ENTER("THD::raise_note_printf");
|
|
DBUG_PRINT("enter",("code: %u", sql_errno));
|
|
if (!(this->options & OPTION_SQL_NOTES))
|
|
DBUG_VOID_RETURN;
|
|
const char* format= ER(sql_errno);
|
|
va_start(args, sql_errno);
|
|
my_vsnprintf(ebuff, sizeof(ebuff), format, args);
|
|
va_end(args);
|
|
(void) raise_condition(sql_errno,
|
|
NULL,
|
|
MYSQL_ERROR::WARN_LEVEL_NOTE,
|
|
ebuff);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
MYSQL_ERROR* THD::raise_condition(uint sql_errno,
|
|
const char* sqlstate,
|
|
MYSQL_ERROR::enum_warning_level level,
|
|
const char* msg)
|
|
{
|
|
MYSQL_ERROR *cond= NULL;
|
|
DBUG_ENTER("THD::raise_condition");
|
|
|
|
if (!(this->options & OPTION_SQL_NOTES) &&
|
|
(level == MYSQL_ERROR::WARN_LEVEL_NOTE))
|
|
DBUG_RETURN(NULL);
|
|
|
|
warning_info->opt_clear_warning_info(query_id);
|
|
|
|
/*
|
|
TODO: replace by DBUG_ASSERT(sql_errno != 0) once all bugs similar to
|
|
Bug#36768 are fixed: a SQL condition must have a real (!=0) error number
|
|
so that it can be caught by handlers.
|
|
*/
|
|
if (sql_errno == 0)
|
|
sql_errno= ER_UNKNOWN_ERROR;
|
|
if (msg == NULL)
|
|
msg= ER(sql_errno);
|
|
if (sqlstate == NULL)
|
|
sqlstate= mysql_errno_to_sqlstate(sql_errno);
|
|
|
|
if ((level == MYSQL_ERROR::WARN_LEVEL_WARN) &&
|
|
really_abort_on_warning())
|
|
{
|
|
/*
|
|
FIXME:
|
|
push_warning and strict SQL_MODE case.
|
|
*/
|
|
level= MYSQL_ERROR::WARN_LEVEL_ERROR;
|
|
killed= THD::KILL_BAD_DATA;
|
|
}
|
|
|
|
switch (level)
|
|
{
|
|
case MYSQL_ERROR::WARN_LEVEL_NOTE:
|
|
case MYSQL_ERROR::WARN_LEVEL_WARN:
|
|
got_warning= 1;
|
|
break;
|
|
case MYSQL_ERROR::WARN_LEVEL_ERROR:
|
|
break;
|
|
default:
|
|
DBUG_ASSERT(FALSE);
|
|
}
|
|
|
|
if (handle_condition(sql_errno, sqlstate, level, msg, &cond))
|
|
DBUG_RETURN(cond);
|
|
|
|
if (level == MYSQL_ERROR::WARN_LEVEL_ERROR)
|
|
{
|
|
is_slave_error= 1; // needed to catch query errors during replication
|
|
|
|
/*
|
|
thd->lex->current_select == 0 if lex structure is not inited
|
|
(not query command (COM_QUERY))
|
|
*/
|
|
if (lex->current_select &&
|
|
lex->current_select->no_error && !is_fatal_error)
|
|
{
|
|
DBUG_PRINT("error",
|
|
("Error converted to warning: current_select: no_error %d "
|
|
"fatal_error: %d",
|
|
(lex->current_select ?
|
|
lex->current_select->no_error : 0),
|
|
(int) is_fatal_error));
|
|
}
|
|
else
|
|
{
|
|
if (! stmt_da->is_error())
|
|
stmt_da->set_error_status(this, sql_errno, msg, sqlstate);
|
|
}
|
|
}
|
|
|
|
/*
|
|
If a continue handler is found, the error message will be cleared
|
|
by the stored procedures code.
|
|
*/
|
|
if (!is_fatal_error && spcont &&
|
|
spcont->handle_condition(this, sql_errno, sqlstate, level, msg, &cond))
|
|
{
|
|
/*
|
|
Do not push any warnings, a handled error must be completely
|
|
silenced.
|
|
*/
|
|
DBUG_RETURN(cond);
|
|
}
|
|
|
|
/* Un-handled conditions */
|
|
|
|
cond= raise_condition_no_handler(sql_errno, sqlstate, level, msg);
|
|
DBUG_RETURN(cond);
|
|
}
|
|
|
|
MYSQL_ERROR*
|
|
THD::raise_condition_no_handler(uint sql_errno,
|
|
const char* sqlstate,
|
|
MYSQL_ERROR::enum_warning_level level,
|
|
const char* msg)
|
|
{
|
|
MYSQL_ERROR *cond= NULL;
|
|
DBUG_ENTER("THD::raise_condition_no_handler");
|
|
|
|
query_cache_abort(&query_cache_tls);
|
|
|
|
/* FIXME: broken special case */
|
|
if (no_warnings_for_error && (level == MYSQL_ERROR::WARN_LEVEL_ERROR))
|
|
DBUG_RETURN(NULL);
|
|
|
|
cond= warning_info->push_warning(this, sql_errno, sqlstate, level, msg);
|
|
DBUG_RETURN(cond);
|
|
}
|
|
extern "C"
|
|
void *thd_alloc(MYSQL_THD thd, unsigned int size)
|
|
{
|
|
return thd->alloc(size);
|
|
}
|
|
|
|
extern "C"
|
|
void *thd_calloc(MYSQL_THD thd, unsigned int size)
|
|
{
|
|
return thd->calloc(size);
|
|
}
|
|
|
|
extern "C"
|
|
char *thd_strdup(MYSQL_THD thd, const char *str)
|
|
{
|
|
return thd->strdup(str);
|
|
}
|
|
|
|
extern "C"
|
|
char *thd_strmake(MYSQL_THD thd, const char *str, unsigned int size)
|
|
{
|
|
return thd->strmake(str, size);
|
|
}
|
|
|
|
extern "C"
|
|
LEX_STRING *thd_make_lex_string(THD *thd, LEX_STRING *lex_str,
|
|
const char *str, unsigned int size,
|
|
int allocate_lex_string)
|
|
{
|
|
return thd->make_lex_string(lex_str, str, size,
|
|
(bool) allocate_lex_string);
|
|
}
|
|
|
|
extern "C"
|
|
void *thd_memdup(MYSQL_THD thd, const void* str, unsigned int size)
|
|
{
|
|
return thd->memdup(str, size);
|
|
}
|
|
|
|
extern "C"
|
|
void thd_get_xid(const MYSQL_THD thd, MYSQL_XID *xid)
|
|
{
|
|
*xid = *(MYSQL_XID *) &thd->transaction.xid_state.xid;
|
|
}
|
|
|
|
#ifdef _WIN32
|
|
extern "C" THD *_current_thd_noinline(void)
|
|
{
|
|
return my_pthread_getspecific_ptr(THD*,THR_THD);
|
|
}
|
|
#endif
|
|
/*
|
|
Init common variables that has to be reset on start and on change_user
|
|
*/
|
|
|
|
void THD::init(void)
|
|
{
|
|
pthread_mutex_lock(&LOCK_global_system_variables);
|
|
plugin_thdvar_init(this);
|
|
variables.time_format= date_time_format_copy((THD*) 0,
|
|
variables.time_format);
|
|
variables.date_format= date_time_format_copy((THD*) 0,
|
|
variables.date_format);
|
|
variables.datetime_format= date_time_format_copy((THD*) 0,
|
|
variables.datetime_format);
|
|
/*
|
|
variables= global_system_variables above has reset
|
|
variables.pseudo_thread_id to 0. We need to correct it here to
|
|
avoid temporary tables replication failure.
|
|
*/
|
|
variables.pseudo_thread_id= thread_id;
|
|
pthread_mutex_unlock(&LOCK_global_system_variables);
|
|
server_status= SERVER_STATUS_AUTOCOMMIT;
|
|
if (variables.sql_mode & MODE_NO_BACKSLASH_ESCAPES)
|
|
server_status|= SERVER_STATUS_NO_BACKSLASH_ESCAPES;
|
|
options= thd_startup_options;
|
|
|
|
if (variables.max_join_size == HA_POS_ERROR)
|
|
options |= OPTION_BIG_SELECTS;
|
|
else
|
|
options &= ~OPTION_BIG_SELECTS;
|
|
|
|
transaction.all.modified_non_trans_table= transaction.stmt.modified_non_trans_table= FALSE;
|
|
open_options=ha_open_options;
|
|
update_lock_default= (variables.low_priority_updates ?
|
|
TL_WRITE_LOW_PRIORITY :
|
|
TL_WRITE);
|
|
session_tx_isolation= (enum_tx_isolation) variables.tx_isolation;
|
|
update_charset();
|
|
reset_current_stmt_binlog_row_based();
|
|
bzero((char *) &status_var, sizeof(status_var));
|
|
sql_log_bin_toplevel= options & OPTION_BIN_LOG;
|
|
|
|
#if defined(ENABLED_DEBUG_SYNC)
|
|
/* Initialize the Debug Sync Facility. See debug_sync.cc. */
|
|
debug_sync_init_thread(this);
|
|
#endif /* defined(ENABLED_DEBUG_SYNC) */
|
|
}
|
|
|
|
|
|
/*
|
|
Init THD for query processing.
|
|
This has to be called once before we call mysql_parse.
|
|
See also comments in sql_class.h.
|
|
*/
|
|
|
|
void THD::init_for_queries()
|
|
{
|
|
set_time();
|
|
ha_enable_transaction(this,TRUE);
|
|
|
|
reset_root_defaults(mem_root, variables.query_alloc_block_size,
|
|
variables.query_prealloc_size);
|
|
reset_root_defaults(&transaction.mem_root,
|
|
variables.trans_alloc_block_size,
|
|
variables.trans_prealloc_size);
|
|
transaction.xid_state.xid.null();
|
|
transaction.xid_state.in_thd=1;
|
|
}
|
|
|
|
|
|
/*
|
|
Do what's needed when one invokes change user
|
|
|
|
SYNOPSIS
|
|
change_user()
|
|
|
|
IMPLEMENTATION
|
|
Reset all resources that are connection specific
|
|
*/
|
|
|
|
|
|
void THD::change_user(void)
|
|
{
|
|
pthread_mutex_lock(&LOCK_status);
|
|
add_to_status(&global_status_var, &status_var);
|
|
pthread_mutex_unlock(&LOCK_status);
|
|
|
|
cleanup();
|
|
killed= NOT_KILLED;
|
|
cleanup_done= 0;
|
|
init();
|
|
stmt_map.reset();
|
|
my_hash_init(&user_vars, system_charset_info, USER_VARS_HASH_SIZE, 0, 0,
|
|
(my_hash_get_key) get_var_key,
|
|
(my_hash_free_key) free_user_var, 0);
|
|
sp_cache_clear(&sp_proc_cache);
|
|
sp_cache_clear(&sp_func_cache);
|
|
}
|
|
|
|
|
|
/* Do operations that may take a long time */
|
|
|
|
void THD::cleanup(void)
|
|
{
|
|
DBUG_ENTER("THD::cleanup");
|
|
DBUG_ASSERT(cleanup_done == 0);
|
|
|
|
killed= KILL_CONNECTION;
|
|
#ifdef ENABLE_WHEN_BINLOG_WILL_BE_ABLE_TO_PREPARE
|
|
if (transaction.xid_state.xa_state == XA_PREPARED)
|
|
{
|
|
#error xid_state in the cache should be replaced by the allocated value
|
|
}
|
|
#endif
|
|
{
|
|
ha_rollback(this);
|
|
xid_cache_delete(&transaction.xid_state);
|
|
}
|
|
if (locked_tables)
|
|
{
|
|
lock=locked_tables; locked_tables=0;
|
|
close_thread_tables(this);
|
|
}
|
|
|
|
#if defined(ENABLED_DEBUG_SYNC)
|
|
/* End the Debug Sync Facility. See debug_sync.cc. */
|
|
debug_sync_end_thread(this);
|
|
#endif /* defined(ENABLED_DEBUG_SYNC) */
|
|
|
|
mysql_ha_cleanup(this);
|
|
delete_dynamic(&user_var_events);
|
|
my_hash_free(&user_vars);
|
|
close_temporary_tables(this);
|
|
my_free((char*) variables.time_format, MYF(MY_ALLOW_ZERO_PTR));
|
|
my_free((char*) variables.date_format, MYF(MY_ALLOW_ZERO_PTR));
|
|
my_free((char*) variables.datetime_format, MYF(MY_ALLOW_ZERO_PTR));
|
|
|
|
sp_cache_clear(&sp_proc_cache);
|
|
sp_cache_clear(&sp_func_cache);
|
|
|
|
if (global_read_lock)
|
|
unlock_global_read_lock(this);
|
|
if (ull)
|
|
{
|
|
pthread_mutex_lock(&LOCK_user_locks);
|
|
item_user_lock_release(ull);
|
|
pthread_mutex_unlock(&LOCK_user_locks);
|
|
ull= NULL;
|
|
}
|
|
|
|
cleanup_done=1;
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
THD::~THD()
|
|
{
|
|
THD_CHECK_SENTRY(this);
|
|
DBUG_ENTER("~THD()");
|
|
/* Ensure that no one is using THD */
|
|
pthread_mutex_lock(&LOCK_thd_data);
|
|
pthread_mutex_unlock(&LOCK_thd_data);
|
|
add_to_status(&global_status_var, &status_var);
|
|
|
|
/* Close connection */
|
|
#ifndef EMBEDDED_LIBRARY
|
|
if (net.vio)
|
|
{
|
|
vio_delete(net.vio);
|
|
net_end(&net);
|
|
}
|
|
#endif
|
|
stmt_map.reset(); /* close all prepared statements */
|
|
DBUG_ASSERT(lock_info.n_cursors == 0);
|
|
if (!cleanup_done)
|
|
cleanup();
|
|
|
|
mdl_context_destroy(&mdl_context);
|
|
mdl_context_destroy(&handler_mdl_context);
|
|
|
|
ha_close_connection(this);
|
|
plugin_thdvar_cleanup(this);
|
|
|
|
DBUG_PRINT("info", ("freeing security context"));
|
|
main_security_ctx.destroy();
|
|
safeFree(db);
|
|
free_root(&transaction.mem_root,MYF(0));
|
|
mysys_var=0; // Safety (shouldn't be needed)
|
|
pthread_mutex_destroy(&LOCK_thd_data);
|
|
#ifndef DBUG_OFF
|
|
dbug_sentry= THD_SENTRY_GONE;
|
|
#endif
|
|
#ifndef EMBEDDED_LIBRARY
|
|
if (rli_fake)
|
|
{
|
|
delete rli_fake;
|
|
rli_fake= NULL;
|
|
}
|
|
#endif
|
|
|
|
free_root(&main_mem_root, MYF(0));
|
|
free_root(&locked_tables_root, MYF(0));
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
/*
|
|
Add all status variables to another status variable array
|
|
|
|
SYNOPSIS
|
|
add_to_status()
|
|
to_var add to this array
|
|
from_var from this array
|
|
|
|
NOTES
|
|
This function assumes that all variables are long/ulong.
|
|
If this assumption will change, then we have to explictely add
|
|
the other variables after the while loop
|
|
*/
|
|
|
|
void add_to_status(STATUS_VAR *to_var, STATUS_VAR *from_var)
|
|
{
|
|
ulong *end= (ulong*) ((uchar*) to_var +
|
|
offsetof(STATUS_VAR, last_system_status_var) +
|
|
sizeof(ulong));
|
|
ulong *to= (ulong*) to_var, *from= (ulong*) from_var;
|
|
|
|
while (to != end)
|
|
*(to++)+= *(from++);
|
|
}
|
|
|
|
/*
|
|
Add the difference between two status variable arrays to another one.
|
|
|
|
SYNOPSIS
|
|
add_diff_to_status
|
|
to_var add to this array
|
|
from_var from this array
|
|
dec_var minus this array
|
|
|
|
NOTE
|
|
This function assumes that all variables are long/ulong.
|
|
*/
|
|
|
|
void add_diff_to_status(STATUS_VAR *to_var, STATUS_VAR *from_var,
|
|
STATUS_VAR *dec_var)
|
|
{
|
|
ulong *end= (ulong*) ((uchar*) to_var + offsetof(STATUS_VAR,
|
|
last_system_status_var) +
|
|
sizeof(ulong));
|
|
ulong *to= (ulong*) to_var, *from= (ulong*) from_var, *dec= (ulong*) dec_var;
|
|
|
|
while (to != end)
|
|
*(to++)+= *(from++) - *(dec++);
|
|
}
|
|
|
|
|
|
void THD::awake(THD::killed_state state_to_set)
|
|
{
|
|
DBUG_ENTER("THD::awake");
|
|
DBUG_PRINT("enter", ("this: 0x%lx", (long) this));
|
|
THD_CHECK_SENTRY(this);
|
|
safe_mutex_assert_owner(&LOCK_thd_data);
|
|
|
|
killed= state_to_set;
|
|
if (state_to_set != THD::KILL_QUERY)
|
|
{
|
|
thr_alarm_kill(thread_id);
|
|
if (!slave_thread)
|
|
thread_scheduler.post_kill_notification(this);
|
|
#ifdef SIGNAL_WITH_VIO_CLOSE
|
|
if (this != current_thd)
|
|
{
|
|
/*
|
|
In addition to a signal, let's close the socket of the thread that
|
|
is being killed. This is to make sure it does not block if the
|
|
signal is lost. This needs to be done only on platforms where
|
|
signals are not a reliable interruption mechanism.
|
|
|
|
If we're killing ourselves, we know that we're not blocked, so this
|
|
hack is not used.
|
|
*/
|
|
|
|
close_active_vio();
|
|
}
|
|
#endif
|
|
}
|
|
if (mysys_var)
|
|
{
|
|
pthread_mutex_lock(&mysys_var->mutex);
|
|
if (!system_thread) // Don't abort locks
|
|
mysys_var->abort=1;
|
|
/*
|
|
This broadcast could be up in the air if the victim thread
|
|
exits the cond in the time between read and broadcast, but that is
|
|
ok since all we want to do is to make the victim thread get out
|
|
of waiting on current_cond.
|
|
If we see a non-zero current_cond: it cannot be an old value (because
|
|
then exit_cond() should have run and it can't because we have mutex); so
|
|
it is the true value but maybe current_mutex is not yet non-zero (we're
|
|
in the middle of enter_cond() and there is a "memory order
|
|
inversion"). So we test the mutex too to not lock 0.
|
|
|
|
Note that there is a small chance we fail to kill. If victim has locked
|
|
current_mutex, but hasn't yet entered enter_cond() (which means that
|
|
current_cond and current_mutex are 0), then the victim will not get
|
|
a signal and it may wait "forever" on the cond (until
|
|
we issue a second KILL or the status it's waiting for happens).
|
|
It's true that we have set its thd->killed but it may not
|
|
see it immediately and so may have time to reach the cond_wait().
|
|
*/
|
|
if (mysys_var->current_cond && mysys_var->current_mutex)
|
|
{
|
|
pthread_mutex_lock(mysys_var->current_mutex);
|
|
pthread_cond_broadcast(mysys_var->current_cond);
|
|
pthread_mutex_unlock(mysys_var->current_mutex);
|
|
}
|
|
pthread_mutex_unlock(&mysys_var->mutex);
|
|
}
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
/*
|
|
Remember the location of thread info, the structure needed for
|
|
sql_alloc() and the structure for the net buffer
|
|
*/
|
|
|
|
bool THD::store_globals()
|
|
{
|
|
/*
|
|
Assert that thread_stack is initialized: it's necessary to be able
|
|
to track stack overrun.
|
|
*/
|
|
DBUG_ASSERT(thread_stack);
|
|
|
|
if (my_pthread_setspecific_ptr(THR_THD, this) ||
|
|
my_pthread_setspecific_ptr(THR_MALLOC, &mem_root))
|
|
return 1;
|
|
mysys_var=my_thread_var;
|
|
/*
|
|
Let mysqld define the thread id (not mysys)
|
|
This allows us to move THD to different threads if needed.
|
|
*/
|
|
mysys_var->id= thread_id;
|
|
real_id= pthread_self(); // For debugging
|
|
|
|
/*
|
|
We have to call thr_lock_info_init() again here as THD may have been
|
|
created in another thread
|
|
*/
|
|
thr_lock_info_init(&lock_info);
|
|
return 0;
|
|
}
|
|
|
|
|
|
/*
|
|
Cleanup after query.
|
|
|
|
SYNOPSIS
|
|
THD::cleanup_after_query()
|
|
|
|
DESCRIPTION
|
|
This function is used to reset thread data to its default state.
|
|
|
|
NOTE
|
|
This function is not suitable for setting thread data to some
|
|
non-default values, as there is only one replication thread, so
|
|
different master threads may overwrite data of each other on
|
|
slave.
|
|
*/
|
|
|
|
void THD::cleanup_after_query()
|
|
{
|
|
/*
|
|
Reset rand_used so that detection of calls to rand() will save random
|
|
seeds if needed by the slave.
|
|
|
|
Do not reset rand_used if inside a stored function or trigger because
|
|
only the call to these operations is logged. Thus only the calling
|
|
statement needs to detect rand() calls made by its substatements. These
|
|
substatements must not set rand_used to 0 because it would remove the
|
|
detection of rand() by the calling statement.
|
|
*/
|
|
if (!in_sub_stmt) /* stored functions and triggers are a special case */
|
|
{
|
|
/* Forget those values, for next binlogger: */
|
|
stmt_depends_on_first_successful_insert_id_in_prev_stmt= 0;
|
|
auto_inc_intervals_in_cur_stmt_for_binlog.empty();
|
|
rand_used= 0;
|
|
}
|
|
if (first_successful_insert_id_in_cur_stmt > 0)
|
|
{
|
|
/* set what LAST_INSERT_ID() will return */
|
|
first_successful_insert_id_in_prev_stmt=
|
|
first_successful_insert_id_in_cur_stmt;
|
|
first_successful_insert_id_in_cur_stmt= 0;
|
|
substitute_null_with_insert_id= TRUE;
|
|
}
|
|
arg_of_last_insert_id_function= 0;
|
|
/* Free Items that were created during this execution */
|
|
free_items();
|
|
/* Reset where. */
|
|
where= THD::DEFAULT_WHERE;
|
|
/* reset table map for multi-table update */
|
|
table_map_for_update= 0;
|
|
}
|
|
|
|
|
|
/**
|
|
Create a LEX_STRING in this connection.
|
|
|
|
@param lex_str pointer to LEX_STRING object to be initialized
|
|
@param str initializer to be copied into lex_str
|
|
@param length length of str, in bytes
|
|
@param allocate_lex_string if TRUE, allocate new LEX_STRING object,
|
|
instead of using lex_str value
|
|
@return NULL on failure, or pointer to the LEX_STRING object
|
|
*/
|
|
LEX_STRING *THD::make_lex_string(LEX_STRING *lex_str,
|
|
const char* str, uint length,
|
|
bool allocate_lex_string)
|
|
{
|
|
if (allocate_lex_string)
|
|
if (!(lex_str= (LEX_STRING *)alloc(sizeof(LEX_STRING))))
|
|
return 0;
|
|
if (!(lex_str->str= strmake_root(mem_root, str, length)))
|
|
return 0;
|
|
lex_str->length= length;
|
|
return lex_str;
|
|
}
|
|
|
|
|
|
/*
|
|
Convert a string to another character set
|
|
|
|
SYNOPSIS
|
|
convert_string()
|
|
to Store new allocated string here
|
|
to_cs New character set for allocated string
|
|
from String to convert
|
|
from_length Length of string to convert
|
|
from_cs Original character set
|
|
|
|
NOTES
|
|
to will be 0-terminated to make it easy to pass to system funcs
|
|
|
|
RETURN
|
|
0 ok
|
|
1 End of memory.
|
|
In this case to->str will point to 0 and to->length will be 0.
|
|
*/
|
|
|
|
bool THD::convert_string(LEX_STRING *to, CHARSET_INFO *to_cs,
|
|
const char *from, uint from_length,
|
|
CHARSET_INFO *from_cs)
|
|
{
|
|
DBUG_ENTER("convert_string");
|
|
size_t new_length= to_cs->mbmaxlen * from_length;
|
|
uint dummy_errors;
|
|
if (!(to->str= (char*) alloc(new_length+1)))
|
|
{
|
|
to->length= 0; // Safety fix
|
|
DBUG_RETURN(1); // EOM
|
|
}
|
|
to->length= copy_and_convert((char*) to->str, new_length, to_cs,
|
|
from, from_length, from_cs, &dummy_errors);
|
|
to->str[to->length]=0; // Safety
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
|
|
/*
|
|
Convert string from source character set to target character set inplace.
|
|
|
|
SYNOPSIS
|
|
THD::convert_string
|
|
|
|
DESCRIPTION
|
|
Convert string using convert_buffer - buffer for character set
|
|
conversion shared between all protocols.
|
|
|
|
RETURN
|
|
0 ok
|
|
!0 out of memory
|
|
*/
|
|
|
|
bool THD::convert_string(String *s, CHARSET_INFO *from_cs, CHARSET_INFO *to_cs)
|
|
{
|
|
uint dummy_errors;
|
|
if (convert_buffer.copy(s->ptr(), s->length(), from_cs, to_cs, &dummy_errors))
|
|
return TRUE;
|
|
/* If convert_buffer >> s copying is more efficient long term */
|
|
if (convert_buffer.alloced_length() >= convert_buffer.length() * 2 ||
|
|
!s->is_alloced())
|
|
{
|
|
return s->copy(convert_buffer);
|
|
}
|
|
s->swap(convert_buffer);
|
|
return FALSE;
|
|
}
|
|
|
|
|
|
/*
|
|
Update some cache variables when character set changes
|
|
*/
|
|
|
|
void THD::update_charset()
|
|
{
|
|
uint32 not_used;
|
|
charset_is_system_charset= !String::needs_conversion(0,charset(),
|
|
system_charset_info,
|
|
¬_used);
|
|
charset_is_collation_connection=
|
|
!String::needs_conversion(0,charset(),variables.collation_connection,
|
|
¬_used);
|
|
charset_is_character_set_filesystem=
|
|
!String::needs_conversion(0, charset(),
|
|
variables.character_set_filesystem, ¬_used);
|
|
}
|
|
|
|
|
|
/* routings to adding tables to list of changed in transaction tables */
|
|
|
|
inline static void list_include(CHANGED_TABLE_LIST** prev,
|
|
CHANGED_TABLE_LIST* curr,
|
|
CHANGED_TABLE_LIST* new_table)
|
|
{
|
|
if (new_table)
|
|
{
|
|
*prev = new_table;
|
|
(*prev)->next = curr;
|
|
}
|
|
}
|
|
|
|
/* add table to list of changed in transaction tables */
|
|
|
|
void THD::add_changed_table(TABLE *table)
|
|
{
|
|
DBUG_ENTER("THD::add_changed_table(table)");
|
|
|
|
DBUG_ASSERT((options & (OPTION_NOT_AUTOCOMMIT | OPTION_BEGIN)) &&
|
|
table->file->has_transactions());
|
|
add_changed_table(table->s->table_cache_key.str,
|
|
(long) table->s->table_cache_key.length);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
void THD::add_changed_table(const char *key, long key_length)
|
|
{
|
|
DBUG_ENTER("THD::add_changed_table(key)");
|
|
CHANGED_TABLE_LIST **prev_changed = &transaction.changed_tables;
|
|
CHANGED_TABLE_LIST *curr = transaction.changed_tables;
|
|
|
|
for (; curr; prev_changed = &(curr->next), curr = curr->next)
|
|
{
|
|
int cmp = (long)curr->key_length - (long)key_length;
|
|
if (cmp < 0)
|
|
{
|
|
list_include(prev_changed, curr, changed_table_dup(key, key_length));
|
|
DBUG_PRINT("info",
|
|
("key_length: %ld %u", key_length,
|
|
(*prev_changed)->key_length));
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
else if (cmp == 0)
|
|
{
|
|
cmp = memcmp(curr->key, key, curr->key_length);
|
|
if (cmp < 0)
|
|
{
|
|
list_include(prev_changed, curr, changed_table_dup(key, key_length));
|
|
DBUG_PRINT("info",
|
|
("key_length: %ld %u", key_length,
|
|
(*prev_changed)->key_length));
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
else if (cmp == 0)
|
|
{
|
|
DBUG_PRINT("info", ("already in list"));
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
}
|
|
}
|
|
*prev_changed = changed_table_dup(key, key_length);
|
|
DBUG_PRINT("info", ("key_length: %ld %u", key_length,
|
|
(*prev_changed)->key_length));
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
CHANGED_TABLE_LIST* THD::changed_table_dup(const char *key, long key_length)
|
|
{
|
|
CHANGED_TABLE_LIST* new_table =
|
|
(CHANGED_TABLE_LIST*) trans_alloc(ALIGN_SIZE(sizeof(CHANGED_TABLE_LIST))+
|
|
key_length + 1);
|
|
if (!new_table)
|
|
{
|
|
my_error(EE_OUTOFMEMORY, MYF(ME_BELL),
|
|
ALIGN_SIZE(sizeof(TABLE_LIST)) + key_length + 1);
|
|
killed= KILL_CONNECTION;
|
|
return 0;
|
|
}
|
|
|
|
new_table->key= ((char*)new_table)+ ALIGN_SIZE(sizeof(CHANGED_TABLE_LIST));
|
|
new_table->next = 0;
|
|
new_table->key_length = key_length;
|
|
::memcpy(new_table->key, key, key_length);
|
|
return new_table;
|
|
}
|
|
|
|
|
|
int THD::send_explain_fields(select_result *result)
|
|
{
|
|
List<Item> field_list;
|
|
Item *item;
|
|
CHARSET_INFO *cs= system_charset_info;
|
|
field_list.push_back(new Item_return_int("id",3, MYSQL_TYPE_LONGLONG));
|
|
field_list.push_back(new Item_empty_string("select_type", 19, cs));
|
|
field_list.push_back(item= new Item_empty_string("table", NAME_CHAR_LEN, cs));
|
|
item->maybe_null= 1;
|
|
if (lex->describe & DESCRIBE_PARTITIONS)
|
|
{
|
|
/* Maximum length of string that make_used_partitions_str() can produce */
|
|
item= new Item_empty_string("partitions", MAX_PARTITIONS * (1 + FN_LEN),
|
|
cs);
|
|
field_list.push_back(item);
|
|
item->maybe_null= 1;
|
|
}
|
|
field_list.push_back(item= new Item_empty_string("type", 10, cs));
|
|
item->maybe_null= 1;
|
|
field_list.push_back(item=new Item_empty_string("possible_keys",
|
|
NAME_CHAR_LEN*MAX_KEY, cs));
|
|
item->maybe_null=1;
|
|
field_list.push_back(item=new Item_empty_string("key", NAME_CHAR_LEN, cs));
|
|
item->maybe_null=1;
|
|
field_list.push_back(item=new Item_empty_string("key_len",
|
|
NAME_CHAR_LEN*MAX_KEY));
|
|
item->maybe_null=1;
|
|
field_list.push_back(item=new Item_empty_string("ref",
|
|
NAME_CHAR_LEN*MAX_REF_PARTS,
|
|
cs));
|
|
item->maybe_null=1;
|
|
field_list.push_back(item= new Item_return_int("rows", 10,
|
|
MYSQL_TYPE_LONGLONG));
|
|
if (lex->describe & DESCRIBE_EXTENDED)
|
|
{
|
|
field_list.push_back(item= new Item_float("filtered", 0.1234, 2, 4));
|
|
item->maybe_null=1;
|
|
}
|
|
item->maybe_null= 1;
|
|
field_list.push_back(new Item_empty_string("Extra", 255, cs));
|
|
return (result->send_result_set_metadata(field_list,
|
|
Protocol::SEND_NUM_ROWS | Protocol::SEND_EOF));
|
|
}
|
|
|
|
#ifdef SIGNAL_WITH_VIO_CLOSE
|
|
void THD::close_active_vio()
|
|
{
|
|
DBUG_ENTER("close_active_vio");
|
|
safe_mutex_assert_owner(&LOCK_thd_data);
|
|
#ifndef EMBEDDED_LIBRARY
|
|
if (active_vio)
|
|
{
|
|
vio_close(active_vio);
|
|
active_vio = 0;
|
|
}
|
|
#endif
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
#endif
|
|
|
|
|
|
struct Item_change_record: public ilink
|
|
{
|
|
Item **place;
|
|
Item *old_value;
|
|
/* Placement new was hidden by `new' in ilink (TODO: check): */
|
|
static void *operator new(size_t size, void *mem) { return mem; }
|
|
static void operator delete(void *ptr, size_t size) {}
|
|
static void operator delete(void *ptr, void *mem) { /* never called */ }
|
|
};
|
|
|
|
|
|
/*
|
|
Register an item tree tree transformation, performed by the query
|
|
optimizer. We need a pointer to runtime_memroot because it may be !=
|
|
thd->mem_root (due to possible set_n_backup_active_arena called for thd).
|
|
*/
|
|
|
|
void THD::nocheck_register_item_tree_change(Item **place, Item *old_value,
|
|
MEM_ROOT *runtime_memroot)
|
|
{
|
|
Item_change_record *change;
|
|
/*
|
|
Now we use one node per change, which adds some memory overhead,
|
|
but still is rather fast as we use alloc_root for allocations.
|
|
A list of item tree changes of an average query should be short.
|
|
*/
|
|
void *change_mem= alloc_root(runtime_memroot, sizeof(*change));
|
|
if (change_mem == 0)
|
|
{
|
|
/*
|
|
OOM, thd->fatal_error() is called by the error handler of the
|
|
memroot. Just return.
|
|
*/
|
|
return;
|
|
}
|
|
change= new (change_mem) Item_change_record;
|
|
change->place= place;
|
|
change->old_value= old_value;
|
|
change_list.append(change);
|
|
}
|
|
|
|
|
|
void THD::rollback_item_tree_changes()
|
|
{
|
|
I_List_iterator<Item_change_record> it(change_list);
|
|
Item_change_record *change;
|
|
DBUG_ENTER("rollback_item_tree_changes");
|
|
|
|
while ((change= it++))
|
|
*change->place= change->old_value;
|
|
/* We can forget about changes memory: it's allocated in runtime memroot */
|
|
change_list.empty();
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
/*****************************************************************************
|
|
** Functions to provide a interface to select results
|
|
*****************************************************************************/
|
|
|
|
select_result::select_result()
|
|
{
|
|
thd=current_thd;
|
|
nest_level= -1;
|
|
}
|
|
|
|
void select_result::send_error(uint errcode,const char *err)
|
|
{
|
|
my_message(errcode, err, MYF(0));
|
|
}
|
|
|
|
|
|
void select_result::cleanup()
|
|
{
|
|
/* do nothing */
|
|
}
|
|
|
|
bool select_result::check_simple_select() const
|
|
{
|
|
my_error(ER_SP_BAD_CURSOR_QUERY, MYF(0));
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
static String default_line_term("\n",default_charset_info);
|
|
static String default_escaped("\\",default_charset_info);
|
|
static String default_field_term("\t",default_charset_info);
|
|
static String default_xml_row_term("<row>", default_charset_info);
|
|
|
|
sql_exchange::sql_exchange(char *name, bool flag,
|
|
enum enum_filetype filetype_arg)
|
|
:file_name(name), opt_enclosed(0), dumpfile(flag), skip_lines(0)
|
|
{
|
|
filetype= filetype_arg;
|
|
field_term= &default_field_term;
|
|
enclosed= line_start= &my_empty_string;
|
|
line_term= filetype == FILETYPE_CSV ?
|
|
&default_line_term : &default_xml_row_term;
|
|
escaped= &default_escaped;
|
|
cs= NULL;
|
|
}
|
|
|
|
bool sql_exchange::escaped_given(void)
|
|
{
|
|
return escaped != &default_escaped;
|
|
}
|
|
|
|
|
|
bool select_send::send_result_set_metadata(List<Item> &list, uint flags)
|
|
{
|
|
bool res;
|
|
if (!(res= thd->protocol->send_result_set_metadata(&list, flags)))
|
|
is_result_set_started= 1;
|
|
return res;
|
|
}
|
|
|
|
void select_send::abort()
|
|
{
|
|
DBUG_ENTER("select_send::abort");
|
|
|
|
if (is_result_set_started && thd->spcont)
|
|
{
|
|
/*
|
|
We're executing a stored procedure, have an open result
|
|
set and an SQL exception condition. In this situation we
|
|
must abort the current statement, silence the error and
|
|
start executing the continue/exit handler if one is found.
|
|
Before aborting the statement, let's end the open result set, as
|
|
otherwise the client will hang due to the violation of the
|
|
client/server protocol.
|
|
*/
|
|
thd->spcont->end_partial_result_set= TRUE;
|
|
}
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
/**
|
|
Cleanup an instance of this class for re-use
|
|
at next execution of a prepared statement/
|
|
stored procedure statement.
|
|
*/
|
|
|
|
void select_send::cleanup()
|
|
{
|
|
is_result_set_started= FALSE;
|
|
}
|
|
|
|
/* Send data to client. Returns 0 if ok */
|
|
|
|
bool select_send::send_data(List<Item> &items)
|
|
{
|
|
Protocol *protocol= thd->protocol;
|
|
DBUG_ENTER("select_send::send_data");
|
|
|
|
if (unit->offset_limit_cnt)
|
|
{ // using limit offset,count
|
|
unit->offset_limit_cnt--;
|
|
DBUG_RETURN(FALSE);
|
|
}
|
|
|
|
/*
|
|
We may be passing the control from mysqld to the client: release the
|
|
InnoDB adaptive hash S-latch to avoid thread deadlocks if it was reserved
|
|
by thd
|
|
*/
|
|
ha_release_temporary_latches(thd);
|
|
|
|
protocol->prepare_for_resend();
|
|
if (protocol->send_result_set_row(&items))
|
|
{
|
|
protocol->remove_last_row();
|
|
DBUG_RETURN(TRUE);
|
|
}
|
|
|
|
thd->sent_row_count++;
|
|
|
|
if (thd->vio_ok())
|
|
DBUG_RETURN(protocol->write());
|
|
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
bool select_send::send_eof()
|
|
{
|
|
/*
|
|
We may be passing the control from mysqld to the client: release the
|
|
InnoDB adaptive hash S-latch to avoid thread deadlocks if it was reserved
|
|
by thd
|
|
*/
|
|
ha_release_temporary_latches(thd);
|
|
|
|
/* Unlock tables before sending packet to gain some speed */
|
|
if (thd->lock)
|
|
{
|
|
mysql_unlock_tables(thd, thd->lock);
|
|
thd->lock=0;
|
|
}
|
|
/*
|
|
Don't send EOF if we're in error condition (which implies we've already
|
|
sent or are sending an error)
|
|
*/
|
|
if (thd->is_error())
|
|
return TRUE;
|
|
::my_eof(thd);
|
|
is_result_set_started= 0;
|
|
return FALSE;
|
|
}
|
|
|
|
|
|
/************************************************************************
|
|
Handling writing to file
|
|
************************************************************************/
|
|
|
|
void select_to_file::send_error(uint errcode,const char *err)
|
|
{
|
|
my_message(errcode, err, MYF(0));
|
|
if (file > 0)
|
|
{
|
|
(void) end_io_cache(&cache);
|
|
(void) my_close(file,MYF(0));
|
|
(void) my_delete(path,MYF(0)); // Delete file on error
|
|
file= -1;
|
|
}
|
|
}
|
|
|
|
|
|
bool select_to_file::send_eof()
|
|
{
|
|
int error= test(end_io_cache(&cache));
|
|
if (my_close(file,MYF(MY_WME)))
|
|
error= 1;
|
|
if (!error)
|
|
{
|
|
/*
|
|
In order to remember the value of affected rows for ROW_COUNT()
|
|
function, SELECT INTO has to have an own SQLCOM.
|
|
TODO: split from SQLCOM_SELECT
|
|
*/
|
|
::my_ok(thd,row_count);
|
|
}
|
|
file= -1;
|
|
return error;
|
|
}
|
|
|
|
|
|
void select_to_file::cleanup()
|
|
{
|
|
/* In case of error send_eof() may be not called: close the file here. */
|
|
if (file >= 0)
|
|
{
|
|
(void) end_io_cache(&cache);
|
|
(void) my_close(file,MYF(0));
|
|
file= -1;
|
|
}
|
|
path[0]= '\0';
|
|
row_count= 0;
|
|
}
|
|
|
|
|
|
select_to_file::~select_to_file()
|
|
{
|
|
if (file >= 0)
|
|
{ // This only happens in case of error
|
|
(void) end_io_cache(&cache);
|
|
(void) my_close(file,MYF(0));
|
|
file= -1;
|
|
}
|
|
}
|
|
|
|
/***************************************************************************
|
|
** Export of select to textfile
|
|
***************************************************************************/
|
|
|
|
select_export::~select_export()
|
|
{
|
|
thd->sent_row_count=row_count;
|
|
}
|
|
|
|
|
|
/*
|
|
Create file with IO cache
|
|
|
|
SYNOPSIS
|
|
create_file()
|
|
thd Thread handle
|
|
path File name
|
|
exchange Excange class
|
|
cache IO cache
|
|
|
|
RETURN
|
|
>= 0 File handle
|
|
-1 Error
|
|
*/
|
|
|
|
|
|
static File create_file(THD *thd, char *path, sql_exchange *exchange,
|
|
IO_CACHE *cache)
|
|
{
|
|
File file;
|
|
uint option= MY_UNPACK_FILENAME | MY_RELATIVE_PATH;
|
|
|
|
#ifdef DONT_ALLOW_FULL_LOAD_DATA_PATHS
|
|
option|= MY_REPLACE_DIR; // Force use of db directory
|
|
#endif
|
|
|
|
if (!dirname_length(exchange->file_name))
|
|
{
|
|
strxnmov(path, FN_REFLEN-1, mysql_real_data_home, thd->db ? thd->db : "",
|
|
NullS);
|
|
(void) fn_format(path, exchange->file_name, path, "", option);
|
|
}
|
|
else
|
|
(void) fn_format(path, exchange->file_name, mysql_real_data_home, "", option);
|
|
|
|
if (opt_secure_file_priv &&
|
|
strncmp(opt_secure_file_priv, path, strlen(opt_secure_file_priv)))
|
|
{
|
|
/* Write only allowed to dir or subdir specified by secure_file_priv */
|
|
my_error(ER_OPTION_PREVENTS_STATEMENT, MYF(0), "--secure-file-priv");
|
|
return -1;
|
|
}
|
|
|
|
if (!access(path, F_OK))
|
|
{
|
|
my_error(ER_FILE_EXISTS_ERROR, MYF(0), exchange->file_name);
|
|
return -1;
|
|
}
|
|
/* Create the file world readable */
|
|
if ((file= my_create(path, 0666, O_WRONLY|O_EXCL, MYF(MY_WME))) < 0)
|
|
return file;
|
|
#ifdef HAVE_FCHMOD
|
|
(void) fchmod(file, 0666); // Because of umask()
|
|
#else
|
|
(void) chmod(path, 0666);
|
|
#endif
|
|
if (init_io_cache(cache, file, 0L, WRITE_CACHE, 0L, 1, MYF(MY_WME)))
|
|
{
|
|
my_close(file, MYF(0));
|
|
my_delete(path, MYF(0)); // Delete file on error, it was just created
|
|
return -1;
|
|
}
|
|
return file;
|
|
}
|
|
|
|
|
|
int
|
|
select_export::prepare(List<Item> &list, SELECT_LEX_UNIT *u)
|
|
{
|
|
bool blob_flag=0;
|
|
bool string_results= FALSE, non_string_results= FALSE;
|
|
unit= u;
|
|
if ((uint) strlen(exchange->file_name) + NAME_LEN >= FN_REFLEN)
|
|
strmake(path,exchange->file_name,FN_REFLEN-1);
|
|
|
|
write_cs= exchange->cs ? exchange->cs : &my_charset_bin;
|
|
|
|
if ((file= create_file(thd, path, exchange, &cache)) < 0)
|
|
return 1;
|
|
/* Check if there is any blobs in data */
|
|
{
|
|
List_iterator_fast<Item> li(list);
|
|
Item *item;
|
|
while ((item=li++))
|
|
{
|
|
if (item->max_length >= MAX_BLOB_WIDTH)
|
|
{
|
|
blob_flag=1;
|
|
break;
|
|
}
|
|
if (item->result_type() == STRING_RESULT)
|
|
string_results= TRUE;
|
|
else
|
|
non_string_results= TRUE;
|
|
}
|
|
}
|
|
if (exchange->escaped->numchars() > 1 || exchange->enclosed->numchars() > 1)
|
|
{
|
|
my_error(ER_WRONG_FIELD_TERMINATORS, MYF(0));
|
|
return TRUE;
|
|
}
|
|
if (exchange->escaped->length() > 1 || exchange->enclosed->length() > 1 ||
|
|
!my_isascii(exchange->escaped->ptr()[0]) ||
|
|
!my_isascii(exchange->enclosed->ptr()[0]) ||
|
|
!exchange->field_term->is_ascii() || !exchange->line_term->is_ascii() ||
|
|
!exchange->line_start->is_ascii())
|
|
{
|
|
/*
|
|
Current LOAD DATA INFILE recognizes field/line separators "as is" without
|
|
converting from client charset to data file charset. So, it is supposed,
|
|
that input file of LOAD DATA INFILE consists of data in one charset and
|
|
separators in other charset. For the compatibility with that [buggy]
|
|
behaviour SELECT INTO OUTFILE implementation has been saved "as is" too,
|
|
but the new warning message has been added:
|
|
|
|
Non-ASCII separator arguments are not fully supported
|
|
*/
|
|
push_warning(thd, MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
WARN_NON_ASCII_SEPARATOR_NOT_IMPLEMENTED,
|
|
ER(WARN_NON_ASCII_SEPARATOR_NOT_IMPLEMENTED));
|
|
}
|
|
field_term_length=exchange->field_term->length();
|
|
field_term_char= field_term_length ?
|
|
(int) (uchar) (*exchange->field_term)[0] : INT_MAX;
|
|
if (!exchange->line_term->length())
|
|
exchange->line_term=exchange->field_term; // Use this if it exists
|
|
field_sep_char= (exchange->enclosed->length() ?
|
|
(int) (uchar) (*exchange->enclosed)[0] : field_term_char);
|
|
if (exchange->escaped->length() && (exchange->escaped_given() ||
|
|
!(thd->variables.sql_mode & MODE_NO_BACKSLASH_ESCAPES)))
|
|
escape_char= (int) (uchar) (*exchange->escaped)[0];
|
|
else
|
|
escape_char= -1;
|
|
is_ambiguous_field_sep= test(strchr(ESCAPE_CHARS, field_sep_char));
|
|
is_unsafe_field_sep= test(strchr(NUMERIC_CHARS, field_sep_char));
|
|
line_sep_char= (exchange->line_term->length() ?
|
|
(int) (uchar) (*exchange->line_term)[0] : INT_MAX);
|
|
if (!field_term_length)
|
|
exchange->opt_enclosed=0;
|
|
if (!exchange->enclosed->length())
|
|
exchange->opt_enclosed=1; // A little quicker loop
|
|
fixed_row_size= (!field_term_length && !exchange->enclosed->length() &&
|
|
!blob_flag);
|
|
if ((is_ambiguous_field_sep && exchange->enclosed->is_empty() &&
|
|
(string_results || is_unsafe_field_sep)) ||
|
|
(exchange->opt_enclosed && non_string_results &&
|
|
field_term_length && strchr(NUMERIC_CHARS, field_term_char)))
|
|
{
|
|
push_warning(thd, MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
ER_AMBIGUOUS_FIELD_TERM, ER(ER_AMBIGUOUS_FIELD_TERM));
|
|
is_ambiguous_field_term= TRUE;
|
|
}
|
|
else
|
|
is_ambiguous_field_term= FALSE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
#define NEED_ESCAPING(x) ((int) (uchar) (x) == escape_char || \
|
|
(enclosed ? (int) (uchar) (x) == field_sep_char \
|
|
: (int) (uchar) (x) == field_term_char) || \
|
|
(int) (uchar) (x) == line_sep_char || \
|
|
!(x))
|
|
|
|
bool select_export::send_data(List<Item> &items)
|
|
{
|
|
|
|
DBUG_ENTER("select_export::send_data");
|
|
char buff[MAX_FIELD_WIDTH],null_buff[2],space[MAX_FIELD_WIDTH];
|
|
char cvt_buff[MAX_FIELD_WIDTH];
|
|
String cvt_str(cvt_buff, sizeof(cvt_buff), write_cs);
|
|
bool space_inited=0;
|
|
String tmp(buff,sizeof(buff),&my_charset_bin),*res;
|
|
tmp.length(0);
|
|
|
|
if (unit->offset_limit_cnt)
|
|
{ // using limit offset,count
|
|
unit->offset_limit_cnt--;
|
|
DBUG_RETURN(0);
|
|
}
|
|
row_count++;
|
|
Item *item;
|
|
uint used_length=0,items_left=items.elements;
|
|
List_iterator_fast<Item> li(items);
|
|
|
|
if (my_b_write(&cache,(uchar*) exchange->line_start->ptr(),
|
|
exchange->line_start->length()))
|
|
goto err;
|
|
while ((item=li++))
|
|
{
|
|
Item_result result_type=item->result_type();
|
|
bool enclosed = (exchange->enclosed->length() &&
|
|
(!exchange->opt_enclosed || result_type == STRING_RESULT));
|
|
res=item->str_result(&tmp);
|
|
if (res && !my_charset_same(write_cs, res->charset()) &&
|
|
!my_charset_same(write_cs, &my_charset_bin))
|
|
{
|
|
const char *well_formed_error_pos;
|
|
const char *cannot_convert_error_pos;
|
|
const char *from_end_pos;
|
|
const char *error_pos;
|
|
uint32 bytes;
|
|
bytes= well_formed_copy_nchars(write_cs, cvt_buff, sizeof(cvt_buff),
|
|
res->charset(), res->ptr(), res->length(),
|
|
sizeof(cvt_buff),
|
|
&well_formed_error_pos,
|
|
&cannot_convert_error_pos,
|
|
&from_end_pos);
|
|
error_pos= well_formed_error_pos ? well_formed_error_pos
|
|
: cannot_convert_error_pos;
|
|
if (error_pos)
|
|
{
|
|
char printable_buff[32];
|
|
convert_to_printable(printable_buff, sizeof(printable_buff),
|
|
error_pos, res->ptr() + res->length() - error_pos,
|
|
res->charset(), 6);
|
|
push_warning_printf(thd, MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
ER_TRUNCATED_WRONG_VALUE_FOR_FIELD,
|
|
ER(ER_TRUNCATED_WRONG_VALUE_FOR_FIELD),
|
|
"string", printable_buff,
|
|
item->name, row_count);
|
|
}
|
|
cvt_str.length(bytes);
|
|
res= &cvt_str;
|
|
}
|
|
if (res && enclosed)
|
|
{
|
|
if (my_b_write(&cache,(uchar*) exchange->enclosed->ptr(),
|
|
exchange->enclosed->length()))
|
|
goto err;
|
|
}
|
|
if (!res)
|
|
{ // NULL
|
|
if (!fixed_row_size)
|
|
{
|
|
if (escape_char != -1) // Use \N syntax
|
|
{
|
|
null_buff[0]=escape_char;
|
|
null_buff[1]='N';
|
|
if (my_b_write(&cache,(uchar*) null_buff,2))
|
|
goto err;
|
|
}
|
|
else if (my_b_write(&cache,(uchar*) "NULL",4))
|
|
goto err;
|
|
}
|
|
else
|
|
{
|
|
used_length=0; // Fill with space
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (fixed_row_size)
|
|
used_length=min(res->length(),item->max_length);
|
|
else
|
|
used_length=res->length();
|
|
if ((result_type == STRING_RESULT || is_unsafe_field_sep) &&
|
|
escape_char != -1)
|
|
{
|
|
char *pos, *start, *end;
|
|
CHARSET_INFO *res_charset= res->charset();
|
|
CHARSET_INFO *character_set_client= thd->variables.
|
|
character_set_client;
|
|
bool check_second_byte= (res_charset == &my_charset_bin) &&
|
|
character_set_client->
|
|
escape_with_backslash_is_dangerous;
|
|
DBUG_ASSERT(character_set_client->mbmaxlen == 2 ||
|
|
!character_set_client->escape_with_backslash_is_dangerous);
|
|
for (start=pos=(char*) res->ptr(),end=pos+used_length ;
|
|
pos != end ;
|
|
pos++)
|
|
{
|
|
#ifdef USE_MB
|
|
if (use_mb(res_charset))
|
|
{
|
|
int l;
|
|
if ((l=my_ismbchar(res_charset, pos, end)))
|
|
{
|
|
pos += l-1;
|
|
continue;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
Special case when dumping BINARY/VARBINARY/BLOB values
|
|
for the clients with character sets big5, cp932, gbk and sjis,
|
|
which can have the escape character (0x5C "\" by default)
|
|
as the second byte of a multi-byte sequence.
|
|
|
|
If
|
|
- pos[0] is a valid multi-byte head (e.g 0xEE) and
|
|
- pos[1] is 0x00, which will be escaped as "\0",
|
|
|
|
then we'll get "0xEE + 0x5C + 0x30" in the output file.
|
|
|
|
If this file is later loaded using this sequence of commands:
|
|
|
|
mysql> create table t1 (a varchar(128)) character set big5;
|
|
mysql> LOAD DATA INFILE 'dump.txt' INTO TABLE t1;
|
|
|
|
then 0x5C will be misinterpreted as the second byte
|
|
of a multi-byte character "0xEE + 0x5C", instead of
|
|
escape character for 0x00.
|
|
|
|
To avoid this confusion, we'll escape the multi-byte
|
|
head character too, so the sequence "0xEE + 0x00" will be
|
|
dumped as "0x5C + 0xEE + 0x5C + 0x30".
|
|
|
|
Note, in the condition below we only check if
|
|
mbcharlen is equal to 2, because there are no
|
|
character sets with mbmaxlen longer than 2
|
|
and with escape_with_backslash_is_dangerous set.
|
|
DBUG_ASSERT before the loop makes that sure.
|
|
*/
|
|
|
|
if ((NEED_ESCAPING(*pos) ||
|
|
(check_second_byte &&
|
|
my_mbcharlen(character_set_client, (uchar) *pos) == 2 &&
|
|
pos + 1 < end &&
|
|
NEED_ESCAPING(pos[1]))) &&
|
|
/*
|
|
Don't escape field_term_char by doubling - doubling is only
|
|
valid for ENCLOSED BY characters:
|
|
*/
|
|
(enclosed || !is_ambiguous_field_term ||
|
|
(int) (uchar) *pos != field_term_char))
|
|
{
|
|
char tmp_buff[2];
|
|
tmp_buff[0]= ((int) (uchar) *pos == field_sep_char &&
|
|
is_ambiguous_field_sep) ?
|
|
field_sep_char : escape_char;
|
|
tmp_buff[1]= *pos ? *pos : '0';
|
|
if (my_b_write(&cache,(uchar*) start,(uint) (pos-start)) ||
|
|
my_b_write(&cache,(uchar*) tmp_buff,2))
|
|
goto err;
|
|
start=pos+1;
|
|
}
|
|
}
|
|
if (my_b_write(&cache,(uchar*) start,(uint) (pos-start)))
|
|
goto err;
|
|
}
|
|
else if (my_b_write(&cache,(uchar*) res->ptr(),used_length))
|
|
goto err;
|
|
}
|
|
if (fixed_row_size)
|
|
{ // Fill with space
|
|
if (item->max_length > used_length)
|
|
{
|
|
/* QQ: Fix by adding a my_b_fill() function */
|
|
if (!space_inited)
|
|
{
|
|
space_inited=1;
|
|
bfill(space,sizeof(space),' ');
|
|
}
|
|
uint length=item->max_length-used_length;
|
|
for (; length > sizeof(space) ; length-=sizeof(space))
|
|
{
|
|
if (my_b_write(&cache,(uchar*) space,sizeof(space)))
|
|
goto err;
|
|
}
|
|
if (my_b_write(&cache,(uchar*) space,length))
|
|
goto err;
|
|
}
|
|
}
|
|
if (res && enclosed)
|
|
{
|
|
if (my_b_write(&cache, (uchar*) exchange->enclosed->ptr(),
|
|
exchange->enclosed->length()))
|
|
goto err;
|
|
}
|
|
if (--items_left)
|
|
{
|
|
if (my_b_write(&cache, (uchar*) exchange->field_term->ptr(),
|
|
field_term_length))
|
|
goto err;
|
|
}
|
|
}
|
|
if (my_b_write(&cache,(uchar*) exchange->line_term->ptr(),
|
|
exchange->line_term->length()))
|
|
goto err;
|
|
DBUG_RETURN(0);
|
|
err:
|
|
DBUG_RETURN(1);
|
|
}
|
|
|
|
|
|
/***************************************************************************
|
|
** Dump of select to a binary file
|
|
***************************************************************************/
|
|
|
|
|
|
int
|
|
select_dump::prepare(List<Item> &list __attribute__((unused)),
|
|
SELECT_LEX_UNIT *u)
|
|
{
|
|
unit= u;
|
|
return (int) ((file= create_file(thd, path, exchange, &cache)) < 0);
|
|
}
|
|
|
|
|
|
bool select_dump::send_data(List<Item> &items)
|
|
{
|
|
List_iterator_fast<Item> li(items);
|
|
char buff[MAX_FIELD_WIDTH];
|
|
String tmp(buff,sizeof(buff),&my_charset_bin),*res;
|
|
tmp.length(0);
|
|
Item *item;
|
|
DBUG_ENTER("select_dump::send_data");
|
|
|
|
if (unit->offset_limit_cnt)
|
|
{ // using limit offset,count
|
|
unit->offset_limit_cnt--;
|
|
DBUG_RETURN(0);
|
|
}
|
|
if (row_count++ > 1)
|
|
{
|
|
my_message(ER_TOO_MANY_ROWS, ER(ER_TOO_MANY_ROWS), MYF(0));
|
|
goto err;
|
|
}
|
|
while ((item=li++))
|
|
{
|
|
res=item->str_result(&tmp);
|
|
if (!res) // If NULL
|
|
{
|
|
if (my_b_write(&cache,(uchar*) "",1))
|
|
goto err;
|
|
}
|
|
else if (my_b_write(&cache,(uchar*) res->ptr(),res->length()))
|
|
{
|
|
my_error(ER_ERROR_ON_WRITE, MYF(0), path, my_errno);
|
|
goto err;
|
|
}
|
|
}
|
|
DBUG_RETURN(0);
|
|
err:
|
|
DBUG_RETURN(1);
|
|
}
|
|
|
|
|
|
select_subselect::select_subselect(Item_subselect *item_arg)
|
|
{
|
|
item= item_arg;
|
|
}
|
|
|
|
|
|
bool select_singlerow_subselect::send_data(List<Item> &items)
|
|
{
|
|
DBUG_ENTER("select_singlerow_subselect::send_data");
|
|
Item_singlerow_subselect *it= (Item_singlerow_subselect *)item;
|
|
if (it->assigned())
|
|
{
|
|
my_message(ER_SUBQUERY_NO_1_ROW, ER(ER_SUBQUERY_NO_1_ROW), MYF(0));
|
|
DBUG_RETURN(1);
|
|
}
|
|
if (unit->offset_limit_cnt)
|
|
{ // Using limit offset,count
|
|
unit->offset_limit_cnt--;
|
|
DBUG_RETURN(0);
|
|
}
|
|
List_iterator_fast<Item> li(items);
|
|
Item *val_item;
|
|
for (uint i= 0; (val_item= li++); i++)
|
|
it->store(i, val_item);
|
|
it->assigned(1);
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
|
|
void select_max_min_finder_subselect::cleanup()
|
|
{
|
|
DBUG_ENTER("select_max_min_finder_subselect::cleanup");
|
|
cache= 0;
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
bool select_max_min_finder_subselect::send_data(List<Item> &items)
|
|
{
|
|
DBUG_ENTER("select_max_min_finder_subselect::send_data");
|
|
Item_maxmin_subselect *it= (Item_maxmin_subselect *)item;
|
|
List_iterator_fast<Item> li(items);
|
|
Item *val_item= li++;
|
|
it->register_value();
|
|
if (it->assigned())
|
|
{
|
|
cache->store(val_item);
|
|
if ((this->*op)())
|
|
it->store(0, cache);
|
|
}
|
|
else
|
|
{
|
|
if (!cache)
|
|
{
|
|
cache= Item_cache::get_cache(val_item);
|
|
switch (val_item->result_type())
|
|
{
|
|
case REAL_RESULT:
|
|
op= &select_max_min_finder_subselect::cmp_real;
|
|
break;
|
|
case INT_RESULT:
|
|
op= &select_max_min_finder_subselect::cmp_int;
|
|
break;
|
|
case STRING_RESULT:
|
|
op= &select_max_min_finder_subselect::cmp_str;
|
|
break;
|
|
case DECIMAL_RESULT:
|
|
op= &select_max_min_finder_subselect::cmp_decimal;
|
|
break;
|
|
case ROW_RESULT:
|
|
// This case should never be choosen
|
|
DBUG_ASSERT(0);
|
|
op= 0;
|
|
}
|
|
}
|
|
cache->store(val_item);
|
|
it->store(0, cache);
|
|
}
|
|
it->assigned(1);
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
bool select_max_min_finder_subselect::cmp_real()
|
|
{
|
|
Item *maxmin= ((Item_singlerow_subselect *)item)->element_index(0);
|
|
double val1= cache->val_real(), val2= maxmin->val_real();
|
|
if (fmax)
|
|
return (cache->null_value && !maxmin->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
val1 > val2);
|
|
return (maxmin->null_value && !cache->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
val1 < val2);
|
|
}
|
|
|
|
bool select_max_min_finder_subselect::cmp_int()
|
|
{
|
|
Item *maxmin= ((Item_singlerow_subselect *)item)->element_index(0);
|
|
longlong val1= cache->val_int(), val2= maxmin->val_int();
|
|
if (fmax)
|
|
return (cache->null_value && !maxmin->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
val1 > val2);
|
|
return (maxmin->null_value && !cache->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
val1 < val2);
|
|
}
|
|
|
|
bool select_max_min_finder_subselect::cmp_decimal()
|
|
{
|
|
Item *maxmin= ((Item_singlerow_subselect *)item)->element_index(0);
|
|
my_decimal cval, *cvalue= cache->val_decimal(&cval);
|
|
my_decimal mval, *mvalue= maxmin->val_decimal(&mval);
|
|
if (fmax)
|
|
return (cache->null_value && !maxmin->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
my_decimal_cmp(cvalue, mvalue) > 0) ;
|
|
return (maxmin->null_value && !cache->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
my_decimal_cmp(cvalue,mvalue) < 0);
|
|
}
|
|
|
|
bool select_max_min_finder_subselect::cmp_str()
|
|
{
|
|
String *val1, *val2, buf1, buf2;
|
|
Item *maxmin= ((Item_singlerow_subselect *)item)->element_index(0);
|
|
/*
|
|
as far as both operand is Item_cache buf1 & buf2 will not be used,
|
|
but added for safety
|
|
*/
|
|
val1= cache->val_str(&buf1);
|
|
val2= maxmin->val_str(&buf1);
|
|
if (fmax)
|
|
return (cache->null_value && !maxmin->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
sortcmp(val1, val2, cache->collation.collation) > 0) ;
|
|
return (maxmin->null_value && !cache->null_value) ||
|
|
(!cache->null_value && !maxmin->null_value &&
|
|
sortcmp(val1, val2, cache->collation.collation) < 0);
|
|
}
|
|
|
|
bool select_exists_subselect::send_data(List<Item> &items)
|
|
{
|
|
DBUG_ENTER("select_exists_subselect::send_data");
|
|
Item_exists_subselect *it= (Item_exists_subselect *)item;
|
|
if (unit->offset_limit_cnt)
|
|
{ // Using limit offset,count
|
|
unit->offset_limit_cnt--;
|
|
DBUG_RETURN(0);
|
|
}
|
|
it->value= 1;
|
|
it->assigned(1);
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
|
|
/***************************************************************************
|
|
Dump of select to variables
|
|
***************************************************************************/
|
|
|
|
int select_dumpvar::prepare(List<Item> &list, SELECT_LEX_UNIT *u)
|
|
{
|
|
unit= u;
|
|
|
|
if (var_list.elements != list.elements)
|
|
{
|
|
my_message(ER_WRONG_NUMBER_OF_COLUMNS_IN_SELECT,
|
|
ER(ER_WRONG_NUMBER_OF_COLUMNS_IN_SELECT), MYF(0));
|
|
return 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
bool select_dumpvar::check_simple_select() const
|
|
{
|
|
my_error(ER_SP_BAD_CURSOR_SELECT, MYF(0));
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
void select_dumpvar::cleanup()
|
|
{
|
|
row_count= 0;
|
|
}
|
|
|
|
|
|
Query_arena::Type Query_arena::type() const
|
|
{
|
|
DBUG_ASSERT(0); /* Should never be called */
|
|
return STATEMENT;
|
|
}
|
|
|
|
|
|
void Query_arena::free_items()
|
|
{
|
|
Item *next;
|
|
DBUG_ENTER("Query_arena::free_items");
|
|
/* This works because items are allocated with sql_alloc() */
|
|
for (; free_list; free_list= next)
|
|
{
|
|
next= free_list->next;
|
|
free_list->delete_self();
|
|
}
|
|
/* Postcondition: free_list is 0 */
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
void Query_arena::set_query_arena(Query_arena *set)
|
|
{
|
|
mem_root= set->mem_root;
|
|
free_list= set->free_list;
|
|
state= set->state;
|
|
}
|
|
|
|
|
|
void Query_arena::cleanup_stmt()
|
|
{
|
|
DBUG_ASSERT(! "Query_arena::cleanup_stmt() not implemented");
|
|
}
|
|
|
|
/*
|
|
Statement functions
|
|
*/
|
|
|
|
Statement::Statement(LEX *lex_arg, MEM_ROOT *mem_root_arg,
|
|
enum enum_state state_arg, ulong id_arg)
|
|
:Query_arena(mem_root_arg, state_arg),
|
|
id(id_arg),
|
|
mark_used_columns(MARK_COLUMNS_READ),
|
|
lex(lex_arg),
|
|
cursor(0),
|
|
db(NULL),
|
|
db_length(0)
|
|
{
|
|
query_string.length= 0;
|
|
query_string.str= NULL;
|
|
name.str= NULL;
|
|
}
|
|
|
|
|
|
Query_arena::Type Statement::type() const
|
|
{
|
|
return STATEMENT;
|
|
}
|
|
|
|
|
|
void Statement::set_statement(Statement *stmt)
|
|
{
|
|
id= stmt->id;
|
|
mark_used_columns= stmt->mark_used_columns;
|
|
lex= stmt->lex;
|
|
query_string= stmt->query_string;
|
|
cursor= stmt->cursor;
|
|
}
|
|
|
|
|
|
void
|
|
Statement::set_n_backup_statement(Statement *stmt, Statement *backup)
|
|
{
|
|
DBUG_ENTER("Statement::set_n_backup_statement");
|
|
backup->set_statement(this);
|
|
set_statement(stmt);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
void Statement::restore_backup_statement(Statement *stmt, Statement *backup)
|
|
{
|
|
DBUG_ENTER("Statement::restore_backup_statement");
|
|
stmt->set_statement(this);
|
|
set_statement(backup);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
/** Assign a new value to thd->query. */
|
|
|
|
void Statement::set_query_inner(char *query_arg, uint32 query_length_arg)
|
|
{
|
|
query_string.str= query_arg;
|
|
query_string.length= query_length_arg;
|
|
}
|
|
|
|
|
|
void THD::end_statement()
|
|
{
|
|
/* Cleanup SQL processing state to reuse this statement in next query. */
|
|
lex_end(lex);
|
|
delete lex->result;
|
|
lex->result= 0;
|
|
/* Note that free_list is freed in cleanup_after_query() */
|
|
|
|
/*
|
|
Don't free mem_root, as mem_root is freed in the end of dispatch_command
|
|
(once for any command).
|
|
*/
|
|
}
|
|
|
|
|
|
void THD::set_n_backup_active_arena(Query_arena *set, Query_arena *backup)
|
|
{
|
|
DBUG_ENTER("THD::set_n_backup_active_arena");
|
|
DBUG_ASSERT(backup->is_backup_arena == FALSE);
|
|
|
|
backup->set_query_arena(this);
|
|
set_query_arena(set);
|
|
#ifndef DBUG_OFF
|
|
backup->is_backup_arena= TRUE;
|
|
#endif
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
void THD::restore_active_arena(Query_arena *set, Query_arena *backup)
|
|
{
|
|
DBUG_ENTER("THD::restore_active_arena");
|
|
DBUG_ASSERT(backup->is_backup_arena);
|
|
set->set_query_arena(this);
|
|
set_query_arena(backup);
|
|
#ifndef DBUG_OFF
|
|
backup->is_backup_arena= FALSE;
|
|
#endif
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
Statement::~Statement()
|
|
{
|
|
}
|
|
|
|
C_MODE_START
|
|
|
|
static uchar *
|
|
get_statement_id_as_hash_key(const uchar *record, size_t *key_length,
|
|
my_bool not_used __attribute__((unused)))
|
|
{
|
|
const Statement *statement= (const Statement *) record;
|
|
*key_length= sizeof(statement->id);
|
|
return (uchar *) &((const Statement *) statement)->id;
|
|
}
|
|
|
|
static void delete_statement_as_hash_key(void *key)
|
|
{
|
|
delete (Statement *) key;
|
|
}
|
|
|
|
static uchar *get_stmt_name_hash_key(Statement *entry, size_t *length,
|
|
my_bool not_used __attribute__((unused)))
|
|
{
|
|
*length= entry->name.length;
|
|
return (uchar*) entry->name.str;
|
|
}
|
|
|
|
C_MODE_END
|
|
|
|
Statement_map::Statement_map() :
|
|
last_found_statement(0)
|
|
{
|
|
enum
|
|
{
|
|
START_STMT_HASH_SIZE = 16,
|
|
START_NAME_HASH_SIZE = 16
|
|
};
|
|
my_hash_init(&st_hash, &my_charset_bin, START_STMT_HASH_SIZE, 0, 0,
|
|
get_statement_id_as_hash_key,
|
|
delete_statement_as_hash_key, MYF(0));
|
|
my_hash_init(&names_hash, system_charset_info, START_NAME_HASH_SIZE, 0, 0,
|
|
(my_hash_get_key) get_stmt_name_hash_key,
|
|
NULL,MYF(0));
|
|
}
|
|
|
|
|
|
/*
|
|
Insert a new statement to the thread-local statement map.
|
|
|
|
DESCRIPTION
|
|
If there was an old statement with the same name, replace it with the
|
|
new one. Otherwise, check if max_prepared_stmt_count is not reached yet,
|
|
increase prepared_stmt_count, and insert the new statement. It's okay
|
|
to delete an old statement and fail to insert the new one.
|
|
|
|
POSTCONDITIONS
|
|
All named prepared statements are also present in names_hash.
|
|
Statement names in names_hash are unique.
|
|
The statement is added only if prepared_stmt_count < max_prepard_stmt_count
|
|
last_found_statement always points to a valid statement or is 0
|
|
|
|
RETURN VALUE
|
|
0 success
|
|
1 error: out of resources or max_prepared_stmt_count limit has been
|
|
reached. An error is sent to the client, the statement is deleted.
|
|
*/
|
|
|
|
int Statement_map::insert(THD *thd, Statement *statement)
|
|
{
|
|
if (my_hash_insert(&st_hash, (uchar*) statement))
|
|
{
|
|
/*
|
|
Delete is needed only in case of an insert failure. In all other
|
|
cases hash_delete will also delete the statement.
|
|
*/
|
|
delete statement;
|
|
my_error(ER_OUT_OF_RESOURCES, MYF(0));
|
|
goto err_st_hash;
|
|
}
|
|
if (statement->name.str && my_hash_insert(&names_hash, (uchar*) statement))
|
|
{
|
|
my_error(ER_OUT_OF_RESOURCES, MYF(0));
|
|
goto err_names_hash;
|
|
}
|
|
pthread_mutex_lock(&LOCK_prepared_stmt_count);
|
|
/*
|
|
We don't check that prepared_stmt_count is <= max_prepared_stmt_count
|
|
because we would like to allow to lower the total limit
|
|
of prepared statements below the current count. In that case
|
|
no new statements can be added until prepared_stmt_count drops below
|
|
the limit.
|
|
*/
|
|
if (prepared_stmt_count >= max_prepared_stmt_count)
|
|
{
|
|
pthread_mutex_unlock(&LOCK_prepared_stmt_count);
|
|
my_error(ER_MAX_PREPARED_STMT_COUNT_REACHED, MYF(0),
|
|
max_prepared_stmt_count);
|
|
goto err_max;
|
|
}
|
|
prepared_stmt_count++;
|
|
pthread_mutex_unlock(&LOCK_prepared_stmt_count);
|
|
|
|
last_found_statement= statement;
|
|
return 0;
|
|
|
|
err_max:
|
|
if (statement->name.str)
|
|
my_hash_delete(&names_hash, (uchar*) statement);
|
|
err_names_hash:
|
|
my_hash_delete(&st_hash, (uchar*) statement);
|
|
err_st_hash:
|
|
return 1;
|
|
}
|
|
|
|
|
|
void Statement_map::close_transient_cursors()
|
|
{
|
|
#ifdef TO_BE_IMPLEMENTED
|
|
Statement *stmt;
|
|
while ((stmt= transient_cursor_list.head()))
|
|
stmt->close_cursor(); /* deletes itself from the list */
|
|
#endif
|
|
}
|
|
|
|
|
|
void Statement_map::erase(Statement *statement)
|
|
{
|
|
if (statement == last_found_statement)
|
|
last_found_statement= 0;
|
|
if (statement->name.str)
|
|
my_hash_delete(&names_hash, (uchar *) statement);
|
|
|
|
my_hash_delete(&st_hash, (uchar *) statement);
|
|
pthread_mutex_lock(&LOCK_prepared_stmt_count);
|
|
DBUG_ASSERT(prepared_stmt_count > 0);
|
|
prepared_stmt_count--;
|
|
pthread_mutex_unlock(&LOCK_prepared_stmt_count);
|
|
}
|
|
|
|
|
|
void Statement_map::reset()
|
|
{
|
|
/* Must be first, hash_free will reset st_hash.records */
|
|
pthread_mutex_lock(&LOCK_prepared_stmt_count);
|
|
DBUG_ASSERT(prepared_stmt_count >= st_hash.records);
|
|
prepared_stmt_count-= st_hash.records;
|
|
pthread_mutex_unlock(&LOCK_prepared_stmt_count);
|
|
|
|
my_hash_reset(&names_hash);
|
|
my_hash_reset(&st_hash);
|
|
last_found_statement= 0;
|
|
}
|
|
|
|
|
|
Statement_map::~Statement_map()
|
|
{
|
|
/* Must go first, hash_free will reset st_hash.records */
|
|
pthread_mutex_lock(&LOCK_prepared_stmt_count);
|
|
DBUG_ASSERT(prepared_stmt_count >= st_hash.records);
|
|
prepared_stmt_count-= st_hash.records;
|
|
pthread_mutex_unlock(&LOCK_prepared_stmt_count);
|
|
|
|
my_hash_free(&names_hash);
|
|
my_hash_free(&st_hash);
|
|
}
|
|
|
|
bool select_dumpvar::send_data(List<Item> &items)
|
|
{
|
|
List_iterator_fast<my_var> var_li(var_list);
|
|
List_iterator<Item> it(items);
|
|
Item *item;
|
|
my_var *mv;
|
|
DBUG_ENTER("select_dumpvar::send_data");
|
|
|
|
if (unit->offset_limit_cnt)
|
|
{ // using limit offset,count
|
|
unit->offset_limit_cnt--;
|
|
DBUG_RETURN(0);
|
|
}
|
|
if (row_count++)
|
|
{
|
|
my_message(ER_TOO_MANY_ROWS, ER(ER_TOO_MANY_ROWS), MYF(0));
|
|
DBUG_RETURN(1);
|
|
}
|
|
while ((mv= var_li++) && (item= it++))
|
|
{
|
|
if (mv->local)
|
|
{
|
|
if (thd->spcont->set_variable(thd, mv->offset, &item))
|
|
DBUG_RETURN(1);
|
|
}
|
|
else
|
|
{
|
|
Item_func_set_user_var *suv= new Item_func_set_user_var(mv->s, item);
|
|
if (suv->fix_fields(thd, 0))
|
|
DBUG_RETURN (1);
|
|
suv->save_item_result(item);
|
|
if (suv->update())
|
|
DBUG_RETURN (1);
|
|
}
|
|
}
|
|
DBUG_RETURN(thd->is_error());
|
|
}
|
|
|
|
bool select_dumpvar::send_eof()
|
|
{
|
|
if (! row_count)
|
|
push_warning(thd, MYSQL_ERROR::WARN_LEVEL_WARN,
|
|
ER_SP_FETCH_NO_DATA, ER(ER_SP_FETCH_NO_DATA));
|
|
/*
|
|
In order to remember the value of affected rows for ROW_COUNT()
|
|
function, SELECT INTO has to have an own SQLCOM.
|
|
TODO: split from SQLCOM_SELECT
|
|
*/
|
|
::my_ok(thd,row_count);
|
|
return 0;
|
|
}
|
|
|
|
/****************************************************************************
|
|
TMP_TABLE_PARAM
|
|
****************************************************************************/
|
|
|
|
void TMP_TABLE_PARAM::init()
|
|
{
|
|
DBUG_ENTER("TMP_TABLE_PARAM::init");
|
|
DBUG_PRINT("enter", ("this: 0x%lx", (ulong)this));
|
|
field_count= sum_func_count= func_count= hidden_field_count= 0;
|
|
group_parts= group_length= group_null_parts= 0;
|
|
quick_group= 1;
|
|
table_charset= 0;
|
|
precomputed_group_by= 0;
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
void thd_increment_bytes_sent(ulong length)
|
|
{
|
|
THD *thd=current_thd;
|
|
if (likely(thd != 0))
|
|
{ /* current_thd==0 when close_connection() calls net_send_error() */
|
|
thd->status_var.bytes_sent+= length;
|
|
}
|
|
}
|
|
|
|
|
|
void thd_increment_bytes_received(ulong length)
|
|
{
|
|
current_thd->status_var.bytes_received+= length;
|
|
}
|
|
|
|
|
|
void thd_increment_net_big_packet_count(ulong length)
|
|
{
|
|
current_thd->status_var.net_big_packet_count+= length;
|
|
}
|
|
|
|
|
|
void THD::set_status_var_init()
|
|
{
|
|
bzero((char*) &status_var, sizeof(status_var));
|
|
}
|
|
|
|
|
|
void Security_context::init()
|
|
{
|
|
host= user= priv_user= ip= 0;
|
|
host_or_ip= "connecting host";
|
|
priv_host[0]= '\0';
|
|
master_access= 0;
|
|
#ifndef NO_EMBEDDED_ACCESS_CHECKS
|
|
db_access= NO_ACCESS;
|
|
#endif
|
|
}
|
|
|
|
|
|
void Security_context::destroy()
|
|
{
|
|
// If not pointer to constant
|
|
if (host != my_localhost)
|
|
safeFree(host);
|
|
if (user != delayed_user)
|
|
safeFree(user);
|
|
safeFree(ip);
|
|
}
|
|
|
|
|
|
void Security_context::skip_grants()
|
|
{
|
|
/* privileges for the user are unknown everything is allowed */
|
|
host_or_ip= (char *)"";
|
|
master_access= ~NO_ACCESS;
|
|
priv_user= (char *)"";
|
|
*priv_host= '\0';
|
|
}
|
|
|
|
|
|
bool Security_context::set_user(char *user_arg)
|
|
{
|
|
safeFree(user);
|
|
user= my_strdup(user_arg, MYF(0));
|
|
return user == 0;
|
|
}
|
|
|
|
#ifndef NO_EMBEDDED_ACCESS_CHECKS
|
|
/**
|
|
Initialize this security context from the passed in credentials
|
|
and activate it in the current thread.
|
|
|
|
@param thd
|
|
@param definer_user
|
|
@param definer_host
|
|
@param db
|
|
@param[out] backup Save a pointer to the current security context
|
|
in the thread. In case of success it points to the
|
|
saved old context, otherwise it points to NULL.
|
|
|
|
|
|
During execution of a statement, multiple security contexts may
|
|
be needed:
|
|
- the security context of the authenticated user, used as the
|
|
default security context for all top-level statements
|
|
- in case of a view or a stored program, possibly the security
|
|
context of the definer of the routine, if the object is
|
|
defined with SQL SECURITY DEFINER option.
|
|
|
|
The currently "active" security context is parameterized in THD
|
|
member security_ctx. By default, after a connection is
|
|
established, this member points at the "main" security context
|
|
- the credentials of the authenticated user.
|
|
|
|
Later, if we would like to execute some sub-statement or a part
|
|
of a statement under credentials of a different user, e.g.
|
|
definer of a procedure, we authenticate this user in a local
|
|
instance of Security_context by means of this method (and
|
|
ultimately by means of acl_getroot_no_password), and make the
|
|
local instance active in the thread by re-setting
|
|
thd->security_ctx pointer.
|
|
|
|
Note, that the life cycle and memory management of the "main" and
|
|
temporary security contexts are different.
|
|
For the main security context, the memory for user/host/ip is
|
|
allocated on system heap, and the THD class frees this memory in
|
|
its destructor. The only case when contents of the main security
|
|
context may change during its life time is when someone issued
|
|
CHANGE USER command.
|
|
Memory management of a "temporary" security context is
|
|
responsibility of the module that creates it.
|
|
|
|
@retval TRUE there is no user with the given credentials. The erro
|
|
is reported in the thread.
|
|
@retval FALSE success
|
|
*/
|
|
|
|
bool
|
|
Security_context::
|
|
change_security_context(THD *thd,
|
|
LEX_STRING *definer_user,
|
|
LEX_STRING *definer_host,
|
|
LEX_STRING *db,
|
|
Security_context **backup)
|
|
{
|
|
bool needs_change;
|
|
|
|
DBUG_ENTER("Security_context::change_security_context");
|
|
|
|
DBUG_ASSERT(definer_user->str && definer_host->str);
|
|
|
|
*backup= NULL;
|
|
/*
|
|
The current security context may have NULL members
|
|
if we have just started the thread and not authenticated
|
|
any user. This use case is currently in events worker thread.
|
|
*/
|
|
needs_change= (thd->security_ctx->priv_user == NULL ||
|
|
strcmp(definer_user->str, thd->security_ctx->priv_user) ||
|
|
thd->security_ctx->priv_host == NULL ||
|
|
my_strcasecmp(system_charset_info, definer_host->str,
|
|
thd->security_ctx->priv_host));
|
|
if (needs_change)
|
|
{
|
|
if (acl_getroot_no_password(this, definer_user->str, definer_host->str,
|
|
definer_host->str, db->str))
|
|
{
|
|
my_error(ER_NO_SUCH_USER, MYF(0), definer_user->str,
|
|
definer_host->str);
|
|
DBUG_RETURN(TRUE);
|
|
}
|
|
*backup= thd->security_ctx;
|
|
thd->security_ctx= this;
|
|
}
|
|
|
|
DBUG_RETURN(FALSE);
|
|
}
|
|
|
|
|
|
void
|
|
Security_context::restore_security_context(THD *thd,
|
|
Security_context *backup)
|
|
{
|
|
if (backup)
|
|
thd->security_ctx= backup;
|
|
}
|
|
#endif
|
|
|
|
|
|
bool Security_context::user_matches(Security_context *them)
|
|
{
|
|
return ((user != NULL) && (them->user != NULL) &&
|
|
!strcmp(user, them->user));
|
|
}
|
|
|
|
|
|
/****************************************************************************
|
|
Handling of open and locked tables states.
|
|
|
|
This is used when we want to open/lock (and then close) some tables when
|
|
we already have a set of tables open and locked. We use these methods for
|
|
access to mysql.proc table to find definitions of stored routines.
|
|
****************************************************************************/
|
|
|
|
void THD::reset_n_backup_open_tables_state(Open_tables_state *backup)
|
|
{
|
|
DBUG_ENTER("reset_n_backup_open_tables_state");
|
|
backup->set_open_tables_state(this);
|
|
reset_open_tables_state(this);
|
|
state_flags|= Open_tables_state::BACKUPS_AVAIL;
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
void THD::restore_backup_open_tables_state(Open_tables_state *backup)
|
|
{
|
|
DBUG_ENTER("restore_backup_open_tables_state");
|
|
/*
|
|
Before we will throw away current open tables state we want
|
|
to be sure that it was properly cleaned up.
|
|
*/
|
|
DBUG_ASSERT(open_tables == 0 && temporary_tables == 0 &&
|
|
handler_tables == 0 && derived_tables == 0 &&
|
|
lock == 0 && locked_tables == 0 &&
|
|
prelocked_mode == NON_PRELOCKED &&
|
|
m_reprepare_observer == NULL);
|
|
mdl_context_destroy(&mdl_context);
|
|
mdl_context_destroy(&handler_mdl_context);
|
|
|
|
set_open_tables_state(backup);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
/**
|
|
Check the killed state of a user thread
|
|
@param thd user thread
|
|
@retval 0 the user thread is active
|
|
@retval 1 the user thread has been killed
|
|
*/
|
|
extern "C" int thd_killed(const MYSQL_THD thd)
|
|
{
|
|
return(thd->killed);
|
|
}
|
|
|
|
/**
|
|
Return the thread id of a user thread
|
|
@param thd user thread
|
|
@return thread id
|
|
*/
|
|
extern "C" unsigned long thd_get_thread_id(const MYSQL_THD thd)
|
|
{
|
|
return((unsigned long)thd->thread_id);
|
|
}
|
|
|
|
|
|
#ifdef INNODB_COMPATIBILITY_HOOKS
|
|
extern "C" struct charset_info_st *thd_charset(MYSQL_THD thd)
|
|
{
|
|
return(thd->charset());
|
|
}
|
|
|
|
/**
|
|
OBSOLETE : there's no way to ensure the string is null terminated.
|
|
Use thd_query_string instead()
|
|
*/
|
|
extern "C" char **thd_query(MYSQL_THD thd)
|
|
{
|
|
return(&thd->query_string.str);
|
|
}
|
|
|
|
/**
|
|
Get the current query string for the thread.
|
|
|
|
@param The MySQL internal thread pointer
|
|
@return query string and length. May be non-null-terminated.
|
|
*/
|
|
extern "C" LEX_STRING * thd_query_string (MYSQL_THD thd)
|
|
{
|
|
return(&thd->query_string);
|
|
}
|
|
|
|
extern "C" int thd_slave_thread(const MYSQL_THD thd)
|
|
{
|
|
return(thd->slave_thread);
|
|
}
|
|
|
|
extern "C" int thd_non_transactional_update(const MYSQL_THD thd)
|
|
{
|
|
return(thd->transaction.all.modified_non_trans_table);
|
|
}
|
|
|
|
extern "C" int thd_binlog_format(const MYSQL_THD thd)
|
|
{
|
|
if (mysql_bin_log.is_open() && (thd->options & OPTION_BIN_LOG))
|
|
return (int) thd->variables.binlog_format;
|
|
else
|
|
return BINLOG_FORMAT_UNSPEC;
|
|
}
|
|
|
|
extern "C" void thd_mark_transaction_to_rollback(MYSQL_THD thd, bool all)
|
|
{
|
|
mark_transaction_to_rollback(thd, all);
|
|
}
|
|
|
|
extern "C" bool thd_binlog_filter_ok(const MYSQL_THD thd)
|
|
{
|
|
return binlog_filter->db_ok(thd->db);
|
|
}
|
|
#endif // INNODB_COMPATIBILITY_HOOKS */
|
|
|
|
/****************************************************************************
|
|
Handling of statement states in functions and triggers.
|
|
|
|
This is used to ensure that the function/trigger gets a clean state
|
|
to work with and does not cause any side effects of the calling statement.
|
|
|
|
It also allows most stored functions and triggers to replicate even
|
|
if they are used items that would normally be stored in the binary
|
|
replication (like last_insert_id() etc...)
|
|
|
|
The following things is done
|
|
- Disable binary logging for the duration of the statement
|
|
- Disable multi-result-sets for the duration of the statement
|
|
- Value of last_insert_id() is saved and restored
|
|
- Value set by 'SET INSERT_ID=#' is reset and restored
|
|
- Value for found_rows() is reset and restored
|
|
- examined_row_count is added to the total
|
|
- cuted_fields is added to the total
|
|
- new savepoint level is created and destroyed
|
|
|
|
NOTES:
|
|
Seed for random() is saved for the first! usage of RAND()
|
|
We reset examined_row_count and cuted_fields and add these to the
|
|
result to ensure that if we have a bug that would reset these within
|
|
a function, we are not loosing any rows from the main statement.
|
|
|
|
We do not reset value of last_insert_id().
|
|
****************************************************************************/
|
|
|
|
void THD::reset_sub_statement_state(Sub_statement_state *backup,
|
|
uint new_state)
|
|
{
|
|
#ifndef EMBEDDED_LIBRARY
|
|
/* BUG#33029, if we are replicating from a buggy master, reset
|
|
auto_inc_intervals_forced to prevent substatement
|
|
(triggers/functions) from using erroneous INSERT_ID value
|
|
*/
|
|
if (rpl_master_erroneous_autoinc(this))
|
|
{
|
|
DBUG_ASSERT(backup->auto_inc_intervals_forced.nb_elements() == 0);
|
|
auto_inc_intervals_forced.swap(&backup->auto_inc_intervals_forced);
|
|
}
|
|
#endif
|
|
|
|
backup->options= options;
|
|
backup->in_sub_stmt= in_sub_stmt;
|
|
backup->enable_slow_log= enable_slow_log;
|
|
backup->limit_found_rows= limit_found_rows;
|
|
backup->examined_row_count= examined_row_count;
|
|
backup->sent_row_count= sent_row_count;
|
|
backup->cuted_fields= cuted_fields;
|
|
backup->client_capabilities= client_capabilities;
|
|
backup->savepoints= transaction.savepoints;
|
|
backup->first_successful_insert_id_in_prev_stmt=
|
|
first_successful_insert_id_in_prev_stmt;
|
|
backup->first_successful_insert_id_in_cur_stmt=
|
|
first_successful_insert_id_in_cur_stmt;
|
|
|
|
if ((!lex->requires_prelocking() || is_update_query(lex->sql_command)) &&
|
|
!current_stmt_binlog_row_based)
|
|
{
|
|
options&= ~OPTION_BIN_LOG;
|
|
}
|
|
|
|
if ((backup->options & OPTION_BIN_LOG) && is_update_query(lex->sql_command)&&
|
|
!current_stmt_binlog_row_based)
|
|
mysql_bin_log.start_union_events(this, this->query_id);
|
|
|
|
/* Disable result sets */
|
|
client_capabilities &= ~CLIENT_MULTI_RESULTS;
|
|
in_sub_stmt|= new_state;
|
|
examined_row_count= 0;
|
|
sent_row_count= 0;
|
|
cuted_fields= 0;
|
|
transaction.savepoints= 0;
|
|
first_successful_insert_id_in_cur_stmt= 0;
|
|
}
|
|
|
|
|
|
void THD::restore_sub_statement_state(Sub_statement_state *backup)
|
|
{
|
|
#ifndef EMBEDDED_LIBRARY
|
|
/* BUG#33029, if we are replicating from a buggy master, restore
|
|
auto_inc_intervals_forced so that the top statement can use the
|
|
INSERT_ID value set before this statement.
|
|
*/
|
|
if (rpl_master_erroneous_autoinc(this))
|
|
{
|
|
backup->auto_inc_intervals_forced.swap(&auto_inc_intervals_forced);
|
|
DBUG_ASSERT(backup->auto_inc_intervals_forced.nb_elements() == 0);
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
To save resources we want to release savepoints which were created
|
|
during execution of function or trigger before leaving their savepoint
|
|
level. It is enough to release first savepoint set on this level since
|
|
all later savepoints will be released automatically.
|
|
*/
|
|
if (transaction.savepoints)
|
|
{
|
|
SAVEPOINT *sv;
|
|
for (sv= transaction.savepoints; sv->prev; sv= sv->prev)
|
|
{}
|
|
/* ha_release_savepoint() never returns error. */
|
|
(void)ha_release_savepoint(this, sv);
|
|
}
|
|
transaction.savepoints= backup->savepoints;
|
|
options= backup->options;
|
|
in_sub_stmt= backup->in_sub_stmt;
|
|
enable_slow_log= backup->enable_slow_log;
|
|
first_successful_insert_id_in_prev_stmt=
|
|
backup->first_successful_insert_id_in_prev_stmt;
|
|
first_successful_insert_id_in_cur_stmt=
|
|
backup->first_successful_insert_id_in_cur_stmt;
|
|
limit_found_rows= backup->limit_found_rows;
|
|
sent_row_count= backup->sent_row_count;
|
|
client_capabilities= backup->client_capabilities;
|
|
/*
|
|
If we've left sub-statement mode, reset the fatal error flag.
|
|
Otherwise keep the current value, to propagate it up the sub-statement
|
|
stack.
|
|
*/
|
|
if (!in_sub_stmt)
|
|
is_fatal_sub_stmt_error= FALSE;
|
|
|
|
if ((options & OPTION_BIN_LOG) && is_update_query(lex->sql_command) &&
|
|
!current_stmt_binlog_row_based)
|
|
mysql_bin_log.stop_union_events(this);
|
|
|
|
/*
|
|
The following is added to the old values as we are interested in the
|
|
total complexity of the query
|
|
*/
|
|
examined_row_count+= backup->examined_row_count;
|
|
cuted_fields+= backup->cuted_fields;
|
|
}
|
|
|
|
|
|
void THD::set_statement(Statement *stmt)
|
|
{
|
|
pthread_mutex_lock(&LOCK_thd_data);
|
|
Statement::set_statement(stmt);
|
|
pthread_mutex_unlock(&LOCK_thd_data);
|
|
}
|
|
|
|
|
|
/** Assign a new value to thd->query. */
|
|
|
|
void THD::set_query(char *query_arg, uint32 query_length_arg)
|
|
{
|
|
pthread_mutex_lock(&LOCK_thd_data);
|
|
set_query_inner(query_arg, query_length_arg);
|
|
pthread_mutex_unlock(&LOCK_thd_data);
|
|
}
|
|
|
|
|
|
/**
|
|
Mark transaction to rollback and mark error as fatal to a sub-statement.
|
|
|
|
@param thd Thread handle
|
|
@param all TRUE <=> rollback main transaction.
|
|
*/
|
|
|
|
void mark_transaction_to_rollback(THD *thd, bool all)
|
|
{
|
|
if (thd)
|
|
{
|
|
thd->is_fatal_sub_stmt_error= TRUE;
|
|
thd->transaction_rollback_request= all;
|
|
/*
|
|
Aborted transactions can not be IGNOREd.
|
|
Switch off the IGNORE flag for the current
|
|
SELECT_LEX. This should allow my_error()
|
|
to report the error and abort the execution
|
|
flow, even in presence
|
|
of IGNORE clause.
|
|
*/
|
|
if (thd->lex->current_select)
|
|
thd->lex->current_select->no_error= FALSE;
|
|
}
|
|
}
|
|
/***************************************************************************
|
|
Handling of XA id cacheing
|
|
***************************************************************************/
|
|
|
|
pthread_mutex_t LOCK_xid_cache;
|
|
HASH xid_cache;
|
|
|
|
extern "C" uchar *xid_get_hash_key(const uchar *, size_t *, my_bool);
|
|
extern "C" void xid_free_hash(void *);
|
|
|
|
uchar *xid_get_hash_key(const uchar *ptr, size_t *length,
|
|
my_bool not_used __attribute__((unused)))
|
|
{
|
|
*length=((XID_STATE*)ptr)->xid.key_length();
|
|
return ((XID_STATE*)ptr)->xid.key();
|
|
}
|
|
|
|
void xid_free_hash(void *ptr)
|
|
{
|
|
if (!((XID_STATE*)ptr)->in_thd)
|
|
my_free((uchar*)ptr, MYF(0));
|
|
}
|
|
|
|
bool xid_cache_init()
|
|
{
|
|
pthread_mutex_init(&LOCK_xid_cache, MY_MUTEX_INIT_FAST);
|
|
return my_hash_init(&xid_cache, &my_charset_bin, 100, 0, 0,
|
|
xid_get_hash_key, xid_free_hash, 0) != 0;
|
|
}
|
|
|
|
void xid_cache_free()
|
|
{
|
|
if (my_hash_inited(&xid_cache))
|
|
{
|
|
my_hash_free(&xid_cache);
|
|
pthread_mutex_destroy(&LOCK_xid_cache);
|
|
}
|
|
}
|
|
|
|
XID_STATE *xid_cache_search(XID *xid)
|
|
{
|
|
pthread_mutex_lock(&LOCK_xid_cache);
|
|
XID_STATE *res=(XID_STATE *)my_hash_search(&xid_cache, xid->key(),
|
|
xid->key_length());
|
|
pthread_mutex_unlock(&LOCK_xid_cache);
|
|
return res;
|
|
}
|
|
|
|
|
|
bool xid_cache_insert(XID *xid, enum xa_states xa_state)
|
|
{
|
|
XID_STATE *xs;
|
|
my_bool res;
|
|
pthread_mutex_lock(&LOCK_xid_cache);
|
|
if (my_hash_search(&xid_cache, xid->key(), xid->key_length()))
|
|
res=0;
|
|
else if (!(xs=(XID_STATE *)my_malloc(sizeof(*xs), MYF(MY_WME))))
|
|
res=1;
|
|
else
|
|
{
|
|
xs->xa_state=xa_state;
|
|
xs->xid.set(xid);
|
|
xs->in_thd=0;
|
|
res=my_hash_insert(&xid_cache, (uchar*)xs);
|
|
}
|
|
pthread_mutex_unlock(&LOCK_xid_cache);
|
|
return res;
|
|
}
|
|
|
|
|
|
bool xid_cache_insert(XID_STATE *xid_state)
|
|
{
|
|
pthread_mutex_lock(&LOCK_xid_cache);
|
|
DBUG_ASSERT(my_hash_search(&xid_cache, xid_state->xid.key(),
|
|
xid_state->xid.key_length())==0);
|
|
my_bool res=my_hash_insert(&xid_cache, (uchar*)xid_state);
|
|
pthread_mutex_unlock(&LOCK_xid_cache);
|
|
return res;
|
|
}
|
|
|
|
|
|
void xid_cache_delete(XID_STATE *xid_state)
|
|
{
|
|
pthread_mutex_lock(&LOCK_xid_cache);
|
|
my_hash_delete(&xid_cache, (uchar *)xid_state);
|
|
pthread_mutex_unlock(&LOCK_xid_cache);
|
|
}
|
|
|
|
/*
|
|
Implementation of interface to write rows to the binary log through the
|
|
thread. The thread is responsible for writing the rows it has
|
|
inserted/updated/deleted.
|
|
*/
|
|
|
|
#ifndef MYSQL_CLIENT
|
|
|
|
/*
|
|
Template member function for ensuring that there is an rows log
|
|
event of the apropriate type before proceeding.
|
|
|
|
PRE CONDITION:
|
|
- Events of type 'RowEventT' have the type code 'type_code'.
|
|
|
|
POST CONDITION:
|
|
If a non-NULL pointer is returned, the pending event for thread 'thd' will
|
|
be an event of type 'RowEventT' (which have the type code 'type_code')
|
|
will either empty or have enough space to hold 'needed' bytes. In
|
|
addition, the columns bitmap will be correct for the row, meaning that
|
|
the pending event will be flushed if the columns in the event differ from
|
|
the columns suppled to the function.
|
|
|
|
RETURNS
|
|
If no error, a non-NULL pending event (either one which already existed or
|
|
the newly created one).
|
|
If error, NULL.
|
|
*/
|
|
|
|
template <class RowsEventT> Rows_log_event*
|
|
THD::binlog_prepare_pending_rows_event(TABLE* table, uint32 serv_id,
|
|
MY_BITMAP const* cols,
|
|
size_t colcnt,
|
|
size_t needed,
|
|
bool is_transactional,
|
|
RowsEventT *hint __attribute__((unused)))
|
|
{
|
|
DBUG_ENTER("binlog_prepare_pending_rows_event");
|
|
/* Pre-conditions */
|
|
DBUG_ASSERT(table->s->table_map_id != ~0UL);
|
|
|
|
/* Fetch the type code for the RowsEventT template parameter */
|
|
int const type_code= RowsEventT::TYPE_CODE;
|
|
|
|
/*
|
|
There is no good place to set up the transactional data, so we
|
|
have to do it here.
|
|
*/
|
|
if (binlog_setup_trx_data())
|
|
DBUG_RETURN(NULL);
|
|
|
|
Rows_log_event* pending= binlog_get_pending_rows_event();
|
|
|
|
if (unlikely(pending && !pending->is_valid()))
|
|
DBUG_RETURN(NULL);
|
|
|
|
/*
|
|
Check if the current event is non-NULL and a write-rows
|
|
event. Also check if the table provided is mapped: if it is not,
|
|
then we have switched to writing to a new table.
|
|
If there is no pending event, we need to create one. If there is a pending
|
|
event, but it's not about the same table id, or not of the same type
|
|
(between Write, Update and Delete), or not the same affected columns, or
|
|
going to be too big, flush this event to disk and create a new pending
|
|
event.
|
|
*/
|
|
if (!pending ||
|
|
pending->server_id != serv_id ||
|
|
pending->get_table_id() != table->s->table_map_id ||
|
|
pending->get_type_code() != type_code ||
|
|
pending->get_data_size() + needed > opt_binlog_rows_event_max_size ||
|
|
pending->get_width() != colcnt ||
|
|
!bitmap_cmp(pending->get_cols(), cols))
|
|
{
|
|
/* Create a new RowsEventT... */
|
|
Rows_log_event* const
|
|
ev= new RowsEventT(this, table, table->s->table_map_id, cols,
|
|
is_transactional);
|
|
if (unlikely(!ev))
|
|
DBUG_RETURN(NULL);
|
|
ev->server_id= serv_id; // I don't like this, it's too easy to forget.
|
|
/*
|
|
flush the pending event and replace it with the newly created
|
|
event...
|
|
*/
|
|
if (unlikely(mysql_bin_log.flush_and_set_pending_rows_event(this, ev)))
|
|
{
|
|
delete ev;
|
|
DBUG_RETURN(NULL);
|
|
}
|
|
|
|
DBUG_RETURN(ev); /* This is the new pending event */
|
|
}
|
|
DBUG_RETURN(pending); /* This is the current pending event */
|
|
}
|
|
|
|
#ifdef HAVE_EXPLICIT_TEMPLATE_INSTANTIATION
|
|
/*
|
|
Instantiate the versions we need, we have -fno-implicit-template as
|
|
compiling option.
|
|
*/
|
|
template Rows_log_event*
|
|
THD::binlog_prepare_pending_rows_event(TABLE*, uint32, MY_BITMAP const*,
|
|
size_t, size_t, bool,
|
|
Write_rows_log_event*);
|
|
|
|
template Rows_log_event*
|
|
THD::binlog_prepare_pending_rows_event(TABLE*, uint32, MY_BITMAP const*,
|
|
size_t colcnt, size_t, bool,
|
|
Delete_rows_log_event *);
|
|
|
|
template Rows_log_event*
|
|
THD::binlog_prepare_pending_rows_event(TABLE*, uint32, MY_BITMAP const*,
|
|
size_t colcnt, size_t, bool,
|
|
Update_rows_log_event *);
|
|
#endif
|
|
|
|
#ifdef NOT_USED
|
|
static char const*
|
|
field_type_name(enum_field_types type)
|
|
{
|
|
switch (type) {
|
|
case MYSQL_TYPE_DECIMAL:
|
|
return "MYSQL_TYPE_DECIMAL";
|
|
case MYSQL_TYPE_TINY:
|
|
return "MYSQL_TYPE_TINY";
|
|
case MYSQL_TYPE_SHORT:
|
|
return "MYSQL_TYPE_SHORT";
|
|
case MYSQL_TYPE_LONG:
|
|
return "MYSQL_TYPE_LONG";
|
|
case MYSQL_TYPE_FLOAT:
|
|
return "MYSQL_TYPE_FLOAT";
|
|
case MYSQL_TYPE_DOUBLE:
|
|
return "MYSQL_TYPE_DOUBLE";
|
|
case MYSQL_TYPE_NULL:
|
|
return "MYSQL_TYPE_NULL";
|
|
case MYSQL_TYPE_TIMESTAMP:
|
|
return "MYSQL_TYPE_TIMESTAMP";
|
|
case MYSQL_TYPE_LONGLONG:
|
|
return "MYSQL_TYPE_LONGLONG";
|
|
case MYSQL_TYPE_INT24:
|
|
return "MYSQL_TYPE_INT24";
|
|
case MYSQL_TYPE_DATE:
|
|
return "MYSQL_TYPE_DATE";
|
|
case MYSQL_TYPE_TIME:
|
|
return "MYSQL_TYPE_TIME";
|
|
case MYSQL_TYPE_DATETIME:
|
|
return "MYSQL_TYPE_DATETIME";
|
|
case MYSQL_TYPE_YEAR:
|
|
return "MYSQL_TYPE_YEAR";
|
|
case MYSQL_TYPE_NEWDATE:
|
|
return "MYSQL_TYPE_NEWDATE";
|
|
case MYSQL_TYPE_VARCHAR:
|
|
return "MYSQL_TYPE_VARCHAR";
|
|
case MYSQL_TYPE_BIT:
|
|
return "MYSQL_TYPE_BIT";
|
|
case MYSQL_TYPE_NEWDECIMAL:
|
|
return "MYSQL_TYPE_NEWDECIMAL";
|
|
case MYSQL_TYPE_ENUM:
|
|
return "MYSQL_TYPE_ENUM";
|
|
case MYSQL_TYPE_SET:
|
|
return "MYSQL_TYPE_SET";
|
|
case MYSQL_TYPE_TINY_BLOB:
|
|
return "MYSQL_TYPE_TINY_BLOB";
|
|
case MYSQL_TYPE_MEDIUM_BLOB:
|
|
return "MYSQL_TYPE_MEDIUM_BLOB";
|
|
case MYSQL_TYPE_LONG_BLOB:
|
|
return "MYSQL_TYPE_LONG_BLOB";
|
|
case MYSQL_TYPE_BLOB:
|
|
return "MYSQL_TYPE_BLOB";
|
|
case MYSQL_TYPE_VAR_STRING:
|
|
return "MYSQL_TYPE_VAR_STRING";
|
|
case MYSQL_TYPE_STRING:
|
|
return "MYSQL_TYPE_STRING";
|
|
case MYSQL_TYPE_GEOMETRY:
|
|
return "MYSQL_TYPE_GEOMETRY";
|
|
}
|
|
return "Unknown";
|
|
}
|
|
#endif
|
|
|
|
|
|
namespace {
|
|
/**
|
|
Class to handle temporary allocation of memory for row data.
|
|
|
|
The responsibilities of the class is to provide memory for
|
|
packing one or two rows of packed data (depending on what
|
|
constructor is called).
|
|
|
|
In order to make the allocation more efficient for "simple" rows,
|
|
i.e., rows that do not contain any blobs, a pointer to the
|
|
allocated memory is of memory is stored in the table structure
|
|
for simple rows. If memory for a table containing a blob field
|
|
is requested, only memory for that is allocated, and subsequently
|
|
released when the object is destroyed.
|
|
|
|
*/
|
|
class Row_data_memory {
|
|
public:
|
|
/**
|
|
Build an object to keep track of a block-local piece of memory
|
|
for storing a row of data.
|
|
|
|
@param table
|
|
Table where the pre-allocated memory is stored.
|
|
|
|
@param length
|
|
Length of data that is needed, if the record contain blobs.
|
|
*/
|
|
Row_data_memory(TABLE *table, size_t const len1)
|
|
: m_memory(0)
|
|
{
|
|
#ifndef DBUG_OFF
|
|
m_alloc_checked= FALSE;
|
|
#endif
|
|
allocate_memory(table, len1);
|
|
m_ptr[0]= has_memory() ? m_memory : 0;
|
|
m_ptr[1]= 0;
|
|
}
|
|
|
|
Row_data_memory(TABLE *table, size_t const len1, size_t const len2)
|
|
: m_memory(0)
|
|
{
|
|
#ifndef DBUG_OFF
|
|
m_alloc_checked= FALSE;
|
|
#endif
|
|
allocate_memory(table, len1 + len2);
|
|
m_ptr[0]= has_memory() ? m_memory : 0;
|
|
m_ptr[1]= has_memory() ? m_memory + len1 : 0;
|
|
}
|
|
|
|
~Row_data_memory()
|
|
{
|
|
if (m_memory != 0 && m_release_memory_on_destruction)
|
|
my_free((uchar*) m_memory, MYF(MY_WME));
|
|
}
|
|
|
|
/**
|
|
Is there memory allocated?
|
|
|
|
@retval true There is memory allocated
|
|
@retval false Memory allocation failed
|
|
*/
|
|
bool has_memory() const {
|
|
#ifndef DBUG_OFF
|
|
m_alloc_checked= TRUE;
|
|
#endif
|
|
return m_memory != 0;
|
|
}
|
|
|
|
uchar *slot(uint s)
|
|
{
|
|
DBUG_ASSERT(s < sizeof(m_ptr)/sizeof(*m_ptr));
|
|
DBUG_ASSERT(m_ptr[s] != 0);
|
|
DBUG_ASSERT(m_alloc_checked == TRUE);
|
|
return m_ptr[s];
|
|
}
|
|
|
|
private:
|
|
void allocate_memory(TABLE *const table, size_t const total_length)
|
|
{
|
|
if (table->s->blob_fields == 0)
|
|
{
|
|
/*
|
|
The maximum length of a packed record is less than this
|
|
length. We use this value instead of the supplied length
|
|
when allocating memory for records, since we don't know how
|
|
the memory will be used in future allocations.
|
|
|
|
Since table->s->reclength is for unpacked records, we have
|
|
to add two bytes for each field, which can potentially be
|
|
added to hold the length of a packed field.
|
|
*/
|
|
size_t const maxlen= table->s->reclength + 2 * table->s->fields;
|
|
|
|
/*
|
|
Allocate memory for two records if memory hasn't been
|
|
allocated. We allocate memory for two records so that it can
|
|
be used when processing update rows as well.
|
|
*/
|
|
if (table->write_row_record == 0)
|
|
table->write_row_record=
|
|
(uchar *) alloc_root(&table->mem_root, 2 * maxlen);
|
|
m_memory= table->write_row_record;
|
|
m_release_memory_on_destruction= FALSE;
|
|
}
|
|
else
|
|
{
|
|
m_memory= (uchar *) my_malloc(total_length, MYF(MY_WME));
|
|
m_release_memory_on_destruction= TRUE;
|
|
}
|
|
}
|
|
|
|
#ifndef DBUG_OFF
|
|
mutable bool m_alloc_checked;
|
|
#endif
|
|
bool m_release_memory_on_destruction;
|
|
uchar *m_memory;
|
|
uchar *m_ptr[2];
|
|
};
|
|
}
|
|
|
|
|
|
int THD::binlog_write_row(TABLE* table, bool is_trans,
|
|
MY_BITMAP const* cols, size_t colcnt,
|
|
uchar const *record)
|
|
{
|
|
DBUG_ASSERT(current_stmt_binlog_row_based && mysql_bin_log.is_open());
|
|
|
|
/*
|
|
Pack records into format for transfer. We are allocating more
|
|
memory than needed, but that doesn't matter.
|
|
*/
|
|
Row_data_memory memory(table, max_row_length(table, record));
|
|
if (!memory.has_memory())
|
|
return HA_ERR_OUT_OF_MEM;
|
|
|
|
uchar *row_data= memory.slot(0);
|
|
|
|
size_t const len= pack_row(table, cols, row_data, record);
|
|
|
|
Rows_log_event* const ev=
|
|
binlog_prepare_pending_rows_event(table, server_id, cols, colcnt,
|
|
len, is_trans,
|
|
static_cast<Write_rows_log_event*>(0));
|
|
|
|
if (unlikely(ev == 0))
|
|
return HA_ERR_OUT_OF_MEM;
|
|
|
|
return ev->add_row_data(row_data, len);
|
|
}
|
|
|
|
int THD::binlog_update_row(TABLE* table, bool is_trans,
|
|
MY_BITMAP const* cols, size_t colcnt,
|
|
const uchar *before_record,
|
|
const uchar *after_record)
|
|
{
|
|
DBUG_ASSERT(current_stmt_binlog_row_based && mysql_bin_log.is_open());
|
|
|
|
size_t const before_maxlen = max_row_length(table, before_record);
|
|
size_t const after_maxlen = max_row_length(table, after_record);
|
|
|
|
Row_data_memory row_data(table, before_maxlen, after_maxlen);
|
|
if (!row_data.has_memory())
|
|
return HA_ERR_OUT_OF_MEM;
|
|
|
|
uchar *before_row= row_data.slot(0);
|
|
uchar *after_row= row_data.slot(1);
|
|
|
|
size_t const before_size= pack_row(table, cols, before_row,
|
|
before_record);
|
|
size_t const after_size= pack_row(table, cols, after_row,
|
|
after_record);
|
|
|
|
/*
|
|
Don't print debug messages when running valgrind since they can
|
|
trigger false warnings.
|
|
*/
|
|
#ifndef HAVE_purify
|
|
DBUG_DUMP("before_record", before_record, table->s->reclength);
|
|
DBUG_DUMP("after_record", after_record, table->s->reclength);
|
|
DBUG_DUMP("before_row", before_row, before_size);
|
|
DBUG_DUMP("after_row", after_row, after_size);
|
|
#endif
|
|
|
|
Rows_log_event* const ev=
|
|
binlog_prepare_pending_rows_event(table, server_id, cols, colcnt,
|
|
before_size + after_size, is_trans,
|
|
static_cast<Update_rows_log_event*>(0));
|
|
|
|
if (unlikely(ev == 0))
|
|
return HA_ERR_OUT_OF_MEM;
|
|
|
|
return
|
|
ev->add_row_data(before_row, before_size) ||
|
|
ev->add_row_data(after_row, after_size);
|
|
}
|
|
|
|
int THD::binlog_delete_row(TABLE* table, bool is_trans,
|
|
MY_BITMAP const* cols, size_t colcnt,
|
|
uchar const *record)
|
|
{
|
|
DBUG_ASSERT(current_stmt_binlog_row_based && mysql_bin_log.is_open());
|
|
|
|
/*
|
|
Pack records into format for transfer. We are allocating more
|
|
memory than needed, but that doesn't matter.
|
|
*/
|
|
Row_data_memory memory(table, max_row_length(table, record));
|
|
if (unlikely(!memory.has_memory()))
|
|
return HA_ERR_OUT_OF_MEM;
|
|
|
|
uchar *row_data= memory.slot(0);
|
|
|
|
size_t const len= pack_row(table, cols, row_data, record);
|
|
|
|
Rows_log_event* const ev=
|
|
binlog_prepare_pending_rows_event(table, server_id, cols, colcnt,
|
|
len, is_trans,
|
|
static_cast<Delete_rows_log_event*>(0));
|
|
|
|
if (unlikely(ev == 0))
|
|
return HA_ERR_OUT_OF_MEM;
|
|
|
|
return ev->add_row_data(row_data, len);
|
|
}
|
|
|
|
|
|
int THD::binlog_remove_pending_rows_event(bool clear_maps)
|
|
{
|
|
DBUG_ENTER("THD::binlog_remove_pending_rows_event");
|
|
|
|
if (!mysql_bin_log.is_open())
|
|
DBUG_RETURN(0);
|
|
|
|
mysql_bin_log.remove_pending_rows_event(this);
|
|
|
|
if (clear_maps)
|
|
binlog_table_maps= 0;
|
|
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
int THD::binlog_flush_pending_rows_event(bool stmt_end)
|
|
{
|
|
DBUG_ENTER("THD::binlog_flush_pending_rows_event");
|
|
/*
|
|
We shall flush the pending event even if we are not in row-based
|
|
mode: it might be the case that we left row-based mode before
|
|
flushing anything (e.g., if we have explicitly locked tables).
|
|
*/
|
|
if (!mysql_bin_log.is_open())
|
|
DBUG_RETURN(0);
|
|
|
|
/*
|
|
Mark the event as the last event of a statement if the stmt_end
|
|
flag is set.
|
|
*/
|
|
int error= 0;
|
|
if (Rows_log_event *pending= binlog_get_pending_rows_event())
|
|
{
|
|
if (stmt_end)
|
|
{
|
|
pending->set_flags(Rows_log_event::STMT_END_F);
|
|
pending->flags|= LOG_EVENT_UPDATE_TABLE_MAP_VERSION_F;
|
|
binlog_table_maps= 0;
|
|
}
|
|
|
|
error= mysql_bin_log.flush_and_set_pending_rows_event(this, 0);
|
|
}
|
|
|
|
DBUG_RETURN(error);
|
|
}
|
|
|
|
|
|
#if !defined(DBUG_OFF) && !defined(_lint)
|
|
static const char *
|
|
show_query_type(THD::enum_binlog_query_type qtype)
|
|
{
|
|
switch (qtype) {
|
|
case THD::ROW_QUERY_TYPE:
|
|
return "ROW";
|
|
case THD::STMT_QUERY_TYPE:
|
|
return "STMT";
|
|
case THD::MYSQL_QUERY_TYPE:
|
|
return "MYSQL";
|
|
case THD::QUERY_TYPE_COUNT:
|
|
default:
|
|
DBUG_ASSERT(0 <= qtype && qtype < THD::QUERY_TYPE_COUNT);
|
|
}
|
|
static char buf[64];
|
|
sprintf(buf, "UNKNOWN#%d", qtype);
|
|
return buf;
|
|
}
|
|
#endif
|
|
|
|
|
|
/*
|
|
Member function that will log query, either row-based or
|
|
statement-based depending on the value of the 'current_stmt_binlog_row_based'
|
|
the value of the 'qtype' flag.
|
|
|
|
This function should be called after the all calls to ha_*_row()
|
|
functions have been issued, but before tables are unlocked and
|
|
closed.
|
|
|
|
OBSERVE
|
|
There shall be no writes to any system table after calling
|
|
binlog_query(), so these writes has to be moved to before the call
|
|
of binlog_query() for correct functioning.
|
|
|
|
This is necessesary not only for RBR, but the master might crash
|
|
after binlogging the query but before changing the system tables.
|
|
This means that the slave and the master are not in the same state
|
|
(after the master has restarted), so therefore we have to
|
|
eliminate this problem.
|
|
|
|
RETURN VALUE
|
|
Error code, or 0 if no error.
|
|
*/
|
|
int THD::binlog_query(THD::enum_binlog_query_type qtype, char const *query_arg,
|
|
ulong query_len, bool is_trans, bool suppress_use,
|
|
int errcode)
|
|
{
|
|
DBUG_ENTER("THD::binlog_query");
|
|
DBUG_PRINT("enter", ("qtype: %s query: '%s'",
|
|
show_query_type(qtype), query_arg));
|
|
DBUG_ASSERT(query_arg && mysql_bin_log.is_open());
|
|
|
|
/*
|
|
If we are not in prelocked mode, mysql_unlock_tables() will be
|
|
called after this binlog_query(), so we have to flush the pending
|
|
rows event with the STMT_END_F set to unlock all tables at the
|
|
slave side as well.
|
|
|
|
If we are in prelocked mode, the flushing will be done inside the
|
|
top-most close_thread_tables().
|
|
*/
|
|
if (this->prelocked_mode == NON_PRELOCKED)
|
|
if (int error= binlog_flush_pending_rows_event(TRUE))
|
|
DBUG_RETURN(error);
|
|
|
|
/*
|
|
If we are in statement mode and trying to log an unsafe statement,
|
|
we should print a warning.
|
|
*/
|
|
if (sql_log_bin_toplevel && lex->is_stmt_unsafe() &&
|
|
variables.binlog_format == BINLOG_FORMAT_STMT &&
|
|
binlog_filter->db_ok(this->db))
|
|
{
|
|
/*
|
|
A warning can be elevated a error when STRICT sql mode.
|
|
But we don't want to elevate binlog warning to error here.
|
|
*/
|
|
push_warning(this, MYSQL_ERROR::WARN_LEVEL_NOTE,
|
|
ER_BINLOG_UNSAFE_STATEMENT,
|
|
ER(ER_BINLOG_UNSAFE_STATEMENT));
|
|
if (global_system_variables.log_warnings &&
|
|
!(binlog_flags & BINLOG_FLAG_UNSAFE_STMT_PRINTED))
|
|
{
|
|
sql_print_warning("%s Statement: %.*s",
|
|
ER(ER_BINLOG_UNSAFE_STATEMENT),
|
|
MYSQL_ERRMSG_SIZE, query_arg);
|
|
binlog_flags|= BINLOG_FLAG_UNSAFE_STMT_PRINTED;
|
|
}
|
|
}
|
|
|
|
switch (qtype) {
|
|
case THD::ROW_QUERY_TYPE:
|
|
DBUG_PRINT("debug",
|
|
("current_stmt_binlog_row_based: %d",
|
|
current_stmt_binlog_row_based));
|
|
if (current_stmt_binlog_row_based)
|
|
DBUG_RETURN(0);
|
|
/* Otherwise, we fall through */
|
|
case THD::MYSQL_QUERY_TYPE:
|
|
/*
|
|
Using this query type is a conveniece hack, since we have been
|
|
moving back and forth between using RBR for replication of
|
|
system tables and not using it.
|
|
|
|
Make sure to change in check_table_binlog_row_based() according
|
|
to how you treat this.
|
|
*/
|
|
case THD::STMT_QUERY_TYPE:
|
|
/*
|
|
The MYSQL_LOG::write() function will set the STMT_END_F flag and
|
|
flush the pending rows event if necessary.
|
|
*/
|
|
{
|
|
Query_log_event qinfo(this, query_arg, query_len, is_trans, suppress_use,
|
|
errcode);
|
|
qinfo.flags|= LOG_EVENT_UPDATE_TABLE_MAP_VERSION_F;
|
|
/*
|
|
Binlog table maps will be irrelevant after a Query_log_event
|
|
(they are just removed on the slave side) so after the query
|
|
log event is written to the binary log, we pretend that no
|
|
table maps were written.
|
|
*/
|
|
int error= mysql_bin_log.write(&qinfo);
|
|
binlog_table_maps= 0;
|
|
DBUG_RETURN(error);
|
|
}
|
|
break;
|
|
|
|
case THD::QUERY_TYPE_COUNT:
|
|
default:
|
|
DBUG_ASSERT(0 <= qtype && qtype < QUERY_TYPE_COUNT);
|
|
}
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
bool Discrete_intervals_list::append(ulonglong start, ulonglong val,
|
|
ulonglong incr)
|
|
{
|
|
DBUG_ENTER("Discrete_intervals_list::append");
|
|
/* first, see if this can be merged with previous */
|
|
if ((head == NULL) || tail->merge_if_contiguous(start, val, incr))
|
|
{
|
|
/* it cannot, so need to add a new interval */
|
|
Discrete_interval *new_interval= new Discrete_interval(start, val, incr);
|
|
DBUG_RETURN(append(new_interval));
|
|
}
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
bool Discrete_intervals_list::append(Discrete_interval *new_interval)
|
|
{
|
|
DBUG_ENTER("Discrete_intervals_list::append");
|
|
if (unlikely(new_interval == NULL))
|
|
DBUG_RETURN(1);
|
|
DBUG_PRINT("info",("adding new auto_increment interval"));
|
|
if (head == NULL)
|
|
head= current= new_interval;
|
|
else
|
|
tail->next= new_interval;
|
|
tail= new_interval;
|
|
elements++;
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
#endif /* !defined(MYSQL_CLIENT) */
|