mariadb/storage/innobase/handler/handler0alter.cc
Sergei Golubchik 51d457f371 compilation failures
with -DPLUGIN_PARTITION=NO and -DPLUGIN_PERFSCHEMA=NO
2017-07-05 17:15:59 +02:00

9519 lines
260 KiB
C++

/*****************************************************************************
Copyright (c) 2005, 2016, Oracle and/or its affiliates. All Rights Reserved.
Copyright (c) 2013, 2017, MariaDB Corporation.
This program is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free Software
Foundation; version 2 of the License.
This program is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with
this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Suite 500, Boston, MA 02110-1335 USA
*****************************************************************************/
/**************************************************//**
@file handler/handler0alter.cc
Smart ALTER TABLE
*******************************************************/
/* Include necessary SQL headers */
#include "ha_prototypes.h"
#include <debug_sync.h>
#include <log.h>
#include <sql_lex.h>
#include <sql_class.h>
#include <sql_table.h>
#include <mysql/plugin.h>
/* Include necessary InnoDB headers */
#include "btr0sea.h"
#include "dict0crea.h"
#include "dict0dict.h"
#include "dict0priv.h"
#include "dict0stats.h"
#include "dict0stats_bg.h"
#include "fsp0sysspace.h"
#include "log0log.h"
#include "rem0types.h"
#include "row0log.h"
#include "row0merge.h"
#include "trx0trx.h"
#include "trx0roll.h"
#include "handler0alter.h"
#include "srv0mon.h"
#include "fts0priv.h"
#include "fts0plugin.h"
#include "pars0pars.h"
#include "row0sel.h"
#include "ha_innodb.h"
#include "ut0new.h"
#include "ut0stage.h"
#ifdef WITH_WSREP
//#include "wsrep_api.h"
#include <sql_acl.h> // PROCESS_ACL
#endif
static const char *MSG_UNSUPPORTED_ALTER_ONLINE_ON_VIRTUAL_COLUMN=
"INPLACE ADD or DROP of virtual columns cannot be "
"combined with other ALTER TABLE actions";
/* For supporting Native InnoDB Partitioning. */
/* JAN: TODO: MySQL 5.7
#include "partition_info.h"
#include "ha_innopart.h"
*/
/** Operations for creating secondary indexes (no rebuild needed) */
static const Alter_inplace_info::HA_ALTER_FLAGS INNOBASE_ONLINE_CREATE
= Alter_inplace_info::ADD_INDEX
| Alter_inplace_info::ADD_UNIQUE_INDEX
| Alter_inplace_info::ADD_SPATIAL_INDEX;
/** Operations for rebuilding a table in place */
static const Alter_inplace_info::HA_ALTER_FLAGS INNOBASE_ALTER_REBUILD
= Alter_inplace_info::ADD_PK_INDEX
| Alter_inplace_info::DROP_PK_INDEX
| Alter_inplace_info::CHANGE_CREATE_OPTION
/* CHANGE_CREATE_OPTION needs to check innobase_need_rebuild() */
| Alter_inplace_info::ALTER_COLUMN_NULLABLE
| Alter_inplace_info::ALTER_COLUMN_NOT_NULLABLE
| Alter_inplace_info::ALTER_STORED_COLUMN_ORDER
| Alter_inplace_info::DROP_STORED_COLUMN
| Alter_inplace_info::ADD_STORED_BASE_COLUMN
| Alter_inplace_info::RECREATE_TABLE
/*
| Alter_inplace_info::ALTER_STORED_COLUMN_TYPE
*/
;
/** Operations that require changes to data */
static const Alter_inplace_info::HA_ALTER_FLAGS INNOBASE_ALTER_DATA
= INNOBASE_ONLINE_CREATE | INNOBASE_ALTER_REBUILD;
/** Operations for altering a table that InnoDB does not care about */
static const Alter_inplace_info::HA_ALTER_FLAGS INNOBASE_INPLACE_IGNORE
= Alter_inplace_info::ALTER_COLUMN_DEFAULT
| Alter_inplace_info::ALTER_PARTITIONED
| Alter_inplace_info::ALTER_COLUMN_COLUMN_FORMAT
| Alter_inplace_info::ALTER_COLUMN_STORAGE_TYPE
| Alter_inplace_info::ALTER_VIRTUAL_GCOL_EXPR
| Alter_inplace_info::ALTER_RENAME;
/** Operations on foreign key definitions (changing the schema only) */
static const Alter_inplace_info::HA_ALTER_FLAGS INNOBASE_FOREIGN_OPERATIONS
= Alter_inplace_info::DROP_FOREIGN_KEY
| Alter_inplace_info::ADD_FOREIGN_KEY;
/** Operations that InnoDB cares about and can perform without rebuild */
static const Alter_inplace_info::HA_ALTER_FLAGS INNOBASE_ALTER_NOREBUILD
= INNOBASE_ONLINE_CREATE
| INNOBASE_FOREIGN_OPERATIONS
| Alter_inplace_info::DROP_INDEX
| Alter_inplace_info::DROP_UNIQUE_INDEX
#ifdef MYSQL_RENAME_INDEX
| Alter_inplace_info::RENAME_INDEX
#endif
| Alter_inplace_info::ALTER_COLUMN_NAME
| Alter_inplace_info::ALTER_COLUMN_EQUAL_PACK_LENGTH
//| Alter_inplace_info::ALTER_INDEX_COMMENT
| Alter_inplace_info::ADD_VIRTUAL_COLUMN
| Alter_inplace_info::DROP_VIRTUAL_COLUMN
| Alter_inplace_info::ALTER_VIRTUAL_COLUMN_ORDER;
struct ha_innobase_inplace_ctx : public inplace_alter_handler_ctx
{
/** Dummy query graph */
que_thr_t* thr;
/** The prebuilt struct of the creating instance */
row_prebuilt_t*& prebuilt;
/** InnoDB indexes being created */
dict_index_t** add_index;
/** MySQL key numbers for the InnoDB indexes that are being created */
const ulint* add_key_numbers;
/** number of InnoDB indexes being created */
ulint num_to_add_index;
/** InnoDB indexes being dropped */
dict_index_t** drop_index;
/** number of InnoDB indexes being dropped */
const ulint num_to_drop_index;
/** InnoDB indexes being renamed */
dict_index_t** rename;
/** number of InnoDB indexes being renamed */
const ulint num_to_rename;
/** InnoDB foreign key constraints being dropped */
dict_foreign_t** drop_fk;
/** number of InnoDB foreign key constraints being dropped */
const ulint num_to_drop_fk;
/** InnoDB foreign key constraints being added */
dict_foreign_t** add_fk;
/** number of InnoDB foreign key constraints being dropped */
const ulint num_to_add_fk;
/** whether to create the indexes online */
bool online;
/** memory heap */
mem_heap_t* heap;
/** dictionary transaction */
trx_t* trx;
/** original table (if rebuilt, differs from indexed_table) */
dict_table_t* old_table;
/** table where the indexes are being created or dropped */
dict_table_t* new_table;
/** mapping of old column numbers to new ones, or NULL */
const ulint* col_map;
/** new column names, or NULL if nothing was renamed */
const char** col_names;
/** added AUTO_INCREMENT column position, or ULINT_UNDEFINED */
const ulint add_autoinc;
/** default values of ADD COLUMN, or NULL */
const dtuple_t* add_cols;
/** autoinc sequence to use */
ib_sequence_t sequence;
/** temporary table name to use for old table when renaming tables */
const char* tmp_name;
/** whether the order of the clustered index is unchanged */
bool skip_pk_sort;
/** number of virtual columns to be added */
ulint num_to_add_vcol;
/** virtual columns to be added */
dict_v_col_t* add_vcol;
const char** add_vcol_name;
/** number of virtual columns to be dropped */
ulint num_to_drop_vcol;
/** virtual columns to be dropped */
dict_v_col_t* drop_vcol;
const char** drop_vcol_name;
/** ALTER TABLE stage progress recorder */
ut_stage_alter_t* m_stage;
ha_innobase_inplace_ctx(row_prebuilt_t*& prebuilt_arg,
dict_index_t** drop_arg,
ulint num_to_drop_arg,
dict_index_t** rename_arg,
ulint num_to_rename_arg,
dict_foreign_t** drop_fk_arg,
ulint num_to_drop_fk_arg,
dict_foreign_t** add_fk_arg,
ulint num_to_add_fk_arg,
bool online_arg,
mem_heap_t* heap_arg,
dict_table_t* new_table_arg,
const char** col_names_arg,
ulint add_autoinc_arg,
ulonglong autoinc_col_min_value_arg,
ulonglong autoinc_col_max_value_arg,
ulint num_to_drop_vcol_arg) :
inplace_alter_handler_ctx(),
prebuilt (prebuilt_arg),
add_index (0), add_key_numbers (0), num_to_add_index (0),
drop_index (drop_arg), num_to_drop_index (num_to_drop_arg),
rename (rename_arg), num_to_rename (num_to_rename_arg),
drop_fk (drop_fk_arg), num_to_drop_fk (num_to_drop_fk_arg),
add_fk (add_fk_arg), num_to_add_fk (num_to_add_fk_arg),
online (online_arg), heap (heap_arg), trx (0),
old_table (prebuilt_arg->table),
new_table (new_table_arg),
col_map (0), col_names (col_names_arg),
add_autoinc (add_autoinc_arg),
add_cols (0),
sequence(prebuilt->trx->mysql_thd,
autoinc_col_min_value_arg, autoinc_col_max_value_arg),
tmp_name (0),
skip_pk_sort(false),
num_to_add_vcol(0),
add_vcol(0),
add_vcol_name(0),
num_to_drop_vcol(0),
drop_vcol(0),
drop_vcol_name(0),
m_stage(NULL)
{
#ifdef UNIV_DEBUG
for (ulint i = 0; i < num_to_add_index; i++) {
ut_ad(!add_index[i]->to_be_dropped);
}
for (ulint i = 0; i < num_to_drop_index; i++) {
ut_ad(drop_index[i]->to_be_dropped);
}
#endif /* UNIV_DEBUG */
thr = pars_complete_graph_for_exec(NULL, prebuilt->trx, heap,
prebuilt);
}
~ha_innobase_inplace_ctx()
{
UT_DELETE(m_stage);
mem_heap_free(heap);
}
/** Determine if the table will be rebuilt.
@return whether the table will be rebuilt */
bool need_rebuild () const { return(old_table != new_table); }
private:
// Disable copying
ha_innobase_inplace_ctx(const ha_innobase_inplace_ctx&);
ha_innobase_inplace_ctx& operator=(const ha_innobase_inplace_ctx&);
};
/********************************************************************//**
Get the upper limit of the MySQL integral and floating-point type.
@return maximum allowed value for the field */
UNIV_INTERN
ulonglong
innobase_get_int_col_max_value(
/*===========================*/
const Field* field); /*!< in: MySQL field */
/* Report an InnoDB error to the client by invoking my_error(). */
static UNIV_COLD MY_ATTRIBUTE((nonnull))
void
my_error_innodb(
/*============*/
dberr_t error, /*!< in: InnoDB error code */
const char* table, /*!< in: table name */
ulint flags) /*!< in: table flags */
{
switch (error) {
case DB_MISSING_HISTORY:
my_error(ER_TABLE_DEF_CHANGED, MYF(0));
break;
case DB_RECORD_NOT_FOUND:
my_error(ER_KEY_NOT_FOUND, MYF(0), table);
break;
case DB_DEADLOCK:
my_error(ER_LOCK_DEADLOCK, MYF(0));
break;
case DB_LOCK_WAIT_TIMEOUT:
my_error(ER_LOCK_WAIT_TIMEOUT, MYF(0));
break;
case DB_INTERRUPTED:
my_error(ER_QUERY_INTERRUPTED, MYF(0));
break;
case DB_OUT_OF_MEMORY:
my_error(ER_OUT_OF_RESOURCES, MYF(0));
break;
case DB_OUT_OF_FILE_SPACE:
my_error(ER_RECORD_FILE_FULL, MYF(0), table);
break;
case DB_TEMP_FILE_WRITE_FAIL:
my_error(ER_TEMP_FILE_WRITE_FAILURE, MYF(0));
break;
case DB_TOO_BIG_INDEX_COL:
my_error(ER_INDEX_COLUMN_TOO_LONG, MYF(0),
DICT_MAX_FIELD_LEN_BY_FORMAT_FLAG(flags));
break;
case DB_TOO_MANY_CONCURRENT_TRXS:
my_error(ER_TOO_MANY_CONCURRENT_TRXS, MYF(0));
break;
case DB_LOCK_TABLE_FULL:
my_error(ER_LOCK_TABLE_FULL, MYF(0));
break;
case DB_UNDO_RECORD_TOO_BIG:
my_error(ER_UNDO_RECORD_TOO_BIG, MYF(0));
break;
case DB_CORRUPTION:
my_error(ER_NOT_KEYFILE, MYF(0), table);
break;
case DB_TOO_BIG_RECORD:
/* We limit max record size to 16k for 64k page size. */
my_error(ER_TOO_BIG_ROWSIZE, MYF(0),
srv_page_size == UNIV_PAGE_SIZE_MAX
? REC_MAX_DATA_SIZE - 1
: page_get_free_space_of_empty(
flags & DICT_TF_COMPACT) / 2);
break;
case DB_INVALID_NULL:
/* TODO: report the row, as we do for DB_DUPLICATE_KEY */
my_error(ER_INVALID_USE_OF_NULL, MYF(0));
break;
case DB_CANT_CREATE_GEOMETRY_OBJECT:
my_error(ER_CANT_CREATE_GEOMETRY_OBJECT, MYF(0));
break;
case DB_TABLESPACE_EXISTS:
my_error(ER_TABLESPACE_EXISTS, MYF(0), table);
break;
#ifdef UNIV_DEBUG
case DB_SUCCESS:
case DB_DUPLICATE_KEY:
case DB_ONLINE_LOG_TOO_BIG:
/* These codes should not be passed here. */
ut_error;
#endif /* UNIV_DEBUG */
default:
my_error(ER_GET_ERRNO, MYF(0), error, "InnoDB");
break;
}
}
/** Determine if fulltext indexes exist in a given table.
@param table MySQL table
@return whether fulltext indexes exist on the table */
static
bool
innobase_fulltext_exist(
/*====================*/
const TABLE* table)
{
for (uint i = 0; i < table->s->keys; i++) {
if (table->key_info[i].flags & HA_FULLTEXT) {
return(true);
}
}
return(false);
}
/** Determine if spatial indexes exist in a given table.
@param table MySQL table
@return whether spatial indexes exist on the table */
static
bool
innobase_spatial_exist(
/*===================*/
const TABLE* table)
{
for (uint i = 0; i < table->s->keys; i++) {
if (table->key_info[i].flags & HA_SPATIAL) {
return(true);
}
}
return(false);
}
/*******************************************************************//**
Determine if ALTER TABLE needs to rebuild the table.
@param ha_alter_info the DDL operation
@param altered_table MySQL original table
@return whether it is necessary to rebuild the table */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_need_rebuild(
/*==================*/
const Alter_inplace_info* ha_alter_info,
const TABLE* altered_table)
{
Alter_inplace_info::HA_ALTER_FLAGS alter_inplace_flags =
ha_alter_info->handler_flags & ~(INNOBASE_INPLACE_IGNORE);
if (alter_inplace_flags
== Alter_inplace_info::CHANGE_CREATE_OPTION
&& !(ha_alter_info->create_info->used_fields
& (HA_CREATE_USED_ROW_FORMAT
| HA_CREATE_USED_KEY_BLOCK_SIZE))) {
// JAN: TODO: MySQL 5.7
// | HA_CREATE_USED_TABLESPACE))) {
/* Any other CHANGE_CREATE_OPTION than changing
ROW_FORMAT, KEY_BLOCK_SIZE or TABLESPACE can be done
without rebuilding the table. */
return(false);
}
/* If alter table changes column name and adds a new
index, we need to check is this new index created
to new column name. This is because column name
changes are done normally after creating indexes. */
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME) &&
((ha_alter_info->handler_flags
& Alter_inplace_info::ADD_INDEX) ||
(ha_alter_info->handler_flags
& Alter_inplace_info::ADD_FOREIGN_KEY))) {
for (ulint i = 0; i < ha_alter_info->index_add_count; i++) {
const KEY* key = &ha_alter_info->key_info_buffer[
ha_alter_info->index_add_buffer[i]];
for (ulint j = 0; j < key->user_defined_key_parts; j++) {
const KEY_PART_INFO* key_part = &(key->key_part[j]);
const Field* field = altered_table->field[key_part->fieldnr];
/* Field used on added index is renamed on
this same alter table. We need table
rebuild. */
if (field && field->flags & FIELD_IS_RENAMED) {
return (true);
}
}
}
}
return(!!(ha_alter_info->handler_flags & INNOBASE_ALTER_REBUILD));
}
/** Check if virtual column in old and new table are in order, excluding
those dropped column. This is needed because when we drop a virtual column,
ALTER_VIRTUAL_COLUMN_ORDER is also turned on, so we can't decide if this
is a real ORDER change or just DROP COLUMN
@param[in] table old TABLE
@param[in] altered_table new TABLE
@param[in] ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@return true is all columns in order, false otherwise. */
static
bool
check_v_col_in_order(
const TABLE* table,
const TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
ulint j = 0;
/* We don't support any adding new virtual column before
existed virtual column. */
if (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_VIRTUAL_COLUMN) {
bool has_new = false;
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
cf_it.rewind();
while (const Create_field* new_field = cf_it++) {
if (!innobase_is_v_fld(new_field)) {
continue;
}
/* Found a new added virtual column. */
if (!new_field->field) {
has_new = true;
continue;
}
/* If there's any old virtual column
after the new added virtual column,
order must be changed. */
if (has_new) {
return(false);
}
}
}
/* directly return true if ALTER_VIRTUAL_COLUMN_ORDER is not on */
if (!(ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_VIRTUAL_COLUMN_ORDER)) {
return(true);
}
for (ulint i = 0; i < table->s->fields; i++) {
Field* field = table->field[i];
if (field->stored_in_db()) {
continue;
}
ut_ad(innobase_is_v_fld(field));
if (field->flags & FIELD_IS_DROPPED) {
continue;
}
/* Now check if the next virtual column in altered table
matches this column */
while (j < altered_table->s->fields) {
Field* new_field = altered_table->s->field[j];
if (new_field->stored_in_db()) {
j++;
continue;
}
if (my_strcasecmp(system_charset_info,
field->field_name,
new_field->field_name) != 0) {
/* different column */
return(false);
} else {
j++;
break;
}
}
if (j > altered_table->s->fields) {
/* there should not be less column in new table
without them being in drop list */
ut_ad(0);
return(false);
}
}
return(true);
}
/** Check if InnoDB supports a particular alter table in-place
@param altered_table TABLE object for new version of table.
@param ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@retval HA_ALTER_INPLACE_NOT_SUPPORTED Not supported
@retval HA_ALTER_INPLACE_NO_LOCK Supported
@retval HA_ALTER_INPLACE_SHARED_LOCK_AFTER_PREPARE Supported, but requires
lock during main phase and exclusive lock during prepare phase.
@retval HA_ALTER_INPLACE_NO_LOCK_AFTER_PREPARE Supported, prepare phase
requires exclusive lock (any transactions that have accessed the table
must commit or roll back first, and no transactions can access the table
while prepare_inplace_alter_table() is executing)
*/
enum_alter_inplace_result
ha_innobase::check_if_supported_inplace_alter(
/*==========================================*/
TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
DBUG_ENTER("check_if_supported_inplace_alter");
if (high_level_read_only
|| srv_sys_space.created_new_raw()
|| srv_force_recovery) {
ha_alter_info->unsupported_reason = (srv_force_recovery)?
"Operation not allowed when innodb_forced_recovery > 0." :
innobase_get_err_msg(ER_READ_ONLY_MODE);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
if (altered_table->s->fields > REC_MAX_N_USER_FIELDS) {
/* Deny the inplace ALTER TABLE. MySQL will try to
re-create the table and ha_innobase::create() will
return an error too. This is how we effectively
deny adding too many columns to a table. */
ha_alter_info->unsupported_reason =
innobase_get_err_msg(ER_TOO_MANY_FIELDS);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
update_thd();
/* Change on engine specific table options require rebuild of the
table */
if (ha_alter_info->handler_flags
& Alter_inplace_info::CHANGE_CREATE_OPTION) {
ha_table_option_struct *new_options= ha_alter_info->create_info->option_struct;
ha_table_option_struct *old_options= table->s->option_struct;
if (new_options->page_compressed != old_options->page_compressed ||
new_options->page_compression_level != old_options->page_compression_level) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
if (new_options->encryption != old_options->encryption ||
new_options->encryption_key_id != old_options->encryption_key_id) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
if (ha_alter_info->handler_flags
& ~(INNOBASE_INPLACE_IGNORE
| INNOBASE_ALTER_NOREBUILD
| INNOBASE_ALTER_REBUILD)) {
if (ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_STORED_COLUMN_TYPE) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_COLUMN_TYPE);
}
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
/* Only support online add foreign key constraint when
check_foreigns is turned off */
if ((ha_alter_info->handler_flags & Alter_inplace_info::ADD_FOREIGN_KEY)
&& m_prebuilt->trx->check_foreigns) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_FK_CHECK);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
#if 0
if (altered_table->file->ht != ht) {
/* Non-native partitioning table engine. No longer supported,
due to implementation of native InnoDB partitioning. */
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
#endif
if (!(ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE)) {
DBUG_RETURN(HA_ALTER_INPLACE_NO_LOCK);
}
/* Only support NULL -> NOT NULL change if strict table sql_mode
is set. Fall back to COPY for conversion if not strict tables.
In-Place will fail with an error when trying to convert
NULL to a NOT NULL value. */
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NOT_NULLABLE)
&& !thd_is_strict_mode(m_user_thd)) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_NOT_NULL);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
/* DROP PRIMARY KEY is only allowed in combination with ADD
PRIMARY KEY. */
if ((ha_alter_info->handler_flags
& (Alter_inplace_info::ADD_PK_INDEX
| Alter_inplace_info::DROP_PK_INDEX))
== Alter_inplace_info::DROP_PK_INDEX) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_NOPK);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
/* If a column change from NOT NULL to NULL,
and there's a implict pk on this column. the
table should be rebuild. The change should
only go through the "Copy" method. */
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NULLABLE)) {
const uint my_primary_key = altered_table->s->primary_key;
/* See if MYSQL table has no pk but we do. */
if (UNIV_UNLIKELY(my_primary_key >= MAX_KEY)
&& !row_table_got_default_clust_index(m_prebuilt->table)) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_PRIMARY_CANT_HAVE_NULL);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
/*
InnoDB in different MariaDB versions was generating different mtype
codes for certain types. In some cases the signed/unsigned bit was
generated differently too.
Online ALTER would change the mtype/unsigned_flag (to what the
current code generates) without changing the underlying data
represenation, and it might result in data corruption.
Don't do online ALTER if mtype/unsigned_flag are wrong.
*/
for (ulint i = 0, icol= 0; i < table->s->fields; i++) {
const Field* field = table->field[i];
const dict_col_t* col = dict_table_get_nth_col(m_prebuilt->table, icol);
ulint unsigned_flag;
if (!field->stored_in_db()) {
continue;
}
icol++;
if (col->mtype != get_innobase_type_from_mysql_type(&unsigned_flag, field)) {
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
if ((col->prtype & DATA_UNSIGNED) != unsigned_flag) {
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
ulint n_indexes = UT_LIST_GET_LEN((m_prebuilt->table)->indexes);
/* If InnoDB dictionary and MySQL frm file are not consistent
use "Copy" method. */
if (m_prebuilt->table->dict_frm_mismatch) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_NO_SUCH_INDEX);
ib_push_frm_error(m_user_thd, m_prebuilt->table, altered_table,
n_indexes, true);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
bool add_drop_v_cols = false;
/* If there is add or drop virtual columns, we will support operations
with these 2 options alone with inplace interface for now */
if (ha_alter_info->handler_flags
& (Alter_inplace_info::ADD_VIRTUAL_COLUMN
| Alter_inplace_info::DROP_VIRTUAL_COLUMN
| Alter_inplace_info::ALTER_VIRTUAL_COLUMN_ORDER)) {
ulonglong flags = ha_alter_info->handler_flags;
/* TODO: uncomment the flags below, once we start to
support them */
flags &= ~(Alter_inplace_info::ADD_VIRTUAL_COLUMN
| Alter_inplace_info::DROP_VIRTUAL_COLUMN
| Alter_inplace_info::ALTER_VIRTUAL_COLUMN_ORDER
| Alter_inplace_info::ALTER_VIRTUAL_GCOL_EXPR
| Alter_inplace_info::ALTER_COLUMN_VCOL
/*
| Alter_inplace_info::ADD_STORED_BASE_COLUMN
| Alter_inplace_info::DROP_STORED_COLUMN
| Alter_inplace_info::ALTER_STORED_COLUMN_ORDER
| Alter_inplace_info::ADD_UNIQUE_INDEX
*/
| Alter_inplace_info::ADD_INDEX
| Alter_inplace_info::DROP_INDEX);
if (flags != 0
|| IF_PARTITIONING((altered_table->s->partition_info_str
&& altered_table->s->partition_info_str_len), 0)
|| (!check_v_col_in_order(
this->table, altered_table, ha_alter_info))) {
ha_alter_info->unsupported_reason =
MSG_UNSUPPORTED_ALTER_ONLINE_ON_VIRTUAL_COLUMN;
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
add_drop_v_cols = true;
}
/* We should be able to do the operation in-place.
See if we can do it online (LOCK=NONE). */
bool online = true;
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
/* Fix the key parts. */
for (KEY* new_key = ha_alter_info->key_info_buffer;
new_key < ha_alter_info->key_info_buffer
+ ha_alter_info->key_count;
new_key++) {
/* Do not support adding/droping a virtual column, while
there is a table rebuild caused by adding a new FTS_DOC_ID */
if ((new_key->flags & HA_FULLTEXT) && add_drop_v_cols
&& !DICT_TF2_FLAG_IS_SET(m_prebuilt->table,
DICT_TF2_FTS_HAS_DOC_ID)) {
ha_alter_info->unsupported_reason =
MSG_UNSUPPORTED_ALTER_ONLINE_ON_VIRTUAL_COLUMN;
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
for (KEY_PART_INFO* key_part = new_key->key_part;
key_part < new_key->key_part + new_key->user_defined_key_parts;
key_part++) {
const Create_field* new_field;
DBUG_ASSERT(key_part->fieldnr
< altered_table->s->fields);
cf_it.rewind();
for (uint fieldnr = 0; (new_field = cf_it++);
fieldnr++) {
if (fieldnr == key_part->fieldnr) {
break;
}
}
DBUG_ASSERT(new_field);
key_part->field = altered_table->field[
key_part->fieldnr];
/* In some special cases InnoDB emits "false"
duplicate key errors with NULL key values. Let
us play safe and ensure that we can correctly
print key values even in such cases. */
key_part->null_offset = key_part->field->null_offset();
key_part->null_bit = key_part->field->null_bit;
if (new_field->field) {
/* This is an existing column. */
continue;
}
/* This is an added column. */
DBUG_ASSERT(ha_alter_info->handler_flags
& Alter_inplace_info::ADD_COLUMN);
/* We cannot replace a hidden FTS_DOC_ID
with a user-visible FTS_DOC_ID. */
if (m_prebuilt->table->fts
&& innobase_fulltext_exist(altered_table)
&& !my_strcasecmp(
system_charset_info,
key_part->field->field_name,
FTS_DOC_ID_COL_NAME)) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_HIDDEN_FTS);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
DBUG_ASSERT((MTYP_TYPENR(key_part->field->unireg_check)
== Field::NEXT_NUMBER)
== !!(key_part->field->flags
& AUTO_INCREMENT_FLAG));
if (key_part->field->flags & AUTO_INCREMENT_FLAG) {
/* We cannot assign an AUTO_INCREMENT
column values during online ALTER. */
DBUG_ASSERT(key_part->field == altered_table
-> found_next_number_field);
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_AUTOINC);
online = false;
}
if (innobase_is_v_fld(key_part->field)) {
/* Do not support adding index on newly added
virtual column, while there is also a drop
virtual column in the same clause */
if (ha_alter_info->handler_flags
& Alter_inplace_info::DROP_VIRTUAL_COLUMN) {
ha_alter_info->unsupported_reason =
MSG_UNSUPPORTED_ALTER_ONLINE_ON_VIRTUAL_COLUMN;
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
ha_alter_info->unsupported_reason =
MSG_UNSUPPORTED_ALTER_ONLINE_ON_VIRTUAL_COLUMN;
online = false;
}
}
}
DBUG_ASSERT(!m_prebuilt->table->fts || m_prebuilt->table->fts->doc_col
<= table->s->fields);
DBUG_ASSERT(!m_prebuilt->table->fts || m_prebuilt->table->fts->doc_col
< dict_table_get_n_user_cols(m_prebuilt->table));
/* Spatial indexes should use copy method for now.
TOO: remove this when below ADD_SPATIAL_INDEX supported. */
for (uint i = 0; i < ha_alter_info->index_add_count; i++) {
const KEY* key =
&ha_alter_info->key_info_buffer[
ha_alter_info->index_add_buffer[i]];
if (key->flags & HA_SPATIAL) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_GIS);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
#ifdef MYSQL_SPATIAL_INDEX
if (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_SPATIAL_INDEX) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_GIS);
online = false;
}
#endif
if (m_prebuilt->table->fts
&& innobase_fulltext_exist(altered_table)) {
/* FULLTEXT indexes are supposed to remain. */
/* Disallow DROP INDEX FTS_DOC_ID_INDEX */
for (uint i = 0; i < ha_alter_info->index_drop_count; i++) {
if (!my_strcasecmp(
system_charset_info,
ha_alter_info->index_drop_buffer[i]->name,
FTS_DOC_ID_INDEX_NAME)) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_CHANGE_FTS);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
/* InnoDB can have a hidden FTS_DOC_ID_INDEX on a
visible FTS_DOC_ID column as well. Prevent dropping or
renaming the FTS_DOC_ID. */
for (Field** fp = table->field; *fp; fp++) {
if (!((*fp)->flags
& (FIELD_IS_RENAMED | FIELD_IS_DROPPED))) {
continue;
}
if (!my_strcasecmp(
system_charset_info,
(*fp)->field_name,
FTS_DOC_ID_COL_NAME)) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_CHANGE_FTS);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
}
m_prebuilt->trx->will_lock++;
if (!online) {
/* We already determined that only a non-locking
operation is possible. */
} else if (((ha_alter_info->handler_flags
& Alter_inplace_info::ADD_PK_INDEX)
|| innobase_need_rebuild(ha_alter_info, table))
&& (innobase_fulltext_exist(altered_table)
|| innobase_spatial_exist(altered_table))) {
/* Refuse to rebuild the table online, if
FULLTEXT OR SPATIAL indexes are to survive the rebuild. */
online = false;
/* If the table already contains fulltext indexes,
refuse to rebuild the table natively altogether. */
if (m_prebuilt->table->fts) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_INNODB_FT_LIMIT);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
if (innobase_spatial_exist(altered_table)) {
ha_alter_info->unsupported_reason =
innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_GIS);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
} else {
ha_alter_info->unsupported_reason =
innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_FTS);
}
} else if ((ha_alter_info->handler_flags
& Alter_inplace_info::ADD_INDEX)) {
/* Building a full-text index requires a lock.
We could do without a lock if the table already contains
an FTS_DOC_ID column, but in that case we would have
to apply the modification log to the full-text indexes. */
for (uint i = 0; i < ha_alter_info->index_add_count; i++) {
const KEY* key =
&ha_alter_info->key_info_buffer[
ha_alter_info->index_add_buffer[i]];
if (key->flags & HA_FULLTEXT) {
DBUG_ASSERT(!(key->flags & HA_KEYFLAG_MASK
& ~(HA_FULLTEXT
| HA_PACK_KEY
| HA_GENERATED_KEY
| HA_BINARY_PACK_KEY)));
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_FTS);
online = false;
break;
}
}
}
/* When changing a NULL column to NOT NULL and specifying a
DEFAULT value, ensure that the DEFAULT expression is a constant.
Also, in ADD COLUMN, for now we only support a
constant DEFAULT expression. */
cf_it.rewind();
Field **af = altered_table->field;
while (Create_field* cf = cf_it++) {
DBUG_ASSERT(cf->field
|| (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_COLUMN));
if (const Field* f = cf->field) {
/* This could be changing an existing column
from NULL to NOT NULL. */
switch ((*af)->type()) {
case MYSQL_TYPE_TIMESTAMP:
case MYSQL_TYPE_TIMESTAMP2:
/* Inserting NULL into a TIMESTAMP column
would cause the DEFAULT value to be
replaced. Ensure that the DEFAULT
expression is not changing during
ALTER TABLE. */
if (!f->real_maybe_null()
|| (*af)->real_maybe_null()) {
/* The column was NOT NULL, or it
will allow NULL after ALTER TABLE. */
goto next_column;
}
if (!(*af)->default_value
&& (*af)->is_real_null()) {
/* No DEFAULT value is
specified. We can report
errors for any NULL values for
the TIMESTAMP.
FIXME: Allow any DEFAULT
expression whose value does
not change during ALTER TABLE.
This would require a fix in
row_merge_read_clustered_index()
to try to replace the DEFAULT
value before reporting
DB_INVALID_NULL. */
goto next_column;
}
break;
default:
/* For any other data type, NULL
values are not converted.
(An AUTO_INCREMENT attribute cannot
be introduced to a column with
ALGORITHM=INPLACE.) */
ut_ad((MTYP_TYPENR((*af)->unireg_check)
== Field::NEXT_NUMBER)
== (MTYP_TYPENR(f->unireg_check)
== Field::NEXT_NUMBER));
goto next_column;
}
ha_alter_info->unsupported_reason
= innobase_get_err_msg(
ER_ALTER_OPERATION_NOT_SUPPORTED_REASON_NOT_NULL);
} else if (!(*af)->default_value
|| !((*af)->default_value->flags
& ~(VCOL_SESSION_FUNC | VCOL_TIME_FUNC))) {
/* The added NOT NULL column lacks a DEFAULT value,
or the DEFAULT is the same for all rows.
(Time functions, such as CURRENT_TIMESTAMP(),
are evaluated from a timestamp that is assigned
at the start of the statement. Session
functions, such as USER(), always evaluate the
same within a statement.) */
/* Compute the DEFAULT values of non-constant columns
(VCOL_SESSION_FUNC | VCOL_TIME_FUNC). */
(*af)->set_default();
goto next_column;
}
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
next_column:
af++;
}
cf_it.rewind();
DBUG_RETURN(online
? HA_ALTER_INPLACE_NO_LOCK_AFTER_PREPARE
: HA_ALTER_INPLACE_SHARED_LOCK_AFTER_PREPARE);
}
/*************************************************************//**
Initialize the dict_foreign_t structure with supplied info
@return true if added, false if duplicate foreign->id */
static MY_ATTRIBUTE((nonnull(1,3,5,7)))
bool
innobase_init_foreign(
/*==================*/
dict_foreign_t* foreign, /*!< in/out: structure to
initialize */
char* constraint_name, /*!< in/out: constraint name if
exists */
dict_table_t* table, /*!< in: foreign table */
dict_index_t* index, /*!< in: foreign key index */
const char** column_names, /*!< in: foreign key column
names */
ulint num_field, /*!< in: number of columns */
const char* referenced_table_name, /*!< in: referenced table
name */
dict_table_t* referenced_table, /*!< in: referenced table */
dict_index_t* referenced_index, /*!< in: referenced index */
const char** referenced_column_names,/*!< in: referenced column
names */
ulint referenced_num_field) /*!< in: number of referenced
columns */
{
ut_ad(mutex_own(&dict_sys->mutex));
if (constraint_name) {
ulint db_len;
/* Catenate 'databasename/' to the constraint name specified
by the user: we conceive the constraint as belonging to the
same MySQL 'database' as the table itself. We store the name
to foreign->id. */
db_len = dict_get_db_name_len(table->name.m_name);
foreign->id = static_cast<char*>(mem_heap_alloc(
foreign->heap, db_len + strlen(constraint_name) + 2));
ut_memcpy(foreign->id, table->name.m_name, db_len);
foreign->id[db_len] = '/';
strcpy(foreign->id + db_len + 1, constraint_name);
/* Check if any existing foreign key has the same id,
this is needed only if user supplies the constraint name */
if (table->foreign_set.find(foreign)
!= table->foreign_set.end()) {
return(false);
}
}
foreign->foreign_table = table;
foreign->foreign_table_name = mem_heap_strdup(
foreign->heap, table->name.m_name);
dict_mem_foreign_table_name_lookup_set(foreign, TRUE);
foreign->foreign_index = index;
foreign->n_fields = (unsigned int) num_field;
foreign->foreign_col_names = static_cast<const char**>(
mem_heap_alloc(foreign->heap, num_field * sizeof(void*)));
for (ulint i = 0; i < foreign->n_fields; i++) {
foreign->foreign_col_names[i] = mem_heap_strdup(
foreign->heap, column_names[i]);
}
foreign->referenced_index = referenced_index;
foreign->referenced_table = referenced_table;
foreign->referenced_table_name = mem_heap_strdup(
foreign->heap, referenced_table_name);
dict_mem_referenced_table_name_lookup_set(foreign, TRUE);
foreign->referenced_col_names = static_cast<const char**>(
mem_heap_alloc(foreign->heap,
referenced_num_field * sizeof(void*)));
for (ulint i = 0; i < foreign->n_fields; i++) {
foreign->referenced_col_names[i]
= mem_heap_strdup(foreign->heap,
referenced_column_names[i]);
}
return(true);
}
/*************************************************************//**
Check whether the foreign key options is legit
@return true if it is */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_check_fk_option(
/*=====================*/
const dict_foreign_t* foreign) /*!< in: foreign key */
{
if (!foreign->foreign_index) {
return(true);
}
if (foreign->type & (DICT_FOREIGN_ON_UPDATE_SET_NULL
| DICT_FOREIGN_ON_DELETE_SET_NULL)) {
for (ulint j = 0; j < foreign->n_fields; j++) {
if ((dict_index_get_nth_col(
foreign->foreign_index, j)->prtype)
& DATA_NOT_NULL) {
/* It is not sensible to define
SET NULL if the column is not
allowed to be NULL! */
return(false);
}
}
}
return(true);
}
/*************************************************************//**
Set foreign key options
@return true if successfully set */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_set_foreign_key_option(
/*============================*/
dict_foreign_t* foreign, /*!< in:InnoDB Foreign key */
Foreign_key* fk_key) /*!< in: Foreign key info from
MySQL */
{
ut_ad(!foreign->type);
switch (fk_key->delete_opt) {
// JAN: TODO: ? MySQL 5.7 used enum fk_option directly from sql_lex.h
case FK_OPTION_NO_ACTION:
case FK_OPTION_RESTRICT:
case FK_OPTION_SET_DEFAULT:
foreign->type = DICT_FOREIGN_ON_DELETE_NO_ACTION;
break;
case FK_OPTION_CASCADE:
foreign->type = DICT_FOREIGN_ON_DELETE_CASCADE;
break;
case FK_OPTION_SET_NULL:
foreign->type = DICT_FOREIGN_ON_DELETE_SET_NULL;
break;
}
switch (fk_key->update_opt) {
case FK_OPTION_NO_ACTION:
case FK_OPTION_RESTRICT:
case FK_OPTION_SET_DEFAULT:
foreign->type |= DICT_FOREIGN_ON_UPDATE_NO_ACTION;
break;
case FK_OPTION_CASCADE:
foreign->type |= DICT_FOREIGN_ON_UPDATE_CASCADE;
break;
case FK_OPTION_SET_NULL:
foreign->type |= DICT_FOREIGN_ON_UPDATE_SET_NULL;
break;
}
return(innobase_check_fk_option(foreign));
}
/*******************************************************************//**
Check if a foreign key constraint can make use of an index
that is being created.
@return useable index, or NULL if none found */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
const KEY*
innobase_find_equiv_index(
/*======================*/
const char*const* col_names,
/*!< in: column names */
uint n_cols, /*!< in: number of columns */
const KEY* keys, /*!< in: index information */
const uint* add, /*!< in: indexes being created */
uint n_add) /*!< in: number of indexes to create */
{
for (uint i = 0; i < n_add; i++) {
const KEY* key = &keys[add[i]];
if (key->user_defined_key_parts < n_cols
|| key->flags & HA_SPATIAL) {
no_match:
continue;
}
for (uint j = 0; j < n_cols; j++) {
const KEY_PART_INFO& key_part = key->key_part[j];
uint32 col_len
= key_part.field->pack_length();
/* Any index on virtual columns cannot be used
for reference constaint */
if (innobase_is_v_fld(key_part.field)) {
goto no_match;
}
/* The MySQL pack length contains 1 or 2 bytes
length field for a true VARCHAR. */
if (key_part.field->type() == MYSQL_TYPE_VARCHAR) {
col_len -= static_cast<const Field_varstring*>(
key_part.field)->length_bytes;
}
if (key_part.length < col_len) {
/* Column prefix indexes cannot be
used for FOREIGN KEY constraints. */
goto no_match;
}
if (innobase_strcasecmp(col_names[j],
key_part.field->field_name)) {
/* Name mismatch */
goto no_match;
}
}
return(key);
}
return(NULL);
}
/*************************************************************//**
Find an index whose first fields are the columns in the array
in the same order and is not marked for deletion
@return matching index, NULL if not found */
static MY_ATTRIBUTE((nonnull(1,2,6), warn_unused_result))
dict_index_t*
innobase_find_fk_index(
/*===================*/
Alter_inplace_info* ha_alter_info,
/*!< in: alter table info */
dict_table_t* table, /*!< in: table */
const char** col_names,
/*!< in: column names, or NULL
to use table->col_names */
dict_index_t** drop_index,
/*!< in: indexes to be dropped */
ulint n_drop_index,
/*!< in: size of drop_index[] */
const char** columns,/*!< in: array of column names */
ulint n_cols) /*!< in: number of columns */
{
dict_index_t* index;
index = dict_table_get_first_index(table);
while (index != NULL) {
if (!(index->type & DICT_FTS)
&& dict_foreign_qualify_index(
table, col_names, columns, n_cols,
index, NULL, true, 0,
NULL, NULL, NULL)) {
for (ulint i = 0; i < n_drop_index; i++) {
if (index == drop_index[i]) {
/* Skip to-be-dropped indexes. */
goto next_rec;
}
}
return(index);
}
next_rec:
index = dict_table_get_next_index(index);
}
return(NULL);
}
/** Check whether given column is a base of stored column.
@param[in] col_name column name
@param[in] table table
@param[in] s_cols list of stored columns
@return true if the given column is a base of stored column,else false. */
static
bool
innobase_col_check_fk(
const char* col_name,
const dict_table_t* table,
dict_s_col_list* s_cols)
{
dict_s_col_list::const_iterator it;
for (it = s_cols->begin();
it != s_cols->end(); ++it) {
dict_s_col_t s_col = *it;
for (ulint j = 0; j < s_col.num_base; j++) {
if (strcmp(col_name, dict_table_get_col_name(
table,
s_col.base_col[j]->ind)) == 0) {
return(true);
}
}
}
return(false);
}
/** Check whether the foreign key constraint is on base of any stored columns.
@param[in] foreign Foriegn key constraing information
@param[in] table table to which the foreign key objects
to be added
@param[in] s_cols list of stored column information in the table.
@return true if yes, otherwise false. */
static
bool
innobase_check_fk_stored(
const dict_foreign_t* foreign,
const dict_table_t* table,
dict_s_col_list* s_cols)
{
ulint type = foreign->type;
type &= ~(DICT_FOREIGN_ON_DELETE_NO_ACTION
| DICT_FOREIGN_ON_UPDATE_NO_ACTION);
if (type == 0 || s_cols == NULL) {
return(false);
}
for (ulint i = 0; i < foreign->n_fields; i++) {
if (innobase_col_check_fk(
foreign->foreign_col_names[i], table, s_cols)) {
return(true);
}
}
return(false);
}
/** Create InnoDB foreign key structure from MySQL alter_info
@param[in] ha_alter_info alter table info
@param[in] table_share TABLE_SHARE
@param[in] table table object
@param[in] col_names column names, or NULL to use
table->col_names
@param[in] drop_index indexes to be dropped
@param[in] n_drop_index size of drop_index
@param[out] add_fk foreign constraint added
@param[out] n_add_fk number of foreign constraints
added
@param[in] trx user transaction
@param[in] s_cols list of stored column information
@retval true if successful
@retval false on error (will call my_error()) */
static MY_ATTRIBUTE((nonnull(1,2,3,7,8), warn_unused_result))
bool
innobase_get_foreign_key_info(
Alter_inplace_info*
ha_alter_info,
const TABLE_SHARE*
table_share,
dict_table_t* table,
const char** col_names,
dict_index_t** drop_index,
ulint n_drop_index,
dict_foreign_t**add_fk,
ulint* n_add_fk,
const trx_t* trx,
dict_s_col_list*s_cols)
{
Key* key;
Foreign_key* fk_key;
dict_table_t* referenced_table = NULL;
char* referenced_table_name = NULL;
ulint num_fk = 0;
Alter_info* alter_info = ha_alter_info->alter_info;
DBUG_ENTER("innobase_get_foreign_key_info");
*n_add_fk = 0;
List_iterator<Key> key_iterator(alter_info->key_list);
while ((key=key_iterator++)) {
if (key->type != Key::FOREIGN_KEY) {
continue;
}
const char* column_names[MAX_NUM_FK_COLUMNS];
dict_index_t* index = NULL;
const char* referenced_column_names[MAX_NUM_FK_COLUMNS];
dict_index_t* referenced_index = NULL;
ulint num_col = 0;
ulint referenced_num_col = 0;
bool correct_option;
char* db_namep = NULL;
char* tbl_namep = NULL;
ulint db_name_len = 0;
ulint tbl_name_len = 0;
char db_name[MAX_DATABASE_NAME_LEN];
char tbl_name[MAX_TABLE_NAME_LEN];
fk_key = static_cast<Foreign_key*>(key);
if (fk_key->columns.elements > 0) {
ulint i = 0;
Key_part_spec* column;
List_iterator<Key_part_spec> key_part_iterator(
fk_key->columns);
/* Get all the foreign key column info for the
current table */
while ((column = key_part_iterator++)) {
column_names[i] = column->field_name.str;
ut_ad(i < MAX_NUM_FK_COLUMNS);
i++;
}
index = innobase_find_fk_index(
ha_alter_info,
table, col_names,
drop_index, n_drop_index,
column_names, i);
/* MySQL would add a index in the creation
list if no such index for foreign table,
so we have to use DBUG_EXECUTE_IF to simulate
the scenario */
DBUG_EXECUTE_IF("innodb_test_no_foreign_idx",
index = NULL;);
/* Check whether there exist such
index in the the index create clause */
if (!index && !innobase_find_equiv_index(
column_names, static_cast<uint>(i),
ha_alter_info->key_info_buffer,
ha_alter_info->index_add_buffer,
ha_alter_info->index_add_count)) {
my_error(
ER_FK_NO_INDEX_CHILD,
MYF(0),
fk_key->name.str
? fk_key->name.str : "",
table_share->table_name.str);
goto err_exit;
}
num_col = i;
}
add_fk[num_fk] = dict_mem_foreign_create();
#ifndef _WIN32
if (fk_key->ref_db.str) {
tablename_to_filename(fk_key->ref_db.str, db_name,
MAX_DATABASE_NAME_LEN);
db_namep = db_name;
db_name_len = strlen(db_name);
}
if (fk_key->ref_table.str) {
tablename_to_filename(fk_key->ref_table.str, tbl_name,
MAX_TABLE_NAME_LEN);
tbl_namep = tbl_name;
tbl_name_len = strlen(tbl_name);
}
#else
ut_ad(fk_key->ref_table.str);
tablename_to_filename(fk_key->ref_table.str, tbl_name,
MAX_TABLE_NAME_LEN);
innobase_casedn_str(tbl_name);
tbl_name_len = strlen(tbl_name);
tbl_namep = &tbl_name[0];
if (fk_key->ref_db.str != NULL) {
tablename_to_filename(fk_key->ref_db.str, db_name,
MAX_DATABASE_NAME_LEN);
innobase_casedn_str(db_name);
db_name_len = strlen(db_name);
db_namep = &db_name[0];
}
#endif
mutex_enter(&dict_sys->mutex);
referenced_table_name = dict_get_referenced_table(
table->name.m_name,
db_namep,
db_name_len,
tbl_namep,
tbl_name_len,
&referenced_table,
add_fk[num_fk]->heap);
/* Test the case when referenced_table failed to
open, if trx->check_foreigns is not set, we should
still be able to add the foreign key */
DBUG_EXECUTE_IF("innodb_test_open_ref_fail",
referenced_table = NULL;);
if (!referenced_table && trx->check_foreigns) {
mutex_exit(&dict_sys->mutex);
my_error(ER_FK_CANNOT_OPEN_PARENT,
MYF(0), tbl_namep);
goto err_exit;
}
if (fk_key->ref_columns.elements > 0) {
ulint i = 0;
Key_part_spec* column;
List_iterator<Key_part_spec> key_part_iterator(
fk_key->ref_columns);
while ((column = key_part_iterator++)) {
referenced_column_names[i] =
column->field_name.str;
ut_ad(i < MAX_NUM_FK_COLUMNS);
i++;
}
if (referenced_table) {
referenced_index =
dict_foreign_find_index(
referenced_table, 0,
referenced_column_names,
i, index,
TRUE, FALSE,
NULL, NULL, NULL);
DBUG_EXECUTE_IF(
"innodb_test_no_reference_idx",
referenced_index = NULL;);
/* Check whether there exist such
index in the the index create clause */
if (!referenced_index) {
mutex_exit(&dict_sys->mutex);
my_error(ER_FK_NO_INDEX_PARENT, MYF(0),
fk_key->name.str
? fk_key->name.str : "",
tbl_namep);
goto err_exit;
}
} else {
ut_a(!trx->check_foreigns);
}
referenced_num_col = i;
} else {
/* Not possible to add a foreign key without a
referenced column */
mutex_exit(&dict_sys->mutex);
my_error(ER_CANNOT_ADD_FOREIGN, MYF(0), tbl_namep);
goto err_exit;
}
if (!innobase_init_foreign(
add_fk[num_fk], fk_key->name.str,
table, index, column_names,
num_col, referenced_table_name,
referenced_table, referenced_index,
referenced_column_names, referenced_num_col)) {
mutex_exit(&dict_sys->mutex);
my_error(
ER_DUP_CONSTRAINT_NAME,
MYF(0),
"FOREIGN KEY", add_fk[num_fk]->id);
goto err_exit;
}
mutex_exit(&dict_sys->mutex);
correct_option = innobase_set_foreign_key_option(
add_fk[num_fk], fk_key);
DBUG_EXECUTE_IF("innodb_test_wrong_fk_option",
correct_option = false;);
if (!correct_option) {
my_error(ER_FK_INCORRECT_OPTION,
MYF(0),
table_share->table_name.str,
add_fk[num_fk]->id);
goto err_exit;
}
if (innobase_check_fk_stored(
add_fk[num_fk], table, s_cols)) {
my_printf_error(
HA_ERR_UNSUPPORTED,
"Cannot add foreign key on the base column "
"of stored column", MYF(0));
goto err_exit;
}
num_fk++;
}
*n_add_fk = num_fk;
DBUG_RETURN(true);
err_exit:
for (ulint i = 0; i <= num_fk; i++) {
if (add_fk[i]) {
dict_foreign_free(add_fk[i]);
}
}
DBUG_RETURN(false);
}
/*************************************************************//**
Copies an InnoDB column to a MySQL field. This function is
adapted from row_sel_field_store_in_mysql_format(). */
static
void
innobase_col_to_mysql(
/*==================*/
const dict_col_t* col, /*!< in: InnoDB column */
const uchar* data, /*!< in: InnoDB column data */
ulint len, /*!< in: length of data, in bytes */
Field* field) /*!< in/out: MySQL field */
{
uchar* ptr;
uchar* dest = field->ptr;
ulint flen = field->pack_length();
switch (col->mtype) {
case DATA_INT:
ut_ad(len == flen);
/* Convert integer data from Innobase to little-endian
format, sign bit restored to normal */
for (ptr = dest + len; ptr != dest; ) {
*--ptr = *data++;
}
if (!(col->prtype & DATA_UNSIGNED)) {
((byte*) dest)[len - 1] ^= 0x80;
}
break;
case DATA_VARCHAR:
case DATA_VARMYSQL:
case DATA_BINARY:
field->reset();
if (field->type() == MYSQL_TYPE_VARCHAR) {
/* This is a >= 5.0.3 type true VARCHAR. Store the
length of the data to the first byte or the first
two bytes of dest. */
dest = row_mysql_store_true_var_len(
dest, len, flen - field->key_length());
}
/* Copy the actual data */
memcpy(dest, data, len);
break;
case DATA_GEOMETRY:
case DATA_BLOB:
/* Skip MySQL BLOBs when reporting an erroneous row
during index creation or table rebuild. */
field->set_null();
break;
#ifdef UNIV_DEBUG
case DATA_MYSQL:
ut_ad(flen >= len);
ut_ad(DATA_MBMAXLEN(col->mbminmaxlen)
>= DATA_MBMINLEN(col->mbminmaxlen));
memcpy(dest, data, len);
break;
default:
case DATA_SYS_CHILD:
case DATA_SYS:
/* These column types should never be shipped to MySQL. */
ut_ad(0);
case DATA_FLOAT:
case DATA_DOUBLE:
case DATA_DECIMAL:
/* Above are the valid column types for MySQL data. */
ut_ad(flen == len);
/* fall through */
case DATA_FIXBINARY:
case DATA_CHAR:
/* We may have flen > len when there is a shorter
prefix on the CHAR and BINARY column. */
ut_ad(flen >= len);
#else /* UNIV_DEBUG */
default:
#endif /* UNIV_DEBUG */
memcpy(dest, data, len);
}
}
/*************************************************************//**
Copies an InnoDB record to table->record[0]. */
void
innobase_rec_to_mysql(
/*==================*/
struct TABLE* table, /*!< in/out: MySQL table */
const rec_t* rec, /*!< in: record */
const dict_index_t* index, /*!< in: index */
const ulint* offsets)/*!< in: rec_get_offsets(
rec, index, ...) */
{
uint n_fields = table->s->fields;
ut_ad(n_fields == dict_table_get_n_user_cols(index->table)
- !!(DICT_TF2_FLAG_IS_SET(index->table,
DICT_TF2_FTS_HAS_DOC_ID)));
for (uint i = 0; i < n_fields; i++) {
Field* field = table->field[i];
ulint ipos;
ulint ilen;
const uchar* ifield;
ulint prefix_col;
field->reset();
ipos = dict_index_get_nth_col_or_prefix_pos(
index, i, true, false, &prefix_col);
if (ipos == ULINT_UNDEFINED
|| rec_offs_nth_extern(offsets, ipos)) {
null_field:
field->set_null();
continue;
}
ifield = rec_get_nth_field(rec, offsets, ipos, &ilen);
/* Assign the NULL flag */
if (ilen == UNIV_SQL_NULL) {
ut_ad(field->real_maybe_null());
goto null_field;
}
field->set_notnull();
innobase_col_to_mysql(
dict_field_get_col(
dict_index_get_nth_field(index, ipos)),
ifield, ilen, field);
}
}
/*************************************************************//**
Copies an InnoDB index entry to table->record[0].
This is used in preparation for print_keydup_error() from
inline add index */
void
innobase_fields_to_mysql(
/*=====================*/
struct TABLE* table, /*!< in/out: MySQL table */
const dict_index_t* index, /*!< in: InnoDB index */
const dfield_t* fields) /*!< in: InnoDB index fields */
{
uint n_fields = table->s->fields;
ulint num_v = 0;
ut_ad(n_fields == dict_table_get_n_user_cols(index->table)
+ dict_table_get_n_v_cols(index->table)
- !!(DICT_TF2_FLAG_IS_SET(index->table,
DICT_TF2_FTS_HAS_DOC_ID)));
for (uint i = 0; i < n_fields; i++) {
Field* field = table->field[i];
ulint ipos;
ulint col_n;
ulint prefix_col;
field->reset();
if (innobase_is_v_fld(field)) {
col_n = num_v;
num_v++;
} else {
col_n = i - num_v;
}
ipos = dict_index_get_nth_col_or_prefix_pos(
index, col_n, true, innobase_is_v_fld(field),
&prefix_col);
if (ipos == ULINT_UNDEFINED
|| dfield_is_ext(&fields[ipos])
|| dfield_is_null(&fields[ipos])) {
field->set_null();
} else {
field->set_notnull();
const dfield_t* df = &fields[ipos];
innobase_col_to_mysql(
dict_field_get_col(
dict_index_get_nth_field(index, ipos)),
static_cast<const uchar*>(dfield_get_data(df)),
dfield_get_len(df), field);
}
}
}
/*************************************************************//**
Copies an InnoDB row to table->record[0].
This is used in preparation for print_keydup_error() from
row_log_table_apply() */
void
innobase_row_to_mysql(
/*==================*/
struct TABLE* table, /*!< in/out: MySQL table */
const dict_table_t* itab, /*!< in: InnoDB table */
const dtuple_t* row) /*!< in: InnoDB row */
{
uint n_fields = table->s->fields;
ulint num_v = 0;
/* The InnoDB row may contain an extra FTS_DOC_ID column at the end. */
ut_ad(row->n_fields == dict_table_get_n_cols(itab));
ut_ad(n_fields == row->n_fields - DATA_N_SYS_COLS
+ dict_table_get_n_v_cols(itab)
- !!(DICT_TF2_FLAG_IS_SET(itab, DICT_TF2_FTS_HAS_DOC_ID)));
for (uint i = 0; i < n_fields; i++) {
Field* field = table->field[i];
field->reset();
if (innobase_is_v_fld(field)) {
/* Virtual column are not stored in InnoDB table, so
skip it */
num_v++;
continue;
}
const dfield_t* df = dtuple_get_nth_field(row, i - num_v);
if (dfield_is_ext(df) || dfield_is_null(df)) {
field->set_null();
} else {
field->set_notnull();
innobase_col_to_mysql(
dict_table_get_nth_col(itab, i - num_v),
static_cast<const uchar*>(dfield_get_data(df)),
dfield_get_len(df), field);
}
}
if (table->vfield) {
my_bitmap_map* old_vcol_set = tmp_use_all_columns(table, table->vcol_set);
table->update_virtual_fields(table->file, VCOL_UPDATE_FOR_READ);
tmp_restore_column_map(table->vcol_set, old_vcol_set);
}
}
/*************************************************************//**
Resets table->record[0]. */
void
innobase_rec_reset(
/*===============*/
TABLE* table) /*!< in/out: MySQL table */
{
uint n_fields = table->s->fields;
uint i;
for (i = 0; i < n_fields; i++) {
table->field[i]->set_default();
}
}
/*******************************************************************//**
This function checks that index keys are sensible.
@return 0 or error number */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
int
innobase_check_index_keys(
/*======================*/
const Alter_inplace_info* info,
/*!< in: indexes to be created or dropped */
const dict_table_t* innodb_table)
/*!< in: Existing indexes */
{
for (uint key_num = 0; key_num < info->index_add_count;
key_num++) {
const KEY& key = info->key_info_buffer[
info->index_add_buffer[key_num]];
/* Check that the same index name does not appear
twice in indexes to be created. */
for (ulint i = 0; i < key_num; i++) {
const KEY& key2 = info->key_info_buffer[
info->index_add_buffer[i]];
if (0 == strcmp(key.name, key2.name)) {
my_error(ER_WRONG_NAME_FOR_INDEX, MYF(0),
key.name);
return(ER_WRONG_NAME_FOR_INDEX);
}
}
/* Check that the same index name does not already exist. */
const dict_index_t* index;
for (index = dict_table_get_first_index(innodb_table);
index; index = dict_table_get_next_index(index)) {
if (index->is_committed()
&& !strcmp(key.name, index->name)) {
break;
}
}
/* Now we are in a situation where we have "ADD INDEX x"
and an index by the same name already exists. We have 4
possible cases:
1. No further clauses for an index x are given. Should reject
the operation.
2. "DROP INDEX x" is given. Should allow the operation.
3. "RENAME INDEX x TO y" is given. Should allow the operation.
4. "DROP INDEX x, RENAME INDEX x TO y" is given. Should allow
the operation, since no name clash occurs. In this particular
case MySQL cancels the operation without calling InnoDB
methods. */
if (index) {
/* If a key by the same name is being created and
dropped, the name clash is OK. */
for (uint i = 0; i < info->index_drop_count;
i++) {
const KEY* drop_key
= info->index_drop_buffer[i];
if (0 == strcmp(key.name, drop_key->name)) {
goto name_ok;
}
}
#ifdef MYSQL_RENAME_INDEX
/* If a key by the same name is being created and
renamed, the name clash is OK. E.g.
ALTER TABLE t ADD INDEX i (col), RENAME INDEX i TO x
where the index "i" exists prior to the ALTER command.
In this case we:
1. rename the existing index from "i" to "x"
2. add the new index "i" */
for (uint i = 0; i < info->index_rename_count; i++) {
const KEY_PAIR* pair
= &info->index_rename_buffer[i];
if (0 == strcmp(key.name, pair->old_key->name)) {
goto name_ok;
}
}
#endif /* MYSQL_RENAME_INDEX */
my_error(ER_WRONG_NAME_FOR_INDEX, MYF(0), key.name);
return(ER_WRONG_NAME_FOR_INDEX);
}
name_ok:
for (ulint i = 0; i < key.user_defined_key_parts; i++) {
const KEY_PART_INFO& key_part1
= key.key_part[i];
const Field* field
= key_part1.field;
ibool is_unsigned;
switch (get_innobase_type_from_mysql_type(
&is_unsigned, field)) {
default:
break;
case DATA_INT:
case DATA_FLOAT:
case DATA_DOUBLE:
case DATA_DECIMAL:
/* Check that MySQL does not try to
create a column prefix index field on
an inappropriate data type. */
if (field->type() == MYSQL_TYPE_VARCHAR) {
if (key_part1.length
>= field->pack_length()
- ((Field_varstring*) field)
->length_bytes) {
break;
}
} else {
if (key_part1.length
>= field->pack_length()) {
break;
}
}
my_error(ER_WRONG_KEY_COLUMN, MYF(0), "InnoDB",
field->field_name);
return(ER_WRONG_KEY_COLUMN);
}
/* Check that the same column does not appear
twice in the index. */
for (ulint j = 0; j < i; j++) {
const KEY_PART_INFO& key_part2
= key.key_part[j];
if (key_part1.fieldnr != key_part2.fieldnr) {
continue;
}
my_error(ER_WRONG_KEY_COLUMN, MYF(0), "InnoDB",
field->field_name);
return(ER_WRONG_KEY_COLUMN);
}
}
}
return(0);
}
/** Create index field definition for key part
@param[in] altered_table MySQL table that is being altered,
or NULL if a new clustered index
is not being created
@param[in] key_part MySQL key definition
@param[in,out] index_field index field
@param[in] new_clustered new cluster */
static
void
innobase_create_index_field_def(
const TABLE* altered_table,
const KEY_PART_INFO* key_part,
index_field_t* index_field,
bool new_clustered)
{
const Field* field;
ibool is_unsigned;
ulint col_type;
ulint num_v = 0;
DBUG_ENTER("innobase_create_index_field_def");
ut_ad(key_part);
ut_ad(index_field);
field = new_clustered
? altered_table->field[key_part->fieldnr]
: key_part->field;
ut_a(field);
for (ulint i = 0; i < key_part->fieldnr; i++) {
if (innobase_is_v_fld(altered_table->field[i])) {
num_v++;
}
}
col_type = get_innobase_type_from_mysql_type(
&is_unsigned, field);
if (innobase_is_v_fld(field)) {
index_field->is_v_col = true;
index_field->col_no = num_v;
} else {
index_field->is_v_col = false;
index_field->col_no = key_part->fieldnr - num_v;
}
if (DATA_LARGE_MTYPE(col_type)
|| (key_part->length < field->pack_length()
&& field->type() != MYSQL_TYPE_VARCHAR)
|| (field->type() == MYSQL_TYPE_VARCHAR
&& key_part->length < field->pack_length()
- ((Field_varstring*) field)->length_bytes)) {
index_field->prefix_len = key_part->length;
} else {
index_field->prefix_len = 0;
}
DBUG_VOID_RETURN;
}
/** Create index definition for key
@param[in] altered_table MySQL table that is being altered
@param[in] keys key definitions
@param[in] key_number MySQL key number
@param[in] new_clustered true if generating a new clustered
index on the table
@param[in] key_clustered true if this is the new clustered index
@param[out] index index definition
@param[in] heap heap where memory is allocated */
static MY_ATTRIBUTE((nonnull))
void
innobase_create_index_def(
const TABLE* altered_table,
const KEY* keys,
ulint key_number,
bool new_clustered,
bool key_clustered,
index_def_t* index,
mem_heap_t* heap)
{
const KEY* key = &keys[key_number];
ulint i;
ulint n_fields = key->user_defined_key_parts;
DBUG_ENTER("innobase_create_index_def");
DBUG_ASSERT(!key_clustered || new_clustered);
index->fields = static_cast<index_field_t*>(
mem_heap_alloc(heap, n_fields * sizeof *index->fields));
memset(index->fields, 0, n_fields * sizeof *index->fields);
index->parser = NULL;
index->key_number = key_number;
index->n_fields = n_fields;
index->name = mem_heap_strdup(heap, key->name);
index->rebuild = new_clustered;
if (key_clustered) {
DBUG_ASSERT(!(key->flags & (HA_FULLTEXT | HA_SPATIAL)));
DBUG_ASSERT(key->flags & HA_NOSAME);
index->ind_type = DICT_CLUSTERED | DICT_UNIQUE;
} else if (key->flags & HA_FULLTEXT) {
DBUG_ASSERT(!(key->flags & (HA_SPATIAL | HA_NOSAME)));
DBUG_ASSERT(!(key->flags & HA_KEYFLAG_MASK
& ~(HA_FULLTEXT
| HA_PACK_KEY
| HA_BINARY_PACK_KEY)));
index->ind_type = DICT_FTS;
/* Note: key->parser is only parser name,
we need to get parser from altered_table instead */
if (key->flags & HA_USES_PARSER) {
for (ulint j = 0; j < altered_table->s->keys; j++) {
if (ut_strcmp(altered_table->key_info[j].name,
key->name) == 0) {
ut_ad(altered_table->key_info[j].flags
& HA_USES_PARSER);
plugin_ref parser =
altered_table->key_info[j].parser;
index->parser =
static_cast<st_mysql_ftparser*>(
plugin_decl(parser)->info);
break;
}
}
DBUG_EXECUTE_IF("fts_instrument_use_default_parser",
index->parser = &fts_default_parser;);
ut_ad(index->parser);
}
} else if (key->flags & HA_SPATIAL) {
DBUG_ASSERT(!(key->flags & HA_NOSAME));
index->ind_type = DICT_SPATIAL;
ut_ad(n_fields == 1);
ulint num_v = 0;
/* Need to count the virtual fields before this spatial
indexed field */
for (ulint i = 0; i < key->key_part->fieldnr; i++) {
if (innobase_is_v_fld(altered_table->field[i])) {
num_v++;
}
}
index->fields[0].col_no = key->key_part[0].fieldnr - num_v;
index->fields[0].prefix_len = 0;
index->fields[0].is_v_col = false;
if (innobase_is_v_fld(key->key_part[0].field)) {
/* Currently, the spatial index cannot be created
on virtual columns. It is blocked in server
layer */
ut_ad(0);
index->fields[0].is_v_col = true;
} else {
index->fields[0].is_v_col = false;
}
} else {
index->ind_type = (key->flags & HA_NOSAME) ? DICT_UNIQUE : 0;
}
if (!(key->flags & HA_SPATIAL)) {
for (i = 0; i < n_fields; i++) {
innobase_create_index_field_def(
altered_table, &key->key_part[i],
&index->fields[i], new_clustered);
if (index->fields[i].is_v_col) {
index->ind_type |= DICT_VIRTUAL;
}
}
}
DBUG_VOID_RETURN;
}
/*******************************************************************//**
Check whether the table has the FTS_DOC_ID column
@return whether there exists an FTS_DOC_ID column */
static
bool
innobase_fts_check_doc_id_col(
/*==========================*/
const dict_table_t* table, /*!< in: InnoDB table with
fulltext index */
const TABLE* altered_table,
/*!< in: MySQL table with
fulltext index */
ulint* fts_doc_col_no,
/*!< out: The column number for
Doc ID, or ULINT_UNDEFINED
if it is of wrong type */
ulint* num_v) /*!< out: number of virtual column */
{
*fts_doc_col_no = ULINT_UNDEFINED;
const uint n_cols = altered_table->s->fields;
ulint i;
*num_v = 0;
for (i = 0; i < n_cols; i++) {
const Field* field = altered_table->field[i];
if (innobase_is_v_fld(field)) {
(*num_v)++;
}
if (my_strcasecmp(system_charset_info,
field->field_name, FTS_DOC_ID_COL_NAME)) {
continue;
}
if (strcmp(field->field_name, FTS_DOC_ID_COL_NAME)) {
my_error(ER_WRONG_COLUMN_NAME, MYF(0),
field->field_name);
} else if (field->type() != MYSQL_TYPE_LONGLONG
|| field->pack_length() != 8
|| field->real_maybe_null()
|| !(field->flags & UNSIGNED_FLAG)
|| innobase_is_v_fld(field)) {
my_error(ER_INNODB_FT_WRONG_DOCID_COLUMN, MYF(0),
field->field_name);
} else {
*fts_doc_col_no = i - *num_v;
}
return(true);
}
if (!table) {
return(false);
}
/* Not to count the virtual columns */
i -= *num_v;
for (; i + DATA_N_SYS_COLS < (uint) table->n_cols; i++) {
const char* name = dict_table_get_col_name(table, i);
if (strcmp(name, FTS_DOC_ID_COL_NAME) == 0) {
#ifdef UNIV_DEBUG
const dict_col_t* col;
col = dict_table_get_nth_col(table, i);
/* Because the FTS_DOC_ID does not exist in
the MySQL data dictionary, this must be the
internally created FTS_DOC_ID column. */
ut_ad(col->mtype == DATA_INT);
ut_ad(col->len == 8);
ut_ad(col->prtype & DATA_NOT_NULL);
ut_ad(col->prtype & DATA_UNSIGNED);
#endif /* UNIV_DEBUG */
*fts_doc_col_no = i;
return(true);
}
}
return(false);
}
/*******************************************************************//**
Check whether the table has a unique index with FTS_DOC_ID_INDEX_NAME
on the Doc ID column.
@return the status of the FTS_DOC_ID index */
enum fts_doc_id_index_enum
innobase_fts_check_doc_id_index(
/*============================*/
const dict_table_t* table, /*!< in: table definition */
const TABLE* altered_table, /*!< in: MySQL table
that is being altered */
ulint* fts_doc_col_no) /*!< out: The column number for
Doc ID, or ULINT_UNDEFINED
if it is being created in
ha_alter_info */
{
const dict_index_t* index;
const dict_field_t* field;
if (altered_table) {
/* Check if a unique index with the name of
FTS_DOC_ID_INDEX_NAME is being created. */
for (uint i = 0; i < altered_table->s->keys; i++) {
const KEY& key = altered_table->key_info[i];
if (innobase_strcasecmp(
key.name, FTS_DOC_ID_INDEX_NAME)) {
continue;
}
if ((key.flags & HA_NOSAME)
&& key.user_defined_key_parts == 1
&& !strcmp(key.name, FTS_DOC_ID_INDEX_NAME)
&& !strcmp(key.key_part[0].field->field_name,
FTS_DOC_ID_COL_NAME)) {
if (fts_doc_col_no) {
*fts_doc_col_no = ULINT_UNDEFINED;
}
return(FTS_EXIST_DOC_ID_INDEX);
} else {
return(FTS_INCORRECT_DOC_ID_INDEX);
}
}
}
if (!table) {
return(FTS_NOT_EXIST_DOC_ID_INDEX);
}
for (index = dict_table_get_first_index(table);
index; index = dict_table_get_next_index(index)) {
/* Check if there exists a unique index with the name of
FTS_DOC_ID_INDEX_NAME */
if (innobase_strcasecmp(index->name, FTS_DOC_ID_INDEX_NAME)) {
continue;
}
if (!dict_index_is_unique(index)
|| dict_index_get_n_unique(index) > 1
|| strcmp(index->name, FTS_DOC_ID_INDEX_NAME)) {
return(FTS_INCORRECT_DOC_ID_INDEX);
}
/* Check whether the index has FTS_DOC_ID as its
first column */
field = dict_index_get_nth_field(index, 0);
/* The column would be of a BIGINT data type */
if (strcmp(field->name, FTS_DOC_ID_COL_NAME) == 0
&& field->col->mtype == DATA_INT
&& field->col->len == 8
&& field->col->prtype & DATA_NOT_NULL
&& !dict_col_is_virtual(field->col)) {
if (fts_doc_col_no) {
*fts_doc_col_no = dict_col_get_no(field->col);
}
return(FTS_EXIST_DOC_ID_INDEX);
} else {
return(FTS_INCORRECT_DOC_ID_INDEX);
}
}
/* Not found */
return(FTS_NOT_EXIST_DOC_ID_INDEX);
}
/*******************************************************************//**
Check whether the table has a unique index with FTS_DOC_ID_INDEX_NAME
on the Doc ID column in MySQL create index definition.
@return FTS_EXIST_DOC_ID_INDEX if there exists the FTS_DOC_ID index,
FTS_INCORRECT_DOC_ID_INDEX if the FTS_DOC_ID index is of wrong format */
enum fts_doc_id_index_enum
innobase_fts_check_doc_id_index_in_def(
/*===================================*/
ulint n_key, /*!< in: Number of keys */
const KEY* key_info) /*!< in: Key definition */
{
/* Check whether there is a "FTS_DOC_ID_INDEX" in the to be built index
list */
for (ulint j = 0; j < n_key; j++) {
const KEY* key = &key_info[j];
if (innobase_strcasecmp(key->name, FTS_DOC_ID_INDEX_NAME)) {
continue;
}
/* Do a check on FTS DOC ID_INDEX, it must be unique,
named as "FTS_DOC_ID_INDEX" and on column "FTS_DOC_ID" */
if (!(key->flags & HA_NOSAME)
|| key->user_defined_key_parts != 1
|| strcmp(key->name, FTS_DOC_ID_INDEX_NAME)
|| strcmp(key->key_part[0].field->field_name,
FTS_DOC_ID_COL_NAME)) {
return(FTS_INCORRECT_DOC_ID_INDEX);
}
return(FTS_EXIST_DOC_ID_INDEX);
}
return(FTS_NOT_EXIST_DOC_ID_INDEX);
}
/*******************************************************************//**
Create an index table where indexes are ordered as follows:
IF a new primary key is defined for the table THEN
1) New primary key
2) The remaining keys in key_info
ELSE
1) All new indexes in the order they arrive from MySQL
ENDIF
@return key definitions */
static MY_ATTRIBUTE((nonnull, warn_unused_result, malloc))
index_def_t*
innobase_create_key_defs(
/*=====================*/
mem_heap_t* heap,
/*!< in/out: memory heap where space for key
definitions are allocated */
const Alter_inplace_info* ha_alter_info,
/*!< in: alter operation */
const TABLE* altered_table,
/*!< in: MySQL table that is being altered */
ulint& n_add,
/*!< in/out: number of indexes to be created */
ulint& n_fts_add,
/*!< out: number of FTS indexes to be created */
bool got_default_clust,
/*!< in: whether the table lacks a primary key */
ulint& fts_doc_id_col,
/*!< in: The column number for Doc ID */
bool& add_fts_doc_id,
/*!< in: whether we need to add new DOC ID
column for FTS index */
bool& add_fts_doc_idx,
/*!< in: whether we need to add new DOC ID
index for FTS index */
const TABLE* table)
/*!< in: MySQL table that is being altered */
{
index_def_t* indexdef;
index_def_t* indexdefs;
bool new_primary;
const uint*const add
= ha_alter_info->index_add_buffer;
const KEY*const key_info
= ha_alter_info->key_info_buffer;
DBUG_ENTER("innobase_create_key_defs");
DBUG_ASSERT(!add_fts_doc_id || add_fts_doc_idx);
DBUG_ASSERT(ha_alter_info->index_add_count == n_add);
/* If there is a primary key, it is always the first index
defined for the innodb_table. */
new_primary = n_add > 0
&& !my_strcasecmp(system_charset_info,
key_info[*add].name, "PRIMARY");
n_fts_add = 0;
/* If there is a UNIQUE INDEX consisting entirely of NOT NULL
columns and if the index does not contain column prefix(es)
(only prefix/part of the column is indexed), MySQL will treat the
index as a PRIMARY KEY unless the table already has one. */
ut_ad(altered_table->s->primary_key == 0
|| altered_table->s->primary_key == MAX_KEY);
if (got_default_clust && !new_primary) {
new_primary = (altered_table->s->primary_key != MAX_KEY);
}
const bool rebuild = new_primary || add_fts_doc_id
|| innobase_need_rebuild(ha_alter_info, table);
/* Reserve one more space if new_primary is true, and we might
need to add the FTS_DOC_ID_INDEX */
indexdef = indexdefs = static_cast<index_def_t*>(
mem_heap_alloc(
heap, sizeof *indexdef
* (ha_alter_info->key_count
+ rebuild
+ got_default_clust)));
if (rebuild) {
ulint primary_key_number;
if (new_primary) {
if (n_add == 0) {
DBUG_ASSERT(got_default_clust);
DBUG_ASSERT(altered_table->s->primary_key
== 0);
primary_key_number = 0;
} else {
primary_key_number = *add;
}
} else if (got_default_clust) {
/* Create the GEN_CLUST_INDEX */
index_def_t* index = indexdef++;
index->fields = NULL;
index->n_fields = 0;
index->ind_type = DICT_CLUSTERED;
index->name = innobase_index_reserve_name;
index->rebuild = true;
index->key_number = ~0;
primary_key_number = ULINT_UNDEFINED;
goto created_clustered;
} else {
primary_key_number = 0;
}
/* Create the PRIMARY key index definition */
innobase_create_index_def(
altered_table, key_info, primary_key_number,
true, true, indexdef++, heap);
created_clustered:
n_add = 1;
for (ulint i = 0; i < ha_alter_info->key_count; i++) {
if (i == primary_key_number) {
continue;
}
/* Copy the index definitions. */
innobase_create_index_def(
altered_table, key_info, i, true,
false, indexdef, heap);
if (indexdef->ind_type & DICT_FTS) {
n_fts_add++;
}
indexdef++;
n_add++;
}
if (n_fts_add > 0) {
ulint num_v = 0;
if (!add_fts_doc_id
&& !innobase_fts_check_doc_id_col(
NULL, altered_table,
&fts_doc_id_col, &num_v)) {
fts_doc_id_col = altered_table->s->fields - num_v;
add_fts_doc_id = true;
}
if (!add_fts_doc_idx) {
fts_doc_id_index_enum ret;
ulint doc_col_no;
ret = innobase_fts_check_doc_id_index(
NULL, altered_table, &doc_col_no);
/* This should have been checked before */
ut_ad(ret != FTS_INCORRECT_DOC_ID_INDEX);
if (ret == FTS_NOT_EXIST_DOC_ID_INDEX) {
add_fts_doc_idx = true;
} else {
ut_ad(ret == FTS_EXIST_DOC_ID_INDEX);
ut_ad(doc_col_no == ULINT_UNDEFINED
|| doc_col_no == fts_doc_id_col);
}
}
}
} else {
/* Create definitions for added secondary indexes. */
for (ulint i = 0; i < n_add; i++) {
innobase_create_index_def(
altered_table, key_info, add[i],
false, false, indexdef, heap);
if (indexdef->ind_type & DICT_FTS) {
n_fts_add++;
}
indexdef++;
}
}
DBUG_ASSERT(indexdefs + n_add == indexdef);
if (add_fts_doc_idx) {
index_def_t* index = indexdef++;
index->fields = static_cast<index_field_t*>(
mem_heap_alloc(heap, sizeof *index->fields));
memset(index->fields, 0, sizeof *index->fields);
index->n_fields = 1;
index->fields->col_no = fts_doc_id_col;
index->fields->prefix_len = 0;
index->fields->is_v_col = false;
index->ind_type = DICT_UNIQUE;
ut_ad(!rebuild
|| !add_fts_doc_id
|| fts_doc_id_col <= altered_table->s->fields);
index->name = FTS_DOC_ID_INDEX_NAME;
index->rebuild = rebuild;
/* TODO: assign a real MySQL key number for this */
index->key_number = ULINT_UNDEFINED;
n_add++;
}
DBUG_ASSERT(indexdef > indexdefs);
DBUG_ASSERT((ulint) (indexdef - indexdefs)
<= ha_alter_info->key_count
+ add_fts_doc_idx + got_default_clust);
DBUG_ASSERT(ha_alter_info->index_add_count <= n_add);
DBUG_RETURN(indexdefs);
}
/*******************************************************************//**
Check each index column size, make sure they do not exceed the max limit
@return true if index column size exceeds limit */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_check_column_length(
/*=========================*/
ulint max_col_len, /*!< in: maximum column length */
const KEY* key_info) /*!< in: Indexes to be created */
{
for (ulint key_part = 0; key_part < key_info->user_defined_key_parts; key_part++) {
if (key_info->key_part[key_part].length > max_col_len) {
return(true);
}
}
return(false);
}
/********************************************************************//**
Drop any indexes that we were not able to free previously due to
open table handles. */
static
void
online_retry_drop_indexes_low(
/*==========================*/
dict_table_t* table, /*!< in/out: table */
trx_t* trx) /*!< in/out: transaction */
{
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
ut_ad(trx_get_dict_operation(trx) == TRX_DICT_OP_INDEX);
/* We can have table->n_ref_count > 1, because other threads
may have prebuilt->table pointing to the table. However, these
other threads should be between statements, waiting for the
next statement to execute, or for a meta-data lock. */
ut_ad(table->get_ref_count() >= 1);
if (table->drop_aborted) {
row_merge_drop_indexes(trx, table, TRUE);
}
}
/********************************************************************//**
Drop any indexes that we were not able to free previously due to
open table handles. */
static MY_ATTRIBUTE((nonnull))
void
online_retry_drop_indexes(
/*======================*/
dict_table_t* table, /*!< in/out: table */
THD* user_thd) /*!< in/out: MySQL connection */
{
if (table->drop_aborted) {
trx_t* trx = innobase_trx_allocate(user_thd);
trx_start_for_ddl(trx, TRX_DICT_OP_INDEX);
row_mysql_lock_data_dictionary(trx);
online_retry_drop_indexes_low(table, trx);
trx_commit_for_mysql(trx);
row_mysql_unlock_data_dictionary(trx);
trx_free_for_mysql(trx);
}
ut_d(mutex_enter(&dict_sys->mutex));
ut_d(dict_table_check_for_dup_indexes(table, CHECK_ALL_COMPLETE));
ut_d(mutex_exit(&dict_sys->mutex));
ut_ad(!table->drop_aborted);
}
/********************************************************************//**
Commit a dictionary transaction and drop any indexes that we were not
able to free previously due to open table handles. */
static MY_ATTRIBUTE((nonnull))
void
online_retry_drop_indexes_with_trx(
/*===============================*/
dict_table_t* table, /*!< in/out: table */
trx_t* trx) /*!< in/out: transaction */
{
ut_ad(trx_state_eq(trx, TRX_STATE_NOT_STARTED)
|| trx_state_eq(trx, TRX_STATE_FORCED_ROLLBACK));
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
/* Now that the dictionary is being locked, check if we can
drop any incompletely created indexes that may have been left
behind in rollback_inplace_alter_table() earlier. */
if (table->drop_aborted) {
trx->table_id = 0;
trx_start_for_ddl(trx, TRX_DICT_OP_INDEX);
online_retry_drop_indexes_low(table, trx);
trx_commit_for_mysql(trx);
}
}
/** Determines if InnoDB is dropping a foreign key constraint.
@param foreign the constraint
@param drop_fk constraints being dropped
@param n_drop_fk number of constraints that are being dropped
@return whether the constraint is being dropped */
MY_ATTRIBUTE((pure, nonnull(1), warn_unused_result))
inline
bool
innobase_dropping_foreign(
const dict_foreign_t* foreign,
dict_foreign_t** drop_fk,
ulint n_drop_fk)
{
while (n_drop_fk--) {
if (*drop_fk++ == foreign) {
return(true);
}
}
return(false);
}
/** Determines if an InnoDB FOREIGN KEY constraint depends on a
column that is being dropped or modified to NOT NULL.
@param user_table InnoDB table as it is before the ALTER operation
@param col_name Name of the column being altered
@param drop_fk constraints being dropped
@param n_drop_fk number of constraints that are being dropped
@param drop true=drop column, false=set NOT NULL
@retval true Not allowed (will call my_error())
@retval false Allowed
*/
MY_ATTRIBUTE((pure, nonnull(1,4), warn_unused_result))
static
bool
innobase_check_foreigns_low(
const dict_table_t* user_table,
dict_foreign_t** drop_fk,
ulint n_drop_fk,
const char* col_name,
bool drop)
{
dict_foreign_t* foreign;
ut_ad(mutex_own(&dict_sys->mutex));
/* Check if any FOREIGN KEY constraints are defined on this
column. */
for (dict_foreign_set::const_iterator it = user_table->foreign_set.begin();
it != user_table->foreign_set.end();
++it) {
foreign = *it;
if (!drop && !(foreign->type
& (DICT_FOREIGN_ON_DELETE_SET_NULL
| DICT_FOREIGN_ON_UPDATE_SET_NULL))) {
continue;
}
if (innobase_dropping_foreign(foreign, drop_fk, n_drop_fk)) {
continue;
}
for (unsigned f = 0; f < foreign->n_fields; f++) {
if (!strcmp(foreign->foreign_col_names[f],
col_name)) {
my_error(drop
? ER_FK_COLUMN_CANNOT_DROP
: ER_FK_COLUMN_NOT_NULL, MYF(0),
col_name, foreign->id);
return(true);
}
}
}
if (!drop) {
/* SET NULL clauses on foreign key constraints of
child tables affect the child tables, not the parent table.
The column can be NOT NULL in the parent table. */
return(false);
}
/* Check if any FOREIGN KEY constraints in other tables are
referring to the column that is being dropped. */
for (dict_foreign_set::const_iterator it
= user_table->referenced_set.begin();
it != user_table->referenced_set.end();
++it) {
foreign = *it;
if (innobase_dropping_foreign(foreign, drop_fk, n_drop_fk)) {
continue;
}
for (unsigned f = 0; f < foreign->n_fields; f++) {
char display_name[FN_REFLEN];
if (strcmp(foreign->referenced_col_names[f],
col_name)) {
continue;
}
char* buf_end = innobase_convert_name(
display_name, (sizeof display_name) - 1,
foreign->foreign_table_name,
strlen(foreign->foreign_table_name),
NULL);
*buf_end = '\0';
my_error(ER_FK_COLUMN_CANNOT_DROP_CHILD,
MYF(0), col_name, foreign->id,
display_name);
return(true);
}
}
return(false);
}
/** Determines if an InnoDB FOREIGN KEY constraint depends on a
column that is being dropped or modified to NOT NULL.
@param ha_alter_info Data used during in-place alter
@param altered_table MySQL table that is being altered
@param old_table MySQL table as it is before the ALTER operation
@param user_table InnoDB table as it is before the ALTER operation
@param drop_fk constraints being dropped
@param n_drop_fk number of constraints that are being dropped
@retval true Not allowed (will call my_error())
@retval false Allowed
*/
MY_ATTRIBUTE((pure, nonnull(1,2,3,4), warn_unused_result))
static
bool
innobase_check_foreigns(
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* old_table,
const dict_table_t* user_table,
dict_foreign_t** drop_fk,
ulint n_drop_fk)
{
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
for (Field** fp = old_table->field; *fp; fp++) {
cf_it.rewind();
const Create_field* new_field;
ut_ad(!(*fp)->real_maybe_null()
== !!((*fp)->flags & NOT_NULL_FLAG));
while ((new_field = cf_it++)) {
if (new_field->field == *fp) {
break;
}
}
if (!new_field || (new_field->flags & NOT_NULL_FLAG)) {
if (innobase_check_foreigns_low(
user_table, drop_fk, n_drop_fk,
(*fp)->field_name, !new_field)) {
return(true);
}
}
}
return(false);
}
/** Convert a default value for ADD COLUMN.
@param heap Memory heap where allocated
@param dfield InnoDB data field to copy to
@param field MySQL value for the column
@param comp nonzero if in compact format */
static MY_ATTRIBUTE((nonnull))
void
innobase_build_col_map_add(
/*=======================*/
mem_heap_t* heap,
dfield_t* dfield,
const Field* field,
ulint comp)
{
if (field->is_real_null()) {
dfield_set_null(dfield);
return;
}
ulint size = field->pack_length();
byte* buf = static_cast<byte*>(mem_heap_alloc(heap, size));
const byte* mysql_data = field->ptr;
row_mysql_store_col_in_innobase_format(
dfield, buf, true, mysql_data, size, comp);
}
/** Construct the translation table for reordering, dropping or
adding columns.
@param ha_alter_info Data used during in-place alter
@param altered_table MySQL table that is being altered
@param table MySQL table as it is before the ALTER operation
@param new_table InnoDB table corresponding to MySQL altered_table
@param old_table InnoDB table corresponding to MYSQL table
@param add_cols Default values for ADD COLUMN, or NULL if no ADD COLUMN
@param heap Memory heap where allocated
@return array of integers, mapping column numbers in the table
to column numbers in altered_table */
static MY_ATTRIBUTE((nonnull(1,2,3,4,5,7), warn_unused_result))
const ulint*
innobase_build_col_map(
/*===================*/
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* table,
const dict_table_t* new_table,
const dict_table_t* old_table,
dtuple_t* add_cols,
mem_heap_t* heap)
{
DBUG_ENTER("innobase_build_col_map");
DBUG_ASSERT(altered_table != table);
DBUG_ASSERT(new_table != old_table);
DBUG_ASSERT(dict_table_get_n_cols(new_table)
+ dict_table_get_n_v_cols(new_table)
>= altered_table->s->fields + DATA_N_SYS_COLS);
DBUG_ASSERT(dict_table_get_n_cols(old_table)
+ dict_table_get_n_v_cols(old_table)
>= table->s->fields + DATA_N_SYS_COLS);
DBUG_ASSERT(!!add_cols == !!(ha_alter_info->handler_flags
& Alter_inplace_info::ADD_COLUMN));
DBUG_ASSERT(!add_cols || dtuple_get_n_fields(add_cols)
== dict_table_get_n_cols(new_table));
ulint* col_map = static_cast<ulint*>(
mem_heap_alloc(
heap, (old_table->n_cols + old_table->n_v_cols)
* sizeof *col_map));
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
uint i = 0;
uint num_v = 0;
/* Any dropped columns will map to ULINT_UNDEFINED. */
for (uint old_i = 0; old_i + DATA_N_SYS_COLS < old_table->n_cols;
old_i++) {
col_map[old_i] = ULINT_UNDEFINED;
}
for (uint old_i = 0; old_i < old_table->n_v_cols; old_i++) {
col_map[old_i + old_table->n_cols] = ULINT_UNDEFINED;
}
while (const Create_field* new_field = cf_it++) {
bool is_v = false;
if (innobase_is_v_fld(new_field)) {
is_v = true;
}
ulint num_old_v = 0;
for (uint old_i = 0; table->field[old_i]; old_i++) {
const Field* field = table->field[old_i];
if (innobase_is_v_fld(field)) {
if (is_v && new_field->field == field) {
col_map[old_table->n_cols + num_v]
= num_old_v;
num_old_v++;
goto found_col;
}
num_old_v++;
continue;
}
if (new_field->field == field) {
col_map[old_i - num_old_v] = i;
goto found_col;
}
}
ut_ad(!is_v);
innobase_build_col_map_add(
heap, dtuple_get_nth_field(add_cols, i),
altered_table->field[i + num_v],
dict_table_is_comp(new_table));
found_col:
if (is_v) {
num_v++;
} else {
i++;
}
}
DBUG_ASSERT(i == altered_table->s->fields - num_v);
i = table->s->fields - old_table->n_v_cols;
/* Add the InnoDB hidden FTS_DOC_ID column, if any. */
if (i + DATA_N_SYS_COLS < old_table->n_cols) {
/* There should be exactly one extra field,
the FTS_DOC_ID. */
DBUG_ASSERT(DICT_TF2_FLAG_IS_SET(old_table,
DICT_TF2_FTS_HAS_DOC_ID));
DBUG_ASSERT(i + DATA_N_SYS_COLS + 1 == old_table->n_cols);
DBUG_ASSERT(!strcmp(dict_table_get_col_name(
old_table, i),
FTS_DOC_ID_COL_NAME));
if (altered_table->s->fields + DATA_N_SYS_COLS
- new_table->n_v_cols
< new_table->n_cols) {
DBUG_ASSERT(DICT_TF2_FLAG_IS_SET(
new_table,
DICT_TF2_FTS_HAS_DOC_ID));
DBUG_ASSERT(altered_table->s->fields
+ DATA_N_SYS_COLS + 1
== static_cast<ulint>(
new_table->n_cols
+ new_table->n_v_cols));
col_map[i] = altered_table->s->fields
- new_table->n_v_cols;
} else {
DBUG_ASSERT(!DICT_TF2_FLAG_IS_SET(
new_table,
DICT_TF2_FTS_HAS_DOC_ID));
col_map[i] = ULINT_UNDEFINED;
}
i++;
} else {
DBUG_ASSERT(!DICT_TF2_FLAG_IS_SET(
old_table,
DICT_TF2_FTS_HAS_DOC_ID));
}
for (; i < old_table->n_cols; i++) {
col_map[i] = i + new_table->n_cols - old_table->n_cols;
}
DBUG_RETURN(col_map);
}
/** Drop newly create FTS index related auxiliary table during
FIC create index process, before fts_add_index is called
@param table table that was being rebuilt online
@param trx transaction
@return DB_SUCCESS if successful, otherwise last error code
*/
static
dberr_t
innobase_drop_fts_index_table(
/*==========================*/
dict_table_t* table,
trx_t* trx)
{
dberr_t ret_err = DB_SUCCESS;
for (dict_index_t* index = dict_table_get_first_index(table);
index != NULL;
index = dict_table_get_next_index(index)) {
if (index->type & DICT_FTS) {
dberr_t err;
err = fts_drop_index_tables(trx, index);
if (err != DB_SUCCESS) {
ret_err = err;
}
}
}
return(ret_err);
}
/** Get the new non-virtual column names if any columns were renamed
@param ha_alter_info Data used during in-place alter
@param altered_table MySQL table that is being altered
@param table MySQL table as it is before the ALTER operation
@param user_table InnoDB table as it is before the ALTER operation
@param heap Memory heap for the allocation
@return array of new column names in rebuilt_table, or NULL if not renamed */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
const char**
innobase_get_col_names(
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* table,
const dict_table_t* user_table,
mem_heap_t* heap)
{
const char** cols;
uint i;
DBUG_ENTER("innobase_get_col_names");
DBUG_ASSERT(user_table->n_t_def > table->s->fields);
DBUG_ASSERT(ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME);
cols = static_cast<const char**>(
mem_heap_zalloc(heap, user_table->n_def * sizeof *cols));
i = 0;
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
while (const Create_field* new_field = cf_it++) {
ulint num_v = 0;
DBUG_ASSERT(i < altered_table->s->fields);
if (innobase_is_v_fld(new_field)) {
continue;
}
for (uint old_i = 0; table->field[old_i]; old_i++) {
if (innobase_is_v_fld(table->field[old_i])) {
num_v++;
}
if (new_field->field == table->field[old_i]) {
cols[old_i - num_v] = new_field->field_name;
break;
}
}
i++;
}
/* Copy the internal column names. */
i = table->s->fields - user_table->n_v_def;
cols[i] = dict_table_get_col_name(user_table, i);
while (++i < user_table->n_def) {
cols[i] = cols[i - 1] + strlen(cols[i - 1]) + 1;
}
DBUG_RETURN(cols);
}
/** Check whether the column prefix is increased, decreased, or unchanged.
@param[in] new_prefix_len new prefix length
@param[in] old_prefix_len new prefix length
@retval 1 prefix is increased
@retval 0 prefix is unchanged
@retval -1 prefix is decreased */
static inline
lint
innobase_pk_col_prefix_compare(
ulint new_prefix_len,
ulint old_prefix_len)
{
ut_ad(new_prefix_len < REC_MAX_DATA_SIZE);
ut_ad(old_prefix_len < REC_MAX_DATA_SIZE);
if (new_prefix_len == old_prefix_len) {
return(0);
}
if (new_prefix_len == 0) {
new_prefix_len = ULINT_MAX;
}
if (old_prefix_len == 0) {
old_prefix_len = ULINT_MAX;
}
if (new_prefix_len > old_prefix_len) {
return(1);
} else {
return(-1);
}
}
/** Check whether the column is existing in old table.
@param[in] new_col_no new column no
@param[in] col_map mapping of old column numbers to new ones
@param[in] col_map_size the column map size
@return true if the column is existing, otherwise false. */
static inline
bool
innobase_pk_col_is_existing(
const ulint new_col_no,
const ulint* col_map,
const ulint col_map_size)
{
for (ulint i = 0; i < col_map_size; i++) {
if (col_map[i] == new_col_no) {
return(true);
}
}
return(false);
}
/** Determine whether both the indexes have same set of primary key
fields arranged in the same order.
Rules when we cannot skip sorting:
(1) Removing existing PK columns somewhere else than at the end of the PK;
(2) Adding existing columns to the PK, except at the end of the PK when no
columns are removed from the PK;
(3) Changing the order of existing PK columns;
(4) Decreasing the prefix length just like removing existing PK columns
follows rule(1), Increasing the prefix length just like adding existing
PK columns follows rule(2).
@param[in] col_map mapping of old column numbers to new ones
@param[in] ha_alter_info Data used during in-place alter
@param[in] old_clust_index index to be compared
@param[in] new_clust_index index to be compared
@retval true if both indexes have same order.
@retval false. */
static MY_ATTRIBUTE((warn_unused_result))
bool
innobase_pk_order_preserved(
const ulint* col_map,
const dict_index_t* old_clust_index,
const dict_index_t* new_clust_index)
{
ulint old_n_uniq
= dict_index_get_n_ordering_defined_by_user(
old_clust_index);
ulint new_n_uniq
= dict_index_get_n_ordering_defined_by_user(
new_clust_index);
ut_ad(dict_index_is_clust(old_clust_index));
ut_ad(dict_index_is_clust(new_clust_index));
ut_ad(old_clust_index->table != new_clust_index->table);
ut_ad(col_map != NULL);
if (old_n_uniq == 0) {
/* There was no PRIMARY KEY in the table.
If there is no PRIMARY KEY after the ALTER either,
no sorting is needed. */
return(new_n_uniq == old_n_uniq);
}
/* DROP PRIMARY KEY is only allowed in combination with
ADD PRIMARY KEY. */
ut_ad(new_n_uniq > 0);
/* The order of the last processed new_clust_index key field,
not counting ADD COLUMN, which are constant. */
lint last_field_order = -1;
ulint existing_field_count = 0;
ulint old_n_cols = dict_table_get_n_cols(old_clust_index->table);
for (ulint new_field = 0; new_field < new_n_uniq; new_field++) {
ulint new_col_no =
new_clust_index->fields[new_field].col->ind;
/* Check if there is a match in old primary key. */
ulint old_field = 0;
while (old_field < old_n_uniq) {
ulint old_col_no =
old_clust_index->fields[old_field].col->ind;
if (col_map[old_col_no] == new_col_no) {
break;
}
old_field++;
}
/* The order of key field in the new primary key.
1. old PK column: idx in old primary key
2. existing column: old_n_uniq + sequence no
3. newly added column: no order */
lint new_field_order;
const bool old_pk_column = old_field < old_n_uniq;
if (old_pk_column) {
new_field_order = old_field;
} else if (innobase_pk_col_is_existing(new_col_no, col_map,
old_n_cols)) {
new_field_order = old_n_uniq + existing_field_count++;
} else {
/* Skip newly added column. */
continue;
}
if (last_field_order + 1 != new_field_order) {
/* Old PK order is not kept, or existing column
is not added at the end of old PK. */
return(false);
}
last_field_order = new_field_order;
if (!old_pk_column) {
continue;
}
/* Check prefix length change. */
const lint prefix_change = innobase_pk_col_prefix_compare(
new_clust_index->fields[new_field].prefix_len,
old_clust_index->fields[old_field].prefix_len);
if (prefix_change < 0) {
/* If a column's prefix length is decreased, it should
be the last old PK column in new PK.
Note: we set last_field_order to -2, so that if there
are any old PK colmns or existing columns after it in
new PK, the comparison to new_field_order will fail in
the next round.*/
last_field_order = -2;
} else if (prefix_change > 0) {
/* If a column's prefix length is increased, it should
be the last PK column in old PK. */
if (old_field != old_n_uniq - 1) {
return(false);
}
}
}
return(true);
}
/** Update the mtype from DATA_BLOB to DATA_GEOMETRY for a specified
GIS column of a table. This is used when we want to create spatial index
on legacy GIS columns coming from 5.6, where we store GIS data as DATA_BLOB
in innodb layer.
@param[in] table_id table id
@param[in] col_name column name
@param[in] trx data dictionary transaction
@retval true Failure
@retval false Success */
static
bool
innobase_update_gis_column_type(
table_id_t table_id,
const char* col_name,
trx_t* trx)
{
pars_info_t* info;
dberr_t error;
DBUG_ENTER("innobase_update_gis_column_type");
DBUG_ASSERT(trx_get_dict_operation(trx) == TRX_DICT_OP_INDEX);
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
info = pars_info_create();
pars_info_add_ull_literal(info, "tableid", table_id);
pars_info_add_str_literal(info, "name", col_name);
pars_info_add_int4_literal(info, "mtype", DATA_GEOMETRY);
trx->op_info = "update column type to DATA_GEOMETRY";
error = que_eval_sql(
info,
"PROCEDURE UPDATE_SYS_COLUMNS_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_COLUMNS SET MTYPE=:mtype\n"
"WHERE TABLE_ID=:tableid AND NAME=:name;\n"
"END;\n",
false, trx);
trx->error_state = DB_SUCCESS;
trx->op_info = "";
DBUG_RETURN(error != DB_SUCCESS);
}
/** Check if we are creating spatial indexes on GIS columns, which are
legacy columns from earlier MySQL, such as 5.6. If so, we have to update
the mtypes of the old GIS columns to DATA_GEOMETRY.
In 5.6, we store GIS columns as DATA_BLOB in InnoDB layer, it will introduce
confusion when we run latest server on older data. That's why we need to
do the upgrade.
@param[in] ha_alter_info Data used during in-place alter
@param[in] table Table on which we want to add indexes
@param[in] trx Transaction
@return DB_SUCCESS if update successfully or no columns need to be updated,
otherwise DB_ERROR, which means we can't update the mtype for some
column, and creating spatial index on it should be dangerous */
static
dberr_t
innobase_check_gis_columns(
Alter_inplace_info* ha_alter_info,
dict_table_t* table,
trx_t* trx)
{
DBUG_ENTER("innobase_check_gis_columns");
for (uint key_num = 0;
key_num < ha_alter_info->index_add_count;
key_num++) {
const KEY& key = ha_alter_info->key_info_buffer[
ha_alter_info->index_add_buffer[key_num]];
if (!(key.flags & HA_SPATIAL)) {
continue;
}
ut_ad(key.user_defined_key_parts == 1);
const KEY_PART_INFO& key_part = key.key_part[0];
/* Does not support spatial index on virtual columns */
if (innobase_is_v_fld(key_part.field)) {
DBUG_RETURN(DB_UNSUPPORTED);
}
ulint col_nr = dict_table_has_column(
table,
key_part.field->field_name,
key_part.fieldnr);
ut_ad(col_nr != table->n_def);
dict_col_t* col = &table->cols[col_nr];
if (col->mtype != DATA_BLOB) {
ut_ad(DATA_GEOMETRY_MTYPE(col->mtype));
continue;
}
const char* col_name = dict_table_get_col_name(
table, col_nr);
if (innobase_update_gis_column_type(
table->id, col_name, trx)) {
DBUG_RETURN(DB_ERROR);
} else {
col->mtype = DATA_GEOMETRY;
ib::info() << "Updated mtype of column" << col_name
<< " in table " << table->name
<< ", whose id is " << table->id
<< " to DATA_GEOMETRY";
}
}
DBUG_RETURN(DB_SUCCESS);
}
/** Collect virtual column info for its addition
@param[in] ha_alter_info Data used during in-place alter
@param[in] altered_table MySQL table that is being altered to
@param[in] table MySQL table as it is before the ALTER operation
@retval true Failure
@retval false Success */
static
bool
prepare_inplace_add_virtual(
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* table)
{
ha_innobase_inplace_ctx* ctx;
ulint i = 0;
ulint j = 0;
const Create_field* new_field;
ctx = static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
ctx->num_to_add_vcol = altered_table->s->fields
+ ctx->num_to_drop_vcol - table->s->fields;
ctx->add_vcol = static_cast<dict_v_col_t*>(
mem_heap_zalloc(ctx->heap, ctx->num_to_add_vcol
* sizeof *ctx->add_vcol));
ctx->add_vcol_name = static_cast<const char**>(
mem_heap_alloc(ctx->heap, ctx->num_to_add_vcol
* sizeof *ctx->add_vcol_name));
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
while ((new_field = (cf_it++)) != NULL) {
const Field* field = new_field->field;
ulint old_i;
for (old_i = 0; table->field[old_i]; old_i++) {
const Field* n_field = table->field[old_i];
if (field == n_field) {
break;
}
}
i++;
if (table->field[old_i]) {
continue;
}
ut_ad(!field);
ulint col_len;
ulint is_unsigned;
ulint field_type;
ulint charset_no;
field = altered_table->field[i - 1];
ulint col_type
= get_innobase_type_from_mysql_type(
&is_unsigned, field);
if (!innobase_is_v_fld(field)) {
continue;
}
col_len = field->pack_length();
field_type = (ulint) field->type();
if (!field->real_maybe_null()) {
field_type |= DATA_NOT_NULL;
}
if (field->binary()) {
field_type |= DATA_BINARY_TYPE;
}
if (is_unsigned) {
field_type |= DATA_UNSIGNED;
}
if (dtype_is_string_type(col_type)) {
charset_no = (ulint) field->charset()->number;
DBUG_EXECUTE_IF(
"ib_alter_add_virtual_fail",
charset_no += MAX_CHAR_COLL_NUM;);
if (charset_no > MAX_CHAR_COLL_NUM) {
my_error(ER_WRONG_KEY_COLUMN, MYF(0), "InnoDB",
field->field_name);
return(true);
}
} else {
charset_no = 0;
}
if (field->type() == MYSQL_TYPE_VARCHAR) {
uint32 length_bytes
= static_cast<const Field_varstring*>(
field)->length_bytes;
col_len -= length_bytes;
if (length_bytes == 2) {
field_type |= DATA_LONG_TRUE_VARCHAR;
}
}
ctx->add_vcol[j].m_col.prtype = dtype_form_prtype(
field_type, charset_no);
ctx->add_vcol[j].m_col.prtype |= DATA_VIRTUAL;
ctx->add_vcol[j].m_col.mtype = col_type;
ctx->add_vcol[j].m_col.len = col_len;
ctx->add_vcol[j].m_col.ind = i - 1;
ctx->add_vcol[j].num_base = 0;
ctx->add_vcol_name[j] = field->field_name;
ctx->add_vcol[j].base_col = NULL;
ctx->add_vcol[j].v_pos = ctx->old_table->n_v_cols
- ctx->num_to_drop_vcol + j;
/* No need to track the list */
ctx->add_vcol[j].v_indexes = NULL;
innodb_base_col_setup(ctx->old_table, field, &ctx->add_vcol[j]);
j++;
}
return(false);
}
/** Collect virtual column info for its addition
@param[in] ha_alter_info Data used during in-place alter
@param[in] altered_table MySQL table that is being altered to
@param[in] table MySQL table as it is before the ALTER operation
@retval true Failure
@retval false Success */
static
bool
prepare_inplace_drop_virtual(
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* table)
{
ha_innobase_inplace_ctx* ctx;
ulint i = 0;
ulint j = 0;
ctx = static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
ctx->num_to_drop_vcol = 0;
for (i = 0; table->field[i]; i++) {
const Field* field = table->field[i];
if (field->flags & FIELD_IS_DROPPED && !field->stored_in_db()) {
ctx->num_to_drop_vcol++;
}
}
ctx->drop_vcol = static_cast<dict_v_col_t*>(
mem_heap_alloc(ctx->heap, ctx->num_to_drop_vcol
* sizeof *ctx->drop_vcol));
ctx->drop_vcol_name = static_cast<const char**>(
mem_heap_alloc(ctx->heap, ctx->num_to_drop_vcol
* sizeof *ctx->drop_vcol_name));
for (i = 0; table->field[i]; i++) {
Field *field = table->field[i];
if (!(field->flags & FIELD_IS_DROPPED) || field->stored_in_db()) {
continue;
}
ulint col_len;
ulint is_unsigned;
ulint field_type;
ulint charset_no;
ulint col_type
= get_innobase_type_from_mysql_type(
&is_unsigned, field);
col_len = field->pack_length();
field_type = (ulint) field->type();
if (!field->real_maybe_null()) {
field_type |= DATA_NOT_NULL;
}
if (field->binary()) {
field_type |= DATA_BINARY_TYPE;
}
if (is_unsigned) {
field_type |= DATA_UNSIGNED;
}
if (dtype_is_string_type(col_type)) {
charset_no = (ulint) field->charset()->number;
DBUG_EXECUTE_IF(
"ib_alter_add_virtual_fail",
charset_no += MAX_CHAR_COLL_NUM;);
if (charset_no > MAX_CHAR_COLL_NUM) {
my_error(ER_WRONG_KEY_COLUMN, MYF(0), "InnoDB",
field->field_name);
return(true);
}
} else {
charset_no = 0;
}
if (field->type() == MYSQL_TYPE_VARCHAR) {
uint32 length_bytes
= static_cast<const Field_varstring*>(
field)->length_bytes;
col_len -= length_bytes;
if (length_bytes == 2) {
field_type |= DATA_LONG_TRUE_VARCHAR;
}
}
ctx->drop_vcol[j].m_col.prtype = dtype_form_prtype(
field_type, charset_no);
ctx->drop_vcol[j].m_col.prtype |= DATA_VIRTUAL;
ctx->drop_vcol[j].m_col.mtype = col_type;
ctx->drop_vcol[j].m_col.len = col_len;
ctx->drop_vcol[j].m_col.ind = i;
ctx->drop_vcol_name[j] = field->field_name;
dict_v_col_t* v_col = dict_table_get_nth_v_col_mysql(
ctx->old_table, i);
ctx->drop_vcol[j].v_pos = v_col->v_pos;
j++;
}
return(false);
}
/** Insert a new record to INNODB SYS_VIRTUAL
@param[in] table InnoDB table
@param[in] pos virtual column column no
@param[in] base_pos base column pos
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_insert_sys_virtual(
const dict_table_t* table,
ulint pos,
ulint base_pos,
trx_t* trx)
{
pars_info_t* info = pars_info_create();
pars_info_add_ull_literal(info, "id", table->id);
pars_info_add_int4_literal(info, "pos", pos);
pars_info_add_int4_literal(info, "base_pos", base_pos);
dberr_t error = que_eval_sql(
info,
"PROCEDURE P () IS\n"
"BEGIN\n"
"INSERT INTO SYS_VIRTUAL VALUES"
"(:id, :pos, :base_pos);\n"
"END;\n",
FALSE, trx);
return(error);
}
/** Update INNODB SYS_COLUMNS on new virtual columns
@param[in] table InnoDB table
@param[in] col_name column name
@param[in] vcol virtual column
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_add_one_virtual(
const dict_table_t* table,
const char* col_name,
dict_v_col_t* vcol,
trx_t* trx)
{
ulint pos = dict_create_v_col_pos(vcol->v_pos,
vcol->m_col.ind);
ulint mtype = vcol->m_col.mtype;
ulint prtype = vcol->m_col.prtype;
ulint len = vcol->m_col.len;
pars_info_t* info = pars_info_create();
pars_info_add_ull_literal(info, "id", table->id);
pars_info_add_int4_literal(info, "pos", pos);
pars_info_add_str_literal(info, "name", col_name);
pars_info_add_int4_literal(info, "mtype", mtype);
pars_info_add_int4_literal(info, "prtype", prtype);
pars_info_add_int4_literal(info, "len", len);
pars_info_add_int4_literal(info, "prec", vcol->num_base);
dberr_t error = que_eval_sql(
info,
"PROCEDURE P () IS\n"
"BEGIN\n"
"INSERT INTO SYS_COLUMNS VALUES"
"(:id, :pos, :name, :mtype, :prtype, :len, :prec);\n"
"END;\n",
FALSE, trx);
if (error != DB_SUCCESS) {
return(error);
}
for (ulint i = 0; i < vcol->num_base; i++) {
error = innobase_insert_sys_virtual(
table, pos, vcol->base_col[i]->ind, trx);
if (error != DB_SUCCESS) {
return(error);
}
}
return(error);
}
/** Update INNODB SYS_TABLES on number of virtual columns
@param[in] user_table InnoDB table
@param[in] n_col number of columns
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_update_n_virtual(
const dict_table_t* table,
ulint n_col,
trx_t* trx)
{
dberr_t err = DB_SUCCESS;
pars_info_t* info = pars_info_create();
pars_info_add_int4_literal(info, "num_col", n_col);
pars_info_add_ull_literal(info, "id", table->id);
err = que_eval_sql(
info,
"PROCEDURE RENUMBER_TABLE_ID_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_TABLES"
" SET N_COLS = :num_col\n"
" WHERE ID = :id;\n"
"END;\n", FALSE, trx);
return(err);
}
/** Update system table for adding virtual column(s)
@param[in] ha_alter_info Data used during in-place alter
@param[in] altered_table MySQL table that is being altered
@param[in] table MySQL table as it is before the ALTER operation
@param[in] user_table InnoDB table
@param[in] trx transaction
@retval true Failure
@retval false Success */
static
bool
innobase_add_virtual_try(
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* table,
const dict_table_t* user_table,
trx_t* trx)
{
ha_innobase_inplace_ctx* ctx;
dberr_t err = DB_SUCCESS;
ctx = static_cast<ha_innobase_inplace_ctx*>(
ha_alter_info->handler_ctx);
for (ulint i = 0; i < ctx->num_to_add_vcol; i++) {
err = innobase_add_one_virtual(
user_table, ctx->add_vcol_name[i],
&ctx->add_vcol[i], trx);
if (err != DB_SUCCESS) {
my_error(ER_INTERNAL_ERROR, MYF(0),
"InnoDB: ADD COLUMN...VIRTUAL");
return(true);
}
}
ulint n_col = user_table->n_cols;
ulint n_v_col = user_table->n_v_cols;
n_v_col += ctx->num_to_add_vcol;
n_col -= dict_table_get_n_sys_cols(user_table);
n_v_col -= ctx->num_to_drop_vcol;
ulint new_n = dict_table_encode_n_col(n_col, n_v_col)
+ ((user_table->flags & DICT_TF_COMPACT) << 31);
err = innobase_update_n_virtual(user_table, new_n, trx);
if (err != DB_SUCCESS) {
my_error(ER_INTERNAL_ERROR, MYF(0),
"InnoDB: ADD COLUMN...VIRTUAL");
return(true);
}
return(false);
}
/** Update INNODB SYS_COLUMNS on new virtual column's position
@param[in] table InnoDB table
@param[in] old_pos old position
@param[in] new_pos new position
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_update_v_pos_sys_columns(
const dict_table_t* table,
ulint old_pos,
ulint new_pos,
trx_t* trx)
{
pars_info_t* info = pars_info_create();
pars_info_add_int4_literal(info, "pos", old_pos);
pars_info_add_int4_literal(info, "val", new_pos);
pars_info_add_ull_literal(info, "id", table->id);
dberr_t error = que_eval_sql(
info,
"PROCEDURE P () IS\n"
"BEGIN\n"
"UPDATE SYS_COLUMNS\n"
"SET POS = :val\n"
"WHERE POS = :pos\n"
"AND TABLE_ID = :id;\n"
"END;\n",
FALSE, trx);
return(error);
}
/** Update INNODB SYS_VIRTUAL table with new virtual column position
@param[in] table InnoDB table
@param[in] old_pos old position
@param[in] new_pos new position
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_update_v_pos_sys_virtual(
const dict_table_t* table,
ulint old_pos,
ulint new_pos,
trx_t* trx)
{
pars_info_t* info = pars_info_create();
pars_info_add_int4_literal(info, "pos", old_pos);
pars_info_add_int4_literal(info, "val", new_pos);
pars_info_add_ull_literal(info, "id", table->id);
dberr_t error = que_eval_sql(
info,
"PROCEDURE P () IS\n"
"BEGIN\n"
"UPDATE SYS_VIRTUAL\n"
"SET POS = :val\n"
"WHERE POS = :pos\n"
"AND TABLE_ID = :id;\n"
"END;\n",
FALSE, trx);
return(error);
}
/** Update InnoDB system tables on dropping a virtual column
@param[in] table InnoDB table
@param[in] col_name column name of the dropping column
@param[in] drop_col col information for the dropping column
@param[in] n_prev_dropped number of previously dropped columns in the
same alter clause
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_drop_one_virtual_sys_columns(
const dict_table_t* table,
const char* col_name,
dict_col_t* drop_col,
ulint n_prev_dropped,
trx_t* trx)
{
pars_info_t* info = pars_info_create();
pars_info_add_ull_literal(info, "id", table->id);
pars_info_add_str_literal(info, "name", col_name);
dberr_t error = que_eval_sql(
info,
"PROCEDURE P () IS\n"
"BEGIN\n"
"DELETE FROM SYS_COLUMNS\n"
"WHERE TABLE_ID = :id\n"
"AND NAME = :name;\n"
"END;\n",
FALSE, trx);
if (error != DB_SUCCESS) {
return(error);
}
dict_v_col_t* v_col = dict_table_get_nth_v_col_mysql(
table, drop_col->ind);
/* Adjust column positions for all subsequent columns */
for (ulint i = v_col->v_pos + 1; i < table->n_v_cols; i++) {
dict_v_col_t* t_col = dict_table_get_nth_v_col(table, i);
ulint old_p = dict_create_v_col_pos(
t_col->v_pos - n_prev_dropped,
t_col->m_col.ind - n_prev_dropped);
ulint new_p = dict_create_v_col_pos(
t_col->v_pos - 1 - n_prev_dropped,
t_col->m_col.ind - 1 - n_prev_dropped);
error = innobase_update_v_pos_sys_columns(
table, old_p, new_p, trx);
if (error != DB_SUCCESS) {
return(error);
}
error = innobase_update_v_pos_sys_virtual(
table, old_p, new_p, trx);
if (error != DB_SUCCESS) {
return(error);
}
}
return(error);
}
/** Delete virtual column's info from INNODB SYS_VIRTUAL
@param[in] table InnoDB table
@param[in] pos position of the virtual column to be deleted
@param[in] trx transaction
@return DB_SUCCESS if successful, otherwise error code */
static
dberr_t
innobase_drop_one_virtual_sys_virtual(
const dict_table_t* table,
ulint pos,
trx_t* trx)
{
pars_info_t* info = pars_info_create();
pars_info_add_ull_literal(info, "id", table->id);
pars_info_add_int4_literal(info, "pos", pos);
dberr_t error = que_eval_sql(
info,
"PROCEDURE P () IS\n"
"BEGIN\n"
"DELETE FROM SYS_VIRTUAL\n"
"WHERE TABLE_ID = :id\n"
"AND POS = :pos;\n"
"END;\n",
FALSE, trx);
return(error);
}
/** Update system table for dropping virtual column(s)
@param[in] ha_alter_info Data used during in-place alter
@param[in] altered_table MySQL table that is being altered
@param[in] table MySQL table as it is before the ALTER operation
@param[in] user_table InnoDB table
@param[in] trx transaction
@retval true Failure
@retval false Success */
static
bool
innobase_drop_virtual_try(
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* table,
const dict_table_t* user_table,
trx_t* trx)
{
ha_innobase_inplace_ctx* ctx;
dberr_t err = DB_SUCCESS;
ctx = static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
for (ulint i = 0; i < ctx->num_to_drop_vcol; i++) {
ulint pos = dict_create_v_col_pos(
ctx->drop_vcol[i].v_pos - i,
ctx->drop_vcol[i].m_col.ind - i);
err = innobase_drop_one_virtual_sys_virtual(
user_table, pos, trx);
if (err != DB_SUCCESS) {
my_error(ER_INTERNAL_ERROR, MYF(0),
"InnoDB: DROP COLUMN...VIRTUAL");
return(true);
}
err = innobase_drop_one_virtual_sys_columns(
user_table, ctx->drop_vcol_name[i],
&(ctx->drop_vcol[i].m_col), i, trx);
if (err != DB_SUCCESS) {
my_error(ER_INTERNAL_ERROR, MYF(0),
"InnoDB: DROP COLUMN...VIRTUAL");
return(true);
}
}
ulint n_col = user_table->n_cols;
ulint n_v_col = user_table->n_v_cols;
n_v_col -= ctx->num_to_drop_vcol;
n_col -= dict_table_get_n_sys_cols(user_table);
ulint new_n = dict_table_encode_n_col(n_col, n_v_col)
+ ((user_table->flags & DICT_TF_COMPACT) << 31);
err = innobase_update_n_virtual(user_table, new_n, trx);
if (err != DB_SUCCESS) {
my_error(ER_INTERNAL_ERROR, MYF(0),
"InnoDB: DROP COLUMN...VIRTUAL");
}
return(false);
}
/** Adjust the create index column number from "New table" to
"old InnoDB table" while we are doing dropping virtual column. Since we do
not create separate new table for the dropping/adding virtual columns.
To correctly find the indexed column, we will need to find its col_no
in the "Old Table", not the "New table".
@param[in] ha_alter_info Data used during in-place alter
@param[in] old_table MySQL table as it is before the ALTER operation
@param[in] num_v_dropped number of virtual column dropped
@param[in,out] index_def index definition */
static
void
innodb_v_adjust_idx_col(
const Alter_inplace_info* ha_alter_info,
const TABLE* old_table,
ulint num_v_dropped,
index_def_t* index_def)
{
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
for (ulint i = 0; i < index_def->n_fields; i++) {
#ifdef UNIV_DEBUG
bool col_found = false;
#endif /* UNIV_DEBUG */
ulint num_v = 0;
index_field_t* index_field = &index_def->fields[i];
/* Only adjust virtual column col_no, since non-virtual
column position (in non-vcol list) won't change unless
table rebuild */
if (!index_field->is_v_col) {
continue;
}
const Field* field = NULL;
cf_it.rewind();
/* Found the field in the new table */
while (const Create_field* new_field = cf_it++) {
if (!innobase_is_v_fld(new_field)) {
continue;
}
field = new_field->field;
if (num_v == index_field->col_no) {
break;
}
num_v++;
}
if (!field) {
/* this means the field is a newly added field, this
should have been blocked when we drop virtual column
at the same time */
ut_ad(num_v_dropped > 0);
ut_a(0);
}
ut_ad(innobase_is_v_fld(field));
num_v = 0;
/* Look for its position in old table */
for (uint old_i = 0; old_table->field[old_i]; old_i++) {
if (old_table->field[old_i] == field) {
/* Found it, adjust its col_no to its position
in old table */
index_def->fields[i].col_no = num_v;
ut_d(col_found = true);
break;
}
if (innobase_is_v_fld(old_table->field[old_i])) {
num_v++;
}
}
ut_ad(col_found);
}
}
/** Update internal structures with concurrent writes blocked,
while preparing ALTER TABLE.
@param ha_alter_info Data used during in-place alter
@param altered_table MySQL table that is being altered
@param old_table MySQL table as it is before the ALTER operation
@param table_name Table name in MySQL
@param flags Table and tablespace flags
@param flags2 Additional table flags
@param fts_doc_id_col The column number of FTS_DOC_ID
@param add_fts_doc_id Flag: add column FTS_DOC_ID?
@param add_fts_doc_id_idx Flag: add index FTS_DOC_ID_INDEX (FTS_DOC_ID)?
@retval true Failure
@retval false Success
*/
static MY_ATTRIBUTE((warn_unused_result, nonnull(1,2,3,4)))
bool
prepare_inplace_alter_table_dict(
/*=============================*/
Alter_inplace_info* ha_alter_info,
const TABLE* altered_table,
const TABLE* old_table,
const char* table_name,
ulint flags,
ulint flags2,
ulint fts_doc_id_col,
bool add_fts_doc_id,
bool add_fts_doc_id_idx)
{
bool dict_locked = false;
ulint* add_key_nums; /* MySQL key numbers */
index_def_t* index_defs; /* index definitions */
dict_table_t* user_table;
dict_index_t* fts_index = NULL;
ulint new_clustered = 0;
dberr_t error;
ulint num_fts_index;
dict_add_v_col_t* add_v = NULL;
ha_innobase_inplace_ctx*ctx;
DBUG_ENTER("prepare_inplace_alter_table_dict");
ctx = static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
DBUG_ASSERT((ctx->add_autoinc != ULINT_UNDEFINED)
== (ctx->sequence.m_max_value > 0));
DBUG_ASSERT(!ctx->num_to_drop_index == !ctx->drop_index);
DBUG_ASSERT(!ctx->num_to_drop_fk == !ctx->drop_fk);
DBUG_ASSERT(!add_fts_doc_id || add_fts_doc_id_idx);
DBUG_ASSERT(!add_fts_doc_id_idx
|| innobase_fulltext_exist(altered_table));
DBUG_ASSERT(!ctx->add_cols);
DBUG_ASSERT(!ctx->add_index);
DBUG_ASSERT(!ctx->add_key_numbers);
DBUG_ASSERT(!ctx->num_to_add_index);
user_table = ctx->new_table;
trx_start_if_not_started_xa(ctx->prebuilt->trx, true);
if (ha_alter_info->handler_flags
& Alter_inplace_info::DROP_VIRTUAL_COLUMN) {
if (prepare_inplace_drop_virtual(
ha_alter_info, altered_table, old_table)) {
DBUG_RETURN(true);
}
}
if (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_VIRTUAL_COLUMN) {
if (prepare_inplace_add_virtual(
ha_alter_info, altered_table, old_table)) {
DBUG_RETURN(true);
}
/* Need information for newly added virtual columns
for create index */
if (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_INDEX) {
add_v = static_cast<dict_add_v_col_t*>(
mem_heap_alloc(ctx->heap, sizeof *add_v));
add_v->n_v_col = ctx->num_to_add_vcol;
add_v->v_col = ctx->add_vcol;
add_v->v_col_name = ctx->add_vcol_name;
}
}
/* There should be no order change for virtual columns coming in
here */
ut_ad(check_v_col_in_order(old_table, altered_table, ha_alter_info));
/* Create a background transaction for the operations on
the data dictionary tables. */
ctx->trx = innobase_trx_allocate(ctx->prebuilt->trx->mysql_thd);
trx_start_for_ddl(ctx->trx, TRX_DICT_OP_INDEX);
/* Create table containing all indexes to be built in this
ALTER TABLE ADD INDEX so that they are in the correct order
in the table. */
ctx->num_to_add_index = ha_alter_info->index_add_count;
ut_ad(ctx->prebuilt->trx->mysql_thd != NULL);
const char* path = thd_innodb_tmpdir(
ctx->prebuilt->trx->mysql_thd);
index_defs = innobase_create_key_defs(
ctx->heap, ha_alter_info, altered_table, ctx->num_to_add_index,
num_fts_index,
row_table_got_default_clust_index(ctx->new_table),
fts_doc_id_col, add_fts_doc_id, add_fts_doc_id_idx,
old_table);
new_clustered = DICT_CLUSTERED & index_defs[0].ind_type;
if (num_fts_index > 1) {
my_error(ER_INNODB_FT_LIMIT, MYF(0));
goto error_handled;
}
if (!ctx->online) {
/* This is not an online operation (LOCK=NONE). */
} else if (ctx->add_autoinc == ULINT_UNDEFINED
&& num_fts_index == 0
&& (!innobase_need_rebuild(ha_alter_info, old_table)
|| !innobase_fulltext_exist(altered_table))) {
/* InnoDB can perform an online operation (LOCK=NONE). */
} else {
size_t query_length;
/* This should have been blocked in
check_if_supported_inplace_alter(). */
ut_ad(0);
my_error(ER_NOT_SUPPORTED_YET, MYF(0),
innobase_get_stmt_unsafe(ctx->prebuilt->trx->mysql_thd,
&query_length));
goto error_handled;
}
/* The primary index would be rebuilt if a FTS Doc ID
column is to be added, and the primary index definition
is just copied from old table and stored in indexdefs[0] */
DBUG_ASSERT(!add_fts_doc_id || new_clustered);
DBUG_ASSERT(!!new_clustered ==
(innobase_need_rebuild(ha_alter_info, old_table)
|| add_fts_doc_id));
/* Allocate memory for dictionary index definitions */
ctx->add_index = static_cast<dict_index_t**>(
mem_heap_alloc(ctx->heap, ctx->num_to_add_index
* sizeof *ctx->add_index));
ctx->add_key_numbers = add_key_nums = static_cast<ulint*>(
mem_heap_alloc(ctx->heap, ctx->num_to_add_index
* sizeof *ctx->add_key_numbers));
/* This transaction should be dictionary operation, so that
the data dictionary will be locked during crash recovery. */
ut_ad(ctx->trx->dict_operation == TRX_DICT_OP_INDEX);
/* Acquire a lock on the table before creating any indexes. */
if (ctx->online) {
error = DB_SUCCESS;
} else {
error = row_merge_lock_table(
ctx->prebuilt->trx, ctx->new_table, LOCK_S);
if (error != DB_SUCCESS) {
goto error_handling;
}
}
/* Latch the InnoDB data dictionary exclusively so that no deadlocks
or lock waits can happen in it during an index create operation. */
row_mysql_lock_data_dictionary(ctx->trx);
dict_locked = true;
/* Wait for background stats processing to stop using the table that
we are going to alter. We know bg stats will not start using it again
until we are holding the data dict locked and we are holding it here
at least until checking ut_ad(user_table->n_ref_count == 1) below.
XXX what may happen if bg stats opens the table after we
have unlocked data dictionary below? */
dict_stats_wait_bg_to_stop_using_table(user_table, ctx->trx);
online_retry_drop_indexes_low(ctx->new_table, ctx->trx);
ut_d(dict_table_check_for_dup_indexes(
ctx->new_table, CHECK_ABORTED_OK));
/* If a new clustered index is defined for the table we need
to rebuild the table with a temporary name. */
if (new_clustered) {
fil_space_crypt_t* crypt_data;
const char* new_table_name
= dict_mem_create_temporary_tablename(
ctx->heap,
ctx->new_table->name.m_name,
ctx->new_table->id);
ulint n_cols = 0;
ulint n_v_cols = 0;
dtuple_t* add_cols;
ulint space_id = 0;
ulint z = 0;
uint32_t key_id = FIL_DEFAULT_ENCRYPTION_KEY;
fil_encryption_t mode = FIL_ENCRYPTION_DEFAULT;
fil_space_t* space = fil_space_acquire(ctx->prebuilt->table->space);
crypt_data = space->crypt_data;
fil_space_release(space);
if (crypt_data) {
key_id = crypt_data->key_id;
mode = crypt_data->encryption;
}
if (innobase_check_foreigns(
ha_alter_info, altered_table, old_table,
user_table, ctx->drop_fk, ctx->num_to_drop_fk)) {
goto new_clustered_failed;
}
for (uint i = 0; i < altered_table->s->fields; i++) {
const Field* field = altered_table->field[i];
if (innobase_is_v_fld(field)) {
n_v_cols++;
} else {
n_cols++;
}
}
ut_ad(n_cols + n_v_cols == altered_table->s->fields);
if (add_fts_doc_id) {
n_cols++;
DBUG_ASSERT(flags2 & DICT_TF2_FTS);
DBUG_ASSERT(add_fts_doc_id_idx);
flags2 |= DICT_TF2_FTS_ADD_DOC_ID
| DICT_TF2_FTS_HAS_DOC_ID
| DICT_TF2_FTS;
}
DBUG_ASSERT(!add_fts_doc_id_idx || (flags2 & DICT_TF2_FTS));
/* Create the table. */
trx_set_dict_operation(ctx->trx, TRX_DICT_OP_TABLE);
if (dict_table_get_low(new_table_name)) {
my_error(ER_TABLE_EXISTS_ERROR, MYF(0),
new_table_name);
goto new_clustered_failed;
}
/* The initial space id 0 may be overridden later if this
table is going to be a file_per_table tablespace. */
ctx->new_table = dict_mem_table_create(
new_table_name, space_id, n_cols + n_v_cols, n_v_cols,
flags, flags2);
/* The rebuilt indexed_table will use the renamed
column names. */
ctx->col_names = NULL;
if (DICT_TF_HAS_DATA_DIR(flags)) {
ctx->new_table->data_dir_path =
mem_heap_strdup(ctx->new_table->heap,
user_table->data_dir_path);
}
for (uint i = 0; i < altered_table->s->fields; i++) {
const Field* field = altered_table->field[i];
ulint is_unsigned;
ulint field_type
= (ulint) field->type();
ulint col_type
= get_innobase_type_from_mysql_type(
&is_unsigned, field);
ulint charset_no;
ulint col_len;
bool is_virtual = innobase_is_v_fld(field);
/* we assume in dtype_form_prtype() that this
fits in two bytes */
ut_a(field_type <= MAX_CHAR_COLL_NUM);
if (!field->real_maybe_null()) {
field_type |= DATA_NOT_NULL;
}
if (field->binary()) {
field_type |= DATA_BINARY_TYPE;
}
if (is_unsigned) {
field_type |= DATA_UNSIGNED;
}
if (dtype_is_string_type(col_type)) {
charset_no = (ulint) field->charset()->number;
if (charset_no > MAX_CHAR_COLL_NUM) {
dict_mem_table_free(
ctx->new_table);
my_error(ER_WRONG_KEY_COLUMN, MYF(0), "InnoDB",
field->field_name);
goto new_clustered_failed;
}
} else {
charset_no = 0;
}
col_len = field->pack_length();
/* The MySQL pack length contains 1 or 2 bytes
length field for a true VARCHAR. Let us
subtract that, so that the InnoDB column
length in the InnoDB data dictionary is the
real maximum byte length of the actual data. */
if (field->type() == MYSQL_TYPE_VARCHAR) {
uint32 length_bytes
= static_cast<const Field_varstring*>(
field)->length_bytes;
col_len -= length_bytes;
if (length_bytes == 2) {
field_type |= DATA_LONG_TRUE_VARCHAR;
}
}
if (dict_col_name_is_reserved(field->field_name)) {
dict_mem_table_free(ctx->new_table);
my_error(ER_WRONG_COLUMN_NAME, MYF(0),
field->field_name);
goto new_clustered_failed;
}
if (is_virtual) {
dict_mem_table_add_v_col(
ctx->new_table, ctx->heap,
field->field_name,
col_type,
dtype_form_prtype(
field_type, charset_no)
| DATA_VIRTUAL,
col_len, i, 0);
} else {
dict_mem_table_add_col(
ctx->new_table, ctx->heap,
field->field_name,
col_type,
dtype_form_prtype(
field_type, charset_no),
col_len);
}
}
if (n_v_cols) {
for (uint i = 0; i < altered_table->s->fields; i++) {
dict_v_col_t* v_col;
const Field* field = altered_table->field[i];
if (!innobase_is_v_fld(field)) {
continue;
}
v_col = dict_table_get_nth_v_col(
ctx->new_table, z);
z++;
innodb_base_col_setup(
ctx->new_table, field, v_col);
}
}
if (add_fts_doc_id) {
fts_add_doc_id_column(ctx->new_table, ctx->heap);
ctx->new_table->fts->doc_col = fts_doc_id_col;
ut_ad(fts_doc_id_col
== altered_table->s->fields - n_v_cols);
} else if (ctx->new_table->fts) {
ctx->new_table->fts->doc_col = fts_doc_id_col;
}
error = row_create_table_for_mysql(
ctx->new_table, ctx->trx, false, mode, key_id);
switch (error) {
dict_table_t* temp_table;
case DB_SUCCESS:
/* We need to bump up the table ref count and
before we can use it we need to open the
table. The new_table must be in the data
dictionary cache, because we are still holding
the dict_sys->mutex. */
ut_ad(mutex_own(&dict_sys->mutex));
temp_table = dict_table_open_on_name(
ctx->new_table->name.m_name, TRUE, FALSE,
DICT_ERR_IGNORE_NONE);
ut_a(ctx->new_table == temp_table);
/* n_ref_count must be 1, because purge cannot
be executing on this very table as we are
holding dict_operation_lock X-latch. */
DBUG_ASSERT(ctx->new_table->get_ref_count() == 1);
break;
case DB_TABLESPACE_EXISTS:
my_error(ER_TABLESPACE_EXISTS, MYF(0),
new_table_name);
goto new_clustered_failed;
case DB_DUPLICATE_KEY:
my_error(HA_ERR_TABLE_EXIST, MYF(0),
altered_table->s->table_name.str);
goto new_clustered_failed;
case DB_UNSUPPORTED:
my_error(ER_UNSUPPORTED_EXTENSION, MYF(0),
ctx->new_table->name.m_name);
goto new_clustered_failed;
default:
my_error_innodb(error, table_name, flags);
new_clustered_failed:
DBUG_ASSERT(ctx->trx != ctx->prebuilt->trx);
trx_rollback_to_savepoint(ctx->trx, NULL);
ut_ad(user_table->get_ref_count() == 1);
online_retry_drop_indexes_with_trx(
user_table, ctx->trx);
goto err_exit;
}
if (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_COLUMN) {
add_cols = dtuple_create_with_vcol(
ctx->heap,
dict_table_get_n_cols(ctx->new_table),
dict_table_get_n_v_cols(ctx->new_table));
dict_table_copy_types(add_cols, ctx->new_table);
} else {
add_cols = NULL;
}
ctx->col_map = innobase_build_col_map(
ha_alter_info, altered_table, old_table,
ctx->new_table, user_table,
add_cols, ctx->heap);
ctx->add_cols = add_cols;
} else {
DBUG_ASSERT(!innobase_need_rebuild(ha_alter_info, old_table));
DBUG_ASSERT(old_table->s->primary_key
== altered_table->s->primary_key);
for (dict_index_t* index
= dict_table_get_first_index(user_table);
index != NULL;
index = dict_table_get_next_index(index)) {
if (!index->to_be_dropped
&& dict_index_is_corrupted(index)) {
my_error(ER_CHECK_NO_SUCH_TABLE, MYF(0));
goto error_handled;
}
}
for (dict_index_t* index
= dict_table_get_first_index(user_table);
index != NULL;
index = dict_table_get_next_index(index)) {
if (!index->to_be_dropped
&& dict_index_is_corrupted(index)) {
my_error(ER_CHECK_NO_SUCH_TABLE, MYF(0));
goto error_handled;
}
}
if (!ctx->new_table->fts
&& innobase_fulltext_exist(altered_table)) {
ctx->new_table->fts = fts_create(
ctx->new_table);
ctx->new_table->fts->doc_col = fts_doc_id_col;
}
/* Check if we need to update mtypes of legacy GIS columns.
This check is only needed when we don't have to rebuild
the table, since rebuild would update all mtypes for GIS
columns */
error = innobase_check_gis_columns(
ha_alter_info, ctx->new_table, ctx->trx);
if (error != DB_SUCCESS) {
ut_ad(error == DB_ERROR);
error = DB_UNSUPPORTED;
goto error_handling;
}
}
/* Assign table_id, so that no table id of
fts_create_index_tables() will be written to the undo logs. */
DBUG_ASSERT(ctx->new_table->id != 0);
ctx->trx->table_id = ctx->new_table->id;
/* Create the indexes in SYS_INDEXES and load into dictionary. */
for (ulint a = 0; a < ctx->num_to_add_index; a++) {
if (index_defs[a].ind_type & DICT_VIRTUAL
&& ctx->num_to_drop_vcol > 0 && !new_clustered) {
innodb_v_adjust_idx_col(ha_alter_info, old_table,
ctx->num_to_drop_vcol,
&index_defs[a]);
}
ctx->add_index[a] = row_merge_create_index(
ctx->trx, ctx->new_table,
&index_defs[a], add_v, ctx->col_names);
add_key_nums[a] = index_defs[a].key_number;
if (!ctx->add_index[a]) {
error = ctx->trx->error_state;
DBUG_ASSERT(error != DB_SUCCESS);
goto error_handling;
}
DBUG_ASSERT(ctx->add_index[a]->is_committed()
== !!new_clustered);
if (ctx->add_index[a]->type & DICT_FTS) {
DBUG_ASSERT(num_fts_index);
DBUG_ASSERT(!fts_index);
DBUG_ASSERT(ctx->add_index[a]->type == DICT_FTS);
fts_index = ctx->add_index[a];
}
/* If only online ALTER TABLE operations have been
requested, allocate a modification log. If the table
will be locked anyway, the modification
log is unnecessary. When rebuilding the table
(new_clustered), we will allocate the log for the
clustered index of the old table, later. */
if (new_clustered
|| !ctx->online
|| !user_table->is_readable()
|| dict_table_is_discarded(user_table)) {
/* No need to allocate a modification log. */
ut_ad(!ctx->add_index[a]->online_log);
} else if (ctx->add_index[a]->type & DICT_FTS) {
/* Fulltext indexes are not covered
by a modification log. */
} else {
DBUG_EXECUTE_IF("innodb_OOM_prepare_inplace_alter",
error = DB_OUT_OF_MEMORY;
goto error_handling;);
rw_lock_x_lock(&ctx->add_index[a]->lock);
bool ok = row_log_allocate(ctx->add_index[a],
NULL, true, NULL, NULL,
path);
rw_lock_x_unlock(&ctx->add_index[a]->lock);
if (!ok) {
error = DB_OUT_OF_MEMORY;
goto error_handling;
}
}
}
ut_ad(new_clustered == ctx->need_rebuild());
DBUG_EXECUTE_IF("innodb_OOM_prepare_inplace_alter",
error = DB_OUT_OF_MEMORY;
goto error_handling;);
if (new_clustered) {
dict_index_t* clust_index = dict_table_get_first_index(
user_table);
dict_index_t* new_clust_index = dict_table_get_first_index(
ctx->new_table);
ctx->skip_pk_sort = innobase_pk_order_preserved(
ctx->col_map, clust_index, new_clust_index);
DBUG_EXECUTE_IF("innodb_alter_table_pk_assert_no_sort",
DBUG_ASSERT(ctx->skip_pk_sort););
DBUG_ASSERT(!ctx->new_table->persistent_autoinc);
if (const Field* ai = altered_table->found_next_number_field) {
const unsigned col_no = innodb_col_no(ai);
ctx->new_table->persistent_autoinc = 1
+ dict_table_get_nth_col_pos(
ctx->new_table, col_no, NULL);
/* Initialize the AUTO_INCREMENT sequence
to the rebuilt table from the old one. */
if (!old_table->found_next_number_field) {
} else if (ib_uint64_t autoinc
= btr_read_autoinc(clust_index)) {
btr_write_autoinc(new_clust_index, autoinc);
}
}
if (ctx->online) {
/* Allocate a log for online table rebuild. */
rw_lock_x_lock(&clust_index->lock);
bool ok = row_log_allocate(
clust_index, ctx->new_table,
!(ha_alter_info->handler_flags
& Alter_inplace_info::ADD_PK_INDEX),
ctx->add_cols, ctx->col_map, path);
rw_lock_x_unlock(&clust_index->lock);
if (!ok) {
error = DB_OUT_OF_MEMORY;
goto error_handling;
}
}
}
if (ctx->online) {
/* Assign a consistent read view for
row_merge_read_clustered_index(). */
trx_assign_read_view(ctx->prebuilt->trx);
}
if (fts_index) {
/* Ensure that the dictionary operation mode will
not change while creating the auxiliary tables. */
trx_dict_op_t op = trx_get_dict_operation(ctx->trx);
#ifdef UNIV_DEBUG
switch (op) {
case TRX_DICT_OP_NONE:
break;
case TRX_DICT_OP_TABLE:
case TRX_DICT_OP_INDEX:
goto op_ok;
}
ut_error;
op_ok:
#endif /* UNIV_DEBUG */
ut_ad(ctx->trx->dict_operation_lock_mode == RW_X_LATCH);
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
DICT_TF2_FLAG_SET(ctx->new_table, DICT_TF2_FTS);
if (new_clustered) {
/* For !new_clustered, this will be set at
commit_cache_norebuild(). */
ctx->new_table->fts_doc_id_index
= dict_table_get_index_on_name(
ctx->new_table, FTS_DOC_ID_INDEX_NAME);
DBUG_ASSERT(ctx->new_table->fts_doc_id_index != NULL);
}
/* This function will commit the transaction and reset
the trx_t::dict_operation flag on success. */
error = fts_create_index_tables(ctx->trx, fts_index);
DBUG_EXECUTE_IF("innodb_test_fail_after_fts_index_table",
error = DB_LOCK_WAIT_TIMEOUT;
goto error_handling;);
if (error != DB_SUCCESS) {
goto error_handling;
}
trx_start_for_ddl(ctx->trx, op);
if (!ctx->new_table->fts
|| ib_vector_size(ctx->new_table->fts->indexes) == 0) {
error = fts_create_common_tables(
ctx->trx, ctx->new_table,
user_table->name.m_name, TRUE);
DBUG_EXECUTE_IF(
"innodb_test_fail_after_fts_common_table",
error = DB_LOCK_WAIT_TIMEOUT;);
if (error != DB_SUCCESS) {
goto error_handling;
}
ctx->new_table->fts->fts_status
|= TABLE_DICT_LOCKED;
error = innobase_fts_load_stopword(
ctx->new_table, ctx->trx,
ctx->prebuilt->trx->mysql_thd)
? DB_SUCCESS : DB_ERROR;
ctx->new_table->fts->fts_status
&= ulint(~TABLE_DICT_LOCKED);
if (error != DB_SUCCESS) {
goto error_handling;
}
}
ut_ad(trx_get_dict_operation(ctx->trx) == op);
}
DBUG_ASSERT(error == DB_SUCCESS);
/* Commit the data dictionary transaction in order to release
the table locks on the system tables. This means that if
MySQL crashes while creating a new primary key inside
row_merge_build_indexes(), ctx->new_table will not be dropped
by trx_rollback_active(). It will have to be recovered or
dropped by the database administrator. */
trx_commit_for_mysql(ctx->trx);
row_mysql_unlock_data_dictionary(ctx->trx);
dict_locked = false;
ut_a(ctx->trx->lock.n_active_thrs == 0);
error_handling:
/* After an error, remove all those index definitions from the
dictionary which were defined. */
switch (error) {
case DB_SUCCESS:
ut_a(!dict_locked);
ut_d(mutex_enter(&dict_sys->mutex));
ut_d(dict_table_check_for_dup_indexes(
user_table, CHECK_PARTIAL_OK));
ut_d(mutex_exit(&dict_sys->mutex));
DBUG_RETURN(false);
case DB_TABLESPACE_EXISTS:
my_error(ER_TABLESPACE_EXISTS, MYF(0), "(unknown)");
break;
case DB_DUPLICATE_KEY:
my_error(ER_DUP_KEY, MYF(0), "SYS_INDEXES");
break;
case DB_UNSUPPORTED:
my_error(ER_TABLE_CANT_HANDLE_SPKEYS, MYF(0), "SYS_COLUMNS");
break;
default:
my_error_innodb(error, table_name, user_table->flags);
}
error_handled:
ctx->prebuilt->trx->error_info = NULL;
ctx->trx->error_state = DB_SUCCESS;
if (!dict_locked) {
row_mysql_lock_data_dictionary(ctx->trx);
}
if (new_clustered) {
if (ctx->need_rebuild()) {
if (DICT_TF2_FLAG_IS_SET(
ctx->new_table, DICT_TF2_FTS)) {
innobase_drop_fts_index_table(
ctx->new_table, ctx->trx);
}
dict_table_close_and_drop(ctx->trx, ctx->new_table);
/* Free the log for online table rebuild, if
one was allocated. */
dict_index_t* clust_index = dict_table_get_first_index(
user_table);
rw_lock_x_lock(&clust_index->lock);
if (clust_index->online_log) {
ut_ad(ctx->online);
row_log_abort_sec(clust_index);
clust_index->online_status
= ONLINE_INDEX_COMPLETE;
}
rw_lock_x_unlock(&clust_index->lock);
}
trx_commit_for_mysql(ctx->trx);
/* n_ref_count must be 1, because purge cannot
be executing on this very table as we are
holding dict_operation_lock X-latch. */
DBUG_ASSERT(user_table->get_ref_count() == 1 || ctx->online);
online_retry_drop_indexes_with_trx(user_table, ctx->trx);
} else {
ut_ad(!ctx->need_rebuild());
row_merge_drop_indexes(ctx->trx, user_table, TRUE);
trx_commit_for_mysql(ctx->trx);
}
ut_d(dict_table_check_for_dup_indexes(user_table, CHECK_ALL_COMPLETE));
ut_ad(!user_table->drop_aborted);
err_exit:
#ifdef UNIV_DEBUG
/* Clear the to_be_dropped flag in the data dictionary cache. */
for (ulint i = 0; i < ctx->num_to_drop_index; i++) {
DBUG_ASSERT(ctx->drop_index[i]->is_committed());
DBUG_ASSERT(ctx->drop_index[i]->to_be_dropped);
ctx->drop_index[i]->to_be_dropped = 0;
}
#endif /* UNIV_DEBUG */
row_mysql_unlock_data_dictionary(ctx->trx);
trx_free_for_mysql(ctx->trx);
trx_commit_for_mysql(ctx->prebuilt->trx);
delete ctx;
ha_alter_info->handler_ctx = NULL;
DBUG_RETURN(true);
}
/* Check whether an index is needed for the foreign key constraint.
If so, if it is dropped, is there an equivalent index can play its role.
@return true if the index is needed and can't be dropped */
static MY_ATTRIBUTE((nonnull(1,2,3,5), warn_unused_result))
bool
innobase_check_foreign_key_index(
/*=============================*/
Alter_inplace_info* ha_alter_info, /*!< in: Structure describing
changes to be done by ALTER
TABLE */
dict_index_t* index, /*!< in: index to check */
dict_table_t* indexed_table, /*!< in: table that owns the
foreign keys */
const char** col_names, /*!< in: column names, or NULL
for indexed_table->col_names */
trx_t* trx, /*!< in/out: transaction */
dict_foreign_t** drop_fk, /*!< in: Foreign key constraints
to drop */
ulint n_drop_fk) /*!< in: Number of foreign keys
to drop */
{
const dict_foreign_set* fks = &indexed_table->referenced_set;
/* Check for all FK references from other tables to the index. */
for (dict_foreign_set::const_iterator it = fks->begin();
it != fks->end(); ++it) {
dict_foreign_t* foreign = *it;
if (foreign->referenced_index != index) {
continue;
}
ut_ad(indexed_table == foreign->referenced_table);
if (NULL == dict_foreign_find_index(
indexed_table, col_names,
foreign->referenced_col_names,
foreign->n_fields, index,
/*check_charsets=*/TRUE,
/*check_null=*/FALSE,
NULL, NULL, NULL)
&& NULL == innobase_find_equiv_index(
foreign->referenced_col_names,
foreign->n_fields,
ha_alter_info->key_info_buffer,
ha_alter_info->index_add_buffer,
ha_alter_info->index_add_count)) {
/* Index cannot be dropped. */
trx->error_info = index;
return(true);
}
}
fks = &indexed_table->foreign_set;
/* Check for all FK references in current table using the index. */
for (dict_foreign_set::const_iterator it = fks->begin();
it != fks->end(); ++it) {
dict_foreign_t* foreign = *it;
if (foreign->foreign_index != index) {
continue;
}
ut_ad(indexed_table == foreign->foreign_table);
if (!innobase_dropping_foreign(
foreign, drop_fk, n_drop_fk)
&& NULL == dict_foreign_find_index(
indexed_table, col_names,
foreign->foreign_col_names,
foreign->n_fields, index,
/*check_charsets=*/TRUE,
/*check_null=*/FALSE,
NULL, NULL, NULL)
&& NULL == innobase_find_equiv_index(
foreign->foreign_col_names,
foreign->n_fields,
ha_alter_info->key_info_buffer,
ha_alter_info->index_add_buffer,
ha_alter_info->index_add_count)) {
/* Index cannot be dropped. */
trx->error_info = index;
return(true);
}
}
return(false);
}
#ifdef MYSQL_RENAME_INDEX
/**
Rename a given index in the InnoDB data dictionary.
@param index index to rename
@param new_name new name of the index
@param[in,out] trx dict transaction to use, not going to be committed here
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((warn_unused_result))
bool
rename_index_in_data_dictionary(
/*============================*/
const dict_index_t* index,
const char* new_name,
trx_t* trx)
{
DBUG_ENTER("rename_index_in_data_dictionary");
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
pars_info_t* pinfo;
dberr_t err;
pinfo = pars_info_create();
pars_info_add_ull_literal(pinfo, "table_id", index->table->id);
pars_info_add_ull_literal(pinfo, "index_id", index->id);
pars_info_add_str_literal(pinfo, "new_name", new_name);
trx->op_info = "Renaming an index in SYS_INDEXES";
DBUG_EXECUTE_IF(
"ib_rename_index_fail1",
DBUG_SET("+d,innodb_report_deadlock");
);
err = que_eval_sql(
pinfo,
"PROCEDURE RENAME_INDEX_IN_SYS_INDEXES () IS\n"
"BEGIN\n"
"UPDATE SYS_INDEXES SET\n"
"NAME = :new_name\n"
"WHERE\n"
"ID = :index_id AND\n"
"TABLE_ID = :table_id;\n"
"END;\n",
FALSE, trx); /* pinfo is freed by que_eval_sql() */
DBUG_EXECUTE_IF(
"ib_rename_index_fail1",
DBUG_SET("-d,innodb_report_deadlock");
);
trx->op_info = "";
if (err != DB_SUCCESS) {
my_error_innodb(err, index->table->name.m_name, 0);
DBUG_RETURN(true);
}
DBUG_RETURN(false);
}
/**
Rename all indexes in data dictionary of a given table that are
specified in ha_alter_info.
@param ctx alter context, used to fetch the list of indexes to
rename
@param ha_alter_info fetch the new names from here
@param[in,out] trx dict transaction to use, not going to be committed here
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((warn_unused_result))
bool
rename_indexes_in_data_dictionary(
/*==============================*/
const ha_innobase_inplace_ctx* ctx,
const Alter_inplace_info* ha_alter_info,
trx_t* trx)
{
DBUG_ENTER("rename_indexes_in_data_dictionary");
ut_ad(ctx->num_to_rename == ha_alter_info->index_rename_count);
for (ulint i = 0; i < ctx->num_to_rename; i++) {
KEY_PAIR* pair = &ha_alter_info->index_rename_buffer[i];
dict_index_t* index;
index = ctx->rename[i];
ut_ad(strcmp(index->name, pair->old_key->name) == 0);
if (rename_index_in_data_dictionary(index,
pair->new_key->name,
trx)) {
/* failed */
DBUG_RETURN(true);
}
}
DBUG_RETURN(false);
}
/**
Rename a given index in the InnoDB data dictionary cache.
@param[in,out] index index to rename
@param new_name new index name
*/
static
void
rename_index_in_cache(
/*==================*/
dict_index_t* index,
const char* new_name)
{
DBUG_ENTER("rename_index_in_cache");
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
size_t old_name_len = strlen(index->name);
size_t new_name_len = strlen(new_name);
if (old_name_len >= new_name_len) {
/* reuse the old buffer for the name if it is large enough */
memcpy(const_cast<char*>(index->name()), new_name,
new_name_len + 1);
} else {
/* Free the old chunk of memory if it is at the topmost
place in the heap, otherwise the old chunk will be freed
when the index is evicted from the cache. This code will
kick-in in a repeated ALTER sequences where the old name is
alternately longer/shorter than the new name:
1. ALTER TABLE t RENAME INDEX a TO aa;
2. ALTER TABLE t RENAME INDEX aa TO a;
3. go to 1. */
index->name = mem_heap_strdup_replace(
index->heap,
/* Presumed topmost element of the heap: */
index->name, old_name_len + 1,
new_name);
}
DBUG_VOID_RETURN;
}
/**
Rename all indexes in data dictionary cache of a given table that are
specified in ha_alter_info.
@param ctx alter context, used to fetch the list of indexes to rename
@param ha_alter_info fetch the new names from here
*/
static
void
rename_indexes_in_cache(
/*====================*/
const ha_innobase_inplace_ctx* ctx,
const Alter_inplace_info* ha_alter_info)
{
DBUG_ENTER("rename_indexes_in_cache");
ut_ad(ctx->num_to_rename == ha_alter_info->index_rename_count);
for (ulint i = 0; i < ctx->num_to_rename; i++) {
KEY_PAIR* pair = &ha_alter_info->index_rename_buffer[i];
dict_index_t* index;
index = ctx->rename[i];
ut_ad(strcmp(index->name, pair->old_key->name) == 0);
rename_index_in_cache(index, pair->new_key->name);
}
DBUG_VOID_RETURN;
}
#endif /* MYSQL_RENAME_INDEX */
/** Fill the stored column information in s_cols list.
@param[in] altered_table mysql table object
@param[in] table innodb table object
@param[out] s_cols list of stored column
@param[out] s_heap heap for storing stored
column information. */
static
void
alter_fill_stored_column(
const TABLE* altered_table,
dict_table_t* table,
dict_s_col_list** s_cols,
mem_heap_t** s_heap)
{
ulint n_cols = altered_table->s->fields;
for (ulint i = 0; i < n_cols; i++) {
Field* field = altered_table->field[i];
dict_s_col_t s_col;
if (!innobase_is_s_fld(field)) {
continue;
}
ulint num_base = 0;
dict_col_t* col = dict_table_get_nth_col(table, i);
s_col.m_col = col;
s_col.s_pos = i;
if (*s_cols == NULL) {
*s_cols = UT_NEW_NOKEY(dict_s_col_list());
*s_heap = mem_heap_create(1000);
}
if (num_base != 0) {
s_col.base_col = static_cast<dict_col_t**>(mem_heap_zalloc(
*s_heap, num_base * sizeof(dict_col_t*)));
} else {
s_col.base_col = NULL;
}
s_col.num_base = num_base;
innodb_base_col_setup_for_stored(table, field, &s_col);
(*s_cols)->push_back(s_col);
}
}
/** Allows InnoDB to update internal structures with concurrent
writes blocked (provided that check_if_supported_inplace_alter()
did not return HA_ALTER_INPLACE_NO_LOCK).
This will be invoked before inplace_alter_table().
@param altered_table TABLE object for new version of table.
@param ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@retval true Failure
@retval false Success
*/
bool
ha_innobase::prepare_inplace_alter_table(
/*=====================================*/
TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
dict_index_t** drop_index; /*!< Index to be dropped */
ulint n_drop_index; /*!< Number of indexes to drop */
dict_index_t** rename_index; /*!< Indexes to be dropped */
ulint n_rename_index; /*!< Number of indexes to rename */
dict_foreign_t**drop_fk; /*!< Foreign key constraints to drop */
ulint n_drop_fk; /*!< Number of foreign keys to drop */
dict_foreign_t**add_fk = NULL; /*!< Foreign key constraints to drop */
ulint n_add_fk; /*!< Number of foreign keys to drop */
dict_table_t* indexed_table; /*!< Table where indexes are created */
mem_heap_t* heap;
const char** col_names;
int error;
ulint max_col_len;
ulint add_autoinc_col_no = ULINT_UNDEFINED;
ulonglong autoinc_col_max_value = 0;
ulint fts_doc_col_no = ULINT_UNDEFINED;
bool add_fts_doc_id = false;
bool add_fts_doc_id_idx = false;
bool add_fts_idx = false;
dict_s_col_list*s_cols = NULL;
mem_heap_t* s_heap = NULL;
DBUG_ENTER("prepare_inplace_alter_table");
DBUG_ASSERT(!ha_alter_info->handler_ctx);
DBUG_ASSERT(ha_alter_info->create_info);
DBUG_ASSERT(!srv_read_only_mode);
/* Init online ddl status variables */
onlineddl_rowlog_rows = 0;
onlineddl_rowlog_pct_used = 0;
onlineddl_pct_progress = 0;
MONITOR_ATOMIC_INC(MONITOR_PENDING_ALTER_TABLE);
#ifdef UNIV_DEBUG
for (dict_index_t* index = dict_table_get_first_index(m_prebuilt->table);
index;
index = dict_table_get_next_index(index)) {
ut_ad(!index->to_be_dropped);
}
#endif /* UNIV_DEBUG */
ut_d(mutex_enter(&dict_sys->mutex));
ut_d(dict_table_check_for_dup_indexes(
m_prebuilt->table, CHECK_ABORTED_OK));
ut_d(mutex_exit(&dict_sys->mutex));
if (!(ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE)) {
/* Nothing to do */
DBUG_ASSERT(m_prebuilt->trx->dict_operation_lock_mode == 0);
if (ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE) {
online_retry_drop_indexes(
m_prebuilt->table, m_user_thd);
}
DBUG_RETURN(false);
}
indexed_table = m_prebuilt->table;
/* ALTER TABLE will not implicitly move a table from a single-table
tablespace to the system tablespace when innodb_file_per_table=OFF.
But it will implicitly move a table from the system tablespace to a
single-table tablespace if innodb_file_per_table = ON. */
create_table_info_t info(m_user_thd,
altered_table,
ha_alter_info->create_info,
NULL,
NULL);
info.set_tablespace_type(indexed_table->space != TRX_SYS_SPACE);
if (ha_alter_info->handler_flags & Alter_inplace_info::ADD_INDEX) {
if (info.gcols_in_fulltext_or_spatial()) {
goto err_exit_no_heap;
}
}
if (indexed_table->is_readable()) {
} else {
if (indexed_table->corrupted) {
/* Handled below */
} else {
FilSpace space(indexed_table->space, true);
if (space()) {
String str;
const char* engine= table_type();
push_warning_printf(
m_user_thd,
Sql_condition::WARN_LEVEL_WARN,
HA_ERR_DECRYPTION_FAILED,
"Table %s in file %s is encrypted but encryption service or"
" used key_id is not available. "
" Can't continue reading table.",
table_share->table_name.str,
space()->chain.start->name);
my_error(ER_GET_ERRMSG, MYF(0), HA_ERR_DECRYPTION_FAILED, str.c_ptr(), engine);
DBUG_RETURN(true);
}
}
}
if (indexed_table->corrupted
|| dict_table_get_first_index(indexed_table) == NULL
|| dict_index_is_corrupted(
dict_table_get_first_index(indexed_table))) {
/* The clustered index is corrupted. */
my_error(ER_CHECK_NO_SUCH_TABLE, MYF(0));
DBUG_RETURN(true);
}
if (ha_alter_info->handler_flags
& Alter_inplace_info::CHANGE_CREATE_OPTION) {
const char* invalid_opt = info.create_options_are_invalid();
/* Check engine specific table options */
if (const char* invalid_tbopt = info.check_table_options()) {
my_error(ER_ILLEGAL_HA_CREATE_OPTION, MYF(0),
table_type(), invalid_tbopt);
goto err_exit_no_heap;
}
if (invalid_opt) {
my_error(ER_ILLEGAL_HA_CREATE_OPTION, MYF(0),
table_type(), invalid_opt);
goto err_exit_no_heap;
}
}
/* Check if any index name is reserved. */
if (innobase_index_name_is_reserved(
m_user_thd,
ha_alter_info->key_info_buffer,
ha_alter_info->key_count)) {
err_exit_no_heap:
DBUG_ASSERT(m_prebuilt->trx->dict_operation_lock_mode == 0);
if (ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE) {
online_retry_drop_indexes(
m_prebuilt->table, m_user_thd);
}
DBUG_RETURN(true);
}
indexed_table = m_prebuilt->table;
/* Check that index keys are sensible */
error = innobase_check_index_keys(ha_alter_info, indexed_table);
if (error) {
goto err_exit_no_heap;
}
/* Prohibit renaming a column to something that the table
already contains. */
if (ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME) {
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
for (Field** fp = table->field; *fp; fp++) {
if (!((*fp)->flags & FIELD_IS_RENAMED)) {
continue;
}
const char* name = 0;
cf_it.rewind();
while (Create_field* cf = cf_it++) {
if (cf->field == *fp) {
name = cf->field_name;
goto check_if_ok_to_rename;
}
}
ut_error;
check_if_ok_to_rename:
/* Prohibit renaming a column from FTS_DOC_ID
if full-text indexes exist. */
if (!my_strcasecmp(system_charset_info,
(*fp)->field_name,
FTS_DOC_ID_COL_NAME)
&& innobase_fulltext_exist(altered_table)) {
my_error(ER_INNODB_FT_WRONG_DOCID_COLUMN,
MYF(0), name);
goto err_exit_no_heap;
}
/* Prohibit renaming a column to an internal column. */
const char* s = m_prebuilt->table->col_names;
unsigned j;
/* Skip user columns.
MySQL should have checked these already.
We want to allow renaming of c1 to c2, c2 to c1. */
for (j = 0; j < table->s->fields; j++) {
if (!innobase_is_v_fld(table->field[j])) {
s += strlen(s) + 1;
}
}
for (; j < m_prebuilt->table->n_def; j++) {
if (!my_strcasecmp(
system_charset_info, name, s)) {
my_error(ER_WRONG_COLUMN_NAME, MYF(0),
s);
goto err_exit_no_heap;
}
s += strlen(s) + 1;
}
}
}
if (!info.innobase_table_flags()) {
goto err_exit_no_heap;
}
max_col_len = DICT_MAX_FIELD_LEN_BY_FORMAT_FLAG(info.flags());
/* Check each index's column length to make sure they do not
exceed limit */
for (ulint i = 0; i < ha_alter_info->index_add_count; i++) {
const KEY* key = &ha_alter_info->key_info_buffer[
ha_alter_info->index_add_buffer[i]];
if (key->flags & HA_FULLTEXT) {
/* The column length does not matter for
fulltext search indexes. But, UNIQUE
fulltext indexes are not supported. */
DBUG_ASSERT(!(key->flags & HA_NOSAME));
DBUG_ASSERT(!(key->flags & HA_KEYFLAG_MASK
& ~(HA_FULLTEXT
| HA_PACK_KEY
| HA_BINARY_PACK_KEY)));
add_fts_idx = true;
continue;
}
if (innobase_check_column_length(max_col_len, key)) {
my_error(ER_INDEX_COLUMN_TOO_LONG, MYF(0),
max_col_len);
goto err_exit_no_heap;
}
}
/* We won't be allowed to add fts index to a table with
fts indexes already but without AUX_HEX_NAME set.
This means the aux tables of the table failed to
rename to hex format but new created aux tables
shall be in hex format, which is contradictory. */
if (!DICT_TF2_FLAG_IS_SET(indexed_table, DICT_TF2_FTS_AUX_HEX_NAME)
&& indexed_table->fts != NULL && add_fts_idx) {
my_error(ER_INNODB_FT_AUX_NOT_HEX_ID, MYF(0));
goto err_exit_no_heap;
}
/* Check existing index definitions for too-long column
prefixes as well, in case max_col_len shrunk. */
for (const dict_index_t* index
= dict_table_get_first_index(indexed_table);
index;
index = dict_table_get_next_index(index)) {
if (index->type & DICT_FTS) {
DBUG_ASSERT(index->type == DICT_FTS
|| (index->type & DICT_CORRUPT));
/* We need to drop any corrupted fts indexes
before we add a new fts index. */
if (add_fts_idx && index->type & DICT_CORRUPT) {
ib_errf(m_user_thd, IB_LOG_LEVEL_ERROR,
ER_INNODB_INDEX_CORRUPT,
"Fulltext index '%s' is corrupt. "
"you should drop this index first.",
index->name());
goto err_exit_no_heap;
}
continue;
}
for (ulint i = 0; i < dict_index_get_n_fields(index); i++) {
const dict_field_t* field
= dict_index_get_nth_field(index, i);
if (field->prefix_len > max_col_len) {
my_error(ER_INDEX_COLUMN_TOO_LONG, MYF(0),
max_col_len);
goto err_exit_no_heap;
}
}
}
n_drop_index = 0;
n_drop_fk = 0;
if (ha_alter_info->handler_flags
& (INNOBASE_ALTER_NOREBUILD | INNOBASE_ALTER_REBUILD)) {
heap = mem_heap_create(1024);
if (ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME) {
col_names = innobase_get_col_names(
ha_alter_info, altered_table, table,
indexed_table, heap);
} else {
col_names = NULL;
}
} else {
heap = NULL;
col_names = NULL;
}
if (ha_alter_info->handler_flags
& Alter_inplace_info::DROP_FOREIGN_KEY) {
DBUG_ASSERT(ha_alter_info->alter_info->drop_list.elements > 0);
drop_fk = static_cast<dict_foreign_t**>(
mem_heap_alloc(
heap,
ha_alter_info->alter_info->drop_list.elements
* sizeof(dict_foreign_t*)));
List_iterator<Alter_drop> drop_it(
ha_alter_info->alter_info->drop_list);
while (Alter_drop* drop = drop_it++) {
if (drop->type != Alter_drop::FOREIGN_KEY) {
continue;
}
for (dict_foreign_set::iterator it
= m_prebuilt->table->foreign_set.begin();
it != m_prebuilt->table->foreign_set.end();
++it) {
dict_foreign_t* foreign = *it;
const char* fid = strchr(foreign->id, '/');
DBUG_ASSERT(fid);
/* If no database/ prefix was present in
the FOREIGN KEY constraint name, compare
to the full constraint name. */
fid = fid ? fid + 1 : foreign->id;
if (!my_strcasecmp(system_charset_info,
fid, drop->name)) {
drop_fk[n_drop_fk++] = foreign;
goto found_fk;
}
}
my_error(ER_CANT_DROP_FIELD_OR_KEY, MYF(0),
drop->type_name(), drop->name);
goto err_exit;
found_fk:
continue;
}
DBUG_ASSERT(n_drop_fk > 0);
DBUG_ASSERT(n_drop_fk
== ha_alter_info->alter_info->drop_list.elements);
} else {
drop_fk = NULL;
}
if (ha_alter_info->index_drop_count) {
dict_index_t* drop_primary = NULL;
DBUG_ASSERT(ha_alter_info->handler_flags
& (Alter_inplace_info::DROP_INDEX
| Alter_inplace_info::DROP_UNIQUE_INDEX
| Alter_inplace_info::DROP_PK_INDEX));
/* Check which indexes to drop. */
drop_index = static_cast<dict_index_t**>(
mem_heap_alloc(
heap, (ha_alter_info->index_drop_count + 1)
* sizeof *drop_index));
for (uint i = 0; i < ha_alter_info->index_drop_count; i++) {
const KEY* key
= ha_alter_info->index_drop_buffer[i];
dict_index_t* index
= dict_table_get_index_on_name(
indexed_table, key->name);
if (!index) {
push_warning_printf(
m_user_thd,
Sql_condition::WARN_LEVEL_WARN,
HA_ERR_WRONG_INDEX,
"InnoDB could not find key"
" with name %s", key->name);
} else {
ut_ad(!index->to_be_dropped);
if (!dict_index_is_clust(index)) {
drop_index[n_drop_index++] = index;
} else {
drop_primary = index;
}
}
}
/* If all FULLTEXT indexes were removed, drop an
internal FTS_DOC_ID_INDEX as well, unless it exists in
the table. */
if (innobase_fulltext_exist(table)
&& !innobase_fulltext_exist(altered_table)
&& !DICT_TF2_FLAG_IS_SET(
indexed_table, DICT_TF2_FTS_HAS_DOC_ID)) {
dict_index_t* fts_doc_index
= indexed_table->fts_doc_id_index;
ut_ad(fts_doc_index);
// Add some fault tolerance for non-debug builds.
if (fts_doc_index == NULL) {
goto check_if_can_drop_indexes;
}
DBUG_ASSERT(!fts_doc_index->to_be_dropped);
for (uint i = 0; i < table->s->keys; i++) {
if (!my_strcasecmp(
system_charset_info,
FTS_DOC_ID_INDEX_NAME,
table->key_info[i].name)) {
/* The index exists in the MySQL
data dictionary. Do not drop it,
even though it is no longer needed
by InnoDB fulltext search. */
goto check_if_can_drop_indexes;
}
}
drop_index[n_drop_index++] = fts_doc_index;
}
check_if_can_drop_indexes:
/* Check if the indexes can be dropped. */
/* Prevent a race condition between DROP INDEX and
CREATE TABLE adding FOREIGN KEY constraints. */
row_mysql_lock_data_dictionary(m_prebuilt->trx);
if (!n_drop_index) {
drop_index = NULL;
} else {
/* Flag all indexes that are to be dropped. */
for (ulint i = 0; i < n_drop_index; i++) {
ut_ad(!drop_index[i]->to_be_dropped);
drop_index[i]->to_be_dropped = 1;
}
}
if (m_prebuilt->trx->check_foreigns) {
for (uint i = 0; i < n_drop_index; i++) {
dict_index_t* index = drop_index[i];
if (innobase_check_foreign_key_index(
ha_alter_info, index,
indexed_table, col_names,
m_prebuilt->trx, drop_fk, n_drop_fk)) {
row_mysql_unlock_data_dictionary(
m_prebuilt->trx);
m_prebuilt->trx->error_info = index;
print_error(HA_ERR_DROP_INDEX_FK,
MYF(0));
goto err_exit;
}
}
/* If a primary index is dropped, need to check
any depending foreign constraints get affected */
if (drop_primary
&& innobase_check_foreign_key_index(
ha_alter_info, drop_primary,
indexed_table, col_names,
m_prebuilt->trx, drop_fk, n_drop_fk)) {
row_mysql_unlock_data_dictionary(m_prebuilt->trx);
print_error(HA_ERR_DROP_INDEX_FK, MYF(0));
goto err_exit;
}
}
row_mysql_unlock_data_dictionary(m_prebuilt->trx);
} else {
drop_index = NULL;
}
/* Check if any of the existing indexes are marked as corruption
and if they are, refuse adding more indexes. */
if (ha_alter_info->handler_flags & Alter_inplace_info::ADD_INDEX) {
for (dict_index_t* index = dict_table_get_first_index(indexed_table);
index != NULL; index = dict_table_get_next_index(index)) {
if (!index->to_be_dropped && dict_index_is_corrupted(index)) {
my_error(ER_INDEX_CORRUPT, MYF(0), index->name());
goto err_exit;
}
}
}
n_rename_index = 0;
rename_index = NULL;
#ifdef MYSQL_RENAME_INDEX
n_rename_index = ha_alter_info->index_rename_count;
/* Create a list of dict_index_t objects that are to be renamed,
also checking for requests to rename nonexistent indexes. If
the table is going to be rebuilt (new_clustered == true in
prepare_inplace_alter_table_dict()), then this can be skipped,
but we don't for simplicity (we have not determined the value of
new_clustered yet). */
if (n_rename_index > 0) {
rename_index = static_cast<dict_index_t**>(
mem_heap_alloc(
heap,
n_rename_index * sizeof(*rename_index)));
for (ulint i = 0; i < n_rename_index; i++) {
dict_index_t* index = NULL;
const char* old_name = NULL;
const char* old_name = ha_alter_info
->index_rename_buffer[i].old_key->name;
index = dict_table_get_index_on_name(indexed_table,
old_name);
if (index == NULL) {
my_error(ER_KEY_DOES_NOT_EXITS, MYF(0),
old_name,
m_prebuilt->table->name.m_name);
goto err_exit;
}
rename_index[i] = index;
}
}
#endif /* MYSQL_RENAME_INDEX */
n_add_fk = 0;
if (ha_alter_info->handler_flags
& Alter_inplace_info::ADD_FOREIGN_KEY) {
ut_ad(!m_prebuilt->trx->check_foreigns);
alter_fill_stored_column(altered_table, m_prebuilt->table,
&s_cols, &s_heap);
add_fk = static_cast<dict_foreign_t**>(
mem_heap_zalloc(
heap,
ha_alter_info->alter_info->key_list.elements
* sizeof(dict_foreign_t*)));
if (!innobase_get_foreign_key_info(
ha_alter_info, table_share,
m_prebuilt->table, col_names,
drop_index, n_drop_index,
add_fk, &n_add_fk, m_prebuilt->trx, s_cols)) {
err_exit:
if (n_drop_index) {
row_mysql_lock_data_dictionary(m_prebuilt->trx);
/* Clear the to_be_dropped flags, which might
have been set at this point. */
for (ulint i = 0; i < n_drop_index; i++) {
ut_ad(drop_index[i]->is_committed());
drop_index[i]->to_be_dropped = 0;
}
row_mysql_unlock_data_dictionary(
m_prebuilt->trx);
}
if (heap) {
mem_heap_free(heap);
}
if (s_cols != NULL) {
UT_DELETE(s_cols);
mem_heap_free(s_heap);
}
goto err_exit_no_heap;
}
if (s_cols != NULL) {
UT_DELETE(s_cols);
mem_heap_free(s_heap);
}
}
if (!(ha_alter_info->handler_flags & INNOBASE_ALTER_DATA)
|| ((ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE)
== Alter_inplace_info::CHANGE_CREATE_OPTION
&& !innobase_need_rebuild(ha_alter_info, table))) {
if (heap) {
ha_alter_info->handler_ctx
= new ha_innobase_inplace_ctx(
(*m_prebuilt_ptr),
drop_index, n_drop_index,
rename_index, n_rename_index,
drop_fk, n_drop_fk,
add_fk, n_add_fk,
ha_alter_info->online,
heap, indexed_table,
col_names, ULINT_UNDEFINED, 0, 0, 0);
}
DBUG_ASSERT(m_prebuilt->trx->dict_operation_lock_mode == 0);
if (ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE) {
online_retry_drop_indexes(
m_prebuilt->table, m_user_thd);
}
if ((ha_alter_info->handler_flags
& Alter_inplace_info::DROP_VIRTUAL_COLUMN)
&& prepare_inplace_drop_virtual(
ha_alter_info, altered_table, table)) {
DBUG_RETURN(true);
}
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ADD_VIRTUAL_COLUMN)
&& prepare_inplace_add_virtual(
ha_alter_info, altered_table, table)) {
DBUG_RETURN(true);
}
DBUG_RETURN(false);
}
/* If we are to build a full-text search index, check whether
the table already has a DOC ID column. If not, we will need to
add a Doc ID hidden column and rebuild the primary index */
if (innobase_fulltext_exist(altered_table)) {
ulint doc_col_no;
ulint num_v = 0;
if (!innobase_fts_check_doc_id_col(
m_prebuilt->table,
altered_table, &fts_doc_col_no, &num_v)) {
fts_doc_col_no = altered_table->s->fields - num_v;
add_fts_doc_id = true;
add_fts_doc_id_idx = true;
push_warning_printf(
m_user_thd,
Sql_condition::WARN_LEVEL_WARN,
HA_ERR_WRONG_INDEX,
"InnoDB rebuilding table to add"
" column " FTS_DOC_ID_COL_NAME);
} else if (fts_doc_col_no == ULINT_UNDEFINED) {
goto err_exit;
}
switch (innobase_fts_check_doc_id_index(
m_prebuilt->table, altered_table,
&doc_col_no)) {
case FTS_NOT_EXIST_DOC_ID_INDEX:
add_fts_doc_id_idx = true;
break;
case FTS_INCORRECT_DOC_ID_INDEX:
my_error(ER_INNODB_FT_WRONG_DOCID_INDEX, MYF(0),
FTS_DOC_ID_INDEX_NAME);
goto err_exit;
case FTS_EXIST_DOC_ID_INDEX:
DBUG_ASSERT(
doc_col_no == fts_doc_col_no
|| doc_col_no == ULINT_UNDEFINED
|| (ha_alter_info->handler_flags
& (Alter_inplace_info::ALTER_STORED_COLUMN_ORDER
| Alter_inplace_info::DROP_STORED_COLUMN
| Alter_inplace_info::ADD_STORED_BASE_COLUMN)));
}
}
/* See if an AUTO_INCREMENT column was added. */
uint i = 0;
ulint num_v = 0;
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
while (const Create_field* new_field = cf_it++) {
const Field* field;
DBUG_ASSERT(i < altered_table->s->fields);
for (uint old_i = 0; table->field[old_i]; old_i++) {
if (new_field->field == table->field[old_i]) {
goto found_col;
}
}
/* This is an added column. */
DBUG_ASSERT(!new_field->field);
DBUG_ASSERT(ha_alter_info->handler_flags
& Alter_inplace_info::ADD_COLUMN);
field = altered_table->field[i];
DBUG_ASSERT((MTYP_TYPENR(field->unireg_check)
== Field::NEXT_NUMBER)
== !!(field->flags & AUTO_INCREMENT_FLAG));
if (field->flags & AUTO_INCREMENT_FLAG) {
if (add_autoinc_col_no != ULINT_UNDEFINED) {
/* This should have been blocked earlier. */
ut_ad(0);
my_error(ER_WRONG_AUTO_KEY, MYF(0));
goto err_exit;
}
/* Get the col no of the old table non-virtual column array */
add_autoinc_col_no = i - num_v;
autoinc_col_max_value = innobase_get_int_col_max_value(field);
}
found_col:
if (innobase_is_v_fld(new_field)) {
++num_v;
}
i++;
}
DBUG_ASSERT(heap);
DBUG_ASSERT(m_user_thd == m_prebuilt->trx->mysql_thd);
DBUG_ASSERT(!ha_alter_info->handler_ctx);
ha_alter_info->handler_ctx = new ha_innobase_inplace_ctx(
(*m_prebuilt_ptr),
drop_index, n_drop_index,
rename_index, n_rename_index,
drop_fk, n_drop_fk, add_fk, n_add_fk,
ha_alter_info->online,
heap, m_prebuilt->table, col_names,
add_autoinc_col_no,
ha_alter_info->create_info->auto_increment_value,
autoinc_col_max_value, 0);
DBUG_RETURN(prepare_inplace_alter_table_dict(
ha_alter_info, altered_table, table,
table_share->table_name.str,
info.flags(), info.flags2(),
fts_doc_col_no, add_fts_doc_id,
add_fts_doc_id_idx));
}
/** Check that the column is part of a virtual index(index contains
virtual column) in the table
@param[in] table Table containing column
@param[in] col column to be checked
@return true if this column is indexed with other virtual columns */
static
bool
dict_col_in_v_indexes(
dict_table_t* table,
dict_col_t* col)
{
for (dict_index_t* index = dict_table_get_next_index(
dict_table_get_first_index(table)); index != NULL;
index = dict_table_get_next_index(index)) {
if (!dict_index_has_virtual(index)) {
continue;
}
for (ulint k = 0; k < index->n_fields; k++) {
dict_field_t* field
= dict_index_get_nth_field(index, k);
if (field->col->ind == col->ind) {
return(true);
}
}
}
return(false);
}
/* Check whether a columnn length change alter operation requires
to rebuild the template.
@param[in] altered_table TABLE object for new version of table.
@param[in] ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used
during in-place alter.
@param[in] table table being altered
@return TRUE if needs rebuild. */
static
bool
alter_templ_needs_rebuild(
TABLE* altered_table,
Alter_inplace_info* ha_alter_info,
dict_table_t* table)
{
ulint i = 0;
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
for (Field** fp = altered_table->field; *fp; fp++, i++) {
cf_it.rewind();
while (const Create_field* cf = cf_it++) {
for (ulint j=0; j < table->n_cols; j++) {
dict_col_t* cols
= dict_table_get_nth_col(table, j);
if (cf->length > cols->len
&& dict_col_in_v_indexes(table, cols)) {
return(true);
}
}
}
}
return(false);
}
/** Get the name of an erroneous key.
@param[in] error_key_num InnoDB number of the erroneus key
@param[in] ha_alter_info changes that were being performed
@param[in] table InnoDB table
@return the name of the erroneous key */
static
const char*
get_error_key_name(
ulint error_key_num,
const Alter_inplace_info* ha_alter_info,
const dict_table_t* table)
{
if (error_key_num == ULINT_UNDEFINED) {
return(FTS_DOC_ID_INDEX_NAME);
} else if (ha_alter_info->key_count == 0) {
return(dict_table_get_first_index(table)->name);
} else {
return(ha_alter_info->key_info_buffer[error_key_num].name);
}
}
/** Alter the table structure in-place with operations
specified using Alter_inplace_info.
The level of concurrency allowed during this operation depends
on the return value from check_if_supported_inplace_alter().
@param altered_table TABLE object for new version of table.
@param ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@retval true Failure
@retval false Success
*/
bool
ha_innobase::inplace_alter_table(
/*=============================*/
TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
dberr_t error;
dict_add_v_col_t* add_v = NULL;
dict_vcol_templ_t* s_templ = NULL;
dict_vcol_templ_t* old_templ = NULL;
struct TABLE* eval_table = altered_table;
bool rebuild_templ = false;
DBUG_ENTER("inplace_alter_table");
DBUG_ASSERT(!srv_read_only_mode);
ut_ad(!rw_lock_own(dict_operation_lock, RW_LOCK_X));
ut_ad(!rw_lock_own(dict_operation_lock, RW_LOCK_S));
DEBUG_SYNC(m_user_thd, "innodb_inplace_alter_table_enter");
if (!(ha_alter_info->handler_flags & INNOBASE_ALTER_DATA)) {
ok_exit:
DEBUG_SYNC(m_user_thd, "innodb_after_inplace_alter_table");
DBUG_RETURN(false);
}
if ((ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE)
== Alter_inplace_info::CHANGE_CREATE_OPTION
&& !innobase_need_rebuild(ha_alter_info, table)) {
goto ok_exit;
}
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
DBUG_ASSERT(ctx);
DBUG_ASSERT(ctx->trx);
DBUG_ASSERT(ctx->prebuilt == m_prebuilt);
dict_index_t* pk = dict_table_get_first_index(m_prebuilt->table);
ut_ad(pk != NULL);
/* For partitioned tables this could be already allocated from a
previous partition invocation. For normal tables this is NULL. */
UT_DELETE(ctx->m_stage);
ctx->m_stage = UT_NEW_NOKEY(ut_stage_alter_t(pk));
if (!m_prebuilt->table->is_readable()
|| dict_table_is_discarded(m_prebuilt->table)) {
goto all_done;
}
/* If we are doing a table rebuilding or having added virtual
columns in the same clause, we will need to build a table template
that carries translation information between MySQL TABLE and InnoDB
table, which indicates the virtual columns and their base columns
info. This is used to do the computation callback, so that the
data in base columns can be extracted send to server.
If the Column length changes and it is a part of virtual
index then we need to rebuild the template. */
rebuild_templ
= ctx->need_rebuild()
|| ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_EQUAL_PACK_LENGTH)
&& alter_templ_needs_rebuild(
altered_table, ha_alter_info, ctx->new_table));
if ((ctx->new_table->n_v_cols > 0) && rebuild_templ) {
/* Save the templ if isn't NULL so as to restore the
original state in case of alter operation failures. */
if (ctx->new_table->vc_templ != NULL && !ctx->need_rebuild()) {
old_templ = ctx->new_table->vc_templ;
}
s_templ = UT_NEW_NOKEY(dict_vcol_templ_t());
innobase_build_v_templ(
altered_table, ctx->new_table, s_templ, NULL, false);
ctx->new_table->vc_templ = s_templ;
} else if (ctx->num_to_add_vcol > 0 && ctx->num_to_drop_vcol == 0) {
/* if there is ongoing drop virtual column, then we disallow
inplace add index on newly added virtual column, so it does
not need to come in here to rebuild template with add_v.
Please also see the assertion in innodb_v_adjust_idx_col() */
s_templ = UT_NEW_NOKEY(dict_vcol_templ_t());
add_v = static_cast<dict_add_v_col_t*>(
mem_heap_alloc(ctx->heap, sizeof *add_v));
add_v->n_v_col = ctx->num_to_add_vcol;
add_v->v_col = ctx->add_vcol;
add_v->v_col_name = ctx->add_vcol_name;
innobase_build_v_templ(
altered_table, ctx->new_table, s_templ, add_v, false);
old_templ = ctx->new_table->vc_templ;
ctx->new_table->vc_templ = s_templ;
}
/* Drop virtual column without rebuild will keep dict table
unchanged, we use old table to evaluate virtual column value
in innobase_get_computed_value(). */
if (!ctx->need_rebuild() && ctx->num_to_drop_vcol > 0) {
eval_table = table;
}
/* Read the clustered index of the table and build
indexes based on this information using temporary
files and merge sort. */
DBUG_EXECUTE_IF("innodb_OOM_inplace_alter",
error = DB_OUT_OF_MEMORY; goto oom;);
error = row_merge_build_indexes(
m_prebuilt->trx,
m_prebuilt->table, ctx->new_table,
ctx->online,
ctx->add_index, ctx->add_key_numbers, ctx->num_to_add_index,
altered_table, ctx->add_cols, ctx->col_map,
ctx->add_autoinc, ctx->sequence, ctx->skip_pk_sort,
ctx->m_stage, add_v, eval_table);
#ifndef DBUG_OFF
oom:
#endif /* !DBUG_OFF */
if (error == DB_SUCCESS && ctx->online && ctx->need_rebuild()) {
DEBUG_SYNC_C("row_log_table_apply1_before");
error = row_log_table_apply(
ctx->thr, m_prebuilt->table, altered_table,
ctx->m_stage);
}
/* Init online ddl status variables */
onlineddl_rowlog_rows = 0;
onlineddl_rowlog_pct_used = 0;
onlineddl_pct_progress = 0;
if (s_templ) {
ut_ad(ctx->need_rebuild() || ctx->num_to_add_vcol > 0
|| rebuild_templ);
dict_free_vc_templ(s_templ);
UT_DELETE(s_templ);
ctx->new_table->vc_templ = old_templ;
}
DEBUG_SYNC_C("inplace_after_index_build");
DBUG_EXECUTE_IF("create_index_fail",
error = DB_DUPLICATE_KEY;
m_prebuilt->trx->error_key_num = ULINT_UNDEFINED;);
/* After an error, remove all those index definitions
from the dictionary which were defined. */
switch (error) {
KEY* dup_key;
all_done:
case DB_SUCCESS:
ut_d(mutex_enter(&dict_sys->mutex));
ut_d(dict_table_check_for_dup_indexes(
m_prebuilt->table, CHECK_PARTIAL_OK));
ut_d(mutex_exit(&dict_sys->mutex));
/* prebuilt->table->n_ref_count can be anything here,
given that we hold at most a shared lock on the table. */
goto ok_exit;
case DB_DUPLICATE_KEY:
if (m_prebuilt->trx->error_key_num == ULINT_UNDEFINED
|| ha_alter_info->key_count == 0) {
/* This should be the hidden index on
FTS_DOC_ID, or there is no PRIMARY KEY in the
table. Either way, we should be seeing and
reporting a bogus duplicate key error. */
dup_key = NULL;
} else {
DBUG_ASSERT(m_prebuilt->trx->error_key_num
< ha_alter_info->key_count);
dup_key = &ha_alter_info->key_info_buffer[
m_prebuilt->trx->error_key_num];
}
print_keydup_error(altered_table, dup_key, MYF(0));
break;
case DB_ONLINE_LOG_TOO_BIG:
DBUG_ASSERT(ctx->online);
my_error(ER_INNODB_ONLINE_LOG_TOO_BIG, MYF(0),
get_error_key_name(m_prebuilt->trx->error_key_num,
ha_alter_info, m_prebuilt->table));
break;
case DB_INDEX_CORRUPT:
my_error(ER_INDEX_CORRUPT, MYF(0),
get_error_key_name(m_prebuilt->trx->error_key_num,
ha_alter_info, m_prebuilt->table));
break;
case DB_DECRYPTION_FAILED: {
String str;
const char* engine= table_type();
get_error_message(HA_ERR_DECRYPTION_FAILED, &str);
my_error(ER_GET_ERRMSG, MYF(0), HA_ERR_DECRYPTION_FAILED, str.c_ptr(), engine);
break;
}
default:
my_error_innodb(error,
table_share->table_name.str,
m_prebuilt->table->flags);
}
/* prebuilt->table->n_ref_count can be anything here, given
that we hold at most a shared lock on the table. */
m_prebuilt->trx->error_info = NULL;
ctx->trx->error_state = DB_SUCCESS;
DBUG_RETURN(true);
}
/** Free the modification log for online table rebuild.
@param table table that was being rebuilt online */
static
void
innobase_online_rebuild_log_free(
/*=============================*/
dict_table_t* table)
{
dict_index_t* clust_index = dict_table_get_first_index(table);
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
rw_lock_x_lock(&clust_index->lock);
if (clust_index->online_log) {
ut_ad(dict_index_get_online_status(clust_index)
== ONLINE_INDEX_CREATION);
clust_index->online_status = ONLINE_INDEX_COMPLETE;
row_log_free(clust_index->online_log);
DEBUG_SYNC_C("innodb_online_rebuild_log_free_aborted");
}
DBUG_ASSERT(dict_index_get_online_status(clust_index)
== ONLINE_INDEX_COMPLETE);
rw_lock_x_unlock(&clust_index->lock);
}
/** For each user column, which is part of an index which is not going to be
dropped, it checks if the column number of the column is same as col_no
argument passed.
@param[in] table table
@param[in] col_no column number
@param[in] is_v if this is a virtual column
@param[in] only_committed whether to consider only committed indexes
@retval true column exists
@retval false column does not exist, true if column is system column or
it is in the index. */
static
bool
check_col_exists_in_indexes(
const dict_table_t* table,
ulint col_no,
bool is_v,
bool only_committed = false)
{
/* This function does not check system columns */
if (!is_v && dict_table_get_nth_col(table, col_no)->mtype == DATA_SYS) {
return(true);
}
for (const dict_index_t* index = dict_table_get_first_index(table);
index;
index = dict_table_get_next_index(index)) {
if (only_committed
? !index->is_committed()
: index->to_be_dropped) {
continue;
}
for (ulint i = 0; i < index->n_user_defined_cols; i++) {
const dict_col_t* idx_col
= dict_index_get_nth_col(index, i);
if (is_v && dict_col_is_virtual(idx_col)) {
const dict_v_col_t* v_col = reinterpret_cast<
const dict_v_col_t*>(idx_col);
if (v_col->v_pos == col_no) {
return(true);
}
}
if (!is_v && !dict_col_is_virtual(idx_col)
&& dict_col_get_no(idx_col) == col_no) {
return(true);
}
}
}
return(false);
}
/** Rollback a secondary index creation, drop the indexes with
temparary index prefix
@param user_table InnoDB table
@param table the TABLE
@param locked TRUE=table locked, FALSE=may need to do a lazy drop
@param trx the transaction
*/
static MY_ATTRIBUTE((nonnull))
void
innobase_rollback_sec_index(
/*========================*/
dict_table_t* user_table,
const TABLE* table,
ibool locked,
trx_t* trx)
{
row_merge_drop_indexes(trx, user_table, locked);
/* Free the table->fts only if there is no FTS_DOC_ID
in the table */
if (user_table->fts
&& !DICT_TF2_FLAG_IS_SET(user_table,
DICT_TF2_FTS_HAS_DOC_ID)
&& !innobase_fulltext_exist(table)) {
fts_free(user_table);
}
}
/** Roll back the changes made during prepare_inplace_alter_table()
and inplace_alter_table() inside the storage engine. Note that the
allowed level of concurrency during this operation will be the same as
for inplace_alter_table() and thus might be higher than during
prepare_inplace_alter_table(). (E.g concurrent writes were blocked
during prepare, but might not be during commit).
@param ha_alter_info Data used during in-place alter.
@param table the TABLE
@param prebuilt the prebuilt struct
@retval true Failure
@retval false Success
*/
inline MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
rollback_inplace_alter_table(
/*=========================*/
Alter_inplace_info* ha_alter_info,
const TABLE* table,
row_prebuilt_t* prebuilt)
{
bool fail = false;
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
DBUG_ENTER("rollback_inplace_alter_table");
if (!ctx || !ctx->trx) {
/* If we have not started a transaction yet,
(almost) nothing has been or needs to be done. */
goto func_exit;
}
trx_start_for_ddl(ctx->trx, TRX_DICT_OP_INDEX);
row_mysql_lock_data_dictionary(ctx->trx);
if (ctx->need_rebuild()) {
dberr_t err = DB_SUCCESS;
ulint flags = ctx->new_table->flags;
/* DML threads can access ctx->new_table via the
online rebuild log. Free it first. */
innobase_online_rebuild_log_free(prebuilt->table);
/* Since the FTS index specific auxiliary tables has
not yet registered with "table->fts" by fts_add_index(),
we will need explicitly delete them here */
if (dict_table_has_fts_index(ctx->new_table)) {
err = innobase_drop_fts_index_table(
ctx->new_table, ctx->trx);
if (err != DB_SUCCESS) {
my_error_innodb(
err, table->s->table_name.str,
flags);
fail = true;
}
}
dict_table_close_and_drop(ctx->trx, ctx->new_table);
switch (err) {
case DB_SUCCESS:
break;
default:
my_error_innodb(err, table->s->table_name.str,
flags);
fail = true;
}
} else {
DBUG_ASSERT(!(ha_alter_info->handler_flags
& Alter_inplace_info::ADD_PK_INDEX));
DBUG_ASSERT(ctx->new_table == prebuilt->table);
innobase_rollback_sec_index(
prebuilt->table, table, FALSE, ctx->trx);
}
trx_commit_for_mysql(ctx->trx);
row_mysql_unlock_data_dictionary(ctx->trx);
trx_free_for_mysql(ctx->trx);
func_exit:
#ifndef DBUG_OFF
dict_index_t* clust_index = dict_table_get_first_index(
prebuilt->table);
DBUG_ASSERT(!clust_index->online_log);
DBUG_ASSERT(dict_index_get_online_status(clust_index)
== ONLINE_INDEX_COMPLETE);
#endif /* !DBUG_OFF */
if (ctx) {
DBUG_ASSERT(ctx->prebuilt == prebuilt);
if (ctx->num_to_add_fk) {
for (ulint i = 0; i < ctx->num_to_add_fk; i++) {
dict_foreign_free(ctx->add_fk[i]);
}
}
if (ctx->num_to_drop_index) {
row_mysql_lock_data_dictionary(prebuilt->trx);
/* Clear the to_be_dropped flags
in the data dictionary cache.
The flags may already have been cleared,
in case an error was detected in
commit_inplace_alter_table(). */
for (ulint i = 0; i < ctx->num_to_drop_index; i++) {
dict_index_t* index = ctx->drop_index[i];
DBUG_ASSERT(index->is_committed());
index->to_be_dropped = 0;
}
row_mysql_unlock_data_dictionary(prebuilt->trx);
}
}
/* Reset dict_col_t::ord_part for those columns fail to be indexed,
we do this by checking every existing column, if any current
index would index them */
for (ulint i = 0; i < dict_table_get_n_cols(prebuilt->table); i++) {
dict_col_t& col = prebuilt->table->cols[i];
if (!col.ord_part) {
continue;
}
if (!check_col_exists_in_indexes(prebuilt->table, i, false,
true)) {
col.ord_part = 0;
}
}
for (ulint i = 0; i < dict_table_get_n_v_cols(prebuilt->table); i++) {
dict_col_t& col = prebuilt->table->v_cols[i].m_col;
if (!col.ord_part) {
continue;
}
if (!check_col_exists_in_indexes(prebuilt->table, i, true,
true)) {
col.ord_part = 0;
}
}
trx_commit_for_mysql(prebuilt->trx);
MONITOR_ATOMIC_DEC(MONITOR_PENDING_ALTER_TABLE);
DBUG_RETURN(fail);
}
/** Drop a FOREIGN KEY constraint from the data dictionary tables.
@param trx data dictionary transaction
@param table_name Table name in MySQL
@param foreign_id Foreign key constraint identifier
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_drop_foreign_try(
/*======================*/
trx_t* trx,
const char* table_name,
const char* foreign_id)
{
DBUG_ENTER("innobase_drop_foreign_try");
DBUG_ASSERT(trx_get_dict_operation(trx) == TRX_DICT_OP_INDEX);
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
/* Drop the constraint from the data dictionary. */
static const char sql[] =
"PROCEDURE DROP_FOREIGN_PROC () IS\n"
"BEGIN\n"
"DELETE FROM SYS_FOREIGN WHERE ID=:id;\n"
"DELETE FROM SYS_FOREIGN_COLS WHERE ID=:id;\n"
"END;\n";
dberr_t error;
pars_info_t* info;
info = pars_info_create();
pars_info_add_str_literal(info, "id", foreign_id);
trx->op_info = "dropping foreign key constraint from dictionary";
error = que_eval_sql(info, sql, FALSE, trx);
trx->op_info = "";
DBUG_EXECUTE_IF("ib_drop_foreign_error",
error = DB_OUT_OF_FILE_SPACE;);
if (error != DB_SUCCESS) {
my_error_innodb(error, table_name, 0);
trx->error_state = DB_SUCCESS;
DBUG_RETURN(true);
}
DBUG_RETURN(false);
}
/** Rename a column in the data dictionary tables.
@param[in] user_table InnoDB table that was being altered
@param[in] trx data dictionary transaction
@param[in] table_name Table name in MySQL
@param[in] nth_col 0-based index of the column
@param[in] from old column name
@param[in] to new column name
@param[in] new_clustered whether the table has been rebuilt
@param[in] is_virtual whether it is a virtual column
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_rename_column_try(
const dict_table_t* user_table,
trx_t* trx,
const char* table_name,
ulint nth_col,
const char* from,
const char* to,
bool new_clustered,
bool is_virtual)
{
pars_info_t* info;
dberr_t error;
DBUG_ENTER("innobase_rename_column_try");
DBUG_ASSERT(trx_get_dict_operation(trx) == TRX_DICT_OP_INDEX);
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
if (new_clustered) {
goto rename_foreign;
}
info = pars_info_create();
pars_info_add_ull_literal(info, "tableid", user_table->id);
pars_info_add_int4_literal(info, "nth", nth_col);
pars_info_add_str_literal(info, "old", from);
pars_info_add_str_literal(info, "new", to);
trx->op_info = "renaming column in SYS_COLUMNS";
error = que_eval_sql(
info,
"PROCEDURE RENAME_SYS_COLUMNS_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_COLUMNS SET NAME=:new\n"
"WHERE TABLE_ID=:tableid AND NAME=:old\n"
"AND POS=:nth;\n"
"END;\n",
FALSE, trx);
DBUG_EXECUTE_IF("ib_rename_column_error",
error = DB_OUT_OF_FILE_SPACE;);
if (error != DB_SUCCESS) {
err_exit:
my_error_innodb(error, table_name, 0);
trx->error_state = DB_SUCCESS;
trx->op_info = "";
DBUG_RETURN(true);
}
trx->op_info = "renaming column in SYS_FIELDS";
for (const dict_index_t* index = dict_table_get_first_index(
user_table);
index != NULL;
index = dict_table_get_next_index(index)) {
for (ulint i = 0; i < dict_index_get_n_fields(index); i++) {
if (strcmp(dict_index_get_nth_field(index, i)->name,
from)) {
continue;
}
info = pars_info_create();
pars_info_add_ull_literal(info, "indexid", index->id);
pars_info_add_int4_literal(info, "nth", i);
pars_info_add_str_literal(info, "old", from);
pars_info_add_str_literal(info, "new", to);
error = que_eval_sql(
info,
"PROCEDURE RENAME_SYS_FIELDS_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_FIELDS SET COL_NAME=:new\n"
"WHERE INDEX_ID=:indexid AND COL_NAME=:old\n"
"AND POS=:nth;\n"
/* Try again, in case there is a prefix_len
encoded in SYS_FIELDS.POS */
"UPDATE SYS_FIELDS SET COL_NAME=:new\n"
"WHERE INDEX_ID=:indexid AND COL_NAME=:old\n"
"AND POS>=65536*:nth AND POS<65536*(:nth+1);\n"
"END;\n",
FALSE, trx);
if (error != DB_SUCCESS) {
goto err_exit;
}
}
}
rename_foreign:
trx->op_info = "renaming column in SYS_FOREIGN_COLS";
std::list<dict_foreign_t*> fk_evict;
bool foreign_modified;
for (dict_foreign_set::const_iterator it = user_table->foreign_set.begin();
it != user_table->foreign_set.end();
++it) {
dict_foreign_t* foreign = *it;
foreign_modified = false;
for (unsigned i = 0; i < foreign->n_fields; i++) {
if (strcmp(foreign->foreign_col_names[i], from)) {
continue;
}
info = pars_info_create();
pars_info_add_str_literal(info, "id", foreign->id);
pars_info_add_int4_literal(info, "nth", i);
pars_info_add_str_literal(info, "old", from);
pars_info_add_str_literal(info, "new", to);
error = que_eval_sql(
info,
"PROCEDURE RENAME_SYS_FOREIGN_F_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_FOREIGN_COLS\n"
"SET FOR_COL_NAME=:new\n"
"WHERE ID=:id AND POS=:nth\n"
"AND FOR_COL_NAME=:old;\n"
"END;\n",
FALSE, trx);
if (error != DB_SUCCESS) {
goto err_exit;
}
foreign_modified = true;
}
if (foreign_modified) {
fk_evict.push_back(foreign);
}
}
for (dict_foreign_set::const_iterator it
= user_table->referenced_set.begin();
it != user_table->referenced_set.end();
++it) {
foreign_modified = false;
dict_foreign_t* foreign = *it;
for (unsigned i = 0; i < foreign->n_fields; i++) {
if (strcmp(foreign->referenced_col_names[i], from)) {
continue;
}
info = pars_info_create();
pars_info_add_str_literal(info, "id", foreign->id);
pars_info_add_int4_literal(info, "nth", i);
pars_info_add_str_literal(info, "old", from);
pars_info_add_str_literal(info, "new", to);
error = que_eval_sql(
info,
"PROCEDURE RENAME_SYS_FOREIGN_R_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_FOREIGN_COLS\n"
"SET REF_COL_NAME=:new\n"
"WHERE ID=:id AND POS=:nth\n"
"AND REF_COL_NAME=:old;\n"
"END;\n",
FALSE, trx);
if (error != DB_SUCCESS) {
goto err_exit;
}
foreign_modified = true;
}
if (foreign_modified) {
fk_evict.push_back(foreign);
}
}
if (new_clustered) {
std::for_each(fk_evict.begin(), fk_evict.end(),
dict_foreign_remove_from_cache);
}
trx->op_info = "";
DBUG_RETURN(false);
}
/** Rename columns in the data dictionary tables.
@param ha_alter_info Data used during in-place alter.
@param ctx In-place ALTER TABLE context
@param table the TABLE
@param trx data dictionary transaction
@param table_name Table name in MySQL
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_rename_columns_try(
/*========================*/
Alter_inplace_info* ha_alter_info,
ha_innobase_inplace_ctx*ctx,
const TABLE* table,
trx_t* trx,
const char* table_name)
{
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
uint i = 0;
ulint num_v = 0;
DBUG_ASSERT(ctx);
DBUG_ASSERT(ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME);
for (Field** fp = table->field; *fp; fp++, i++) {
bool is_virtual = innobase_is_v_fld(*fp);
if (!((*fp)->flags & FIELD_IS_RENAMED)) {
goto processed_field;
}
cf_it.rewind();
while (Create_field* cf = cf_it++) {
if (cf->field == *fp) {
ulint col_n = is_virtual
? dict_create_v_col_pos(
num_v, i)
: i - num_v;
if (innobase_rename_column_try(
ctx->old_table, trx, table_name,
col_n,
cf->field->field_name,
cf->field_name,
ctx->need_rebuild(),
is_virtual)) {
return(true);
}
goto processed_field;
}
}
ut_error;
processed_field:
if (is_virtual) {
num_v++;
}
continue;
}
return(false);
}
/** Enlarge a column in the data dictionary tables.
@param user_table InnoDB table that was being altered
@param trx data dictionary transaction
@param table_name Table name in MySQL
@param nth_col 0-based index of the column
@param new_len new column length, in bytes
@param is_v if it's a virtual column
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_enlarge_column_try(
/*========================*/
const dict_table_t* user_table,
trx_t* trx,
const char* table_name,
ulint nth_col,
ulint new_len,
bool is_v)
{
pars_info_t* info;
dberr_t error;
#ifdef UNIV_DEBUG
dict_col_t* col;
#endif /* UNIV_DEBUG */
dict_v_col_t* v_col;
ulint pos;
DBUG_ENTER("innobase_enlarge_column_try");
DBUG_ASSERT(trx_get_dict_operation(trx) == TRX_DICT_OP_INDEX);
ut_ad(trx->dict_operation_lock_mode == RW_X_LATCH);
ut_ad(mutex_own(&dict_sys->mutex));
ut_ad(rw_lock_own(dict_operation_lock, RW_LOCK_X));
if (is_v) {
v_col = dict_table_get_nth_v_col(user_table, nth_col);
pos = dict_create_v_col_pos(v_col->v_pos, v_col->m_col.ind);
#ifdef UNIV_DEBUG
col = &v_col->m_col;
#endif /* UNIV_DEBUG */
} else {
#ifdef UNIV_DEBUG
col = dict_table_get_nth_col(user_table, nth_col);
#endif /* UNIV_DEBUG */
pos = nth_col;
}
#ifdef UNIV_DEBUG
ut_ad(col->len < new_len);
switch (col->mtype) {
case DATA_MYSQL:
/* NOTE: we could allow this when !(prtype & DATA_BINARY_TYPE)
and ROW_FORMAT is not REDUNDANT and mbminlen<mbmaxlen.
That is, we treat a UTF-8 CHAR(n) column somewhat like
a VARCHAR. */
ut_error;
case DATA_BINARY:
case DATA_VARCHAR:
case DATA_VARMYSQL:
case DATA_DECIMAL:
case DATA_BLOB:
break;
default:
ut_error;
}
#endif /* UNIV_DEBUG */
info = pars_info_create();
pars_info_add_ull_literal(info, "tableid", user_table->id);
pars_info_add_int4_literal(info, "nth", pos);
pars_info_add_int4_literal(info, "new", new_len);
trx->op_info = "resizing column in SYS_COLUMNS";
error = que_eval_sql(
info,
"PROCEDURE RESIZE_SYS_COLUMNS_PROC () IS\n"
"BEGIN\n"
"UPDATE SYS_COLUMNS SET LEN=:new\n"
"WHERE TABLE_ID=:tableid AND POS=:nth;\n"
"END;\n",
FALSE, trx);
DBUG_EXECUTE_IF("ib_resize_column_error",
error = DB_OUT_OF_FILE_SPACE;);
trx->op_info = "";
trx->error_state = DB_SUCCESS;
if (error != DB_SUCCESS) {
my_error_innodb(error, table_name, 0);
DBUG_RETURN(true);
}
DBUG_RETURN(false);
}
/** Enlarge columns in the data dictionary tables.
@param ha_alter_info Data used during in-place alter.
@param table the TABLE
@param user_table InnoDB table that was being altered
@param trx data dictionary transaction
@param table_name Table name in MySQL
@retval true Failure
@retval false Success */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_enlarge_columns_try(
/*=========================*/
Alter_inplace_info* ha_alter_info,
const TABLE* table,
const dict_table_t* user_table,
trx_t* trx,
const char* table_name)
{
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
ulint i = 0;
ulint num_v = 0;
bool is_v;
for (Field** fp = table->field; *fp; fp++, i++) {
ulint idx;
if (innobase_is_v_fld(*fp)) {
is_v = true;
idx = num_v;
num_v++;
} else {
idx = i - num_v;
is_v = false;
}
cf_it.rewind();
while (Create_field* cf = cf_it++) {
if (cf->field == *fp) {
if ((*fp)->is_equal(cf)
== IS_EQUAL_PACK_LENGTH
&& innobase_enlarge_column_try(
user_table, trx, table_name,
idx, static_cast<ulint>(cf->length), is_v)) {
return(true);
}
break;
}
}
}
return(false);
}
/** Rename or enlarge columns in the data dictionary cache
as part of commit_cache_norebuild().
@param ha_alter_info Data used during in-place alter.
@param table the TABLE
@param user_table InnoDB table that was being altered */
static MY_ATTRIBUTE((nonnull))
void
innobase_rename_or_enlarge_columns_cache(
/*=====================================*/
Alter_inplace_info* ha_alter_info,
const TABLE* table,
dict_table_t* user_table)
{
if (!(ha_alter_info->handler_flags
& (Alter_inplace_info::ALTER_COLUMN_EQUAL_PACK_LENGTH
| Alter_inplace_info::ALTER_COLUMN_NAME))) {
return;
}
List_iterator_fast<Create_field> cf_it(
ha_alter_info->alter_info->create_list);
uint i = 0;
ulint num_v = 0;
for (Field** fp = table->field; *fp; fp++, i++) {
bool is_virtual = innobase_is_v_fld(*fp);
cf_it.rewind();
while (Create_field* cf = cf_it++) {
if (cf->field != *fp) {
continue;
}
ulint col_n = is_virtual ? num_v : i - num_v;
if ((*fp)->is_equal(cf) == IS_EQUAL_PACK_LENGTH) {
if (is_virtual) {
dict_table_get_nth_v_col(
user_table, col_n)->m_col.len
= cf->length;
} else {
dict_table_get_nth_col(
user_table, col_n)->len
= cf->length;
}
}
if ((*fp)->flags & FIELD_IS_RENAMED) {
dict_mem_table_col_rename(
user_table, col_n,
cf->field->field_name,
cf->field_name, is_virtual);
}
break;
}
if (is_virtual) {
num_v++;
}
}
}
/** Set the auto-increment value of the table on commit.
@param ha_alter_info Data used during in-place alter
@param ctx In-place ALTER TABLE context
@param altered_table MySQL table that is being altered
@param old_table MySQL table as it is before the ALTER operation */
static MY_ATTRIBUTE((nonnull))
void
commit_set_autoinc(
Alter_inplace_info* ha_alter_info,
ha_innobase_inplace_ctx*ctx,
const TABLE* altered_table,
const TABLE* old_table)
{
DBUG_ENTER("commit_get_autoinc");
if (!altered_table->found_next_number_field) {
/* There is no AUTO_INCREMENT column in the table
after the ALTER operation. */
} else if (ctx->add_autoinc != ULINT_UNDEFINED) {
ut_ad(ctx->need_rebuild());
/* An AUTO_INCREMENT column was added. Get the last
value from the sequence, which may be based on a
supplied AUTO_INCREMENT value. */
ib_uint64_t autoinc = ctx->sequence.last();
ctx->new_table->autoinc = autoinc;
/* Bulk index creation does not update
PAGE_ROOT_AUTO_INC, so we must persist the "last used"
value here. */
btr_write_autoinc(dict_table_get_first_index(ctx->new_table),
autoinc - 1, true);
} else if ((ha_alter_info->handler_flags
& Alter_inplace_info::CHANGE_CREATE_OPTION)
&& (ha_alter_info->create_info->used_fields
& HA_CREATE_USED_AUTO)) {
/* An AUTO_INCREMENT value was supplied by the user.
It must be persisted to the data file. */
const Field* ai = old_table->found_next_number_field;
ut_ad(!strcmp(dict_table_get_col_name(ctx->old_table,
innodb_col_no(ai)),
ai->field_name));
ib_uint64_t autoinc
= ha_alter_info->create_info->auto_increment_value;
if (autoinc == 0) {
autoinc = 1;
}
if (autoinc >= ctx->old_table->autoinc) {
/* Persist the predecessor of the
AUTO_INCREMENT value as the last used one. */
ctx->new_table->autoinc = autoinc--;
} else {
/* Mimic ALGORITHM=COPY in the following scenario:
CREATE TABLE t (a SERIAL);
INSERT INTO t SET a=100;
ALTER TABLE t AUTO_INCREMENT = 1;
INSERT INTO t SET a=NULL;
SELECT * FROM t;
By default, ALGORITHM=INPLACE would reset the
sequence to 1, while after ALGORITHM=COPY, the
last INSERT would use a value larger than 100.
We could only search the tree to know current
max counter in the table and compare. */
const dict_col_t* autoinc_col
= dict_table_get_nth_col(ctx->old_table,
innodb_col_no(ai));
dict_index_t* index
= dict_table_get_first_index(ctx->old_table);
while (index != NULL
&& index->fields[0].col != autoinc_col) {
index = dict_table_get_next_index(index);
}
ut_ad(index);
ib_uint64_t max_in_table = index
? row_search_max_autoinc(index)
: 0;
if (autoinc <= max_in_table) {
ctx->new_table->autoinc = innobase_next_autoinc(
max_in_table, 1,
ctx->prebuilt->autoinc_increment,
ctx->prebuilt->autoinc_offset,
innobase_get_int_col_max_value(ai));
/* Persist the maximum value as the
last used one. */
autoinc = max_in_table;
} else {
/* Persist the predecessor of the
AUTO_INCREMENT value as the last used one. */
ctx->new_table->autoinc = autoinc--;
}
}
btr_write_autoinc(dict_table_get_first_index(ctx->new_table),
autoinc, true);
} else if (ctx->need_rebuild()) {
/* No AUTO_INCREMENT value was specified.
Copy it from the old table. */
ctx->new_table->autoinc = ctx->old_table->autoinc;
/* The persistent value was already copied in
prepare_inplace_alter_table_dict() when ctx->new_table
was created. If this was a LOCK=NONE operation, the
AUTO_INCREMENT values would be updated during
row_log_table_apply(). If this was LOCK!=NONE,
the table contents could not possibly have changed
between prepare_inplace and commit_inplace. */
}
DBUG_VOID_RETURN;
}
/** Add or drop foreign key constraints to the data dictionary tables,
but do not touch the data dictionary cache.
@param ha_alter_info Data used during in-place alter
@param ctx In-place ALTER TABLE context
@param trx Data dictionary transaction
@param table_name Table name in MySQL
@retval true Failure
@retval false Success
*/
static MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
innobase_update_foreign_try(
/*========================*/
ha_innobase_inplace_ctx*ctx,
trx_t* trx,
const char* table_name)
{
ulint foreign_id;
ulint i;
DBUG_ENTER("innobase_update_foreign_try");
DBUG_ASSERT(ctx);
foreign_id = dict_table_get_highest_foreign_id(ctx->new_table);
foreign_id++;
for (i = 0; i < ctx->num_to_add_fk; i++) {
dict_foreign_t* fk = ctx->add_fk[i];
ut_ad(fk->foreign_table == ctx->new_table
|| fk->foreign_table == ctx->old_table);
dberr_t error = dict_create_add_foreign_id(
&foreign_id, ctx->old_table->name.m_name, fk);
if (error != DB_SUCCESS) {
my_error(ER_TOO_LONG_IDENT, MYF(0),
fk->id);
DBUG_RETURN(true);
}
if (!fk->foreign_index) {
fk->foreign_index = dict_foreign_find_index(
ctx->new_table, ctx->col_names,
fk->foreign_col_names,
fk->n_fields, fk->referenced_index, TRUE,
fk->type
& (DICT_FOREIGN_ON_DELETE_SET_NULL
| DICT_FOREIGN_ON_UPDATE_SET_NULL),
NULL, NULL, NULL);
if (!fk->foreign_index) {
my_error(ER_FK_INCORRECT_OPTION,
MYF(0), table_name, fk->id);
DBUG_RETURN(true);
}
}
/* The fk->foreign_col_names[] uses renamed column
names, while the columns in ctx->old_table have not
been renamed yet. */
error = dict_create_add_foreign_to_dictionary(
ctx->old_table->name.m_name, fk, trx);
DBUG_EXECUTE_IF(
"innodb_test_cannot_add_fk_system",
error = DB_ERROR;);
if (error != DB_SUCCESS) {
my_error(ER_FK_FAIL_ADD_SYSTEM, MYF(0),
fk->id);
DBUG_RETURN(true);
}
}
for (i = 0; i < ctx->num_to_drop_fk; i++) {
dict_foreign_t* fk = ctx->drop_fk[i];
DBUG_ASSERT(fk->foreign_table == ctx->old_table);
if (innobase_drop_foreign_try(trx, table_name, fk->id)) {
DBUG_RETURN(true);
}
}
DBUG_RETURN(false);
}
/** Update the foreign key constraint definitions in the data dictionary cache
after the changes to data dictionary tables were committed.
@param ctx In-place ALTER TABLE context
@param user_thd MySQL connection
@return InnoDB error code (should always be DB_SUCCESS) */
static MY_ATTRIBUTE((nonnull, warn_unused_result))
dberr_t
innobase_update_foreign_cache(
/*==========================*/
ha_innobase_inplace_ctx* ctx,
THD* user_thd)
{
dict_table_t* user_table;
dberr_t err = DB_SUCCESS;
DBUG_ENTER("innobase_update_foreign_cache");
ut_ad(mutex_own(&dict_sys->mutex));
user_table = ctx->old_table;
/* Discard the added foreign keys, because we will
load them from the data dictionary. */
for (ulint i = 0; i < ctx->num_to_add_fk; i++) {
dict_foreign_t* fk = ctx->add_fk[i];
dict_foreign_free(fk);
}
if (ctx->need_rebuild()) {
/* The rebuilt table is already using the renamed
column names. No need to pass col_names or to drop
constraints from the data dictionary cache. */
DBUG_ASSERT(!ctx->col_names);
DBUG_ASSERT(user_table->foreign_set.empty());
DBUG_ASSERT(user_table->referenced_set.empty());
user_table = ctx->new_table;
} else {
/* Drop the foreign key constraints if the
table was not rebuilt. If the table is rebuilt,
there would not be any foreign key contraints for
it yet in the data dictionary cache. */
for (ulint i = 0; i < ctx->num_to_drop_fk; i++) {
dict_foreign_t* fk = ctx->drop_fk[i];
dict_foreign_remove_from_cache(fk);
}
}
/* Load the old or added foreign keys from the data dictionary
and prevent the table from being evicted from the data
dictionary cache (work around the lack of WL#6049). */
dict_names_t fk_tables;
err = dict_load_foreigns(user_table->name.m_name,
ctx->col_names, false, true,
DICT_ERR_IGNORE_NONE,
fk_tables);
if (err == DB_CANNOT_ADD_CONSTRAINT) {
fk_tables.clear();
/* It is possible there are existing foreign key are
loaded with "foreign_key checks" off,
so let's retry the loading with charset_check is off */
err = dict_load_foreigns(user_table->name.m_name,
ctx->col_names, false, false,
DICT_ERR_IGNORE_NONE,
fk_tables);
/* The load with "charset_check" off is successful, warn
the user that the foreign key has loaded with mis-matched
charset */
if (err == DB_SUCCESS) {
push_warning_printf(
user_thd,
Sql_condition::WARN_LEVEL_WARN,
ER_ALTER_INFO,
"Foreign key constraints for table '%s'"
" are loaded with charset check off",
user_table->name.m_name);
}
}
/* For complete loading of foreign keys, all associated tables must
also be loaded. */
while (err == DB_SUCCESS && !fk_tables.empty()) {
dict_table_t* table = dict_load_table(
fk_tables.front(), true, DICT_ERR_IGNORE_NONE);
if (table == NULL) {
table_name_t table_name;
table_name.m_name = const_cast<char*>(
fk_tables.front());
err = DB_TABLE_NOT_FOUND;
ib::error()
<< "Failed to load table '" << table_name
<< "' which has a foreign key constraint with"
<< " table '" << user_table->name << "'.";
break;
}
fk_tables.pop_front();
}
DBUG_RETURN(err);
}
/** Commit the changes made during prepare_inplace_alter_table()
and inplace_alter_table() inside the data dictionary tables,
when rebuilding the table.
@param ha_alter_info Data used during in-place alter
@param ctx In-place ALTER TABLE context
@param altered_table MySQL table that is being altered
@param old_table MySQL table as it is before the ALTER operation
@param trx Data dictionary transaction
@param table_name Table name in MySQL
@retval true Failure
@retval false Success
*/
inline MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
commit_try_rebuild(
/*===============*/
Alter_inplace_info* ha_alter_info,
ha_innobase_inplace_ctx*ctx,
TABLE* altered_table,
const TABLE* old_table,
trx_t* trx,
const char* table_name)
{
dict_table_t* rebuilt_table = ctx->new_table;
dict_table_t* user_table = ctx->old_table;
DBUG_ENTER("commit_try_rebuild");
DBUG_ASSERT(ctx->need_rebuild());
DBUG_ASSERT(trx->dict_operation_lock_mode == RW_X_LATCH);
DBUG_ASSERT(!(ha_alter_info->handler_flags
& Alter_inplace_info::DROP_FOREIGN_KEY)
|| ctx->num_to_drop_fk > 0);
for (dict_index_t* index = dict_table_get_first_index(rebuilt_table);
index;
index = dict_table_get_next_index(index)) {
DBUG_ASSERT(dict_index_get_online_status(index)
== ONLINE_INDEX_COMPLETE);
DBUG_ASSERT(index->is_committed());
if (dict_index_is_corrupted(index)) {
my_error(ER_INDEX_CORRUPT, MYF(0), index->name());
DBUG_RETURN(true);
}
}
if (innobase_update_foreign_try(ctx, trx, table_name)) {
DBUG_RETURN(true);
}
dberr_t error;
/* Clear the to_be_dropped flag in the data dictionary cache
of user_table. */
for (ulint i = 0; i < ctx->num_to_drop_index; i++) {
dict_index_t* index = ctx->drop_index[i];
DBUG_ASSERT(index->table == user_table);
DBUG_ASSERT(index->is_committed());
DBUG_ASSERT(index->to_be_dropped);
index->to_be_dropped = 0;
}
/* We copied the table. Any indexes that were requested to be
dropped were not created in the copy of the table. Apply any
last bit of the rebuild log and then rename the tables. */
if (ctx->online) {
DEBUG_SYNC_C("row_log_table_apply2_before");
dict_vcol_templ_t* s_templ = NULL;
if (ctx->new_table->n_v_cols > 0) {
s_templ = UT_NEW_NOKEY(
dict_vcol_templ_t());
s_templ->vtempl = NULL;
innobase_build_v_templ(
altered_table, ctx->new_table, s_templ,
NULL, true);
ctx->new_table->vc_templ = s_templ;
}
error = row_log_table_apply(
ctx->thr, user_table, altered_table,
static_cast<ha_innobase_inplace_ctx*>(
ha_alter_info->handler_ctx)->m_stage);
if (s_templ) {
ut_ad(ctx->need_rebuild());
dict_free_vc_templ(s_templ);
UT_DELETE(s_templ);
ctx->new_table->vc_templ = NULL;
}
ulint err_key = thr_get_trx(ctx->thr)->error_key_num;
switch (error) {
KEY* dup_key;
case DB_SUCCESS:
break;
case DB_DUPLICATE_KEY:
if (err_key == ULINT_UNDEFINED) {
/* This should be the hidden index on
FTS_DOC_ID. */
dup_key = NULL;
} else {
DBUG_ASSERT(err_key <
ha_alter_info->key_count);
dup_key = &ha_alter_info
->key_info_buffer[err_key];
}
print_keydup_error(altered_table, dup_key, MYF(0));
DBUG_RETURN(true);
case DB_ONLINE_LOG_TOO_BIG:
my_error(ER_INNODB_ONLINE_LOG_TOO_BIG, MYF(0),
get_error_key_name(err_key, ha_alter_info,
rebuilt_table));
DBUG_RETURN(true);
case DB_INDEX_CORRUPT:
my_error(ER_INDEX_CORRUPT, MYF(0),
get_error_key_name(err_key, ha_alter_info,
rebuilt_table));
DBUG_RETURN(true);
default:
my_error_innodb(error, table_name, user_table->flags);
DBUG_RETURN(true);
}
}
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME)
&& innobase_rename_columns_try(ha_alter_info, ctx, old_table,
trx, table_name)) {
DBUG_RETURN(true);
}
DBUG_EXECUTE_IF("ib_ddl_crash_before_rename", DBUG_SUICIDE(););
/* The new table must inherit the flag from the
"parent" table. */
if (dict_table_is_discarded(user_table)) {
rebuilt_table->file_unreadable = true;
rebuilt_table->flags2 |= DICT_TF2_DISCARDED;
}
/* We can now rename the old table as a temporary table,
rename the new temporary table as the old table and drop the
old table. First, we only do this in the data dictionary
tables. The actual renaming will be performed in
commit_cache_rebuild(), once the data dictionary transaction
has been successfully committed. */
error = row_merge_rename_tables_dict(
user_table, rebuilt_table, ctx->tmp_name, trx);
/* We must be still holding a table handle. */
DBUG_ASSERT(user_table->get_ref_count() == 1);
DBUG_EXECUTE_IF("ib_ddl_crash_after_rename", DBUG_SUICIDE(););
DBUG_EXECUTE_IF("ib_rebuild_cannot_rename", error = DB_ERROR;);
switch (error) {
case DB_SUCCESS:
DBUG_RETURN(false);
case DB_TABLESPACE_EXISTS:
ut_a(rebuilt_table->get_ref_count() == 1);
my_error(ER_TABLESPACE_EXISTS, MYF(0), ctx->tmp_name);
DBUG_RETURN(true);
case DB_DUPLICATE_KEY:
ut_a(rebuilt_table->get_ref_count() == 1);
my_error(ER_TABLE_EXISTS_ERROR, MYF(0), ctx->tmp_name);
DBUG_RETURN(true);
default:
my_error_innodb(error, table_name, user_table->flags);
DBUG_RETURN(true);
}
}
/** Apply the changes made during commit_try_rebuild(),
to the data dictionary cache and the file system.
@param ctx In-place ALTER TABLE context */
inline MY_ATTRIBUTE((nonnull))
void
commit_cache_rebuild(
/*=================*/
ha_innobase_inplace_ctx* ctx)
{
dberr_t error;
DBUG_ENTER("commit_cache_rebuild");
DEBUG_SYNC_C("commit_cache_rebuild");
DBUG_ASSERT(ctx->need_rebuild());
DBUG_ASSERT(dict_table_is_discarded(ctx->old_table)
== dict_table_is_discarded(ctx->new_table));
const char* old_name = mem_heap_strdup(
ctx->heap, ctx->old_table->name.m_name);
/* We already committed and redo logged the renames,
so this must succeed. */
error = dict_table_rename_in_cache(
ctx->old_table, ctx->tmp_name, FALSE);
ut_a(error == DB_SUCCESS);
error = dict_table_rename_in_cache(
ctx->new_table, old_name, FALSE);
ut_a(error == DB_SUCCESS);
DBUG_VOID_RETURN;
}
/** Set of column numbers */
typedef std::set<ulint, std::less<ulint>, ut_allocator<ulint> > col_set;
/** Store the column number of the columns in a list belonging
to indexes which are not being dropped.
@param[in] ctx In-place ALTER TABLE context
@param[in, out] drop_col_list list which will be set, containing columns
which is part of index being dropped
@param[in, out] drop_v_col_list list which will be set, containing
virtual columns which is part of index
being dropped */
static
void
get_col_list_to_be_dropped(
const ha_innobase_inplace_ctx* ctx,
col_set& drop_col_list,
col_set& drop_v_col_list)
{
for (ulint index_count = 0; index_count < ctx->num_to_drop_index;
index_count++) {
const dict_index_t* index = ctx->drop_index[index_count];
for (ulint col = 0; col < index->n_user_defined_cols; col++) {
const dict_col_t* idx_col
= dict_index_get_nth_col(index, col);
if (dict_col_is_virtual(idx_col)) {
const dict_v_col_t* v_col
= reinterpret_cast<
const dict_v_col_t*>(idx_col);
drop_v_col_list.insert(v_col->v_pos);
} else {
ulint col_no = dict_col_get_no(idx_col);
drop_col_list.insert(col_no);
}
}
}
}
/** Commit the changes made during prepare_inplace_alter_table()
and inplace_alter_table() inside the data dictionary tables,
when not rebuilding the table.
@param ha_alter_info Data used during in-place alter
@param ctx In-place ALTER TABLE context
@param old_table MySQL table as it is before the ALTER operation
@param trx Data dictionary transaction
@param table_name Table name in MySQL
@retval true Failure
@retval false Success
*/
inline MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
commit_try_norebuild(
/*=================*/
Alter_inplace_info* ha_alter_info,
ha_innobase_inplace_ctx*ctx,
TABLE* altered_table,
const TABLE* old_table,
trx_t* trx,
const char* table_name)
{
DBUG_ENTER("commit_try_norebuild");
DBUG_ASSERT(!ctx->need_rebuild());
DBUG_ASSERT(trx->dict_operation_lock_mode == RW_X_LATCH);
DBUG_ASSERT(!(ha_alter_info->handler_flags
& Alter_inplace_info::DROP_FOREIGN_KEY)
|| ctx->num_to_drop_fk > 0);
DBUG_ASSERT(ctx->num_to_drop_fk
== ha_alter_info->alter_info->drop_list.elements
|| ctx->num_to_drop_vcol
== ha_alter_info->alter_info->drop_list.elements);
for (ulint i = 0; i < ctx->num_to_add_index; i++) {
dict_index_t* index = ctx->add_index[i];
DBUG_ASSERT(dict_index_get_online_status(index)
== ONLINE_INDEX_COMPLETE);
DBUG_ASSERT(!index->is_committed());
if (dict_index_is_corrupted(index)) {
/* Report a duplicate key
error for the index that was
flagged corrupted, most likely
because a duplicate value was
inserted (directly or by
rollback) after
ha_innobase::inplace_alter_table()
completed.
TODO: report this as a corruption
with a detailed reason once
WL#6379 has been implemented. */
my_error(ER_DUP_UNKNOWN_IN_INDEX,
MYF(0), index->name());
DBUG_RETURN(true);
}
}
if (innobase_update_foreign_try(ctx, trx, table_name)) {
DBUG_RETURN(true);
}
dberr_t error;
/* We altered the table in place. Mark the indexes as committed. */
for (ulint i = 0; i < ctx->num_to_add_index; i++) {
dict_index_t* index = ctx->add_index[i];
DBUG_ASSERT(dict_index_get_online_status(index)
== ONLINE_INDEX_COMPLETE);
DBUG_ASSERT(!index->is_committed());
error = row_merge_rename_index_to_add(
trx, ctx->new_table->id, index->id);
switch (error) {
case DB_SUCCESS:
break;
case DB_TOO_MANY_CONCURRENT_TRXS:
/* If we wrote some undo log here, then the
persistent data dictionary for this table may
probably be corrupted. This is because a
'trigger' on SYS_INDEXES could already have invoked
btr_free_if_exists(), which cannot be rolled back. */
DBUG_ASSERT(trx->undo_no == 0);
my_error(ER_TOO_MANY_CONCURRENT_TRXS, MYF(0));
DBUG_RETURN(true);
default:
sql_print_error(
"InnoDB: rename index to add: %lu\n",
(ulong) error);
DBUG_ASSERT(0);
my_error(ER_INTERNAL_ERROR, MYF(0),
"rename index to add");
DBUG_RETURN(true);
}
}
/* Drop any indexes that were requested to be dropped.
Flag them in the data dictionary first. */
for (ulint i = 0; i < ctx->num_to_drop_index; i++) {
dict_index_t* index = ctx->drop_index[i];
DBUG_ASSERT(index->is_committed());
DBUG_ASSERT(index->table == ctx->new_table);
DBUG_ASSERT(index->to_be_dropped);
error = row_merge_rename_index_to_drop(
trx, index->table->id, index->id);
if (error != DB_SUCCESS) {
sql_print_error(
"InnoDB: rename index to drop: %lu\n",
(ulong) error);
DBUG_ASSERT(0);
my_error(ER_INTERNAL_ERROR, MYF(0),
"rename index to drop");
DBUG_RETURN(true);
}
}
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_NAME)
&& innobase_rename_columns_try(ha_alter_info, ctx, old_table,
trx, table_name)) {
DBUG_RETURN(true);
}
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_COLUMN_EQUAL_PACK_LENGTH)
&& innobase_enlarge_columns_try(ha_alter_info, old_table,
ctx->old_table, trx, table_name)) {
DBUG_RETURN(true);
}
#ifdef MYSQL_RENAME_INDEX
if ((ha_alter_info->handler_flags
& Alter_inplace_info::RENAME_INDEX)
&& rename_indexes_in_data_dictionary(ctx, ha_alter_info, trx)) {
DBUG_RETURN(true);
}
#endif /* MYSQL_RENAME_INDEX */
if ((ha_alter_info->handler_flags
& Alter_inplace_info::DROP_VIRTUAL_COLUMN)
&& innobase_drop_virtual_try(
ha_alter_info, altered_table, old_table,
ctx->old_table, trx)) {
DBUG_RETURN(true);
}
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ADD_VIRTUAL_COLUMN)
&& innobase_add_virtual_try(
ha_alter_info, altered_table, old_table,
ctx->old_table, trx)) {
DBUG_RETURN(true);
}
DBUG_RETURN(false);
}
/** Commit the changes to the data dictionary cache
after a successful commit_try_norebuild() call.
@param ctx In-place ALTER TABLE context
@param table the TABLE before the ALTER
@param trx Data dictionary transaction object
(will be started and committed)
@return whether all replacements were found for dropped indexes */
inline MY_ATTRIBUTE((nonnull, warn_unused_result))
bool
commit_cache_norebuild(
/*===================*/
ha_innobase_inplace_ctx*ctx,
const TABLE* table,
trx_t* trx)
{
DBUG_ENTER("commit_cache_norebuild");
bool found = true;
DBUG_ASSERT(!ctx->need_rebuild());
col_set drop_list;
col_set v_drop_list;
col_set::const_iterator col_it;
/* Check if the column, part of an index to be dropped is part of any
other index which is not being dropped. If it so, then set the ord_part
of the column to 0. */
get_col_list_to_be_dropped(ctx, drop_list, v_drop_list);
for (col_it = drop_list.begin(); col_it != drop_list.end(); ++col_it) {
if (!check_col_exists_in_indexes(ctx->new_table,
*col_it, false)) {
ctx->new_table->cols[*col_it].ord_part = 0;
}
}
for (col_it = v_drop_list.begin();
col_it != v_drop_list.end(); ++col_it) {
if (!check_col_exists_in_indexes(ctx->new_table,
*col_it, true)) {
ctx->new_table->v_cols[*col_it].m_col.ord_part = 0;
}
}
for (ulint i = 0; i < ctx->num_to_add_index; i++) {
dict_index_t* index = ctx->add_index[i];
DBUG_ASSERT(dict_index_get_online_status(index)
== ONLINE_INDEX_COMPLETE);
DBUG_ASSERT(!index->is_committed());
index->set_committed(true);
}
if (ctx->num_to_drop_index) {
/* Really drop the indexes that were dropped.
The transaction had to be committed first
(after renaming the indexes), so that in the
event of a crash, crash recovery will drop the
indexes, because it drops all indexes whose
names start with TEMP_INDEX_PREFIX. Once we
have started dropping an index tree, there is
no way to roll it back. */
for (ulint i = 0; i < ctx->num_to_drop_index; i++) {
dict_index_t* index = ctx->drop_index[i];
DBUG_ASSERT(index->is_committed());
DBUG_ASSERT(index->table == ctx->new_table);
DBUG_ASSERT(index->to_be_dropped);
/* Replace the indexes in foreign key
constraints if needed. */
if (!dict_foreign_replace_index(
index->table, ctx->col_names, index)) {
found = false;
}
/* Mark the index dropped
in the data dictionary cache. */
rw_lock_x_lock(dict_index_get_lock(index));
index->page = FIL_NULL;
rw_lock_x_unlock(dict_index_get_lock(index));
}
trx_start_for_ddl(trx, TRX_DICT_OP_INDEX);
row_merge_drop_indexes_dict(trx, ctx->new_table->id);
for (ulint i = 0; i < ctx->num_to_drop_index; i++) {
dict_index_t* index = ctx->drop_index[i];
DBUG_ASSERT(index->is_committed());
DBUG_ASSERT(index->table == ctx->new_table);
if (index->type & DICT_FTS) {
DBUG_ASSERT(index->type == DICT_FTS
|| (index->type
& DICT_CORRUPT));
DBUG_ASSERT(index->table->fts);
fts_drop_index(index->table, index, trx);
}
dict_index_remove_from_cache(index->table, index);
}
trx_commit_for_mysql(trx);
}
ctx->new_table->fts_doc_id_index
= ctx->new_table->fts
? dict_table_get_index_on_name(
ctx->new_table, FTS_DOC_ID_INDEX_NAME)
: NULL;
DBUG_ASSERT((ctx->new_table->fts == NULL)
== (ctx->new_table->fts_doc_id_index == NULL));
DBUG_RETURN(found);
}
/** Adjust the persistent statistics after non-rebuilding ALTER TABLE.
Remove statistics for dropped indexes, add statistics for created indexes
and rename statistics for renamed indexes.
@param ha_alter_info Data used during in-place alter
@param ctx In-place ALTER TABLE context
@param altered_table MySQL table that is being altered
@param table_name Table name in MySQL
@param thd MySQL connection
*/
static
void
alter_stats_norebuild(
/*==================*/
Alter_inplace_info* ha_alter_info,
ha_innobase_inplace_ctx* ctx,
TABLE* altered_table,
const char* table_name,
THD* thd)
{
ulint i;
DBUG_ENTER("alter_stats_norebuild");
DBUG_ASSERT(!ctx->need_rebuild());
if (!dict_stats_is_persistent_enabled(ctx->new_table)) {
DBUG_VOID_RETURN;
}
/* Delete corresponding rows from the stats table. We do this
in a separate transaction from trx, because lock waits are not
allowed in a data dictionary transaction. (Lock waits are possible
on the statistics table, because it is directly accessible by users,
not covered by the dict_operation_lock.)
Because the data dictionary changes were already committed, orphaned
rows may be left in the statistics table if the system crashes.
FIXME: each change to the statistics tables is being committed in a
separate transaction, meaning that the operation is not atomic
FIXME: This will not drop the (unused) statistics for
FTS_DOC_ID_INDEX if it was a hidden index, dropped together
with the last renamining FULLTEXT index. */
for (i = 0; i < ha_alter_info->index_drop_count; i++) {
const KEY* key = ha_alter_info->index_drop_buffer[i];
if (key->flags & HA_FULLTEXT) {
/* There are no index cardinality
statistics for FULLTEXT indexes. */
continue;
}
char errstr[1024];
if (dict_stats_drop_index(
ctx->new_table->name.m_name, key->name,
errstr, sizeof errstr) != DB_SUCCESS) {
push_warning(thd,
Sql_condition::WARN_LEVEL_WARN,
ER_LOCK_WAIT_TIMEOUT, errstr);
}
}
#ifdef MYSQL_RENAME_INDEX
for (i = 0; i < ha_alter_info->index_rename_count; i++) {
KEY_PAIR* pair = &ha_alter_info->index_rename_buffer[i];
dberr_t err;
err = dict_stats_rename_index(ctx->new_table,
pair->old_key->name,
pair->new_key->name);
if (err != DB_SUCCESS) {
push_warning_printf(
thd,
Sql_condition::WARN_LEVEL_WARN,
ER_ERROR_ON_RENAME,
"Error renaming an index of table '%s'"
" from '%s' to '%s' in InnoDB persistent"
" statistics storage: %s",
table_name,
pair->old_key->name,
pair->new_key->name,
ut_strerr(err));
}
}
#endif /* MYSQL_RENAME_INDEX */
for (i = 0; i < ctx->num_to_add_index; i++) {
dict_index_t* index = ctx->add_index[i];
DBUG_ASSERT(index->table == ctx->new_table);
if (!(index->type & DICT_FTS)) {
dict_stats_init(ctx->new_table);
dict_stats_update_for_index(index);
}
}
DBUG_VOID_RETURN;
}
/** Adjust the persistent statistics after rebuilding ALTER TABLE.
Remove statistics for dropped indexes, add statistics for created indexes
and rename statistics for renamed indexes.
@param table InnoDB table that was rebuilt by ALTER TABLE
@param table_name Table name in MySQL
@param thd MySQL connection
*/
static
void
alter_stats_rebuild(
/*================*/
dict_table_t* table,
const char* table_name,
THD* thd)
{
DBUG_ENTER("alter_stats_rebuild");
if (dict_table_is_discarded(table)
|| !dict_stats_is_persistent_enabled(table)) {
DBUG_VOID_RETURN;
}
#ifndef DBUG_OFF
bool file_unreadable_orig = false;
#endif /* DBUG_OFF */
DBUG_EXECUTE_IF(
"ib_rename_index_fail2",
file_unreadable_orig = table->file_unreadable;
table->file_unreadable = true;
);
dberr_t ret = dict_stats_update(table, DICT_STATS_RECALC_PERSISTENT);
DBUG_EXECUTE_IF(
"ib_rename_index_fail2",
table->file_unreadable = file_unreadable_orig;
);
if (ret != DB_SUCCESS) {
push_warning_printf(
thd,
Sql_condition::WARN_LEVEL_WARN,
ER_ALTER_INFO,
"Error updating stats for table '%s'"
" after table rebuild: %s",
table_name, ut_strerr(ret));
}
DBUG_VOID_RETURN;
}
#ifndef DBUG_OFF
# define DBUG_INJECT_CRASH(prefix, count) \
do { \
char buf[32]; \
ut_snprintf(buf, sizeof buf, prefix "_%u", count); \
DBUG_EXECUTE_IF(buf, DBUG_SUICIDE();); \
} while (0)
#else
# define DBUG_INJECT_CRASH(prefix, count)
#endif
/** Commit or rollback the changes made during
prepare_inplace_alter_table() and inplace_alter_table() inside
the storage engine. Note that the allowed level of concurrency
during this operation will be the same as for
inplace_alter_table() and thus might be higher than during
prepare_inplace_alter_table(). (E.g concurrent writes were
blocked during prepare, but might not be during commit).
@param altered_table TABLE object for new version of table.
@param ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@param commit true => Commit, false => Rollback.
@retval true Failure
@retval false Success
*/
bool
ha_innobase::commit_inplace_alter_table(
/*====================================*/
TABLE* altered_table,
Alter_inplace_info* ha_alter_info,
bool commit)
{
dberr_t error;
ha_innobase_inplace_ctx*ctx0;
struct mtr_buf_copy_t logs;
ctx0 = static_cast<ha_innobase_inplace_ctx*>
(ha_alter_info->handler_ctx);
#ifndef DBUG_OFF
uint crash_inject_count = 1;
uint crash_fail_inject_count = 1;
uint failure_inject_count = 1;
#endif /* DBUG_OFF */
DBUG_ENTER("commit_inplace_alter_table");
DBUG_ASSERT(!srv_read_only_mode);
DBUG_ASSERT(!ctx0 || ctx0->prebuilt == m_prebuilt);
DBUG_ASSERT(!ctx0 || ctx0->old_table == m_prebuilt->table);
DEBUG_SYNC_C("innodb_commit_inplace_alter_table_enter");
DEBUG_SYNC_C("innodb_commit_inplace_alter_table_wait");
if (ctx0 != NULL && ctx0->m_stage != NULL) {
ctx0->m_stage->begin_phase_end();
}
if (!commit) {
/* A rollback is being requested. So far we may at
most have created some indexes. If any indexes were to
be dropped, they would actually be dropped in this
method if commit=true. */
const bool ret = rollback_inplace_alter_table(
ha_alter_info, table, m_prebuilt);
DBUG_RETURN(ret);
}
if (!(ha_alter_info->handler_flags & ~INNOBASE_INPLACE_IGNORE)) {
DBUG_ASSERT(!ctx0);
MONITOR_ATOMIC_DEC(MONITOR_PENDING_ALTER_TABLE);
ha_alter_info->group_commit_ctx = NULL;
DBUG_RETURN(false);
}
DBUG_ASSERT(ctx0);
inplace_alter_handler_ctx** ctx_array;
inplace_alter_handler_ctx* ctx_single[2];
if (ha_alter_info->group_commit_ctx) {
ctx_array = ha_alter_info->group_commit_ctx;
} else {
ctx_single[0] = ctx0;
ctx_single[1] = NULL;
ctx_array = ctx_single;
}
DBUG_ASSERT(ctx0 == ctx_array[0]);
ut_ad(m_prebuilt->table == ctx0->old_table);
ha_alter_info->group_commit_ctx = NULL;
/* Free the ctx->trx of other partitions, if any. We will only
use the ctx0->trx here. Others may have been allocated in
the prepare stage. */
for (inplace_alter_handler_ctx** pctx = &ctx_array[1]; *pctx;
pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
if (ctx->trx) {
trx_free_for_mysql(ctx->trx);
ctx->trx = NULL;
}
}
trx_start_if_not_started_xa(m_prebuilt->trx, true);
for (inplace_alter_handler_ctx** pctx = ctx_array; *pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
DBUG_ASSERT(ctx->prebuilt->trx == m_prebuilt->trx);
/* If decryption failed for old table or new table
fail here. */
if ((!ctx->old_table->is_readable()
&& fil_space_get(ctx->old_table->space))
|| (!ctx->new_table->is_readable()
&& fil_space_get(ctx->new_table->space))) {
String str;
const char* engine= table_type();
get_error_message(HA_ERR_DECRYPTION_FAILED, &str);
my_error(ER_GET_ERRMSG, MYF(0), HA_ERR_DECRYPTION_FAILED, str.c_ptr(), engine);
DBUG_RETURN(true);
}
/* Exclusively lock the table, to ensure that no other
transaction is holding locks on the table while we
change the table definition. The MySQL meta-data lock
should normally guarantee that no conflicting locks
exist. However, FOREIGN KEY constraints checks and any
transactions collected during crash recovery could be
holding InnoDB locks only, not MySQL locks. */
error = row_merge_lock_table(
m_prebuilt->trx, ctx->old_table, LOCK_X);
if (error != DB_SUCCESS) {
my_error_innodb(
error, table_share->table_name.str, 0);
DBUG_RETURN(true);
}
}
DEBUG_SYNC(m_user_thd, "innodb_alter_commit_after_lock_table");
const bool new_clustered = ctx0->need_rebuild();
trx_t* trx = ctx0->trx;
bool fail = false;
if (new_clustered) {
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
DBUG_ASSERT(ctx->need_rebuild());
if (ctx->old_table->fts) {
ut_ad(!ctx->old_table->fts->add_wq);
fts_optimize_remove_table(
ctx->old_table);
}
if (ctx->new_table->fts) {
ut_ad(!ctx->new_table->fts->add_wq);
fts_optimize_remove_table(
ctx->new_table);
}
}
}
if (!trx) {
DBUG_ASSERT(!new_clustered);
trx = innobase_trx_allocate(m_user_thd);
}
trx_start_for_ddl(trx, TRX_DICT_OP_INDEX);
/* Latch the InnoDB data dictionary exclusively so that no deadlocks
or lock waits can happen in it during the data dictionary operation. */
row_mysql_lock_data_dictionary(trx);
ut_ad(log_append_on_checkpoint(NULL) == NULL);
/* Prevent the background statistics collection from accessing
the tables. */
for (;;) {
bool retry = false;
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
DBUG_ASSERT(new_clustered == ctx->need_rebuild());
if (new_clustered
&& !dict_stats_stop_bg(ctx->old_table)) {
retry = true;
}
if (!dict_stats_stop_bg(ctx->new_table)) {
retry = true;
}
}
if (!retry) {
break;
}
DICT_STATS_BG_YIELD(trx);
}
/* Apply the changes to the data dictionary tables, for all
partitions. */
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx && !fail; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
DBUG_ASSERT(new_clustered == ctx->need_rebuild());
commit_set_autoinc(ha_alter_info, ctx, altered_table, table);
if (ctx->need_rebuild()) {
ctx->tmp_name = dict_mem_create_temporary_tablename(
ctx->heap, ctx->new_table->name.m_name,
ctx->new_table->id);
fail = commit_try_rebuild(
ha_alter_info, ctx, altered_table, table,
trx, table_share->table_name.str);
} else {
fail = commit_try_norebuild(
ha_alter_info, ctx, altered_table, table, trx,
table_share->table_name.str);
}
DBUG_INJECT_CRASH("ib_commit_inplace_crash",
crash_inject_count++);
#ifndef DBUG_OFF
{
/* Generate a dynamic dbug text. */
char buf[32];
ut_snprintf(buf, sizeof buf,
"ib_commit_inplace_fail_%u",
failure_inject_count++);
DBUG_EXECUTE_IF(buf,
my_error(ER_INTERNAL_ERROR, MYF(0),
"Injected error!");
fail = true;
);
}
#endif
}
/* Commit or roll back the changes to the data dictionary. */
if (fail) {
trx_rollback_for_mysql(trx);
} else if (!new_clustered) {
trx_commit_for_mysql(trx);
} else {
mtr_t mtr;
mtr_start(&mtr);
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
DBUG_ASSERT(ctx->need_rebuild());
/* Check for any possible problems for any
file operations that will be performed in
commit_cache_rebuild(), and if none, generate
the redo log for these operations. */
error = fil_mtr_rename_log(ctx->old_table,
ctx->new_table,
ctx->tmp_name, &mtr);
if (error != DB_SUCCESS) {
/* Out of memory or a problem will occur
when renaming files. */
fail = true;
my_error_innodb(error, ctx->old_table->name.m_name,
ctx->old_table->flags);
}
DBUG_INJECT_CRASH("ib_commit_inplace_crash",
crash_inject_count++);
}
/* Test what happens on crash if the redo logs
are flushed to disk here. The log records
about the rename should not be committed, and
the data dictionary transaction should be
rolled back, restoring the old table. */
DBUG_EXECUTE_IF("innodb_alter_commit_crash_before_commit",
log_buffer_flush_to_disk();
DBUG_SUICIDE(););
ut_ad(!trx->fts_trx);
if (fail) {
mtr.set_log_mode(MTR_LOG_NO_REDO);
mtr_commit(&mtr);
trx_rollback_for_mysql(trx);
} else {
ut_ad(trx_state_eq(trx, TRX_STATE_ACTIVE));
ut_ad(trx->has_logged());
if (mtr.get_log()->size() > 0) {
ut_ad(*mtr.get_log()->front()->begin()
== MLOG_FILE_RENAME2);
/* Append the MLOG_FILE_RENAME2
records on checkpoint, as a separate
mini-transaction before the one that
contains the MLOG_CHECKPOINT marker. */
static const byte multi
= MLOG_MULTI_REC_END;
mtr.get_log()->for_each_block(logs);
logs.m_buf.push(&multi, sizeof multi);
log_append_on_checkpoint(&logs.m_buf);
}
/* The following call commits the
mini-transaction, making the data dictionary
transaction committed at mtr.end_lsn. The
transaction becomes 'durable' by the time when
log_buffer_flush_to_disk() returns. In the
logical sense the commit in the file-based
data structures happens here. */
trx_commit_low(trx, &mtr);
}
/* If server crashes here, the dictionary in
InnoDB and MySQL will differ. The .ibd files
and the .frm files must be swapped manually by
the administrator. No loss of data. */
DBUG_EXECUTE_IF("innodb_alter_commit_crash_after_commit",
log_make_checkpoint_at(LSN_MAX, TRUE);
log_buffer_flush_to_disk();
DBUG_SUICIDE(););
}
/* Flush the log to reduce probability that the .frm files and
the InnoDB data dictionary get out-of-sync if the user runs
with innodb_flush_log_at_trx_commit = 0 */
log_buffer_flush_to_disk();
/* At this point, the changes to the persistent storage have
been committed or rolled back. What remains to be done is to
update the in-memory structures, close some handles, release
temporary files, and (unless we rolled back) update persistent
statistics. */
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>(*pctx);
DBUG_ASSERT(ctx->need_rebuild() == new_clustered);
if (new_clustered) {
innobase_online_rebuild_log_free(ctx->old_table);
}
if (fail) {
if (new_clustered) {
trx_start_for_ddl(trx, TRX_DICT_OP_TABLE);
dict_table_close_and_drop(trx, ctx->new_table);
trx_commit_for_mysql(trx);
ctx->new_table = NULL;
} else {
/* We failed, but did not rebuild the table.
Roll back any ADD INDEX, or get rid of garbage
ADD INDEX that was left over from a previous
ALTER TABLE statement. */
trx_start_for_ddl(trx, TRX_DICT_OP_INDEX);
innobase_rollback_sec_index(
ctx->new_table, table, TRUE, trx);
trx_commit_for_mysql(trx);
}
DBUG_INJECT_CRASH("ib_commit_inplace_crash_fail",
crash_fail_inject_count++);
continue;
}
innobase_copy_frm_flags_from_table_share(
ctx->new_table, altered_table->s);
if (new_clustered) {
/* We will reload and refresh the
in-memory foreign key constraint
metadata. This is a rename operation
in preparing for dropping the old
table. Set the table to_be_dropped bit
here, so to make sure DML foreign key
constraint check does not use the
stale dict_foreign_t. This is done
because WL#6049 (FK MDL) has not been
implemented yet. */
ctx->old_table->to_be_dropped = true;
DBUG_PRINT("to_be_dropped",
("table: %s", ctx->old_table->name.m_name));
/* Rename the tablespace files. */
commit_cache_rebuild(ctx);
error = innobase_update_foreign_cache(ctx, m_user_thd);
if (error != DB_SUCCESS) {
goto foreign_fail;
}
} else {
error = innobase_update_foreign_cache(ctx, m_user_thd);
if (error != DB_SUCCESS) {
foreign_fail:
/* The data dictionary cache
should be corrupted now. The
best solution should be to
kill and restart the server,
but the *.frm file has not
been replaced yet. */
push_warning_printf(
m_user_thd,
Sql_condition::WARN_LEVEL_WARN,
ER_ALTER_INFO,
"InnoDB: Could not add foreign"
" key constraints.");
} else {
if (!commit_cache_norebuild(
ctx, table, trx)) {
ut_a(!m_prebuilt->trx->check_foreigns);
}
innobase_rename_or_enlarge_columns_cache(
ha_alter_info, table,
ctx->new_table);
#ifdef MYSQL_RENAME_INDEX
rename_indexes_in_cache(ctx, ha_alter_info);
#endif
}
}
dict_mem_table_free_foreign_vcol_set(ctx->new_table);
dict_mem_table_fill_foreign_vcol_set(ctx->new_table);
DBUG_INJECT_CRASH("ib_commit_inplace_crash",
crash_inject_count++);
}
log_append_on_checkpoint(NULL);
/* Invalidate the index translation table. In partitioned
tables, there is no share. */
if (m_share) {
m_share->idx_trans_tbl.index_count = 0;
}
if (trx == ctx0->trx) {
ctx0->trx = NULL;
}
/* Tell the InnoDB server that there might be work for
utility threads: */
srv_active_wake_master_thread();
if (fail) {
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>
(*pctx);
DBUG_ASSERT(ctx->need_rebuild() == new_clustered);
ut_d(dict_table_check_for_dup_indexes(
ctx->old_table,
CHECK_ABORTED_OK));
ut_a(fts_check_cached_index(ctx->old_table));
DBUG_INJECT_CRASH("ib_commit_inplace_crash_fail",
crash_fail_inject_count++);
}
row_mysql_unlock_data_dictionary(trx);
trx_free_for_mysql(trx);
DBUG_RETURN(true);
}
if (ctx0->num_to_drop_vcol || ctx0->num_to_add_vcol) {
DBUG_ASSERT(ctx0->old_table->get_ref_count() == 1);
trx_commit_for_mysql(m_prebuilt->trx);
#ifdef BTR_CUR_HASH_ADAPT
if (btr_search_enabled) {
btr_search_disable(false);
btr_search_enable();
}
#endif /* BTR_CUR_HASH_ADAPT */
char tb_name[FN_REFLEN];
ut_strcpy(tb_name, m_prebuilt->table->name.m_name);
tb_name[strlen(m_prebuilt->table->name.m_name)] = 0;
dict_table_close(m_prebuilt->table, true, false);
dict_table_remove_from_cache(m_prebuilt->table);
m_prebuilt->table = dict_table_open_on_name(
tb_name, TRUE, TRUE, DICT_ERR_IGNORE_NONE);
/* Drop outdated table stats. */
char errstr[1024];
if (dict_stats_drop_table(
m_prebuilt->table->name.m_name,
errstr, sizeof(errstr))
!= DB_SUCCESS) {
push_warning_printf(
m_user_thd,
Sql_condition::WARN_LEVEL_WARN,
ER_ALTER_INFO,
"Deleting persistent statistics"
" for table '%s' in"
" InnoDB failed: %s",
table->s->table_name.str,
errstr);
}
row_mysql_unlock_data_dictionary(trx);
trx_free_for_mysql(trx);
MONITOR_ATOMIC_DEC(MONITOR_PENDING_ALTER_TABLE);
DBUG_RETURN(false);
}
/* Release the table locks. */
trx_commit_for_mysql(m_prebuilt->trx);
DBUG_EXECUTE_IF("ib_ddl_crash_after_user_trx_commit", DBUG_SUICIDE(););
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>
(*pctx);
DBUG_ASSERT(ctx->need_rebuild() == new_clustered);
bool add_fts = false;
/* Publish the created fulltext index, if any.
Note that a fulltext index can be created without
creating the clustered index, if there already exists
a suitable FTS_DOC_ID column. If not, one will be
created, implying new_clustered */
for (ulint i = 0; i < ctx->num_to_add_index; i++) {
dict_index_t* index = ctx->add_index[i];
if (index->type & DICT_FTS) {
DBUG_ASSERT(index->type == DICT_FTS);
/* We reset DICT_TF2_FTS here because the bit
is left unset when a drop proceeds the add. */
DICT_TF2_FLAG_SET(ctx->new_table, DICT_TF2_FTS);
fts_add_index(index, ctx->new_table);
add_fts = true;
}
}
ut_d(dict_table_check_for_dup_indexes(
ctx->new_table, CHECK_ALL_COMPLETE));
if (add_fts) {
fts_optimize_add_table(ctx->new_table);
}
ut_d(dict_table_check_for_dup_indexes(
ctx->new_table, CHECK_ABORTED_OK));
ut_a(fts_check_cached_index(ctx->new_table));
if (new_clustered) {
/* Since the table has been rebuilt, we remove
all persistent statistics corresponding to the
old copy of the table (which was renamed to
ctx->tmp_name). */
char errstr[1024];
DBUG_ASSERT(0 == strcmp(ctx->old_table->name.m_name,
ctx->tmp_name));
DBUG_EXECUTE_IF(
"ib_rename_index_fail3",
DBUG_SET("+d,innodb_report_deadlock");
);
if (dict_stats_drop_table(
ctx->new_table->name.m_name,
errstr, sizeof(errstr))
!= DB_SUCCESS) {
push_warning_printf(
m_user_thd,
Sql_condition::WARN_LEVEL_WARN,
ER_ALTER_INFO,
"Deleting persistent statistics"
" for rebuilt table '%s' in"
" InnoDB failed: %s",
table->s->table_name.str,
errstr);
}
DBUG_EXECUTE_IF(
"ib_rename_index_fail3",
DBUG_SET("-d,innodb_report_deadlock");
);
DBUG_EXECUTE_IF("ib_ddl_crash_before_commit",
DBUG_SUICIDE(););
ut_ad(m_prebuilt != ctx->prebuilt
|| ctx == ctx0);
bool update_own_prebuilt =
(m_prebuilt == ctx->prebuilt);
trx_t* const user_trx = m_prebuilt->trx;
row_prebuilt_free(ctx->prebuilt, TRUE);
/* Drop the copy of the old table, which was
renamed to ctx->tmp_name at the atomic DDL
transaction commit. If the system crashes
before this is completed, some orphan tables
with ctx->tmp_name may be recovered. */
trx_start_for_ddl(trx, TRX_DICT_OP_TABLE);
error = row_merge_drop_table(trx, ctx->old_table);
if (error != DB_SUCCESS) {
ib::error() << "Inplace alter table " << ctx->old_table->name.m_name
<< " dropping copy of the old table failed error "
<< error
<< ". tmp_name " << (ctx->tmp_name ? ctx->tmp_name : "N/A")
<< " new_table " << (ctx->new_table ? ctx->new_table->name.m_name
: "N/A");
}
trx_commit_for_mysql(trx);
/* Rebuild the prebuilt object. */
ctx->prebuilt = row_create_prebuilt(
ctx->new_table, altered_table->s->reclength);
if (update_own_prebuilt) {
m_prebuilt = ctx->prebuilt;
}
trx_start_if_not_started(user_trx, true);
user_trx->will_lock++;
m_prebuilt->trx = user_trx;
}
DBUG_INJECT_CRASH("ib_commit_inplace_crash",
crash_inject_count++);
}
row_mysql_unlock_data_dictionary(trx);
trx_free_for_mysql(trx);
/* TODO: The following code could be executed
while allowing concurrent access to the table
(MDL downgrade). */
if (new_clustered) {
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>
(*pctx);
DBUG_ASSERT(ctx->need_rebuild());
alter_stats_rebuild(
ctx->new_table, table->s->table_name.str,
m_user_thd);
DBUG_INJECT_CRASH("ib_commit_inplace_crash",
crash_inject_count++);
}
} else {
for (inplace_alter_handler_ctx** pctx = ctx_array;
*pctx; pctx++) {
ha_innobase_inplace_ctx* ctx
= static_cast<ha_innobase_inplace_ctx*>
(*pctx);
DBUG_ASSERT(!ctx->need_rebuild());
alter_stats_norebuild(
ha_alter_info, ctx, altered_table,
table->s->table_name.str, m_user_thd);
DBUG_INJECT_CRASH("ib_commit_inplace_crash",
crash_inject_count++);
}
}
innobase_parse_hint_from_comment(
m_user_thd, m_prebuilt->table, altered_table->s);
/* TODO: Also perform DROP TABLE and DROP INDEX after
the MDL downgrade. */
#ifndef DBUG_OFF
dict_index_t* clust_index = dict_table_get_first_index(
ctx0->prebuilt->table);
DBUG_ASSERT(!clust_index->online_log);
DBUG_ASSERT(dict_index_get_online_status(clust_index)
== ONLINE_INDEX_COMPLETE);
for (dict_index_t* index = clust_index;
index;
index = dict_table_get_next_index(index)) {
DBUG_ASSERT(!index->to_be_dropped);
}
#endif /* DBUG_OFF */
MONITOR_ATOMIC_DEC(MONITOR_PENDING_ALTER_TABLE);
DBUG_RETURN(false);
}
/** Helper class for in-place alter, see handler.h */
class ha_innopart_inplace_ctx : public inplace_alter_handler_ctx
{
/* Only used locally in this file, so have everything public for
conveniance. */
public:
/** Total number of partitions. */
uint m_tot_parts;
/** Array of inplace contexts for all partitions. */
inplace_alter_handler_ctx** ctx_array;
/** Array of prebuilt for all partitions. */
row_prebuilt_t** prebuilt_array;
ha_innopart_inplace_ctx(THD *thd, uint tot_parts)
: inplace_alter_handler_ctx(),
m_tot_parts(tot_parts),
ctx_array(),
prebuilt_array()
{}
~ha_innopart_inplace_ctx()
{
if (ctx_array) {
for (uint i = 0; i < m_tot_parts; i++) {
delete ctx_array[i];
}
ut_free(ctx_array);
}
if (prebuilt_array) {
/* First entry is the original prebuilt! */
for (uint i = 1; i < m_tot_parts; i++) {
/* Don't close the tables. */
prebuilt_array[i]->table = NULL;
row_prebuilt_free(prebuilt_array[i], false);
}
ut_free(prebuilt_array);
}
}
};
#ifdef MYSQL_INNODB_PARTITIONING
/** Check if supported inplace alter table.
@param[in] altered_table Altered MySQL table.
@param[in] ha_alter_info Information about inplace operations to do.
@return Lock level, not supported or error */
enum_alter_inplace_result
ha_innopart::check_if_supported_inplace_alter(
TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
DBUG_ENTER("ha_innopart::check_if_supported_inplace_alter");
DBUG_ASSERT(ha_alter_info->handler_ctx == NULL);
/* Not supporting these for partitioned tables yet! */
/* FK not yet supported. */
if (ha_alter_info->handler_flags
& (Alter_inplace_info::ADD_FOREIGN_KEY
| Alter_inplace_info::DROP_FOREIGN_KEY)) {
ha_alter_info->unsupported_reason = innobase_get_err_msg(
ER_FOREIGN_KEY_ON_PARTITIONED);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
/* FTS not yet supported either. */
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ADD_INDEX)) {
for (uint i = 0; i < ha_alter_info->index_add_count; i++) {
const KEY* key =
&ha_alter_info->key_info_buffer[
ha_alter_info->index_add_buffer[i]];
if (key->flags & HA_FULLTEXT) {
DBUG_ASSERT(!(key->flags & HA_KEYFLAG_MASK
& ~(HA_FULLTEXT
| HA_PACK_KEY
| HA_GENERATED_KEY
| HA_BINARY_PACK_KEY)));
ha_alter_info->unsupported_reason =
innobase_get_err_msg(
ER_FULLTEXT_NOT_SUPPORTED_WITH_PARTITIONING);
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
}
/* We cannot allow INPLACE to change order of KEY partitioning fields! */
if ((ha_alter_info->handler_flags
& Alter_inplace_info::ALTER_STORED_COLUMN_ORDER)
&& !m_part_info->same_key_column_order(
&ha_alter_info->alter_info->create_list)) {
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
/* Cannot allow INPLACE for drop and create PRIMARY KEY if partition is
on Primary Key - PARTITION BY KEY() */
if ((ha_alter_info->handler_flags
& (Alter_inplace_info::ADD_PK_INDEX
| Alter_inplace_info::DROP_PK_INDEX))) {
/* Check partition by key(). */
if ((m_part_info->part_type == HASH_PARTITION)
&& m_part_info->list_of_part_fields
&& m_part_info->part_field_list.is_empty()) {
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
/* Check sub-partition by key(). */
if ((m_part_info->subpart_type == HASH_PARTITION)
&& m_part_info->list_of_subpart_fields
&& m_part_info->subpart_field_list.is_empty()) {
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
}
}
/* Check for PK and UNIQUE should already be done when creating the
new table metadata.
(fix_partition_info/check_primary_key+check_unique_key) */
set_partition(0);
DBUG_RETURN(ha_innobase::check_if_supported_inplace_alter(altered_table,
ha_alter_info));
}
/** Prepare inplace alter table.
Allows InnoDB to update internal structures with concurrent
writes blocked (provided that check_if_supported_inplace_alter()
did not return HA_ALTER_INPLACE_NO_LOCK).
This will be invoked before inplace_alter_table().
@param[in] altered_table TABLE object for new version of table.
@param[in] ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@retval true Failure.
@retval false Success. */
bool
ha_innopart::prepare_inplace_alter_table(
TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
THD* thd;
ha_innopart_inplace_ctx* ctx_parts;
bool res = true;
DBUG_ENTER("ha_innopart::prepare_inplace_alter_table");
DBUG_ASSERT(ha_alter_info->handler_ctx == NULL);
thd = ha_thd();
/* Clean up all ins/upd nodes. */
clear_ins_upd_nodes();
/* Based on Sql_alloc class, return NULL for new on failure. */
ctx_parts = new ha_innopart_inplace_ctx(thd, m_tot_parts);
if (!ctx_parts) {
DBUG_RETURN(HA_ALTER_ERROR);
}
uint ctx_array_size = sizeof(inplace_alter_handler_ctx*)
* (m_tot_parts + 1);
ctx_parts->ctx_array =
static_cast<inplace_alter_handler_ctx**>(
ut_malloc(ctx_array_size,
mem_key_partitioning));
if (!ctx_parts->ctx_array) {
DBUG_RETURN(HA_ALTER_ERROR);
}
/* Set all to NULL, including the terminating one. */
memset(ctx_parts->ctx_array, 0, ctx_array_size);
ctx_parts->prebuilt_array = static_cast<row_prebuilt_t**>(
ut_malloc(sizeof(row_prebuilt_t*)
* m_tot_parts,
mem_key_partitioning));
if (!ctx_parts->prebuilt_array) {
DBUG_RETURN(HA_ALTER_ERROR);
}
/* For the first partition use the current prebuilt. */
ctx_parts->prebuilt_array[0] = m_prebuilt;
/* Create new prebuilt for the rest of the partitions.
It is needed for the current implementation of
ha_innobase::commit_inplace_alter_table(). */
for (uint i = 1; i < m_tot_parts; i++) {
row_prebuilt_t* tmp_prebuilt;
tmp_prebuilt = row_create_prebuilt(
m_part_share->get_table_part(i),
table_share->reclength);
/* Use same trx as original prebuilt. */
tmp_prebuilt->trx = m_prebuilt->trx;
ctx_parts->prebuilt_array[i] = tmp_prebuilt;
}
const char* save_tablespace =
ha_alter_info->create_info->tablespace;
const char* save_data_file_name =
ha_alter_info->create_info->data_file_name;
for (uint i = 0; i < m_tot_parts; i++) {
m_prebuilt = ctx_parts->prebuilt_array[i];
m_prebuilt_ptr = ctx_parts->prebuilt_array + i;
ha_alter_info->handler_ctx = ctx_parts->ctx_array[i];
set_partition(i);
/* Set the tablespace and data_file_name value of the
alter_info to the tablespace value and data_file_name
value that was existing for the partition originally,
so that for ALTER TABLE the tablespace clause in create
option is ignored for existing partitions, and later
set it back to its old value */
ha_alter_info->create_info->tablespace =
m_prebuilt->table->tablespace;
ha_alter_info->create_info->data_file_name =
m_prebuilt->table->data_dir_path;
res = ha_innobase::prepare_inplace_alter_table(altered_table,
ha_alter_info);
update_partition(i);
ctx_parts->ctx_array[i] = ha_alter_info->handler_ctx;
if (res) {
break;
}
}
m_prebuilt = ctx_parts->prebuilt_array[0];
m_prebuilt_ptr = &m_prebuilt;
ha_alter_info->handler_ctx = ctx_parts;
ha_alter_info->group_commit_ctx = ctx_parts->ctx_array;
ha_alter_info->create_info->tablespace = save_tablespace;
ha_alter_info->create_info->data_file_name = save_data_file_name;
DBUG_RETURN(res);
}
/** Inplace alter table.
Alter the table structure in-place with operations
specified using Alter_inplace_info.
The level of concurrency allowed during this operation depends
on the return value from check_if_supported_inplace_alter().
@param[in] altered_table TABLE object for new version of table.
@param[in] ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@retval true Failure.
@retval false Success. */
bool
ha_innopart::inplace_alter_table(
TABLE* altered_table,
Alter_inplace_info* ha_alter_info)
{
bool res = true;
ha_innopart_inplace_ctx* ctx_parts;
ctx_parts = static_cast<ha_innopart_inplace_ctx*>(
ha_alter_info->handler_ctx);
for (uint i = 0; i < m_tot_parts; i++) {
m_prebuilt = ctx_parts->prebuilt_array[i];
ha_alter_info->handler_ctx = ctx_parts->ctx_array[i];
set_partition(i);
res = ha_innobase::inplace_alter_table(altered_table,
ha_alter_info);
ut_ad(ctx_parts->ctx_array[i] == ha_alter_info->handler_ctx);
ctx_parts->ctx_array[i] = ha_alter_info->handler_ctx;
if (res) {
break;
}
}
m_prebuilt = ctx_parts->prebuilt_array[0];
ha_alter_info->handler_ctx = ctx_parts;
return(res);
}
/** Commit or rollback inplace alter table.
Commit or rollback the changes made during
prepare_inplace_alter_table() and inplace_alter_table() inside
the storage engine. Note that the allowed level of concurrency
during this operation will be the same as for
inplace_alter_table() and thus might be higher than during
prepare_inplace_alter_table(). (E.g concurrent writes were
blocked during prepare, but might not be during commit).
@param[in] altered_table TABLE object for new version of table.
@param[in] ha_alter_info Structure describing changes to be done
by ALTER TABLE and holding data used during in-place alter.
@param[in] commit true => Commit, false => Rollback.
@retval true Failure.
@retval false Success. */
bool
ha_innopart::commit_inplace_alter_table(
TABLE* altered_table,
Alter_inplace_info* ha_alter_info,
bool commit)
{
bool res = false;
ha_innopart_inplace_ctx* ctx_parts;
ctx_parts = static_cast<ha_innopart_inplace_ctx*>(
ha_alter_info->handler_ctx);
ut_ad(ctx_parts);
ut_ad(ctx_parts->prebuilt_array);
ut_ad(ctx_parts->prebuilt_array[0] == m_prebuilt);
if (commit) {
/* Commit is done through first partition (group commit). */
ut_ad(ha_alter_info->group_commit_ctx == ctx_parts->ctx_array);
ha_alter_info->handler_ctx = ctx_parts->ctx_array[0];
set_partition(0);
res = ha_innobase::commit_inplace_alter_table(altered_table,
ha_alter_info,
commit);
ut_ad(res || !ha_alter_info->group_commit_ctx);
goto end;
}
/* Rollback is done for each partition. */
for (uint i = 0; i < m_tot_parts; i++) {
m_prebuilt = ctx_parts->prebuilt_array[i];
ha_alter_info->handler_ctx = ctx_parts->ctx_array[i];
set_partition(i);
if (ha_innobase::commit_inplace_alter_table(altered_table,
ha_alter_info, commit)) {
res = true;
}
ut_ad(ctx_parts->ctx_array[i] == ha_alter_info->handler_ctx);
ctx_parts->ctx_array[i] = ha_alter_info->handler_ctx;
}
end:
/* Move the ownership of the new tables back to
the m_part_share. */
ha_innobase_inplace_ctx* ctx;
for (uint i = 0; i < m_tot_parts; i++) {
/* TODO: Fix to only use one prebuilt (i.e. make inplace
alter partition aware instead of using multiple prebuilt
copies... */
ctx = static_cast<ha_innobase_inplace_ctx*>(
ctx_parts->ctx_array[i]);
if (ctx) {
m_part_share->set_table_part(i, ctx->prebuilt->table);
ctx->prebuilt->table = NULL;
ctx_parts->prebuilt_array[i] = ctx->prebuilt;
}
}
/* The above juggling of prebuilt must be reset here. */
m_prebuilt = ctx_parts->prebuilt_array[0];
m_prebuilt->table = m_part_share->get_table_part(0);
ha_alter_info->handler_ctx = ctx_parts;
return(res);
}
/** Notify the storage engine that the table structure (.frm) has
been updated.
ha_partition allows inplace operations that also upgrades the engine
if it supports partitioning natively. So if this is the case then
we will remove the .par file since it is not used with ha_innopart
(we use the internal data dictionary instead). */
void
ha_innopart::notify_table_changed()
{
char tmp_par_path[FN_REFLEN + 1];
strxnmov(tmp_par_path, FN_REFLEN, table->s->normalized_path.str,
".par", NullS);
if (my_access(tmp_par_path, W_OK) == 0)
{
my_delete(tmp_par_path, MYF(0));
}
}
#endif /* MYSQL_INNODB_PARTITIONING */
/**
@param thd the session
@param start_value the lower bound
@param max_value the upper bound (inclusive) */
ib_sequence_t::ib_sequence_t(
THD* thd,
ulonglong start_value,
ulonglong max_value)
:
m_max_value(max_value),
m_increment(0),
m_offset(0),
m_next_value(start_value),
m_eof(false)
{
if (thd != 0 && m_max_value > 0) {
thd_get_autoinc(thd, &m_offset, &m_increment);
if (m_increment > 1 || m_offset > 1) {
/* If there is an offset or increment specified
then we need to work out the exact next value. */
m_next_value = innobase_next_autoinc(
start_value, 1,
m_increment, m_offset, m_max_value);
} else if (start_value == 0) {
/* The next value can never be 0. */
m_next_value = 1;
}
} else {
m_eof = true;
}
}
/**
Postfix increment
@return the next value to insert */
ulonglong
ib_sequence_t::operator++(int) UNIV_NOTHROW
{
ulonglong current = m_next_value;
ut_ad(!m_eof);
ut_ad(m_max_value > 0);
m_next_value = innobase_next_autoinc(
current, 1, m_increment, m_offset, m_max_value);
if (m_next_value == m_max_value && current == m_next_value) {
m_eof = true;
}
return(current);
}