mariadb/storage/innobase/dict/dict0dict.cc
2024-11-29 13:43:17 +02:00

4850 lines
129 KiB
C++

/*****************************************************************************
Copyright (c) 1996, 2016, Oracle and/or its affiliates. All Rights Reserved.
Copyright (c) 2012, Facebook Inc.
Copyright (c) 2013, 2023, MariaDB Corporation.
This program is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free Software
Foundation; version 2 of the License.
This program is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with
this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
*****************************************************************************/
/******************************************************************//**
@file dict/dict0dict.cc
Data dictionary system
Created 1/8/1996 Heikki Tuuri
***********************************************************************/
#include <my_config.h>
#include <string>
#include "ha_prototypes.h"
#include <mysqld.h>
#include <strfunc.h>
#include "dict0dict.h"
#include "fts0fts.h"
#include "fil0fil.h"
#include <algorithm>
#include "sql_class.h"
#include "sql_table.h"
#include <mysql/service_thd_mdl.h>
#include "btr0btr.h"
#include "btr0cur.h"
#include "btr0sea.h"
#include "buf0buf.h"
#include "data0type.h"
#include "dict0boot.h"
#include "dict0load.h"
#include "dict0crea.h"
#include "dict0mem.h"
#include "dict0stats.h"
#include "fts0fts.h"
#include "fts0types.h"
#include "lock0lock.h"
#include "mach0data.h"
#include "mem0mem.h"
#include "page0page.h"
#include "page0zip.h"
#include "pars0pars.h"
#include "pars0sym.h"
#include "que0que.h"
#include "rem0cmp.h"
#include "row0log.h"
#include "row0merge.h"
#include "row0mysql.h"
#include "row0upd.h"
#include "srv0mon.h"
#include "srv0start.h"
#include "trx0undo.h"
#include "trx0purge.h"
#include <vector>
#include <algorithm>
/** the dictionary system */
dict_sys_t dict_sys;
/** System table names; @see dict_system_id_t */
const span<const char> dict_sys_t::SYS_TABLE[]=
{
{C_STRING_WITH_LEN("SYS_TABLES")},{C_STRING_WITH_LEN("SYS_INDEXES")},
{C_STRING_WITH_LEN("SYS_COLUMNS")},{C_STRING_WITH_LEN("SYS_FIELDS")},
{C_STRING_WITH_LEN("SYS_FOREIGN")},{C_STRING_WITH_LEN("SYS_FOREIGN_COLS")},
{C_STRING_WITH_LEN("SYS_VIRTUAL")}
};
/** Diagnostic message for exceeding the mutex_lock_wait() timeout */
const char dict_sys_t::fatal_msg[]=
"innodb_fatal_semaphore_wait_threshold was exceeded for dict_sys.latch. "
"Please refer to "
"https://mariadb.com/kb/en/how-to-produce-a-full-stack-trace-for-mysqld/";
/** Percentage of compression failures that are allowed in a single
round */
ulong zip_failure_threshold_pct = 5;
/** Maximum percentage of a page that can be allowed as a pad to avoid
compression failures */
ulong zip_pad_max = 50;
#define DICT_HEAP_SIZE 100 /*!< initial memory heap size when
creating a table or index object */
#define DICT_POOL_PER_TABLE_HASH 512 /*!< buffer pool max size per table
hash table fixed size in bytes */
#define DICT_POOL_PER_VARYING 4 /*!< buffer pool max size per data
dictionary varying size in bytes */
/** Identifies generated InnoDB foreign key names */
static char dict_ibfk[] = "_ibfk_";
/*******************************************************************//**
Tries to find column names for the index and sets the col field of the
index.
@param[in] index index
@param[in] add_v new virtual columns added along with an add index call
@return whether the column names were found */
static
bool
dict_index_find_cols(
dict_index_t* index,
const dict_add_v_col_t* add_v);
/*******************************************************************//**
Builds the internal dictionary cache representation for a clustered
index, containing also system fields not defined by the user.
@return own: the internal representation of the clustered index */
static
dict_index_t*
dict_index_build_internal_clust(
/*============================*/
dict_index_t* index); /*!< in: user representation of
a clustered index */
/*******************************************************************//**
Builds the internal dictionary cache representation for a non-clustered
index, containing also system fields not defined by the user.
@return own: the internal representation of the non-clustered index */
static
dict_index_t*
dict_index_build_internal_non_clust(
/*================================*/
dict_index_t* index); /*!< in: user representation of
a non-clustered index */
/**********************************************************************//**
Builds the internal dictionary cache representation for an FTS index.
@return own: the internal representation of the FTS index */
static
dict_index_t*
dict_index_build_internal_fts(
/*==========================*/
dict_index_t* index); /*!< in: user representation of an FTS index */
/**********************************************************************//**
Removes an index from the dictionary cache. */
static
void
dict_index_remove_from_cache_low(
/*=============================*/
dict_table_t* table, /*!< in/out: table */
dict_index_t* index, /*!< in, own: index */
ibool lru_evict); /*!< in: TRUE if page being evicted
to make room in the table LRU list */
#ifdef UNIV_DEBUG
/**********************************************************************//**
Validate the dictionary table LRU list.
@return TRUE if validate OK */
static
ibool
dict_lru_validate(void);
/*===================*/
#endif /* UNIV_DEBUG */
/* Stream for storing detailed information about the latest foreign key
and unique key errors. Only created if !srv_read_only_mode */
FILE* dict_foreign_err_file = NULL;
/* mutex protecting the foreign and unique error buffers */
mysql_mutex_t dict_foreign_err_mutex;
/********************************************************************//**
Checks if the database name in two table names is the same.
@return TRUE if same db name */
ibool
dict_tables_have_same_db(
/*=====================*/
const char* name1, /*!< in: table name in the form
dbname '/' tablename */
const char* name2) /*!< in: table name in the form
dbname '/' tablename */
{
for (; *name1 == *name2; name1++, name2++) {
if (*name1 == '/') {
return(TRUE);
}
ut_a(*name1); /* the names must contain '/' */
}
return(FALSE);
}
/** Decrement the count of open handles */
void dict_table_close(dict_table_t *table)
{
if (table->get_ref_count() == 1 &&
dict_stats_is_persistent_enabled(table) &&
strchr(table->name.m_name, '/'))
{
/* It looks like we are closing the last handle. The user could
have executed FLUSH TABLES in order to have the statistics reloaded
from the InnoDB persistent statistics tables. We must acquire
exclusive dict_sys.latch to prevent a race condition with another
thread concurrently acquiring a handle on the table. */
dict_sys.lock(SRW_LOCK_CALL);
if (table->release())
{
table->stats_mutex_lock();
if (table->get_ref_count() == 0)
dict_stats_deinit(table);
table->stats_mutex_unlock();
}
dict_sys.unlock();
}
else
table->release();
}
/** Decrements the count of open handles of a table.
@param[in,out] table table
@param[in] dict_locked whether dict_sys.latch is being held
@param[in] thd thread to release MDL
@param[in] mdl metadata lock or NULL if the thread
is a foreground one. */
void
dict_table_close(
dict_table_t* table,
bool dict_locked,
THD* thd,
MDL_ticket* mdl)
{
if (!dict_locked)
dict_table_close(table);
else
{
if (table->release() && dict_stats_is_persistent_enabled(table) &&
strchr(table->name.m_name, '/'))
{
/* Force persistent stats re-read upon next open of the table so
that FLUSH TABLE can be used to forcibly fetch stats from disk if
they have been manually modified. */
table->stats_mutex_lock();
if (table->get_ref_count() == 0)
dict_stats_deinit(table);
table->stats_mutex_unlock();
}
ut_ad(dict_lru_validate());
ut_ad(dict_sys.find(table));
}
if (!thd || !mdl);
else if (MDL_context *mdl_context= static_cast<MDL_context*>
(thd_mdl_context(thd)))
mdl_context->release_lock(mdl);
}
/** Check if the table has a given (non_virtual) column.
@param[in] table table object
@param[in] col_name column name
@param[in] col_nr column number guessed, 0 as default
@return column number if the table has the specified column,
otherwise table->n_def */
ulint
dict_table_has_column(
const dict_table_t* table,
const char* col_name,
ulint col_nr)
{
ulint col_max = table->n_def;
ut_ad(table);
ut_ad(col_name);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
if (col_nr < col_max
&& innobase_strcasecmp(
col_name, dict_table_get_col_name(table, col_nr)) == 0) {
return(col_nr);
}
/** The order of column may changed, check it with other columns */
for (ulint i = 0; i < col_max; i++) {
if (i != col_nr
&& innobase_strcasecmp(
col_name, dict_table_get_col_name(table, i)) == 0) {
return(i);
}
}
return(col_max);
}
/** Retrieve the column name.
@param[in] table the table of this column */
const char* dict_col_t::name(const dict_table_t& table) const
{
ut_ad(table.magic_n == DICT_TABLE_MAGIC_N);
size_t col_nr;
const char *s;
if (is_virtual()) {
col_nr = size_t(reinterpret_cast<const dict_v_col_t*>(this)
- table.v_cols);
ut_ad(col_nr < table.n_v_def);
s = table.v_col_names;
} else {
col_nr = size_t(this - table.cols);
ut_ad(col_nr < table.n_def);
s = table.col_names;
}
if (s) {
for (size_t i = 0; i < col_nr; i++) {
s += strlen(s) + 1;
}
}
return(s);
}
/** Returns a virtual column's name.
@param[in] table target table
@param[in] col_nr virtual column number (nth virtual column)
@return column name or NULL if column number out of range. */
const char*
dict_table_get_v_col_name(
const dict_table_t* table,
ulint col_nr)
{
const char* s;
ut_ad(table);
ut_ad(col_nr < table->n_v_def);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
if (col_nr >= table->n_v_def) {
return(NULL);
}
s = table->v_col_names;
if (s != NULL) {
for (ulint i = 0; i < col_nr; i++) {
s += strlen(s) + 1;
}
}
return(s);
}
/** Search virtual column's position in InnoDB according to its position
in original table's position
@param[in] table target table
@param[in] col_nr column number (nth column in the MySQL table)
@return virtual column's position in InnoDB, ULINT_UNDEFINED if not find */
static
ulint
dict_table_get_v_col_pos_for_mysql(
const dict_table_t* table,
ulint col_nr)
{
ulint i;
ut_ad(table);
ut_ad(col_nr < static_cast<ulint>(table->n_t_def));
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
for (i = 0; i < table->n_v_def; i++) {
if (col_nr == dict_get_v_col_mysql_pos(
table->v_cols[i].m_col.ind)) {
break;
}
}
if (i == table->n_v_def) {
return(ULINT_UNDEFINED);
}
return(i);
}
/** Returns a virtual column's name according to its original
MySQL table position.
@param[in] table target table
@param[in] col_nr column number (nth column in the table)
@return column name. */
static
const char*
dict_table_get_v_col_name_mysql(
const dict_table_t* table,
ulint col_nr)
{
ulint i = dict_table_get_v_col_pos_for_mysql(table, col_nr);
if (i == ULINT_UNDEFINED) {
return(NULL);
}
return(dict_table_get_v_col_name(table, i));
}
/** Get nth virtual column according to its original MySQL table position
@param[in] table target table
@param[in] col_nr column number in MySQL Table definition
@return dict_v_col_t ptr */
dict_v_col_t*
dict_table_get_nth_v_col_mysql(
const dict_table_t* table,
ulint col_nr)
{
ulint i = dict_table_get_v_col_pos_for_mysql(table, col_nr);
if (i == ULINT_UNDEFINED) {
return(NULL);
}
return(dict_table_get_nth_v_col(table, i));
}
/** Get all the FTS indexes on a table.
@param[in] table table
@param[out] indexes all FTS indexes on this table
@return number of FTS indexes */
ulint
dict_table_get_all_fts_indexes(
const dict_table_t* table,
ib_vector_t* indexes)
{
dict_index_t* index;
ut_a(ib_vector_size(indexes) == 0);
for (index = dict_table_get_first_index(table);
index;
index = dict_table_get_next_index(index)) {
if (index->type == DICT_FTS) {
ib_vector_push(indexes, &index);
}
}
return(ib_vector_size(indexes));
}
/** Looks for column n in an index.
@param[in] index index
@param[in] n column number
@param[in] inc_prefix true=consider column prefixes too
@param[in] is_virtual true==virtual column
@param[out] prefix_col_pos col num if prefix
@return position in internal representation of the index;
ULINT_UNDEFINED if not contained */
ulint
dict_index_get_nth_col_or_prefix_pos(
const dict_index_t* index,
ulint n,
bool inc_prefix,
bool is_virtual,
ulint* prefix_col_pos)
{
const dict_field_t* field;
const dict_col_t* col;
ulint pos;
ulint n_fields;
ut_ad(index);
ut_ad(index->magic_n == DICT_INDEX_MAGIC_N);
if (prefix_col_pos) {
*prefix_col_pos = ULINT_UNDEFINED;
}
if (is_virtual) {
col = &(dict_table_get_nth_v_col(index->table, n)->m_col);
} else {
col = dict_table_get_nth_col(index->table, n);
}
if (dict_index_is_clust(index)) {
return(dict_col_get_clust_pos(col, index));
}
n_fields = dict_index_get_n_fields(index);
for (pos = 0; pos < n_fields; pos++) {
field = dict_index_get_nth_field(index, pos);
if (col == field->col) {
if (prefix_col_pos) {
*prefix_col_pos = pos;
}
if (inc_prefix || field->prefix_len == 0) {
return(pos);
}
}
}
return(ULINT_UNDEFINED);
}
/** Check if the index contains a column or a prefix of that column.
@param[in] n column number
@param[in] is_virtual whether it is a virtual col
@return whether the index contains the column or its prefix */
bool dict_index_t::contains_col_or_prefix(ulint n, bool is_virtual) const
{
ut_ad(magic_n == DICT_INDEX_MAGIC_N);
if (is_primary()) {
return(!is_virtual);
}
const dict_col_t* col = is_virtual
? &dict_table_get_nth_v_col(table, n)->m_col
: dict_table_get_nth_col(table, n);
for (ulint pos = 0; pos < n_fields; pos++) {
if (col == fields[pos].col) {
return true;
}
}
return false;
}
/********************************************************************//**
Looks for a matching field in an index. The column has to be the same. The
column in index must be complete, or must contain a prefix longer than the
column in index2. That is, we must be able to construct the prefix in index2
from the prefix in index.
@return position in internal representation of the index;
ULINT_UNDEFINED if not contained */
ulint
dict_index_get_nth_field_pos(
/*=========================*/
const dict_index_t* index, /*!< in: index from which to search */
const dict_index_t* index2, /*!< in: index */
ulint n) /*!< in: field number in index2 */
{
const dict_field_t* field;
const dict_field_t* field2;
ulint n_fields;
ulint pos;
ut_ad(index->magic_n == DICT_INDEX_MAGIC_N);
field2 = dict_index_get_nth_field(index2, n);
n_fields = dict_index_get_n_fields(index);
/* Are we looking for a MBR (Minimum Bound Box) field of
a spatial index */
bool is_mbr_fld = (n == 0 && dict_index_is_spatial(index2));
for (pos = 0; pos < n_fields; pos++) {
field = dict_index_get_nth_field(index, pos);
/* The first field of a spatial index is a transformed
MBR (Minimum Bound Box) field made out of original column,
so its field->col still points to original cluster index
col, but the actual content is different. So we cannot
consider them equal if neither of them is MBR field */
if (pos == 0 && dict_index_is_spatial(index) && !is_mbr_fld) {
continue;
}
if (field->col == field2->col
&& (field->prefix_len == 0
|| (field->prefix_len >= field2->prefix_len
&& field2->prefix_len != 0))) {
return(pos);
}
}
return(ULINT_UNDEFINED);
}
/** Parse the table file name into table name and database name.
@tparam dict_frozen whether the caller holds dict_sys.latch
@param[in,out] db_name database name buffer
@param[in,out] tbl_name table name buffer
@param[out] db_name_len database name length
@param[out] tbl_name_len table name length
@return whether the table name is visible to SQL */
template<bool dict_frozen>
bool dict_table_t::parse_name(char (&db_name)[NAME_LEN + 1],
char (&tbl_name)[NAME_LEN + 1],
size_t *db_name_len, size_t *tbl_name_len) const
{
char db_buf[MAX_DATABASE_NAME_LEN + 1];
char tbl_buf[MAX_TABLE_NAME_LEN + 1];
if (!dict_frozen)
dict_sys.freeze(SRW_LOCK_CALL); /* protect against renaming */
ut_ad(dict_sys.frozen());
const size_t db_len= name.dblen();
ut_ad(db_len <= MAX_DATABASE_NAME_LEN);
memcpy(db_buf, mdl_name.m_name, db_len);
db_buf[db_len]= 0;
size_t tbl_len= strlen(mdl_name.m_name + db_len + 1);
const bool is_temp= mdl_name.is_temporary();
if (is_temp);
else if (const char *is_part= static_cast<const char*>
(memchr(mdl_name.m_name + db_len + 1, '#', tbl_len)))
tbl_len= static_cast<size_t>(is_part - &mdl_name.m_name[db_len + 1]);
memcpy(tbl_buf, mdl_name.m_name + db_len + 1, tbl_len);
tbl_buf[tbl_len]= 0;
if (!dict_frozen)
dict_sys.unfreeze();
*db_name_len= filename_to_tablename(db_buf, db_name,
MAX_DATABASE_NAME_LEN + 1, true);
if (is_temp)
return false;
*tbl_name_len= filename_to_tablename(tbl_buf, tbl_name,
MAX_TABLE_NAME_LEN + 1, true);
return true;
}
template bool
dict_table_t::parse_name<>(char(&)[NAME_LEN + 1], char(&)[NAME_LEN + 1],
size_t*, size_t*) const;
dict_table_t *dict_sys_t::acquire_temporary_table(table_id_t id) const noexcept
{
ut_ad(frozen());
ut_ad(id >= DICT_HDR_FIRST_ID);
return temp_id_hash.cell_get(ut_fold_ull(id))->
find(&dict_table_t::id_hash, [id](dict_table_t *t)
{
ut_ad(t->is_temporary());
ut_ad(t->cached);
if (t->id != id)
return false;
t->acquire();
return true;
});
}
dict_table_t *dict_sys_t::find_table(table_id_t id) const noexcept
{
ut_ad(frozen());
return table_id_hash.cell_get(ut_fold_ull(id))->
find(&dict_table_t::id_hash, [id](const dict_table_t *t)
{
ut_ad(!t->is_temporary());
ut_ad(t->cached);
return t->id == id;
});
}
dict_table_t *dict_sys_t::find_table(const span<const char> &name)
const noexcept
{
ut_ad(frozen());
return table_hash.cell_get(my_crc32c(0, name.data(), name.size()))->
find(&dict_table_t::name_hash, [name](const dict_table_t *t)
{
return strlen(t->name.m_name) == name.size() &&
!memcmp(t->name.m_name, name.data(), name.size());
});
}
/** Acquire MDL shared for the table name.
@tparam trylock whether to use non-blocking operation
@param[in,out] table table object
@param[in,out] mdl_context MDL context
@param[out] mdl MDL ticket
@param[in] table_op operation to perform when opening
@return table object after locking MDL shared
@retval nullptr if the table is not readable, or if trylock && MDL blocked */
template<bool trylock>
__attribute__((nonnull, warn_unused_result))
dict_table_t*
dict_acquire_mdl_shared(dict_table_t *table,
MDL_context *mdl_context, MDL_ticket **mdl,
dict_table_op_t table_op)
{
table_id_t table_id= table->id;
char db_buf[NAME_LEN + 1], db_buf1[NAME_LEN + 1];
char tbl_buf[NAME_LEN + 1], tbl_buf1[NAME_LEN + 1];
size_t db_len, tbl_len;
bool unaccessible= false;
if (!table->parse_name<!trylock>(db_buf, tbl_buf, &db_len, &tbl_len))
/* The name of an intermediate table starts with #sql */
return table;
retry:
if (!unaccessible && (!table->is_readable() || table->corrupted))
{
if (*mdl)
{
mdl_context->release_lock(*mdl);
*mdl= nullptr;
}
unaccessible= true;
}
if (!trylock)
table->release();
if (unaccessible)
return nullptr;
if (!trylock)
dict_sys.unfreeze();
{
MDL_request request;
MDL_REQUEST_INIT(&request,MDL_key::TABLE, db_buf, tbl_buf, MDL_SHARED,
MDL_EXPLICIT);
if (trylock
? mdl_context->try_acquire_lock(&request)
: mdl_context->acquire_lock(&request,
/* FIXME: use compatible type, and maybe
remove this parameter altogether! */
static_cast<double>(global_system_variables
.lock_wait_timeout)))
{
*mdl= nullptr;
if (trylock)
return nullptr;
}
else
{
*mdl= request.ticket;
if (trylock && !*mdl)
return nullptr;
}
}
dict_sys.freeze(SRW_LOCK_CALL);
table= dict_sys.find_table(table_id);
if (table)
table->acquire();
if (!table && table_op != DICT_TABLE_OP_OPEN_ONLY_IF_CACHED)
{
dict_sys.unfreeze();
dict_sys.lock(SRW_LOCK_CALL);
table= dict_load_table_on_id(table_id,
table_op == DICT_TABLE_OP_LOAD_TABLESPACE
? DICT_ERR_IGNORE_RECOVER_LOCK
: DICT_ERR_IGNORE_FK_NOKEY);
if (table)
table->acquire();
dict_sys.unlock();
dict_sys.freeze(SRW_LOCK_CALL);
}
if (!table || !table->is_accessible())
{
return_without_mdl:
if (trylock)
dict_sys.unfreeze();
if (*mdl)
{
mdl_context->release_lock(*mdl);
*mdl= nullptr;
}
return nullptr;
}
size_t db1_len, tbl1_len;
if (!table->parse_name<true>(db_buf1, tbl_buf1, &db1_len, &tbl1_len))
{
/* The table was renamed to #sql prefix.
Release MDL (if any) for the old name and return. */
goto return_without_mdl;
}
if (*mdl)
{
if (db_len == db1_len && tbl_len == tbl1_len &&
!memcmp(db_buf, db_buf1, db_len) &&
!memcmp(tbl_buf, tbl_buf1, tbl_len))
{
if (trylock)
dict_sys.unfreeze();
return table;
}
/* The table was renamed. Release MDL for the old name and
try to acquire MDL for the new name. */
mdl_context->release_lock(*mdl);
*mdl= nullptr;
}
db_len= db1_len;
tbl_len= tbl1_len;
memcpy(tbl_buf, tbl_buf1, tbl_len + 1);
memcpy(db_buf, db_buf1, db_len + 1);
goto retry;
}
template dict_table_t* dict_acquire_mdl_shared<false>
(dict_table_t*,MDL_context*,MDL_ticket**,dict_table_op_t);
/** Acquire MDL shared for the table name.
@tparam trylock whether to use non-blocking operation
@param[in,out] table table object
@param[in,out] thd background thread
@param[out] mdl mdl ticket
@param[in] table_op operation to perform when opening
@return table object after locking MDL shared
@retval nullptr if the table is not readable, or if trylock && MDL blocked */
template<bool trylock>
dict_table_t*
dict_acquire_mdl_shared(dict_table_t *table,
THD *thd,
MDL_ticket **mdl,
dict_table_op_t table_op)
{
if (!table || !mdl)
return table;
MDL_context *mdl_context= static_cast<MDL_context*>(thd_mdl_context(thd));
size_t db_len;
if (trylock)
{
dict_sys.freeze(SRW_LOCK_CALL);
db_len= dict_get_db_name_len(table->name.m_name);
dict_sys.unfreeze();
}
else
{
ut_ad(dict_sys.frozen_not_locked());
db_len= dict_get_db_name_len(table->name.m_name);
}
if (db_len == 0)
return table; /* InnoDB system tables are not covered by MDL */
return mdl_context
? dict_acquire_mdl_shared<trylock>(table, mdl_context, mdl, table_op)
: nullptr;
}
template dict_table_t* dict_acquire_mdl_shared<false>
(dict_table_t*,THD*,MDL_ticket**,dict_table_op_t);
template dict_table_t* dict_acquire_mdl_shared<true>
(dict_table_t*,THD*,MDL_ticket**,dict_table_op_t);
/** Look up a table by numeric identifier.
@param[in] table_id table identifier
@param[in] dict_locked data dictionary locked
@param[in] table_op operation to perform when opening
@param[in,out] thd background thread, or NULL to not acquire MDL
@param[out] mdl mdl ticket, or NULL
@return table, NULL if does not exist */
dict_table_t *dict_table_open_on_id(table_id_t table_id, bool dict_locked,
dict_table_op_t table_op, THD *thd,
MDL_ticket **mdl)
{
if (!dict_locked)
dict_sys.freeze(SRW_LOCK_CALL);
dict_table_t *table= dict_sys.find_table(table_id);
if (table)
{
table->acquire();
if (thd && !dict_locked)
table= dict_acquire_mdl_shared<false>(table, thd, mdl, table_op);
}
else if (table_op != DICT_TABLE_OP_OPEN_ONLY_IF_CACHED)
{
if (!dict_locked)
{
dict_sys.unfreeze();
dict_sys.lock(SRW_LOCK_CALL);
}
table= dict_load_table_on_id(table_id,
table_op == DICT_TABLE_OP_LOAD_TABLESPACE
? DICT_ERR_IGNORE_RECOVER_LOCK
: DICT_ERR_IGNORE_FK_NOKEY);
if (table)
table->acquire();
if (!dict_locked)
{
dict_sys.unlock();
if (table && thd)
{
dict_sys.freeze(SRW_LOCK_CALL);
table= dict_acquire_mdl_shared<false>(table, thd, mdl, table_op);
dict_sys.unfreeze();
}
return table;
}
}
if (!dict_locked)
dict_sys.unfreeze();
return table;
}
/********************************************************************//**
Looks for column n position in the clustered index.
@return position in internal representation of the clustered index */
unsigned
dict_table_get_nth_col_pos(
/*=======================*/
const dict_table_t* table, /*!< in: table */
ulint n, /*!< in: column number */
ulint* prefix_col_pos)
{
ulint pos= dict_index_get_nth_col_pos(dict_table_get_first_index(table),
n, prefix_col_pos);
DBUG_ASSERT(pos <= dict_index_t::MAX_N_FIELDS);
return static_cast<unsigned>(pos);
}
/********************************************************************//**
Checks if a column is in the ordering columns of the clustered index of a
table. Column prefixes are treated like whole columns.
@return TRUE if the column, or its prefix, is in the clustered key */
ibool
dict_table_col_in_clustered_key(
/*============================*/
const dict_table_t* table, /*!< in: table */
ulint n) /*!< in: column number */
{
const dict_index_t* index;
const dict_field_t* field;
const dict_col_t* col;
ulint pos;
ulint n_fields;
col = dict_table_get_nth_col(table, n);
index = dict_table_get_first_index(table);
n_fields = dict_index_get_n_unique(index);
for (pos = 0; pos < n_fields; pos++) {
field = dict_index_get_nth_field(index, pos);
if (col == field->col) {
return(TRUE);
}
}
return(FALSE);
}
/** Initialise the data dictionary cache. */
void dict_sys_t::create() noexcept
{
ut_ad(this == &dict_sys);
ut_ad(!is_initialised());
m_initialised= true;
UT_LIST_INIT(table_LRU, &dict_table_t::table_LRU);
UT_LIST_INIT(table_non_LRU, &dict_table_t::table_LRU);
const ulint hash_size = buf_pool_get_curr_size()
/ (DICT_POOL_PER_TABLE_HASH * UNIV_WORD_SIZE);
table_hash.create(hash_size);
table_id_hash.create(hash_size);
temp_id_hash.create(hash_size);
latch.SRW_LOCK_INIT(dict_operation_lock_key);
if (!srv_read_only_mode)
{
dict_foreign_err_file= os_file_create_tmpfile();
ut_a(dict_foreign_err_file);
}
mysql_mutex_init(dict_foreign_err_mutex_key, &dict_foreign_err_mutex,
nullptr);
}
void dict_sys_t::lock_wait(SRW_LOCK_ARGS(const char *file, unsigned line)) noexcept
{
ulonglong now= my_hrtime_coarse().val, old= 0;
if (latch_ex_wait_start.compare_exchange_strong
(old, now, std::memory_order_relaxed, std::memory_order_relaxed))
{
latch.wr_lock(SRW_LOCK_ARGS(file, line));
latch_ex_wait_start.store(0, std::memory_order_relaxed);
return;
}
ut_ad(old);
/* We could have old > now due to our use of my_hrtime_coarse(). */
ulong waited= old <= now ? static_cast<ulong>((now - old) / 1000000) : 0;
const ulong threshold= srv_fatal_semaphore_wait_threshold;
if (waited >= threshold)
ib::fatal() << fatal_msg;
if (waited > threshold / 4)
ib::warn() << "A long wait (" << waited
<< " seconds) was observed for dict_sys.latch";
latch.wr_lock(SRW_LOCK_ARGS(file, line));
}
#ifdef UNIV_PFS_RWLOCK
ATTRIBUTE_NOINLINE void dict_sys_t::unlock() noexcept
{
latch.wr_unlock();
}
ATTRIBUTE_NOINLINE void dict_sys_t::freeze(const char *file, unsigned line) noexcept
{
latch.rd_lock(file, line);
}
ATTRIBUTE_NOINLINE void dict_sys_t::unfreeze() noexcept
{
latch.rd_unlock();
}
#endif /* UNIV_PFS_RWLOCK */
/** Report an error about failing to open a table.
@param name table name */
static void dict_table_open_failed(const table_name_t &name)
{
my_printf_error(ER_TABLE_CORRUPT,
"Table %`.*s.%`s is corrupted."
" Please drop the table and recreate.",
MYF(ME_ERROR_LOG),
int(name.dblen()), name.m_name, name.basename());
}
/**********************************************************************//**
Returns a table object and increments its open handle count.
NOTE! This is a high-level function to be used mainly from outside the
'dict' directory. Inside this directory dict_table_get_low
is usually the appropriate function.
@param[in] table_name Table name
@param[in] dict_locked whether dict_sys.latch is being held exclusively
@param[in] ignore_err error to be ignored when loading the table
@return table
@retval nullptr if does not exist */
dict_table_t*
dict_table_open_on_name(
const char* table_name,
bool dict_locked,
dict_err_ignore_t ignore_err)
{
dict_table_t *table;
DBUG_ENTER("dict_table_open_on_name");
DBUG_PRINT("dict_table_open_on_name", ("table: '%s'", table_name));
const span<const char> name{table_name, strlen(table_name)};
if (!dict_locked)
{
dict_sys.freeze(SRW_LOCK_CALL);
table= dict_sys.find_table(name);
if (table)
{
ut_ad(table->cached);
if (!(ignore_err & ~DICT_ERR_IGNORE_FK_NOKEY) &&
!table->is_readable() && table->corrupted)
{
ulint algo= table->space->get_compression_algo();
if (algo <= PAGE_ALGORITHM_LAST && !fil_comp_algo_loaded(algo))
my_printf_error(ER_PROVIDER_NOT_LOADED,
"Table %`.*s.%`s is compressed with %s,"
" which is not currently loaded. "
"Please load the %s provider plugin"
" to open the table",
MYF(ME_ERROR_LOG),
int(table->name.dblen()), table->name.m_name,
table->name.basename(),
page_compression_algorithms[algo],
page_compression_algorithms[algo]);
else
dict_table_open_failed(table->name);
dict_sys.unfreeze();
DBUG_RETURN(nullptr);
}
table->acquire();
dict_sys.unfreeze();
DBUG_RETURN(table);
}
dict_sys.unfreeze();
dict_sys.lock(SRW_LOCK_CALL);
}
table= dict_sys.load_table(name, ignore_err);
if (table)
{
ut_ad(table->cached);
if (!(ignore_err & ~DICT_ERR_IGNORE_FK_NOKEY) &&
!table->is_readable() && table->corrupted)
{
dict_table_open_failed(table->name);
if (!dict_locked)
dict_sys.unlock();
DBUG_RETURN(nullptr);
}
table->acquire();
}
ut_ad(dict_lru_validate());
if (!dict_locked)
dict_sys.unlock();
DBUG_RETURN(table);
}
/**********************************************************************//**
Adds system columns to a table object. */
void
dict_table_add_system_columns(
/*==========================*/
dict_table_t* table, /*!< in/out: table */
mem_heap_t* heap) /*!< in: temporary heap */
{
ut_ad(table->n_def == table->n_cols - DATA_N_SYS_COLS);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(!table->cached);
/* NOTE: the system columns MUST be added in the following order
(so that they can be indexed by the numerical value of DATA_ROW_ID,
etc.) and as the last columns of the table memory object.
The clustered index will not always physically contain all system
columns. */
dict_mem_table_add_col(table, heap, "DB_ROW_ID", DATA_SYS,
DATA_ROW_ID | DATA_NOT_NULL,
DATA_ROW_ID_LEN);
compile_time_assert(DATA_ROW_ID == 0);
dict_mem_table_add_col(table, heap, "DB_TRX_ID", DATA_SYS,
DATA_TRX_ID | DATA_NOT_NULL,
DATA_TRX_ID_LEN);
compile_time_assert(DATA_TRX_ID == 1);
dict_mem_table_add_col(table, heap, "DB_ROLL_PTR", DATA_SYS,
DATA_ROLL_PTR | DATA_NOT_NULL,
DATA_ROLL_PTR_LEN);
compile_time_assert(DATA_ROLL_PTR == 2);
/* This check reminds that if a new system column is added to
the program, it should be dealt with here */
compile_time_assert(DATA_N_SYS_COLS == 3);
}
/** Add the table definition to the data dictionary cache */
void dict_table_t::add_to_cache()
{
cached = TRUE;
dict_sys.add(this);
}
/** Add a table definition to the data dictionary cache */
inline void dict_sys_t::add(dict_table_t *table) noexcept
{
ut_ad(!table->name_hash);
ut_ad(!table->id_hash);
table->autoinc_mutex.init();
table->lock_mutex_init();
const char *name= table->name.m_name;
dict_table_t **prev= table_hash.cell_get(my_crc32c(0, name, strlen(name)))->
search(&dict_table_t::name_hash, [name](const dict_table_t *t)
{
if (!t) return true;
ut_ad(t->cached);
ut_a(strcmp(t->name.m_name, name));
return false;
});
*prev= table;
prev= (table->is_temporary() ? temp_id_hash : table_id_hash).
cell_get(ut_fold_ull(table->id))->
search(&dict_table_t::id_hash, [table](const dict_table_t *t)
{
if (!t) return true;
ut_ad(t->cached);
ut_a(t->id != table->id);
return false;
});
*prev= table;
UT_LIST_ADD_FIRST(table->can_be_evicted ? table_LRU : table_non_LRU, table);
ut_ad(dict_lru_validate());
}
/** Test whether a table can be evicted from dict_sys.table_LRU.
@param table table to be considered for eviction
@return whether the table can be evicted */
TRANSACTIONAL_TARGET
static bool dict_table_can_be_evicted(dict_table_t *table)
{
ut_ad(dict_sys.locked());
ut_a(table->can_be_evicted);
ut_a(table->foreign_set.empty());
ut_a(table->referenced_set.empty());
if (table->get_ref_count() == 0) {
/* The transaction commit and rollback are called from
outside the handler interface. This means that there is
a window where the table->n_ref_count can be zero but
the table instance is in "use". */
if (lock_table_has_locks(table)) {
return false;
}
#ifdef BTR_CUR_HASH_ADAPT
/* We cannot really evict the table if adaptive hash
index entries are pointing to any of its indexes. */
for (const dict_index_t* index
= dict_table_get_first_index(table);
index; index = dict_table_get_next_index(index)) {
if (index->n_ahi_pages()) {
return false;
}
}
#endif /* BTR_CUR_HASH_ADAPT */
ut_ad(!table->fts);
return true;
}
return false;
}
#ifdef BTR_CUR_HASH_ADAPT
/** @return a clone of this */
dict_index_t *dict_index_t::clone() const
{
ut_ad(n_fields);
ut_ad(is_btree());
ut_ad(online_status == ONLINE_INDEX_COMPLETE);
ut_ad(is_committed());
ut_ad(!is_dummy);
ut_ad(!parser);
ut_ad(!online_log);
ut_ad(!rtr_track);
const size_t size= sizeof *this + n_fields * sizeof(*fields) +
#ifdef BTR_CUR_ADAPT
sizeof *search_info +
#endif
1 + strlen(name) +
n_uniq * (sizeof *stat_n_diff_key_vals +
sizeof *stat_n_sample_sizes +
sizeof *stat_n_non_null_key_vals);
mem_heap_t* heap= mem_heap_create(size);
dict_index_t *index= static_cast<dict_index_t*>
(mem_heap_alloc(heap, sizeof *this));
*index= *this;
index->lock.SRW_LOCK_INIT(index_tree_rw_lock_key);
index->heap= heap;
index->name= mem_heap_strdup(heap, name);
index->fields= static_cast<dict_field_t*>
(mem_heap_dup(heap, fields, n_fields * sizeof *fields));
#ifdef BTR_CUR_ADAPT
index->search_info= btr_search_info_create(index->heap);
#endif /* BTR_CUR_ADAPT */
index->stat_n_diff_key_vals= static_cast<ib_uint64_t*>
(mem_heap_zalloc(heap, n_uniq * sizeof *stat_n_diff_key_vals));
index->stat_n_sample_sizes= static_cast<ib_uint64_t*>
(mem_heap_zalloc(heap, n_uniq * sizeof *stat_n_sample_sizes));
index->stat_n_non_null_key_vals= static_cast<ib_uint64_t*>
(mem_heap_zalloc(heap, n_uniq * sizeof *stat_n_non_null_key_vals));
new (&index->zip_pad.mutex) std::mutex();
return index;
}
/** Clone this index for lazy dropping of the adaptive hash.
@return this or a clone */
dict_index_t *dict_index_t::clone_if_needed()
{
if (!search_info->ref_count)
return this;
dict_index_t *prev= UT_LIST_GET_PREV(indexes, this);
table->autoinc_mutex.wr_lock();
UT_LIST_REMOVE(table->indexes, this);
UT_LIST_ADD_LAST(table->freed_indexes, this);
dict_index_t *index= clone();
set_freed();
if (prev)
UT_LIST_INSERT_AFTER(table->indexes, prev, index);
else
UT_LIST_ADD_FIRST(table->indexes, index);
table->autoinc_mutex.wr_unlock();
return index;
}
#endif /* BTR_CUR_HASH_ADAPT */
/** Evict unused, unlocked tables from table_LRU.
@param half whether to consider half the tables only (instead of all)
@return number of tables evicted */
ulint dict_sys_t::evict_table_LRU(bool half) noexcept
{
#ifdef MYSQL_DYNAMIC_PLUGIN
constexpr ulint max_tables = 400;
#else
extern ulong tdc_size;
const ulint max_tables = tdc_size;
#endif
ulint n_evicted = 0;
lock(SRW_LOCK_CALL);
ut_ad(dict_lru_validate());
const ulint len = UT_LIST_GET_LEN(table_LRU);
if (len < max_tables) {
func_exit:
unlock();
return(n_evicted);
}
const ulint check_up_to = half ? len / 2 : 0;
ulint i = len;
/* Find a suitable candidate to evict from the cache. Don't scan the
entire LRU list. Only scan pct_check list entries. */
for (dict_table_t *table = UT_LIST_GET_LAST(table_LRU);
table && i > check_up_to && (len - n_evicted) > max_tables; --i) {
dict_table_t* prev_table = UT_LIST_GET_PREV(table_LRU, table);
if (dict_table_can_be_evicted(table)) {
remove(table, true);
++n_evicted;
}
table = prev_table;
}
goto func_exit;
}
/** Looks for an index with the given id given a table instance.
@param[in] table table instance
@param[in] id index id
@return index or NULL */
dict_index_t*
dict_table_find_index_on_id(
const dict_table_t* table,
index_id_t id)
{
dict_index_t* index;
for (index = dict_table_get_first_index(table);
index != NULL;
index = dict_table_get_next_index(index)) {
if (id == index->id) {
/* Found */
return(index);
}
}
return(NULL);
}
/** Function object to remove a foreign key constraint from the
referenced_set of the referenced table. The foreign key object is
also removed from the dictionary cache. The foreign key constraint
is not removed from the foreign_set of the table containing the
constraint. */
struct dict_foreign_remove_partial
{
void operator()(dict_foreign_t* foreign) {
dict_table_t* table = foreign->referenced_table;
if (table != NULL) {
table->referenced_set.erase(foreign);
}
dict_foreign_free(foreign);
}
};
/** This function returns a new path name after replacing the basename
in an old path with a new basename. The old_path is a full path
name including the extension. The tablename is in the normal
form "databasename/tablename". The new base name is found after
the forward slash. Both input strings are null terminated.
This function allocates memory to be returned. It is the callers
responsibility to free the return value after it is no longer needed.
@param[in] old_path Pathname
@param[in] tablename Contains new base name
@return own: new full pathname */
static char *dir_pathname(const char *old_path, span<const char> tablename)
{
/* Split the tablename into its database and table name components.
They are separated by a '/'. */
const char *base_name= tablename.data();
for (const char *last= tablename.end(); last > tablename.data(); last--)
{
if (last[-1] == '/')
{
base_name= last;
break;
}
}
const size_t base_name_len= tablename.end() - base_name;
/* Find the offset of the last slash. We will strip off the
old basename.ibd which starts after that slash. */
const char *last_slash= strrchr(old_path, '/');
#ifdef _WIN32
if (const char *last= strrchr(old_path, '\\'))
if (last > last_slash)
last_slash= last;
#endif
size_t dir_len= last_slash
? size_t(last_slash - old_path)
: strlen(old_path);
/* allocate a new path and move the old directory path to it. */
size_t new_path_len= dir_len + base_name_len + sizeof "/.ibd";
char *new_path= static_cast<char*>(ut_malloc_nokey(new_path_len));
memcpy(new_path, old_path, dir_len);
snprintf(new_path + dir_len, new_path_len - dir_len, "/%.*s.ibd",
int(base_name_len), base_name);
return new_path;
}
/** Rename the data file.
@param new_name name of the table
@param replace whether to replace the file with the new name
(as part of rolling back TRUNCATE) */
dberr_t
dict_table_t::rename_tablespace(span<const char> new_name, bool replace) const
{
ut_ad(dict_table_is_file_per_table(this));
ut_ad(!is_temporary());
if (!space)
return DB_SUCCESS;
const char *old_path= UT_LIST_GET_FIRST(space->chain)->name;
const bool data_dir= DICT_TF_HAS_DATA_DIR(flags);
char *path= data_dir
? dir_pathname(old_path, new_name)
: fil_make_filepath(nullptr, new_name, IBD, false);
dberr_t err;
if (!path)
err= DB_OUT_OF_MEMORY;
else if (!strcmp(path, old_path))
err= DB_SUCCESS;
else if (data_dir &&
DB_SUCCESS != RemoteDatafile::create_link_file(new_name, path))
err= DB_TABLESPACE_EXISTS;
else
{
space->x_lock();
err= space->rename(path, true, replace);
if (data_dir)
{
if (err == DB_SUCCESS)
new_name= {name.m_name, strlen(name.m_name)};
RemoteDatafile::delete_link_file(new_name);
}
space->x_unlock();
}
ut_free(path);
return err;
}
/**********************************************************************//**
Renames a table object.
@return TRUE if success */
dberr_t
dict_table_rename_in_cache(
/*=======================*/
dict_table_t* table, /*!< in/out: table */
span<const char> new_name, /*!< in: new name */
bool replace_new_file)
/*!< in: whether to replace the
file with the new name
(as part of rolling back TRUNCATE) */
{
dict_foreign_t* foreign;
char old_name[MAX_FULL_NAME_LEN + 1];
ut_ad(dict_sys.locked());
/* store the old/current name to an automatic variable */
const size_t old_name_len = strlen(table->name.m_name);
ut_a(old_name_len < sizeof old_name);
strcpy(old_name, table->name.m_name);
if (!dict_table_is_file_per_table(table)) {
} else if (dberr_t err = table->rename_tablespace(new_name,
replace_new_file)) {
return err;
}
/* Remove table from the hash tables of tables */
dict_sys.table_hash.cell_get(my_crc32c(0, table->name.m_name,
old_name_len))
->remove(*table, &dict_table_t::name_hash);
bool keep_mdl_name = !table->name.is_temporary();
if (!keep_mdl_name) {
} else if (const char* s = static_cast<const char*>
(memchr(new_name.data(), '/', new_name.size()))) {
keep_mdl_name = new_name.end() - s >= 5
&& !memcmp(s, "/#sql", 5);
}
if (keep_mdl_name) {
/* Preserve the original table name for
dict_table_t::parse_name() and dict_acquire_mdl_shared(). */
table->mdl_name.m_name = mem_heap_strdup(table->heap,
table->name.m_name);
}
if (new_name.size() > strlen(table->name.m_name)) {
/* We allocate MAX_FULL_NAME_LEN + 1 bytes here to avoid
memory fragmentation, we assume a repeated calls of
ut_realloc() with the same size do not cause fragmentation */
ut_a(new_name.size() <= MAX_FULL_NAME_LEN);
table->name.m_name = static_cast<char*>(
ut_realloc(table->name.m_name, MAX_FULL_NAME_LEN + 1));
}
memcpy(table->name.m_name, new_name.data(), new_name.size());
table->name.m_name[new_name.size()] = '\0';
if (!keep_mdl_name) {
table->mdl_name.m_name = table->name.m_name;
}
/* Add table to hash table of tables */
ut_ad(!table->name_hash);
dict_table_t** after = reinterpret_cast<dict_table_t**>(
&dict_sys.table_hash.cell_get(my_crc32c(0, new_name.data(),
new_name.size()))
->node);
for (; *after; after = &(*after)->name_hash) {
ut_ad((*after)->cached);
ut_a(strcmp((*after)->name.m_name, new_name.data()));
}
*after = table;
if (table->name.is_temporary()) {
/* In ALTER TABLE we think of the rename table operation
in the direction table -> temporary table (#sql...)
as dropping the table with the old name and creating
a new with the new name. Thus we kind of drop the
constraints from the dictionary cache here. The foreign key
constraints will be inherited to the new table from the
system tables through a call of dict_load_foreigns. */
/* Remove the foreign constraints from the cache */
std::for_each(table->foreign_set.begin(),
table->foreign_set.end(),
dict_foreign_remove_partial());
table->foreign_set.clear();
/* Reset table field in referencing constraints */
for (dict_foreign_set::iterator it
= table->referenced_set.begin();
it != table->referenced_set.end();
++it) {
foreign = *it;
foreign->referenced_table = NULL;
foreign->referenced_index = NULL;
}
/* Make the set of referencing constraints empty */
table->referenced_set.clear();
return(DB_SUCCESS);
}
/* Update the table name fields in foreign constraints, and update also
the constraint id of new format >= 4.0.18 constraints. Note that at
this point we have already changed table->name to the new name. */
dict_foreign_set fk_set;
for (;;) {
dict_foreign_set::iterator it
= table->foreign_set.begin();
if (it == table->foreign_set.end()) {
break;
}
foreign = *it;
if (foreign->referenced_table) {
foreign->referenced_table->referenced_set.erase(foreign);
}
if (strlen(foreign->foreign_table_name)
< strlen(table->name.m_name)) {
/* Allocate a longer name buffer;
TODO: store buf len to save memory */
foreign->foreign_table_name = mem_heap_strdup(
foreign->heap, table->name.m_name);
dict_mem_foreign_table_name_lookup_set(foreign, TRUE);
} else {
strcpy(foreign->foreign_table_name,
table->name.m_name);
dict_mem_foreign_table_name_lookup_set(foreign, FALSE);
}
if (strchr(foreign->id, '/')) {
/* This is a >= 4.0.18 format id */
ulint db_len;
char* old_id;
char old_name_cs_filename[MAX_FULL_NAME_LEN+1];
uint errors = 0;
/* All table names are internally stored in charset
my_charset_filename (except the temp tables and the
partition identifier suffix in partition tables). The
foreign key constraint names are internally stored
in UTF-8 charset. The variable fkid here is used
to store foreign key constraint name in charset
my_charset_filename for comparison further below. */
char fkid[MAX_TABLE_NAME_LEN * 2 + 20];
/* The old table name in my_charset_filename is stored
in old_name_cs_filename */
strcpy(old_name_cs_filename, old_name);
old_name_cs_filename[MAX_FULL_NAME_LEN] = '\0';
if (!dict_table_t::is_temporary_name(old_name)) {
innobase_convert_to_system_charset(
strchr(old_name_cs_filename, '/') + 1,
strchr(old_name, '/') + 1,
MAX_TABLE_NAME_LEN, &errors);
if (errors) {
/* There has been an error to convert
old table into UTF-8. This probably
means that the old table name is
actually in UTF-8. */
innobase_convert_to_filename_charset(
strchr(old_name_cs_filename,
'/') + 1,
strchr(old_name, '/') + 1,
MAX_TABLE_NAME_LEN);
} else {
/* Old name already in
my_charset_filename */
strcpy(old_name_cs_filename, old_name);
old_name_cs_filename[MAX_FULL_NAME_LEN]
= '\0';
}
}
strncpy(fkid, foreign->id, (sizeof fkid) - 1);
fkid[(sizeof fkid) - 1] = '\0';
const bool on_tmp = dict_table_t::is_temporary_name(
fkid);
if (!on_tmp) {
innobase_convert_to_filename_charset(
strchr(fkid, '/') + 1,
strchr(foreign->id, '/') + 1,
MAX_TABLE_NAME_LEN+20);
}
old_id = mem_strdup(foreign->id);
if (strlen(fkid) > strlen(old_name_cs_filename)
+ ((sizeof dict_ibfk) - 1)
&& !memcmp(fkid, old_name_cs_filename,
strlen(old_name_cs_filename))
&& !memcmp(fkid + strlen(old_name_cs_filename),
dict_ibfk, (sizeof dict_ibfk) - 1)) {
/* This is a generated >= 4.0.18 format id */
char table_name[MAX_TABLE_NAME_LEN + 1];
uint errors = 0;
if (strlen(table->name.m_name)
> strlen(old_name)) {
foreign->id = static_cast<char*>(
mem_heap_alloc(
foreign->heap,
strlen(table->name.m_name)
+ strlen(old_id) + 1));
}
/* Convert the table name to UTF-8 */
strncpy(table_name, table->name.m_name,
MAX_TABLE_NAME_LEN);
table_name[MAX_TABLE_NAME_LEN] = '\0';
innobase_convert_to_system_charset(
strchr(table_name, '/') + 1,
strchr(table->name.m_name, '/') + 1,
MAX_TABLE_NAME_LEN, &errors);
if (errors) {
/* Table name could not be converted
from charset my_charset_filename to
UTF-8. This means that the table name
is already in UTF-8 (#mysql50#). */
strncpy(table_name, table->name.m_name,
MAX_TABLE_NAME_LEN);
table_name[MAX_TABLE_NAME_LEN] = '\0';
}
/* Replace the prefix 'databasename/tablename'
with the new names */
strcpy(foreign->id, table_name);
if (on_tmp) {
strcat(foreign->id,
old_id + strlen(old_name));
} else {
sprintf(strchr(foreign->id, '/') + 1,
"%s%s",
strchr(table_name, '/') +1,
strstr(old_id, "_ibfk_") );
}
} else {
/* This is a >= 4.0.18 format id where the user
gave the id name */
db_len = dict_get_db_name_len(
table->name.m_name) + 1;
if (db_len - 1
> dict_get_db_name_len(foreign->id)) {
foreign->id = static_cast<char*>(
mem_heap_alloc(
foreign->heap,
db_len + strlen(old_id) + 1));
}
/* Replace the database prefix in id with the
one from table->name */
memcpy(foreign->id,
table->name.m_name, db_len);
strcpy(foreign->id + db_len,
dict_remove_db_name(old_id));
}
ut_free(old_id);
}
table->foreign_set.erase(it);
fk_set.insert(foreign);
if (foreign->referenced_table) {
foreign->referenced_table->referenced_set.insert(foreign);
}
}
ut_a(table->foreign_set.empty());
table->foreign_set.swap(fk_set);
for (dict_foreign_set::iterator it = table->referenced_set.begin();
it != table->referenced_set.end();
++it) {
foreign = *it;
if (strlen(foreign->referenced_table_name)
< strlen(table->name.m_name)) {
/* Allocate a longer name buffer;
TODO: store buf len to save memory */
foreign->referenced_table_name = mem_heap_strdup(
foreign->heap, table->name.m_name);
dict_mem_referenced_table_name_lookup_set(
foreign, TRUE);
} else {
/* Use the same buffer */
strcpy(foreign->referenced_table_name,
table->name.m_name);
dict_mem_referenced_table_name_lookup_set(
foreign, FALSE);
}
}
return(DB_SUCCESS);
}
/** Evict a table definition from the InnoDB data dictionary cache.
@param[in,out] table cached table definition to be evicted
@param[in] lru whether this is part of least-recently-used eviction
@param[in] keep whether to keep (not free) the object */
void dict_sys_t::remove(dict_table_t* table, bool lru, bool keep) noexcept
{
dict_foreign_t* foreign;
dict_index_t* index;
ut_ad(dict_lru_validate());
ut_a(table->get_ref_count() == 0);
ut_a(table->n_rec_locks == 0);
ut_ad(find(table));
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
/* Remove the foreign constraints from the cache */
std::for_each(table->foreign_set.begin(), table->foreign_set.end(),
dict_foreign_remove_partial());
table->foreign_set.clear();
/* Reset table field in referencing constraints */
for (dict_foreign_set::iterator it = table->referenced_set.begin();
it != table->referenced_set.end();
++it) {
foreign = *it;
foreign->referenced_table = NULL;
foreign->referenced_index = NULL;
}
/* Remove the indexes from the cache */
for (index = UT_LIST_GET_LAST(table->indexes);
index != NULL;
index = UT_LIST_GET_LAST(table->indexes)) {
dict_index_remove_from_cache_low(table, index, lru);
}
/* Remove table from the hash tables of tables */
table_hash.cell_get(my_crc32c(0, table->name.m_name,
strlen(table->name.m_name)))
->remove(*table, &dict_table_t::name_hash);
(table->is_temporary() ? temp_id_hash : table_id_hash)
.cell_get(ut_fold_ull(table->id))
->remove(*table, &dict_table_t::id_hash);
/* Remove table from LRU or non-LRU list. */
if (table->can_be_evicted) {
UT_LIST_REMOVE(table_LRU, table);
} else {
UT_LIST_REMOVE(table_non_LRU, table);
}
/* Free virtual column template if any */
if (table->vc_templ != NULL) {
dict_free_vc_templ(table->vc_templ);
UT_DELETE(table->vc_templ);
}
table->lock_mutex_destroy();
if (keep) {
table->autoinc_mutex.destroy();
return;
}
#ifdef BTR_CUR_HASH_ADAPT
if (table->fts) {
fts_optimize_remove_table(table);
table->fts->~fts_t();
table->fts = nullptr;
}
table->autoinc_mutex.wr_lock();
ulint freed = UT_LIST_GET_LEN(table->freed_indexes);
table->vc_templ = NULL;
table->id = 0;
table->autoinc_mutex.wr_unlock();
if (UNIV_UNLIKELY(freed != 0)) {
return;
}
#endif /* BTR_CUR_HASH_ADAPT */
table->autoinc_mutex.destroy();
dict_mem_table_free(table);
}
/****************************************************************//**
If the given column name is reserved for InnoDB system columns, return
TRUE.
@return TRUE if name is reserved */
ibool
dict_col_name_is_reserved(
/*======================*/
const char* name) /*!< in: column name */
{
static const char* reserved_names[] = {
"DB_ROW_ID", "DB_TRX_ID", "DB_ROLL_PTR"
};
compile_time_assert(UT_ARR_SIZE(reserved_names) == DATA_N_SYS_COLS);
for (ulint i = 0; i < UT_ARR_SIZE(reserved_names); i++) {
if (innobase_strcasecmp(name, reserved_names[i]) == 0) {
return(TRUE);
}
}
return(FALSE);
}
/** Adds an index to the dictionary cache, with possible indexing newly
added column.
@param[in,out] index index; NOTE! The index memory
object is freed in this function!
@param[in] page_no root page number of the index
@param[in] add_v virtual columns being added along with ADD INDEX
@return DB_SUCCESS, or DB_CORRUPTION */
dberr_t
dict_index_add_to_cache(
dict_index_t*& index,
ulint page_no,
const dict_add_v_col_t* add_v)
{
dict_index_t* new_index;
ulint n_ord;
ulint i;
ut_ad(dict_sys.locked());
ut_ad(index->n_def == index->n_fields);
ut_ad(index->magic_n == DICT_INDEX_MAGIC_N);
ut_ad(!dict_index_is_online_ddl(index));
ut_ad(!dict_index_is_ibuf(index));
ut_d(mem_heap_validate(index->heap));
ut_a(!dict_index_is_clust(index)
|| UT_LIST_GET_LEN(index->table->indexes) == 0);
ut_ad(dict_index_is_clust(index) || !index->table->no_rollback());
if (!dict_index_find_cols(index, add_v)) {
dict_mem_index_free(index);
index = NULL;
return DB_CORRUPTION;
}
/* Build the cache internal representation of the index,
containing also the added system fields */
if (dict_index_is_clust(index)) {
new_index = dict_index_build_internal_clust(index);
} else {
new_index = (index->type & DICT_FTS)
? dict_index_build_internal_fts(index)
: dict_index_build_internal_non_clust(index);
new_index->n_core_null_bytes = static_cast<uint8_t>(
UT_BITS_IN_BYTES(unsigned(new_index->n_nullable)));
}
/* Set the n_fields value in new_index to the actual defined
number of fields in the cache internal representation */
new_index->n_fields = new_index->n_def;
new_index->trx_id = index->trx_id;
new_index->set_committed(index->is_committed());
n_ord = new_index->n_uniq;
/* Flag the ordering columns and also set column max_prefix */
for (i = 0; i < n_ord; i++) {
const dict_field_t* field
= dict_index_get_nth_field(new_index, i);
/* Check the column being added in the index for
the first time and flag the ordering column. */
if (field->col->ord_part == 0 ) {
field->col->max_prefix = field->prefix_len;
field->col->ord_part = 1;
} else if (field->prefix_len == 0) {
/* Set the max_prefix for a column to 0 if
its prefix length is 0 (for this index)
even if it was a part of any other index
with some prefix length. */
field->col->max_prefix = 0;
} else if (field->col->max_prefix != 0
&& field->prefix_len
> field->col->max_prefix) {
/* Set the max_prefix value based on the
prefix_len. */
ut_ad(field->col->is_binary()
|| field->prefix_len % field->col->mbmaxlen == 0
|| field->prefix_len % 4 == 0);
field->col->max_prefix = field->prefix_len;
}
ut_ad(field->col->ord_part == 1);
}
new_index->stat_n_diff_key_vals =
static_cast<ib_uint64_t*>(mem_heap_zalloc(
new_index->heap,
dict_index_get_n_unique(new_index)
* sizeof(*new_index->stat_n_diff_key_vals)));
new_index->stat_n_sample_sizes =
static_cast<ib_uint64_t*>(mem_heap_zalloc(
new_index->heap,
dict_index_get_n_unique(new_index)
* sizeof(*new_index->stat_n_sample_sizes)));
new_index->stat_n_non_null_key_vals =
static_cast<ib_uint64_t*>(mem_heap_zalloc(
new_index->heap,
dict_index_get_n_unique(new_index)
* sizeof(*new_index->stat_n_non_null_key_vals)));
new_index->stat_index_size = 1;
new_index->stat_n_leaf_pages = 1;
new_index->stat_defrag_n_pages_freed = 0;
new_index->stat_defrag_n_page_split = 0;
new_index->stat_defrag_sample_next_slot = 0;
memset(&new_index->stat_defrag_data_size_sample,
0x0, sizeof(ulint) * STAT_DEFRAG_DATA_SIZE_N_SAMPLE);
/* Add the new index as the last index for the table */
UT_LIST_ADD_LAST(new_index->table->indexes, new_index);
#ifdef BTR_CUR_ADAPT
new_index->search_info = btr_search_info_create(new_index->heap);
#endif /* BTR_CUR_ADAPT */
new_index->page = unsigned(page_no);
new_index->lock.SRW_LOCK_INIT(index_tree_rw_lock_key);
new_index->n_core_fields = new_index->n_fields;
dict_mem_index_free(index);
index = new_index;
return DB_SUCCESS;
}
/**********************************************************************//**
Removes an index from the dictionary cache. */
TRANSACTIONAL_TARGET
static
void
dict_index_remove_from_cache_low(
/*=============================*/
dict_table_t* table, /*!< in/out: table */
dict_index_t* index, /*!< in, own: index */
ibool lru_evict) /*!< in: TRUE if index being evicted
to make room in the table LRU list */
{
ut_ad(table && index);
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(index->magic_n == DICT_INDEX_MAGIC_N);
ut_ad(dict_sys.locked());
ut_ad(table->id);
#ifdef BTR_CUR_HASH_ADAPT
ut_ad(!index->freed());
#endif /* BTR_CUR_HASH_ADAPT */
/* No need to acquire the dict_index_t::lock here because
there can't be any active operations on this index (or table). */
if (index->online_log) {
row_log_free(index->online_log);
index->online_log = NULL;
}
/* Remove the index from the list of indexes of the table */
UT_LIST_REMOVE(table->indexes, index);
/* The index is being dropped, remove any compression stats for it. */
if (!lru_evict && DICT_TF_GET_ZIP_SSIZE(index->table->flags)) {
mysql_mutex_lock(&page_zip_stat_per_index_mutex);
page_zip_stat_per_index.erase(index->id);
mysql_mutex_unlock(&page_zip_stat_per_index_mutex);
}
/* Remove the index from affected virtual column index list */
index->detach_columns();
#ifdef BTR_CUR_HASH_ADAPT
/* We always create search info whether or not adaptive
hash index is enabled or not. */
/* We are not allowed to free the in-memory index struct
dict_index_t until all entries in the adaptive hash index
that point to any of the page belonging to his b-tree index
are dropped. This is so because dropping of these entries
require access to dict_index_t struct. To avoid such scenario
We keep a count of number of such pages in the search_info and
only free the dict_index_t struct when this count drops to
zero. See also: dict_table_can_be_evicted() */
if (index->n_ahi_pages()) {
table->autoinc_mutex.wr_lock();
index->set_freed();
UT_LIST_ADD_LAST(table->freed_indexes, index);
table->autoinc_mutex.wr_unlock();
return;
}
#endif /* BTR_CUR_HASH_ADAPT */
index->lock.free();
dict_mem_index_free(index);
}
/**********************************************************************//**
Removes an index from the dictionary cache. */
void
dict_index_remove_from_cache(
/*=========================*/
dict_table_t* table, /*!< in/out: table */
dict_index_t* index) /*!< in, own: index */
{
dict_index_remove_from_cache_low(table, index, FALSE);
}
/** Tries to find column names for the index and sets the col field of the
index.
@param[in] table table
@param[in,out] index index
@param[in] add_v new virtual columns added along with an add index call
@return whether the column names were found */
static
bool
dict_index_find_cols(
dict_index_t* index,
const dict_add_v_col_t* add_v)
{
std::vector<ulint, ut_allocator<ulint> > col_added;
std::vector<ulint, ut_allocator<ulint> > v_col_added;
const dict_table_t* table = index->table;
ut_ad(table->magic_n == DICT_TABLE_MAGIC_N);
ut_ad(dict_sys.locked());
for (ulint i = 0; i < index->n_fields; i++) {
ulint j;
dict_field_t* field = dict_index_get_nth_field(index, i);
for (j = 0; j < table->n_cols; j++) {
if (!innobase_strcasecmp(dict_table_get_col_name(table, j),
field->name)) {
/* Check if same column is being assigned again
which suggest that column has duplicate name. */
bool exists =
std::find(col_added.begin(),
col_added.end(), j)
!= col_added.end();
if (exists) {
/* Duplicate column found. */
goto dup_err;
}
field->col = dict_table_get_nth_col(table, j);
col_added.push_back(j);
goto found;
}
}
/* Let's check if it is a virtual column */
for (j = 0; j < table->n_v_cols; j++) {
if (!strcmp(dict_table_get_v_col_name(table, j),
field->name)) {
/* Check if same column is being assigned again
which suggest that column has duplicate name. */
bool exists =
std::find(v_col_added.begin(),
v_col_added.end(), j)
!= v_col_added.end();
if (exists) {
/* Duplicate column found. */
break;
}
field->col = reinterpret_cast<dict_col_t*>(
dict_table_get_nth_v_col(table, j));
v_col_added.push_back(j);
goto found;
}
}
if (add_v) {
for (j = 0; j < add_v->n_v_col; j++) {
if (!strcmp(add_v->v_col_name[j],
field->name)) {
field->col = const_cast<dict_col_t*>(
&add_v->v_col[j].m_col);
goto found;
}
}
}
dup_err:
#ifdef UNIV_DEBUG
/* It is an error not to find a matching column. */
ib::error() << "No matching column for " << field->name
<< " in index " << index->name
<< " of table " << table->name;
#endif /* UNIV_DEBUG */
return(FALSE);
found:
;
}
return(TRUE);
}
/** Add a column to an index.
@param index index
@param table table
@param col column
@param prefix_len column prefix length
@param descending whether to use descending order */
void dict_index_add_col(dict_index_t *index, const dict_table_t *table,
dict_col_t *col, ulint prefix_len, bool descending)
{
dict_field_t* field;
const char* col_name;
if (col->is_virtual()) {
dict_v_col_t* v_col = reinterpret_cast<dict_v_col_t*>(col);
/* Register the index with the virtual column index list */
v_col->v_indexes.push_front(dict_v_idx_t(index, index->n_def));
col_name = dict_table_get_v_col_name_mysql(
table, dict_col_get_no(col));
} else {
col_name = dict_table_get_col_name(table, dict_col_get_no(col));
}
dict_mem_index_add_field(index, col_name, prefix_len);
field = dict_index_get_nth_field(index, unsigned(index->n_def) - 1);
field->col = col;
field->fixed_len = static_cast<uint16_t>(
dict_col_get_fixed_size(
col, dict_table_is_comp(table)))
& ((1U << 10) - 1);
if (prefix_len && field->fixed_len > prefix_len) {
field->fixed_len = static_cast<uint16_t>(prefix_len)
& ((1U << 10) - 1);
}
/* Long fixed-length fields that need external storage are treated as
variable-length fields, so that the extern flag can be embedded in
the length word. */
if (field->fixed_len > DICT_MAX_FIXED_COL_LEN) {
field->fixed_len = 0;
}
field->descending = descending;
/* The comparison limit above must be constant. If it were
changed, the disk format of some fixed-length columns would
change, which would be a disaster. */
compile_time_assert(DICT_MAX_FIXED_COL_LEN == 768);
if (!(col->prtype & DATA_NOT_NULL)) {
index->n_nullable++;
}
}
/*******************************************************************//**
Copies fields contained in index2 to index1. */
static
void
dict_index_copy(
/*============*/
dict_index_t* index1, /*!< in: index to copy to */
const dict_index_t* index2, /*!< in: index to copy from */
ulint start, /*!< in: first position to copy */
ulint end) /*!< in: last position to copy */
{
dict_field_t* field;
ulint i;
/* Copy fields contained in index2 */
for (i = start; i < end; i++) {
field = dict_index_get_nth_field(index2, i);
dict_index_add_col(index1, index2->table, field->col,
field->prefix_len, field->descending);
}
}
/*******************************************************************//**
Copies types of fields contained in index to tuple. */
void
dict_index_copy_types(
/*==================*/
dtuple_t* tuple, /*!< in/out: data tuple */
const dict_index_t* index, /*!< in: index */
ulint n_fields) /*!< in: number of
field types to copy */
{
ulint i;
if (dict_index_is_ibuf(index)) {
dtuple_set_types_binary(tuple, n_fields);
return;
}
for (i = 0; i < n_fields; i++) {
const dict_field_t* ifield;
dtype_t* dfield_type;
ifield = dict_index_get_nth_field(index, i);
dfield_type = dfield_get_type(dtuple_get_nth_field(tuple, i));
dict_col_copy_type(dict_field_get_col(ifield), dfield_type);
if (dict_index_is_spatial(index)
&& DATA_GEOMETRY_MTYPE(dfield_type->mtype)) {
dfield_type->prtype |= DATA_GIS_MBR;
}
}
}
/** Copies types of virtual columns contained in table to tuple and sets all
fields of the tuple to the SQL NULL value. This function should
be called right after dtuple_create().
@param[in,out] tuple data tuple
@param[in] table table
*/
void
dict_table_copy_v_types(
dtuple_t* tuple,
const dict_table_t* table)
{
/* tuple could have more virtual columns than existing table,
if we are calling this for creating index along with adding
virtual columns */
ulint n_fields = ut_min(dtuple_get_n_v_fields(tuple),
static_cast<ulint>(table->n_v_def));
for (ulint i = 0; i < n_fields; i++) {
dfield_t* dfield = dtuple_get_nth_v_field(tuple, i);
dtype_t* dtype = dfield_get_type(dfield);
dfield_set_null(dfield);
dict_col_copy_type(
&(dict_table_get_nth_v_col(table, i)->m_col),
dtype);
}
}
/*******************************************************************//**
Copies types of columns contained in table to tuple and sets all
fields of the tuple to the SQL NULL value. This function should
be called right after dtuple_create(). */
void
dict_table_copy_types(
/*==================*/
dtuple_t* tuple, /*!< in/out: data tuple */
const dict_table_t* table) /*!< in: table */
{
ulint i;
for (i = 0; i < dtuple_get_n_fields(tuple); i++) {
dfield_t* dfield = dtuple_get_nth_field(tuple, i);
dtype_t* dtype = dfield_get_type(dfield);
dfield_set_null(dfield);
dict_col_copy_type(dict_table_get_nth_col(table, i), dtype);
}
dict_table_copy_v_types(tuple, table);
}
/*******************************************************************//**
Builds the internal dictionary cache representation for a clustered
index, containing also system fields not defined by the user.
@return own: the internal representation of the clustered index */
static
dict_index_t*
dict_index_build_internal_clust(
/*============================*/
dict_index_t* index) /*!< in: user representation of
a clustered index */
{
dict_table_t* table = index->table;
dict_index_t* new_index;
dict_field_t* field;
ulint trx_id_pos;
ulint i;
ibool* indexed;
ut_ad(index->is_primary());
ut_ad(!index->has_virtual());
ut_ad(dict_sys.locked());
/* Create a new index object with certainly enough fields */
new_index = dict_mem_index_create(index->table, index->name,
index->type,
unsigned(index->n_fields
+ table->n_cols));
/* Copy other relevant data from the old index struct to the new
struct: it inherits the values */
new_index->n_user_defined_cols = index->n_fields;
new_index->id = index->id;
/* Copy the fields of index */
dict_index_copy(new_index, index, 0, index->n_fields);
if (dict_index_is_unique(index)) {
/* Only the fields defined so far are needed to identify
the index entry uniquely */
new_index->n_uniq = new_index->n_def;
} else {
/* Also the row id is needed to identify the entry */
new_index->n_uniq = unsigned(new_index->n_def + 1)
& dict_index_t::MAX_N_FIELDS;
}
new_index->trx_id_offset = 0;
/* Add system columns, trx id first */
trx_id_pos = new_index->n_def;
compile_time_assert(DATA_ROW_ID == 0);
compile_time_assert(DATA_TRX_ID == 1);
compile_time_assert(DATA_ROLL_PTR == 2);
if (!dict_index_is_unique(index)) {
dict_index_add_col(new_index, table,
dict_table_get_sys_col(
table, DATA_ROW_ID),
0);
trx_id_pos++;
}
dict_index_add_col(
new_index, table,
dict_table_get_sys_col(table, DATA_TRX_ID), 0);
for (i = 0; i < trx_id_pos; i++) {
ulint fixed_size = dict_col_get_fixed_size(
dict_index_get_nth_col(new_index, i),
dict_table_is_comp(table));
if (fixed_size == 0) {
new_index->trx_id_offset = 0;
break;
}
dict_field_t* field = dict_index_get_nth_field(
new_index, i);
if (field->prefix_len > 0) {
new_index->trx_id_offset = 0;
break;
}
/* Add fixed_size to new_index->trx_id_offset.
Because the latter is a bit-field, an overflow
can theoretically occur. Check for it. */
fixed_size += new_index->trx_id_offset;
new_index->trx_id_offset = static_cast<unsigned>(fixed_size)
& ((1U << 12) - 1);
if (new_index->trx_id_offset != fixed_size) {
/* Overflow. Pretend that this is a
variable-length PRIMARY KEY. */
ut_ad(0);
new_index->trx_id_offset = 0;
break;
}
}
dict_index_add_col(
new_index, table,
dict_table_get_sys_col(table, DATA_ROLL_PTR), 0);
/* Remember the table columns already contained in new_index */
indexed = static_cast<ibool*>(
ut_zalloc_nokey(table->n_cols * sizeof *indexed));
/* Mark the table columns already contained in new_index */
for (i = 0; i < new_index->n_def; i++) {
field = dict_index_get_nth_field(new_index, i);
/* If there is only a prefix of the column in the index
field, do not mark the column as contained in the index */
if (field->prefix_len == 0) {
indexed[field->col->ind] = TRUE;
}
}
/* Add to new_index non-system columns of table not yet included
there */
for (i = 0; i + DATA_N_SYS_COLS < ulint(table->n_cols); i++) {
dict_col_t* col = dict_table_get_nth_col(table, i);
ut_ad(col->mtype != DATA_SYS);
if (!indexed[col->ind]) {
dict_index_add_col(new_index, table, col, 0);
}
}
ut_free(indexed);
ut_ad(UT_LIST_GET_LEN(table->indexes) == 0);
new_index->n_core_null_bytes = table->supports_instant()
? dict_index_t::NO_CORE_NULL_BYTES
: static_cast<uint8_t>(
UT_BITS_IN_BYTES(unsigned(new_index->n_nullable)));
new_index->cached = TRUE;
return(new_index);
}
/*******************************************************************//**
Builds the internal dictionary cache representation for a non-clustered
index, containing also system fields not defined by the user.
@return own: the internal representation of the non-clustered index */
static
dict_index_t*
dict_index_build_internal_non_clust(
/*================================*/
dict_index_t* index) /*!< in: user representation of
a non-clustered index */
{
dict_field_t* field;
dict_index_t* new_index;
dict_index_t* clust_index;
dict_table_t* table = index->table;
ulint i;
ibool* indexed;
ut_ad(table && index);
ut_ad(!dict_index_is_clust(index));
ut_ad(!dict_index_is_ibuf(index));
ut_ad(dict_sys.locked());
/* The clustered index should be the first in the list of indexes */
clust_index = UT_LIST_GET_FIRST(table->indexes);
ut_ad(clust_index);
ut_ad(dict_index_is_clust(clust_index));
ut_ad(!dict_index_is_ibuf(clust_index));
/* Create a new index */
new_index = dict_mem_index_create(
index->table, index->name, index->type,
ulint(index->n_fields + 1 + clust_index->n_uniq));
/* Copy other relevant data from the old index
struct to the new struct: it inherits the values */
new_index->n_user_defined_cols = index->n_fields;
new_index->id = index->id;
/* Copy fields from index to new_index */
dict_index_copy(new_index, index, 0, index->n_fields);
/* Remember the table columns already contained in new_index */
indexed = static_cast<ibool*>(
ut_zalloc_nokey(table->n_cols * sizeof *indexed));
/* Mark the table columns already contained in new_index */
for (i = 0; i < new_index->n_def; i++) {
field = dict_index_get_nth_field(new_index, i);
if (field->col->is_virtual()) {
continue;
}
/* If there is only a prefix of the column in the index
field, do not mark the column as contained in the index */
if (field->prefix_len == 0) {
indexed[field->col->ind] = TRUE;
}
}
/* Add to new_index the columns necessary to determine the clustered
index entry uniquely */
for (i = 0; i < clust_index->n_uniq; i++) {
field = dict_index_get_nth_field(clust_index, i);
if (!indexed[field->col->ind] || index->is_spatial()) {
dict_index_add_col(new_index, table, field->col,
field->prefix_len,
field->descending);
}
}
ut_free(indexed);
if (dict_index_is_unique(index)) {
new_index->n_uniq = index->n_fields;
} else {
new_index->n_uniq = new_index->n_def;
}
/* Set the n_fields value in new_index to the actual defined
number of fields */
new_index->n_fields = new_index->n_def;
new_index->cached = TRUE;
return(new_index);
}
/***********************************************************************
Builds the internal dictionary cache representation for an FTS index.
@return own: the internal representation of the FTS index */
static
dict_index_t*
dict_index_build_internal_fts(
/*==========================*/
dict_index_t* index) /*!< in: user representation of an FTS index */
{
dict_index_t* new_index;
ut_ad(index->type & DICT_FTS);
ut_ad(dict_sys.locked());
/* Create a new index */
new_index = dict_mem_index_create(index->table, index->name,
index->type, index->n_fields);
/* Copy other relevant data from the old index struct to the new
struct: it inherits the values */
new_index->n_user_defined_cols = index->n_fields;
new_index->id = index->id;
/* Copy fields from index to new_index */
dict_index_copy(new_index, index, 0, index->n_fields);
new_index->n_uniq = 0;
new_index->cached = TRUE;
dict_table_t* table = index->table;
if (table->fts->cache == NULL) {
table->fts->cache = fts_cache_create(table);
}
mysql_mutex_lock(&table->fts->cache->init_lock);
/* Notify the FTS cache about this index. */
fts_cache_index_cache_create(table, new_index);
mysql_mutex_unlock(&table->fts->cache->init_lock);
return(new_index);
}
/*====================== FOREIGN KEY PROCESSING ========================*/
/**********************************************************************//**
Removes a foreign constraint struct from the dictionary cache. */
void
dict_foreign_remove_from_cache(
/*===========================*/
dict_foreign_t* foreign) /*!< in, own: foreign constraint */
{
ut_ad(dict_sys.locked());
ut_a(foreign);
if (foreign->referenced_table != NULL) {
foreign->referenced_table->referenced_set.erase(foreign);
}
if (foreign->foreign_table != NULL) {
foreign->foreign_table->foreign_set.erase(foreign);
}
dict_foreign_free(foreign);
}
/**********************************************************************//**
Looks for the foreign constraint from the foreign and referenced lists
of a table.
@return foreign constraint */
static
dict_foreign_t*
dict_foreign_find(
/*==============*/
dict_table_t* table, /*!< in: table object */
dict_foreign_t* foreign) /*!< in: foreign constraint */
{
ut_ad(dict_sys.frozen());
ut_ad(dict_foreign_set_validate(table->foreign_set));
ut_ad(dict_foreign_set_validate(table->referenced_set));
dict_foreign_set::iterator it = table->foreign_set.find(foreign);
if (it != table->foreign_set.end()) {
return(*it);
}
it = table->referenced_set.find(foreign);
if (it != table->referenced_set.end()) {
return(*it);
}
return(NULL);
}
/*********************************************************************//**
Tries to find an index whose first fields are the columns in the array,
in the same order and is not marked for deletion and is not the same
as types_idx.
@return matching index, NULL if not found */
dict_index_t*
dict_foreign_find_index(
/*====================*/
const dict_table_t* table, /*!< in: table */
const char** col_names,
/*!< in: column names, or NULL
to use table->col_names */
const char** columns,/*!< in: array of column names */
ulint n_cols, /*!< in: number of columns */
const dict_index_t* types_idx,
/*!< in: NULL or an index
whose types the column types
must match */
bool check_charsets,
/*!< in: whether to check
charsets. only has an effect
if types_idx != NULL */
ulint check_null,
/*!< in: nonzero if none of
the columns must be declared
NOT NULL */
fkerr_t* error, /*!< out: error code */
ulint* err_col_no,
/*!< out: column number where
error happened */
dict_index_t** err_index)
/*!< out: index where error
happened */
{
ut_ad(dict_sys.frozen());
if (error) {
*error = FK_INDEX_NOT_FOUND;
}
for (dict_index_t* index = dict_table_get_first_index(table);
index;
index = dict_table_get_next_index(index)) {
if (!index->to_be_dropped
&& !dict_index_is_online_ddl(index)
&& dict_foreign_qualify_index(
table, col_names, columns, n_cols,
index, types_idx,
check_charsets, check_null,
error, err_col_no, err_index)) {
if (error) {
*error = FK_SUCCESS;
}
return(index);
}
}
return(NULL);
}
/**********************************************************************//**
Report an error in a foreign key definition. */
static
void
dict_foreign_error_report_low(
/*==========================*/
FILE* file, /*!< in: output stream */
const char* name) /*!< in: table name */
{
rewind(file);
ut_print_timestamp(file);
fprintf(file, " Error in foreign key constraint of table %s:\n",
name);
}
/**********************************************************************//**
Report an error in a foreign key definition. */
static
void
dict_foreign_error_report(
/*======================*/
FILE* file, /*!< in: output stream */
dict_foreign_t* fk, /*!< in: foreign key constraint */
const char* msg) /*!< in: the error message */
{
std::string fk_str;
mysql_mutex_lock(&dict_foreign_err_mutex);
dict_foreign_error_report_low(file, fk->foreign_table_name);
fputs(msg, file);
fputs(" Constraint:\n", file);
fk_str = dict_print_info_on_foreign_key_in_create_format(NULL, fk, TRUE);
fputs(fk_str.c_str(), file);
putc('\n', file);
if (fk->foreign_index) {
fprintf(file, "The index in the foreign key in table is"
" %s\n%s\n", fk->foreign_index->name(),
FOREIGN_KEY_CONSTRAINTS_MSG);
}
mysql_mutex_unlock(&dict_foreign_err_mutex);
}
/**********************************************************************//**
Adds a foreign key constraint object to the dictionary cache. May free
the object if there already is an object with the same identifier in.
At least one of the foreign table and the referenced table must already
be in the dictionary cache!
@return DB_SUCCESS or error code */
dberr_t
dict_foreign_add_to_cache(
/*======================*/
dict_foreign_t* foreign,
/*!< in, own: foreign key constraint */
const char** col_names,
/*!< in: column names, or NULL to use
foreign->foreign_table->col_names */
bool check_charsets,
/*!< in: whether to check charset
compatibility */
dict_err_ignore_t ignore_err)
/*!< in: error to be ignored */
{
dict_table_t* for_table;
dict_table_t* ref_table;
dict_foreign_t* for_in_cache = NULL;
dict_index_t* index;
ibool added_to_referenced_list= FALSE;
FILE* ef = dict_foreign_err_file;
DBUG_ENTER("dict_foreign_add_to_cache");
DBUG_PRINT("dict_foreign_add_to_cache", ("id: %s", foreign->id));
ut_ad(dict_sys.locked());
for_table = dict_sys.find_table(
{foreign->foreign_table_name_lookup,
strlen(foreign->foreign_table_name_lookup)});
ref_table = dict_sys.find_table(
{foreign->referenced_table_name_lookup,
strlen(foreign->referenced_table_name_lookup)});
ut_a(for_table || ref_table);
if (for_table) {
for_in_cache = dict_foreign_find(for_table, foreign);
}
if (!for_in_cache && ref_table) {
for_in_cache = dict_foreign_find(ref_table, foreign);
}
if (for_in_cache) {
dict_foreign_free(foreign);
} else {
for_in_cache = foreign;
}
if (ref_table && !for_in_cache->referenced_table) {
index = dict_foreign_find_index(
ref_table, NULL,
for_in_cache->referenced_col_names,
for_in_cache->n_fields, for_in_cache->foreign_index,
check_charsets, false);
if (index == NULL
&& !(ignore_err & DICT_ERR_IGNORE_FK_NOKEY)) {
dict_foreign_error_report(
ef, for_in_cache,
"there is no index in referenced table"
" which would contain\n"
"the columns as the first columns,"
" or the data types in the\n"
"referenced table do not match"
" the ones in table.");
if (for_in_cache == foreign) {
dict_foreign_free(foreign);
}
DBUG_RETURN(DB_CANNOT_ADD_CONSTRAINT);
}
for_in_cache->referenced_table = ref_table;
for_in_cache->referenced_index = index;
std::pair<dict_foreign_set::iterator, bool> ret
= ref_table->referenced_set.insert(for_in_cache);
ut_a(ret.second); /* second is true if the insertion
took place */
added_to_referenced_list = TRUE;
}
if (for_table && !for_in_cache->foreign_table) {
index = dict_foreign_find_index(
for_table, col_names,
for_in_cache->foreign_col_names,
for_in_cache->n_fields,
for_in_cache->referenced_index, check_charsets,
for_in_cache->type
& (foreign->DELETE_SET_NULL
| foreign->UPDATE_SET_NULL));
if (index == NULL
&& !(ignore_err & DICT_ERR_IGNORE_FK_NOKEY)) {
dict_foreign_error_report(
ef, for_in_cache,
"there is no index in the table"
" which would contain\n"
"the columns as the first columns,"
" or the data types in the\n"
"table do not match"
" the ones in the referenced table\n"
"or one of the ON ... SET NULL columns"
" is declared NOT NULL.");
if (for_in_cache == foreign) {
if (added_to_referenced_list) {
const dict_foreign_set::size_type
n = ref_table->referenced_set
.erase(for_in_cache);
ut_a(n == 1); /* the number of
elements removed must
be one */
}
dict_foreign_free(foreign);
}
DBUG_RETURN(DB_CANNOT_ADD_CONSTRAINT);
}
for_in_cache->foreign_table = for_table;
for_in_cache->foreign_index = index;
std::pair<dict_foreign_set::iterator, bool> ret
= for_table->foreign_set.insert(for_in_cache);
ut_a(ret.second); /* second is true if the insertion
took place */
}
/* We need to move the table to the non-LRU end of the table LRU
list. Otherwise it will be evicted from the cache. */
if (ref_table != NULL) {
dict_sys.prevent_eviction(ref_table);
}
if (for_table != NULL) {
dict_sys.prevent_eviction(for_table);
}
ut_ad(dict_lru_validate());
DBUG_RETURN(DB_SUCCESS);
}
/*********************************************************************//**
Scans from pointer onwards. Stops if is at the start of a copy of
'string' where characters are compared without case sensitivity, and
only outside `` or "" quotes. Stops also at NUL.
@return scanned up to this */
static
const char*
dict_scan_to(
/*=========*/
const char* ptr, /*!< in: scan from */
const char* string) /*!< in: look for this */
{
char quote = '\0';
bool escape = false;
for (; *ptr; ptr++) {
if (*ptr == quote) {
/* Closing quote character: do not look for
starting quote or the keyword. */
/* If the quote character is escaped by a
backslash, ignore it. */
if (escape) {
escape = false;
} else {
quote = '\0';
}
} else if (quote) {
/* Within quotes: do nothing. */
if (escape) {
escape = false;
} else if (*ptr == '\\') {
escape = true;
}
} else if (*ptr == '`' || *ptr == '"' || *ptr == '\'') {
/* Starting quote: remember the quote character. */
quote = *ptr;
} else {
/* Outside quotes: look for the keyword. */
ulint i;
for (i = 0; string[i]; i++) {
if (toupper((int)(unsigned char)(ptr[i]))
!= toupper((int)(unsigned char)
(string[i]))) {
goto nomatch;
}
}
break;
nomatch:
;
}
}
return(ptr);
}
/*********************************************************************//**
Accepts a specified string. Comparisons are case-insensitive.
@return if string was accepted, the pointer is moved after that, else
ptr is returned */
static
const char*
dict_accept(
/*========*/
CHARSET_INFO* cs, /*!< in: the character set of ptr */
const char* ptr, /*!< in: scan from this */
const char* string, /*!< in: accept only this string as the next
non-whitespace string */
ibool* success)/*!< out: TRUE if accepted */
{
const char* old_ptr = ptr;
const char* old_ptr2;
*success = FALSE;
while (my_isspace(cs, *ptr)) {
ptr++;
}
old_ptr2 = ptr;
ptr = dict_scan_to(ptr, string);
if (*ptr == '\0' || old_ptr2 != ptr) {
return(old_ptr);
}
*success = TRUE;
return ptr + strlen(string);
}
/*********************************************************************//**
Scans an id. For the lexical definition of an 'id', see the code below.
Strips backquotes or double quotes from around the id.
@return scanned to */
static
const char*
dict_scan_id(
/*=========*/
CHARSET_INFO* cs, /*!< in: the character set of ptr */
const char* ptr, /*!< in: scanned to */
mem_heap_t* heap, /*!< in: heap where to allocate the id
(NULL=id will not be allocated, but it
will point to string near ptr) */
const char** id, /*!< out,own: the id; NULL if no id was
scannable */
ibool table_id,/*!< in: TRUE=convert the allocated id
as a table name; FALSE=convert to UTF-8 */
ibool accept_also_dot)
/*!< in: TRUE if also a dot can appear in a
non-quoted id; in a quoted id it can appear
always */
{
char quote = '\0';
ulint len = 0;
const char* s;
char* str;
char* dst;
*id = NULL;
while (my_isspace(cs, *ptr)) {
ptr++;
}
if (*ptr == '\0') {
return(ptr);
}
if (*ptr == '`' || *ptr == '"') {
quote = *ptr++;
}
s = ptr;
if (quote) {
for (;;) {
if (!*ptr) {
/* Syntax error */
return(ptr);
}
if (*ptr == quote) {
ptr++;
if (*ptr != quote) {
break;
}
}
ptr++;
len++;
}
} else {
while (!my_isspace(cs, *ptr) && *ptr != '(' && *ptr != ')'
&& (accept_also_dot || *ptr != '.')
&& *ptr != ',' && *ptr != '\0') {
ptr++;
}
len = ulint(ptr - s);
}
if (heap == NULL) {
/* no heap given: id will point to source string */
*id = s;
return(ptr);
}
if (quote) {
char* d;
str = d = static_cast<char*>(
mem_heap_alloc(heap, len + 1));
while (len--) {
if ((*d++ = *s++) == quote) {
s++;
}
}
*d++ = 0;
len = ulint(d - str);
ut_ad(*s == quote);
ut_ad(s + 1 == ptr);
} else {
str = mem_heap_strdupl(heap, s, len);
}
if (!table_id) {
convert_id:
/* Convert the identifier from connection character set
to UTF-8. */
len = 3 * len + 1;
*id = dst = static_cast<char*>(mem_heap_alloc(heap, len));
innobase_convert_from_id(cs, dst, str, len);
} else if (!strncmp(str, srv_mysql50_table_name_prefix,
sizeof(srv_mysql50_table_name_prefix) - 1)) {
/* This is a pre-5.1 table name
containing chars other than [A-Za-z0-9].
Discard the prefix and use raw UTF-8 encoding. */
str += sizeof(srv_mysql50_table_name_prefix) - 1;
len -= sizeof(srv_mysql50_table_name_prefix) - 1;
goto convert_id;
} else {
/* Encode using filename-safe characters. */
len = 5 * len + 1;
*id = dst = static_cast<char*>(mem_heap_alloc(heap, len));
innobase_convert_from_table_id(cs, dst, str, len);
}
return(ptr);
}
/*********************************************************************//**
Open a table from its database and table name, this is currently used by
foreign constraint parser to get the referenced table.
@return complete table name with database and table name, allocated from
heap memory passed in */
char*
dict_get_referenced_table(
const char* name, /*!< in: foreign key table name */
const char* database_name, /*!< in: table db name */
ulint database_name_len, /*!< in: db name length */
const char* table_name, /*!< in: table name */
ulint table_name_len, /*!< in: table name length */
dict_table_t** table, /*!< out: table object or NULL */
mem_heap_t* heap, /*!< in/out: heap memory */
CHARSET_INFO* from_cs) /*!< in: table name charset */
{
char* ref;
char db_name[MAX_DATABASE_NAME_LEN];
char tbl_name[MAX_TABLE_NAME_LEN];
CHARSET_INFO* to_cs = &my_charset_filename;
uint errors;
ut_ad(database_name || name);
ut_ad(table_name);
if (!strncmp(table_name, srv_mysql50_table_name_prefix,
sizeof(srv_mysql50_table_name_prefix) - 1)) {
/* This is a pre-5.1 table name
containing chars other than [A-Za-z0-9].
Discard the prefix and use raw UTF-8 encoding. */
table_name += sizeof(srv_mysql50_table_name_prefix) - 1;
table_name_len -= sizeof(srv_mysql50_table_name_prefix) - 1;
to_cs = system_charset_info;
}
table_name_len = strconvert(from_cs, table_name, table_name_len, to_cs,
tbl_name, MAX_TABLE_NAME_LEN, &errors);
table_name = tbl_name;
if (database_name) {
to_cs = &my_charset_filename;
if (!strncmp(database_name, srv_mysql50_table_name_prefix,
sizeof(srv_mysql50_table_name_prefix) - 1)) {
database_name
+= sizeof(srv_mysql50_table_name_prefix) - 1;
database_name_len
-= sizeof(srv_mysql50_table_name_prefix) - 1;
to_cs = system_charset_info;
}
database_name_len = strconvert(
from_cs, database_name, database_name_len, to_cs,
db_name, MAX_DATABASE_NAME_LEN, &errors);
database_name = db_name;
} else {
/* Use the database name of the foreign key table */
database_name = name;
database_name_len = dict_get_db_name_len(name);
}
/* Copy database_name, '/', table_name, '\0' */
const size_t len = database_name_len + table_name_len + 1;
ref = static_cast<char*>(mem_heap_alloc(heap, len + 1));
memcpy(ref, database_name, database_name_len);
ref[database_name_len] = '/';
memcpy(ref + database_name_len + 1, table_name, table_name_len + 1);
/* Values; 0 = Store and compare as given; case sensitive
1 = Store and compare in lower; case insensitive
2 = Store as given, compare in lower; case semi-sensitive */
if (lower_case_table_names == 2) {
innobase_casedn_str(ref);
*table = dict_sys.load_table({ref, len});
memcpy(ref, database_name, database_name_len);
ref[database_name_len] = '/';
memcpy(ref + database_name_len + 1, table_name, table_name_len + 1);
} else {
#ifndef _WIN32
if (lower_case_table_names == 1) {
innobase_casedn_str(ref);
}
#else
innobase_casedn_str(ref);
#endif /* !_WIN32 */
*table = dict_sys.load_table({ref, len});
}
return(ref);
}
/*********************************************************************//**
Removes MySQL comments from an SQL string. A comment is either
(a) '#' to the end of the line,
(b) '--[space]' to the end of the line, or
(c) '[slash][asterisk]' till the next '[asterisk][slash]' (like the familiar
C comment syntax).
@return own: SQL string stripped from comments; the caller must free
this with ut_free()! */
static
char*
dict_strip_comments(
/*================*/
const char* sql_string, /*!< in: SQL string */
size_t sql_length) /*!< in: length of sql_string */
{
char* str;
const char* sptr;
const char* eptr = sql_string + sql_length;
char* ptr;
/* unclosed quote character (0 if none) */
char quote = 0;
bool escape = false;
DBUG_ENTER("dict_strip_comments");
DBUG_PRINT("dict_strip_comments", ("%s", sql_string));
str = static_cast<char*>(ut_malloc_nokey(sql_length + 1));
sptr = sql_string;
ptr = str;
for (;;) {
scan_more:
if (sptr >= eptr || *sptr == '\0') {
end_of_string:
*ptr = '\0';
ut_a(ptr <= str + sql_length);
DBUG_PRINT("dict_strip_comments", ("%s", str));
DBUG_RETURN(str);
}
if (*sptr == quote) {
/* Closing quote character: do not look for
starting quote or comments. */
/* If the quote character is escaped by a
backslash, ignore it. */
if (escape) {
escape = false;
} else {
quote = 0;
}
} else if (quote) {
/* Within quotes: do not look for
starting quotes or comments. */
if (escape) {
escape = false;
} else if (*sptr == '\\') {
escape = true;
}
} else if (*sptr == '"' || *sptr == '`' || *sptr == '\'') {
/* Starting quote: remember the quote character. */
quote = *sptr;
} else if (*sptr == '#'
|| (sptr[0] == '-' && sptr[1] == '-'
&& sptr[2] == ' ')) {
for (;;) {
if (++sptr >= eptr) {
goto end_of_string;
}
/* In Unix a newline is 0x0A while in Windows
it is 0x0D followed by 0x0A */
switch (*sptr) {
case (char) 0X0A:
case (char) 0x0D:
case '\0':
goto scan_more;
}
}
} else if (!quote && *sptr == '/' && *(sptr + 1) == '*') {
sptr += 2;
for (;;) {
if (sptr >= eptr) {
goto end_of_string;
}
switch (*sptr) {
case '\0':
goto scan_more;
case '*':
if (sptr[1] == '/') {
sptr += 2;
goto scan_more;
}
}
sptr++;
}
}
*ptr = *sptr;
ptr++;
sptr++;
}
}
/*********************************************************************//**
Finds the highest [number] for foreign key constraints of the table. Looks
only at the >= 4.0.18-format id's, which are of the form
databasename/tablename_ibfk_[number].
@return highest number, 0 if table has no new format foreign key constraints */
ulint
dict_table_get_highest_foreign_id(
/*==============================*/
dict_table_t* table) /*!< in: table in the dictionary memory cache */
{
dict_foreign_t* foreign;
char* endp;
ulint biggest_id = 0;
ulint id;
ulint len;
DBUG_ENTER("dict_table_get_highest_foreign_id");
ut_a(table);
len = strlen(table->name.m_name);
for (dict_foreign_set::iterator it = table->foreign_set.begin();
it != table->foreign_set.end();
++it) {
char fkid[MAX_TABLE_NAME_LEN * 2 + 20];
foreign = *it;
strncpy(fkid, foreign->id, (sizeof fkid) - 1);
fkid[(sizeof fkid) - 1] = '\0';
/* Convert foreign key identifier on dictionary memory
cache to filename charset. */
innobase_convert_to_filename_charset(
strchr(fkid, '/') + 1,
strchr(foreign->id, '/') + 1,
MAX_TABLE_NAME_LEN);
if (strlen(fkid) > ((sizeof dict_ibfk) - 1) + len
&& 0 == memcmp(fkid, table->name.m_name, len)
&& 0 == memcmp(fkid + len,
dict_ibfk, (sizeof dict_ibfk) - 1)
&& fkid[len + ((sizeof dict_ibfk) - 1)] != '0') {
/* It is of the >= 4.0.18 format */
id = strtoul(fkid + len
+ ((sizeof dict_ibfk) - 1),
&endp, 10);
if (*endp == '\0') {
ut_a(id != biggest_id);
if (id > biggest_id) {
biggest_id = id;
}
}
}
}
DBUG_PRINT("dict_table_get_highest_foreign_id",
("id: " ULINTPF, biggest_id));
DBUG_RETURN(biggest_id);
}
/**********************************************************************//**
Parses the CONSTRAINT id's to be dropped in an ALTER TABLE statement.
@return DB_SUCCESS or DB_CANNOT_DROP_CONSTRAINT if syntax error or the
constraint id does not match */
dberr_t
dict_foreign_parse_drop_constraints(
/*================================*/
mem_heap_t* heap, /*!< in: heap from which we can
allocate memory */
trx_t* trx, /*!< in: transaction */
dict_table_t* table, /*!< in: table */
ulint* n, /*!< out: number of constraints
to drop */
const char*** constraints_to_drop) /*!< out: id's of the
constraints to drop */
{
ibool success;
char* str;
size_t len;
const char* ptr;
const char* ptr1;
const char* id;
CHARSET_INFO* cs;
bool if_exists = false;
ut_a(trx->mysql_thd);
cs = thd_charset(trx->mysql_thd);
*n = 0;
*constraints_to_drop = static_cast<const char**>(
mem_heap_alloc(heap, 1000 * sizeof(char*)));
ptr = innobase_get_stmt_unsafe(trx->mysql_thd, &len);
str = dict_strip_comments(ptr, len);
ptr = str;
ut_ad(dict_sys.locked());
loop:
ptr = dict_scan_to(ptr, "DROP");
if (*ptr == '\0') {
ut_free(str);
return(DB_SUCCESS);
}
ptr = dict_accept(cs, ptr, "DROP", &success);
if (!my_isspace(cs, *ptr)) {
goto loop;
}
ptr = dict_accept(cs, ptr, "FOREIGN", &success);
if (!success || !my_isspace(cs, *ptr)) {
goto loop;
}
ptr = dict_accept(cs, ptr, "KEY", &success);
if (!success) {
goto syntax_error;
}
ptr1 = dict_accept(cs, ptr, "IF", &success);
if (success && my_isspace(cs, *ptr1)) {
ptr1 = dict_accept(cs, ptr1, "EXISTS", &success);
if (success) {
ptr = ptr1;
if_exists = true;
}
}
ptr = dict_scan_id(cs, ptr, heap, &id, FALSE, TRUE);
if (id == NULL) {
goto syntax_error;
}
if (std::find_if(table->foreign_set.begin(),
table->foreign_set.end(),
dict_foreign_matches_id(id))
== table->foreign_set.end()) {
if (if_exists) {
goto loop;
}
if (!srv_read_only_mode) {
FILE* ef = dict_foreign_err_file;
mysql_mutex_lock(&dict_foreign_err_mutex);
rewind(ef);
ut_print_timestamp(ef);
fputs(" Error in dropping of a foreign key"
" constraint of table ", ef);
ut_print_name(ef, NULL, table->name.m_name);
fprintf(ef, ",\nin SQL command\n%s"
"\nCannot find a constraint with the"
" given id %s.\n", str, id);
mysql_mutex_unlock(&dict_foreign_err_mutex);
}
ut_free(str);
return(DB_CANNOT_DROP_CONSTRAINT);
}
ut_a(*n < 1000);
(*constraints_to_drop)[*n] = id;
(*n)++;
goto loop;
syntax_error:
if (!srv_read_only_mode) {
FILE* ef = dict_foreign_err_file;
mysql_mutex_lock(&dict_foreign_err_mutex);
rewind(ef);
ut_print_timestamp(ef);
fputs(" Syntax error in dropping of a"
" foreign key constraint of table ", ef);
ut_print_name(ef, NULL, table->name.m_name);
fprintf(ef, ",\n"
"close to:\n%s\n in SQL command\n%s\n", ptr, str);
mysql_mutex_unlock(&dict_foreign_err_mutex);
}
ut_free(str);
return(DB_CANNOT_DROP_CONSTRAINT);
}
/*==================== END OF FOREIGN KEY PROCESSING ====================*/
/**********************************************************************//**
Returns an index object if it is found in the dictionary cache.
Assumes that dict_sys.latch is already being held.
@return index, NULL if not found */
dict_index_t*
dict_index_get_if_in_cache_low(
/*===========================*/
index_id_t index_id) /*!< in: index id */
{
ut_ad(dict_sys.frozen());
for (dict_table_t *table= UT_LIST_GET_FIRST(dict_sys.table_LRU);
table; table= UT_LIST_GET_NEXT(table_LRU, table))
if (dict_index_t *index= dict_table_find_index_on_id(table, index_id))
return index;
for (dict_table_t *table = UT_LIST_GET_FIRST(dict_sys.table_non_LRU);
table; table= UT_LIST_GET_NEXT(table_LRU, table))
if (dict_index_t *index= dict_table_find_index_on_id(table, index_id))
return index;
return nullptr;
}
#ifdef UNIV_DEBUG
/**********************************************************************//**
Returns an index object if it is found in the dictionary cache.
@return index, NULL if not found */
dict_index_t*
dict_index_get_if_in_cache(
/*=======================*/
index_id_t index_id) /*!< in: index id */
{
dict_index_t* index;
if (!dict_sys.is_initialised()) {
return(NULL);
}
dict_sys.freeze(SRW_LOCK_CALL);
index = dict_index_get_if_in_cache_low(index_id);
dict_sys.unfreeze();
return(index);
}
/**********************************************************************//**
Checks that a tuple has n_fields_cmp value in a sensible range, so that
no comparison can occur with the page number field in a node pointer.
@return TRUE if ok */
ibool
dict_index_check_search_tuple(
/*==========================*/
const dict_index_t* index, /*!< in: index tree */
const dtuple_t* tuple) /*!< in: tuple used in a search */
{
ut_ad(dtuple_get_n_fields_cmp(tuple)
<= dict_index_get_n_unique_in_tree(index));
return(TRUE);
}
#endif /* UNIV_DEBUG */
/**********************************************************************//**
Builds a node pointer out of a physical record and a page number.
@return own: node pointer */
dtuple_t*
dict_index_build_node_ptr(
/*======================*/
const dict_index_t* index, /*!< in: index */
const rec_t* rec, /*!< in: record for which to build node
pointer */
ulint page_no,/*!< in: page number to put in node
pointer */
mem_heap_t* heap, /*!< in: memory heap where pointer
created */
ulint level) /*!< in: level of rec in tree:
0 means leaf level */
{
dtuple_t* tuple;
dfield_t* field;
byte* buf;
ulint n_unique;
if (dict_index_is_ibuf(index)) {
/* In a universal index tree, we take the whole record as
the node pointer if the record is on the leaf level,
on non-leaf levels we remove the last field, which
contains the page number of the child page */
ut_a(!dict_table_is_comp(index->table));
n_unique = rec_get_n_fields_old(rec);
if (level > 0) {
ut_a(n_unique > 1);
n_unique--;
}
} else {
n_unique = dict_index_get_n_unique_in_tree_nonleaf(index);
}
tuple = dtuple_create(heap, n_unique + 1);
/* When searching in the tree for the node pointer, we must not do
comparison on the last field, the page number field, as on upper
levels in the tree there may be identical node pointers with a
different page number; therefore, we set the n_fields_cmp to one
less: */
dtuple_set_n_fields_cmp(tuple, n_unique);
dict_index_copy_types(tuple, index, n_unique);
buf = static_cast<byte*>(mem_heap_alloc(heap, 4));
mach_write_to_4(buf, page_no);
field = dtuple_get_nth_field(tuple, n_unique);
dfield_set_data(field, buf, 4);
dtype_set(dfield_get_type(field), DATA_SYS_CHILD, DATA_NOT_NULL, 4);
rec_copy_prefix_to_dtuple(tuple, rec, index,
level ? 0 : index->n_core_fields,
n_unique, heap);
dtuple_set_info_bits(tuple, dtuple_get_info_bits(tuple)
| REC_STATUS_NODE_PTR);
ut_ad(dtuple_check_typed(tuple));
return(tuple);
}
/** Convert a physical record into a search tuple.
@param[in] rec index record (not necessarily in an index page)
@param[in] index index
@param[in] leaf whether rec is in a leaf page
@param[in] n_fields number of data fields
@param[in,out] heap memory heap for allocation
@return own: data tuple */
dtuple_t*
dict_index_build_data_tuple(
const rec_t* rec,
const dict_index_t* index,
bool leaf,
ulint n_fields,
mem_heap_t* heap)
{
ut_ad(!index->is_clust());
dtuple_t* tuple = dtuple_create(heap, n_fields);
dict_index_copy_types(tuple, index, n_fields);
rec_copy_prefix_to_dtuple(tuple, rec, index,
leaf ? n_fields : 0, n_fields, heap);
ut_ad(dtuple_check_typed(tuple));
return(tuple);
}
/*********************************************************************//**
Calculates the minimum record length in an index. */
ulint
dict_index_calc_min_rec_len(
/*========================*/
const dict_index_t* index) /*!< in: index */
{
ulint sum = 0;
ulint i;
ulint comp = dict_table_is_comp(index->table);
if (comp) {
ulint nullable = 0;
sum = REC_N_NEW_EXTRA_BYTES;
for (i = 0; i < dict_index_get_n_fields(index); i++) {
const dict_col_t* col
= dict_index_get_nth_col(index, i);
ulint size = dict_col_get_fixed_size(col, comp);
sum += size;
if (!size) {
size = col->len;
sum += size < 128 ? 1 : 2;
}
if (!(col->prtype & DATA_NOT_NULL)) {
nullable++;
}
}
/* round the NULL flags up to full bytes */
sum += UT_BITS_IN_BYTES(nullable);
return(sum);
}
for (i = 0; i < dict_index_get_n_fields(index); i++) {
sum += dict_col_get_fixed_size(
dict_index_get_nth_col(index, i), comp);
}
if (sum > 127) {
sum += 2 * dict_index_get_n_fields(index);
} else {
sum += dict_index_get_n_fields(index);
}
sum += REC_N_OLD_EXTRA_BYTES;
return(sum);
}
std::string
dict_print_info_on_foreign_key_in_create_format(const trx_t *trx,
const dict_foreign_t *foreign,
bool add_newline)
{
const char* stripped_id;
ulint i;
std::string str;
if (strchr(foreign->id, '/')) {
/* Strip the preceding database name from the constraint id */
stripped_id = foreign->id + 1
+ dict_get_db_name_len(foreign->id);
} else {
stripped_id = foreign->id;
}
str.append(",");
if (add_newline) {
/* SHOW CREATE TABLE wants constraints each printed nicely
on its own line, while error messages want no newlines
inserted. */
str.append("\n ");
}
str.append(" CONSTRAINT ");
str.append(innobase_quote_identifier(trx, stripped_id));
str.append(" FOREIGN KEY (");
for (i = 0;;) {
str.append(innobase_quote_identifier(trx, foreign->foreign_col_names[i]));
if (++i < foreign->n_fields) {
str.append(", ");
} else {
break;
}
}
str.append(") REFERENCES ");
if (dict_tables_have_same_db(foreign->foreign_table_name_lookup,
foreign->referenced_table_name_lookup)) {
/* Do not print the database name of the referenced table */
str.append(ut_get_name(trx,
dict_remove_db_name(
foreign->referenced_table_name)));
} else {
str.append(ut_get_name(trx,
foreign->referenced_table_name));
}
str.append(" (");
for (i = 0;;) {
str.append(innobase_quote_identifier(trx,
foreign->referenced_col_names[i]));
if (++i < foreign->n_fields) {
str.append(", ");
} else {
break;
}
}
str.append(")");
if (foreign->type & foreign->DELETE_CASCADE) {
str.append(" ON DELETE CASCADE");
}
if (foreign->type & foreign->DELETE_SET_NULL) {
str.append(" ON DELETE SET NULL");
}
if (foreign->type & foreign->DELETE_NO_ACTION) {
str.append(" ON DELETE NO ACTION");
}
if (foreign->type & foreign->UPDATE_CASCADE) {
str.append(" ON UPDATE CASCADE");
}
if (foreign->type & foreign->UPDATE_SET_NULL) {
str.append(" ON UPDATE SET NULL");
}
if (foreign->type & foreign->UPDATE_NO_ACTION) {
str.append(" ON UPDATE NO ACTION");
}
return str;
}
/**********************************************************************//**
Outputs info on foreign keys of a table. */
std::string
dict_print_info_on_foreign_keys(
/*============================*/
ibool create_table_format, /*!< in: if TRUE then print in
a format suitable to be inserted into
a CREATE TABLE, otherwise in the format
of SHOW TABLE STATUS */
trx_t* trx, /*!< in: transaction */
dict_table_t* table) /*!< in: table */
{
dict_foreign_t* foreign;
std::string str;
dict_sys.freeze(SRW_LOCK_CALL);
for (dict_foreign_set::iterator it = table->foreign_set.begin();
it != table->foreign_set.end();
++it) {
foreign = *it;
if (create_table_format) {
str.append(
dict_print_info_on_foreign_key_in_create_format(
trx, foreign, TRUE));
} else {
ulint i;
str.append("; (");
for (i = 0; i < foreign->n_fields; i++) {
if (i) {
str.append(" ");
}
str.append(innobase_quote_identifier(trx,
foreign->foreign_col_names[i]));
}
str.append(") REFER ");
str.append(ut_get_name(trx,
foreign->referenced_table_name));
str.append(")");
for (i = 0; i < foreign->n_fields; i++) {
if (i) {
str.append(" ");
}
str.append(innobase_quote_identifier(
trx,
foreign->referenced_col_names[i]));
}
str.append(")");
if (foreign->type == foreign->DELETE_CASCADE) {
str.append(" ON DELETE CASCADE");
}
if (foreign->type == foreign->DELETE_SET_NULL) {
str.append(" ON DELETE SET NULL");
}
if (foreign->type & foreign->DELETE_NO_ACTION) {
str.append(" ON DELETE NO ACTION");
}
if (foreign->type & foreign->UPDATE_CASCADE) {
str.append(" ON UPDATE CASCADE");
}
if (foreign->type & foreign->UPDATE_SET_NULL) {
str.append(" ON UPDATE SET NULL");
}
if (foreign->type & foreign->UPDATE_NO_ACTION) {
str.append(" ON UPDATE NO ACTION");
}
}
}
dict_sys.unfreeze();
return str;
}
/**********************************************************************//**
Flags an index corrupted both in the data dictionary cache
and in the SYS_INDEXES */
void dict_set_corrupted(dict_index_t *index, const char *ctx)
{
mem_heap_t* heap;
mtr_t mtr;
dict_index_t* sys_index;
dtuple_t* tuple;
dfield_t* dfield;
byte* buf;
const char* status;
btr_cur_t cursor;
dict_sys.lock(SRW_LOCK_CALL);
ut_ad(!dict_table_is_comp(dict_sys.sys_tables));
ut_ad(!dict_table_is_comp(dict_sys.sys_indexes));
/* Mark the table as corrupted only if the clustered index
is corrupted */
if (dict_index_is_clust(index)) {
index->table->corrupted = TRUE;
goto func_exit;
}
if (index->type & DICT_CORRUPT) {
/* The index was already flagged corrupted. */
ut_ad(!dict_index_is_clust(index) || index->table->corrupted);
goto func_exit;
}
/* If this is read only mode, do not update SYS_INDEXES, just
mark it as corrupted in memory */
if (high_level_read_only) {
index->type |= DICT_CORRUPT;
goto func_exit;
}
heap = mem_heap_create(sizeof(dtuple_t) + 2 * (sizeof(dfield_t)
+ sizeof(que_fork_t) + sizeof(upd_node_t)
+ sizeof(upd_t) + 12));
mtr_start(&mtr);
index->type |= DICT_CORRUPT;
sys_index = UT_LIST_GET_FIRST(dict_sys.sys_indexes->indexes);
/* Find the index row in SYS_INDEXES */
tuple = dtuple_create(heap, 2);
dfield = dtuple_get_nth_field(tuple, 0);
buf = static_cast<byte*>(mem_heap_alloc(heap, 8));
mach_write_to_8(buf, index->table->id);
dfield_set_data(dfield, buf, 8);
dfield = dtuple_get_nth_field(tuple, 1);
buf = static_cast<byte*>(mem_heap_alloc(heap, 8));
mach_write_to_8(buf, index->id);
dfield_set_data(dfield, buf, 8);
dict_index_copy_types(tuple, sys_index, 2);
cursor.page_cur.index = sys_index;
if (cursor.search_leaf(tuple, PAGE_CUR_LE, BTR_MODIFY_LEAF, &mtr)
!= DB_SUCCESS) {
goto fail;
}
if (cursor.low_match == dtuple_get_n_fields(tuple)) {
/* UPDATE SYS_INDEXES SET TYPE=index->type
WHERE TABLE_ID=index->table->id AND INDEX_ID=index->id */
ulint len;
byte* field = rec_get_nth_field_old(
btr_cur_get_rec(&cursor),
DICT_FLD__SYS_INDEXES__TYPE, &len);
if (len != 4) {
goto fail;
}
mtr.write<4>(*btr_cur_get_block(&cursor), field, index->type);
status = "Flagged";
} else {
fail:
status = "Unable to flag";
}
mtr_commit(&mtr);
mem_heap_free(heap);
ib::error() << status << " corruption of " << index->name
<< " in table " << index->table->name << " in " << ctx;
func_exit:
dict_sys.unlock();
}
/** Sets merge_threshold in the SYS_INDEXES
@param[in,out] index index
@param[in] merge_threshold value to set */
void
dict_index_set_merge_threshold(
dict_index_t* index,
ulint merge_threshold)
{
mem_heap_t* heap;
mtr_t mtr;
dict_index_t* sys_index;
dtuple_t* tuple;
dfield_t* dfield;
byte* buf;
btr_cur_t cursor;
ut_ad(index != NULL);
ut_ad(!dict_table_is_comp(dict_sys.sys_tables));
ut_ad(!dict_table_is_comp(dict_sys.sys_indexes));
heap = mem_heap_create(sizeof(dtuple_t) + 2 * (sizeof(dfield_t)
+ sizeof(que_fork_t) + sizeof(upd_node_t)
+ sizeof(upd_t) + 12));
mtr.start();
sys_index = UT_LIST_GET_FIRST(dict_sys.sys_indexes->indexes);
/* Find the index row in SYS_INDEXES */
tuple = dtuple_create(heap, 2);
dfield = dtuple_get_nth_field(tuple, 0);
buf = static_cast<byte*>(mem_heap_alloc(heap, 8));
mach_write_to_8(buf, index->table->id);
dfield_set_data(dfield, buf, 8);
dfield = dtuple_get_nth_field(tuple, 1);
buf = static_cast<byte*>(mem_heap_alloc(heap, 8));
mach_write_to_8(buf, index->id);
dfield_set_data(dfield, buf, 8);
dict_index_copy_types(tuple, sys_index, 2);
cursor.page_cur.index = sys_index;
if (cursor.search_leaf(tuple, PAGE_CUR_GE, BTR_MODIFY_LEAF, &mtr)
!= DB_SUCCESS) {
goto func_exit;
}
if (cursor.up_match == dtuple_get_n_fields(tuple)
&& rec_get_n_fields_old(btr_cur_get_rec(&cursor))
== DICT_NUM_FIELDS__SYS_INDEXES) {
ulint len;
byte* field = rec_get_nth_field_old(
btr_cur_get_rec(&cursor),
DICT_FLD__SYS_INDEXES__MERGE_THRESHOLD, &len);
ut_ad(len == 4);
mtr.write<4,mtr_t::MAYBE_NOP>(*btr_cur_get_block(&cursor),
field, merge_threshold);
}
func_exit:
mtr_commit(&mtr);
mem_heap_free(heap);
}
#ifdef UNIV_DEBUG
/** Sets merge_threshold for all indexes in the list of tables
@param[in] list pointer to the list of tables */
inline
void
dict_set_merge_threshold_list_debug(
UT_LIST_BASE_NODE_T(dict_table_t)* list,
uint merge_threshold_all)
{
for (dict_table_t* table = UT_LIST_GET_FIRST(*list);
table != NULL;
table = UT_LIST_GET_NEXT(table_LRU, table)) {
for (dict_index_t* index = UT_LIST_GET_FIRST(table->indexes);
index != NULL;
index = UT_LIST_GET_NEXT(indexes, index)) {
index->lock.x_lock(SRW_LOCK_CALL);
index->merge_threshold = merge_threshold_all
& ((1U << 6) - 1);
index->lock.x_unlock();
}
}
}
/** Sets merge_threshold for all indexes in dictionary cache for debug.
@param[in] merge_threshold_all value to set for all indexes */
void
dict_set_merge_threshold_all_debug(
uint merge_threshold_all)
{
dict_sys.freeze(SRW_LOCK_CALL);
dict_set_merge_threshold_list_debug(
&dict_sys.table_LRU, merge_threshold_all);
dict_set_merge_threshold_list_debug(
&dict_sys.table_non_LRU, merge_threshold_all);
dict_sys.unfreeze();
}
#endif /* UNIV_DEBUG */
/** Get an index by name.
@param[in] table the table where to look for the index
@param[in] name the index name to look for
@return index, NULL if does not exist */
dict_index_t*
dict_table_get_index_on_name(dict_table_t* table, const char* name)
{
dict_index_t* index;
index = dict_table_get_first_index(table);
while (index != NULL) {
if (index->is_committed() && !strcmp(index->name, name)) {
return(index);
}
index = dict_table_get_next_index(index);
}
return(NULL);
}
/**********************************************************************//**
Replace the index passed in with another equivalent index in the
foreign key lists of the table.
@return whether all replacements were found */
bool
dict_foreign_replace_index(
/*=======================*/
dict_table_t* table, /*!< in/out: table */
const char** col_names,
/*!< in: column names, or NULL
to use table->col_names */
const dict_index_t* index) /*!< in: index to be replaced */
{
bool found = true;
dict_foreign_t* foreign;
ut_ad(index->to_be_dropped);
ut_ad(index->table == table);
for (dict_foreign_set::iterator it = table->foreign_set.begin();
it != table->foreign_set.end();
++it) {
foreign = *it;
if (foreign->foreign_index == index) {
ut_ad(foreign->foreign_table == index->table);
dict_index_t* new_index = dict_foreign_find_index(
foreign->foreign_table, col_names,
foreign->foreign_col_names,
foreign->n_fields, index,
/*check_charsets=*/TRUE, /*check_null=*/FALSE,
NULL, NULL, NULL);
if (new_index) {
ut_ad(new_index->table == index->table);
ut_ad(!new_index->to_be_dropped);
} else {
found = false;
}
foreign->foreign_index = new_index;
}
}
for (dict_foreign_set::iterator it = table->referenced_set.begin();
it != table->referenced_set.end();
++it) {
foreign = *it;
if (foreign->referenced_index == index) {
ut_ad(foreign->referenced_table == index->table);
dict_index_t* new_index = dict_foreign_find_index(
foreign->referenced_table, NULL,
foreign->referenced_col_names,
foreign->n_fields, index,
/*check_charsets=*/TRUE, /*check_null=*/FALSE,
NULL, NULL, NULL);
/* There must exist an alternative index,
since this must have been checked earlier. */
if (new_index) {
ut_ad(new_index->table == index->table);
ut_ad(!new_index->to_be_dropped);
} else {
found = false;
}
foreign->referenced_index = new_index;
}
}
return(found);
}
#ifdef UNIV_DEBUG
/**********************************************************************//**
Check for duplicate index entries in a table [using the index name] */
void
dict_table_check_for_dup_indexes(
/*=============================*/
const dict_table_t* table, /*!< in: Check for dup indexes
in this table */
enum check_name check) /*!< in: whether and when to allow
temporary index names */
{
/* Check for duplicates, ignoring indexes that are marked
as to be dropped */
const dict_index_t* index1;
const dict_index_t* index2;
ut_ad(dict_sys.frozen());
/* The primary index _must_ exist */
ut_a(UT_LIST_GET_LEN(table->indexes) > 0);
index1 = UT_LIST_GET_FIRST(table->indexes);
do {
if (!index1->is_committed()) {
ut_a(!dict_index_is_clust(index1));
switch (check) {
case CHECK_ALL_COMPLETE:
ut_error;
case CHECK_ABORTED_OK:
switch (dict_index_get_online_status(index1)) {
case ONLINE_INDEX_COMPLETE:
case ONLINE_INDEX_CREATION:
ut_error;
break;
case ONLINE_INDEX_ABORTED:
case ONLINE_INDEX_ABORTED_DROPPED:
break;
}
/* fall through */
case CHECK_PARTIAL_OK:
break;
}
}
for (index2 = UT_LIST_GET_NEXT(indexes, index1);
index2 != NULL;
index2 = UT_LIST_GET_NEXT(indexes, index2)) {
ut_ad(index1->is_committed()
!= index2->is_committed()
|| strcmp(index1->name, index2->name) != 0);
}
index1 = UT_LIST_GET_NEXT(indexes, index1);
} while (index1);
}
#endif /* UNIV_DEBUG */
/*********************************************************************//**
Converts a database and table name from filesystem encoding
(e.g. d@i1b/a@q1b@1Kc, same format as used in dict_table_t::name) in two
strings in UTF8 encoding (e.g. dцb and aюbØc). The output buffers must be
at least MAX_DB_UTF8_LEN and MAX_TABLE_UTF8_LEN bytes. */
void
dict_fs2utf8(
/*=========*/
const char* db_and_table, /*!< in: database and table names,
e.g. d@i1b/a@q1b@1Kc */
char* db_utf8, /*!< out: database name, e.g. dцb */
size_t db_utf8_size, /*!< in: dbname_utf8 size */
char* table_utf8, /*!< out: table name, e.g. aюbØc */
size_t table_utf8_size)/*!< in: table_utf8 size */
{
char db[MAX_DATABASE_NAME_LEN + 1];
ulint db_len;
uint errors;
db_len = dict_get_db_name_len(db_and_table);
ut_a(db_len <= sizeof(db));
memcpy(db, db_and_table, db_len);
db[db_len] = '\0';
strconvert(
&my_charset_filename, db, uint(db_len), system_charset_info,
db_utf8, uint(db_utf8_size), &errors);
/* convert each # to @0023 in table name and store the result in buf */
const char* table = dict_remove_db_name(db_and_table);
const char* table_p;
char buf[MAX_TABLE_NAME_LEN * 5 + 1];
char* buf_p;
for (table_p = table, buf_p = buf; table_p[0] != '\0'; table_p++) {
if (table_p[0] != '#') {
buf_p[0] = table_p[0];
buf_p++;
} else {
buf_p[0] = '@';
buf_p[1] = '0';
buf_p[2] = '0';
buf_p[3] = '2';
buf_p[4] = '3';
buf_p += 5;
}
ut_a((size_t) (buf_p - buf) < sizeof(buf));
}
buf_p[0] = '\0';
errors = 0;
strconvert(
&my_charset_filename, buf, (uint) (buf_p - buf),
system_charset_info,
table_utf8, uint(table_utf8_size),
&errors);
if (errors != 0) {
snprintf(table_utf8, table_utf8_size, "%s%s",
srv_mysql50_table_name_prefix, table);
}
}
/** Insert a table into the hash tables
@param table the table
@param id_hash dict_sys.table_id_hash or dict_sys.temp_id_hash */
static void hash_insert(dict_table_t *table, hash_table_t& id_hash) noexcept
{
ut_ad(table->cached);
dict_sys.table_hash.cell_get(my_crc32c(0, table->name.m_name,
strlen(table->name.m_name)))->
append(*table, &dict_table_t::name_hash);
id_hash.cell_get(ut_fold_ull(table->id))->append(*table,
&dict_table_t::id_hash);
}
/** Resize the hash tables based on the current buffer pool size. */
void dict_sys_t::resize() noexcept
{
ut_ad(this == &dict_sys);
ut_ad(is_initialised());
lock(SRW_LOCK_CALL);
/* all table entries are in table_LRU and table_non_LRU lists */
table_hash.free();
table_id_hash.free();
temp_id_hash.free();
const ulint hash_size = buf_pool_get_curr_size()
/ (DICT_POOL_PER_TABLE_HASH * UNIV_WORD_SIZE);
table_hash.create(hash_size);
table_id_hash.create(hash_size);
temp_id_hash.create(hash_size);
for (dict_table_t *table= UT_LIST_GET_FIRST(table_LRU); table;
table= UT_LIST_GET_NEXT(table_LRU, table))
{
ut_ad(!table->is_temporary());
hash_insert(table, table_id_hash);
}
for (dict_table_t *table = UT_LIST_GET_FIRST(table_non_LRU); table;
table= UT_LIST_GET_NEXT(table_LRU, table))
hash_insert(table, table->is_temporary() ? temp_id_hash : table_id_hash);
unlock();
}
/** Close the data dictionary cache on shutdown. */
void dict_sys_t::close() noexcept
{
ut_ad(this == &dict_sys);
if (!is_initialised()) return;
lock(SRW_LOCK_CALL);
/* Free the hash elements. We don't remove them from table_hash
because we are invoking table_hash.free() below. */
for (ulint i= table_hash.n_cells; i--; )
while (auto table= static_cast<dict_table_t*>(table_hash.array[i].node))
dict_sys.remove(table);
table_hash.free();
/* table_id_hash contains the same elements as in table_hash,
therefore we don't delete the individual elements. */
table_id_hash.free();
/* No temporary tables should exist at this point. */
temp_id_hash.free();
unlock();
latch.destroy();
mysql_mutex_destroy(&dict_foreign_err_mutex);
if (dict_foreign_err_file)
{
my_fclose(dict_foreign_err_file, MYF(MY_WME));
dict_foreign_err_file = NULL;
}
m_initialised= false;
}
#ifdef UNIV_DEBUG
/**********************************************************************//**
Validate the dictionary table LRU list.
@return TRUE if valid */
static
ibool
dict_lru_validate(void)
/*===================*/
{
dict_table_t* table;
ut_ad(dict_sys.frozen());
for (table = UT_LIST_GET_FIRST(dict_sys.table_LRU);
table != NULL;
table = UT_LIST_GET_NEXT(table_LRU, table)) {
ut_a(table->can_be_evicted);
}
for (table = UT_LIST_GET_FIRST(dict_sys.table_non_LRU);
table != NULL;
table = UT_LIST_GET_NEXT(table_LRU, table)) {
ut_a(!table->can_be_evicted);
}
return(TRUE);
}
#endif /* UNIV_DEBUG */
/*********************************************************************//**
Check an index to see whether its first fields are the columns in the array,
in the same order and is not marked for deletion and is not the same
as types_idx.
@return true if the index qualifies, otherwise false */
bool
dict_foreign_qualify_index(
/*=======================*/
const dict_table_t* table, /*!< in: table */
const char** col_names,
/*!< in: column names, or NULL
to use table->col_names */
const char** columns,/*!< in: array of column names */
ulint n_cols, /*!< in: number of columns */
const dict_index_t* index, /*!< in: index to check */
const dict_index_t* types_idx,
/*!< in: NULL or an index
whose types the column types
must match */
bool check_charsets,
/*!< in: whether to check
charsets. only has an effect
if types_idx != NULL */
ulint check_null,
/*!< in: nonzero if none of
the columns must be declared
NOT NULL */
fkerr_t* error, /*!< out: error code */
ulint* err_col_no,
/*!< out: column number where
error happened */
dict_index_t** err_index)
/*!< out: index where error
happened */
{
if (dict_index_get_n_fields(index) < n_cols) {
return(false);
}
if (!index->is_btree()) {
return false;
}
if (index->online_status >= ONLINE_INDEX_ABORTED) {
return false;
}
for (ulint i = 0; i < n_cols; i++) {
dict_field_t* field;
const char* col_name;
ulint col_no;
field = dict_index_get_nth_field(index, i);
col_no = dict_col_get_no(field->col);
if (field->prefix_len != 0) {
/* We do not accept column prefix
indexes here */
if (error && err_col_no && err_index) {
*error = FK_IS_PREFIX_INDEX;
*err_col_no = i;
*err_index = (dict_index_t*)index;
}
return(false);
}
if (check_null
&& (field->col->prtype & DATA_NOT_NULL)) {
if (error && err_col_no && err_index) {
*error = FK_COL_NOT_NULL;
*err_col_no = i;
*err_index = (dict_index_t*)index;
}
return(false);
}
if (field->col->is_virtual()) {
col_name = "";
for (ulint j = 0; j < table->n_v_def; j++) {
col_name = dict_table_get_v_col_name(table, j);
if (innobase_strcasecmp(field->name,col_name) == 0) {
break;
}
}
} else {
col_name = col_names
? col_names[col_no]
: dict_table_get_col_name(table, col_no);
}
if (0 != innobase_strcasecmp(columns[i], col_name)) {
return(false);
}
if (types_idx && !cmp_cols_are_equal(
dict_index_get_nth_col(index, i),
dict_index_get_nth_col(types_idx, i),
check_charsets)) {
if (error && err_col_no && err_index) {
*error = FK_COLS_NOT_EQUAL;
*err_col_no = i;
*err_index = (dict_index_t*)index;
}
return(false);
}
}
return(true);
}
/*********************************************************************//**
Update the state of compression failure padding heuristics. This is
called whenever a compression operation succeeds or fails.
The caller must be holding info->mutex */
static
void
dict_index_zip_pad_update(
/*======================*/
zip_pad_info_t* info, /*<! in/out: info to be updated */
ulint zip_threshold) /*<! in: zip threshold value */
{
ulint total;
ulint fail_pct;
ut_ad(info);
ut_ad(info->pad % ZIP_PAD_INCR == 0);
total = info->success + info->failure;
ut_ad(total > 0);
if (zip_threshold == 0) {
/* User has just disabled the padding. */
return;
}
if (total < ZIP_PAD_ROUND_LEN) {
/* We are in middle of a round. Do nothing. */
return;
}
/* We are at a 'round' boundary. Reset the values but first
calculate fail rate for our heuristic. */
fail_pct = (info->failure * 100) / total;
info->failure = 0;
info->success = 0;
if (fail_pct > zip_threshold) {
/* Compression failures are more then user defined
threshold. Increase the pad size to reduce chances of
compression failures.
Only do increment if it won't increase padding
beyond max pad size. */
if (info->pad + ZIP_PAD_INCR
< (srv_page_size * zip_pad_max) / 100) {
info->pad.fetch_add(ZIP_PAD_INCR);
MONITOR_INC(MONITOR_PAD_INCREMENTS);
}
info->n_rounds = 0;
} else {
/* Failure rate was OK. Another successful round
completed. */
++info->n_rounds;
/* If enough successful rounds are completed with
compression failure rate in control, decrease the
padding. */
if (info->n_rounds >= ZIP_PAD_SUCCESSFUL_ROUND_LIMIT
&& info->pad > 0) {
info->pad.fetch_sub(ZIP_PAD_INCR);
info->n_rounds = 0;
MONITOR_INC(MONITOR_PAD_DECREMENTS);
}
}
}
/*********************************************************************//**
This function should be called whenever a page is successfully
compressed. Updates the compression padding information. */
void
dict_index_zip_success(
/*===================*/
dict_index_t* index) /*!< in/out: index to be updated. */
{
ulint zip_threshold = zip_failure_threshold_pct;
if (!zip_threshold) {
/* Disabled by user. */
return;
}
index->zip_pad.mutex.lock();
++index->zip_pad.success;
dict_index_zip_pad_update(&index->zip_pad, zip_threshold);
index->zip_pad.mutex.unlock();
}
/*********************************************************************//**
This function should be called whenever a page compression attempt
fails. Updates the compression padding information. */
void
dict_index_zip_failure(
/*===================*/
dict_index_t* index) /*!< in/out: index to be updated. */
{
ulint zip_threshold = zip_failure_threshold_pct;
if (!zip_threshold) {
/* Disabled by user. */
return;
}
index->zip_pad.mutex.lock();
++index->zip_pad.failure;
dict_index_zip_pad_update(&index->zip_pad, zip_threshold);
index->zip_pad.mutex.unlock();
}
/*********************************************************************//**
Return the optimal page size, for which page will likely compress.
@return page size beyond which page might not compress */
ulint
dict_index_zip_pad_optimal_page_size(
/*=================================*/
dict_index_t* index) /*!< in: index for which page size
is requested */
{
ulint pad;
ulint min_sz;
ulint sz;
if (!zip_failure_threshold_pct) {
/* Disabled by user. */
return(srv_page_size);
}
pad = index->zip_pad.pad;
ut_ad(pad < srv_page_size);
sz = srv_page_size - pad;
/* Min size allowed by user. */
ut_ad(zip_pad_max < 100);
min_sz = (srv_page_size * (100 - zip_pad_max)) / 100;
return(ut_max(sz, min_sz));
}
/*************************************************************//**
Convert table flag to row format string.
@return row format name. */
const char*
dict_tf_to_row_format_string(
/*=========================*/
ulint table_flag) /*!< in: row format setting */
{
switch (dict_tf_get_rec_format(table_flag)) {
case REC_FORMAT_REDUNDANT:
return("ROW_TYPE_REDUNDANT");
case REC_FORMAT_COMPACT:
return("ROW_TYPE_COMPACT");
case REC_FORMAT_COMPRESSED:
return("ROW_TYPE_COMPRESSED");
case REC_FORMAT_DYNAMIC:
return("ROW_TYPE_DYNAMIC");
}
ut_error;
return(0);
}