2014-02-26 19:11:54 +01:00
|
|
|
/*****************************************************************************
|
|
|
|
|
2016-04-26 23:05:26 +02:00
|
|
|
Copyright (c) 1996, 2016, Oracle and/or its affiliates. All Rights Reserved.
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
This program is free software; you can redistribute it and/or modify it under
|
|
|
|
the terms of the GNU General Public License as published by the Free Software
|
|
|
|
Foundation; version 2 of the License.
|
|
|
|
|
|
|
|
This program is distributed in the hope that it will be useful, but WITHOUT
|
|
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
|
|
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License along with
|
|
|
|
this program; if not, write to the Free Software Foundation, Inc.,
|
|
|
|
51 Franklin Street, Suite 500, Boston, MA 02110-1335 USA
|
|
|
|
|
|
|
|
*****************************************************************************/
|
|
|
|
|
|
|
|
/**************************************************//**
|
|
|
|
@file dict/dict0crea.cc
|
|
|
|
Database object creation
|
|
|
|
|
|
|
|
Created 1/8/1996 Heikki Tuuri
|
|
|
|
*******************************************************/
|
|
|
|
|
|
|
|
#include "dict0crea.h"
|
|
|
|
|
|
|
|
#ifdef UNIV_NONINL
|
|
|
|
#include "dict0crea.ic"
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#include "btr0pcur.h"
|
|
|
|
#include "btr0btr.h"
|
|
|
|
#include "page0page.h"
|
|
|
|
#include "mach0data.h"
|
|
|
|
#include "dict0boot.h"
|
|
|
|
#include "dict0dict.h"
|
|
|
|
#include "que0que.h"
|
|
|
|
#include "row0ins.h"
|
|
|
|
#include "row0mysql.h"
|
|
|
|
#include "pars0pars.h"
|
|
|
|
#include "trx0roll.h"
|
|
|
|
#include "usr0sess.h"
|
|
|
|
#include "ut0vec.h"
|
|
|
|
#include "dict0priv.h"
|
|
|
|
#include "fts0priv.h"
|
|
|
|
#include "ha_prototypes.h"
|
|
|
|
|
|
|
|
/*****************************************************************//**
|
|
|
|
Based on a table object, this function builds the entry to be inserted
|
|
|
|
in the SYS_TABLES system table.
|
|
|
|
@return the tuple which should be inserted */
|
|
|
|
static
|
|
|
|
dtuple_t*
|
|
|
|
dict_create_sys_tables_tuple(
|
|
|
|
/*=========================*/
|
|
|
|
const dict_table_t* table, /*!< in: table */
|
|
|
|
mem_heap_t* heap) /*!< in: memory heap from
|
|
|
|
which the memory for the built
|
|
|
|
tuple is allocated */
|
|
|
|
{
|
|
|
|
dict_table_t* sys_tables;
|
|
|
|
dtuple_t* entry;
|
|
|
|
dfield_t* dfield;
|
|
|
|
byte* ptr;
|
|
|
|
ulint type;
|
|
|
|
|
|
|
|
ut_ad(table);
|
|
|
|
ut_ad(heap);
|
|
|
|
|
|
|
|
sys_tables = dict_sys->sys_tables;
|
|
|
|
|
|
|
|
entry = dtuple_create(heap, 8 + DATA_N_SYS_COLS);
|
|
|
|
|
|
|
|
dict_table_copy_types(entry, sys_tables);
|
|
|
|
|
|
|
|
/* 0: NAME -----------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__NAME);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, table->name, ut_strlen(table->name));
|
|
|
|
|
|
|
|
/* 1: DB_TRX_ID added later */
|
|
|
|
/* 2: DB_ROLL_PTR added later */
|
|
|
|
/* 3: ID -------------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__ID);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 8));
|
|
|
|
mach_write_to_8(ptr, table->id);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 8);
|
|
|
|
|
|
|
|
/* 4: N_COLS ---------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__N_COLS);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, table->n_def
|
|
|
|
| ((table->flags & DICT_TF_COMPACT) << 31));
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
2017-07-25 00:41:54 +03:00
|
|
|
|
2014-02-26 19:11:54 +01:00
|
|
|
/* 5: TYPE (table flags) -----------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__TYPE);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
|
|
|
|
/* Validate the table flags and convert them to what is saved in
|
|
|
|
SYS_TABLES.TYPE. Table flag values 0 and 1 are both written to
|
|
|
|
SYS_TABLES.TYPE as 1. */
|
|
|
|
type = dict_tf_to_sys_tables_type(table->flags);
|
|
|
|
mach_write_to_4(ptr, type);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 6: MIX_ID (obsolete) ---------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__MIX_ID);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_zalloc(heap, 8));
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 8);
|
|
|
|
|
|
|
|
/* 7: MIX_LEN (additional flags) --------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__MIX_LEN);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
/* Be sure all non-used bits are zero. */
|
|
|
|
ut_a(!(table->flags2 & ~DICT_TF2_BIT_MASK));
|
|
|
|
mach_write_to_4(ptr, table->flags2);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 8: CLUSTER_NAME ---------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__CLUSTER_ID);
|
|
|
|
dfield_set_null(dfield); /* not supported */
|
|
|
|
|
|
|
|
/* 9: SPACE ----------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_TABLES__SPACE);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, table->space);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
/*----------------------------------*/
|
|
|
|
|
|
|
|
return(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*****************************************************************//**
|
|
|
|
Based on a table object, this function builds the entry to be inserted
|
|
|
|
in the SYS_COLUMNS system table.
|
|
|
|
@return the tuple which should be inserted */
|
|
|
|
static
|
|
|
|
dtuple_t*
|
|
|
|
dict_create_sys_columns_tuple(
|
|
|
|
/*==========================*/
|
|
|
|
const dict_table_t* table, /*!< in: table */
|
|
|
|
ulint i, /*!< in: column number */
|
|
|
|
mem_heap_t* heap) /*!< in: memory heap from
|
|
|
|
which the memory for the built
|
|
|
|
tuple is allocated */
|
|
|
|
{
|
|
|
|
dict_table_t* sys_columns;
|
|
|
|
dtuple_t* entry;
|
|
|
|
const dict_col_t* column;
|
|
|
|
dfield_t* dfield;
|
|
|
|
byte* ptr;
|
|
|
|
const char* col_name;
|
|
|
|
|
|
|
|
ut_ad(table);
|
|
|
|
ut_ad(heap);
|
|
|
|
|
|
|
|
column = dict_table_get_nth_col(table, i);
|
|
|
|
|
|
|
|
sys_columns = dict_sys->sys_columns;
|
|
|
|
|
|
|
|
entry = dtuple_create(heap, 7 + DATA_N_SYS_COLS);
|
|
|
|
|
|
|
|
dict_table_copy_types(entry, sys_columns);
|
|
|
|
|
|
|
|
/* 0: TABLE_ID -----------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__TABLE_ID);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 8));
|
|
|
|
mach_write_to_8(ptr, table->id);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 8);
|
|
|
|
|
|
|
|
/* 1: POS ----------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__POS);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, i);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 2: DB_TRX_ID added later */
|
|
|
|
/* 3: DB_ROLL_PTR added later */
|
|
|
|
/* 4: NAME ---------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__NAME);
|
|
|
|
|
|
|
|
col_name = dict_table_get_col_name(table, i);
|
|
|
|
dfield_set_data(dfield, col_name, ut_strlen(col_name));
|
|
|
|
|
|
|
|
/* 5: MTYPE --------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__MTYPE);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, column->mtype);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 6: PRTYPE -------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__PRTYPE);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, column->prtype);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 7: LEN ----------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__LEN);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, column->len);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 8: PREC ---------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_COLUMNS__PREC);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, 0/* unused */);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
/*---------------------------------*/
|
|
|
|
|
|
|
|
return(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***************************************************************//**
|
|
|
|
Builds a table definition to insert.
|
|
|
|
@return DB_SUCCESS or error code */
|
2016-06-21 14:21:03 +02:00
|
|
|
static MY_ATTRIBUTE((nonnull, warn_unused_result))
|
2014-02-26 19:11:54 +01:00
|
|
|
dberr_t
|
|
|
|
dict_build_table_def_step(
|
|
|
|
/*======================*/
|
|
|
|
que_thr_t* thr, /*!< in: query thread */
|
|
|
|
tab_node_t* node) /*!< in: table create node */
|
|
|
|
{
|
|
|
|
dict_table_t* table;
|
|
|
|
dtuple_t* row;
|
|
|
|
dberr_t error;
|
|
|
|
const char* path;
|
|
|
|
mtr_t mtr;
|
|
|
|
ulint space = 0;
|
|
|
|
bool use_tablespace;
|
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
|
|
|
|
table = node->table;
|
|
|
|
use_tablespace = DICT_TF2_FLAG_IS_SET(table, DICT_TF2_USE_TABLESPACE);
|
|
|
|
|
|
|
|
dict_hdr_get_new_id(&table->id, NULL, NULL);
|
|
|
|
|
|
|
|
thr_get_trx(thr)->table_id = table->id;
|
|
|
|
|
2014-05-05 18:20:28 +02:00
|
|
|
/* Always set this bit for all new created tables */
|
|
|
|
DICT_TF2_FLAG_SET(table, DICT_TF2_FTS_AUX_HEX_NAME);
|
|
|
|
DBUG_EXECUTE_IF("innodb_test_wrong_fts_aux_table_name",
|
|
|
|
DICT_TF2_FLAG_UNSET(table,
|
|
|
|
DICT_TF2_FTS_AUX_HEX_NAME););
|
|
|
|
|
2014-02-26 19:11:54 +01:00
|
|
|
if (use_tablespace) {
|
|
|
|
/* This table will not use the system tablespace.
|
|
|
|
Get a new space id. */
|
|
|
|
dict_hdr_get_new_id(NULL, NULL, &space);
|
|
|
|
|
|
|
|
DBUG_EXECUTE_IF(
|
|
|
|
"ib_create_table_fail_out_of_space_ids",
|
|
|
|
space = ULINT_UNDEFINED;
|
|
|
|
);
|
|
|
|
|
|
|
|
if (UNIV_UNLIKELY(space == ULINT_UNDEFINED)) {
|
|
|
|
return(DB_ERROR);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* We create a new single-table tablespace for the table.
|
|
|
|
We initially let it be 4 pages:
|
|
|
|
- page 0 is the fsp header and an extent descriptor page,
|
|
|
|
- page 1 is an ibuf bitmap page,
|
|
|
|
- page 2 is the first inode page,
|
|
|
|
- page 3 will contain the root of the clustered index of the
|
|
|
|
table we create here. */
|
|
|
|
|
|
|
|
path = table->data_dir_path ? table->data_dir_path
|
|
|
|
: table->dir_path_of_temp_table;
|
|
|
|
|
|
|
|
ut_ad(dict_table_get_format(table) <= UNIV_FORMAT_MAX);
|
|
|
|
ut_ad(!dict_table_zip_size(table)
|
|
|
|
|| dict_table_get_format(table) >= UNIV_FORMAT_B);
|
|
|
|
|
|
|
|
error = fil_create_new_single_table_tablespace(
|
|
|
|
space, table->name, path,
|
|
|
|
dict_tf_to_fsp_flags(table->flags),
|
|
|
|
table->flags2,
|
2015-08-14 11:09:06 +03:00
|
|
|
FIL_IBD_FILE_INITIAL_SIZE,
|
|
|
|
node->mode, node->key_id);
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
table->space = (unsigned int) space;
|
|
|
|
|
|
|
|
if (error != DB_SUCCESS) {
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
mtr_start(&mtr);
|
|
|
|
|
2017-06-06 14:59:42 +03:00
|
|
|
fsp_header_init(table->space, FIL_IBD_FILE_INITIAL_SIZE, &mtr);
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
mtr_commit(&mtr);
|
|
|
|
} else {
|
|
|
|
/* Create in the system tablespace: disallow Barracuda
|
|
|
|
features by keeping only the first bit which says whether
|
|
|
|
the row format is redundant or compact */
|
|
|
|
table->flags &= DICT_TF_COMPACT;
|
|
|
|
}
|
|
|
|
|
|
|
|
row = dict_create_sys_tables_tuple(table, node->heap);
|
|
|
|
|
|
|
|
ins_node_set_new_row(node->tab_def, row);
|
|
|
|
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***************************************************************//**
|
|
|
|
Builds a column definition to insert. */
|
|
|
|
static
|
|
|
|
void
|
|
|
|
dict_build_col_def_step(
|
|
|
|
/*====================*/
|
|
|
|
tab_node_t* node) /*!< in: table create node */
|
|
|
|
{
|
|
|
|
dtuple_t* row;
|
|
|
|
|
|
|
|
row = dict_create_sys_columns_tuple(node->table, node->col_no,
|
|
|
|
node->heap);
|
|
|
|
ins_node_set_new_row(node->col_def, row);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*****************************************************************//**
|
|
|
|
Based on an index object, this function builds the entry to be inserted
|
|
|
|
in the SYS_INDEXES system table.
|
|
|
|
@return the tuple which should be inserted */
|
|
|
|
static
|
|
|
|
dtuple_t*
|
|
|
|
dict_create_sys_indexes_tuple(
|
|
|
|
/*==========================*/
|
|
|
|
const dict_index_t* index, /*!< in: index */
|
|
|
|
mem_heap_t* heap) /*!< in: memory heap from
|
|
|
|
which the memory for the built
|
|
|
|
tuple is allocated */
|
|
|
|
{
|
|
|
|
dict_table_t* sys_indexes;
|
|
|
|
dict_table_t* table;
|
|
|
|
dtuple_t* entry;
|
|
|
|
dfield_t* dfield;
|
|
|
|
byte* ptr;
|
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
ut_ad(index);
|
|
|
|
ut_ad(heap);
|
|
|
|
|
|
|
|
sys_indexes = dict_sys->sys_indexes;
|
|
|
|
|
|
|
|
table = dict_table_get_low(index->table_name);
|
|
|
|
|
|
|
|
entry = dtuple_create(heap, 7 + DATA_N_SYS_COLS);
|
|
|
|
|
|
|
|
dict_table_copy_types(entry, sys_indexes);
|
|
|
|
|
|
|
|
/* 0: TABLE_ID -----------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__TABLE_ID);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 8));
|
|
|
|
mach_write_to_8(ptr, table->id);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 8);
|
|
|
|
|
|
|
|
/* 1: ID ----------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__ID);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 8));
|
|
|
|
mach_write_to_8(ptr, index->id);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 8);
|
|
|
|
|
|
|
|
/* 2: DB_TRX_ID added later */
|
|
|
|
/* 3: DB_ROLL_PTR added later */
|
|
|
|
/* 4: NAME --------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__NAME);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, index->name, ut_strlen(index->name));
|
|
|
|
|
|
|
|
/* 5: N_FIELDS ----------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__N_FIELDS);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, index->n_fields);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 6: TYPE --------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__TYPE);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, index->type);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 7: SPACE --------------------------*/
|
|
|
|
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__SPACE);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, index->space);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 8: PAGE_NO --------------------------*/
|
|
|
|
|
|
|
|
dfield = dtuple_get_nth_field(
|
|
|
|
entry, DICT_COL__SYS_INDEXES__PAGE_NO);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
mach_write_to_4(ptr, FIL_NULL);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/*--------------------------------*/
|
|
|
|
|
|
|
|
return(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*****************************************************************//**
|
|
|
|
Based on an index object, this function builds the entry to be inserted
|
|
|
|
in the SYS_FIELDS system table.
|
|
|
|
@return the tuple which should be inserted */
|
|
|
|
static
|
|
|
|
dtuple_t*
|
|
|
|
dict_create_sys_fields_tuple(
|
|
|
|
/*=========================*/
|
|
|
|
const dict_index_t* index, /*!< in: index */
|
|
|
|
ulint fld_no, /*!< in: field number */
|
|
|
|
mem_heap_t* heap) /*!< in: memory heap from
|
|
|
|
which the memory for the built
|
|
|
|
tuple is allocated */
|
|
|
|
{
|
|
|
|
dict_table_t* sys_fields;
|
|
|
|
dtuple_t* entry;
|
|
|
|
dict_field_t* field;
|
|
|
|
dfield_t* dfield;
|
|
|
|
byte* ptr;
|
|
|
|
ibool index_contains_column_prefix_field = FALSE;
|
|
|
|
ulint j;
|
|
|
|
|
|
|
|
ut_ad(index);
|
|
|
|
ut_ad(heap);
|
|
|
|
|
|
|
|
for (j = 0; j < index->n_fields; j++) {
|
|
|
|
if (dict_index_get_nth_field(index, j)->prefix_len > 0) {
|
|
|
|
index_contains_column_prefix_field = TRUE;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
field = dict_index_get_nth_field(index, fld_no);
|
|
|
|
|
|
|
|
sys_fields = dict_sys->sys_fields;
|
|
|
|
|
|
|
|
entry = dtuple_create(heap, 3 + DATA_N_SYS_COLS);
|
|
|
|
|
|
|
|
dict_table_copy_types(entry, sys_fields);
|
|
|
|
|
|
|
|
/* 0: INDEX_ID -----------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_FIELDS__INDEX_ID);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 8));
|
|
|
|
mach_write_to_8(ptr, index->id);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 8);
|
|
|
|
|
|
|
|
/* 1: POS; FIELD NUMBER & PREFIX LENGTH -----------------------*/
|
|
|
|
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_FIELDS__POS);
|
|
|
|
|
|
|
|
ptr = static_cast<byte*>(mem_heap_alloc(heap, 4));
|
|
|
|
|
|
|
|
if (index_contains_column_prefix_field) {
|
|
|
|
/* If there are column prefix fields in the index, then
|
|
|
|
we store the number of the field to the 2 HIGH bytes
|
|
|
|
and the prefix length to the 2 low bytes, */
|
|
|
|
|
|
|
|
mach_write_to_4(ptr, (fld_no << 16) + field->prefix_len);
|
|
|
|
} else {
|
|
|
|
/* Else we store the number of the field to the 2 LOW bytes.
|
|
|
|
This is to keep the storage format compatible with
|
|
|
|
InnoDB versions < 4.0.14. */
|
|
|
|
|
|
|
|
mach_write_to_4(ptr, fld_no);
|
|
|
|
}
|
|
|
|
|
|
|
|
dfield_set_data(dfield, ptr, 4);
|
|
|
|
|
|
|
|
/* 2: DB_TRX_ID added later */
|
|
|
|
/* 3: DB_ROLL_PTR added later */
|
|
|
|
/* 4: COL_NAME -------------------------*/
|
|
|
|
dfield = dtuple_get_nth_field(entry, DICT_COL__SYS_FIELDS__COL_NAME);
|
|
|
|
|
|
|
|
dfield_set_data(dfield, field->name,
|
|
|
|
ut_strlen(field->name));
|
|
|
|
/*---------------------------------*/
|
|
|
|
|
|
|
|
return(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*****************************************************************//**
|
|
|
|
Creates the tuple with which the index entry is searched for writing the index
|
|
|
|
tree root page number, if such a tree is created.
|
|
|
|
@return the tuple for search */
|
|
|
|
static
|
|
|
|
dtuple_t*
|
|
|
|
dict_create_search_tuple(
|
|
|
|
/*=====================*/
|
|
|
|
const dtuple_t* tuple, /*!< in: the tuple inserted in the SYS_INDEXES
|
|
|
|
table */
|
|
|
|
mem_heap_t* heap) /*!< in: memory heap from which the memory for
|
|
|
|
the built tuple is allocated */
|
|
|
|
{
|
|
|
|
dtuple_t* search_tuple;
|
|
|
|
const dfield_t* field1;
|
|
|
|
dfield_t* field2;
|
|
|
|
|
|
|
|
ut_ad(tuple && heap);
|
|
|
|
|
|
|
|
search_tuple = dtuple_create(heap, 2);
|
|
|
|
|
|
|
|
field1 = dtuple_get_nth_field(tuple, 0);
|
|
|
|
field2 = dtuple_get_nth_field(search_tuple, 0);
|
|
|
|
|
|
|
|
dfield_copy(field2, field1);
|
|
|
|
|
|
|
|
field1 = dtuple_get_nth_field(tuple, 1);
|
|
|
|
field2 = dtuple_get_nth_field(search_tuple, 1);
|
|
|
|
|
|
|
|
dfield_copy(field2, field1);
|
|
|
|
|
|
|
|
ut_ad(dtuple_validate(search_tuple));
|
|
|
|
|
|
|
|
return(search_tuple);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***************************************************************//**
|
|
|
|
Builds an index definition row to insert.
|
|
|
|
@return DB_SUCCESS or error code */
|
2016-06-21 14:21:03 +02:00
|
|
|
static MY_ATTRIBUTE((nonnull, warn_unused_result))
|
2014-02-26 19:11:54 +01:00
|
|
|
dberr_t
|
|
|
|
dict_build_index_def_step(
|
|
|
|
/*======================*/
|
|
|
|
que_thr_t* thr, /*!< in: query thread */
|
|
|
|
ind_node_t* node) /*!< in: index create node */
|
|
|
|
{
|
|
|
|
dict_table_t* table;
|
|
|
|
dict_index_t* index;
|
|
|
|
dtuple_t* row;
|
|
|
|
trx_t* trx;
|
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
|
|
|
|
trx = thr_get_trx(thr);
|
|
|
|
|
|
|
|
index = node->index;
|
|
|
|
|
|
|
|
table = dict_table_get_low(index->table_name);
|
|
|
|
|
|
|
|
if (table == NULL) {
|
|
|
|
return(DB_TABLE_NOT_FOUND);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!trx->table_id) {
|
|
|
|
/* Record only the first table id. */
|
|
|
|
trx->table_id = table->id;
|
|
|
|
}
|
|
|
|
|
|
|
|
node->table = table;
|
|
|
|
|
|
|
|
ut_ad((UT_LIST_GET_LEN(table->indexes) > 0)
|
|
|
|
|| dict_index_is_clust(index));
|
|
|
|
|
|
|
|
dict_hdr_get_new_id(NULL, &index->id, NULL);
|
|
|
|
|
|
|
|
/* Inherit the space id from the table; we store all indexes of a
|
|
|
|
table in the same tablespace */
|
|
|
|
|
|
|
|
index->space = table->space;
|
|
|
|
node->page_no = FIL_NULL;
|
|
|
|
row = dict_create_sys_indexes_tuple(index, node->heap);
|
|
|
|
node->ind_row = row;
|
|
|
|
|
|
|
|
ins_node_set_new_row(node->ind_def, row);
|
|
|
|
|
|
|
|
/* Note that the index was created by this transaction. */
|
|
|
|
index->trx_id = trx->id;
|
|
|
|
ut_ad(table->def_trx_id <= trx->id);
|
|
|
|
table->def_trx_id = trx->id;
|
|
|
|
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***************************************************************//**
|
|
|
|
Builds a field definition row to insert. */
|
|
|
|
static
|
|
|
|
void
|
|
|
|
dict_build_field_def_step(
|
|
|
|
/*======================*/
|
|
|
|
ind_node_t* node) /*!< in: index create node */
|
|
|
|
{
|
|
|
|
dict_index_t* index;
|
|
|
|
dtuple_t* row;
|
|
|
|
|
|
|
|
index = node->index;
|
|
|
|
|
|
|
|
row = dict_create_sys_fields_tuple(index, node->field_no, node->heap);
|
|
|
|
|
|
|
|
ins_node_set_new_row(node->field_def, row);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***************************************************************//**
|
|
|
|
Creates an index tree for the index if it is not a member of a cluster.
|
|
|
|
@return DB_SUCCESS or DB_OUT_OF_FILE_SPACE */
|
2016-06-21 14:21:03 +02:00
|
|
|
static MY_ATTRIBUTE((nonnull, warn_unused_result))
|
2014-02-26 19:11:54 +01:00
|
|
|
dberr_t
|
|
|
|
dict_create_index_tree_step(
|
|
|
|
/*========================*/
|
|
|
|
ind_node_t* node) /*!< in: index create node */
|
|
|
|
{
|
|
|
|
dict_index_t* index;
|
|
|
|
dict_table_t* sys_indexes;
|
|
|
|
dtuple_t* search_tuple;
|
|
|
|
btr_pcur_t pcur;
|
|
|
|
mtr_t mtr;
|
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
|
|
|
|
index = node->index;
|
|
|
|
|
|
|
|
sys_indexes = dict_sys->sys_indexes;
|
|
|
|
|
|
|
|
if (index->type == DICT_FTS) {
|
|
|
|
/* FTS index does not need an index tree */
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Run a mini-transaction in which the index tree is allocated for
|
|
|
|
the index and its root address is written to the index entry in
|
|
|
|
sys_indexes */
|
|
|
|
|
|
|
|
mtr_start(&mtr);
|
|
|
|
|
|
|
|
search_tuple = dict_create_search_tuple(node->ind_row, node->heap);
|
|
|
|
|
|
|
|
btr_pcur_open(UT_LIST_GET_FIRST(sys_indexes->indexes),
|
|
|
|
search_tuple, PAGE_CUR_L, BTR_MODIFY_LEAF,
|
|
|
|
&pcur, &mtr);
|
|
|
|
|
|
|
|
btr_pcur_move_to_next_user_rec(&pcur, &mtr);
|
|
|
|
|
|
|
|
|
|
|
|
dberr_t err = DB_SUCCESS;
|
|
|
|
ulint zip_size = dict_table_zip_size(index->table);
|
|
|
|
|
MDEV-12253: Buffer pool blocks are accessed after they have been freed
Problem was that bpage was referenced after it was already freed
from LRU. Fixed by adding a new variable encrypted that is
passed down to buf_page_check_corrupt() and used in
buf_page_get_gen() to stop processing page read.
This patch should also address following test failures and
bugs:
MDEV-12419: IMPORT should not look up tablespace in
PageConverter::validate(). This is now removed.
MDEV-10099: encryption.innodb_onlinealter_encryption fails
sporadically in buildbot
MDEV-11420: encryption.innodb_encryption-page-compression
failed in buildbot
MDEV-11222: encryption.encrypt_and_grep failed in buildbot on P8
Removed dict_table_t::is_encrypted and dict_table_t::ibd_file_missing
and replaced these with dict_table_t::file_unreadable. Table
ibd file is missing if fil_get_space(space_id) returns NULL
and encrypted if not. Removed dict_table_t::is_corrupted field.
Ported FilSpace class from 10.2 and using that on buf_page_check_corrupt(),
buf_page_decrypt_after_read(), buf_page_encrypt_before_write(),
buf_dblwr_process(), buf_read_page(), dict_stats_save_defrag_stats().
Added test cases when enrypted page could be read while doing
redo log crash recovery. Also added test case for row compressed
blobs.
btr_cur_open_at_index_side_func(),
btr_cur_open_at_rnd_pos_func(): Avoid referencing block that is
NULL.
buf_page_get_zip(): Issue error if page read fails.
buf_page_get_gen(): Use dberr_t for error detection and
do not reference bpage after we hare freed it.
buf_mark_space_corrupt(): remove bpage from LRU also when
it is encrypted.
buf_page_check_corrupt(): @return DB_SUCCESS if page has
been read and is not corrupted,
DB_PAGE_CORRUPTED if page based on checksum check is corrupted,
DB_DECRYPTION_FAILED if page post encryption checksum matches but
after decryption normal page checksum does not match. In read
case only DB_SUCCESS is possible.
buf_page_io_complete(): use dberr_t for error handling.
buf_flush_write_block_low(),
buf_read_ahead_random(),
buf_read_page_async(),
buf_read_ahead_linear(),
buf_read_ibuf_merge_pages(),
buf_read_recv_pages(),
fil_aio_wait():
Issue error if page read fails.
btr_pcur_move_to_next_page(): Do not reference page if it is
NULL.
Introduced dict_table_t::is_readable() and dict_index_t::is_readable()
that will return true if tablespace exists and pages read from
tablespace are not corrupted or page decryption failed.
Removed buf_page_t::key_version. After page decryption the
key version is not removed from page frame. For unencrypted
pages, old key_version is removed at buf_page_encrypt_before_write()
dict_stats_update_transient_for_index(),
dict_stats_update_transient()
Do not continue if table decryption failed or table
is corrupted.
dict0stats.cc: Introduced a dict_stats_report_error function
to avoid code duplication.
fil_parse_write_crypt_data():
Check that key read from redo log entry is found from
encryption plugin and if it is not, refuse to start.
PageConverter::validate(): Removed access to fil_space_t as
tablespace is not available during import.
Fixed error code on innodb.innodb test.
Merged test cased innodb-bad-key-change5 and innodb-bad-key-shutdown
to innodb-bad-key-change2. Removed innodb-bad-key-change5 test.
Decreased unnecessary complexity on some long lasting tests.
Removed fil_inc_pending_ops(), fil_decr_pending_ops(),
fil_get_first_space(), fil_get_next_space(),
fil_get_first_space_safe(), fil_get_next_space_safe()
functions.
fil_space_verify_crypt_checksum(): Fixed bug found using ASAN
where FIL_PAGE_END_LSN_OLD_CHECKSUM field was incorrectly
accessed from row compressed tables. Fixed out of page frame
bug for row compressed tables in
fil_space_verify_crypt_checksum() found using ASAN. Incorrect
function was called for compressed table.
Added new tests for discard, rename table and drop (we should allow them
even when page decryption fails). Alter table rename is not allowed.
Added test for restart with innodb-force-recovery=1 when page read on
redo-recovery cant be decrypted. Added test for corrupted table where
both page data and FIL_PAGE_FILE_FLUSH_LSN_OR_KEY_VERSION is corrupted.
Adjusted the test case innodb_bug14147491 so that it does not anymore
expect crash. Instead table is just mostly not usable.
fil0fil.h: fil_space_acquire_low is not visible function
and fil_space_acquire and fil_space_acquire_silent are
inline functions. FilSpace class uses fil_space_acquire_low
directly.
recv_apply_hashed_log_recs() does not return anything.
2017-04-26 15:19:16 +03:00
|
|
|
if (node->index->table->file_unreadable
|
2014-02-26 19:11:54 +01:00
|
|
|
|| dict_table_is_discarded(node->index->table)) {
|
|
|
|
|
|
|
|
node->page_no = FIL_NULL;
|
|
|
|
} else {
|
|
|
|
node->page_no = btr_create(
|
|
|
|
index->type, index->space, zip_size,
|
|
|
|
index->id, index, &mtr);
|
|
|
|
|
|
|
|
if (node->page_no == FIL_NULL) {
|
|
|
|
err = DB_OUT_OF_FILE_SPACE;
|
|
|
|
}
|
|
|
|
|
|
|
|
DBUG_EXECUTE_IF("ib_import_create_index_failure_1",
|
|
|
|
node->page_no = FIL_NULL;
|
|
|
|
err = DB_OUT_OF_FILE_SPACE; );
|
|
|
|
}
|
|
|
|
|
|
|
|
page_rec_write_field(
|
|
|
|
btr_pcur_get_rec(&pcur), DICT_FLD__SYS_INDEXES__PAGE_NO,
|
|
|
|
node->page_no, &mtr);
|
|
|
|
|
|
|
|
btr_pcur_close(&pcur);
|
|
|
|
|
|
|
|
mtr_commit(&mtr);
|
|
|
|
|
|
|
|
return(err);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*******************************************************************//**
|
|
|
|
Drops the index tree associated with a row in SYS_INDEXES table. */
|
|
|
|
UNIV_INTERN
|
|
|
|
void
|
|
|
|
dict_drop_index_tree(
|
|
|
|
/*=================*/
|
|
|
|
rec_t* rec, /*!< in/out: record in the clustered index
|
|
|
|
of SYS_INDEXES table */
|
|
|
|
mtr_t* mtr) /*!< in: mtr having the latch on the record page */
|
|
|
|
{
|
|
|
|
ulint root_page_no;
|
|
|
|
ulint space;
|
|
|
|
ulint zip_size;
|
|
|
|
const byte* ptr;
|
|
|
|
ulint len;
|
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
ut_a(!dict_table_is_comp(dict_sys->sys_indexes));
|
|
|
|
ptr = rec_get_nth_field_old(
|
|
|
|
rec, DICT_FLD__SYS_INDEXES__PAGE_NO, &len);
|
|
|
|
|
|
|
|
ut_ad(len == 4);
|
|
|
|
|
|
|
|
root_page_no = mtr_read_ulint(ptr, MLOG_4BYTES, mtr);
|
|
|
|
|
|
|
|
if (root_page_no == FIL_NULL) {
|
|
|
|
/* The tree has already been freed */
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
ptr = rec_get_nth_field_old(
|
|
|
|
rec, DICT_FLD__SYS_INDEXES__SPACE, &len);
|
|
|
|
|
|
|
|
ut_ad(len == 4);
|
|
|
|
|
|
|
|
space = mtr_read_ulint(ptr, MLOG_4BYTES, mtr);
|
|
|
|
zip_size = fil_space_get_zip_size(space);
|
|
|
|
|
|
|
|
if (UNIV_UNLIKELY(zip_size == ULINT_UNDEFINED)) {
|
|
|
|
/* It is a single table tablespace and the .ibd file is
|
|
|
|
missing: do nothing */
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* We free all the pages but the root page first; this operation
|
|
|
|
may span several mini-transactions */
|
|
|
|
|
|
|
|
btr_free_but_not_root(space, zip_size, root_page_no);
|
|
|
|
|
|
|
|
/* Then we free the root page in the same mini-transaction where
|
|
|
|
we write FIL_NULL to the appropriate field in the SYS_INDEXES
|
|
|
|
record: this mini-transaction marks the B-tree totally freed */
|
|
|
|
|
|
|
|
/* printf("Dropping index tree in space %lu root page %lu\n", space,
|
|
|
|
root_page_no); */
|
|
|
|
btr_free_root(space, zip_size, root_page_no, mtr);
|
|
|
|
|
|
|
|
page_rec_write_field(rec, DICT_FLD__SYS_INDEXES__PAGE_NO,
|
|
|
|
FIL_NULL, mtr);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*******************************************************************//**
|
|
|
|
Truncates the index tree associated with a row in SYS_INDEXES table.
|
|
|
|
@return new root page number, or FIL_NULL on failure */
|
|
|
|
UNIV_INTERN
|
|
|
|
ulint
|
|
|
|
dict_truncate_index_tree(
|
|
|
|
/*=====================*/
|
|
|
|
dict_table_t* table, /*!< in: the table the index belongs to */
|
|
|
|
ulint space, /*!< in: 0=truncate,
|
|
|
|
nonzero=create the index tree in the
|
|
|
|
given tablespace */
|
|
|
|
btr_pcur_t* pcur, /*!< in/out: persistent cursor pointing to
|
|
|
|
record in the clustered index of
|
|
|
|
SYS_INDEXES table. The cursor may be
|
|
|
|
repositioned in this call. */
|
|
|
|
mtr_t* mtr) /*!< in: mtr having the latch
|
|
|
|
on the record page. The mtr may be
|
|
|
|
committed and restarted in this call. */
|
|
|
|
{
|
|
|
|
ulint root_page_no;
|
|
|
|
ibool drop = !space;
|
|
|
|
ulint zip_size;
|
|
|
|
ulint type;
|
|
|
|
index_id_t index_id;
|
|
|
|
rec_t* rec;
|
|
|
|
const byte* ptr;
|
|
|
|
ulint len;
|
|
|
|
dict_index_t* index;
|
2014-02-26 19:23:04 +01:00
|
|
|
bool has_been_dropped = false;
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
ut_a(!dict_table_is_comp(dict_sys->sys_indexes));
|
|
|
|
rec = btr_pcur_get_rec(pcur);
|
|
|
|
ptr = rec_get_nth_field_old(
|
|
|
|
rec, DICT_FLD__SYS_INDEXES__PAGE_NO, &len);
|
|
|
|
|
|
|
|
ut_ad(len == 4);
|
|
|
|
|
|
|
|
root_page_no = mtr_read_ulint(ptr, MLOG_4BYTES, mtr);
|
|
|
|
|
|
|
|
if (drop && root_page_no == FIL_NULL) {
|
2014-02-26 19:23:04 +01:00
|
|
|
has_been_dropped = true;
|
2014-02-26 19:11:54 +01:00
|
|
|
drop = FALSE;
|
|
|
|
}
|
|
|
|
|
|
|
|
ptr = rec_get_nth_field_old(
|
|
|
|
rec, DICT_FLD__SYS_INDEXES__SPACE, &len);
|
|
|
|
|
|
|
|
ut_ad(len == 4);
|
|
|
|
|
|
|
|
if (drop) {
|
|
|
|
space = mtr_read_ulint(ptr, MLOG_4BYTES, mtr);
|
|
|
|
}
|
|
|
|
|
|
|
|
zip_size = fil_space_get_zip_size(space);
|
|
|
|
|
|
|
|
if (UNIV_UNLIKELY(zip_size == ULINT_UNDEFINED)) {
|
|
|
|
/* It is a single table tablespace and the .ibd file is
|
|
|
|
missing: do nothing */
|
|
|
|
|
|
|
|
ut_print_timestamp(stderr);
|
|
|
|
fprintf(stderr, " InnoDB: Trying to TRUNCATE"
|
|
|
|
" a missing .ibd file of table %s!\n", table->name);
|
|
|
|
return(FIL_NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
ptr = rec_get_nth_field_old(
|
|
|
|
rec, DICT_FLD__SYS_INDEXES__TYPE, &len);
|
|
|
|
ut_ad(len == 4);
|
|
|
|
type = mach_read_from_4(ptr);
|
|
|
|
|
|
|
|
ptr = rec_get_nth_field_old(rec, DICT_FLD__SYS_INDEXES__ID, &len);
|
|
|
|
ut_ad(len == 8);
|
|
|
|
index_id = mach_read_from_8(ptr);
|
|
|
|
|
|
|
|
if (!drop) {
|
|
|
|
|
|
|
|
goto create;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* We free all the pages but the root page first; this operation
|
|
|
|
may span several mini-transactions */
|
|
|
|
|
|
|
|
btr_free_but_not_root(space, zip_size, root_page_no);
|
|
|
|
|
|
|
|
/* Then we free the root page in the same mini-transaction where
|
|
|
|
we create the b-tree and write its new root page number to the
|
|
|
|
appropriate field in the SYS_INDEXES record: this mini-transaction
|
|
|
|
marks the B-tree totally truncated */
|
|
|
|
|
|
|
|
btr_block_get(space, zip_size, root_page_no, RW_X_LATCH, NULL, mtr);
|
|
|
|
|
|
|
|
btr_free_root(space, zip_size, root_page_no, mtr);
|
|
|
|
create:
|
|
|
|
/* We will temporarily write FIL_NULL to the PAGE_NO field
|
|
|
|
in SYS_INDEXES, so that the database will not get into an
|
|
|
|
inconsistent state in case it crashes between the mtr_commit()
|
|
|
|
below and the following mtr_commit() call. */
|
|
|
|
page_rec_write_field(rec, DICT_FLD__SYS_INDEXES__PAGE_NO,
|
|
|
|
FIL_NULL, mtr);
|
|
|
|
|
|
|
|
/* We will need to commit the mini-transaction in order to avoid
|
|
|
|
deadlocks in the btr_create() call, because otherwise we would
|
|
|
|
be freeing and allocating pages in the same mini-transaction. */
|
|
|
|
btr_pcur_store_position(pcur, mtr);
|
|
|
|
mtr_commit(mtr);
|
|
|
|
|
|
|
|
mtr_start(mtr);
|
|
|
|
btr_pcur_restore_position(BTR_MODIFY_LEAF, pcur, mtr);
|
|
|
|
|
|
|
|
/* Find the index corresponding to this SYS_INDEXES record. */
|
|
|
|
for (index = UT_LIST_GET_FIRST(table->indexes);
|
|
|
|
index;
|
|
|
|
index = UT_LIST_GET_NEXT(indexes, index)) {
|
|
|
|
if (index->id == index_id) {
|
|
|
|
if (index->type & DICT_FTS) {
|
|
|
|
return(FIL_NULL);
|
|
|
|
} else {
|
2014-02-26 19:23:04 +01:00
|
|
|
if (has_been_dropped) {
|
|
|
|
fprintf(stderr, " InnoDB: Trying to"
|
|
|
|
" TRUNCATE a missing index of"
|
|
|
|
" table %s!\n",
|
|
|
|
index->table->name);
|
|
|
|
}
|
|
|
|
|
2014-02-26 19:11:54 +01:00
|
|
|
root_page_no = btr_create(type, space, zip_size,
|
|
|
|
index_id, index, mtr);
|
|
|
|
index->page = (unsigned int) root_page_no;
|
|
|
|
return(root_page_no);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ut_print_timestamp(stderr);
|
|
|
|
fprintf(stderr,
|
|
|
|
" InnoDB: Index %llu of table %s is missing\n"
|
|
|
|
"InnoDB: from the data dictionary during TRUNCATE!\n",
|
|
|
|
(ullint) index_id,
|
|
|
|
table->name);
|
|
|
|
|
|
|
|
return(FIL_NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*********************************************************************//**
|
|
|
|
Creates a table create graph.
|
|
|
|
@return own: table create node */
|
|
|
|
UNIV_INTERN
|
|
|
|
tab_node_t*
|
|
|
|
tab_create_graph_create(
|
|
|
|
/*====================*/
|
|
|
|
dict_table_t* table, /*!< in: table to create, built as a memory data
|
|
|
|
structure */
|
|
|
|
mem_heap_t* heap, /*!< in: heap where created */
|
2015-08-14 11:09:06 +03:00
|
|
|
bool commit, /*!< in: true if the commit node should be
|
2014-02-26 19:11:54 +01:00
|
|
|
added to the query graph */
|
2015-08-14 11:09:06 +03:00
|
|
|
fil_encryption_t mode, /*!< in: encryption mode */
|
|
|
|
ulint key_id) /*!< in: encryption key_id */
|
2014-02-26 19:11:54 +01:00
|
|
|
{
|
|
|
|
tab_node_t* node;
|
|
|
|
|
|
|
|
node = static_cast<tab_node_t*>(
|
|
|
|
mem_heap_alloc(heap, sizeof(tab_node_t)));
|
|
|
|
|
|
|
|
node->common.type = QUE_NODE_CREATE_TABLE;
|
|
|
|
|
|
|
|
node->table = table;
|
|
|
|
|
|
|
|
node->state = TABLE_BUILD_TABLE_DEF;
|
|
|
|
node->heap = mem_heap_create(256);
|
2015-08-14 11:09:06 +03:00
|
|
|
node->mode = mode;
|
|
|
|
node->key_id = key_id;
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
node->tab_def = ins_node_create(INS_DIRECT, dict_sys->sys_tables,
|
|
|
|
heap);
|
|
|
|
node->tab_def->common.parent = node;
|
|
|
|
|
|
|
|
node->col_def = ins_node_create(INS_DIRECT, dict_sys->sys_columns,
|
|
|
|
heap);
|
|
|
|
node->col_def->common.parent = node;
|
|
|
|
|
|
|
|
if (commit) {
|
|
|
|
node->commit_node = trx_commit_node_create(heap);
|
|
|
|
node->commit_node->common.parent = node;
|
|
|
|
} else {
|
|
|
|
node->commit_node = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return(node);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*********************************************************************//**
|
|
|
|
Creates an index create graph.
|
|
|
|
@return own: index create node */
|
|
|
|
UNIV_INTERN
|
|
|
|
ind_node_t*
|
|
|
|
ind_create_graph_create(
|
|
|
|
/*====================*/
|
|
|
|
dict_index_t* index, /*!< in: index to create, built as a memory data
|
|
|
|
structure */
|
|
|
|
mem_heap_t* heap, /*!< in: heap where created */
|
|
|
|
bool commit) /*!< in: true if the commit node should be
|
|
|
|
added to the query graph */
|
|
|
|
{
|
|
|
|
ind_node_t* node;
|
|
|
|
|
|
|
|
node = static_cast<ind_node_t*>(
|
|
|
|
mem_heap_alloc(heap, sizeof(ind_node_t)));
|
|
|
|
|
|
|
|
node->common.type = QUE_NODE_CREATE_INDEX;
|
|
|
|
|
|
|
|
node->index = index;
|
|
|
|
|
|
|
|
node->state = INDEX_BUILD_INDEX_DEF;
|
|
|
|
node->page_no = FIL_NULL;
|
|
|
|
node->heap = mem_heap_create(256);
|
|
|
|
|
|
|
|
node->ind_def = ins_node_create(INS_DIRECT,
|
|
|
|
dict_sys->sys_indexes, heap);
|
|
|
|
node->ind_def->common.parent = node;
|
|
|
|
|
|
|
|
node->field_def = ins_node_create(INS_DIRECT,
|
|
|
|
dict_sys->sys_fields, heap);
|
|
|
|
node->field_def->common.parent = node;
|
|
|
|
|
|
|
|
if (commit) {
|
|
|
|
node->commit_node = trx_commit_node_create(heap);
|
|
|
|
node->commit_node->common.parent = node;
|
|
|
|
} else {
|
|
|
|
node->commit_node = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return(node);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***********************************************************//**
|
|
|
|
Creates a table. This is a high-level function used in SQL execution graphs.
|
|
|
|
@return query thread to run next or NULL */
|
|
|
|
UNIV_INTERN
|
|
|
|
que_thr_t*
|
|
|
|
dict_create_table_step(
|
|
|
|
/*===================*/
|
|
|
|
que_thr_t* thr) /*!< in: query thread */
|
|
|
|
{
|
|
|
|
tab_node_t* node;
|
|
|
|
dberr_t err = DB_ERROR;
|
|
|
|
trx_t* trx;
|
|
|
|
|
|
|
|
ut_ad(thr);
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
|
|
|
|
trx = thr_get_trx(thr);
|
|
|
|
|
|
|
|
node = static_cast<tab_node_t*>(thr->run_node);
|
|
|
|
|
|
|
|
ut_ad(que_node_get_type(node) == QUE_NODE_CREATE_TABLE);
|
|
|
|
|
|
|
|
if (thr->prev_node == que_node_get_parent(node)) {
|
|
|
|
node->state = TABLE_BUILD_TABLE_DEF;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == TABLE_BUILD_TABLE_DEF) {
|
|
|
|
|
|
|
|
/* DO THE CHECKS OF THE CONSISTENCY CONSTRAINTS HERE */
|
|
|
|
|
|
|
|
err = dict_build_table_def_step(thr, node);
|
|
|
|
if (err != DB_SUCCESS) {
|
|
|
|
|
|
|
|
goto function_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
node->state = TABLE_BUILD_COL_DEF;
|
|
|
|
node->col_no = 0;
|
|
|
|
|
|
|
|
thr->run_node = node->tab_def;
|
|
|
|
|
|
|
|
return(thr);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == TABLE_BUILD_COL_DEF) {
|
|
|
|
|
|
|
|
if (node->col_no < (node->table)->n_def) {
|
|
|
|
|
|
|
|
dict_build_col_def_step(node);
|
|
|
|
|
|
|
|
node->col_no++;
|
|
|
|
|
|
|
|
thr->run_node = node->col_def;
|
|
|
|
|
|
|
|
return(thr);
|
|
|
|
} else {
|
|
|
|
node->state = TABLE_COMMIT_WORK;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == TABLE_COMMIT_WORK) {
|
|
|
|
|
|
|
|
/* Table was correctly defined: do NOT commit the transaction
|
|
|
|
(CREATE TABLE does NOT do an implicit commit of the current
|
|
|
|
transaction) */
|
|
|
|
|
|
|
|
node->state = TABLE_ADD_TO_CACHE;
|
|
|
|
|
|
|
|
/* thr->run_node = node->commit_node;
|
|
|
|
|
|
|
|
return(thr); */
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == TABLE_ADD_TO_CACHE) {
|
|
|
|
|
|
|
|
dict_table_add_to_cache(node->table, TRUE, node->heap);
|
|
|
|
|
|
|
|
err = DB_SUCCESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
function_exit:
|
|
|
|
trx->error_state = err;
|
|
|
|
|
|
|
|
if (err == DB_SUCCESS) {
|
|
|
|
/* Ok: do nothing */
|
|
|
|
|
|
|
|
} else if (err == DB_LOCK_WAIT) {
|
|
|
|
|
|
|
|
return(NULL);
|
|
|
|
} else {
|
|
|
|
/* SQL error detected */
|
|
|
|
|
|
|
|
return(NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
thr->run_node = que_node_get_parent(node);
|
|
|
|
|
|
|
|
return(thr);
|
|
|
|
}
|
|
|
|
|
|
|
|
/***********************************************************//**
|
|
|
|
Creates an index. This is a high-level function used in SQL execution
|
|
|
|
graphs.
|
|
|
|
@return query thread to run next or NULL */
|
|
|
|
UNIV_INTERN
|
|
|
|
que_thr_t*
|
|
|
|
dict_create_index_step(
|
|
|
|
/*===================*/
|
|
|
|
que_thr_t* thr) /*!< in: query thread */
|
|
|
|
{
|
|
|
|
ind_node_t* node;
|
|
|
|
dberr_t err = DB_ERROR;
|
|
|
|
trx_t* trx;
|
|
|
|
|
|
|
|
ut_ad(thr);
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
|
|
|
|
trx = thr_get_trx(thr);
|
|
|
|
|
|
|
|
node = static_cast<ind_node_t*>(thr->run_node);
|
|
|
|
|
|
|
|
ut_ad(que_node_get_type(node) == QUE_NODE_CREATE_INDEX);
|
|
|
|
|
|
|
|
if (thr->prev_node == que_node_get_parent(node)) {
|
|
|
|
node->state = INDEX_BUILD_INDEX_DEF;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == INDEX_BUILD_INDEX_DEF) {
|
|
|
|
/* DO THE CHECKS OF THE CONSISTENCY CONSTRAINTS HERE */
|
|
|
|
err = dict_build_index_def_step(thr, node);
|
|
|
|
|
|
|
|
if (err != DB_SUCCESS) {
|
|
|
|
|
|
|
|
goto function_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
node->state = INDEX_BUILD_FIELD_DEF;
|
|
|
|
node->field_no = 0;
|
|
|
|
|
|
|
|
thr->run_node = node->ind_def;
|
|
|
|
|
|
|
|
return(thr);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == INDEX_BUILD_FIELD_DEF) {
|
|
|
|
|
|
|
|
if (node->field_no < (node->index)->n_fields) {
|
|
|
|
|
|
|
|
dict_build_field_def_step(node);
|
|
|
|
|
|
|
|
node->field_no++;
|
|
|
|
|
|
|
|
thr->run_node = node->field_def;
|
|
|
|
|
|
|
|
return(thr);
|
|
|
|
} else {
|
|
|
|
node->state = INDEX_ADD_TO_CACHE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == INDEX_ADD_TO_CACHE) {
|
|
|
|
|
|
|
|
index_id_t index_id = node->index->id;
|
|
|
|
|
|
|
|
err = dict_index_add_to_cache(
|
|
|
|
node->table, node->index, FIL_NULL,
|
|
|
|
trx_is_strict(trx)
|
|
|
|
|| dict_table_get_format(node->table)
|
|
|
|
>= UNIV_FORMAT_B);
|
|
|
|
|
|
|
|
node->index = dict_index_get_if_in_cache_low(index_id);
|
2016-04-26 23:05:26 +02:00
|
|
|
ut_a((node->index == NULL) == (err != DB_SUCCESS));
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
if (err != DB_SUCCESS) {
|
|
|
|
|
|
|
|
goto function_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
node->state = INDEX_CREATE_INDEX_TREE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == INDEX_CREATE_INDEX_TREE) {
|
|
|
|
|
|
|
|
err = dict_create_index_tree_step(node);
|
|
|
|
|
|
|
|
DBUG_EXECUTE_IF("ib_dict_create_index_tree_fail",
|
|
|
|
err = DB_OUT_OF_MEMORY;);
|
|
|
|
|
|
|
|
if (err != DB_SUCCESS) {
|
|
|
|
/* If this is a FTS index, we will need to remove
|
|
|
|
it from fts->cache->indexes list as well */
|
|
|
|
if ((node->index->type & DICT_FTS)
|
|
|
|
&& node->table->fts) {
|
|
|
|
fts_index_cache_t* index_cache;
|
|
|
|
|
|
|
|
rw_lock_x_lock(
|
|
|
|
&node->table->fts->cache->init_lock);
|
|
|
|
|
|
|
|
index_cache = (fts_index_cache_t*)
|
|
|
|
fts_find_index_cache(
|
|
|
|
node->table->fts->cache,
|
|
|
|
node->index);
|
|
|
|
|
|
|
|
if (index_cache->words) {
|
|
|
|
rbt_free(index_cache->words);
|
|
|
|
index_cache->words = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
ib_vector_remove(
|
|
|
|
node->table->fts->cache->indexes,
|
|
|
|
*reinterpret_cast<void**>(index_cache));
|
|
|
|
|
|
|
|
rw_lock_x_unlock(
|
|
|
|
&node->table->fts->cache->init_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
dict_index_remove_from_cache(node->table, node->index);
|
|
|
|
node->index = NULL;
|
|
|
|
|
|
|
|
goto function_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
node->index->page = node->page_no;
|
|
|
|
/* These should have been set in
|
|
|
|
dict_build_index_def_step() and
|
|
|
|
dict_index_add_to_cache(). */
|
|
|
|
ut_ad(node->index->trx_id == trx->id);
|
|
|
|
ut_ad(node->index->table->def_trx_id == trx->id);
|
|
|
|
node->state = INDEX_COMMIT_WORK;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (node->state == INDEX_COMMIT_WORK) {
|
|
|
|
|
|
|
|
/* Index was correctly defined: do NOT commit the transaction
|
|
|
|
(CREATE INDEX does NOT currently do an implicit commit of
|
|
|
|
the current transaction) */
|
|
|
|
|
|
|
|
node->state = INDEX_CREATE_INDEX_TREE;
|
|
|
|
|
|
|
|
/* thr->run_node = node->commit_node;
|
|
|
|
|
|
|
|
return(thr); */
|
|
|
|
}
|
|
|
|
|
|
|
|
function_exit:
|
|
|
|
trx->error_state = err;
|
|
|
|
|
|
|
|
if (err == DB_SUCCESS) {
|
|
|
|
/* Ok: do nothing */
|
|
|
|
|
|
|
|
} else if (err == DB_LOCK_WAIT) {
|
|
|
|
|
|
|
|
return(NULL);
|
|
|
|
} else {
|
|
|
|
/* SQL error detected */
|
|
|
|
|
|
|
|
return(NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
thr->run_node = que_node_get_parent(node);
|
|
|
|
|
|
|
|
return(thr);
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************//**
|
|
|
|
Check whether a system table exists. Additionally, if it exists,
|
|
|
|
move it to the non-LRU end of the table LRU list. This is oly used
|
|
|
|
for system tables that can be upgraded or added to an older database,
|
|
|
|
which include SYS_FOREIGN, SYS_FOREIGN_COLS, SYS_TABLESPACES and
|
|
|
|
SYS_DATAFILES.
|
|
|
|
@return DB_SUCCESS if the sys table exists, DB_CORRUPTION if it exists
|
|
|
|
but is not current, DB_TABLE_NOT_FOUND if it does not exist*/
|
|
|
|
static
|
|
|
|
dberr_t
|
|
|
|
dict_check_if_system_table_exists(
|
|
|
|
/*==============================*/
|
|
|
|
const char* tablename, /*!< in: name of table */
|
|
|
|
ulint num_fields, /*!< in: number of fields */
|
|
|
|
ulint num_indexes) /*!< in: number of indexes */
|
|
|
|
{
|
|
|
|
dict_table_t* sys_table;
|
|
|
|
dberr_t error = DB_SUCCESS;
|
|
|
|
|
|
|
|
ut_a(srv_get_active_thread_type() == SRV_NONE);
|
|
|
|
|
|
|
|
mutex_enter(&dict_sys->mutex);
|
|
|
|
|
|
|
|
sys_table = dict_table_get_low(tablename);
|
|
|
|
|
|
|
|
if (sys_table == NULL) {
|
|
|
|
error = DB_TABLE_NOT_FOUND;
|
|
|
|
|
|
|
|
} else if (UT_LIST_GET_LEN(sys_table->indexes) != num_indexes
|
|
|
|
|| sys_table->n_cols != num_fields) {
|
|
|
|
error = DB_CORRUPTION;
|
|
|
|
|
|
|
|
} else {
|
|
|
|
/* This table has already been created, and it is OK.
|
|
|
|
Ensure that it can't be evicted from the table LRU cache. */
|
|
|
|
|
|
|
|
dict_table_move_from_lru_to_non_lru(sys_table);
|
|
|
|
}
|
|
|
|
|
|
|
|
mutex_exit(&dict_sys->mutex);
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************//**
|
|
|
|
Creates the foreign key constraints system tables inside InnoDB
|
|
|
|
at server bootstrap or server start if they are not found or are
|
|
|
|
not of the right form.
|
|
|
|
@return DB_SUCCESS or error code */
|
|
|
|
UNIV_INTERN
|
|
|
|
dberr_t
|
|
|
|
dict_create_or_check_foreign_constraint_tables(void)
|
|
|
|
/*================================================*/
|
|
|
|
{
|
|
|
|
trx_t* trx;
|
|
|
|
my_bool srv_file_per_table_backup;
|
|
|
|
dberr_t err;
|
|
|
|
dberr_t sys_foreign_err;
|
|
|
|
dberr_t sys_foreign_cols_err;
|
|
|
|
|
|
|
|
ut_a(srv_get_active_thread_type() == SRV_NONE);
|
|
|
|
|
|
|
|
/* Note: The master thread has not been started at this point. */
|
|
|
|
|
|
|
|
|
|
|
|
sys_foreign_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_FOREIGN", DICT_NUM_FIELDS__SYS_FOREIGN + 1, 3);
|
|
|
|
sys_foreign_cols_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_FOREIGN_COLS", DICT_NUM_FIELDS__SYS_FOREIGN_COLS + 1, 1);
|
|
|
|
|
|
|
|
if (sys_foreign_err == DB_SUCCESS
|
|
|
|
&& sys_foreign_cols_err == DB_SUCCESS) {
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
trx = trx_allocate_for_mysql();
|
|
|
|
|
|
|
|
trx_set_dict_operation(trx, TRX_DICT_OP_TABLE);
|
|
|
|
|
|
|
|
trx->op_info = "creating foreign key sys tables";
|
|
|
|
|
|
|
|
row_mysql_lock_data_dictionary(trx);
|
|
|
|
|
|
|
|
/* Check which incomplete table definition to drop. */
|
|
|
|
|
|
|
|
if (sys_foreign_err == DB_CORRUPTION) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_WARN,
|
|
|
|
"Dropping incompletely created "
|
|
|
|
"SYS_FOREIGN table.");
|
2016-06-23 07:42:40 +03:00
|
|
|
row_drop_table_for_mysql("SYS_FOREIGN", trx, TRUE, TRUE);
|
2014-02-26 19:11:54 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
if (sys_foreign_cols_err == DB_CORRUPTION) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_WARN,
|
|
|
|
"Dropping incompletely created "
|
|
|
|
"SYS_FOREIGN_COLS table.");
|
|
|
|
|
2016-06-23 07:42:40 +03:00
|
|
|
row_drop_table_for_mysql("SYS_FOREIGN_COLS", trx, TRUE, TRUE);
|
2014-02-26 19:11:54 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
ib_logf(IB_LOG_LEVEL_WARN,
|
|
|
|
"Creating foreign key constraint system tables.");
|
|
|
|
|
|
|
|
/* NOTE: in dict_load_foreigns we use the fact that
|
|
|
|
there are 2 secondary indexes on SYS_FOREIGN, and they
|
|
|
|
are defined just like below */
|
|
|
|
|
|
|
|
/* NOTE: when designing InnoDB's foreign key support in 2001, we made
|
|
|
|
an error and made the table names and the foreign key id of type
|
|
|
|
'CHAR' (internally, really a VARCHAR). We should have made the type
|
|
|
|
VARBINARY, like in other InnoDB system tables, to get a clean
|
|
|
|
design. */
|
|
|
|
|
|
|
|
srv_file_per_table_backup = srv_file_per_table;
|
|
|
|
|
|
|
|
/* We always want SYSTEM tables to be created inside the system
|
|
|
|
tablespace. */
|
|
|
|
|
|
|
|
srv_file_per_table = 0;
|
|
|
|
|
|
|
|
err = que_eval_sql(
|
|
|
|
NULL,
|
|
|
|
"PROCEDURE CREATE_FOREIGN_SYS_TABLES_PROC () IS\n"
|
|
|
|
"BEGIN\n"
|
|
|
|
"CREATE TABLE\n"
|
|
|
|
"SYS_FOREIGN(ID CHAR, FOR_NAME CHAR,"
|
|
|
|
" REF_NAME CHAR, N_COLS INT);\n"
|
|
|
|
"CREATE UNIQUE CLUSTERED INDEX ID_IND"
|
|
|
|
" ON SYS_FOREIGN (ID);\n"
|
|
|
|
"CREATE INDEX FOR_IND"
|
|
|
|
" ON SYS_FOREIGN (FOR_NAME);\n"
|
|
|
|
"CREATE INDEX REF_IND"
|
|
|
|
" ON SYS_FOREIGN (REF_NAME);\n"
|
|
|
|
"CREATE TABLE\n"
|
|
|
|
"SYS_FOREIGN_COLS(ID CHAR, POS INT,"
|
|
|
|
" FOR_COL_NAME CHAR, REF_COL_NAME CHAR);\n"
|
|
|
|
"CREATE UNIQUE CLUSTERED INDEX ID_IND"
|
|
|
|
" ON SYS_FOREIGN_COLS (ID, POS);\n"
|
|
|
|
"END;\n",
|
|
|
|
FALSE, trx);
|
|
|
|
|
|
|
|
if (err != DB_SUCCESS) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_ERROR,
|
|
|
|
"Creation of SYS_FOREIGN and SYS_FOREIGN_COLS "
|
|
|
|
"has failed with error %lu. Tablespace is full. "
|
|
|
|
"Dropping incompletely created tables.",
|
|
|
|
(ulong) err);
|
|
|
|
|
|
|
|
ut_ad(err == DB_OUT_OF_FILE_SPACE
|
|
|
|
|| err == DB_TOO_MANY_CONCURRENT_TRXS);
|
|
|
|
|
2016-06-23 07:42:40 +03:00
|
|
|
row_drop_table_for_mysql("SYS_FOREIGN", trx, TRUE, TRUE);
|
|
|
|
row_drop_table_for_mysql("SYS_FOREIGN_COLS", trx, TRUE, TRUE);
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
if (err == DB_OUT_OF_FILE_SPACE) {
|
|
|
|
err = DB_MUST_GET_MORE_FILE_SPACE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
trx_commit_for_mysql(trx);
|
|
|
|
|
|
|
|
row_mysql_unlock_data_dictionary(trx);
|
|
|
|
|
|
|
|
trx_free_for_mysql(trx);
|
|
|
|
|
|
|
|
srv_file_per_table = srv_file_per_table_backup;
|
|
|
|
|
|
|
|
if (err == DB_SUCCESS) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_INFO,
|
|
|
|
"Foreign key constraint system tables created");
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Note: The master thread has not been started at this point. */
|
|
|
|
/* Confirm and move to the non-LRU part of the table LRU list. */
|
|
|
|
sys_foreign_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_FOREIGN", DICT_NUM_FIELDS__SYS_FOREIGN + 1, 3);
|
|
|
|
ut_a(sys_foreign_err == DB_SUCCESS);
|
|
|
|
|
|
|
|
sys_foreign_cols_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_FOREIGN_COLS", DICT_NUM_FIELDS__SYS_FOREIGN_COLS + 1, 1);
|
|
|
|
ut_a(sys_foreign_cols_err == DB_SUCCESS);
|
|
|
|
|
|
|
|
return(err);
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************//**
|
|
|
|
Evaluate the given foreign key SQL statement.
|
|
|
|
@return error code or DB_SUCCESS */
|
2016-06-21 14:21:03 +02:00
|
|
|
static MY_ATTRIBUTE((nonnull, warn_unused_result))
|
2014-02-26 19:11:54 +01:00
|
|
|
dberr_t
|
|
|
|
dict_foreign_eval_sql(
|
|
|
|
/*==================*/
|
|
|
|
pars_info_t* info, /*!< in: info struct */
|
|
|
|
const char* sql, /*!< in: SQL string to evaluate */
|
|
|
|
const char* name, /*!< in: table name (for diagnostics) */
|
|
|
|
const char* id, /*!< in: foreign key id */
|
|
|
|
trx_t* trx) /*!< in/out: transaction */
|
|
|
|
{
|
|
|
|
dberr_t error;
|
|
|
|
FILE* ef = dict_foreign_err_file;
|
|
|
|
|
|
|
|
error = que_eval_sql(info, sql, FALSE, trx);
|
|
|
|
|
|
|
|
if (error == DB_DUPLICATE_KEY) {
|
|
|
|
mutex_enter(&dict_foreign_err_mutex);
|
|
|
|
rewind(ef);
|
|
|
|
ut_print_timestamp(ef);
|
|
|
|
fputs(" Error in foreign key constraint creation for table ",
|
|
|
|
ef);
|
|
|
|
ut_print_name(ef, trx, TRUE, name);
|
|
|
|
fputs(".\nA foreign key constraint of name ", ef);
|
|
|
|
ut_print_name(ef, trx, TRUE, id);
|
|
|
|
fputs("\nalready exists."
|
|
|
|
" (Note that internally InnoDB adds 'databasename'\n"
|
|
|
|
"in front of the user-defined constraint name.)\n"
|
|
|
|
"Note that InnoDB's FOREIGN KEY system tables store\n"
|
|
|
|
"constraint names as case-insensitive, with the\n"
|
|
|
|
"MySQL standard latin1_swedish_ci collation. If you\n"
|
|
|
|
"create tables or databases whose names differ only in\n"
|
|
|
|
"the character case, then collisions in constraint\n"
|
|
|
|
"names can occur. Workaround: name your constraints\n"
|
|
|
|
"explicitly with unique names.\n",
|
|
|
|
ef);
|
|
|
|
|
|
|
|
mutex_exit(&dict_foreign_err_mutex);
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (error != DB_SUCCESS) {
|
|
|
|
fprintf(stderr,
|
|
|
|
"InnoDB: Foreign key constraint creation failed:\n"
|
|
|
|
"InnoDB: internal error number %lu\n", (ulong) error);
|
|
|
|
|
|
|
|
mutex_enter(&dict_foreign_err_mutex);
|
|
|
|
ut_print_timestamp(ef);
|
|
|
|
fputs(" Internal error in foreign key constraint creation"
|
|
|
|
" for table ", ef);
|
|
|
|
ut_print_name(ef, trx, TRUE, name);
|
|
|
|
fputs(".\n"
|
|
|
|
"See the MySQL .err log in the datadir"
|
|
|
|
" for more information.\n", ef);
|
|
|
|
mutex_exit(&dict_foreign_err_mutex);
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/********************************************************************//**
|
|
|
|
Add a single foreign key field definition to the data dictionary tables in
|
|
|
|
the database.
|
|
|
|
@return error code or DB_SUCCESS */
|
2016-06-21 14:21:03 +02:00
|
|
|
static MY_ATTRIBUTE((nonnull, warn_unused_result))
|
2014-02-26 19:11:54 +01:00
|
|
|
dberr_t
|
|
|
|
dict_create_add_foreign_field_to_dictionary(
|
|
|
|
/*========================================*/
|
|
|
|
ulint field_nr, /*!< in: field number */
|
|
|
|
const char* table_name, /*!< in: table name */
|
|
|
|
const dict_foreign_t* foreign, /*!< in: foreign */
|
|
|
|
trx_t* trx) /*!< in/out: transaction */
|
|
|
|
{
|
|
|
|
pars_info_t* info = pars_info_create();
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "id", foreign->id);
|
|
|
|
|
|
|
|
pars_info_add_int4_literal(info, "pos", field_nr);
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "for_col_name",
|
|
|
|
foreign->foreign_col_names[field_nr]);
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "ref_col_name",
|
|
|
|
foreign->referenced_col_names[field_nr]);
|
|
|
|
|
|
|
|
return(dict_foreign_eval_sql(
|
|
|
|
info,
|
|
|
|
"PROCEDURE P () IS\n"
|
|
|
|
"BEGIN\n"
|
|
|
|
"INSERT INTO SYS_FOREIGN_COLS VALUES"
|
|
|
|
"(:id, :pos, :for_col_name, :ref_col_name);\n"
|
|
|
|
"END;\n",
|
|
|
|
table_name, foreign->id, trx));
|
|
|
|
}
|
|
|
|
|
2015-07-29 05:58:45 +03:00
|
|
|
/********************************************************************//**
|
|
|
|
Construct foreign key constraint defintion from data dictionary information.
|
|
|
|
*/
|
2015-07-31 08:52:24 +03:00
|
|
|
UNIV_INTERN
|
2015-07-29 05:58:45 +03:00
|
|
|
char*
|
|
|
|
dict_foreign_def_get(
|
2015-07-31 08:52:24 +03:00
|
|
|
/*=================*/
|
2015-07-29 05:58:45 +03:00
|
|
|
dict_foreign_t* foreign,/*!< in: foreign */
|
|
|
|
trx_t* trx) /*!< in: trx */
|
|
|
|
{
|
2015-08-03 23:09:43 +03:00
|
|
|
char* fk_def = (char *)mem_heap_alloc(foreign->heap, 4*1024);
|
2015-07-29 05:58:45 +03:00
|
|
|
const char* tbname;
|
|
|
|
char tablebuf[MAX_TABLE_NAME_LEN + 1] = "";
|
|
|
|
int i;
|
2015-07-31 12:06:29 +03:00
|
|
|
char* bufend;
|
2015-07-29 05:58:45 +03:00
|
|
|
|
|
|
|
tbname = dict_remove_db_name(foreign->id);
|
2015-07-31 12:06:29 +03:00
|
|
|
bufend = innobase_convert_name(tablebuf, MAX_TABLE_NAME_LEN,
|
2015-07-29 05:58:45 +03:00
|
|
|
tbname, strlen(tbname), trx->mysql_thd, FALSE);
|
2015-07-31 12:06:29 +03:00
|
|
|
tablebuf[bufend - tablebuf] = '\0';
|
2015-07-29 05:58:45 +03:00
|
|
|
|
|
|
|
sprintf(fk_def,
|
|
|
|
(char *)"CONSTRAINT %s FOREIGN KEY (", (char *)tablebuf);
|
|
|
|
|
|
|
|
for(i = 0; i < foreign->n_fields; i++) {
|
|
|
|
char buf[MAX_TABLE_NAME_LEN + 1] = "";
|
|
|
|
innobase_convert_name(buf, MAX_TABLE_NAME_LEN,
|
|
|
|
foreign->foreign_col_names[i],
|
|
|
|
strlen(foreign->foreign_col_names[i]),
|
|
|
|
trx->mysql_thd, FALSE);
|
|
|
|
strcat(fk_def, buf);
|
|
|
|
if (i < foreign->n_fields-1) {
|
|
|
|
strcat(fk_def, (char *)",");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
strcat(fk_def,(char *)") REFERENCES ");
|
|
|
|
|
2015-07-31 12:06:29 +03:00
|
|
|
bufend = innobase_convert_name(tablebuf, MAX_TABLE_NAME_LEN,
|
|
|
|
foreign->referenced_table_name,
|
|
|
|
strlen(foreign->referenced_table_name),
|
|
|
|
trx->mysql_thd, TRUE);
|
|
|
|
tablebuf[bufend - tablebuf] = '\0';
|
2015-07-29 05:58:45 +03:00
|
|
|
|
|
|
|
strcat(fk_def, tablebuf);
|
|
|
|
strcat(fk_def, " (");
|
|
|
|
|
|
|
|
for(i = 0; i < foreign->n_fields; i++) {
|
|
|
|
char buf[MAX_TABLE_NAME_LEN + 1] = "";
|
2015-07-31 12:06:29 +03:00
|
|
|
bufend = innobase_convert_name(buf, MAX_TABLE_NAME_LEN,
|
2015-07-29 05:58:45 +03:00
|
|
|
foreign->referenced_col_names[i],
|
|
|
|
strlen(foreign->referenced_col_names[i]),
|
|
|
|
trx->mysql_thd, FALSE);
|
2015-07-31 12:06:29 +03:00
|
|
|
buf[bufend - buf] = '\0';
|
2015-07-29 05:58:45 +03:00
|
|
|
strcat(fk_def, buf);
|
|
|
|
if (i < foreign->n_fields-1) {
|
|
|
|
strcat(fk_def, (char *)",");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
strcat(fk_def, (char *)")");
|
|
|
|
|
|
|
|
return fk_def;
|
|
|
|
}
|
|
|
|
|
|
|
|
/********************************************************************//**
|
|
|
|
Convert foreign key column names from data dictionary to SQL-layer.
|
|
|
|
*/
|
|
|
|
static
|
|
|
|
void
|
|
|
|
dict_foreign_def_get_fields(
|
2015-07-31 08:52:24 +03:00
|
|
|
/*========================*/
|
2015-07-29 05:58:45 +03:00
|
|
|
dict_foreign_t* foreign,/*!< in: foreign */
|
|
|
|
trx_t* trx, /*!< in: trx */
|
|
|
|
char** field, /*!< out: foreign column */
|
|
|
|
char** field2, /*!< out: referenced column */
|
|
|
|
int col_no) /*!< in: column number */
|
|
|
|
{
|
2015-07-31 12:06:29 +03:00
|
|
|
char* bufend;
|
2015-08-03 23:09:43 +03:00
|
|
|
char* fieldbuf = (char *)mem_heap_alloc(foreign->heap, MAX_TABLE_NAME_LEN+1);
|
|
|
|
char* fieldbuf2 = (char *)mem_heap_alloc(foreign->heap, MAX_TABLE_NAME_LEN+1);
|
2015-07-31 12:06:29 +03:00
|
|
|
|
|
|
|
bufend = innobase_convert_name(fieldbuf, MAX_TABLE_NAME_LEN,
|
|
|
|
foreign->foreign_col_names[col_no],
|
|
|
|
strlen(foreign->foreign_col_names[col_no]),
|
|
|
|
trx->mysql_thd, FALSE);
|
|
|
|
|
|
|
|
fieldbuf[bufend - fieldbuf] = '\0';
|
|
|
|
|
|
|
|
bufend = innobase_convert_name(fieldbuf2, MAX_TABLE_NAME_LEN,
|
|
|
|
foreign->referenced_col_names[col_no],
|
|
|
|
strlen(foreign->referenced_col_names[col_no]),
|
|
|
|
trx->mysql_thd, FALSE);
|
|
|
|
|
|
|
|
fieldbuf2[bufend - fieldbuf2] = '\0';
|
|
|
|
*field = fieldbuf;
|
|
|
|
*field2 = fieldbuf2;
|
2015-07-29 05:58:45 +03:00
|
|
|
}
|
|
|
|
|
2014-02-26 19:11:54 +01:00
|
|
|
/********************************************************************//**
|
|
|
|
Add a foreign key definition to the data dictionary tables.
|
|
|
|
@return error code or DB_SUCCESS */
|
|
|
|
UNIV_INTERN
|
|
|
|
dberr_t
|
|
|
|
dict_create_add_foreign_to_dictionary(
|
|
|
|
/*==================================*/
|
2015-08-03 23:09:43 +03:00
|
|
|
dict_table_t* table, /*!< in: table */
|
2014-02-26 19:11:54 +01:00
|
|
|
const char* name, /*!< in: table name */
|
|
|
|
const dict_foreign_t* foreign,/*!< in: foreign key */
|
|
|
|
trx_t* trx) /*!< in/out: dictionary transaction */
|
|
|
|
{
|
|
|
|
dberr_t error;
|
|
|
|
pars_info_t* info = pars_info_create();
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "id", foreign->id);
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "for_name", name);
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "ref_name",
|
|
|
|
foreign->referenced_table_name);
|
|
|
|
|
|
|
|
pars_info_add_int4_literal(info, "n_cols",
|
|
|
|
foreign->n_fields + (foreign->type << 24));
|
|
|
|
|
|
|
|
error = dict_foreign_eval_sql(info,
|
|
|
|
"PROCEDURE P () IS\n"
|
|
|
|
"BEGIN\n"
|
|
|
|
"INSERT INTO SYS_FOREIGN VALUES"
|
|
|
|
"(:id, :for_name, :ref_name, :n_cols);\n"
|
|
|
|
"END;\n"
|
|
|
|
, name, foreign->id, trx);
|
|
|
|
|
|
|
|
if (error != DB_SUCCESS) {
|
|
|
|
|
2015-07-29 05:58:45 +03:00
|
|
|
if (error == DB_DUPLICATE_KEY) {
|
|
|
|
char buf[MAX_TABLE_NAME_LEN + 1] = "";
|
2015-07-31 08:52:24 +03:00
|
|
|
char tablename[MAX_TABLE_NAME_LEN + 1] = "";
|
2015-07-29 05:58:45 +03:00
|
|
|
char* fk_def;
|
|
|
|
|
2015-07-31 08:52:24 +03:00
|
|
|
innobase_convert_name(tablename, MAX_TABLE_NAME_LEN,
|
|
|
|
table->name, strlen(table->name),
|
|
|
|
trx->mysql_thd, TRUE);
|
|
|
|
|
2015-07-29 05:58:45 +03:00
|
|
|
innobase_convert_name(buf, MAX_TABLE_NAME_LEN,
|
|
|
|
foreign->id, strlen(foreign->id), trx->mysql_thd, FALSE);
|
|
|
|
|
2015-08-03 23:09:43 +03:00
|
|
|
fk_def = dict_foreign_def_get((dict_foreign_t*)foreign, trx);
|
2015-07-29 05:58:45 +03:00
|
|
|
|
2015-07-31 08:52:24 +03:00
|
|
|
ib_push_warning(trx, error,
|
|
|
|
"Create or Alter table %s with foreign key constraint"
|
|
|
|
" failed. Foreign key constraint %s"
|
|
|
|
" already exists on data dictionary."
|
2015-07-29 05:58:45 +03:00
|
|
|
" Foreign key constraint names need to be unique in database."
|
|
|
|
" Error in foreign key definition: %s.",
|
2015-07-31 08:52:24 +03:00
|
|
|
tablename, buf, fk_def);
|
2015-07-29 05:58:45 +03:00
|
|
|
}
|
|
|
|
|
2014-02-26 19:11:54 +01:00
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (ulint i = 0; i < foreign->n_fields; i++) {
|
|
|
|
error = dict_create_add_foreign_field_to_dictionary(
|
|
|
|
i, name, foreign, trx);
|
|
|
|
|
|
|
|
if (error != DB_SUCCESS) {
|
2015-07-29 05:58:45 +03:00
|
|
|
char buf[MAX_TABLE_NAME_LEN + 1] = "";
|
2015-07-31 08:52:24 +03:00
|
|
|
char tablename[MAX_TABLE_NAME_LEN + 1] = "";
|
2015-07-29 05:58:45 +03:00
|
|
|
char* field=NULL;
|
|
|
|
char* field2=NULL;
|
|
|
|
char* fk_def;
|
|
|
|
|
2015-07-31 08:52:24 +03:00
|
|
|
innobase_convert_name(tablename, MAX_TABLE_NAME_LEN,
|
|
|
|
table->name, strlen(table->name),
|
|
|
|
trx->mysql_thd, TRUE);
|
2015-07-29 05:58:45 +03:00
|
|
|
innobase_convert_name(buf, MAX_TABLE_NAME_LEN,
|
|
|
|
foreign->id, strlen(foreign->id), trx->mysql_thd, FALSE);
|
2015-08-03 23:09:43 +03:00
|
|
|
fk_def = dict_foreign_def_get((dict_foreign_t*)foreign, trx);
|
|
|
|
dict_foreign_def_get_fields((dict_foreign_t*)foreign, trx, &field, &field2, i);
|
2015-07-29 05:58:45 +03:00
|
|
|
|
|
|
|
ib_push_warning(trx, error,
|
2015-07-31 08:52:24 +03:00
|
|
|
"Create or Alter table %s with foreign key constraint"
|
|
|
|
" failed. Error adding foreign key constraint name %s"
|
|
|
|
" fields %s or %s to the dictionary."
|
2015-07-29 05:58:45 +03:00
|
|
|
" Error in foreign key definition: %s.",
|
2015-07-31 08:52:24 +03:00
|
|
|
tablename, buf, i+1, fk_def);
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
2014-09-11 10:13:35 +02:00
|
|
|
/** Adds the given set of foreign key objects to the dictionary tables
|
|
|
|
in the database. This function does not modify the dictionary cache. The
|
|
|
|
caller must ensure that all foreign key objects contain a valid constraint
|
|
|
|
name in foreign->id.
|
|
|
|
@param[in] local_fk_set set of foreign key objects, to be added to
|
|
|
|
the dictionary tables
|
|
|
|
@param[in] table table to which the foreign key objects in
|
|
|
|
local_fk_set belong to
|
|
|
|
@param[in,out] trx transaction
|
|
|
|
@return error code or DB_SUCCESS */
|
2014-02-26 19:11:54 +01:00
|
|
|
UNIV_INTERN
|
|
|
|
dberr_t
|
|
|
|
dict_create_add_foreigns_to_dictionary(
|
|
|
|
/*===================================*/
|
2014-09-11 10:13:35 +02:00
|
|
|
const dict_foreign_set& local_fk_set,
|
|
|
|
const dict_table_t* table,
|
|
|
|
trx_t* trx)
|
2014-02-26 19:11:54 +01:00
|
|
|
{
|
|
|
|
dict_foreign_t* foreign;
|
|
|
|
dberr_t error;
|
|
|
|
|
|
|
|
ut_ad(mutex_own(&(dict_sys->mutex)));
|
|
|
|
|
|
|
|
if (NULL == dict_table_get_low("SYS_FOREIGN")) {
|
|
|
|
fprintf(stderr,
|
|
|
|
"InnoDB: table SYS_FOREIGN not found"
|
|
|
|
" in internal data dictionary\n");
|
|
|
|
|
|
|
|
return(DB_ERROR);
|
|
|
|
}
|
|
|
|
|
2014-09-11 10:13:35 +02:00
|
|
|
for (dict_foreign_set::const_iterator it = local_fk_set.begin();
|
|
|
|
it != local_fk_set.end();
|
|
|
|
++it) {
|
2014-02-26 19:11:54 +01:00
|
|
|
|
2014-09-11 10:13:35 +02:00
|
|
|
foreign = *it;
|
|
|
|
ut_ad(foreign->id != NULL);
|
2014-02-26 19:11:54 +01:00
|
|
|
|
2015-08-03 23:09:43 +03:00
|
|
|
error = dict_create_add_foreign_to_dictionary((dict_table_t*)table, table->name,
|
2014-02-26 19:11:54 +01:00
|
|
|
foreign, trx);
|
|
|
|
|
|
|
|
if (error != DB_SUCCESS) {
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
trx->op_info = "committing foreign key definitions";
|
|
|
|
|
|
|
|
trx_commit(trx);
|
|
|
|
|
|
|
|
trx->op_info = "";
|
|
|
|
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/****************************************************************//**
|
|
|
|
Creates the tablespaces and datafiles system tables inside InnoDB
|
|
|
|
at server bootstrap or server start if they are not found or are
|
|
|
|
not of the right form.
|
|
|
|
@return DB_SUCCESS or error code */
|
|
|
|
UNIV_INTERN
|
|
|
|
dberr_t
|
|
|
|
dict_create_or_check_sys_tablespace(void)
|
|
|
|
/*=====================================*/
|
|
|
|
{
|
|
|
|
trx_t* trx;
|
|
|
|
my_bool srv_file_per_table_backup;
|
|
|
|
dberr_t err;
|
|
|
|
dberr_t sys_tablespaces_err;
|
|
|
|
dberr_t sys_datafiles_err;
|
|
|
|
|
|
|
|
ut_a(srv_get_active_thread_type() == SRV_NONE);
|
|
|
|
|
|
|
|
/* Note: The master thread has not been started at this point. */
|
|
|
|
|
|
|
|
sys_tablespaces_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_TABLESPACES", DICT_NUM_FIELDS__SYS_TABLESPACES + 1, 1);
|
|
|
|
sys_datafiles_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_DATAFILES", DICT_NUM_FIELDS__SYS_DATAFILES + 1, 1);
|
|
|
|
|
|
|
|
if (sys_tablespaces_err == DB_SUCCESS
|
|
|
|
&& sys_datafiles_err == DB_SUCCESS) {
|
|
|
|
return(DB_SUCCESS);
|
|
|
|
}
|
|
|
|
|
|
|
|
trx = trx_allocate_for_mysql();
|
|
|
|
|
|
|
|
trx_set_dict_operation(trx, TRX_DICT_OP_TABLE);
|
|
|
|
|
|
|
|
trx->op_info = "creating tablepace and datafile sys tables";
|
|
|
|
|
|
|
|
row_mysql_lock_data_dictionary(trx);
|
|
|
|
|
|
|
|
/* Check which incomplete table definition to drop. */
|
|
|
|
|
|
|
|
if (sys_tablespaces_err == DB_CORRUPTION) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_WARN,
|
|
|
|
"Dropping incompletely created "
|
|
|
|
"SYS_TABLESPACES table.");
|
2016-06-23 07:42:40 +03:00
|
|
|
row_drop_table_for_mysql("SYS_TABLESPACES", trx, TRUE, TRUE);
|
2014-02-26 19:11:54 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
if (sys_datafiles_err == DB_CORRUPTION) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_WARN,
|
|
|
|
"Dropping incompletely created "
|
|
|
|
"SYS_DATAFILES table.");
|
|
|
|
|
2016-06-23 07:42:40 +03:00
|
|
|
row_drop_table_for_mysql("SYS_DATAFILES", trx, TRUE, TRUE);
|
2014-02-26 19:11:54 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
ib_logf(IB_LOG_LEVEL_INFO,
|
|
|
|
"Creating tablespace and datafile system tables.");
|
|
|
|
|
|
|
|
/* We always want SYSTEM tables to be created inside the system
|
|
|
|
tablespace. */
|
|
|
|
srv_file_per_table_backup = srv_file_per_table;
|
|
|
|
srv_file_per_table = 0;
|
|
|
|
|
|
|
|
err = que_eval_sql(
|
|
|
|
NULL,
|
|
|
|
"PROCEDURE CREATE_SYS_TABLESPACE_PROC () IS\n"
|
|
|
|
"BEGIN\n"
|
|
|
|
"CREATE TABLE SYS_TABLESPACES(\n"
|
|
|
|
" SPACE INT, NAME CHAR, FLAGS INT);\n"
|
|
|
|
"CREATE UNIQUE CLUSTERED INDEX SYS_TABLESPACES_SPACE"
|
|
|
|
" ON SYS_TABLESPACES (SPACE);\n"
|
|
|
|
"CREATE TABLE SYS_DATAFILES(\n"
|
|
|
|
" SPACE INT, PATH CHAR);\n"
|
|
|
|
"CREATE UNIQUE CLUSTERED INDEX SYS_DATAFILES_SPACE"
|
|
|
|
" ON SYS_DATAFILES (SPACE);\n"
|
|
|
|
"END;\n",
|
|
|
|
FALSE, trx);
|
|
|
|
|
|
|
|
if (err != DB_SUCCESS) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_ERROR,
|
|
|
|
"Creation of SYS_TABLESPACES and SYS_DATAFILES "
|
|
|
|
"has failed with error %lu. Tablespace is full. "
|
|
|
|
"Dropping incompletely created tables.",
|
|
|
|
(ulong) err);
|
|
|
|
|
|
|
|
ut_a(err == DB_OUT_OF_FILE_SPACE
|
|
|
|
|| err == DB_TOO_MANY_CONCURRENT_TRXS);
|
|
|
|
|
2016-06-23 07:42:40 +03:00
|
|
|
row_drop_table_for_mysql("SYS_TABLESPACES", trx, TRUE, TRUE);
|
|
|
|
row_drop_table_for_mysql("SYS_DATAFILES", trx, TRUE, TRUE);
|
2014-02-26 19:11:54 +01:00
|
|
|
|
|
|
|
if (err == DB_OUT_OF_FILE_SPACE) {
|
|
|
|
err = DB_MUST_GET_MORE_FILE_SPACE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
trx_commit_for_mysql(trx);
|
|
|
|
|
|
|
|
row_mysql_unlock_data_dictionary(trx);
|
|
|
|
|
|
|
|
trx_free_for_mysql(trx);
|
|
|
|
|
|
|
|
srv_file_per_table = srv_file_per_table_backup;
|
|
|
|
|
|
|
|
if (err == DB_SUCCESS) {
|
|
|
|
ib_logf(IB_LOG_LEVEL_INFO,
|
|
|
|
"Tablespace and datafile system tables created.");
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Note: The master thread has not been started at this point. */
|
|
|
|
/* Confirm and move to the non-LRU part of the table LRU list. */
|
|
|
|
|
|
|
|
sys_tablespaces_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_TABLESPACES", DICT_NUM_FIELDS__SYS_TABLESPACES + 1, 1);
|
|
|
|
ut_a(sys_tablespaces_err == DB_SUCCESS);
|
|
|
|
|
|
|
|
sys_datafiles_err = dict_check_if_system_table_exists(
|
|
|
|
"SYS_DATAFILES", DICT_NUM_FIELDS__SYS_DATAFILES + 1, 1);
|
|
|
|
ut_a(sys_datafiles_err == DB_SUCCESS);
|
|
|
|
|
|
|
|
return(err);
|
|
|
|
}
|
|
|
|
|
|
|
|
/********************************************************************//**
|
|
|
|
Add a single tablespace definition to the data dictionary tables in the
|
|
|
|
database.
|
|
|
|
@return error code or DB_SUCCESS */
|
|
|
|
UNIV_INTERN
|
|
|
|
dberr_t
|
|
|
|
dict_create_add_tablespace_to_dictionary(
|
|
|
|
/*=====================================*/
|
|
|
|
ulint space, /*!< in: tablespace id */
|
|
|
|
const char* name, /*!< in: tablespace name */
|
|
|
|
ulint flags, /*!< in: tablespace flags */
|
|
|
|
const char* path, /*!< in: tablespace path */
|
|
|
|
trx_t* trx, /*!< in/out: transaction */
|
|
|
|
bool commit) /*!< in: if true then commit the
|
|
|
|
transaction */
|
|
|
|
{
|
|
|
|
dberr_t error;
|
|
|
|
|
|
|
|
pars_info_t* info = pars_info_create();
|
|
|
|
|
|
|
|
ut_a(space > TRX_SYS_SPACE);
|
|
|
|
|
|
|
|
pars_info_add_int4_literal(info, "space", space);
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "name", name);
|
|
|
|
|
|
|
|
pars_info_add_int4_literal(info, "flags", flags);
|
|
|
|
|
|
|
|
pars_info_add_str_literal(info, "path", path);
|
|
|
|
|
|
|
|
error = que_eval_sql(info,
|
|
|
|
"PROCEDURE P () IS\n"
|
|
|
|
"BEGIN\n"
|
|
|
|
"INSERT INTO SYS_TABLESPACES VALUES"
|
|
|
|
"(:space, :name, :flags);\n"
|
|
|
|
"INSERT INTO SYS_DATAFILES VALUES"
|
|
|
|
"(:space, :path);\n"
|
|
|
|
"END;\n",
|
|
|
|
FALSE, trx);
|
|
|
|
|
|
|
|
if (error != DB_SUCCESS) {
|
|
|
|
return(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (commit) {
|
|
|
|
trx->op_info = "committing tablespace and datafile definition";
|
|
|
|
trx_commit(trx);
|
|
|
|
}
|
|
|
|
|
|
|
|
trx->op_info = "";
|
|
|
|
|
|
|
|
return(error);
|
|
|
|
}
|