2023-10-25 10:08:20 +03:00
|
|
|
SET GLOBAL innodb_max_purge_lag_wait=0;
|
|
|
|
connect stop_purge,localhost,root;
|
|
|
|
START TRANSACTION WITH CONSISTENT SNAPSHOT;
|
|
|
|
connection default;
|
2017-08-29 11:27:28 +03:00
|
|
|
CREATE TABLE t1(c1 INT NOT NULL, c2 INT, PRIMARY KEY(c1)) Engine=InnoDB;
|
|
|
|
SHOW CREATE TABLE t1;
|
|
|
|
Table Create Table
|
|
|
|
t1 CREATE TABLE `t1` (
|
|
|
|
`c1` int(11) NOT NULL,
|
|
|
|
`c2` int(11) DEFAULT NULL,
|
|
|
|
PRIMARY KEY (`c1`)
|
2022-09-02 17:32:14 +04:00
|
|
|
) ENGINE=InnoDB DEFAULT CHARSET=latin1 COLLATE=latin1_swedish_ci
|
2017-08-29 11:27:28 +03:00
|
|
|
INSERT INTO t1 VALUES (1,1),(2,2),(3,3),(4,4),(5,5);
|
2017-08-29 15:40:37 +03:00
|
|
|
SET @saved_debug_dbug = @@SESSION.debug_dbug;
|
|
|
|
SET DEBUG_DBUG='+d,ib_build_indexes_too_many_concurrent_trxs, ib_rename_indexes_too_many_concurrent_trxs, ib_drop_index_too_many_concurrent_trxs';
|
2017-08-29 11:27:28 +03:00
|
|
|
ALTER TABLE t1 ADD UNIQUE INDEX(c2);
|
|
|
|
ERROR HY000: Too many active concurrent transactions
|
2017-08-29 15:40:37 +03:00
|
|
|
SET DEBUG_DBUG = @saved_debug_dbug;
|
2017-08-29 11:27:28 +03:00
|
|
|
SHOW CREATE TABLE t1;
|
|
|
|
Table Create Table
|
|
|
|
t1 CREATE TABLE `t1` (
|
|
|
|
`c1` int(11) NOT NULL,
|
|
|
|
`c2` int(11) DEFAULT NULL,
|
|
|
|
PRIMARY KEY (`c1`)
|
2022-09-02 17:32:14 +04:00
|
|
|
) ENGINE=InnoDB DEFAULT CHARSET=latin1 COLLATE=latin1_swedish_ci
|
2017-08-29 11:27:28 +03:00
|
|
|
DROP TABLE t1;
|
|
|
|
CREATE TABLE bug13861218 (c1 INT NOT NULL, c2 INT NOT NULL, INDEX(c2))
|
|
|
|
ENGINE=InnoDB;
|
|
|
|
INSERT INTO bug13861218 VALUES (8, 0), (4, 0), (0, 0);
|
2017-08-29 15:40:37 +03:00
|
|
|
SET DEBUG_DBUG = '+d,ib_row_merge_buf_add_two';
|
2017-08-29 11:27:28 +03:00
|
|
|
CREATE UNIQUE INDEX ui ON bug13861218(c1);
|
2017-08-29 15:40:37 +03:00
|
|
|
SET DEBUG_DBUG = @saved_debug_dbug;
|
2017-08-29 11:27:28 +03:00
|
|
|
DROP TABLE bug13861218;
|
|
|
|
CREATE TABLE bug13861218 (c1 INT NOT NULL, c2 INT UNIQUE) ENGINE=InnoDB;
|
|
|
|
INSERT INTO bug13861218 VALUES (8, NULL), (4, NULL), (0, NULL);
|
2017-08-29 15:40:37 +03:00
|
|
|
SET DEBUG_DBUG = '+d,ib_row_merge_buf_add_two';
|
2017-08-29 11:27:28 +03:00
|
|
|
CREATE UNIQUE INDEX ui ON bug13861218(c1);
|
2017-08-29 15:40:37 +03:00
|
|
|
SET DEBUG_DBUG = @saved_debug_dbug;
|
2017-08-29 11:27:28 +03:00
|
|
|
DROP TABLE bug13861218;
|
|
|
|
#
|
2017-09-14 16:07:10 +03:00
|
|
|
# Bug #17657223 EXCESSIVE TEMPORARY FILE USAGE IN ALTER TABLE
|
|
|
|
#
|
2017-08-29 11:27:28 +03:00
|
|
|
create table t480(a serial)engine=innodb;
|
|
|
|
insert into t480
|
|
|
|
values(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),(),
|
|
|
|
(),(),(),(),(),(),(),();
|
|
|
|
insert into t480 select 0 from t480;
|
|
|
|
insert into t480 select 0 from t480;
|
|
|
|
insert into t480 select 0 from t480;
|
|
|
|
insert into t480 select 0 from t480;
|
|
|
|
create table t1(f1 int auto_increment not null,
|
|
|
|
f2 char(200) not null, f3 char(200) not null,
|
2017-09-14 16:07:10 +03:00
|
|
|
primary key(f1,f2,f3), key(f1))engine=innodb;
|
|
|
|
insert into t1 select NULL,'aaa','bbb' from t480;
|
|
|
|
insert into t1 select NULL,'aaaa','bbbb' from t480;
|
|
|
|
insert into t1 select NULL,'aaaaa','bbbbb' from t480;
|
|
|
|
insert into t1 select NULL,'aaaaaa','bbbbbb' from t480;
|
|
|
|
SET DEBUG_DBUG = '+d,row_merge_write_failure';
|
|
|
|
alter table t1 drop primary key,add primary key(f2,f1);
|
|
|
|
ERROR HY000: Temporary file write failure
|
|
|
|
SET DEBUG_DBUG = @saved_debug_dbug;
|
|
|
|
drop table t1;
|
|
|
|
connect con1,localhost,root;
|
|
|
|
create table t1(k1 int auto_increment primary key,
|
|
|
|
k2 char(200),k3 char(200))engine=innodb;
|
|
|
|
insert into t1 values(NULL,'a','b'), (NULL,'aa','bb');
|
|
|
|
SET DEBUG_SYNC= 'row_merge_after_scan
|
|
|
|
SIGNAL opened WAIT_FOR flushed';
|
|
|
|
optimize table t1;
|
|
|
|
connection default;
|
|
|
|
SET DEBUG_SYNC= 'now WAIT_FOR opened';
|
|
|
|
INSERT INTO t1 select NULL,'aaa','bbb' from t480;
|
|
|
|
SET DEBUG_SYNC= 'now SIGNAL flushed';
|
|
|
|
connection con1;
|
|
|
|
/*con1 reap*/ Optimize table t1;
|
|
|
|
Table Op Msg_type Msg_text
|
|
|
|
test.t1 optimize note Table does not support optimize, doing recreate + analyze instead
|
|
|
|
test.t1 optimize status OK
|
|
|
|
affected rows: 2
|
|
|
|
SELECT COUNT(k1),k2,k3 FROM t1 GROUP BY k2,k3;
|
|
|
|
COUNT(k1) k2 k3
|
|
|
|
1 a b
|
|
|
|
1 aa bb
|
|
|
|
480 aaa bbb
|
|
|
|
drop table t1;
|
|
|
|
create table t1(k1 int auto_increment primary key,
|
|
|
|
k2 char(200),k3 char(200))engine=innodb;
|
2021-11-12 17:46:35 +05:30
|
|
|
INSERT INTO t1 VALUES(1, "test", "test");
|
2017-09-14 16:07:10 +03:00
|
|
|
SET DEBUG_SYNC= 'row_merge_after_scan
|
|
|
|
SIGNAL opened WAIT_FOR flushed';
|
MDEV-11369 Instant ADD COLUMN for InnoDB
For InnoDB tables, adding, dropping and reordering columns has
required a rebuild of the table and all its indexes. Since MySQL 5.6
(and MariaDB 10.0) this has been supported online (LOCK=NONE), allowing
concurrent modification of the tables.
This work revises the InnoDB ROW_FORMAT=REDUNDANT, ROW_FORMAT=COMPACT
and ROW_FORMAT=DYNAMIC so that columns can be appended instantaneously,
with only minor changes performed to the table structure. The counter
innodb_instant_alter_column in INFORMATION_SCHEMA.GLOBAL_STATUS
is incremented whenever a table rebuild operation is converted into
an instant ADD COLUMN operation.
ROW_FORMAT=COMPRESSED tables will not support instant ADD COLUMN.
Some usability limitations will be addressed in subsequent work:
MDEV-13134 Introduce ALTER TABLE attributes ALGORITHM=NOCOPY
and ALGORITHM=INSTANT
MDEV-14016 Allow instant ADD COLUMN, ADD INDEX, LOCK=NONE
The format of the clustered index (PRIMARY KEY) is changed as follows:
(1) The FIL_PAGE_TYPE of the root page will be FIL_PAGE_TYPE_INSTANT,
and a new field PAGE_INSTANT will contain the original number of fields
in the clustered index ('core' fields).
If instant ADD COLUMN has not been used or the table becomes empty,
or the very first instant ADD COLUMN operation is rolled back,
the fields PAGE_INSTANT and FIL_PAGE_TYPE will be reset
to 0 and FIL_PAGE_INDEX.
(2) A special 'default row' record is inserted into the leftmost leaf,
between the page infimum and the first user record. This record is
distinguished by the REC_INFO_MIN_REC_FLAG, and it is otherwise in the
same format as records that contain values for the instantly added
columns. This 'default row' always has the same number of fields as
the clustered index according to the table definition. The values of
'core' fields are to be ignored. For other fields, the 'default row'
will contain the default values as they were during the ALTER TABLE
statement. (If the column default values are changed later, those
values will only be stored in the .frm file. The 'default row' will
contain the original evaluated values, which must be the same for
every row.) The 'default row' must be completely hidden from
higher-level access routines. Assertions have been added to ensure
that no 'default row' is ever present in the adaptive hash index
or in locked records. The 'default row' is never delete-marked.
(3) In clustered index leaf page records, the number of fields must
reside between the number of 'core' fields (dict_index_t::n_core_fields
introduced in this work) and dict_index_t::n_fields. If the number
of fields is less than dict_index_t::n_fields, the missing fields
are replaced with the column value of the 'default row'.
Note: The number of fields in the record may shrink if some of the
last instantly added columns are updated to the value that is
in the 'default row'. The function btr_cur_trim() implements this
'compression' on update and rollback; dtuple::trim() implements it
on insert.
(4) In ROW_FORMAT=COMPACT and ROW_FORMAT=DYNAMIC records, the new
status value REC_STATUS_COLUMNS_ADDED will indicate the presence of
a new record header that will encode n_fields-n_core_fields-1 in
1 or 2 bytes. (In ROW_FORMAT=REDUNDANT records, the record header
always explicitly encodes the number of fields.)
We introduce the undo log record type TRX_UNDO_INSERT_DEFAULT for
covering the insert of the 'default row' record when instant ADD COLUMN
is used for the first time. Subsequent instant ADD COLUMN can use
TRX_UNDO_UPD_EXIST_REC.
This is joint work with Vin Chen (陈福荣) from Tencent. The design
that was discussed in April 2017 would not have allowed import or
export of data files, because instead of the 'default row' it would
have introduced a data dictionary table. The test
rpl.rpl_alter_instant is exactly as contributed in pull request #408.
The test innodb.instant_alter is based on a contributed test.
The redo log record format changes for ROW_FORMAT=DYNAMIC and
ROW_FORMAT=COMPACT are as contributed. (With this change present,
crash recovery from MariaDB 10.3.1 will fail in spectacular ways!)
Also the semantics of higher-level redo log records that modify the
PAGE_INSTANT field is changed. The redo log format version identifier
was already changed to LOG_HEADER_FORMAT_CURRENT=103 in MariaDB 10.3.1.
Everything else has been rewritten by me. Thanks to Elena Stepanova,
the code has been tested extensively.
When rolling back an instant ADD COLUMN operation, we must empty the
PAGE_FREE list after deleting or shortening the 'default row' record,
by calling either btr_page_empty() or btr_page_reorganize(). We must
know the size of each entry in the PAGE_FREE list. If rollback left a
freed copy of the 'default row' in the PAGE_FREE list, we would be
unable to determine its size (if it is in ROW_FORMAT=COMPACT or
ROW_FORMAT=DYNAMIC) because it would contain more fields than the
rolled-back definition of the clustered index.
UNIV_SQL_DEFAULT: A new special constant that designates an instantly
added column that is not present in the clustered index record.
len_is_stored(): Check if a length is an actual length. There are
two magic length values: UNIV_SQL_DEFAULT, UNIV_SQL_NULL.
dict_col_t::def_val: The 'default row' value of the column. If the
column is not added instantly, def_val.len will be UNIV_SQL_DEFAULT.
dict_col_t: Add the accessors is_virtual(), is_nullable(), is_instant(),
instant_value().
dict_col_t::remove_instant(): Remove the 'instant ADD' status of
a column.
dict_col_t::name(const dict_table_t& table): Replaces
dict_table_get_col_name().
dict_index_t::n_core_fields: The original number of fields.
For secondary indexes and if instant ADD COLUMN has not been used,
this will be equal to dict_index_t::n_fields.
dict_index_t::n_core_null_bytes: Number of bytes needed to
represent the null flags; usually equal to UT_BITS_IN_BYTES(n_nullable).
dict_index_t::NO_CORE_NULL_BYTES: Magic value signalling that
n_core_null_bytes was not initialized yet from the clustered index
root page.
dict_index_t: Add the accessors is_instant(), is_clust(),
get_n_nullable(), instant_field_value().
dict_index_t::instant_add_field(): Adjust clustered index metadata
for instant ADD COLUMN.
dict_index_t::remove_instant(): Remove the 'instant ADD' status
of a clustered index when the table becomes empty, or the very first
instant ADD COLUMN operation is rolled back.
dict_table_t: Add the accessors is_instant(), is_temporary(),
supports_instant().
dict_table_t::instant_add_column(): Adjust metadata for
instant ADD COLUMN.
dict_table_t::rollback_instant(): Adjust metadata on the rollback
of instant ADD COLUMN.
prepare_inplace_alter_table_dict(): First create the ctx->new_table,
and only then decide if the table really needs to be rebuilt.
We must split the creation of table or index metadata from the
creation of the dictionary table records and the creation of
the data. In this way, we can transform a table-rebuilding operation
into an instant ADD COLUMN operation. Dictionary objects will only
be added to cache when table rebuilding or index creation is needed.
The ctx->instant_table will never be added to cache.
dict_table_t::add_to_cache(): Modified and renamed from
dict_table_add_to_cache(). Do not modify the table metadata.
Let the callers invoke dict_table_add_system_columns() and if needed,
set can_be_evicted.
dict_create_sys_tables_tuple(), dict_create_table_step(): Omit the
system columns (which will now exist in the dict_table_t object
already at this point).
dict_create_table_step(): Expect the callers to invoke
dict_table_add_system_columns().
pars_create_table(): Before creating the table creation execution
graph, invoke dict_table_add_system_columns().
row_create_table_for_mysql(): Expect all callers to invoke
dict_table_add_system_columns().
create_index_dict(): Replaces row_merge_create_index_graph().
innodb_update_n_cols(): Renamed from innobase_update_n_virtual().
Call my_error() if an error occurs.
btr_cur_instant_init(), btr_cur_instant_init_low(),
btr_cur_instant_root_init():
Load additional metadata from the clustered index and set
dict_index_t::n_core_null_bytes. This is invoked
when table metadata is first loaded into the data dictionary.
dict_boot(): Initialize n_core_null_bytes for the four hard-coded
dictionary tables.
dict_create_index_step(): Initialize n_core_null_bytes. This is
executed as part of CREATE TABLE.
dict_index_build_internal_clust(): Initialize n_core_null_bytes to
NO_CORE_NULL_BYTES if table->supports_instant().
row_create_index_for_mysql(): Initialize n_core_null_bytes for
CREATE TEMPORARY TABLE.
commit_cache_norebuild(): Call the code to rename or enlarge columns
in the cache only if instant ADD COLUMN is not being used.
(Instant ADD COLUMN would copy all column metadata from
instant_table to old_table, including the names and lengths.)
PAGE_INSTANT: A new 13-bit field for storing dict_index_t::n_core_fields.
This is repurposing the 16-bit field PAGE_DIRECTION, of which only the
least significant 3 bits were used. The original byte containing
PAGE_DIRECTION will be accessible via the new constant PAGE_DIRECTION_B.
page_get_instant(), page_set_instant(): Accessors for the PAGE_INSTANT.
page_ptr_get_direction(), page_get_direction(),
page_ptr_set_direction(): Accessors for PAGE_DIRECTION.
page_direction_reset(): Reset PAGE_DIRECTION, PAGE_N_DIRECTION.
page_direction_increment(): Increment PAGE_N_DIRECTION
and set PAGE_DIRECTION.
rec_get_offsets(): Use the 'leaf' parameter for non-debug purposes,
and assume that heap_no is always set.
Initialize all dict_index_t::n_fields for ROW_FORMAT=REDUNDANT records,
even if the record contains fewer fields.
rec_offs_make_valid(): Add the parameter 'leaf'.
rec_copy_prefix_to_dtuple(): Assert that the tuple is only built
on the core fields. Instant ADD COLUMN only applies to the
clustered index, and we should never build a search key that has
more than the PRIMARY KEY and possibly DB_TRX_ID,DB_ROLL_PTR.
All these columns are always present.
dict_index_build_data_tuple(): Remove assertions that would be
duplicated in rec_copy_prefix_to_dtuple().
rec_init_offsets(): Support ROW_FORMAT=REDUNDANT records whose
number of fields is between n_core_fields and n_fields.
cmp_rec_rec_with_match(): Implement the comparison between two
MIN_REC_FLAG records.
trx_t::in_rollback: Make the field available in non-debug builds.
trx_start_for_ddl_low(): Remove dangerous error-tolerance.
A dictionary transaction must be flagged as such before it has generated
any undo log records. This is because trx_undo_assign_undo() will mark
the transaction as a dictionary transaction in the undo log header
right before the very first undo log record is being written.
btr_index_rec_validate(): Account for instant ADD COLUMN
row_undo_ins_remove_clust_rec(): On the rollback of an insert into
SYS_COLUMNS, revert instant ADD COLUMN in the cache by removing the
last column from the table and the clustered index.
row_search_on_row_ref(), row_undo_mod_parse_undo_rec(), row_undo_mod(),
trx_undo_update_rec_get_update(): Handle the 'default row'
as a special case.
dtuple_t::trim(index): Omit a redundant suffix of an index tuple right
before insert or update. After instant ADD COLUMN, if the last fields
of a clustered index tuple match the 'default row', there is no
need to store them. While trimming the entry, we must hold a page latch,
so that the table cannot be emptied and the 'default row' be deleted.
btr_cur_optimistic_update(), btr_cur_pessimistic_update(),
row_upd_clust_rec_by_insert(), row_ins_clust_index_entry_low():
Invoke dtuple_t::trim() if needed.
row_ins_clust_index_entry(): Restore dtuple_t::n_fields after calling
row_ins_clust_index_entry_low().
rec_get_converted_size(), rec_get_converted_size_comp(): Allow the number
of fields to be between n_core_fields and n_fields. Do not support
infimum,supremum. They are never supposed to be stored in dtuple_t,
because page creation nowadays uses a lower-level method for initializing
them.
rec_convert_dtuple_to_rec_comp(): Assign the status bits based on the
number of fields.
btr_cur_trim(): In an update, trim the index entry as needed. For the
'default row', handle rollback specially. For user records, omit
fields that match the 'default row'.
btr_cur_optimistic_delete_func(), btr_cur_pessimistic_delete():
Skip locking and adaptive hash index for the 'default row'.
row_log_table_apply_convert_mrec(): Replace 'default row' values if needed.
In the temporary file that is applied by row_log_table_apply(),
we must identify whether the records contain the extra header for
instantly added columns. For now, we will allocate an additional byte
for this for ROW_T_INSERT and ROW_T_UPDATE records when the source table
has been subject to instant ADD COLUMN. The ROW_T_DELETE records are
fine, as they will be converted and will only contain 'core' columns
(PRIMARY KEY and some system columns) that are converted from dtuple_t.
rec_get_converted_size_temp(), rec_init_offsets_temp(),
rec_convert_dtuple_to_temp(): Add the parameter 'status'.
REC_INFO_DEFAULT_ROW = REC_INFO_MIN_REC_FLAG | REC_STATUS_COLUMNS_ADDED:
An info_bits constant for distinguishing the 'default row' record.
rec_comp_status_t: An enum of the status bit values.
rec_leaf_format: An enum that replaces the bool parameter of
rec_init_offsets_comp_ordinary().
2017-10-06 07:00:05 +03:00
|
|
|
ALTER TABLE t1 FORCE, ADD COLUMN k4 int;
|
2017-09-14 16:07:10 +03:00
|
|
|
connection default;
|
|
|
|
SET DEBUG_SYNC= 'now WAIT_FOR opened';
|
|
|
|
SET debug = '+d,row_log_tmpfile_fail';
|
|
|
|
INSERT INTO t1 select NULL,'aaa','bbb' from t480;
|
|
|
|
INSERT INTO t1 select NULL,'aaaa','bbbb' from t480;
|
|
|
|
SET DEBUG_SYNC= 'now SIGNAL flushed';
|
|
|
|
SET DEBUG_DBUG = @saved_debug_dbug;
|
|
|
|
connection con1;
|
|
|
|
/*con1 reap*/ ALTER TABLE t1 ADD COLUMN k4 int;
|
2018-01-29 20:04:13 +00:00
|
|
|
ERROR HY000: Out of memory.
|
2017-09-14 16:07:10 +03:00
|
|
|
SELECT COUNT(k1),k2,k3 FROM t1 GROUP BY k2,k3;
|
|
|
|
COUNT(k1) k2 k3
|
|
|
|
480 aaa bbb
|
|
|
|
480 aaaa bbbb
|
2021-11-12 17:46:35 +05:30
|
|
|
1 test test
|
2017-09-14 16:07:10 +03:00
|
|
|
disconnect con1;
|
|
|
|
connection default;
|
|
|
|
show create table t1;
|
|
|
|
Table Create Table
|
|
|
|
t1 CREATE TABLE `t1` (
|
|
|
|
`k1` int(11) NOT NULL AUTO_INCREMENT,
|
|
|
|
`k2` char(200) DEFAULT NULL,
|
|
|
|
`k3` char(200) DEFAULT NULL,
|
|
|
|
PRIMARY KEY (`k1`)
|
2022-09-13 16:36:38 +03:00
|
|
|
) ENGINE=InnoDB AUTO_INCREMENT=1024 DEFAULT CHARSET=latin1 COLLATE=latin1_swedish_ci
|
2017-09-14 16:07:10 +03:00
|
|
|
drop table t1;
|
|
|
|
drop table t480;
|
2017-12-21 17:19:13 +02:00
|
|
|
#
|
|
|
|
# MDEV-12827 Assertion failure when reporting duplicate key error
|
|
|
|
# in online table rebuild
|
|
|
|
#
|
MDEV-15250 UPSERT during ALTER TABLE results in 'Duplicate entry' error for alter
- InnoDB DDL results in `Duplicate entry' if concurrent DML throws
duplicate key error. The following scenario explains the problem
connection con1:
ALTER TABLE t1 FORCE;
connection con2:
INSERT INTO t1(pk, uk) VALUES (2, 2), (3, 2);
In connection con2, InnoDB throws the 'DUPLICATE KEY' error because
of unique index. Alter operation will throw the error when applying
the concurrent DML log.
- Inserting the duplicate key for unique index logs the insert
operation for online ALTER TABLE. When insertion fails,
transaction does rollback and it leads to logging of
delete operation for online ALTER TABLE.
While applying the insert log entries, alter operation
encounters 'DUPLICATE KEY' error.
- To avoid the above fake duplicate scenario, InnoDB should
not write any log for online ALTER TABLE before DML transaction
commit.
- User thread which does DML can apply the online log if
InnoDB ran out of online log and index is marked as completed.
Set online log error if apply phase encountered any error.
It can also clear all other indexes log, marks the newly
added indexes as corrupted.
- Removed the old online code which was a part of DML operations
commit_inplace_alter_table() : Does apply the online log
for the last batch of secondary index log and does frees
the log for the completed index.
trx_t::apply_online_log: Set to true while writing the undo
log if the modified table has active DDL
trx_t::apply_log(): Apply the DML changes to online DDL tables
dict_table_t::is_active_ddl(): Returns true if the table
has an active DDL
dict_index_t::online_log_make_dummy(): Assign dummy value
for clustered index online log to indicate the secondary
indexes are being rebuild.
dict_index_t::online_log_is_dummy(): Check whether the online
log has dummy value
ha_innobase_inplace_ctx::log_failure(): Handle the apply log
failure for online DDL transaction
row_log_mark_other_online_index_abort(): Clear out all other
online index log after encountering the error during
row_log_apply()
row_log_get_error(): Get the error happened during row_log_apply()
row_log_online_op(): Does apply the online log if index is
completed and ran out of memory. Returns false if apply log fails
UndorecApplier: Introduced a class to maintain the undo log
record, latched undo buffer page, parse the undo log record,
maintain the undo record type, info bits and update vector
UndorecApplier::get_old_rec(): Get the correct version of the
clustered index record that was modified by the current undo
log record
UndorecApplier::clear_undo_rec(): Clear the undo log related
information after applying the undo log record
UndorecApplier::log_update(): Handle the update, delete undo
log and apply it on online indexes
UndorecApplier::log_insert(): Handle the insert undo log
and apply it on online indexes
UndorecApplier::is_same(): Check whether the given roll pointer
is generated by the current undo log record information
trx_t::rollback_low(): Set apply_online_log for the transaction
after partially rollbacked transaction has any active DDL
prepare_inplace_alter_table_dict(): After allocating the online
log, InnoDB does create fulltext common tables. Fulltext index
doesn't allow the index to be online. So removed the dead
code of online log removal
Thanks to Marko Mäkelä for providing the initial prototype and
Matthias Leich for testing the issue patiently.
2022-04-25 13:36:56 +05:30
|
|
|
CREATE TABLE t1 (j INT UNIQUE, i INT) ENGINE=InnoDB;
|
2021-11-12 17:46:35 +05:30
|
|
|
INSERT INTO t1 VALUES(2, 2);
|
2017-12-22 12:23:39 +02:00
|
|
|
connect con1,localhost,root,,test;
|
2017-12-21 17:19:13 +02:00
|
|
|
SET DEBUG_SYNC='row_log_table_apply1_before SIGNAL built WAIT_FOR log';
|
MDEV-15250 UPSERT during ALTER TABLE results in 'Duplicate entry' error for alter
- InnoDB DDL results in `Duplicate entry' if concurrent DML throws
duplicate key error. The following scenario explains the problem
connection con1:
ALTER TABLE t1 FORCE;
connection con2:
INSERT INTO t1(pk, uk) VALUES (2, 2), (3, 2);
In connection con2, InnoDB throws the 'DUPLICATE KEY' error because
of unique index. Alter operation will throw the error when applying
the concurrent DML log.
- Inserting the duplicate key for unique index logs the insert
operation for online ALTER TABLE. When insertion fails,
transaction does rollback and it leads to logging of
delete operation for online ALTER TABLE.
While applying the insert log entries, alter operation
encounters 'DUPLICATE KEY' error.
- To avoid the above fake duplicate scenario, InnoDB should
not write any log for online ALTER TABLE before DML transaction
commit.
- User thread which does DML can apply the online log if
InnoDB ran out of online log and index is marked as completed.
Set online log error if apply phase encountered any error.
It can also clear all other indexes log, marks the newly
added indexes as corrupted.
- Removed the old online code which was a part of DML operations
commit_inplace_alter_table() : Does apply the online log
for the last batch of secondary index log and does frees
the log for the completed index.
trx_t::apply_online_log: Set to true while writing the undo
log if the modified table has active DDL
trx_t::apply_log(): Apply the DML changes to online DDL tables
dict_table_t::is_active_ddl(): Returns true if the table
has an active DDL
dict_index_t::online_log_make_dummy(): Assign dummy value
for clustered index online log to indicate the secondary
indexes are being rebuild.
dict_index_t::online_log_is_dummy(): Check whether the online
log has dummy value
ha_innobase_inplace_ctx::log_failure(): Handle the apply log
failure for online DDL transaction
row_log_mark_other_online_index_abort(): Clear out all other
online index log after encountering the error during
row_log_apply()
row_log_get_error(): Get the error happened during row_log_apply()
row_log_online_op(): Does apply the online log if index is
completed and ran out of memory. Returns false if apply log fails
UndorecApplier: Introduced a class to maintain the undo log
record, latched undo buffer page, parse the undo log record,
maintain the undo record type, info bits and update vector
UndorecApplier::get_old_rec(): Get the correct version of the
clustered index record that was modified by the current undo
log record
UndorecApplier::clear_undo_rec(): Clear the undo log related
information after applying the undo log record
UndorecApplier::log_update(): Handle the update, delete undo
log and apply it on online indexes
UndorecApplier::log_insert(): Handle the insert undo log
and apply it on online indexes
UndorecApplier::is_same(): Check whether the given roll pointer
is generated by the current undo log record information
trx_t::rollback_low(): Set apply_online_log for the transaction
after partially rollbacked transaction has any active DDL
prepare_inplace_alter_table_dict(): After allocating the online
log, InnoDB does create fulltext common tables. Fulltext index
doesn't allow the index to be online. So removed the dead
code of online log removal
Thanks to Marko Mäkelä for providing the initial prototype and
Matthias Leich for testing the issue patiently.
2022-04-25 13:36:56 +05:30
|
|
|
ALTER TABLE t1 DROP j, ADD UNIQUE INDEX(i), FORCE;
|
2017-12-22 12:23:39 +02:00
|
|
|
connection default;
|
2017-12-21 17:19:13 +02:00
|
|
|
SET DEBUG_SYNC='now WAIT_FOR built';
|
2021-02-05 08:32:57 +02:00
|
|
|
SET DEBUG_DBUG='+d,row_ins_row_level';
|
2017-12-21 17:19:13 +02:00
|
|
|
INSERT INTO t1 (i) VALUES (0),(0);
|
|
|
|
SET DEBUG_SYNC='now SIGNAL log';
|
2021-02-05 08:32:57 +02:00
|
|
|
SET DEBUG_DBUG=@saved_debug_dbug;
|
2017-12-22 12:23:39 +02:00
|
|
|
connection con1;
|
2017-12-21 17:19:13 +02:00
|
|
|
ERROR 23000: Duplicate entry '0' for key 'i'
|
MDEV-15250 UPSERT during ALTER TABLE results in 'Duplicate entry' error for alter
- InnoDB DDL results in `Duplicate entry' if concurrent DML throws
duplicate key error. The following scenario explains the problem
connection con1:
ALTER TABLE t1 FORCE;
connection con2:
INSERT INTO t1(pk, uk) VALUES (2, 2), (3, 2);
In connection con2, InnoDB throws the 'DUPLICATE KEY' error because
of unique index. Alter operation will throw the error when applying
the concurrent DML log.
- Inserting the duplicate key for unique index logs the insert
operation for online ALTER TABLE. When insertion fails,
transaction does rollback and it leads to logging of
delete operation for online ALTER TABLE.
While applying the insert log entries, alter operation
encounters 'DUPLICATE KEY' error.
- To avoid the above fake duplicate scenario, InnoDB should
not write any log for online ALTER TABLE before DML transaction
commit.
- User thread which does DML can apply the online log if
InnoDB ran out of online log and index is marked as completed.
Set online log error if apply phase encountered any error.
It can also clear all other indexes log, marks the newly
added indexes as corrupted.
- Removed the old online code which was a part of DML operations
commit_inplace_alter_table() : Does apply the online log
for the last batch of secondary index log and does frees
the log for the completed index.
trx_t::apply_online_log: Set to true while writing the undo
log if the modified table has active DDL
trx_t::apply_log(): Apply the DML changes to online DDL tables
dict_table_t::is_active_ddl(): Returns true if the table
has an active DDL
dict_index_t::online_log_make_dummy(): Assign dummy value
for clustered index online log to indicate the secondary
indexes are being rebuild.
dict_index_t::online_log_is_dummy(): Check whether the online
log has dummy value
ha_innobase_inplace_ctx::log_failure(): Handle the apply log
failure for online DDL transaction
row_log_mark_other_online_index_abort(): Clear out all other
online index log after encountering the error during
row_log_apply()
row_log_get_error(): Get the error happened during row_log_apply()
row_log_online_op(): Does apply the online log if index is
completed and ran out of memory. Returns false if apply log fails
UndorecApplier: Introduced a class to maintain the undo log
record, latched undo buffer page, parse the undo log record,
maintain the undo record type, info bits and update vector
UndorecApplier::get_old_rec(): Get the correct version of the
clustered index record that was modified by the current undo
log record
UndorecApplier::clear_undo_rec(): Clear the undo log related
information after applying the undo log record
UndorecApplier::log_update(): Handle the update, delete undo
log and apply it on online indexes
UndorecApplier::log_insert(): Handle the insert undo log
and apply it on online indexes
UndorecApplier::is_same(): Check whether the given roll pointer
is generated by the current undo log record information
trx_t::rollback_low(): Set apply_online_log for the transaction
after partially rollbacked transaction has any active DDL
prepare_inplace_alter_table_dict(): After allocating the online
log, InnoDB does create fulltext common tables. Fulltext index
doesn't allow the index to be online. So removed the dead
code of online log removal
Thanks to Marko Mäkelä for providing the initial prototype and
Matthias Leich for testing the issue patiently.
2022-04-25 13:36:56 +05:30
|
|
|
DELETE FROM t1;
|
|
|
|
ALTER TABLE t1 ADD UNIQUE INDEX(i);
|
2017-12-21 17:19:13 +02:00
|
|
|
SET DEBUG_SYNC='row_log_table_apply1_before SIGNAL built2 WAIT_FOR log2';
|
|
|
|
ALTER TABLE t1 DROP j, FORCE;
|
2017-12-22 12:23:39 +02:00
|
|
|
connection default;
|
2017-12-21 17:19:13 +02:00
|
|
|
SET DEBUG_SYNC='now WAIT_FOR built2';
|
|
|
|
INSERT INTO t1 (i) VALUES (0),(1);
|
|
|
|
UPDATE t1 SET i=0;
|
|
|
|
ERROR 23000: Duplicate entry '0' for key 'i'
|
|
|
|
SET DEBUG_SYNC='now SIGNAL log2';
|
2017-12-22 12:23:39 +02:00
|
|
|
connection con1;
|
|
|
|
disconnect con1;
|
2023-10-25 10:08:20 +03:00
|
|
|
disconnect stop_purge;
|
2017-12-22 12:23:39 +02:00
|
|
|
connection default;
|
2017-12-21 17:19:13 +02:00
|
|
|
SET DEBUG_SYNC='RESET';
|
|
|
|
DROP TABLE t1;
|
2017-09-14 16:07:10 +03:00
|
|
|
SET DEBUG_SYNC='RESET';
|
|
|
|
#
|
|
|
|
# BUG#21612714 ALTER TABLE SORTING SKIPPED WHEN CHANGE PK AND DROP
|
|
|
|
# LAST COLUMN OF OLD PK
|
|
|
|
#
|
|
|
|
SET DEBUG_DBUG = '+d,innodb_alter_table_pk_assert_no_sort';
|
|
|
|
create table t1(o1 varchar(10), primary key(o1(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd'), ('acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(3)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), primary key(o1(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd'), ('acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), primary key(o1(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd'), ('acd');
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1(3), n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), primary key(o1(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd'), ('acd');
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1, n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int not null, primary key(o1(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1(3), o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int not null, primary key(o1(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1, o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), primary key(o1(3))) engine = innodb;
|
|
|
|
insert into t1 values('abd'), ('acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values('abd'), ('acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, primary key(o1(3), o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 drop primary key, add primary key(o1(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, primary key(o1, o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 drop primary key, add primary key(o1(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, primary key(o1(3), o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1(2),n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, primary key(o1, o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1(2),n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, primary key(o1(3), o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1(3),n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, primary key(o1, o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1), ('acd', 2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 varchar(10), primary key(o1,o2(3))) engine = innodb;
|
|
|
|
insert into t1 values(1,'abd'), (2,'acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 varchar(10), primary key(o1,o2)) engine = innodb;
|
|
|
|
insert into t1 values(1,'abd'), (2,'acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 varchar(10), primary key(o1,o2(2))) engine = innodb;
|
|
|
|
insert into t1 values(1, 'abd'), (2, 'acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2(3)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 varchar(10), primary key(o1,o2(2))) engine = innodb;
|
|
|
|
insert into t1 values(1, 'abd'), (2, 'acd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 varchar(10), o3 int, primary key(o1,o2(3),o3)) engine = innodb;
|
|
|
|
insert into t1 values(1, 'abd', 1), (2, 'acd', 2);
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 varchar(10), o3 int, primary key(o1,o2,o3)) engine = innodb;
|
|
|
|
insert into t1 values(1, 'abd', 1), (2, 'acd', 2);
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 varchar(10), primary key(o1(3),o2(3))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 'acd'), ('acd', 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(3),o2(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 varchar(10), primary key(o1,o2)) engine = innodb;
|
|
|
|
insert into t1 values('abd', 'acd'), ('acd', 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 varchar(10), primary key(o1(3),o2(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 'acd'), ('acd', 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(3),o2(3)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 varchar(10), primary key(o1,o2(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 'acd'), ('acd', 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, o3 varchar(10), primary key(o1(3),o2,o3(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1, 'acd'), ('acd', 2, 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(3),o2,o3(3)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, o3 varchar(10), primary key(o1,o2,o3(2))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1, 'acd'), ('acd', 2, 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2,o3), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, o3 varchar(10), primary key(o1(3),o2,o3(3))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1, 'acd'), ('acd', 2, 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1(3),o2,o3(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 varchar(10), o2 int, o3 varchar(10), primary key(o1,o2,o3(3))) engine = innodb;
|
|
|
|
insert into t1 values('abd', 1, 'acd'), ('acd', 2, 'abd');
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2,o3(2)), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(n1,o1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, add n2 int not null, drop primary key, add primary key(n1,o1,n2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, add n2 int not null, drop primary key, add primary key(n1,n2,o1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, add n2 int not null, drop primary key, add primary key(o1,n1,n2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,o2,n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,n1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,2);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(n1,o1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, o3 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2),(2,2,1);
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2,o3), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int not null, o3 int not null, primary key(o1)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2),(2,2,1);
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o3,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int, o3 int, primary key(o1,o2,o3)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2),(2,2,1);
|
|
|
|
alter table t1 drop primary key, add primary key(o1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int, o3 int, o4 int not null, primary key(o1,o2,o3)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2,2),(2,2,1,1);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,o2,o3,o4), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int, o3 int, primary key(o1,o2,o3)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2),(2,2,1);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,o2,n1), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int, o3 int, primary key(o1,o2,o3)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2),(2,2,1);
|
|
|
|
alter table t1 add n1 int not null, drop primary key, add primary key(o1,n1,o2), lock=none;
|
|
|
|
drop table t1;
|
|
|
|
create table t1(o1 int, o2 int, o3 int, primary key(o1,o2,o3)) engine = innodb;
|
|
|
|
insert into t1 values(1,1,2),(2,2,1);
|
|
|
|
alter table t1 drop primary key, add primary key(o1), lock=none;
|
|
|
|
drop table t1;
|
2020-08-11 18:52:38 +03:00
|
|
|
create table t1(o1 int, o2 int, primary key(o1,o2)) engine = innodb;
|
|
|
|
insert into t1 values(1,1),(2,1);
|
|
|
|
alter table t1 drop primary key, add column a int unique auto_increment,
|
|
|
|
add primary key(o1,o2,a), algorithm=inplace;
|
|
|
|
drop table t1;
|
2017-09-14 16:07:10 +03:00
|
|
|
SET DEBUG_DBUG = @saved_debug_dbug;
|