mirror of
https://github.com/MariaDB/server.git
synced 2025-01-16 03:52:35 +01:00
Merge 10.3 into 10.4
This commit is contained in:
commit
589cf8dbf3
100 changed files with 2274 additions and 387 deletions
|
@ -864,21 +864,14 @@ datafile_rsync_backup(const char *filepath, bool save_to_list, FILE *f)
|
|||
return(true);
|
||||
}
|
||||
|
||||
|
||||
static
|
||||
bool
|
||||
backup_file_vprintf(const char *filename, const char *fmt, va_list ap)
|
||||
bool backup_file_print_buf(const char *filename, const char *buf, int buf_len)
|
||||
{
|
||||
ds_file_t *dstfile = NULL;
|
||||
MY_STAT stat; /* unused for now */
|
||||
char *buf = 0;
|
||||
int buf_len;
|
||||
const char *action;
|
||||
|
||||
memset(&stat, 0, sizeof(stat));
|
||||
|
||||
buf_len = vasprintf(&buf, fmt, ap);
|
||||
|
||||
stat.st_size = buf_len;
|
||||
stat.st_mtime = my_time(0);
|
||||
|
||||
|
@ -902,7 +895,6 @@ backup_file_vprintf(const char *filename, const char *fmt, va_list ap)
|
|||
|
||||
/* close */
|
||||
msg(" ...done");
|
||||
free(buf);
|
||||
|
||||
if (ds_close(dstfile)) {
|
||||
goto error_close;
|
||||
|
@ -911,7 +903,6 @@ backup_file_vprintf(const char *filename, const char *fmt, va_list ap)
|
|||
return(true);
|
||||
|
||||
error:
|
||||
free(buf);
|
||||
if (dstfile != NULL) {
|
||||
ds_close(dstfile);
|
||||
}
|
||||
|
@ -919,8 +910,21 @@ error:
|
|||
error_close:
|
||||
msg("Error: backup file failed.");
|
||||
return(false); /*ERROR*/
|
||||
}
|
||||
|
||||
return true;
|
||||
};
|
||||
|
||||
static
|
||||
bool
|
||||
backup_file_vprintf(const char *filename, const char *fmt, va_list ap)
|
||||
{
|
||||
char *buf = 0;
|
||||
int buf_len;
|
||||
buf_len = vasprintf(&buf, fmt, ap);
|
||||
bool result = backup_file_print_buf(filename, buf, buf_len);
|
||||
free(buf);
|
||||
return result;
|
||||
}
|
||||
|
||||
bool
|
||||
backup_file_printf(const char *filename, const char *fmt, ...)
|
||||
|
@ -1442,7 +1446,7 @@ out:
|
|||
return(ret);
|
||||
}
|
||||
|
||||
void backup_fix_ddl(void);
|
||||
void backup_fix_ddl(CorruptedPages &);
|
||||
|
||||
lsn_t get_current_lsn(MYSQL *connection)
|
||||
{
|
||||
|
@ -1467,7 +1471,7 @@ lsn_t get_current_lsn(MYSQL *connection)
|
|||
lsn_t server_lsn_after_lock;
|
||||
extern void backup_wait_for_lsn(lsn_t lsn);
|
||||
/** Start --backup */
|
||||
bool backup_start()
|
||||
bool backup_start(CorruptedPages &corrupted_pages)
|
||||
{
|
||||
if (!opt_no_lock) {
|
||||
if (opt_safe_slave_backup) {
|
||||
|
@ -1502,7 +1506,7 @@ bool backup_start()
|
|||
|
||||
msg("Waiting for log copy thread to read lsn %llu", (ulonglong)server_lsn_after_lock);
|
||||
backup_wait_for_lsn(server_lsn_after_lock);
|
||||
backup_fix_ddl();
|
||||
backup_fix_ddl(corrupted_pages);
|
||||
|
||||
// There is no need to stop slave thread before coping non-Innodb data when
|
||||
// --no-lock option is used because --no-lock option requires that no DDL or
|
||||
|
|
|
@ -33,7 +33,7 @@ copy_file(ds_ctxt_t *datasink,
|
|||
uint thread_n);
|
||||
|
||||
/** Start --backup */
|
||||
bool backup_start();
|
||||
bool backup_start(CorruptedPages &corrupted_pages);
|
||||
/** Release resources after backup_start() */
|
||||
void backup_release();
|
||||
/** Finish after backup_start() and backup_release() */
|
||||
|
@ -51,5 +51,6 @@ directory_exists(const char *dir, bool create);
|
|||
|
||||
lsn_t
|
||||
get_current_lsn(MYSQL *connection);
|
||||
bool backup_file_print_buf(const char *filename, const char *buf, int buf_len);
|
||||
|
||||
#endif
|
||||
|
|
32
extra/mariabackup/backup_debug.h
Normal file
32
extra/mariabackup/backup_debug.h
Normal file
|
@ -0,0 +1,32 @@
|
|||
#pragma once
|
||||
#include "my_dbug.h"
|
||||
#ifndef DBUG_OFF
|
||||
extern char *dbug_mariabackup_get_val(const char *event, const char *key);
|
||||
/*
|
||||
In debug mode, execute SQL statement that was passed via environment.
|
||||
To use this facility, you need to
|
||||
|
||||
1. Add code DBUG_EXECUTE_MARIABACKUP_EVENT("my_event_name", key););
|
||||
to the code. key is usually a table name
|
||||
2. Set environment variable my_event_name_$key SQL statement you want to execute
|
||||
when event occurs, in DBUG_EXECUTE_IF from above.
|
||||
In mtr , you can set environment via 'let' statement (do not use $ as the first char
|
||||
for the variable)
|
||||
3. start mariabackup with --dbug=+d,debug_mariabackup_events
|
||||
*/
|
||||
extern void dbug_mariabackup_event(
|
||||
const char *event,const char *key);
|
||||
#define DBUG_MARIABACKUP_EVENT(A, B) \
|
||||
DBUG_EXECUTE_IF("mariabackup_events", \
|
||||
dbug_mariabackup_event(A,B););
|
||||
#define DBUG_EXECUTE_FOR_KEY(EVENT, KEY, CODE) \
|
||||
DBUG_EXECUTE_IF("mariabackup_inject_code", {\
|
||||
char *dbug_val = dbug_mariabackup_get_val(EVENT, KEY); \
|
||||
if (dbug_val && *dbug_val) CODE \
|
||||
})
|
||||
#else
|
||||
#define DBUG_MARIABACKUP_EVENT(A,B)
|
||||
#define DBUG_MARIABACKUP_EVENT_LOCK(A,B)
|
||||
#define DBUG_EXECUTE_FOR_KEY(EVENT, KEY, CODE)
|
||||
#endif
|
||||
|
|
@ -18,7 +18,6 @@
|
|||
#include <mysql.h>
|
||||
#include <xtrabackup.h>
|
||||
#include <encryption_plugin.h>
|
||||
#include <backup_copy.h>
|
||||
#include <sql_plugin.h>
|
||||
#include <sstream>
|
||||
#include <vector>
|
||||
|
|
|
@ -35,6 +35,7 @@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
#include "common.h"
|
||||
#include "read_filt.h"
|
||||
#include "xtrabackup.h"
|
||||
#include "backup_debug.h"
|
||||
|
||||
/* Size of read buffer in pages (640 pages = 10M for 16K sized pages) */
|
||||
#define XB_FIL_CUR_PAGES 640
|
||||
|
@ -375,16 +376,15 @@ static bool page_is_corrupted(const byte *page, ulint page_no,
|
|||
return buf_page_is_corrupted(true, page, space->flags);
|
||||
}
|
||||
|
||||
/************************************************************************
|
||||
Reads and verifies the next block of pages from the source
|
||||
/** Reads and verifies the next block of pages from the source
|
||||
file. Positions the cursor after the last read non-corrupted page.
|
||||
|
||||
@param[in,out] cursor source file cursor
|
||||
@param[out] corrupted_pages adds corrupted pages if
|
||||
opt_log_innodb_page_corruption is set
|
||||
@return XB_FIL_CUR_SUCCESS if some have been read successfully, XB_FIL_CUR_EOF
|
||||
if there are no more pages to read and XB_FIL_CUR_ERROR on error. */
|
||||
xb_fil_cur_result_t
|
||||
xb_fil_cur_read(
|
||||
/*============*/
|
||||
xb_fil_cur_t* cursor) /*!< in/out: source file cursor */
|
||||
xb_fil_cur_result_t xb_fil_cur_read(xb_fil_cur_t* cursor,
|
||||
CorruptedPages &corrupted_pages)
|
||||
{
|
||||
byte* page;
|
||||
ulint i;
|
||||
|
@ -458,20 +458,40 @@ read_retry:
|
|||
retry_count--;
|
||||
|
||||
if (retry_count == 0) {
|
||||
const char *ignore_corruption_warn = opt_log_innodb_page_corruption ?
|
||||
" WARNING!!! The corruption is ignored due to"
|
||||
" log-innodb-page-corruption option, the backup can contain"
|
||||
" corrupted data." : "";
|
||||
msg(cursor->thread_n,
|
||||
"Error: failed to read page after "
|
||||
"10 retries. File %s seems to be "
|
||||
"corrupted.", cursor->abs_path);
|
||||
ret = XB_FIL_CUR_ERROR;
|
||||
"corrupted.%s", cursor->abs_path, ignore_corruption_warn);
|
||||
ut_print_buf(stderr, page, page_size);
|
||||
break;
|
||||
if (opt_log_innodb_page_corruption) {
|
||||
corrupted_pages.add_page(cursor->node->name, cursor->node->space->id,
|
||||
page_no);
|
||||
retry_count = 1;
|
||||
}
|
||||
else {
|
||||
ret = XB_FIL_CUR_ERROR;
|
||||
break;
|
||||
}
|
||||
}
|
||||
else {
|
||||
msg(cursor->thread_n, "Database page corruption detected at page "
|
||||
ULINTPF ", retrying...",
|
||||
page_no);
|
||||
os_thread_sleep(100000);
|
||||
goto read_retry;
|
||||
}
|
||||
msg(cursor->thread_n, "Database page corruption detected at page "
|
||||
ULINTPF ", retrying...",
|
||||
page_no);
|
||||
os_thread_sleep(100000);
|
||||
goto read_retry;
|
||||
}
|
||||
DBUG_EXECUTE_FOR_KEY("add_corrupted_page_for", cursor->node->space->name,
|
||||
{
|
||||
ulint corrupted_page_no = strtoul(dbug_val, NULL, 10);
|
||||
if (page_no == corrupted_page_no)
|
||||
corrupted_pages.add_page(cursor->node->name, cursor->node->space->id,
|
||||
corrupted_page_no);
|
||||
});
|
||||
cursor->buf_read += page_size;
|
||||
cursor->buf_npages++;
|
||||
}
|
||||
|
|
|
@ -29,6 +29,7 @@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
#include "read_filt.h"
|
||||
#include "srv0start.h"
|
||||
#include "srv0srv.h"
|
||||
#include "xtrabackup.h"
|
||||
|
||||
struct xb_fil_cur_t {
|
||||
pfs_os_file_t file; /*!< source file handle */
|
||||
|
@ -89,17 +90,15 @@ xb_fil_cur_open(
|
|||
uint thread_n, /*!< thread number for diagnostics */
|
||||
ulonglong max_file_size = ULLONG_MAX);
|
||||
|
||||
/************************************************************************
|
||||
Reads and verifies the next block of pages from the source
|
||||
/** Reads and verifies the next block of pages from the source
|
||||
file. Positions the cursor after the last read non-corrupted page.
|
||||
|
||||
@param[in,out] cursor source file cursor
|
||||
@param[out] corrupted_pages adds corrupted pages if
|
||||
opt_log_innodb_page_corruption is set
|
||||
@return XB_FIL_CUR_SUCCESS if some have been read successfully, XB_FIL_CUR_EOF
|
||||
if there are no more pages to read and XB_FIL_CUR_ERROR on error. */
|
||||
xb_fil_cur_result_t
|
||||
xb_fil_cur_read(
|
||||
/*============*/
|
||||
xb_fil_cur_t* cursor); /*!< in/out: source file cursor */
|
||||
|
||||
xb_fil_cur_result_t xb_fil_cur_read(xb_fil_cur_t *cursor,
|
||||
CorruptedPages &corrupted_pages);
|
||||
/************************************************************************
|
||||
Close the source file cursor opened with xb_fil_cur_open() and its
|
||||
associated read filter. */
|
||||
|
|
|
@ -27,13 +27,12 @@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
#include "common.h"
|
||||
#include "write_filt.h"
|
||||
#include "fil_cur.h"
|
||||
#include "xtrabackup.h"
|
||||
#include <os0proc.h>
|
||||
|
||||
/************************************************************************
|
||||
Write-through page write filter. */
|
||||
static my_bool wf_wt_init(xb_write_filt_ctxt_t *ctxt, char *dst_name,
|
||||
xb_fil_cur_t *cursor);
|
||||
xb_fil_cur_t *cursor, CorruptedPages *corrupted_pages);
|
||||
static my_bool wf_wt_process(xb_write_filt_ctxt_t *ctxt, ds_file_t *dstfile);
|
||||
|
||||
xb_write_filt_t wf_write_through = {
|
||||
|
@ -46,7 +45,7 @@ xb_write_filt_t wf_write_through = {
|
|||
/************************************************************************
|
||||
Incremental page write filter. */
|
||||
static my_bool wf_incremental_init(xb_write_filt_ctxt_t *ctxt, char *dst_name,
|
||||
xb_fil_cur_t *cursor);
|
||||
xb_fil_cur_t *cursor, CorruptedPages *corrupted_pages);
|
||||
static my_bool wf_incremental_process(xb_write_filt_ctxt_t *ctxt,
|
||||
ds_file_t *dstfile);
|
||||
static my_bool wf_incremental_finalize(xb_write_filt_ctxt_t *ctxt,
|
||||
|
@ -66,11 +65,11 @@ Initialize incremental page write filter.
|
|||
@return TRUE on success, FALSE on error. */
|
||||
static my_bool
|
||||
wf_incremental_init(xb_write_filt_ctxt_t *ctxt, char *dst_name,
|
||||
xb_fil_cur_t *cursor)
|
||||
xb_fil_cur_t *cursor, CorruptedPages *corrupted_pages)
|
||||
{
|
||||
char meta_name[FN_REFLEN];
|
||||
xb_wf_incremental_ctxt_t *cp =
|
||||
&(ctxt->u.wf_incremental_ctxt);
|
||||
&(ctxt->wf_incremental_ctxt);
|
||||
|
||||
ctxt->cursor = cursor;
|
||||
|
||||
|
@ -101,7 +100,9 @@ wf_incremental_init(xb_write_filt_ctxt_t *ctxt, char *dst_name,
|
|||
strcat(dst_name, ".delta");
|
||||
|
||||
mach_write_to_4(cp->delta_buf, 0x78747261UL); /*"xtra"*/
|
||||
|
||||
cp->npages = 1;
|
||||
cp->corrupted_pages = corrupted_pages;
|
||||
|
||||
return(TRUE);
|
||||
}
|
||||
|
@ -117,15 +118,16 @@ wf_incremental_process(xb_write_filt_ctxt_t *ctxt, ds_file_t *dstfile)
|
|||
xb_fil_cur_t *cursor = ctxt->cursor;
|
||||
byte *page;
|
||||
const ulint page_size = cursor->page_size;
|
||||
xb_wf_incremental_ctxt_t *cp = &(ctxt->u.wf_incremental_ctxt);
|
||||
xb_wf_incremental_ctxt_t *cp = &(ctxt->wf_incremental_ctxt);
|
||||
|
||||
for (i = 0, page = cursor->buf; i < cursor->buf_npages;
|
||||
i++, page += page_size) {
|
||||
|
||||
if (incremental_lsn >= mach_read_from_8(page + FIL_PAGE_LSN)) {
|
||||
|
||||
if ((!cp->corrupted_pages ||
|
||||
!cp->corrupted_pages->contains(cursor->node->space->id,
|
||||
cursor->buf_page_no + i)) &&
|
||||
incremental_lsn >= mach_read_from_8(page + FIL_PAGE_LSN))
|
||||
continue;
|
||||
}
|
||||
|
||||
/* updated page */
|
||||
if (cp->npages == page_size / 4) {
|
||||
|
@ -162,7 +164,7 @@ wf_incremental_finalize(xb_write_filt_ctxt_t *ctxt, ds_file_t *dstfile)
|
|||
{
|
||||
xb_fil_cur_t *cursor = ctxt->cursor;
|
||||
const ulint page_size = cursor->page_size;
|
||||
xb_wf_incremental_ctxt_t *cp = &(ctxt->u.wf_incremental_ctxt);
|
||||
xb_wf_incremental_ctxt_t *cp = &(ctxt->wf_incremental_ctxt);
|
||||
|
||||
if (cp->npages != page_size / 4) {
|
||||
mach_write_to_4(cp->delta_buf + cp->npages * 4, 0xFFFFFFFFUL);
|
||||
|
@ -184,7 +186,7 @@ Free the incremental page write filter's buffer. */
|
|||
static void
|
||||
wf_incremental_deinit(xb_write_filt_ctxt_t *ctxt)
|
||||
{
|
||||
xb_wf_incremental_ctxt_t *cp = &(ctxt->u.wf_incremental_ctxt);
|
||||
xb_wf_incremental_ctxt_t *cp = &(ctxt->wf_incremental_ctxt);
|
||||
os_mem_free_large(cp->delta_buf, cp->delta_buf_size);
|
||||
}
|
||||
|
||||
|
@ -194,7 +196,7 @@ Initialize the write-through page write filter.
|
|||
@return TRUE on success, FALSE on error. */
|
||||
static my_bool
|
||||
wf_wt_init(xb_write_filt_ctxt_t *ctxt, char *dst_name __attribute__((unused)),
|
||||
xb_fil_cur_t *cursor)
|
||||
xb_fil_cur_t *cursor, CorruptedPages *)
|
||||
{
|
||||
ctxt->cursor = cursor;
|
||||
|
||||
|
|
|
@ -27,26 +27,26 @@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
|
||||
#include "fil_cur.h"
|
||||
#include "datasink.h"
|
||||
#include "xtrabackup.h"
|
||||
|
||||
/* Incremental page filter context */
|
||||
typedef struct {
|
||||
ulint delta_buf_size;
|
||||
byte *delta_buf;
|
||||
ulint npages;
|
||||
CorruptedPages *corrupted_pages;
|
||||
} xb_wf_incremental_ctxt_t;
|
||||
|
||||
/* Page filter context used as an opaque structure by callers */
|
||||
typedef struct {
|
||||
xb_fil_cur_t *cursor;
|
||||
union {
|
||||
xb_wf_incremental_ctxt_t wf_incremental_ctxt;
|
||||
} u;
|
||||
xb_wf_incremental_ctxt_t wf_incremental_ctxt;
|
||||
} xb_write_filt_ctxt_t;
|
||||
|
||||
|
||||
typedef struct {
|
||||
my_bool (*init)(xb_write_filt_ctxt_t *ctxt, char *dst_name,
|
||||
xb_fil_cur_t *cursor);
|
||||
xb_fil_cur_t *cursor, CorruptedPages *corrupted_pages);
|
||||
my_bool (*process)(xb_write_filt_ctxt_t *ctxt, ds_file_t *dstfile);
|
||||
my_bool (*finalize)(xb_write_filt_ctxt_t *, ds_file_t *dstfile);
|
||||
void (*deinit)(xb_write_filt_ctxt_t *);
|
||||
|
|
|
@ -77,6 +77,7 @@ Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
#include <list>
|
||||
#include <sstream>
|
||||
#include <set>
|
||||
#include <fstream>
|
||||
#include <mysql.h>
|
||||
|
||||
#define G_PTR uchar*
|
||||
|
@ -104,6 +105,9 @@ Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
#include <crc_glue.h>
|
||||
#include <log.h>
|
||||
#include <derror.h>
|
||||
#include "backup_debug.h"
|
||||
|
||||
#define MB_CORRUPTED_PAGES_FILE "innodb_corrupted_pages"
|
||||
|
||||
int sys_var_init();
|
||||
|
||||
|
@ -303,6 +307,7 @@ my_bool opt_noversioncheck = FALSE;
|
|||
my_bool opt_no_backup_locks = FALSE;
|
||||
my_bool opt_decompress = FALSE;
|
||||
my_bool opt_remove_original;
|
||||
my_bool opt_log_innodb_page_corruption;
|
||||
|
||||
my_bool opt_lock_ddl_per_table = FALSE;
|
||||
static my_bool opt_check_privileges;
|
||||
|
@ -367,6 +372,208 @@ struct ddl_tracker_t {
|
|||
|
||||
static ddl_tracker_t ddl_tracker;
|
||||
|
||||
// Convert non-null terminated filename to space name
|
||||
std::string filename_to_spacename(const byte *filename, size_t len);
|
||||
|
||||
CorruptedPages::CorruptedPages() { ut_a(!pthread_mutex_init(&m_mutex, NULL)); }
|
||||
|
||||
CorruptedPages::~CorruptedPages() { ut_a(!pthread_mutex_destroy(&m_mutex)); }
|
||||
|
||||
void CorruptedPages::add_page_no_lock(const char *space_name, ulint space_id,
|
||||
ulint page_no, bool convert_space_name)
|
||||
{
|
||||
space_info_t &space_info = m_spaces[space_id];
|
||||
if (space_info.space_name.empty())
|
||||
space_info.space_name=
|
||||
convert_space_name
|
||||
? filename_to_spacename(reinterpret_cast<const byte *>(space_name),
|
||||
strlen(space_name))
|
||||
: space_name;
|
||||
(void)space_info.pages.insert(page_no);
|
||||
}
|
||||
|
||||
void CorruptedPages::add_page(const char *file_name, ulint space_id,
|
||||
ulint page_no)
|
||||
{
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
add_page_no_lock(file_name, space_id, page_no, true);
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
}
|
||||
|
||||
bool CorruptedPages::contains(ulint space_id, ulint page_no) const
|
||||
{
|
||||
bool result = false;
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
container_t::const_iterator space_it= m_spaces.find(space_id);
|
||||
if (space_it != m_spaces.end())
|
||||
result = space_it->second.pages.count(page_no);
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
return result;
|
||||
}
|
||||
|
||||
void CorruptedPages::drop_space(ulint space_id)
|
||||
{
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
m_spaces.erase(space_id);
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
}
|
||||
|
||||
void CorruptedPages::rename_space(ulint space_id, const std::string &new_name)
|
||||
{
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
container_t::iterator space_it = m_spaces.find(space_id);
|
||||
if (space_it != m_spaces.end())
|
||||
space_it->second.space_name = new_name;
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
}
|
||||
|
||||
bool CorruptedPages::print_to_file(const char *filename) const
|
||||
{
|
||||
std::ostringstream out;
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
if (!m_spaces.size())
|
||||
{
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
return true;
|
||||
}
|
||||
for (container_t::const_iterator space_it=
|
||||
m_spaces.begin();
|
||||
space_it != m_spaces.end(); ++space_it)
|
||||
{
|
||||
out << space_it->second.space_name << " " << space_it->first << "\n";
|
||||
bool first_page_no= true;
|
||||
for (std::set<ulint>::const_iterator page_it=
|
||||
space_it->second.pages.begin();
|
||||
page_it != space_it->second.pages.end(); ++page_it)
|
||||
if (first_page_no)
|
||||
{
|
||||
out << *page_it;
|
||||
first_page_no= false;
|
||||
}
|
||||
else
|
||||
out << " " << *page_it;
|
||||
out << "\n";
|
||||
}
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
if (xtrabackup_backup)
|
||||
return backup_file_print_buf(filename, out.str().c_str(),
|
||||
static_cast<int>(out.str().size()));
|
||||
std::ofstream outfile;
|
||||
outfile.open(filename);
|
||||
if (!outfile.is_open())
|
||||
die("Can't open %s, error number: %d, error message: %s", filename, errno,
|
||||
strerror(errno));
|
||||
outfile << out.str();
|
||||
return true;
|
||||
}
|
||||
|
||||
void CorruptedPages::read_from_file(const char *file_name)
|
||||
{
|
||||
MY_STAT mystat;
|
||||
if (!my_stat(file_name, &mystat, MYF(0)))
|
||||
return;
|
||||
std::ifstream infile;
|
||||
infile.open(file_name);
|
||||
if (!infile.is_open())
|
||||
die("Can't open %s, error number: %d, error message: %s", file_name, errno,
|
||||
strerror(errno));
|
||||
std::string line;
|
||||
std::string space_name;
|
||||
ulint space_id;
|
||||
ulint line_number= 0;
|
||||
while (std::getline(infile, line))
|
||||
{
|
||||
++line_number;
|
||||
std::istringstream iss(line);
|
||||
if (line_number & 1) {
|
||||
if (!(iss >> space_name))
|
||||
die("Can't parse space name from corrupted pages file at "
|
||||
"line " ULINTPF,
|
||||
line_number);
|
||||
if (!(iss >> space_id))
|
||||
die("Can't parse space id from corrupted pages file at line " ULINTPF,
|
||||
line_number);
|
||||
}
|
||||
else
|
||||
{
|
||||
ulint page_no;
|
||||
while ((iss >> page_no))
|
||||
add_page_no_lock(space_name.c_str(), space_id, page_no, false);
|
||||
if (!iss.eof())
|
||||
die("Corrupted pages file parse error on line number " ULINTPF,
|
||||
line_number);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
bool CorruptedPages::empty() const
|
||||
{
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
bool result= !m_spaces.size();
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
return result;
|
||||
}
|
||||
|
||||
static void xb_load_single_table_tablespace(const std::string &space_name,
|
||||
bool set_size);
|
||||
static void xb_data_files_close();
|
||||
static fil_space_t* fil_space_get_by_name(const char* name);
|
||||
|
||||
void CorruptedPages::zero_out_free_pages()
|
||||
{
|
||||
container_t non_free_pages;
|
||||
byte* buf= static_cast<byte*>(ut_malloc_nokey(2 * srv_page_size));
|
||||
byte* zero_page = static_cast<byte*>(ut_align(buf, srv_page_size));
|
||||
memset(zero_page, 0, srv_page_size);
|
||||
|
||||
ut_a(!pthread_mutex_lock(&m_mutex));
|
||||
for (container_t::const_iterator space_it= m_spaces.begin();
|
||||
space_it != m_spaces.end(); ++space_it)
|
||||
{
|
||||
ulint space_id = space_it->first;
|
||||
const std::string &space_name = space_it->second.space_name;
|
||||
// There is no need to close tablespaces explixitly as they will be closed
|
||||
// in innodb_shutdown().
|
||||
xb_load_single_table_tablespace(space_name, false);
|
||||
mutex_enter(&fil_system.mutex);
|
||||
fil_space_t *space = fil_space_get_by_name(space_name.c_str());
|
||||
mutex_exit(&fil_system.mutex);
|
||||
if (!space)
|
||||
die("Can't find space object for space name %s to check corrupted page",
|
||||
space_name.c_str());
|
||||
for (std::set<ulint>::const_iterator page_it=
|
||||
space_it->second.pages.begin();
|
||||
page_it != space_it->second.pages.end(); ++page_it)
|
||||
{
|
||||
bool is_free= fseg_page_is_free(space, static_cast<unsigned>(*page_it));
|
||||
if (!is_free) {
|
||||
space_info_t &space_info = non_free_pages[space_id];
|
||||
space_info.pages.insert(*page_it);
|
||||
if (space_info.space_name.empty())
|
||||
space_info.space_name = space_name;
|
||||
msg("Error: corrupted page " ULINTPF
|
||||
" of tablespace %s can not be fixed",
|
||||
*page_it, space_name.c_str());
|
||||
}
|
||||
else
|
||||
{
|
||||
const page_id_t page_id(space->id, *page_it);
|
||||
dberr_t err= fil_io(IORequestWrite, true, page_id, 0, 0,
|
||||
srv_page_size, zero_page, NULL);
|
||||
if (err != DB_SUCCESS)
|
||||
die("Can't zero out corrupted page " ULINTPF " of tablespace %s",
|
||||
*page_it, space_name.c_str());
|
||||
msg("Corrupted page " ULINTPF
|
||||
" of tablespace %s was successfuly fixed.",
|
||||
*page_it, space_name.c_str());
|
||||
}
|
||||
}
|
||||
}
|
||||
m_spaces.swap(non_free_pages);
|
||||
ut_a(!pthread_mutex_unlock(&m_mutex));
|
||||
ut_free(buf);
|
||||
}
|
||||
|
||||
/* Simple datasink creation tracking...add datasinks in the reverse order you
|
||||
want them destroyed. */
|
||||
#define XTRABACKUP_MAX_DATASINKS 10
|
||||
|
@ -380,11 +587,12 @@ xtrabackup_add_datasink(ds_ctxt_t *ds)
|
|||
datasinks[actual_datasinks] = ds; actual_datasinks++;
|
||||
}
|
||||
|
||||
|
||||
typedef void (*process_single_tablespace_func_t)(const char *dirname, const char *filname, bool is_remote);
|
||||
typedef void (*process_single_tablespace_func_t)(const char *dirname,
|
||||
const char *filname,
|
||||
bool is_remote,
|
||||
bool set_size);
|
||||
static dberr_t enumerate_ibd_files(process_single_tablespace_func_t callback);
|
||||
|
||||
|
||||
/* ======== Datafiles iterator ======== */
|
||||
struct datafiles_iter_t {
|
||||
fil_space_t *space;
|
||||
|
@ -733,6 +941,7 @@ typedef struct {
|
|||
uint *count;
|
||||
pthread_mutex_t* count_mutex;
|
||||
os_thread_id_t id;
|
||||
CorruptedPages *corrupted_pages;
|
||||
} data_thread_ctxt_t;
|
||||
|
||||
/* ======== for option and variables ======== */
|
||||
|
@ -836,7 +1045,8 @@ enum options_xtrabackup
|
|||
OPT_ROCKSDB_DATADIR,
|
||||
OPT_BACKUP_ROCKSDB,
|
||||
OPT_XTRA_CHECK_PRIVILEGES,
|
||||
OPT_XTRA_MYSQLD_ARGS
|
||||
OPT_XTRA_MYSQLD_ARGS,
|
||||
OPT_XB_IGNORE_INNODB_PAGE_CORRUPTION
|
||||
};
|
||||
|
||||
struct my_option xb_client_options[]= {
|
||||
|
@ -1233,6 +1443,17 @@ struct my_option xb_client_options[]= {
|
|||
" uses old (pre-4.1.1) protocol.",
|
||||
&opt_secure_auth, &opt_secure_auth, 0, GET_BOOL, NO_ARG, 1, 0, 0, 0, 0,
|
||||
0},
|
||||
|
||||
{"log-innodb-page-corruption", OPT_XB_IGNORE_INNODB_PAGE_CORRUPTION,
|
||||
"Continue backup if innodb corrupted pages are found. The pages are "
|
||||
"logged in " MB_CORRUPTED_PAGES_FILE
|
||||
" and backup is finished with error. "
|
||||
"--prepare will try to fix corrupted pages. If " MB_CORRUPTED_PAGES_FILE
|
||||
" exists after --prepare in base backup directory, backup still contains "
|
||||
"corrupted pages and can not be considered as consistent.",
|
||||
&opt_log_innodb_page_corruption, &opt_log_innodb_page_corruption, 0,
|
||||
GET_BOOL, NO_ARG, 0, 0, 0, 0, 0, 0},
|
||||
|
||||
#define MYSQL_CLIENT
|
||||
#include "sslopt-longopts.h"
|
||||
#undef MYSQL_CLIENT
|
||||
|
@ -1526,7 +1747,8 @@ debug_sync_point(const char *name)
|
|||
|
||||
static std::set<std::string> tables_for_export;
|
||||
|
||||
static void append_export_table(const char *dbname, const char *tablename, bool is_remote)
|
||||
static void append_export_table(const char *dbname, const char *tablename,
|
||||
bool is_remote, bool set_size)
|
||||
{
|
||||
if(dbname && tablename && !is_remote)
|
||||
{
|
||||
|
@ -2541,7 +2763,8 @@ for full backup, pages filter for incremental backup, etc.
|
|||
@return FALSE on success and TRUE on error */
|
||||
static my_bool xtrabackup_copy_datafile(fil_node_t *node, uint thread_n,
|
||||
const char *dest_name,
|
||||
const xb_write_filt_t &write_filter)
|
||||
const xb_write_filt_t &write_filter,
|
||||
CorruptedPages &corrupted_pages)
|
||||
{
|
||||
char dst_name[FN_REFLEN];
|
||||
ds_file_t *dstfile = NULL;
|
||||
|
@ -2606,7 +2829,8 @@ static my_bool xtrabackup_copy_datafile(fil_node_t *node, uint thread_n,
|
|||
ut_a(write_filter.process != NULL);
|
||||
|
||||
if (write_filter.init != NULL &&
|
||||
!write_filter.init(&write_filt_ctxt, dst_name, &cursor)) {
|
||||
!write_filter.init(&write_filt_ctxt, dst_name, &cursor,
|
||||
opt_log_innodb_page_corruption ? &corrupted_pages : NULL)) {
|
||||
msg (thread_n, "mariabackup: error: failed to initialize page write filter.");
|
||||
goto error;
|
||||
}
|
||||
|
@ -2626,7 +2850,8 @@ static my_bool xtrabackup_copy_datafile(fil_node_t *node, uint thread_n,
|
|||
}
|
||||
|
||||
/* The main copy loop */
|
||||
while ((res = xb_fil_cur_read(&cursor)) == XB_FIL_CUR_SUCCESS) {
|
||||
while ((res = xb_fil_cur_read(&cursor, corrupted_pages)) ==
|
||||
XB_FIL_CUR_SUCCESS) {
|
||||
if (!write_filter.process(&write_filt_ctxt, dstfile)) {
|
||||
goto error;
|
||||
}
|
||||
|
@ -2910,6 +3135,21 @@ static os_thread_ret_t DECLARE_THREAD(io_watching_thread)(void*)
|
|||
}
|
||||
|
||||
#ifndef DBUG_OFF
|
||||
char *dbug_mariabackup_get_val(const char *event, const char *key)
|
||||
{
|
||||
char envvar[FN_REFLEN];
|
||||
if (key) {
|
||||
snprintf(envvar, sizeof(envvar), "%s_%s", event, key);
|
||||
char *slash = strchr(envvar, '/');
|
||||
if (slash)
|
||||
*slash = '_';
|
||||
} else {
|
||||
strncpy(envvar, event, sizeof envvar - 1);
|
||||
envvar[sizeof envvar - 1] = '\0';
|
||||
}
|
||||
return getenv(envvar);
|
||||
}
|
||||
|
||||
/*
|
||||
In debug mode, execute SQL statement that was passed via environment.
|
||||
To use this facility, you need to
|
||||
|
@ -2922,35 +3162,15 @@ To use this facility, you need to
|
|||
for the variable)
|
||||
3. start mariabackup with --dbug=+d,debug_mariabackup_events
|
||||
*/
|
||||
static void dbug_mariabackup_event(const char *event,const char *key)
|
||||
void dbug_mariabackup_event(const char *event,const char *key)
|
||||
{
|
||||
char envvar[FN_REFLEN];
|
||||
if (key) {
|
||||
snprintf(envvar, sizeof(envvar), "%s_%s", event, key);
|
||||
char *slash = strchr(envvar, '/');
|
||||
if (slash)
|
||||
*slash = '_';
|
||||
} else {
|
||||
strncpy(envvar, event, sizeof envvar - 1);
|
||||
envvar[sizeof envvar - 1] = '\0';
|
||||
}
|
||||
char *sql = getenv(envvar);
|
||||
if (sql) {
|
||||
char *sql = dbug_mariabackup_get_val(event, key);
|
||||
if (sql && *sql) {
|
||||
msg("dbug_mariabackup_event : executing '%s'", sql);
|
||||
xb_mysql_query(mysql_connection, sql, false, true);
|
||||
}
|
||||
|
||||
}
|
||||
#define DBUG_MARIABACKUP_EVENT(A, B) DBUG_EXECUTE_IF("mariabackup_events", dbug_mariabackup_event(A,B););
|
||||
#define DBUG_MB_INJECT_CODE(EVENT, KEY, CODE) \
|
||||
DBUG_EXECUTE_IF("mariabackup_inject_code", {\
|
||||
char *env = getenv(EVENT); \
|
||||
if (env && !strcmp(env, KEY)) { CODE } \
|
||||
})
|
||||
#else
|
||||
#define DBUG_MARIABACKUP_EVENT(A,B)
|
||||
#define DBUG_MB_INJECT_CODE(EVENT, KEY, CODE)
|
||||
#endif
|
||||
#endif // DBUG_OFF
|
||||
|
||||
/**************************************************************************
|
||||
Datafiles copying thread.*/
|
||||
|
@ -2963,6 +3183,7 @@ DECLARE_THREAD(data_copy_thread_func)(
|
|||
data_thread_ctxt_t *ctxt = (data_thread_ctxt_t *) arg;
|
||||
uint num = ctxt->num;
|
||||
fil_node_t* node;
|
||||
ut_ad(ctxt->corrupted_pages);
|
||||
|
||||
/*
|
||||
Initialize mysys thread-specific memory so we can
|
||||
|
@ -2974,11 +3195,12 @@ DECLARE_THREAD(data_copy_thread_func)(
|
|||
|
||||
while ((node = datafiles_iter_next(ctxt->it)) != NULL) {
|
||||
DBUG_MARIABACKUP_EVENT("before_copy", node->space->name);
|
||||
DBUG_MB_INJECT_CODE("wait_innodb_redo_before_copy", node->space->name,
|
||||
DBUG_EXECUTE_FOR_KEY("wait_innodb_redo_before_copy", node->space->name,
|
||||
backup_wait_for_lsn(get_current_lsn(mysql_connection)););
|
||||
/* copy the datafile */
|
||||
if (xtrabackup_copy_datafile(node, num, NULL,
|
||||
xtrabackup_incremental ? wf_incremental : wf_write_through))
|
||||
xtrabackup_incremental ? wf_incremental : wf_write_through,
|
||||
*ctxt->corrupted_pages))
|
||||
die("failed to copy datafile.");
|
||||
|
||||
DBUG_MARIABACKUP_EVENT("after_copy", node->space->name);
|
||||
|
@ -3113,15 +3335,22 @@ xb_new_datafile(const char *name, bool is_remote)
|
|||
}
|
||||
|
||||
|
||||
static
|
||||
void
|
||||
xb_load_single_table_tablespace(
|
||||
const char *dirname,
|
||||
const char *filname,
|
||||
bool is_remote)
|
||||
/** Load tablespace.
|
||||
|
||||
@param[in] dirname directory name of the tablespace to open
|
||||
@param[in] filname file name of the tablespece to open
|
||||
@param[in] is_remote true if tablespace file is .isl
|
||||
@param[in] set_size true if we need to set tablespace size in pages explixitly.
|
||||
If this parameter is set, the size and free pages limit will not be read
|
||||
from page 0.
|
||||
*/
|
||||
static void xb_load_single_table_tablespace(const char *dirname,
|
||||
const char *filname,
|
||||
bool is_remote, bool set_size)
|
||||
{
|
||||
ut_ad(srv_operation == SRV_OPERATION_BACKUP
|
||||
|| srv_operation == SRV_OPERATION_RESTORE_DELTA);
|
||||
|| srv_operation == SRV_OPERATION_RESTORE_DELTA
|
||||
|| srv_operation == SRV_OPERATION_RESTORE);
|
||||
/* Ignore .isl files on XtraBackup recovery. All tablespaces must be
|
||||
local. */
|
||||
if (is_remote && srv_operation == SRV_OPERATION_RESTORE_DELTA) {
|
||||
|
@ -3169,13 +3398,12 @@ xb_load_single_table_tablespace(
|
|||
bool is_empty_file = file->exists() && file->is_empty_file();
|
||||
|
||||
if (err == DB_SUCCESS && file->space_id() != SRV_TMP_SPACE_ID) {
|
||||
os_offset_t node_size = os_file_get_size(file->handle());
|
||||
os_offset_t n_pages;
|
||||
|
||||
ut_a(node_size != (os_offset_t) -1);
|
||||
|
||||
n_pages = node_size / fil_space_t::physical_size(file->flags());
|
||||
|
||||
os_offset_t n_pages = 0;
|
||||
if (set_size) {
|
||||
os_offset_t node_size = os_file_get_size(file->handle());
|
||||
ut_a(node_size != (os_offset_t) -1);
|
||||
n_pages = node_size / fil_space_t::physical_size(file->flags());
|
||||
}
|
||||
space = fil_space_create(
|
||||
name, file->space_id(), file->flags(),
|
||||
FIL_TYPE_TABLESPACE, NULL/* TODO: crypt_data */);
|
||||
|
@ -3203,6 +3431,27 @@ xb_load_single_table_tablespace(
|
|||
ut_free(name);
|
||||
}
|
||||
|
||||
static void xb_load_single_table_tablespace(const std::string &space_name,
|
||||
bool set_size)
|
||||
{
|
||||
std::string name(space_name);
|
||||
bool is_remote= access((name + ".ibd").c_str(), R_OK) != 0;
|
||||
const char *extension= is_remote ? ".isl" : ".ibd";
|
||||
name.append(extension);
|
||||
char buf[FN_REFLEN];
|
||||
strncpy(buf, name.c_str(), sizeof buf - 1);
|
||||
buf[sizeof buf - 1]= '\0';
|
||||
const char *dbname= buf;
|
||||
char *p= strchr(buf, '/');
|
||||
if (p == 0)
|
||||
die("Unexpected tablespace %s filename %s", space_name.c_str(),
|
||||
name.c_str());
|
||||
ut_a(p);
|
||||
*p= 0;
|
||||
const char *tablename= p + 1;
|
||||
xb_load_single_table_tablespace(dbname, tablename, is_remote, set_size);
|
||||
}
|
||||
|
||||
/** Scan the database directories under the MySQL datadir, looking for
|
||||
.ibd files and determining the space id in each of them.
|
||||
@return DB_SUCCESS or error number */
|
||||
|
@ -3244,7 +3493,7 @@ static dberr_t enumerate_ibd_files(process_single_tablespace_func_t callback)
|
|||
bool is_ibd = !is_isl && ends_with(dbinfo.name,".ibd");
|
||||
|
||||
if (is_isl || is_ibd) {
|
||||
(*callback)(NULL, dbinfo.name, is_isl);
|
||||
(*callback)(NULL, dbinfo.name, is_isl, false);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -3301,7 +3550,7 @@ static dberr_t enumerate_ibd_files(process_single_tablespace_func_t callback)
|
|||
if (strlen(fileinfo.name) > 4) {
|
||||
bool is_isl= false;
|
||||
if (ends_with(fileinfo.name, ".ibd") || ((is_isl = ends_with(fileinfo.name, ".isl"))))
|
||||
(*callback)(dbinfo.name, fileinfo.name, is_isl);
|
||||
(*callback)(dbinfo.name, fileinfo.name, is_isl, false);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -4070,6 +4319,7 @@ static bool xtrabackup_backup_func()
|
|||
uint i;
|
||||
uint count;
|
||||
pthread_mutex_t count_mutex;
|
||||
CorruptedPages corrupted_pages;
|
||||
data_thread_ctxt_t *data_threads;
|
||||
pthread_mutex_init(&backup_mutex, NULL);
|
||||
pthread_cond_init(&scanned_lsn_cond, NULL);
|
||||
|
@ -4354,6 +4604,7 @@ fail_before_log_copying_thread_start:
|
|||
data_threads[i].num = i+1;
|
||||
data_threads[i].count = &count;
|
||||
data_threads[i].count_mutex = &count_mutex;
|
||||
data_threads[i].corrupted_pages = &corrupted_pages;
|
||||
os_thread_create(data_copy_thread_func, data_threads + i,
|
||||
&data_threads[i].id);
|
||||
}
|
||||
|
@ -4374,7 +4625,7 @@ fail_before_log_copying_thread_start:
|
|||
datafiles_iter_free(it);
|
||||
}
|
||||
|
||||
bool ok = backup_start();
|
||||
bool ok = backup_start(corrupted_pages);
|
||||
|
||||
if (ok) {
|
||||
ok = xtrabackup_backup_low();
|
||||
|
@ -4391,6 +4642,9 @@ fail_before_log_copying_thread_start:
|
|||
}
|
||||
}
|
||||
|
||||
if (opt_log_innodb_page_corruption)
|
||||
ok = corrupted_pages.print_to_file(MB_CORRUPTED_PAGES_FILE);
|
||||
|
||||
if (!ok) {
|
||||
goto fail;
|
||||
}
|
||||
|
@ -4418,7 +4672,13 @@ fail_before_log_copying_thread_start:
|
|||
log_file_op = NULL;
|
||||
pthread_mutex_destroy(&backup_mutex);
|
||||
pthread_cond_destroy(&scanned_lsn_cond);
|
||||
return(true);
|
||||
if (opt_log_innodb_page_corruption && !corrupted_pages.empty()) {
|
||||
msg("Error: corrupted innodb pages are found and logged to "
|
||||
MB_CORRUPTED_PAGES_FILE " file");
|
||||
return false;
|
||||
}
|
||||
else
|
||||
return(true);
|
||||
}
|
||||
|
||||
|
||||
|
@ -4440,7 +4700,7 @@ FTWRL. This ensures consistent backup in presence of DDL.
|
|||
It is the responsibility of the prepare phase to deal with .new, .ren, and .del
|
||||
files.
|
||||
*/
|
||||
void backup_fix_ddl(void)
|
||||
void backup_fix_ddl(CorruptedPages &corrupted_pages)
|
||||
{
|
||||
std::set<std::string> new_tables;
|
||||
std::set<std::string> dropped_tables;
|
||||
|
@ -4463,6 +4723,7 @@ void backup_fix_ddl(void)
|
|||
|
||||
if (ddl_tracker.drops.find(id) != ddl_tracker.drops.end()) {
|
||||
dropped_tables.insert(name);
|
||||
corrupted_pages.drop_space(id);
|
||||
continue;
|
||||
}
|
||||
|
||||
|
@ -4483,15 +4744,21 @@ void backup_fix_ddl(void)
|
|||
/* table was renamed, but we need a full copy
|
||||
of it because of optimized DDL. We emulate a drop/create.*/
|
||||
dropped_tables.insert(name);
|
||||
if (opt_log_innodb_page_corruption)
|
||||
corrupted_pages.drop_space(id);
|
||||
new_tables.insert(new_name);
|
||||
} else {
|
||||
/* Renamed, and no optimized DDL*/
|
||||
renamed_tables[name] = new_name;
|
||||
if (opt_log_innodb_page_corruption)
|
||||
corrupted_pages.rename_space(id, new_name);
|
||||
}
|
||||
} else if (has_optimized_ddl) {
|
||||
/* Table was recreated, or optimized DDL ran.
|
||||
In both cases we need a full copy in the backup.*/
|
||||
new_tables.insert(name);
|
||||
if (opt_log_innodb_page_corruption)
|
||||
corrupted_pages.drop_space(id);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -4511,6 +4778,8 @@ void backup_fix_ddl(void)
|
|||
if (ddl_tracker.drops.find(id) == ddl_tracker.drops.end()) {
|
||||
dropped_tables.erase(name);
|
||||
new_tables.insert(name);
|
||||
if (opt_log_innodb_page_corruption)
|
||||
corrupted_pages.drop_space(id);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -4558,23 +4827,7 @@ void backup_fix_ddl(void)
|
|||
const char *space_name = iter->c_str();
|
||||
if (check_if_skip_table(space_name))
|
||||
continue;
|
||||
std::string name(*iter);
|
||||
bool is_remote = access((name + ".ibd").c_str(), R_OK) != 0;
|
||||
const char *extension = is_remote ? ".isl" : ".ibd";
|
||||
name.append(extension);
|
||||
char buf[FN_REFLEN];
|
||||
strncpy(buf, name.c_str(), sizeof buf - 1);
|
||||
buf[sizeof buf - 1] = '\0';
|
||||
const char *dbname = buf;
|
||||
char *p = strchr(buf, '/');
|
||||
if (p == 0) {
|
||||
msg("Unexpected tablespace %s filename %s", space_name, name.c_str());
|
||||
ut_a(0);
|
||||
}
|
||||
ut_a(p);
|
||||
*p = 0;
|
||||
const char *tablename = p + 1;
|
||||
xb_load_single_table_tablespace(dbname, tablename, is_remote);
|
||||
xb_load_single_table_tablespace(*iter, false);
|
||||
}
|
||||
|
||||
it = datafiles_iter_new();
|
||||
|
@ -4587,7 +4840,8 @@ void backup_fix_ddl(void)
|
|||
continue;
|
||||
std::string dest_name(node->space->name);
|
||||
dest_name.append(".new");
|
||||
xtrabackup_copy_datafile(node, 0, dest_name.c_str(), wf_write_through);
|
||||
xtrabackup_copy_datafile(node, 0, dest_name.c_str(), wf_write_through,
|
||||
corrupted_pages);
|
||||
}
|
||||
|
||||
datafiles_iter_free(it);
|
||||
|
@ -5502,6 +5756,7 @@ static ibool prepare_handle_del_files(const char *datadir, const char *db, const
|
|||
@return whether the operation succeeded */
|
||||
static bool xtrabackup_prepare_func(char** argv)
|
||||
{
|
||||
CorruptedPages corrupted_pages;
|
||||
char metadata_path[FN_REFLEN];
|
||||
|
||||
/* cd to target-dir */
|
||||
|
@ -5674,6 +5929,30 @@ static bool xtrabackup_prepare_func(char** argv)
|
|||
goto error_cleanup;
|
||||
}
|
||||
|
||||
corrupted_pages.read_from_file(MB_CORRUPTED_PAGES_FILE);
|
||||
if (xtrabackup_incremental)
|
||||
{
|
||||
char inc_filename[FN_REFLEN];
|
||||
sprintf(inc_filename, "%s/%s", xtrabackup_incremental_dir,
|
||||
MB_CORRUPTED_PAGES_FILE);
|
||||
corrupted_pages.read_from_file(inc_filename);
|
||||
}
|
||||
if (!corrupted_pages.empty())
|
||||
corrupted_pages.zero_out_free_pages();
|
||||
if (corrupted_pages.empty())
|
||||
{
|
||||
if (!xtrabackup_incremental && unlink(MB_CORRUPTED_PAGES_FILE) &&
|
||||
errno != ENOENT)
|
||||
{
|
||||
char errbuf[MYSYS_STRERROR_SIZE];
|
||||
my_strerror(errbuf, sizeof(errbuf), errno);
|
||||
die("Error: unlink %s failed: %s", MB_CORRUPTED_PAGES_FILE,
|
||||
errbuf);
|
||||
}
|
||||
}
|
||||
else
|
||||
corrupted_pages.print_to_file(MB_CORRUPTED_PAGES_FILE);
|
||||
|
||||
if (xtrabackup_rollback_xa)
|
||||
{
|
||||
/* Please do not merge MDEV-21168 fix in 10.5+ */
|
||||
|
@ -5785,7 +6064,7 @@ static bool xtrabackup_prepare_func(char** argv)
|
|||
|
||||
error_cleanup:
|
||||
xb_filters_free();
|
||||
return ok && !ib::error::was_logged();
|
||||
return ok && !ib::error::was_logged() && corrupted_pages.empty();
|
||||
}
|
||||
|
||||
/**************************************************************************
|
||||
|
|
|
@ -25,6 +25,7 @@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
|
|||
#include "datasink.h"
|
||||
#include "xbstream.h"
|
||||
#include "changed_page_bitmap.h"
|
||||
#include <set>
|
||||
|
||||
struct xb_delta_info_t
|
||||
{
|
||||
|
@ -36,6 +37,32 @@ struct xb_delta_info_t
|
|||
ulint space_id;
|
||||
};
|
||||
|
||||
class CorruptedPages
|
||||
{
|
||||
public:
|
||||
CorruptedPages();
|
||||
~CorruptedPages();
|
||||
void add_page(const char *file_name, ulint space_id, ulint page_no);
|
||||
bool contains(ulint space_id, ulint page_no) const;
|
||||
void drop_space(ulint space_id);
|
||||
void rename_space(ulint space_id, const std::string &new_name);
|
||||
bool print_to_file(const char *file_name) const;
|
||||
void read_from_file(const char *file_name);
|
||||
bool empty() const;
|
||||
void zero_out_free_pages();
|
||||
|
||||
private:
|
||||
void add_page_no_lock(const char *space_name, ulint space_id, ulint page_no,
|
||||
bool convert_space_name);
|
||||
struct space_info_t {
|
||||
std::string space_name;
|
||||
std::set<ulint> pages;
|
||||
};
|
||||
typedef std::map<ulint, space_info_t> container_t;
|
||||
mutable pthread_mutex_t m_mutex;
|
||||
container_t m_spaces;
|
||||
};
|
||||
|
||||
/* value of the --incremental option */
|
||||
extern lsn_t incremental_lsn;
|
||||
|
||||
|
@ -111,6 +138,7 @@ extern my_bool opt_remove_original;
|
|||
extern my_bool opt_extended_validation;
|
||||
extern my_bool opt_encrypted_backup;
|
||||
extern my_bool opt_lock_ddl_per_table;
|
||||
extern my_bool opt_log_innodb_page_corruption;
|
||||
|
||||
extern char *opt_incremental_history_name;
|
||||
extern char *opt_incremental_history_uuid;
|
||||
|
|
|
@ -514,6 +514,10 @@ sub mtr_report_stats ($$$$) {
|
|||
# if a test case has to be retried it should have the result MTR_RES_FAILED in jUnit XML
|
||||
if ($test->{'result'} eq "MTR_RES_FAILED" || $test->{'retries'} > 0) {
|
||||
my $logcontents = $test->{'logfile-failed'} || $test->{'logfile'};
|
||||
# remove any double ] that would end the cdata
|
||||
$logcontents =~ s/]]/\x{fffd}/g;
|
||||
# replace wide characters that aren't allowed in XML 1.0
|
||||
$logcontents =~ s/[\x00-\x08\x0B\x0C\x0E-\x1F]/\x{fffd}/g;
|
||||
|
||||
$xml_report .= qq(>\n\t\t\t<failure message="" type="MTR_RES_FAILED">\n<![CDATA[$logcontents]]>\n\t\t\t</failure>\n\t\t</testcase>\n);
|
||||
} elsif ($test->{'result'} eq "MTR_RES_SKIPPED" && $test->{'disable'}) {
|
||||
|
@ -530,9 +534,9 @@ sub mtr_report_stats ($$$$) {
|
|||
# save to file
|
||||
my $xml_file = $::opt_xml_report;
|
||||
|
||||
open XML_FILE, ">", $xml_file or die "Cannot create file $xml_file: $!";
|
||||
print XML_FILE $xml_report;
|
||||
close XML_FILE;
|
||||
open (my $XML_UFILE, '>:encoding(UTF-8)', $xml_file) or die 'Cannot create file $xml_file: $!';
|
||||
print $XML_UFILE $xml_report;
|
||||
close $XML_UFILE or warn "File close failed!";
|
||||
}
|
||||
|
||||
if (@$extra_warnings)
|
||||
|
|
|
@ -289,3 +289,51 @@ pk f
|
|||
5 a
|
||||
6 <===
|
||||
drop table t1;
|
||||
#
|
||||
# MDEV-21842: auto_increment does not increment with compound primary
|
||||
# key on partitioned table
|
||||
#
|
||||
create or replace table `t` (
|
||||
`id` bigint(20) unsigned not null auto_increment,
|
||||
`a` int(10) not null ,
|
||||
`dt` date not null,
|
||||
primary key (`id`, `dt`) ,
|
||||
unique key (`a`, `dt`)
|
||||
)
|
||||
partition by range columns(`dt`)
|
||||
(
|
||||
partition `p202002` values less than ('2020-03-01'),
|
||||
partition `P202003` values less than ('2020-04-01')
|
||||
);
|
||||
connect con1, localhost, root,,;
|
||||
connect con2, localhost, root,,;
|
||||
connection con1;
|
||||
start transaction;
|
||||
insert into t (a, dt) values (1, '2020-02-29');
|
||||
connection con2;
|
||||
start transaction;
|
||||
insert into t (a, dt) values (1, '2020-02-29');
|
||||
connection con1;
|
||||
insert into t (a, dt) values (2, '2020-02-29');
|
||||
select auto_increment from information_schema.tables where table_name='t';
|
||||
auto_increment
|
||||
4
|
||||
commit;
|
||||
connection con2;
|
||||
ERROR 23000: Duplicate entry '1-2020-02-29' for key 'a'
|
||||
connection con1;
|
||||
select auto_increment from information_schema.tables where table_name='t';
|
||||
auto_increment
|
||||
4
|
||||
insert into t (a, dt) values (3, '2020-02-29');
|
||||
insert into t (a, dt) values (4, '2020-02-29');
|
||||
disconnect con1;
|
||||
disconnect con2;
|
||||
connection default;
|
||||
select * from t;
|
||||
id a dt
|
||||
1 1 2020-02-29
|
||||
3 2 2020-02-29
|
||||
4 3 2020-02-29
|
||||
5 4 2020-02-29
|
||||
drop table t;
|
||||
|
|
|
@ -18,3 +18,62 @@ select * from t1;
|
|||
drop table t1;
|
||||
--let $datadir=`select @@datadir`
|
||||
--remove_file $datadir/test/load.data
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-21842: auto_increment does not increment with compound primary
|
||||
--echo # key on partitioned table
|
||||
--echo #
|
||||
|
||||
create or replace table `t` (
|
||||
`id` bigint(20) unsigned not null auto_increment,
|
||||
`a` int(10) not null ,
|
||||
`dt` date not null,
|
||||
primary key (`id`, `dt`) ,
|
||||
unique key (`a`, `dt`)
|
||||
)
|
||||
partition by range columns(`dt`)
|
||||
(
|
||||
partition `p202002` values less than ('2020-03-01'),
|
||||
partition `P202003` values less than ('2020-04-01')
|
||||
);
|
||||
|
||||
connect (con1, localhost, root,,);
|
||||
connect (con2, localhost, root,,);
|
||||
|
||||
--connection con1
|
||||
start transaction;
|
||||
insert into t (a, dt) values (1, '2020-02-29');
|
||||
|
||||
--connection con2
|
||||
start transaction;
|
||||
let $conn2_id= `SELECT CONNECTION_ID()`;
|
||||
send insert into t (a, dt) values (1, '2020-02-29');
|
||||
|
||||
--connection con1
|
||||
# Ensure that the above insert via conn2 increments next_auto_inc_val
|
||||
# before the following insert via conn1 starts.
|
||||
let $wait_condition=select 1 from Information_schema.INNODB_TRX
|
||||
where trx_mysql_thread_id = $conn2_id and trx_state = 'LOCK WAIT'
|
||||
and trx_query = "insert into t (a, dt) values (1, '2020-02-29')";
|
||||
--source include/wait_condition.inc
|
||||
|
||||
insert into t (a, dt) values (2, '2020-02-29');
|
||||
select auto_increment from information_schema.tables where table_name='t';
|
||||
commit;
|
||||
|
||||
--connection con2
|
||||
--error ER_DUP_ENTRY
|
||||
reap;
|
||||
|
||||
--connection con1
|
||||
select auto_increment from information_schema.tables where table_name='t';
|
||||
insert into t (a, dt) values (3, '2020-02-29');
|
||||
insert into t (a, dt) values (4, '2020-02-29');
|
||||
|
||||
disconnect con1;
|
||||
disconnect con2;
|
||||
|
||||
--connection default
|
||||
select * from t;
|
||||
drop table t;
|
||||
|
||||
|
|
|
@ -10590,6 +10590,47 @@ a
|
|||
abc
|
||||
DROP VIEW v1;
|
||||
DROP TABLE t1;
|
||||
#
|
||||
# MDEV-19179: pushdown into UNION of aggregation selects whose
|
||||
# corresponding columns have different names
|
||||
#
|
||||
create table t1 (a int);
|
||||
insert into t1 values (3), (7), (1);
|
||||
select *
|
||||
from (select min(a) as x from t1 union all select max(a) as y from t1) t
|
||||
where x>0;
|
||||
x
|
||||
1
|
||||
7
|
||||
explain extended select *
|
||||
from (select min(a) as x from t1 union all select max(a) as y from t1) t
|
||||
where x>0;
|
||||
id select_type table type possible_keys key key_len ref rows filtered Extra
|
||||
1 PRIMARY <derived2> ALL NULL NULL NULL NULL 6 100.00 Using where
|
||||
2 DERIVED t1 ALL NULL NULL NULL NULL 3 100.00
|
||||
3 UNION t1 ALL NULL NULL NULL NULL 3 100.00
|
||||
Warnings:
|
||||
Note 1003 /* select#1 */ select `t`.`x` AS `x` from (/* select#2 */ select min(`test`.`t1`.`a`) AS `x` from `test`.`t1` having `x` > 0 union all /* select#3 */ select max(`test`.`t1`.`a`) AS `x` from `test`.`t1` having `x` > 0) `t` where `t`.`x` > 0
|
||||
prepare stmt from "select *
|
||||
from (select min(a) as x from t1 union all select max(a) as y from t1) t
|
||||
where x>0";
|
||||
execute stmt;
|
||||
x
|
||||
1
|
||||
7
|
||||
execute stmt;
|
||||
x
|
||||
1
|
||||
7
|
||||
deallocate prepare stmt;
|
||||
create view v1(m) as
|
||||
select min(a) as x from t1 union all select max(a) as y from t1;
|
||||
select * from v1 where m > 0;
|
||||
m
|
||||
1
|
||||
7
|
||||
drop view v1;
|
||||
drop table t1;
|
||||
# End of 10.2 tests
|
||||
#
|
||||
# MDEV-14579: pushdown conditions into materialized views/derived tables
|
||||
|
|
|
@ -2185,6 +2185,34 @@ SELECT * FROM v1 WHERE IF( a REGEXP 'def', 'foo', a ) IN ('abc', 'foobar');
|
|||
DROP VIEW v1;
|
||||
DROP TABLE t1;
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-19179: pushdown into UNION of aggregation selects whose
|
||||
--echo # corresponding columns have different names
|
||||
--echo #
|
||||
|
||||
create table t1 (a int);
|
||||
insert into t1 values (3), (7), (1);
|
||||
|
||||
let $q=
|
||||
select *
|
||||
from (select min(a) as x from t1 union all select max(a) as y from t1) t
|
||||
where x>0;
|
||||
|
||||
eval $q;
|
||||
eval explain extended $q;
|
||||
|
||||
eval prepare stmt from "$q";
|
||||
execute stmt;
|
||||
execute stmt;
|
||||
deallocate prepare stmt;
|
||||
|
||||
create view v1(m) as
|
||||
select min(a) as x from t1 union all select max(a) as y from t1;
|
||||
select * from v1 where m > 0;
|
||||
|
||||
drop view v1;
|
||||
drop table t1;
|
||||
|
||||
--echo # End of 10.2 tests
|
||||
|
||||
--echo #
|
||||
|
|
|
@ -2219,8 +2219,6 @@ SCHEMA_NAME
|
|||
# End of 10.1 tests
|
||||
#
|
||||
#
|
||||
# Start of 10.2 Test
|
||||
#
|
||||
# MDEV-14836: Assertion `m_status == DA_ERROR' failed in
|
||||
# Diagnostics_area::sql_errno upon query from I_S with LIMIT ROWS EXAMINED
|
||||
#
|
||||
|
@ -2305,5 +2303,12 @@ mysql global_priv Priv json_valid(`Priv`) def mysql
|
|||
test t a `i` > 0 def test
|
||||
drop table t;
|
||||
#
|
||||
# MDEV-24230 subquery on information_schema fails with error message
|
||||
#
|
||||
create table t1 (n int);
|
||||
create table t2 (n int);
|
||||
insert into t1 set n = (select table_rows from information_schema.tables where table_name='t2');
|
||||
drop table t1, t2;
|
||||
#
|
||||
# End of 10.3 tests
|
||||
#
|
||||
|
|
|
@ -1926,8 +1926,6 @@ SELECT SCHEMA_NAME from information_schema.schemata where schema_name=REPEAT('a'
|
|||
--echo # End of 10.1 tests
|
||||
--echo #
|
||||
|
||||
--echo #
|
||||
--echo # Start of 10.2 Test
|
||||
--echo #
|
||||
--echo # MDEV-14836: Assertion `m_status == DA_ERROR' failed in
|
||||
--echo # Diagnostics_area::sql_errno upon query from I_S with LIMIT ROWS EXAMINED
|
||||
|
@ -2001,6 +1999,14 @@ from information_schema.TABLE_CONSTRAINTS tc
|
|||
|
||||
drop table t;
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-24230 subquery on information_schema fails with error message
|
||||
--echo #
|
||||
create table t1 (n int);
|
||||
create table t2 (n int);
|
||||
insert into t1 set n = (select table_rows from information_schema.tables where table_name='t2');
|
||||
drop table t1, t2;
|
||||
|
||||
--echo #
|
||||
--echo # End of 10.3 tests
|
||||
--echo #
|
||||
|
|
|
@ -1,4 +1,5 @@
|
|||
source include/not_embedded.inc;
|
||||
source include/have_perfschema.inc;
|
||||
#
|
||||
# LOCK TABLES and privileges on views
|
||||
#
|
||||
|
|
|
@ -3,6 +3,10 @@
|
|||
--source include/have_udf.inc
|
||||
--source include/platform.inc
|
||||
|
||||
if (!$AUTH_SOCKET_SO) {
|
||||
--skip Need auth socket plugin
|
||||
}
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-23630: mysqldump to logically dump system tables
|
||||
--echo #
|
||||
|
|
|
@ -1087,10 +1087,10 @@ INSERT INTO t1 VALUES ();
|
|||
SELECT * FROM t1;
|
||||
a
|
||||
-1
|
||||
1
|
||||
3
|
||||
4
|
||||
6
|
||||
7
|
||||
DROP TABLE t1;
|
||||
#
|
||||
# End of 10.3 tests
|
||||
|
|
|
@ -8450,8 +8450,25 @@ ERROR 22007: Incorrect integer value: 'y' for column ``.``.`a` at row 1
|
|||
DROP TABLE t1;
|
||||
SET sql_mode=DEFAULT;
|
||||
#
|
||||
# Start of 10.3 tests
|
||||
# MDEV-24220: error when opening a table for the second call of SP
|
||||
#
|
||||
CREATE TABLE t1 (a INT, b INT);
|
||||
INSERT INTO t1 VALUES (1,1),(2,2);
|
||||
CREATE VIEW v1 AS SELECT MAX(a) as f FROM t1;
|
||||
CREATE PROCEDURE p1()
|
||||
BEGIN
|
||||
SELECT * FROM v1;
|
||||
END $
|
||||
CALL p1;
|
||||
f
|
||||
2
|
||||
ALTER TABLE t1 DROP a;
|
||||
CALL p1;
|
||||
ERROR HY000: View 'test.v1' references invalid table(s) or column(s) or function(s) or definer/invoker of view lack rights to use them
|
||||
DROP PROCEDURE p1;
|
||||
DROP VIEW v1;
|
||||
DROP TABLE t1;
|
||||
#End of 10.2 tests
|
||||
#
|
||||
# MDEV-12007 Allow ROW variables as a cursor FETCH target
|
||||
#
|
||||
|
|
|
@ -9991,9 +9991,30 @@ DROP TABLE t1;
|
|||
SET sql_mode=DEFAULT;
|
||||
|
||||
--echo #
|
||||
--echo # Start of 10.3 tests
|
||||
--echo # MDEV-24220: error when opening a table for the second call of SP
|
||||
--echo #
|
||||
|
||||
CREATE TABLE t1 (a INT, b INT);
|
||||
INSERT INTO t1 VALUES (1,1),(2,2);
|
||||
CREATE VIEW v1 AS SELECT MAX(a) as f FROM t1;
|
||||
--delimiter $
|
||||
CREATE PROCEDURE p1()
|
||||
BEGIN
|
||||
SELECT * FROM v1;
|
||||
END $
|
||||
--delimiter ;
|
||||
|
||||
CALL p1;
|
||||
ALTER TABLE t1 DROP a;
|
||||
-- error ER_VIEW_INVALID
|
||||
CALL p1;
|
||||
|
||||
DROP PROCEDURE p1;
|
||||
DROP VIEW v1;
|
||||
DROP TABLE t1;
|
||||
|
||||
--echo #End of 10.2 tests
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-12007 Allow ROW variables as a cursor FETCH target
|
||||
--echo #
|
||||
|
|
|
@ -2719,6 +2719,40 @@ SET join_cache_level= @save_join_cache_level;
|
|||
DROP TABLE t1,t2,t3,t4;
|
||||
# End of 10.2 tests
|
||||
#
|
||||
# MDEV-21265: IN predicate conversion to IN subquery should be allowed for a broader set of datatype comparison
|
||||
#
|
||||
CREATE TABLE t1(a VARCHAR(50) collate utf8_general_ci, b INT);
|
||||
INSERT INTO t1 VALUES ('abc',1), ('def', 2), ('ghi', 3), ('jkl', 4), ('mno', 5);
|
||||
CREATE TABLE t2(a VARCHAR(50) collate utf8mb4_general_ci, b INT);
|
||||
INSERT INTO t2 VALUES ('abc',1), ('def', 2), ('ghi', 3), ('jkl', 4), ('mno', 5);
|
||||
set @save_in_predicate_conversion_threshold= @@in_predicate_conversion_threshold;
|
||||
set in_predicate_conversion_threshold=2;
|
||||
set names 'utf8mb4';
|
||||
#
|
||||
# IN predicate to IN subquery is not allowed as materialization is not allowed
|
||||
# The character set on the inner side is not equal to or a proper subset of the outer side
|
||||
#
|
||||
EXPLAIN
|
||||
SELECT * FROM t1 WHERE (t1.a,t1.b) IN (('abx',1),('def',2), ('abc', 3));
|
||||
id select_type table type possible_keys key key_len ref rows Extra
|
||||
1 SIMPLE t1 ALL NULL NULL NULL NULL 5 Using where
|
||||
set names 'utf8';
|
||||
#
|
||||
# IN predicate to IN subquery is performed as materialization is llowed
|
||||
# The character set on the inner side is a proper subset of the outer side
|
||||
#
|
||||
EXPLAIN
|
||||
SELECT * FROM t2 WHERE (t2.a,t2.b) IN (('abx',1),('def',2), ('abc', 3));
|
||||
id select_type table type possible_keys key key_len ref rows Extra
|
||||
1 PRIMARY t2 ALL NULL NULL NULL NULL 5
|
||||
1 PRIMARY <subquery2> eq_ref distinct_key distinct_key 16 func,func 1 Using where
|
||||
2 MATERIALIZED <derived3> ALL NULL NULL NULL NULL 3
|
||||
3 DERIVED NULL NULL NULL NULL NULL NULL NULL No tables used
|
||||
set names default;
|
||||
set @@in_predicate_conversion_threshold= @save_in_predicate_conversion_threshold;
|
||||
DROP TABLE t1,t2;
|
||||
# End of 10.3 tests
|
||||
#
|
||||
# MDEV-19134: EXISTS() slower if ORDER BY is defined
|
||||
#
|
||||
create table t0 (a int);
|
||||
|
|
|
@ -2239,6 +2239,43 @@ DROP TABLE t1,t2,t3,t4;
|
|||
|
||||
--echo # End of 10.2 tests
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-21265: IN predicate conversion to IN subquery should be allowed for a broader set of datatype comparison
|
||||
--echo #
|
||||
|
||||
CREATE TABLE t1(a VARCHAR(50) collate utf8_general_ci, b INT);
|
||||
INSERT INTO t1 VALUES ('abc',1), ('def', 2), ('ghi', 3), ('jkl', 4), ('mno', 5);
|
||||
|
||||
CREATE TABLE t2(a VARCHAR(50) collate utf8mb4_general_ci, b INT);
|
||||
INSERT INTO t2 VALUES ('abc',1), ('def', 2), ('ghi', 3), ('jkl', 4), ('mno', 5);
|
||||
|
||||
set @save_in_predicate_conversion_threshold= @@in_predicate_conversion_threshold;
|
||||
set in_predicate_conversion_threshold=2;
|
||||
|
||||
set names 'utf8mb4';
|
||||
--echo #
|
||||
--echo # IN predicate to IN subquery is not allowed as materialization is not allowed
|
||||
--echo # The character set on the inner side is not equal to or a proper subset of the outer side
|
||||
--echo #
|
||||
|
||||
EXPLAIN
|
||||
SELECT * FROM t1 WHERE (t1.a,t1.b) IN (('abx',1),('def',2), ('abc', 3));
|
||||
|
||||
set names 'utf8';
|
||||
--echo #
|
||||
--echo # IN predicate to IN subquery is performed as materialization is llowed
|
||||
--echo # The character set on the inner side is a proper subset of the outer side
|
||||
--echo #
|
||||
|
||||
EXPLAIN
|
||||
SELECT * FROM t2 WHERE (t2.a,t2.b) IN (('abx',1),('def',2), ('abc', 3));
|
||||
|
||||
set names default;
|
||||
set @@in_predicate_conversion_threshold= @save_in_predicate_conversion_threshold;
|
||||
DROP TABLE t1,t2;
|
||||
|
||||
--echo # End of 10.3 tests
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-19134: EXISTS() slower if ORDER BY is defined
|
||||
--echo #
|
||||
|
|
|
@ -294,7 +294,7 @@ DROP TABLE t1;
|
|||
#
|
||||
# Bug#12352846 - TRANS_XA_START(THD*):
|
||||
# ASSERTION THD->TRANSACTION.XID_STATE.XID.IS_NULL()
|
||||
# FAILED
|
||||
# FAILED
|
||||
#
|
||||
CREATE TABLE t1 (a INT) ENGINE=InnoDB;
|
||||
CREATE TABLE t2 (a INT) ENGINE=InnoDB;
|
||||
|
@ -345,6 +345,32 @@ connection default;
|
|||
XA END 'xid1';
|
||||
XA ROLLBACK 'xid1';
|
||||
DROP TABLE t1, t2, t3;
|
||||
#
|
||||
# MDEV 15532 XA: Assertion `!log->same_pk' failed in
|
||||
# row_log_table_apply_delete
|
||||
#
|
||||
CREATE TABLE t1 (a INT) ENGINE=InnoDB;
|
||||
INSERT INTO t1 VALUES (1),(2);
|
||||
connect con1,localhost,root,,test;
|
||||
XA START 'xid';
|
||||
UPDATE t1 SET a = 5;
|
||||
connection default;
|
||||
SET innodb_lock_wait_timeout= 2, lock_wait_timeout= 2;
|
||||
ALTER TABLE non_existing_table1;
|
||||
ERROR 42S02: Table 'test.non_existing_table1' doesn't exist
|
||||
ALTER TABLE t1 FORCE;;
|
||||
connection con1;
|
||||
ALTER TABLE non_existing_table2;
|
||||
ERROR XAE07: XAER_RMFAIL: The command cannot be executed when global transaction is in the ACTIVE state
|
||||
DELETE FROM t1 LIMIT 1;
|
||||
connection default;
|
||||
ERROR HY000: Lock wait timeout exceeded; try restarting transaction
|
||||
connection con1;
|
||||
XA END 'xid';
|
||||
XA ROLLBACK 'xid';
|
||||
DROP TABLE t1;
|
||||
disconnect con1;
|
||||
connection default;
|
||||
XA BEGIN 'xid';
|
||||
CREATE TEMPORARY SEQUENCE s;
|
||||
ERROR XAE07: XAER_RMFAIL: The command cannot be executed when global transaction is in the ACTIVE state
|
||||
|
|
|
@ -390,7 +390,7 @@ DROP TABLE t1;
|
|||
--echo #
|
||||
--echo # Bug#12352846 - TRANS_XA_START(THD*):
|
||||
--echo # ASSERTION THD->TRANSACTION.XID_STATE.XID.IS_NULL()
|
||||
--echo # FAILED
|
||||
--echo # FAILED
|
||||
--echo #
|
||||
|
||||
CREATE TABLE t1 (a INT) ENGINE=InnoDB;
|
||||
|
@ -447,7 +447,7 @@ CREATE TABLE t1 (pk INT PRIMARY KEY) ENGINE=InnoDB;
|
|||
CREATE TABLE t2 (pk INT PRIMARY KEY) ENGINE=InnoDB;
|
||||
INSERT INTO t2 VALUES (1),(2);
|
||||
CREATE TABLE t3 (i INT) ENGINE=InnoDB;
|
||||
|
||||
|
||||
XA BEGIN 'xid1';
|
||||
REPLACE INTO t1 SELECT * FROM t2;
|
||||
|
||||
|
@ -476,6 +476,45 @@ XA END 'xid1';
|
|||
XA ROLLBACK 'xid1';
|
||||
DROP TABLE t1, t2, t3;
|
||||
|
||||
--echo #
|
||||
--echo # MDEV 15532 XA: Assertion `!log->same_pk' failed in
|
||||
--echo # row_log_table_apply_delete
|
||||
--echo #
|
||||
|
||||
CREATE TABLE t1 (a INT) ENGINE=InnoDB;
|
||||
INSERT INTO t1 VALUES (1),(2);
|
||||
|
||||
--connect (con1,localhost,root,,test)
|
||||
|
||||
XA START 'xid';
|
||||
UPDATE t1 SET a = 5;
|
||||
|
||||
--connection default
|
||||
SET innodb_lock_wait_timeout= 2, lock_wait_timeout= 2;
|
||||
|
||||
--error ER_NO_SUCH_TABLE
|
||||
ALTER TABLE non_existing_table1;
|
||||
|
||||
--send ALTER TABLE t1 FORCE;
|
||||
|
||||
--connection con1
|
||||
--error ER_XAER_RMFAIL
|
||||
|
||||
ALTER TABLE non_existing_table2;
|
||||
DELETE FROM t1 LIMIT 1;
|
||||
|
||||
--connection default
|
||||
--error ER_LOCK_WAIT_TIMEOUT
|
||||
--reap
|
||||
|
||||
# Cleanup
|
||||
--connection con1
|
||||
XA END 'xid';
|
||||
XA ROLLBACK 'xid';
|
||||
DROP TABLE t1;
|
||||
--disconnect con1
|
||||
connection default;
|
||||
|
||||
--source include/wait_until_count_sessions.inc
|
||||
|
||||
#
|
||||
|
|
95
mysql-test/suite/galera/r/galera_as_slave_replay.result
Normal file
95
mysql-test/suite/galera/r/galera_as_slave_replay.result
Normal file
|
@ -0,0 +1,95 @@
|
|||
connect node_2a, 127.0.0.1, root, , test, $NODE_MYPORT_2;
|
||||
connection node_2a;
|
||||
connection node_1;
|
||||
RESET MASTER;
|
||||
connection node_2a;
|
||||
START SLAVE;
|
||||
connection node_1;
|
||||
CREATE TABLE t1 (f1 INTEGER PRIMARY KEY, f2 CHAR(1)) engine=innodb;
|
||||
INSERT INTO t1 VALUES (1, 'a');
|
||||
INSERT INTO t1 VALUES (3, 'a');
|
||||
set binlog_format=STATEMENT;
|
||||
SET AUTOCOMMIT=ON;
|
||||
START TRANSACTION;
|
||||
SELECT * FROM t1 FOR UPDATE;
|
||||
f1 f2
|
||||
1 a
|
||||
3 a
|
||||
UPDATE t1 SET f2 = 'c' WHERE f1 > 1;
|
||||
connection node_2a;
|
||||
SET SESSION wsrep_sync_wait = 0;
|
||||
connect node_3, 127.0.0.1, root, , test, $NODE_MYPORT_3;
|
||||
connection node_3;
|
||||
SET SESSION wsrep_sync_wait = 0;
|
||||
connection node_2a;
|
||||
SET GLOBAL wsrep_provider_options = 'dbug=d,commit_monitor_enter_sync';
|
||||
SET GLOBAL debug_dbug = "d,sync.wsrep_apply_cb";
|
||||
connection node_3;
|
||||
INSERT INTO test.t1 VALUES (2, 'b');
|
||||
connection node_1;
|
||||
COMMIT;
|
||||
connection node_2a;
|
||||
SET SESSION wsrep_on = 0;
|
||||
SET SESSION wsrep_on = 1;
|
||||
SET GLOBAL debug_dbug = "";
|
||||
SET DEBUG_SYNC = "now SIGNAL signal.wsrep_apply_cb";
|
||||
connection node_2a;
|
||||
SET GLOBAL wsrep_provider_options = 'dbug=';
|
||||
SET GLOBAL wsrep_provider_options = 'signal=commit_monitor_enter_sync';
|
||||
connection node_1;
|
||||
SELECT COUNT(*) = 1 FROM t1 WHERE f2 = 'a';
|
||||
COUNT(*) = 1
|
||||
1
|
||||
SELECT COUNT(*) = 1 FROM t1 WHERE f2 = 'c';
|
||||
COUNT(*) = 1
|
||||
1
|
||||
SELECT * FROM t1;
|
||||
f1 f2
|
||||
1 a
|
||||
3 c
|
||||
connection node_2a;
|
||||
set session wsrep_sync_wait=15;
|
||||
set session wsrep_sync_wait=0;
|
||||
wsrep_local_replays
|
||||
1
|
||||
SELECT * FROM t1;
|
||||
f1 f2
|
||||
1 a
|
||||
2 b
|
||||
3 c
|
||||
SET DEBUG_SYNC = "RESET";
|
||||
#
|
||||
# test phase with real abort
|
||||
#
|
||||
connection node_1;
|
||||
set binlog_format=ROW;
|
||||
insert into t1 values (4, 'd');
|
||||
SET AUTOCOMMIT=ON;
|
||||
START TRANSACTION;
|
||||
UPDATE t1 SET f2 = 'd' WHERE f1 = 3;
|
||||
connection node_2a;
|
||||
SET GLOBAL wsrep_provider_options = 'dbug=d,commit_monitor_enter_sync';
|
||||
SET GLOBAL debug_dbug = "d,sync.wsrep_apply_cb";
|
||||
connection node_3;
|
||||
UPDATE test.t1 SET f2 = 'e' WHERE f1 = 3;
|
||||
connection node_1;
|
||||
COMMIT;
|
||||
connection node_2a;
|
||||
SET GLOBAL debug_dbug = "";
|
||||
SET DEBUG_SYNC = "now SIGNAL signal.wsrep_apply_cb";
|
||||
connection node_2a;
|
||||
SET GLOBAL wsrep_provider_options = 'dbug=';
|
||||
SET GLOBAL wsrep_provider_options = 'signal=commit_monitor_enter_sync';
|
||||
SET DEBUG_SYNC = "RESET";
|
||||
connection node_2a;
|
||||
set session wsrep_sync_wait=15;
|
||||
SELECT COUNT(*) = 1 FROM test.t1 WHERE f2 = 'e';
|
||||
COUNT(*) = 1
|
||||
1
|
||||
set session wsrep_sync_wait=0;
|
||||
STOP SLAVE;
|
||||
RESET SLAVE;
|
||||
DROP TABLE t1;
|
||||
connection node_1;
|
||||
DROP TABLE t1;
|
||||
RESET MASTER;
|
|
@ -1,7 +1,7 @@
|
|||
connection node_2;
|
||||
connection node_1;
|
||||
connection node_1;
|
||||
CREATE TABLE ten (f1 INTEGER) ENGINE=InnoDB;
|
||||
CREATE TABLE ten (f1 INTEGER NOT NULL PRIMARY KEY) ENGINE=InnoDB;
|
||||
INSERT INTO ten VALUES (1),(2),(3),(4),(5),(6),(7),(8),(9),(10);
|
||||
CREATE TABLE t1 (f1 INTEGER AUTO_INCREMENT PRIMARY KEY, f2 INTEGER) ENGINE=InnoDB;
|
||||
INSERT INTO t1 (f2) SELECT 1 FROM ten;
|
||||
|
|
11
mysql-test/suite/galera/t/galera_as_slave_replay.cnf
Normal file
11
mysql-test/suite/galera/t/galera_as_slave_replay.cnf
Normal file
|
@ -0,0 +1,11 @@
|
|||
!include ../galera_2nodes_as_slave.cnf
|
||||
|
||||
[mysqld]
|
||||
binlog-format=row
|
||||
|
||||
[mysqld.1]
|
||||
wsrep_restart_slave=1
|
||||
|
||||
[mysqld.2]
|
||||
wsrep_restart_slave=1
|
||||
|
200
mysql-test/suite/galera/t/galera_as_slave_replay.test
Normal file
200
mysql-test/suite/galera/t/galera_as_slave_replay.test
Normal file
|
@ -0,0 +1,200 @@
|
|||
#
|
||||
# This test tests the operation of transaction replay for async replication slave.
|
||||
# If a potentially conflicting galera transaction arrives at
|
||||
# just the right time during the commit and has lock conflict with async replication transaction
|
||||
# applied by slave SQL thread, then the async replication transaction should either abort
|
||||
# or rollback and replay (depending on the nature of lock conflict).
|
||||
#
|
||||
|
||||
--source include/have_innodb.inc
|
||||
--source include/have_debug.inc
|
||||
--source include/have_debug_sync.inc
|
||||
--source include/galera_have_debug_sync.inc
|
||||
|
||||
--connect node_2a, 127.0.0.1, root, , test, $NODE_MYPORT_2
|
||||
|
||||
--connection node_2a
|
||||
--source include/galera_cluster.inc
|
||||
#--source suite/galera/include/galera_have_debug_sync.inc
|
||||
|
||||
#
|
||||
# node 1 is native MariaDB server operating as async replication master
|
||||
#
|
||||
--connection node_1
|
||||
RESET MASTER;
|
||||
|
||||
--connection node_2a
|
||||
#
|
||||
# count the number of wsrep replay's done in the node
|
||||
#
|
||||
--let $wsrep_local_replays_old = `SELECT VARIABLE_VALUE FROM INFORMATION_SCHEMA.GLOBAL_STATUS WHERE VARIABLE_NAME = 'wsrep_local_replays'`
|
||||
|
||||
|
||||
#
|
||||
# nodes 2 and 3 form a galera cluster, node 2 operates as slave for native MariaDB naster in node 1
|
||||
#
|
||||
--disable_query_log
|
||||
--eval CHANGE MASTER TO MASTER_HOST='127.0.0.1', MASTER_USER='root', MASTER_PORT=$NODE_MYPORT_1;
|
||||
--enable_query_log
|
||||
START SLAVE;
|
||||
|
||||
--connection node_1
|
||||
CREATE TABLE t1 (f1 INTEGER PRIMARY KEY, f2 CHAR(1)) engine=innodb;
|
||||
INSERT INTO t1 VALUES (1, 'a');
|
||||
INSERT INTO t1 VALUES (3, 'a');
|
||||
|
||||
#
|
||||
# use statement format replication to cause a false positive conflict with async replication transaction
|
||||
# and galera replication. The conflict will be on GAP lock, and slave SQL thread should rollback
|
||||
# and replay
|
||||
#
|
||||
set binlog_format=STATEMENT;
|
||||
|
||||
SET AUTOCOMMIT=ON;
|
||||
START TRANSACTION;
|
||||
|
||||
SELECT * FROM t1 FOR UPDATE;
|
||||
UPDATE t1 SET f2 = 'c' WHERE f1 > 1;
|
||||
|
||||
--connection node_2a
|
||||
# wait for create table and inserts to be replicated from master
|
||||
SET SESSION wsrep_sync_wait = 0;
|
||||
--let $wait_condition = SELECT COUNT(*) = 2 FROM test.t1;
|
||||
--source include/wait_condition.inc
|
||||
|
||||
# wait for create table and inserts to be replicated in cluster
|
||||
--connect node_3, 127.0.0.1, root, , test, $NODE_MYPORT_3
|
||||
--connection node_3
|
||||
SET SESSION wsrep_sync_wait = 0;
|
||||
--let $wait_condition = SELECT COUNT(*) = 2 FROM test.t1;
|
||||
--source include/wait_condition.inc
|
||||
|
||||
--connection node_2a
|
||||
# Block the future commit of async replication
|
||||
--let $galera_sync_point = commit_monitor_enter_sync
|
||||
--source include/galera_set_sync_point.inc
|
||||
|
||||
# block also the applier before applying begins
|
||||
SET GLOBAL debug_dbug = "d,sync.wsrep_apply_cb";
|
||||
|
||||
#
|
||||
# now inject a conflicting insert from node 3, it will replicate with
|
||||
# earlier seqno (than async transaction) and pause before applying in node 2
|
||||
#
|
||||
--connection node_3
|
||||
INSERT INTO test.t1 VALUES (2, 'b');
|
||||
|
||||
#
|
||||
# send the update from master, this will succeed here, beceuase of async replication.
|
||||
# async replication will apply this in node 2 and pause before commit phase,
|
||||
--connection node_1
|
||||
--error 0
|
||||
COMMIT;
|
||||
|
||||
# Wait until async slave commit is blocked in node_2
|
||||
--connection node_2a
|
||||
--source include/galera_wait_sync_point.inc
|
||||
|
||||
#
|
||||
# release the applier
|
||||
# note: have to clear wsrep_apply_cb sync point first, as async replication will go for replay
|
||||
# and as this sync point, after BF applier is released to progress
|
||||
#
|
||||
SET GLOBAL debug_dbug = "";
|
||||
SET DEBUG_SYNC = "now SIGNAL signal.wsrep_apply_cb";
|
||||
|
||||
# Unblock the async slave commit
|
||||
--connection node_2a
|
||||
--source include/galera_clear_sync_point.inc
|
||||
--source include/galera_signal_sync_point.inc
|
||||
|
||||
--connection node_1
|
||||
|
||||
SELECT COUNT(*) = 1 FROM t1 WHERE f2 = 'a';
|
||||
SELECT COUNT(*) = 1 FROM t1 WHERE f2 = 'c';
|
||||
SELECT * FROM t1;
|
||||
|
||||
--connection node_2a
|
||||
|
||||
# wsrep_local_replays has increased by 1
|
||||
set session wsrep_sync_wait=15;
|
||||
--let $wsrep_local_replays_new = `SELECT VARIABLE_VALUE FROM INFORMATION_SCHEMA.GLOBAL_STATUS WHERE VARIABLE_NAME = 'wsrep_local_replays'`
|
||||
set session wsrep_sync_wait=0;
|
||||
|
||||
--disable_query_log
|
||||
--eval SELECT $wsrep_local_replays_new - $wsrep_local_replays_old = 1 AS wsrep_local_replays;
|
||||
--enable_query_log
|
||||
|
||||
#
|
||||
# replaying of async transaction should be effective, and row 3 having 'c' in f2
|
||||
#
|
||||
SELECT * FROM t1;
|
||||
SET DEBUG_SYNC = "RESET";
|
||||
|
||||
#********************************************************************************
|
||||
# test phase 2
|
||||
#********************************************************************************
|
||||
|
||||
--echo #
|
||||
--echo # test phase with real abort
|
||||
--echo #
|
||||
|
||||
--connection node_1
|
||||
|
||||
set binlog_format=ROW;
|
||||
|
||||
insert into t1 values (4, 'd');
|
||||
|
||||
SET AUTOCOMMIT=ON;
|
||||
START TRANSACTION;
|
||||
|
||||
UPDATE t1 SET f2 = 'd' WHERE f1 = 3;
|
||||
|
||||
--connection node_2a
|
||||
# wait for the last insert to be replicated from master
|
||||
--let $wait_condition = SELECT COUNT(*) = 4 FROM test.t1;
|
||||
--source include/wait_condition.inc
|
||||
|
||||
# Block the commit
|
||||
--let $galera_sync_point = commit_monitor_enter_sync
|
||||
--source include/galera_set_sync_point.inc
|
||||
|
||||
# block applier
|
||||
SET GLOBAL debug_dbug = "d,sync.wsrep_apply_cb";
|
||||
|
||||
# Inject a conflicting update from node 3
|
||||
--connection node_3
|
||||
UPDATE test.t1 SET f2 = 'e' WHERE f1 = 3;
|
||||
|
||||
# send the update from master
|
||||
--connection node_1
|
||||
--error 0
|
||||
COMMIT;
|
||||
|
||||
--connection node_2a
|
||||
|
||||
# release the applier
|
||||
SET GLOBAL debug_dbug = "";
|
||||
SET DEBUG_SYNC = "now SIGNAL signal.wsrep_apply_cb";
|
||||
|
||||
|
||||
# Unblock the async slave commit
|
||||
--connection node_2a
|
||||
--source include/galera_clear_sync_point.inc
|
||||
--source include/galera_signal_sync_point.inc
|
||||
SET DEBUG_SYNC = "RESET";
|
||||
|
||||
--connection node_2a
|
||||
|
||||
set session wsrep_sync_wait=15;
|
||||
SELECT COUNT(*) = 1 FROM test.t1 WHERE f2 = 'e';
|
||||
set session wsrep_sync_wait=0;
|
||||
|
||||
STOP SLAVE;
|
||||
RESET SLAVE;
|
||||
|
||||
DROP TABLE t1;
|
||||
|
||||
--connection node_1
|
||||
DROP TABLE t1;
|
||||
RESET MASTER;
|
|
@ -40,11 +40,19 @@ set wsrep_sync_wait=0;
|
|||
|
||||
--let $wait_condition = SELECT COUNT(*) = 2 FROM child;
|
||||
--source include/wait_condition.inc
|
||||
--let $wait_condition = SELECT COUNT(*) = 2 FROM parent;
|
||||
--source include/wait_condition.inc
|
||||
--let $wait_condition = SELECT COUNT(*) = 2 FROM grandparent;
|
||||
--source include/wait_condition.inc
|
||||
DELETE FROM grandparent WHERE id = 1;
|
||||
|
||||
--connection node_1
|
||||
--let $wait_condition = SELECT COUNT(*) = 1 FROM child;
|
||||
--source include/wait_condition.inc
|
||||
--let $wait_condition = SELECT COUNT(*) = 1 FROM parent;
|
||||
--source include/wait_condition.inc
|
||||
--let $wait_condition = SELECT COUNT(*) = 1 FROM grandparent;
|
||||
--source include/wait_condition.inc
|
||||
SELECT COUNT(*), COUNT(*) = 0 FROM parent WHERE grandparent_id = 1;
|
||||
SELECT COUNT(*), COUNT(*) = 0 FROM child WHERE parent_id = 1;
|
||||
|
||||
|
|
|
@ -8,6 +8,9 @@
|
|||
CREATE TABLE t1 (f1 INTEGER PRIMARY KEY) Engine=InnoDB;
|
||||
|
||||
--connection node_2
|
||||
--let $wait_condition = SELECT COUNT(*) = 1 FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME = 't1'
|
||||
--source include/wait_condition.inc
|
||||
|
||||
SET SESSION wsrep_OSU_method = "RSU";
|
||||
ALTER TABLE t1 ADD COLUMN f2 INTEGER;
|
||||
SELECT COUNT(*) = 2 FROM INFORMATION_SCHEMA.COLUMNS WHERE TABLE_NAME = 't1';
|
||||
|
|
|
@ -7,7 +7,7 @@
|
|||
--source include/have_innodb.inc
|
||||
|
||||
--connection node_1
|
||||
CREATE TABLE ten (f1 INTEGER) ENGINE=InnoDB;
|
||||
CREATE TABLE ten (f1 INTEGER NOT NULL PRIMARY KEY) ENGINE=InnoDB;
|
||||
INSERT INTO ten VALUES (1),(2),(3),(4),(5),(6),(7),(8),(9),(10);
|
||||
|
||||
CREATE TABLE t1 (f1 INTEGER AUTO_INCREMENT PRIMARY KEY, f2 INTEGER) ENGINE=InnoDB;
|
||||
|
@ -83,6 +83,8 @@ SET GLOBAL auto_increment_offset = 1;
|
|||
CREATE TABLE t1 (f1 INTEGER AUTO_INCREMENT PRIMARY KEY, f2 INTEGER) ENGINE=InnoDB;
|
||||
|
||||
--connection node_2a
|
||||
--let $wait_condition = SELECT COUNT(*) = 1 FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME = 't1'
|
||||
--source include/wait_condition.inc
|
||||
|
||||
ALTER TABLE t1 AUTO_INCREMENT=100;
|
||||
|
||||
|
|
|
@ -413,6 +413,7 @@ CREATE TABLE x AS SELECT * FROM t1;
|
|||
ERROR XAE07: XAER_RMFAIL: The command cannot be executed when global transaction is in the ACTIVE state
|
||||
connect con1,localhost,root,,test;
|
||||
SET foreign_key_checks= OFF, innodb_lock_wait_timeout= 1;
|
||||
SET lock_wait_timeout=5;
|
||||
ALTER TABLE t1 ADD FOREIGN KEY f (a) REFERENCES t1 (pk), LOCK=EXCLUSIVE;
|
||||
ERROR HY000: Lock wait timeout exceeded; try restarting transaction
|
||||
disconnect con1;
|
||||
|
|
|
@ -74,7 +74,6 @@ DELETE FROM t1;
|
|||
# Kill the server
|
||||
disconnect ddl;
|
||||
# restart
|
||||
SET @saved_frequency= @@GLOBAL.innodb_purge_rseg_truncate_frequency;
|
||||
SET GLOBAL innodb_purge_rseg_truncate_frequency=1;
|
||||
FOUND 3 /\[Note\] InnoDB: Rolled back recovered transaction / in mysqld.1.err
|
||||
SELECT * FROM t1;
|
||||
|
@ -138,6 +137,23 @@ header=0x080008030000 (id=0x000000000000000100)
|
|||
UNLOCK TABLES;
|
||||
DELETE FROM t2;
|
||||
InnoDB 0 transactions not purged
|
||||
#
|
||||
# MDEV-24323 Crash on recovery after kill during instant ADD COLUMN
|
||||
#
|
||||
connect ddl, localhost, root;
|
||||
CREATE TABLE t3(id INT PRIMARY KEY, c2 INT, v2 INT AS(c2) VIRTUAL, UNIQUE(v2))
|
||||
ENGINE=InnoDB;
|
||||
INSERT INTO t3 SET id=1,c2=1;
|
||||
SET DEBUG_SYNC='innodb_alter_inplace_before_commit SIGNAL ddl WAIT_FOR ever';
|
||||
ALTER TABLE t3 ADD COLUMN c3 TEXT NOT NULL DEFAULT 'sic transit gloria mundi';
|
||||
connection default;
|
||||
SET DEBUG_SYNC='now WAIT_FOR ddl';
|
||||
SET GLOBAL innodb_flush_log_at_trx_commit=1;
|
||||
SET debug_dbug='+d,dict_sys_mutex_avoid';
|
||||
INSERT INTO t1 VALUES(0,0);
|
||||
# Kill the server
|
||||
disconnect ddl;
|
||||
# restart
|
||||
SHOW CREATE TABLE t1;
|
||||
Table Create Table
|
||||
t1 CREATE TABLE `t1` (
|
||||
|
@ -154,6 +170,14 @@ t2 CREATE TABLE `t2` (
|
|||
PRIMARY KEY (`id`),
|
||||
UNIQUE KEY `c2` (`c2`)
|
||||
) ENGINE=InnoDB DEFAULT CHARSET=latin1 ROW_FORMAT=REDUNDANT
|
||||
DROP TABLE t1,t2;
|
||||
SHOW CREATE TABLE t3;
|
||||
Table Create Table
|
||||
t3 CREATE TABLE `t3` (
|
||||
`id` int(11) NOT NULL,
|
||||
`c2` int(11) DEFAULT NULL,
|
||||
`v2` int(11) GENERATED ALWAYS AS (`c2`) VIRTUAL,
|
||||
PRIMARY KEY (`id`),
|
||||
UNIQUE KEY `v2` (`v2`)
|
||||
) ENGINE=InnoDB DEFAULT CHARSET=latin1
|
||||
DROP TABLE t1,t2,t3;
|
||||
db.opt
|
||||
SET GLOBAL innodb_purge_rseg_truncate_frequency=@saved_frequency;
|
||||
|
|
|
@ -414,6 +414,7 @@ INSERT INTO t1 VALUES (1,2);
|
|||
CREATE TABLE x AS SELECT * FROM t1;
|
||||
--connect (con1,localhost,root,,test)
|
||||
SET foreign_key_checks= OFF, innodb_lock_wait_timeout= 1;
|
||||
SET lock_wait_timeout=5;
|
||||
--error ER_LOCK_WAIT_TIMEOUT
|
||||
ALTER TABLE t1 ADD FOREIGN KEY f (a) REFERENCES t1 (pk), LOCK=EXCLUSIVE;# Cleanup
|
||||
--disconnect con1
|
||||
|
|
|
@ -91,7 +91,6 @@ DELETE FROM t1;
|
|||
disconnect ddl;
|
||||
--source include/start_mysqld.inc
|
||||
|
||||
SET @saved_frequency= @@GLOBAL.innodb_purge_rseg_truncate_frequency;
|
||||
SET GLOBAL innodb_purge_rseg_truncate_frequency=1;
|
||||
|
||||
let SEARCH_FILE= $MYSQLTEST_VARDIR/log/mysqld.1.err;
|
||||
|
@ -177,11 +176,32 @@ UNLOCK TABLES;
|
|||
DELETE FROM t2;
|
||||
--source include/wait_all_purged.inc
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-24323 Crash on recovery after kill during instant ADD COLUMN
|
||||
--echo #
|
||||
connect ddl, localhost, root;
|
||||
CREATE TABLE t3(id INT PRIMARY KEY, c2 INT, v2 INT AS(c2) VIRTUAL, UNIQUE(v2))
|
||||
ENGINE=InnoDB;
|
||||
INSERT INTO t3 SET id=1,c2=1;
|
||||
|
||||
SET DEBUG_SYNC='innodb_alter_inplace_before_commit SIGNAL ddl WAIT_FOR ever';
|
||||
--send
|
||||
ALTER TABLE t3 ADD COLUMN c3 TEXT NOT NULL DEFAULT 'sic transit gloria mundi';
|
||||
|
||||
connection default;
|
||||
SET DEBUG_SYNC='now WAIT_FOR ddl';
|
||||
SET GLOBAL innodb_flush_log_at_trx_commit=1;
|
||||
SET debug_dbug='+d,dict_sys_mutex_avoid';
|
||||
INSERT INTO t1 VALUES(0,0);
|
||||
|
||||
--source include/kill_mysqld.inc
|
||||
disconnect ddl;
|
||||
--source include/start_mysqld.inc
|
||||
|
||||
SHOW CREATE TABLE t1;
|
||||
SHOW CREATE TABLE t2;
|
||||
DROP TABLE t1,t2;
|
||||
SHOW CREATE TABLE t3;
|
||||
DROP TABLE t1,t2,t3;
|
||||
|
||||
--remove_files_wildcard $MYSQLD_DATADIR/test #sql*.frm
|
||||
--list_files $MYSQLD_DATADIR/test
|
||||
|
||||
SET GLOBAL innodb_purge_rseg_truncate_frequency=@saved_frequency;
|
||||
|
|
146
mysql-test/suite/mariabackup/include/corrupt-page.pl
Normal file
146
mysql-test/suite/mariabackup/include/corrupt-page.pl
Normal file
|
@ -0,0 +1,146 @@
|
|||
use strict;
|
||||
use warnings;
|
||||
use Fcntl qw(:DEFAULT :seek);
|
||||
do "$ENV{MTR_SUITE_DIR}/../innodb/include/crc32.pl";
|
||||
|
||||
sub corrupt_space_page_id {
|
||||
my $file_name = shift;
|
||||
my @pages_to_corrupt = @_;
|
||||
|
||||
my $page_size = $ENV{INNODB_PAGE_SIZE};
|
||||
|
||||
sysopen my $ibd_file, $file_name, O_RDWR || die "Cannot open $file_name\n";
|
||||
sysread($ibd_file, $_, 38) || die "Cannot read $file_name\n";
|
||||
my $space = unpack("x[34]N", $_);
|
||||
foreach my $page_no (@pages_to_corrupt) {
|
||||
$space += 10; # generate wrong space id
|
||||
sysseek($ibd_file, $page_size * $page_no, SEEK_SET)
|
||||
|| die "Cannot seek $file_name\n";
|
||||
|
||||
my $head = pack("Nx[18]", $page_no + 10); # generate wrong page number
|
||||
my $body = chr(0) x ($page_size - 38 - 8);
|
||||
|
||||
# Calculate innodb_checksum_algorithm=crc32 for the unencrypted page.
|
||||
# The following bytes are excluded:
|
||||
# bytes 0..3 (the checksum is stored there)
|
||||
# bytes 26..37 (encryption key version, post-encryption checksum, tablespace id)
|
||||
# bytes $page_size-8..$page_size-1 (checksum, LSB of FIL_PAGE_LSN)
|
||||
my $polynomial = 0x82f63b78; # CRC-32C
|
||||
my $ck = mycrc32($head, 0, $polynomial) ^ mycrc32($body, 0, $polynomial);
|
||||
|
||||
my $page= pack("N",$ck).$head.pack("NNN",1,$ck,$space).$body.pack("Nx[4]",$ck);
|
||||
die unless syswrite($ibd_file, $page, $page_size) == $page_size;
|
||||
}
|
||||
close $ibd_file;
|
||||
}
|
||||
|
||||
sub extend_space {
|
||||
my $file_name = shift;
|
||||
my $n_pages = shift;
|
||||
|
||||
my $page_size = $ENV{INNODB_PAGE_SIZE};
|
||||
my $page;
|
||||
|
||||
sysopen my $ibd_file, $file_name, O_RDWR || die "Cannot open $file_name\n";
|
||||
sysread($ibd_file, $page, $page_size)
|
||||
|| die "Cannot read $file_name\n";
|
||||
my $size = unpack("N", substr($page, 46, 4));
|
||||
my $packed_new_size = pack("N", $size + $n_pages);
|
||||
substr($page, 46, 4, $packed_new_size);
|
||||
|
||||
my $head = substr($page, 4, 22);
|
||||
my $body = substr($page, 38, $page_size - 38 - 8);
|
||||
my $polynomial = 0x82f63b78; # CRC-32C
|
||||
my $ck = mycrc32($head, 0, $polynomial) ^ mycrc32($body, 0, $polynomial);
|
||||
my $packed_ck = pack("N", $ck);
|
||||
substr($page, 0, 4, $packed_ck);
|
||||
substr($page, $page_size - 8, 4, $packed_ck);
|
||||
|
||||
sysseek($ibd_file, 0, SEEK_SET)
|
||||
|| die "Cannot seek $file_name\n";
|
||||
die unless syswrite($ibd_file, $page, $page_size) == $page_size;
|
||||
|
||||
sysseek($ibd_file, 0, SEEK_END)
|
||||
|| die "Cannot seek $file_name\n";
|
||||
my $pages_size = $page_size*$n_pages;
|
||||
my $pages = chr(0) x $pages_size;
|
||||
die unless syswrite($ibd_file, $pages, $pages_size) == $pages_size;
|
||||
close $ibd_file;
|
||||
return $size;
|
||||
}
|
||||
|
||||
sub die_if_page_is_not_zero {
|
||||
my $file_name = shift;
|
||||
my @pages_to_check = @_;
|
||||
|
||||
no locale;
|
||||
my $page_size = $ENV{INNODB_PAGE_SIZE};
|
||||
my $zero_page = chr(0) x $page_size;
|
||||
sysopen my $ibd_file, $file_name, O_RDWR || die "Cannot open $file_name\n";
|
||||
foreach my $page_no_to_check (@pages_to_check) {
|
||||
sysseek($ibd_file, $page_size*$page_no_to_check, SEEK_SET) ||
|
||||
die "Cannot seek $file_name\n";
|
||||
sysread($ibd_file, my $read_page, $page_size) ||
|
||||
die "Cannot read $file_name\n";
|
||||
die "The page $page_no_to_check is not zero-filed in $file_name"
|
||||
if ($read_page cmp $zero_page);
|
||||
}
|
||||
close $ibd_file;
|
||||
}
|
||||
|
||||
sub print_corrupted_pages_file {
|
||||
my $file_in = shift;
|
||||
my $file_out = shift;
|
||||
open my $fh, '<', $file_in || die $!;
|
||||
my $line_number = 0;
|
||||
my $space = {};
|
||||
my @spaces;
|
||||
while (my $line = <$fh>) {
|
||||
++$line_number;
|
||||
if ($line_number & 1) {
|
||||
my ($name, $id) = split(/ /, $line);
|
||||
$space->{name} = $name;
|
||||
}
|
||||
else {
|
||||
$space->{pages} = $line;
|
||||
push (@spaces, $space);
|
||||
$space = {};
|
||||
}
|
||||
}
|
||||
close $fh;
|
||||
my @sorted_spaces = sort { $a->{name} cmp $b->{name} } @spaces;
|
||||
open $fh, '>', $file_out || die $!;
|
||||
foreach my $space (@sorted_spaces) {
|
||||
print $fh $space->{name};
|
||||
print $fh "\n";
|
||||
print $fh $space->{pages};
|
||||
}
|
||||
close $fh;
|
||||
}
|
||||
|
||||
sub append_corrupted_pages {
|
||||
my $file_name = shift;
|
||||
my $space_name = shift;
|
||||
my $pages = shift;
|
||||
open my $fh, '<', $file_name || die $!;
|
||||
my $line_number = 0;
|
||||
my $space_line;
|
||||
while (my $line = <$fh>) {
|
||||
++$line_number;
|
||||
if ($line_number & 1) {
|
||||
my ($name, $id) = split(/ /, $line);
|
||||
if ($name eq $space_name) {
|
||||
$space_line = $line;
|
||||
last;
|
||||
}
|
||||
}
|
||||
}
|
||||
close $fh;
|
||||
if (not defined $space_line) {
|
||||
die "Can't find requested space $space_name in file $file_name";
|
||||
}
|
||||
open $fh, '>>', $file_name || die $!;
|
||||
print $fh $space_line;
|
||||
print $fh "$pages\n";
|
||||
close $fh;
|
||||
}
|
|
@ -22,7 +22,7 @@ INSERT into t1 values(1);
|
|||
--let after_copy_test_t2=DROP TABLE test.t2
|
||||
--let after_copy_test_t3=CREATE INDEX a_i ON test.t3(i);
|
||||
--let before_copy_test_t10=DROP TABLE test.t10
|
||||
--let wait_innodb_redo_before_copy=test/t10
|
||||
--let wait_innodb_redo_before_copy_test_t10 = 1
|
||||
|
||||
# mariabackup should crash with assertion if MDEV-24026 is not fixed
|
||||
exec $XTRABACKUP --defaults-file=$MYSQLTEST_VARDIR/my.cnf --backup --target-dir=$incremental_dir --incremental-basedir=$basedir --dbug=+d,mariabackup_events,mariabackup_inject_code;
|
||||
|
|
1
mysql-test/suite/mariabackup/log_page_corruption.opt
Normal file
1
mysql-test/suite/mariabackup/log_page_corruption.opt
Normal file
|
@ -0,0 +1 @@
|
|||
--innodb-checksum-algorithm=crc32
|
145
mysql-test/suite/mariabackup/log_page_corruption.result
Normal file
145
mysql-test/suite/mariabackup/log_page_corruption.result
Normal file
|
@ -0,0 +1,145 @@
|
|||
########
|
||||
# Test for generating "innodb_corrupted_pages" file during full and
|
||||
# incremental backup, including DDL processing
|
||||
###
|
||||
|
||||
CREATE TABLE t1_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t2_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t3(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t5_corrupted_to_rename(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t6_corrupted_to_drop(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t7_corrupted_to_alter(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t1_inc_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t2_inc_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t3_inc(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t5_inc_corrupted_to_rename(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t6_inc_corrupted_to_drop(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t7_inc_corrupted_to_alter(c INT) ENGINE INNODB;
|
||||
INSERT INTO t1_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t2_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t3 VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t5_corrupted_to_rename VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t6_corrupted_to_drop VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t7_corrupted_to_alter VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
# Corrupt tables
|
||||
# restart
|
||||
# Backup must fail due to page corruption
|
||||
FOUND 1 /Database page corruption detected.*/ in backup.log
|
||||
# "innodb_corrupted_pages" file must not exist
|
||||
# Backup must fail, but "innodb_corrupted_pages" file must be created due to --log-innodb-page-corruption option
|
||||
FOUND 1 /Database page corruption detected.*/ in backup.log
|
||||
--- "innodb_corrupted_pages" file content: ---
|
||||
test/t1_corrupted
|
||||
6 8 9
|
||||
test/t2_corrupted
|
||||
7 8 10
|
||||
test/t4_corrupted_new
|
||||
1
|
||||
test/t5_corrupted_to_rename_renamed
|
||||
6
|
||||
test/t7_corrupted_to_alter
|
||||
3
|
||||
------
|
||||
INSERT INTO t1_inc_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t2_inc_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t3_inc VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
# restart
|
||||
# Backup must fail, but "innodb_corrupted_pages" file must be created due to --log-innodb-page-corruption option
|
||||
--- "innodb_corrupted_pages" file content: ---
|
||||
test/t1_corrupted
|
||||
6 8 9
|
||||
test/t1_inc_corrupted
|
||||
6 8 9
|
||||
test/t2_corrupted
|
||||
7 8 10
|
||||
test/t2_inc_corrupted
|
||||
7 8 10
|
||||
test/t4_inc_corrupted_new
|
||||
1
|
||||
test/t5_corrupted_to_rename_renamed
|
||||
6
|
||||
test/t5_inc_corrupted_to_rename_renamed
|
||||
6
|
||||
test/t7_inc_corrupted_to_alter
|
||||
3
|
||||
------
|
||||
# Check if corrupted pages were copied to delta files, and non-corrupted pages are not copied.
|
||||
DROP TABLE t1_corrupted;
|
||||
DROP TABLE t2_corrupted;
|
||||
DROP TABLE t4_corrupted_new;
|
||||
DROP TABLE t5_corrupted_to_rename_renamed;
|
||||
DROP TABLE t7_corrupted_to_alter;
|
||||
DROP TABLE t1_inc_corrupted;
|
||||
DROP TABLE t2_inc_corrupted;
|
||||
DROP TABLE t4_inc_corrupted_new;
|
||||
DROP TABLE t5_inc_corrupted_to_rename_renamed;
|
||||
DROP TABLE t7_inc_corrupted_to_alter;
|
||||
|
||||
########
|
||||
# Test for --prepare with "innodb_corrupted_pages" file
|
||||
###
|
||||
|
||||
# Extend some tablespace and corrupt extended pages for full backup
|
||||
# restart
|
||||
# Full backup with --log-innodb-page-corruption
|
||||
--- "innodb_corrupted_pages" file content: ---
|
||||
test/t3
|
||||
6 8
|
||||
------
|
||||
# Extend some tablespace and corrupt extended pages for incremental backup
|
||||
# restart
|
||||
# Incremental backup --log-innodb-page-corruption
|
||||
--- "innodb_corrupted_pages" file content: ---
|
||||
test/t3
|
||||
6 8
|
||||
test/t3_inc
|
||||
6 8
|
||||
------
|
||||
# Full backup prepare
|
||||
# "innodb_corrupted_pages" file must not exist after successful prepare
|
||||
FOUND 1 /was successfuly fixed.*/ in backup.log
|
||||
# Check that fixed pages are zero-filled
|
||||
# Incremental backup prepare
|
||||
# "innodb_corrupted_pages" file must not exist after successful prepare
|
||||
# do not remove "innodb_corrupted_pages" in incremental dir
|
||||
FOUND 1 /was successfuly fixed.*/ in backup.log
|
||||
# Check that fixed pages are zero-filled
|
||||
# shutdown server
|
||||
# remove datadir
|
||||
# xtrabackup move back
|
||||
# restart
|
||||
SELECT * FROM t3;
|
||||
c
|
||||
3
|
||||
4
|
||||
5
|
||||
6
|
||||
7
|
||||
8
|
||||
9
|
||||
SELECT * FROM t3_inc;
|
||||
c
|
||||
3
|
||||
4
|
||||
5
|
||||
6
|
||||
7
|
||||
8
|
||||
9
|
||||
# Test the case when not all corrupted pages are fixed
|
||||
|
||||
# Add some fake corrupted pages
|
||||
# Full backup prepare
|
||||
FOUND 1 /Error: corrupted page.*/ in backup.log
|
||||
--- "innodb_corrupted_pages" file content: ---
|
||||
test/t3
|
||||
3
|
||||
------
|
||||
# Incremental backup prepare
|
||||
FOUND 1 /Error: corrupted page.*/ in backup.log
|
||||
--- "innodb_corrupted_pages" file content: ---
|
||||
test/t3
|
||||
3
|
||||
------
|
||||
DROP TABLE t3;
|
||||
DROP TABLE t3_inc;
|
426
mysql-test/suite/mariabackup/log_page_corruption.test
Normal file
426
mysql-test/suite/mariabackup/log_page_corruption.test
Normal file
|
@ -0,0 +1,426 @@
|
|||
--source include/have_debug.inc
|
||||
|
||||
--echo ########
|
||||
--echo # Test for generating "innodb_corrupted_pages" file during full and
|
||||
--echo # incremental backup, including DDL processing
|
||||
--echo ###
|
||||
--echo
|
||||
|
||||
CREATE TABLE t1_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t2_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t3(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t5_corrupted_to_rename(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t6_corrupted_to_drop(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t7_corrupted_to_alter(c INT) ENGINE INNODB;
|
||||
|
||||
CREATE TABLE t1_inc_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t2_inc_corrupted(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t3_inc(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t5_inc_corrupted_to_rename(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t6_inc_corrupted_to_drop(c INT) ENGINE INNODB;
|
||||
CREATE TABLE t7_inc_corrupted_to_alter(c INT) ENGINE INNODB;
|
||||
|
||||
# Fill tables with several pages
|
||||
INSERT INTO t1_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t2_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t3 VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t5_corrupted_to_rename VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t6_corrupted_to_drop VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t7_corrupted_to_alter VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
|
||||
--let MYSQLD_DATADIR=`select @@datadir`
|
||||
--let INNODB_PAGE_SIZE=`select @@innodb_page_size`
|
||||
|
||||
--source include/shutdown_mysqld.inc
|
||||
--echo # Corrupt tables
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
my $schema = "$ENV{MYSQLD_DATADIR}/test";
|
||||
|
||||
my $last_page_no = extend_space("$schema/t1_corrupted.ibd", 4);
|
||||
corrupt_space_page_id("$schema/t1_corrupted.ibd",
|
||||
$last_page_no, $last_page_no + 2, $last_page_no + 3);
|
||||
|
||||
$last_page_no = extend_space("$schema/t2_corrupted.ibd", 5);
|
||||
corrupt_space_page_id("$schema/t2_corrupted.ibd",
|
||||
$last_page_no + 1, $last_page_no + 2, $last_page_no + 4);
|
||||
|
||||
$last_page_no = extend_space("$schema/t5_corrupted_to_rename.ibd", 1);
|
||||
corrupt_space_page_id("$schema/t5_corrupted_to_rename.ibd", $last_page_no);
|
||||
|
||||
$last_page_no = extend_space("$schema/t6_corrupted_to_drop.ibd", );
|
||||
corrupt_space_page_id("$schema/t6_corrupted_to_drop.ibd", $last_page_no);
|
||||
EOF
|
||||
--source include/start_mysqld.inc
|
||||
|
||||
--let targetdir=$MYSQLTEST_VARDIR/tmp/backup
|
||||
--let $backuplog=$MYSQLTEST_VARDIR/tmp/backup.log
|
||||
--let corrupted_pages_file = $targetdir/innodb_corrupted_pages
|
||||
--let corrupted_pages_file_filt = $MYSQLTEST_VARDIR/tmp/innodb_corrupted_pages_filt
|
||||
--let perl_result_file=$MYSQLTEST_VARDIR/tmp/perl_result
|
||||
|
||||
--echo # Backup must fail due to page corruption
|
||||
--disable_result_log
|
||||
--error 1
|
||||
exec $XTRABACKUP --defaults-file=$MYSQLTEST_VARDIR/my.cnf --backup --target-dir=$targetdir > $backuplog;
|
||||
--enable_result_log
|
||||
|
||||
--let SEARCH_PATTERN=Database page corruption detected.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
--echo # "innodb_corrupted_pages" file must not exist
|
||||
--error 1
|
||||
--file_exists $corrupted_pages_file
|
||||
--rmdir $targetdir
|
||||
|
||||
--let after_load_tablespaces=CREATE TABLE test.t4_corrupted_new ENGINE=INNODB SELECT UUID() from test.seq_1_to_10
|
||||
--let add_corrupted_page_for_test_t4_corrupted_new=1
|
||||
--let after_copy_test_t5_corrupted_to_rename=RENAME TABLE test.t5_corrupted_to_rename TO test.t5_corrupted_to_rename_renamed
|
||||
--let after_copy_test_t6_corrupted_to_drop=DROP TABLE test.t6_corrupted_to_drop
|
||||
--let after_copy_test_t7_corrupted_to_alter=ALTER TABLE test.t7_corrupted_to_alter ADD COLUMN (d INT)
|
||||
--let add_corrupted_page_for_test_t7_corrupted_to_alter=3
|
||||
|
||||
--echo # Backup must fail, but "innodb_corrupted_pages" file must be created due to --log-innodb-page-corruption option
|
||||
--disable_result_log
|
||||
--error 1
|
||||
--exec $XTRABACKUP --defaults-file=$MYSQLTEST_VARDIR/my.cnf --backup --log-innodb-page-corruption --target-dir=$targetdir --dbug=+d,mariabackup_events,mariabackup_inject_code > $backuplog
|
||||
--enable_result_log
|
||||
|
||||
--let SEARCH_PATTERN=Database page corruption detected.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
--echo --- "innodb_corrupted_pages" file content: ---
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
print_corrupted_pages_file($ENV{corrupted_pages_file},
|
||||
$ENV{corrupted_pages_file_filt});
|
||||
EOF
|
||||
--cat_file $corrupted_pages_file_filt
|
||||
--echo ------
|
||||
--let after_load_tablespaces=
|
||||
--let add_corrupted_page_for_test_t4_corrupted_new=
|
||||
--let after_copy_test_t5_corrupted_to_rename=
|
||||
--let after_copy_test_t6_corrupted_to_drop=
|
||||
--let after_copy_test_t7_corrupted_to_alter=
|
||||
--let add_corrupted_page_for_test_t7_corrupted_to_alter=
|
||||
# Fill tables for incremental backup with several pages
|
||||
INSERT INTO t1_inc_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t2_inc_corrupted VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
INSERT INTO t3_inc VALUES (3), (4), (5), (6), (7), (8), (9);
|
||||
|
||||
--source include/shutdown_mysqld.inc
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
my $schema="$ENV{MYSQLD_DATADIR}/test";
|
||||
|
||||
open(my $fh, '>', $ENV{perl_result_file}) or die $!;
|
||||
|
||||
my $last_page_no = extend_space("$schema/t1_inc_corrupted.ibd", 4);
|
||||
corrupt_space_page_id("$schema/t1_inc_corrupted.ibd",
|
||||
$last_page_no, $last_page_no + 2, $last_page_no + 3);
|
||||
print $fh "$last_page_no\n";
|
||||
|
||||
$last_page_no = extend_space("$schema/t2_inc_corrupted.ibd", 5);
|
||||
corrupt_space_page_id("$schema/t2_inc_corrupted.ibd",
|
||||
$last_page_no + 1, $last_page_no + 2, $last_page_no + 4);
|
||||
print $fh "$last_page_no\n";
|
||||
|
||||
$last_page_no = extend_space("$schema/t5_inc_corrupted_to_rename.ibd", 1);
|
||||
corrupt_space_page_id("$schema/t5_inc_corrupted_to_rename.ibd", $last_page_no);
|
||||
print $fh "$last_page_no\n";
|
||||
|
||||
$last_page_no = extend_space("$schema/t6_inc_corrupted_to_drop.ibd", );
|
||||
corrupt_space_page_id("$schema/t6_inc_corrupted_to_drop.ibd", $last_page_no);
|
||||
|
||||
close $fh;
|
||||
EOF
|
||||
--source include/start_mysqld.inc
|
||||
|
||||
--let incdir=$MYSQLTEST_VARDIR/tmp/backup_inc
|
||||
|
||||
--let after_load_tablespaces=CREATE TABLE test.t4_inc_corrupted_new ENGINE=INNODB SELECT UUID() from test.seq_1_to_10
|
||||
--let add_corrupted_page_for_test_t4_inc_corrupted_new=1
|
||||
--let after_copy_test_t5_inc_corrupted_to_rename=RENAME TABLE test.t5_inc_corrupted_to_rename TO test.t5_inc_corrupted_to_rename_renamed
|
||||
--let after_copy_test_t6_inc_corrupted_to_drop=DROP TABLE test.t6_inc_corrupted_to_drop
|
||||
--let after_copy_test_t7_inc_corrupted_to_alter=ALTER TABLE test.t7_inc_corrupted_to_alter ADD COLUMN (d INT)
|
||||
--let add_corrupted_page_for_test_t7_inc_corrupted_to_alter=3
|
||||
|
||||
--echo # Backup must fail, but "innodb_corrupted_pages" file must be created due to --log-innodb-page-corruption option
|
||||
--disable_result_log
|
||||
--error 1
|
||||
--exec $XTRABACKUP --defaults-file=$MYSQLTEST_VARDIR/my.cnf --backup --log-innodb-page-corruption --target-dir=$incdir --incremental-basedir=$targetdir --dbug=+d,mariabackup_events,mariabackup_inject_code > $backuplog
|
||||
--disable_result_log
|
||||
|
||||
--let after_load_tablespaces=
|
||||
--let add_corrupted_page_for_test_t4_inc_corrupted_new=
|
||||
--let after_copy_test_t5_inc_corrupted_to_rename=
|
||||
--let after_copy_test_t6_inc_corrupted_to_drop=
|
||||
--let after_copy_test_t7_inc_corrupted_to_alter=
|
||||
--let add_corrupted_page_for_test_t7_inc_corrupted_to_alter=
|
||||
|
||||
--let SEARCH_PATTERN=Database page corruption detected.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
--let corrupted_pages_file = $incdir/innodb_corrupted_pages
|
||||
--echo --- "innodb_corrupted_pages" file content: ---
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
print_corrupted_pages_file($ENV{corrupted_pages_file},
|
||||
$ENV{corrupted_pages_file_filt});
|
||||
EOF
|
||||
--cat_file $corrupted_pages_file_filt
|
||||
--echo ------
|
||||
|
||||
--echo # Check if corrupted pages were copied to delta files, and non-corrupted pages are not copied.
|
||||
perl;
|
||||
use strict;
|
||||
use warnings;
|
||||
my $schema = "$ENV{incdir}/test";
|
||||
|
||||
open(my $fh, '<', $ENV{perl_result_file}) or die $!;
|
||||
|
||||
my $last_page_no = <$fh>;
|
||||
die_if_no_pages("$schema/t1_corrupted.ibd.delta",
|
||||
$last_page_no, $last_page_no + 2, $last_page_no + 3);
|
||||
|
||||
$last_page_no = <$fh>;
|
||||
die_if_no_pages("$schema/t2_corrupted.ibd.delta",
|
||||
$last_page_no + 1, $last_page_no + 2, $last_page_no + 4);
|
||||
|
||||
$last_page_no = <$fh>;
|
||||
die_if_no_pages("$schema/t5_corrupted_to_rename_renamed.ibd.delta",
|
||||
$last_page_no);
|
||||
|
||||
close $fh;
|
||||
|
||||
die_if_not_empty("$schema/t3.ibd.delta");
|
||||
|
||||
sub read_first_page_from_delta {
|
||||
my $file_name = shift;
|
||||
my $pages_count = shift;
|
||||
|
||||
open my $file, '<:raw', $file_name || die "Cannot open $file_name\n";
|
||||
read $file, my $buffer, $pages_count*4 || die "Cannot read $file_name\n";
|
||||
close $file;
|
||||
|
||||
return unpack("N[$pages_count]", $buffer);
|
||||
}
|
||||
|
||||
sub die_if_no_pages {
|
||||
my $file_name = shift;
|
||||
my @check_pages = @_;
|
||||
my @read_pages =
|
||||
read_first_page_from_delta($file_name, scalar(@check_pages) + 1);
|
||||
for (my $i = 1; $i < @check_pages + 1; ++$i) {
|
||||
my $check_page_no = $check_pages[$i - 1];
|
||||
die "Corrupted page $check_page_no was not copied to $file_name."
|
||||
if ($i >= @read_pages || $read_pages[$i] != $check_page_no);
|
||||
}
|
||||
}
|
||||
|
||||
sub die_if_not_empty {
|
||||
my $file_name = shift;
|
||||
my ($magic, $full) = read_first_page_from_delta($file_name, 2);
|
||||
die "Delta $file_name must be empty."
|
||||
if ($full != 0xFFFFFFFF);
|
||||
}
|
||||
EOF
|
||||
--rmdir $incdir
|
||||
--rmdir $targetdir
|
||||
|
||||
DROP TABLE t1_corrupted;
|
||||
DROP TABLE t2_corrupted;
|
||||
DROP TABLE t4_corrupted_new;
|
||||
DROP TABLE t5_corrupted_to_rename_renamed;
|
||||
DROP TABLE t7_corrupted_to_alter;
|
||||
DROP TABLE t1_inc_corrupted;
|
||||
DROP TABLE t2_inc_corrupted;
|
||||
DROP TABLE t4_inc_corrupted_new;
|
||||
DROP TABLE t5_inc_corrupted_to_rename_renamed;
|
||||
DROP TABLE t7_inc_corrupted_to_alter;
|
||||
|
||||
--echo
|
||||
--echo ########
|
||||
--echo # Test for --prepare with "innodb_corrupted_pages" file
|
||||
--echo ###
|
||||
--echo
|
||||
|
||||
--echo # Extend some tablespace and corrupt extended pages for full backup
|
||||
--source include/shutdown_mysqld.inc
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
my $schema="$ENV{MYSQLD_DATADIR}/test";
|
||||
my $last_page_no = extend_space("$schema/t3.ibd", 3);
|
||||
corrupt_space_page_id("$schema/t3.ibd", $last_page_no, $last_page_no + 2);
|
||||
open(my $fh, '>', $ENV{perl_result_file}) or die $!;
|
||||
print $fh "$last_page_no\n";
|
||||
close $fh;
|
||||
EOF
|
||||
--source include/start_mysqld.inc
|
||||
|
||||
--echo # Full backup with --log-innodb-page-corruption
|
||||
--disable_result_log
|
||||
--error 1
|
||||
--exec $XTRABACKUP --defaults-file=$MYSQLTEST_VARDIR/my.cnf --backup --log-innodb-page-corruption --target-dir=$targetdir
|
||||
--enable_result_log
|
||||
--let corrupted_pages_file = $targetdir/innodb_corrupted_pages
|
||||
--echo --- "innodb_corrupted_pages" file content: ---
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
print_corrupted_pages_file($ENV{corrupted_pages_file},
|
||||
$ENV{corrupted_pages_file_filt});
|
||||
EOF
|
||||
--cat_file $corrupted_pages_file_filt
|
||||
--echo ------
|
||||
|
||||
--echo # Extend some tablespace and corrupt extended pages for incremental backup
|
||||
--source include/shutdown_mysqld.inc
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
my $schema="$ENV{MYSQLD_DATADIR}/test";
|
||||
my $last_page_no = extend_space("$schema/t3_inc.ibd", 3);
|
||||
corrupt_space_page_id("$schema/t3_inc.ibd", $last_page_no, $last_page_no + 2);
|
||||
open(my $fh, '>>', $ENV{perl_result_file}) or die $!;
|
||||
print $fh "$last_page_no";
|
||||
close $fh;
|
||||
EOF
|
||||
--source include/start_mysqld.inc
|
||||
|
||||
--echo # Incremental backup --log-innodb-page-corruption
|
||||
--disable_result_log
|
||||
--error 1
|
||||
--exec $XTRABACKUP --defaults-file=$MYSQLTEST_VARDIR/my.cnf --backup --log-innodb-page-corruption --target-dir=$incdir --incremental-basedir=$targetdir --dbug=+d,mariabackup_events,mariabackup_inject_code > $backuplog
|
||||
--disable_result_log
|
||||
--let corrupted_pages_file = $incdir/innodb_corrupted_pages
|
||||
--echo --- "innodb_corrupted_pages" file content: ---
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
print_corrupted_pages_file($ENV{corrupted_pages_file},
|
||||
$ENV{corrupted_pages_file_filt});
|
||||
EOF
|
||||
--cat_file $corrupted_pages_file_filt
|
||||
--echo ------
|
||||
|
||||
--let targetdir2=$targetdir-2
|
||||
--let incdir2=$incdir-2
|
||||
perl;
|
||||
use lib "lib";
|
||||
use My::Handles { suppress_init_messages => 1 };
|
||||
use My::File::Path;
|
||||
copytree($ENV{'targetdir'}, $ENV{'targetdir2'});
|
||||
copytree($ENV{'incdir'}, $ENV{'incdir2'});
|
||||
EOF
|
||||
|
||||
--echo # Full backup prepare
|
||||
--disable_result_log
|
||||
exec $XTRABACKUP --prepare --target-dir=$targetdir > $backuplog;
|
||||
--enable_result_log
|
||||
|
||||
--echo # "innodb_corrupted_pages" file must not exist after successful prepare
|
||||
--error 1
|
||||
--file_exists $targetdir/innodb_corrupted_pages
|
||||
--let SEARCH_PATTERN=was successfuly fixed.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
|
||||
--echo # Check that fixed pages are zero-filled
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
open(my $fh, '<', $ENV{perl_result_file}) or die $!;
|
||||
my $last_page_no = <$fh>;
|
||||
close $fh;
|
||||
my $schema = "$ENV{targetdir}/test";
|
||||
die_if_page_is_not_zero("$schema/t3.ibd", $last_page_no, $last_page_no + 2);
|
||||
EOF
|
||||
|
||||
--echo # Incremental backup prepare
|
||||
--disable_result_log
|
||||
exec $XTRABACKUP --prepare --target-dir=$targetdir --incremental-dir=$incdir > $backuplog;
|
||||
--enable_result_log
|
||||
|
||||
--echo # "innodb_corrupted_pages" file must not exist after successful prepare
|
||||
--error 1
|
||||
--file_exists $targetdir/innodb_corrupted_pages
|
||||
--echo # do not remove "innodb_corrupted_pages" in incremental dir
|
||||
--file_exists $incdir/innodb_corrupted_pages
|
||||
--let SEARCH_PATTERN=was successfuly fixed.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
|
||||
--echo # Check that fixed pages are zero-filled
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
open(my $fh, '<', $ENV{perl_result_file}) or die $!;
|
||||
my $last_page_no_full = <$fh>;
|
||||
my $last_page_no_inc = <$fh>;
|
||||
close $fh;
|
||||
my $schema = "$ENV{targetdir}/test";
|
||||
die_if_page_is_not_zero("$schema/t3.ibd",
|
||||
$last_page_no_full, $last_page_no_full + 2);
|
||||
die_if_page_is_not_zero("$schema/t3_inc.ibd",
|
||||
$last_page_no_inc, $last_page_no_inc + 2);
|
||||
EOF
|
||||
|
||||
--source include/restart_and_restore.inc
|
||||
|
||||
SELECT * FROM t3;
|
||||
SELECT * FROM t3_inc;
|
||||
|
||||
--echo # Test the case when not all corrupted pages are fixed
|
||||
--echo
|
||||
--echo # Add some fake corrupted pages
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
append_corrupted_pages(
|
||||
"$ENV{targetdir2}/innodb_corrupted_pages", 'test/t3', '3 4');
|
||||
append_corrupted_pages(
|
||||
"$ENV{incdir2}/innodb_corrupted_pages", 'test/t3_inc', '4 5');
|
||||
EOF
|
||||
|
||||
--echo # Full backup prepare
|
||||
--disable_result_log
|
||||
--error 1
|
||||
exec $XTRABACKUP --prepare --target-dir=$targetdir2 > $backuplog;
|
||||
--enable_result_log
|
||||
|
||||
--let SEARCH_PATTERN=Error: corrupted page.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
--let corrupted_pages_file = $targetdir2/innodb_corrupted_pages
|
||||
--echo --- "innodb_corrupted_pages" file content: ---
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
print_corrupted_pages_file($ENV{corrupted_pages_file},
|
||||
$ENV{corrupted_pages_file_filt});
|
||||
EOF
|
||||
--cat_file $corrupted_pages_file_filt
|
||||
--echo ------
|
||||
|
||||
--echo # Incremental backup prepare
|
||||
--disable_result_log
|
||||
--error 1
|
||||
exec $XTRABACKUP --prepare --target-dir=$targetdir2 --incremental-dir=$incdir2 > $backuplog;
|
||||
--enable_result_log
|
||||
|
||||
--let SEARCH_PATTERN=Error: corrupted page.*
|
||||
--let SEARCH_FILE=$backuplog
|
||||
--source include/search_pattern_in_file.inc
|
||||
--let corrupted_pages_file = $targetdir2/innodb_corrupted_pages
|
||||
--echo --- "innodb_corrupted_pages" file content: ---
|
||||
perl;
|
||||
do "$ENV{MTR_SUITE_DIR}/include/corrupt-page.pl";
|
||||
print_corrupted_pages_file($ENV{corrupted_pages_file},
|
||||
$ENV{corrupted_pages_file_filt});
|
||||
EOF
|
||||
--cat_file $corrupted_pages_file_filt
|
||||
--echo ------
|
||||
|
||||
DROP TABLE t3;
|
||||
DROP TABLE t3_inc;
|
||||
--remove_file $backuplog
|
||||
--remove_file $perl_result_file
|
||||
--remove_file $corrupted_pages_file_filt
|
||||
--rmdir $targetdir
|
||||
--rmdir $targetdir2
|
||||
--rmdir $incdir
|
||||
--rmdir $incdir2
|
|
@ -147,3 +147,18 @@ drop role test_role2;
|
|||
delete from mysql.roles_mapping where Role='test_role1';
|
||||
delete from mysql.roles_mapping where Role='test_role2';
|
||||
flush privileges;
|
||||
#
|
||||
# MDEV-24289: show grants missing with grant option
|
||||
#
|
||||
create role anel;
|
||||
GRANT SELECT, UPDATE, DELETE, ALTER ON *.* TO 'anel';
|
||||
SHOW GRANTS for 'anel';
|
||||
Grants for anel
|
||||
GRANT SELECT, UPDATE, DELETE, ALTER ON *.* TO `anel`
|
||||
create role MariaDB_admin;
|
||||
GRANT SELECT, UPDATE, DELETE, ALTER ON *.* TO 'MariaDB_admin' WITH GRANT OPTION;
|
||||
SHOW GRANTS for 'MariaDB_admin';
|
||||
Grants for MariaDB_admin
|
||||
GRANT SELECT, UPDATE, DELETE, ALTER ON *.* TO `MariaDB_admin` WITH GRANT OPTION
|
||||
drop role MariaDB_admin;
|
||||
drop role anel;
|
||||
|
|
|
@ -88,3 +88,16 @@ drop role test_role2;
|
|||
delete from mysql.roles_mapping where Role='test_role1';
|
||||
delete from mysql.roles_mapping where Role='test_role2';
|
||||
flush privileges;
|
||||
|
||||
--echo #
|
||||
--echo # MDEV-24289: show grants missing with grant option
|
||||
--echo #
|
||||
create role anel;
|
||||
GRANT SELECT, UPDATE, DELETE, ALTER ON *.* TO 'anel';
|
||||
SHOW GRANTS for 'anel';
|
||||
|
||||
create role MariaDB_admin;
|
||||
GRANT SELECT, UPDATE, DELETE, ALTER ON *.* TO 'MariaDB_admin' WITH GRANT OPTION;
|
||||
SHOW GRANTS for 'MariaDB_admin';
|
||||
drop role MariaDB_admin;
|
||||
drop role anel;
|
||||
|
|
|
@ -401,6 +401,14 @@ then
|
|||
MODULE="rsync_sst"
|
||||
|
||||
RSYNC_PID="$WSREP_SST_OPT_DATA/$MODULE.pid"
|
||||
# give some time for lingering rsync from previous SST to complete
|
||||
check_round=0
|
||||
while check_pid $RSYNC_PID && [ $check_round -lt 10 ]
|
||||
do
|
||||
wsrep_log_info "lingering rsync daemon found at startup, waiting for it to exit"
|
||||
check_round=$(( check_round + 1 ))
|
||||
sleep 1
|
||||
done
|
||||
|
||||
if check_pid $RSYNC_PID
|
||||
then
|
||||
|
|
|
@ -1887,13 +1887,8 @@ bool Field::compatible_field_size(uint field_metadata,
|
|||
int Field::store(const char *to, size_t length, CHARSET_INFO *cs,
|
||||
enum_check_fields check_level)
|
||||
{
|
||||
int res;
|
||||
THD *thd= get_thd();
|
||||
enum_check_fields old_check_level= thd->count_cuted_fields;
|
||||
thd->count_cuted_fields= check_level;
|
||||
res= store(to, length, cs);
|
||||
thd->count_cuted_fields= old_check_level;
|
||||
return res;
|
||||
Check_level_instant_set check_level_save(get_thd(), check_level);
|
||||
return store(to, length, cs);
|
||||
}
|
||||
|
||||
|
||||
|
|
|
@ -4345,7 +4345,7 @@ int ha_partition::write_row(const uchar * buf)
|
|||
|
||||
tmp_disable_binlog(thd); /* Do not replicate the low-level changes. */
|
||||
error= m_file[part_id]->ha_write_row(buf);
|
||||
if (have_auto_increment && !table->s->next_number_keypart)
|
||||
if (!error && have_auto_increment && !table->s->next_number_keypart)
|
||||
set_auto_increment_if_higher(table->next_number_field);
|
||||
reenable_binlog(thd);
|
||||
|
||||
|
|
|
@ -92,7 +92,6 @@ public:
|
|||
bool auto_inc_initialized;
|
||||
mysql_mutex_t auto_inc_mutex; /**< protecting auto_inc val */
|
||||
ulonglong next_auto_inc_val; /**< first non reserved value */
|
||||
ulonglong prev_auto_inc_val; /**< stored next_auto_inc_val */
|
||||
/**
|
||||
Hash of partition names. Initialized in the first ha_partition::open()
|
||||
for the table_share. After that it is read-only, i.e. no locking required.
|
||||
|
@ -104,7 +103,6 @@ public:
|
|||
Partition_share()
|
||||
: auto_inc_initialized(false),
|
||||
next_auto_inc_val(0),
|
||||
prev_auto_inc_val(0),
|
||||
partition_name_hash_initialized(false),
|
||||
partition_names(NULL)
|
||||
{
|
||||
|
@ -430,24 +428,6 @@ private:
|
|||
MY_BITMAP m_locked_partitions;
|
||||
/** Stores shared auto_increment etc. */
|
||||
Partition_share *part_share;
|
||||
/** Fix spurious -Werror=overloaded-virtual in GCC 9 */
|
||||
virtual void restore_auto_increment(ulonglong prev_insert_id) override
|
||||
{
|
||||
handler::restore_auto_increment(prev_insert_id);
|
||||
}
|
||||
/** Store and restore next_auto_inc_val over duplicate key errors. */
|
||||
void store_auto_increment() override
|
||||
{
|
||||
DBUG_ASSERT(part_share);
|
||||
part_share->prev_auto_inc_val= part_share->next_auto_inc_val;
|
||||
handler::store_auto_increment();
|
||||
}
|
||||
void restore_auto_increment() override
|
||||
{
|
||||
DBUG_ASSERT(part_share);
|
||||
part_share->next_auto_inc_val= part_share->prev_auto_inc_val;
|
||||
handler::restore_auto_increment();
|
||||
}
|
||||
void sum_copy_info(handler *file);
|
||||
void sum_copy_infos();
|
||||
void reset_copy_info() override;
|
||||
|
|
|
@ -3310,7 +3310,6 @@ int handler::update_auto_increment()
|
|||
THD *thd= table->in_use;
|
||||
struct system_variables *variables= &thd->variables;
|
||||
int result=0, tmp;
|
||||
enum enum_check_fields save_count_cuted_fields;
|
||||
DBUG_ENTER("handler::update_auto_increment");
|
||||
|
||||
/*
|
||||
|
@ -3452,10 +3451,10 @@ int handler::update_auto_increment()
|
|||
nr, append ? nb_reserved_values : 0));
|
||||
|
||||
/* Store field without warning (Warning will be printed by insert) */
|
||||
save_count_cuted_fields= thd->count_cuted_fields;
|
||||
thd->count_cuted_fields= CHECK_FIELD_IGNORE;
|
||||
tmp= table->next_number_field->store((longlong)nr, TRUE);
|
||||
thd->count_cuted_fields= save_count_cuted_fields;
|
||||
{
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
tmp= table->next_number_field->store((longlong)nr, TRUE);
|
||||
}
|
||||
|
||||
if (unlikely(tmp)) // Out of range value in store
|
||||
{
|
||||
|
|
|
@ -3149,10 +3149,6 @@ private:
|
|||
*/
|
||||
Handler_share **ha_share;
|
||||
|
||||
/** Stores next_insert_id for handling duplicate key errors. */
|
||||
ulonglong m_prev_insert_id;
|
||||
|
||||
|
||||
public:
|
||||
handler(handlerton *ht_arg, TABLE_SHARE *share_arg)
|
||||
:table_share(share_arg), table(0),
|
||||
|
@ -3177,7 +3173,7 @@ public:
|
|||
auto_inc_intervals_count(0),
|
||||
m_psi(NULL), set_top_table_fields(FALSE), top_table(0),
|
||||
top_table_field(0), top_table_fields(0),
|
||||
m_lock_type(F_UNLCK), ha_share(NULL), m_prev_insert_id(0)
|
||||
m_lock_type(F_UNLCK), ha_share(NULL)
|
||||
{
|
||||
DBUG_PRINT("info",
|
||||
("handler created F_UNLCK %d F_RDLCK %d F_WRLCK %d",
|
||||
|
@ -3858,16 +3854,6 @@ public:
|
|||
insert_id_for_cur_row;
|
||||
}
|
||||
|
||||
/** Store and restore next_insert_id over duplicate key errors. */
|
||||
virtual void store_auto_increment()
|
||||
{
|
||||
m_prev_insert_id= next_insert_id;
|
||||
}
|
||||
virtual void restore_auto_increment()
|
||||
{
|
||||
restore_auto_increment(m_prev_insert_id);
|
||||
}
|
||||
|
||||
virtual void update_create_info(HA_CREATE_INFO *create_info) {}
|
||||
int check_old_types();
|
||||
virtual int assign_to_keycache(THD* thd, HA_CHECK_OPT* check_opt)
|
||||
|
|
11
sql/item.cc
11
sql/item.cc
|
@ -1419,18 +1419,13 @@ int Item::save_in_field_no_warnings(Field *field, bool no_conversions)
|
|||
int res;
|
||||
TABLE *table= field->table;
|
||||
THD *thd= table->in_use;
|
||||
enum_check_fields tmp= thd->count_cuted_fields;
|
||||
my_bitmap_map *old_map= dbug_tmp_use_all_columns(table, table->write_set);
|
||||
sql_mode_t sql_mode= thd->variables.sql_mode;
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
Sql_mode_save sql_mode(thd);
|
||||
thd->variables.sql_mode&= ~(MODE_NO_ZERO_IN_DATE | MODE_NO_ZERO_DATE);
|
||||
thd->variables.sql_mode|= MODE_INVALID_DATES;
|
||||
thd->count_cuted_fields= CHECK_FIELD_IGNORE;
|
||||
|
||||
my_bitmap_map *old_map= dbug_tmp_use_all_columns(table, table->write_set);
|
||||
res= save_in_field(field, no_conversions);
|
||||
|
||||
thd->count_cuted_fields= tmp;
|
||||
dbug_tmp_restore_column_map(table->write_set, old_map);
|
||||
thd->variables.sql_mode= sql_mode;
|
||||
return res;
|
||||
}
|
||||
|
||||
|
|
|
@ -951,6 +951,11 @@ public:
|
|||
void set_name_no_truncate(THD *thd, const char *str, uint length,
|
||||
CHARSET_INFO *cs);
|
||||
void init_make_send_field(Send_field *tmp_field, const Type_handler *h);
|
||||
void share_name_with(const Item *item)
|
||||
{
|
||||
name= item->name;
|
||||
is_autogenerated_name= item->is_autogenerated_name;
|
||||
}
|
||||
virtual void cleanup();
|
||||
virtual void make_send_field(THD *thd, Send_field *field);
|
||||
|
||||
|
|
|
@ -319,19 +319,18 @@ static bool convert_const_to_int(THD *thd, Item_field *field_item,
|
|||
if ((*item)->const_item() && !(*item)->is_expensive())
|
||||
{
|
||||
TABLE *table= field->table;
|
||||
sql_mode_t orig_sql_mode= thd->variables.sql_mode;
|
||||
enum_check_fields orig_count_cuted_fields= thd->count_cuted_fields;
|
||||
Sql_mode_save sql_mode(thd);
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
my_bitmap_map *old_maps[2] = { NULL, NULL };
|
||||
ulonglong UNINIT_VAR(orig_field_val); /* original field value if valid */
|
||||
|
||||
/* table->read_set may not be set if we come here from a CREATE TABLE */
|
||||
if (table && table->read_set)
|
||||
dbug_tmp_use_all_columns(table, old_maps,
|
||||
dbug_tmp_use_all_columns(table, old_maps,
|
||||
table->read_set, table->write_set);
|
||||
/* For comparison purposes allow invalid dates like 2000-01-32 */
|
||||
thd->variables.sql_mode= (orig_sql_mode & ~MODE_NO_ZERO_DATE) |
|
||||
thd->variables.sql_mode= (thd->variables.sql_mode & ~MODE_NO_ZERO_DATE) |
|
||||
MODE_INVALID_DATES;
|
||||
thd->count_cuted_fields= CHECK_FIELD_IGNORE;
|
||||
|
||||
/*
|
||||
Store the value of the field/constant because the call to save_in_field
|
||||
|
@ -368,8 +367,6 @@ static bool convert_const_to_int(THD *thd, Item_field *field_item,
|
|||
/* orig_field_val must be a valid value that can be restored back. */
|
||||
DBUG_ASSERT(!result);
|
||||
}
|
||||
thd->variables.sql_mode= orig_sql_mode;
|
||||
thd->count_cuted_fields= orig_count_cuted_fields;
|
||||
if (table && table->read_set)
|
||||
dbug_tmp_restore_column_maps(table->read_set, table->write_set, old_maps);
|
||||
}
|
||||
|
|
|
@ -1484,7 +1484,8 @@ Item_in_subselect::Item_in_subselect(THD *thd, Item * left_exp,
|
|||
pushed_cond_guards(NULL), do_not_convert_to_sj(FALSE), is_jtbm_merged(FALSE),
|
||||
is_jtbm_const_tab(FALSE), is_flattenable_semijoin(FALSE),
|
||||
is_registered_semijoin(FALSE),
|
||||
upper_item(0)
|
||||
upper_item(0),
|
||||
converted_from_in_predicate(FALSE)
|
||||
{
|
||||
DBUG_ENTER("Item_in_subselect::Item_in_subselect");
|
||||
DBUG_PRINT("info", ("in_strategy: %u", (uint)in_strategy));
|
||||
|
|
|
@ -610,12 +610,18 @@ public:
|
|||
|
||||
Item_func_not_all *upper_item; // point on NOT/NOP before ALL/SOME subquery
|
||||
|
||||
/*
|
||||
SET to TRUE if IN subquery is converted from an IN predicate
|
||||
*/
|
||||
bool converted_from_in_predicate;
|
||||
|
||||
Item_in_subselect(THD *thd_arg, Item * left_expr, st_select_lex *select_lex);
|
||||
Item_in_subselect(THD *thd_arg):
|
||||
Item_exists_subselect(thd_arg), left_expr_cache(0), first_execution(TRUE),
|
||||
in_strategy(SUBS_NOT_TRANSFORMED),
|
||||
pushed_cond_guards(NULL), func(NULL), do_not_convert_to_sj(FALSE),
|
||||
is_jtbm_merged(FALSE), is_jtbm_const_tab(FALSE), upper_item(0) {}
|
||||
is_jtbm_merged(FALSE), is_jtbm_const_tab(FALSE), upper_item(0),
|
||||
converted_from_in_predicate(FALSE) {}
|
||||
void cleanup();
|
||||
subs_type substype() { return IN_SUBS; }
|
||||
void reset()
|
||||
|
|
|
@ -7570,10 +7570,10 @@ error:
|
|||
if (thd->transaction_rollback_request)
|
||||
{
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
else if (! thd->in_multi_stmt_transaction_mode())
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
else
|
||||
thd->mdl_context.release_statement_locks();
|
||||
|
||||
|
@ -9047,7 +9047,8 @@ int Xid_log_event::do_apply_event(rpl_group_info *rgi)
|
|||
"COMMIT /* implicit, from Xid_log_event */");
|
||||
thd->variables.option_bits&= ~OPTION_GTID_BEGIN;
|
||||
res= trans_commit(thd); /* Automatically rolls back on error. */
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
#ifdef WITH_WSREP
|
||||
if (WSREP(thd)) mysql_mutex_lock(&thd->LOCK_thd_data);
|
||||
if ((!res || (WSREP(thd) && thd->wsrep_trx().state() == wsrep::transaction::s_must_replay )) && sub_id)
|
||||
|
|
|
@ -3036,6 +3036,9 @@ void MDL_context::rollback_to_savepoint(const MDL_savepoint &mdl_savepoint)
|
|||
void MDL_context::release_transactional_locks()
|
||||
{
|
||||
DBUG_ENTER("MDL_context::release_transactional_locks");
|
||||
/* Fail if there are active transactions */
|
||||
DBUG_ASSERT(!(current_thd->server_status &
|
||||
(SERVER_STATUS_IN_TRANS | SERVER_STATUS_IN_TRANS_READONLY)));
|
||||
release_locks_stored_before(MDL_STATEMENT, NULL);
|
||||
release_locks_stored_before(MDL_TRANSACTION, NULL);
|
||||
DBUG_VOID_RETURN;
|
||||
|
|
|
@ -868,6 +868,7 @@ bool subquery_types_allow_materialization(THD* thd, Item_in_subselect *in_subs)
|
|||
|
||||
bool all_are_fields= TRUE;
|
||||
uint32 total_key_length = 0;
|
||||
bool converted_from_in_predicate= in_subs->converted_from_in_predicate;
|
||||
for (uint i= 0; i < elements; i++)
|
||||
{
|
||||
Item *outer= in_subs->left_expr->element_index(i);
|
||||
|
@ -875,8 +876,11 @@ bool subquery_types_allow_materialization(THD* thd, Item_in_subselect *in_subs)
|
|||
all_are_fields &= (outer->real_item()->type() == Item::FIELD_ITEM &&
|
||||
inner->real_item()->type() == Item::FIELD_ITEM);
|
||||
total_key_length += inner->max_length;
|
||||
if (!inner->type_handler()->subquery_type_allows_materialization(inner,
|
||||
outer))
|
||||
if (!inner->
|
||||
type_handler()->
|
||||
subquery_type_allows_materialization(inner,
|
||||
outer,
|
||||
converted_from_in_predicate))
|
||||
{
|
||||
trace_transform.add("possible", false);
|
||||
trace_transform.add("cause", "types mismatch");
|
||||
|
|
|
@ -441,7 +441,7 @@ rpl_slave_state::truncate_state_table(THD *thd)
|
|||
close_thread_tables(thd);
|
||||
ha_commit_trans(thd, TRUE);
|
||||
}
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
|
||||
reenable_binlog(thd);
|
||||
|
@ -988,7 +988,7 @@ end:
|
|||
ha_rollback_trans(thd, FALSE);
|
||||
}
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
thd->lex->restore_backup_query_tables_list(&lex_backup);
|
||||
|
||||
if (err)
|
||||
|
|
|
@ -68,34 +68,34 @@ injector::transaction::~transaction()
|
|||
*/
|
||||
int injector::transaction::commit()
|
||||
{
|
||||
DBUG_ENTER("injector::transaction::commit()");
|
||||
int error= m_thd->binlog_flush_pending_rows_event(true);
|
||||
/*
|
||||
Cluster replication does not preserve statement or
|
||||
transaction boundaries of the master. Instead, a new
|
||||
transaction on replication slave is started when a new GCI
|
||||
(global checkpoint identifier) is issued, and is committed
|
||||
when the last event of the check point has been received and
|
||||
processed. This ensures consistency of each cluster in
|
||||
cluster replication, and there is no requirement for stronger
|
||||
consistency: MySQL replication is asynchronous with other
|
||||
engines as well.
|
||||
DBUG_ENTER("injector::transaction::commit()");
|
||||
int error= m_thd->binlog_flush_pending_rows_event(true);
|
||||
/*
|
||||
Cluster replication does not preserve statement or
|
||||
transaction boundaries of the master. Instead, a new
|
||||
transaction on replication slave is started when a new GCI
|
||||
(global checkpoint identifier) is issued, and is committed
|
||||
when the last event of the check point has been received and
|
||||
processed. This ensures consistency of each cluster in
|
||||
cluster replication, and there is no requirement for stronger
|
||||
consistency: MySQL replication is asynchronous with other
|
||||
engines as well.
|
||||
|
||||
A practical consequence of that is that row level replication
|
||||
stream passed through the injector thread never contains
|
||||
COMMIT events.
|
||||
Here we should preserve the server invariant that there is no
|
||||
outstanding statement transaction when the normal transaction
|
||||
is committed by committing the statement transaction
|
||||
explicitly.
|
||||
*/
|
||||
trans_commit_stmt(m_thd);
|
||||
if (!trans_commit(m_thd))
|
||||
{
|
||||
close_thread_tables(m_thd);
|
||||
m_thd->mdl_context.release_transactional_locks();
|
||||
}
|
||||
DBUG_RETURN(error);
|
||||
A practical consequence of that is that row level replication
|
||||
stream passed through the injector thread never contains
|
||||
COMMIT events.
|
||||
Here we should preserve the server invariant that there is no
|
||||
outstanding statement transaction when the normal transaction
|
||||
is committed by committing the statement transaction
|
||||
explicitly.
|
||||
*/
|
||||
trans_commit_stmt(m_thd);
|
||||
if (!trans_commit(m_thd))
|
||||
{
|
||||
close_thread_tables(m_thd);
|
||||
m_thd->release_transactional_locks();
|
||||
}
|
||||
DBUG_RETURN(error);
|
||||
}
|
||||
|
||||
|
||||
|
|
|
@ -1702,7 +1702,7 @@ scan_all_gtid_slave_pos_table(THD *thd, int (*cb)(THD *, LEX_CSTRING *, void *),
|
|||
{
|
||||
my_error(ER_FILE_NOT_FOUND, MYF(0), path, my_errno);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
return 1;
|
||||
}
|
||||
else
|
||||
|
@ -1715,7 +1715,7 @@ scan_all_gtid_slave_pos_table(THD *thd, int (*cb)(THD *, LEX_CSTRING *, void *),
|
|||
err= ha_discover_table_names(thd, &MYSQL_SCHEMA_NAME, dirp, &tl, false);
|
||||
my_dirend(dirp);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
|
@ -2000,7 +2000,7 @@ end:
|
|||
ha_commit_trans(thd, FALSE);
|
||||
ha_commit_trans(thd, TRUE);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
|
||||
return err;
|
||||
|
@ -2282,7 +2282,7 @@ void rpl_group_info::cleanup_context(THD *thd, bool error)
|
|||
|
||||
if (unlikely(error))
|
||||
{
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
if (thd == rli->sql_driver_thd)
|
||||
{
|
||||
|
@ -2396,10 +2396,10 @@ void rpl_group_info::slave_close_thread_tables(THD *thd)
|
|||
if (thd->transaction_rollback_request)
|
||||
{
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
else if (! thd->in_multi_stmt_transaction_mode())
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
else
|
||||
thd->mdl_context.release_statement_locks();
|
||||
|
||||
|
|
|
@ -1181,8 +1181,6 @@ Sp_handler::sp_create_routine(THD *thd, const sp_head *sp) const
|
|||
|
||||
CHARSET_INFO *db_cs= get_default_db_collation(thd, sp->m_db.str);
|
||||
|
||||
enum_check_fields saved_count_cuted_fields;
|
||||
|
||||
bool store_failed= FALSE;
|
||||
DBUG_ENTER("sp_create_routine");
|
||||
DBUG_PRINT("enter", ("type: %s name: %.*s",
|
||||
|
@ -1216,8 +1214,7 @@ Sp_handler::sp_create_routine(THD *thd, const sp_head *sp) const
|
|||
/* Reset sql_mode during data dictionary operations. */
|
||||
thd->variables.sql_mode= 0;
|
||||
|
||||
saved_count_cuted_fields= thd->count_cuted_fields;
|
||||
thd->count_cuted_fields= CHECK_FIELD_WARN;
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_WARN);
|
||||
|
||||
if (!(table= open_proc_table_for_update(thd)))
|
||||
{
|
||||
|
@ -1477,7 +1474,6 @@ log:
|
|||
ret= FALSE;
|
||||
|
||||
done:
|
||||
thd->count_cuted_fields= saved_count_cuted_fields;
|
||||
thd->variables.sql_mode= saved_mode;
|
||||
DBUG_ASSERT(!thd->is_current_stmt_binlog_format_row());
|
||||
DBUG_RETURN(ret);
|
||||
|
|
|
@ -2311,10 +2311,10 @@ sp_head::execute_procedure(THD *thd, List<Item> *args)
|
|||
if (thd->transaction_rollback_request)
|
||||
{
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
else if (! thd->in_multi_stmt_transaction_mode())
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
else
|
||||
thd->mdl_context.release_statement_locks();
|
||||
}
|
||||
|
@ -3442,10 +3442,10 @@ sp_lex_keeper::reset_lex_and_exec_core(THD *thd, uint *nextp,
|
|||
if (thd->transaction_rollback_request)
|
||||
{
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
else if (! thd->in_multi_stmt_transaction_mode())
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
else
|
||||
thd->mdl_context.release_statement_locks();
|
||||
}
|
||||
|
|
|
@ -9312,6 +9312,8 @@ static bool show_global_privileges(THD *thd, ACL_USER_BASE *acl_entry,
|
|||
add_user_parameters(thd, &global, (ACL_USER *)acl_entry,
|
||||
(want_access & GRANT_ACL));
|
||||
|
||||
else if (want_access & GRANT_ACL)
|
||||
global.append(STRING_WITH_LEN(" WITH GRANT OPTION"));
|
||||
protocol->prepare_for_resend();
|
||||
protocol->store(global.ptr(),global.length(),global.charset());
|
||||
if (protocol->write())
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
/* Copyright (c) 2010, 2015, Oracle and/or its affiliates.
|
||||
Copyright (c) 2011, 2019, MariaDB
|
||||
Copyright (c) 2011, 2020, MariaDB
|
||||
|
||||
This program is free software; you can redistribute it and/or modify
|
||||
it under the terms of the GNU General Public License as published by
|
||||
|
@ -43,7 +43,7 @@ static bool admin_recreate_table(THD *thd, TABLE_LIST *table_list)
|
|||
trans_rollback_stmt(thd);
|
||||
trans_rollback(thd);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
/*
|
||||
table_list->table has been closed and freed. Do not reference
|
||||
|
@ -116,7 +116,7 @@ static int prepare_for_repair(THD *thd, TABLE_LIST *table_list,
|
|||
acquire the exclusive lock to satisfy MDL asserts and avoid
|
||||
deadlocks.
|
||||
*/
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
/*
|
||||
Attempt to do full-blown table open in mysql_admin_table() has failed.
|
||||
Let us try to open at least a .FRM for this table.
|
||||
|
@ -285,7 +285,7 @@ end:
|
|||
}
|
||||
/* In case of a temporary table there will be no metadata lock. */
|
||||
if (unlikely(error) && has_mdl_lock)
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
DBUG_RETURN(error);
|
||||
}
|
||||
|
@ -614,7 +614,7 @@ static bool mysql_admin_table(THD* thd, TABLE_LIST* tables,
|
|||
trans_rollback(thd);
|
||||
close_thread_tables(thd);
|
||||
table->table= NULL;
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
table->mdl_request.init(MDL_key::TABLE, table->db.str, table->table_name.str,
|
||||
MDL_SHARED_NO_READ_WRITE, MDL_TRANSACTION);
|
||||
}
|
||||
|
@ -674,7 +674,7 @@ static bool mysql_admin_table(THD* thd, TABLE_LIST* tables,
|
|||
trans_rollback_stmt(thd);
|
||||
trans_rollback(thd);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
DBUG_PRINT("admin", ("simple error, admin next table"));
|
||||
continue;
|
||||
case -1: // error, message could be written to net
|
||||
|
@ -741,7 +741,7 @@ static bool mysql_admin_table(THD* thd, TABLE_LIST* tables,
|
|||
trans_commit_stmt(thd);
|
||||
trans_commit(thd);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
lex->reset_query_tables_list(FALSE);
|
||||
/*
|
||||
Restore Query_tables_list::sql_command value to make statement
|
||||
|
@ -874,7 +874,7 @@ static bool mysql_admin_table(THD* thd, TABLE_LIST* tables,
|
|||
thd->open_options|= extra_open_options;
|
||||
close_thread_tables(thd);
|
||||
table->table= NULL;
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
table->mdl_request.init(MDL_key::TABLE, table->db.str, table->table_name.str,
|
||||
MDL_SHARED_NO_READ_WRITE, MDL_TRANSACTION);
|
||||
table->mdl_request.set_type(MDL_SHARED_READ);
|
||||
|
@ -1106,7 +1106,7 @@ send_result_message:
|
|||
trans_commit_stmt(thd);
|
||||
trans_commit(thd);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
/* Clear references to TABLE and MDL_ticket after releasing them. */
|
||||
table->mdl_request.ticket= NULL;
|
||||
|
||||
|
@ -1259,7 +1259,7 @@ send_result_message:
|
|||
goto err;
|
||||
}
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
/*
|
||||
If it is CHECK TABLE v1, v2, v3, and v1, v2, v3 are views, we will run
|
||||
|
@ -1297,7 +1297,7 @@ err:
|
|||
table->table= 0;
|
||||
}
|
||||
close_thread_tables(thd); // Shouldn't be needed
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
thd->resume_subsequent_commits(suspended_wfc);
|
||||
DBUG_RETURN(TRUE);
|
||||
}
|
||||
|
|
|
@ -9149,7 +9149,7 @@ close_mysql_tables(THD *thd)
|
|||
if (! thd->in_sub_stmt)
|
||||
trans_commit_stmt(thd);
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -4824,7 +4824,7 @@ void destroy_thd(MYSQL_THD thd)
|
|||
void reset_thd(MYSQL_THD thd)
|
||||
{
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
thd->free_items();
|
||||
free_root(thd->mem_root, MYF(MY_KEEP_PREALLOC));
|
||||
}
|
||||
|
|
|
@ -4574,6 +4574,13 @@ public:
|
|||
locked_tables_mode= mode_arg;
|
||||
}
|
||||
void leave_locked_tables_mode();
|
||||
/* Relesae transactional locks if there are no active transactions */
|
||||
void release_transactional_locks()
|
||||
{
|
||||
if (!(server_status &
|
||||
(SERVER_STATUS_IN_TRANS | SERVER_STATUS_IN_TRANS_READONLY)))
|
||||
mdl_context.release_transactional_locks();
|
||||
}
|
||||
int decide_logging_format(TABLE_LIST *tables);
|
||||
/*
|
||||
In Some cases when decide_logging_format is called it does not have all
|
||||
|
@ -6745,6 +6752,23 @@ class Switch_to_definer_security_ctx
|
|||
};
|
||||
|
||||
|
||||
class Check_level_instant_set
|
||||
{
|
||||
THD *m_thd;
|
||||
enum_check_fields m_check_level;
|
||||
public:
|
||||
Check_level_instant_set(THD *thd, enum_check_fields temporary_value)
|
||||
:m_thd(thd), m_check_level(thd->count_cuted_fields)
|
||||
{
|
||||
thd->count_cuted_fields= temporary_value;
|
||||
}
|
||||
~Check_level_instant_set()
|
||||
{
|
||||
m_thd->count_cuted_fields= m_check_level;
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
/**
|
||||
This class resembles the SQL Standard schema qualified object name:
|
||||
<schema qualified name> ::= [ <schema name> <period> ] <qualified identifier>
|
||||
|
|
|
@ -260,7 +260,7 @@ int update_portion_of_time(THD *thd, TABLE *table,
|
|||
uint dst_fieldno= lcond ? table->s->period.end_fieldno
|
||||
: table->s->period.start_fieldno;
|
||||
|
||||
table->file->store_auto_increment();
|
||||
ulonglong prev_insert_id= table->file->next_insert_id;
|
||||
store_record(table, record[1]);
|
||||
if (likely(!res))
|
||||
res= src->save_in_field(table->field[dst_fieldno], true);
|
||||
|
@ -276,7 +276,7 @@ int update_portion_of_time(THD *thd, TABLE *table,
|
|||
TRG_ACTION_AFTER, true);
|
||||
restore_record(table, record[1]);
|
||||
if (res)
|
||||
table->file->restore_auto_increment();
|
||||
table->file->restore_auto_increment(prev_insert_id);
|
||||
|
||||
if (likely(!res) && lcond && rcond)
|
||||
res= table->period_make_insert(period_conds.end.item,
|
||||
|
|
|
@ -1402,7 +1402,8 @@ bool pushdown_cond_for_derived(THD *thd, Item *cond, TABLE_LIST *derived)
|
|||
DBUG_RETURN(false);
|
||||
|
||||
st_select_lex_unit *unit= derived->get_unit();
|
||||
st_select_lex *sl= unit->first_select();
|
||||
st_select_lex *first_sl= unit->first_select();
|
||||
st_select_lex *sl= first_sl;
|
||||
|
||||
if (derived->prohibit_cond_pushdown)
|
||||
DBUG_RETURN(false);
|
||||
|
@ -1461,6 +1462,20 @@ bool pushdown_cond_for_derived(THD *thd, Item *cond, TABLE_LIST *derived)
|
|||
if (!extracted_cond_copy)
|
||||
continue;
|
||||
|
||||
/*
|
||||
Rename the columns of all non-first selects of a union to be compatible
|
||||
by names with the columns of the first select. It will allow to use copies
|
||||
of the same expression pushed into having clauses of different selects.
|
||||
*/
|
||||
if (sl != first_sl)
|
||||
{
|
||||
DBUG_ASSERT(sl->item_list.elements == first_sl->item_list.elements);
|
||||
List_iterator_fast<Item> it(sl->item_list);
|
||||
List_iterator_fast<Item> nm_it(unit->types);
|
||||
while (Item *item= it++)
|
||||
item->share_name_with(nm_it++);
|
||||
}
|
||||
|
||||
/* Collect fields that are used in the GROUP BY of sl */
|
||||
if (sl->have_window_funcs())
|
||||
{
|
||||
|
@ -1481,7 +1496,7 @@ bool pushdown_cond_for_derived(THD *thd, Item *cond, TABLE_LIST *derived)
|
|||
&remaining_cond,
|
||||
&Item::derived_field_transformer_for_where,
|
||||
(uchar *) sl);
|
||||
|
||||
|
||||
if (!remaining_cond)
|
||||
continue;
|
||||
/*
|
||||
|
|
|
@ -1721,7 +1721,7 @@ int write_record(THD *thd, TABLE *table,COPY_INFO *info)
|
|||
int error, trg_error= 0;
|
||||
char *key=0;
|
||||
MY_BITMAP *save_read_set, *save_write_set;
|
||||
table->file->store_auto_increment();
|
||||
ulonglong prev_insert_id= table->file->next_insert_id;
|
||||
ulonglong insert_id_for_cur_row= 0;
|
||||
ulonglong prev_insert_id_for_cur_row= 0;
|
||||
DBUG_ENTER("write_record");
|
||||
|
@ -1870,7 +1870,7 @@ int write_record(THD *thd, TABLE *table,COPY_INFO *info)
|
|||
if (res == VIEW_CHECK_ERROR)
|
||||
goto before_trg_err;
|
||||
|
||||
table->file->restore_auto_increment();
|
||||
table->file->restore_auto_increment(prev_insert_id);
|
||||
info->touched++;
|
||||
if (different_records)
|
||||
{
|
||||
|
@ -2064,7 +2064,7 @@ int write_record(THD *thd, TABLE *table,COPY_INFO *info)
|
|||
if (!(thd->variables.old_behavior &
|
||||
OLD_MODE_NO_DUP_KEY_WARNINGS_WITH_IGNORE))
|
||||
table->file->print_error(error, MYF(ME_WARNING));
|
||||
table->file->restore_auto_increment();
|
||||
table->file->restore_auto_increment(prev_insert_id);
|
||||
goto ok_or_after_trg_err;
|
||||
}
|
||||
|
||||
|
@ -2087,7 +2087,7 @@ err:
|
|||
table->file->print_error(error,MYF(0));
|
||||
|
||||
before_trg_err:
|
||||
table->file->restore_auto_increment();
|
||||
table->file->restore_auto_increment(prev_insert_id);
|
||||
if (key)
|
||||
my_safe_afree(key, table->s->max_unique_length);
|
||||
table->column_bitmaps_set(save_read_set, save_write_set);
|
||||
|
@ -3061,7 +3061,7 @@ pthread_handler_t handle_delayed_insert(void *arg)
|
|||
if (thd->mdl_context.clone_ticket(&di->grl_protection) ||
|
||||
thd->mdl_context.clone_ticket(&di->table_list.mdl_request))
|
||||
{
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
di->handler_thread_initialized= TRUE;
|
||||
goto err;
|
||||
}
|
||||
|
@ -3276,7 +3276,7 @@ pthread_handler_t handle_delayed_insert(void *arg)
|
|||
thd->set_killed(KILL_CONNECTION_HARD); // If error
|
||||
|
||||
close_thread_tables(thd); // Free the table
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
mysql_cond_broadcast(&di->cond_client); // Safety
|
||||
|
||||
mysql_mutex_lock(&LOCK_delayed_create); // Because of delayed_get_table
|
||||
|
@ -4766,6 +4766,7 @@ bool select_create::send_eof()
|
|||
WSREP_ERROR("Appending table key for CTAS failed: %s, %d",
|
||||
(wsrep_thd_query(thd)) ?
|
||||
wsrep_thd_query(thd) : "void", rcode);
|
||||
abort_result_set();
|
||||
DBUG_RETURN(true);
|
||||
}
|
||||
/* If commit fails, we should be able to reset the OK status. */
|
||||
|
|
|
@ -2057,7 +2057,7 @@ bool dispatch_command(enum enum_server_command command, THD *thd,
|
|||
locks.
|
||||
*/
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
|
||||
thd->cleanup_after_query();
|
||||
|
@ -2122,7 +2122,7 @@ bool dispatch_command(enum enum_server_command command, THD *thd,
|
|||
ulonglong options= (ulonglong) (uchar) packet[0];
|
||||
if (trans_commit_implicit(thd))
|
||||
break;
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
if (check_global_access(thd,RELOAD_ACL))
|
||||
break;
|
||||
general_log_print(thd, command, NullS);
|
||||
|
@ -2157,7 +2157,7 @@ bool dispatch_command(enum enum_server_command command, THD *thd,
|
|||
if (trans_commit_implicit(thd))
|
||||
break;
|
||||
close_thread_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
my_ok(thd);
|
||||
break;
|
||||
}
|
||||
|
@ -2993,7 +2993,7 @@ err:
|
|||
/* Close tables and release metadata locks. */
|
||||
close_thread_tables(thd);
|
||||
DBUG_ASSERT(!thd->locked_tables_mode);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
return TRUE;
|
||||
}
|
||||
|
||||
|
@ -3733,7 +3733,7 @@ mysql_execute_command(THD *thd)
|
|||
/* Commit the normal transaction if one is active. */
|
||||
bool commit_failed= trans_commit_implicit(thd);
|
||||
/* Release metadata locks acquired in this transaction. */
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
if (commit_failed)
|
||||
{
|
||||
WSREP_DEBUG("implicit commit failed, MDL released: %lld",
|
||||
|
@ -5011,7 +5011,7 @@ mysql_execute_command(THD *thd)
|
|||
{
|
||||
res= trans_commit_implicit(thd);
|
||||
thd->locked_tables_list.unlock_locked_tables(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
thd->variables.option_bits&= ~(OPTION_TABLE_LOCK);
|
||||
}
|
||||
if (thd->global_read_lock.is_acquired() &&
|
||||
|
@ -5026,7 +5026,7 @@ mysql_execute_command(THD *thd)
|
|||
res= trans_commit_implicit(thd);
|
||||
thd->locked_tables_list.unlock_locked_tables(thd);
|
||||
/* Release transactional metadata locks. */
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
if (res)
|
||||
goto error;
|
||||
|
||||
|
@ -5644,7 +5644,7 @@ mysql_execute_command(THD *thd)
|
|||
DBUG_PRINT("info", ("Executing SQLCOM_BEGIN thd: %p", thd));
|
||||
if (trans_begin(thd, lex->start_transaction_opt))
|
||||
{
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
WSREP_DEBUG("BEGIN failed, MDL released: %lld",
|
||||
(longlong) thd->thread_id);
|
||||
WSREP_DEBUG("stmt_da, sql_errno: %d", (thd->get_stmt_da()->is_error()) ? thd->get_stmt_da()->sql_errno() : 0);
|
||||
|
@ -5663,7 +5663,7 @@ mysql_execute_command(THD *thd)
|
|||
(thd->variables.completion_type == 2 &&
|
||||
lex->tx_release != TVL_NO));
|
||||
bool commit_failed= trans_commit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
if (commit_failed)
|
||||
{
|
||||
WSREP_DEBUG("COMMIT failed, MDL released: %lld",
|
||||
|
@ -5701,7 +5701,7 @@ mysql_execute_command(THD *thd)
|
|||
(thd->variables.completion_type == 2 &&
|
||||
lex->tx_release != TVL_NO));
|
||||
bool rollback_failed= trans_rollback(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
if (rollback_failed)
|
||||
{
|
||||
|
@ -6004,7 +6004,6 @@ mysql_execute_command(THD *thd)
|
|||
case SQLCOM_XA_COMMIT:
|
||||
{
|
||||
bool commit_failed= trans_xa_commit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
if (commit_failed)
|
||||
{
|
||||
WSREP_DEBUG("XA commit failed, MDL released: %lld",
|
||||
|
@ -6022,7 +6021,6 @@ mysql_execute_command(THD *thd)
|
|||
case SQLCOM_XA_ROLLBACK:
|
||||
{
|
||||
bool rollback_failed= trans_xa_rollback(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
if (rollback_failed)
|
||||
{
|
||||
WSREP_DEBUG("XA rollback failed, MDL released: %lld",
|
||||
|
@ -6227,7 +6225,7 @@ finish:
|
|||
*/
|
||||
THD_STAGE_INFO(thd, stage_rollback_implicit);
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
else if (stmt_causes_implicit_commit(thd, CF_IMPLICIT_COMMIT_END))
|
||||
{
|
||||
|
@ -6241,7 +6239,7 @@ finish:
|
|||
/* Commit the normal transaction if one is active. */
|
||||
trans_commit_implicit(thd);
|
||||
thd->get_stmt_da()->set_overwrite_status(false);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
}
|
||||
else if (! thd->in_sub_stmt && ! thd->in_multi_stmt_transaction_mode())
|
||||
|
@ -6256,7 +6254,7 @@ finish:
|
|||
- If in autocommit mode, or outside a transactional context,
|
||||
automatically release metadata locks of the current statement.
|
||||
*/
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
else if (! thd->in_sub_stmt)
|
||||
{
|
||||
|
@ -6281,7 +6279,7 @@ finish:
|
|||
{
|
||||
WSREP_DEBUG("Forcing release of transactional locks for thd: %lld",
|
||||
(longlong) thd->thread_id);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -4270,7 +4270,7 @@ bool Prepared_statement::prepare(const char *packet, uint packet_len)
|
|||
if (thd->transaction_rollback_request)
|
||||
{
|
||||
trans_rollback_implicit(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
}
|
||||
|
||||
/* Preserve CHANGE MASTER attributes */
|
||||
|
|
|
@ -24035,8 +24035,7 @@ cmp_buffer_with_ref(THD *thd, TABLE *table, TABLE_REF *tab_ref)
|
|||
bool
|
||||
cp_buffer_from_ref(THD *thd, TABLE *table, TABLE_REF *ref)
|
||||
{
|
||||
enum enum_check_fields save_count_cuted_fields= thd->count_cuted_fields;
|
||||
thd->count_cuted_fields= CHECK_FIELD_IGNORE;
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
my_bitmap_map *old_map= dbug_tmp_use_all_columns(table, table->write_set);
|
||||
bool result= 0;
|
||||
|
||||
|
@ -24048,7 +24047,6 @@ cp_buffer_from_ref(THD *thd, TABLE *table, TABLE_REF *ref)
|
|||
break;
|
||||
}
|
||||
}
|
||||
thd->count_cuted_fields= save_count_cuted_fields;
|
||||
dbug_tmp_restore_column_map(table->write_set, old_map);
|
||||
return result;
|
||||
}
|
||||
|
|
|
@ -1900,18 +1900,11 @@ public:
|
|||
{
|
||||
enum store_key_result result;
|
||||
THD *thd= to_field->table->in_use;
|
||||
enum_check_fields saved_count_cuted_fields= thd->count_cuted_fields;
|
||||
sql_mode_t orig_sql_mode= thd->variables.sql_mode;
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
Sql_mode_save sql_mode(thd);
|
||||
thd->variables.sql_mode&= ~(MODE_NO_ZERO_IN_DATE | MODE_NO_ZERO_DATE);
|
||||
thd->variables.sql_mode|= MODE_INVALID_DATES;
|
||||
|
||||
thd->count_cuted_fields= CHECK_FIELD_IGNORE;
|
||||
|
||||
result= copy_inner();
|
||||
|
||||
thd->count_cuted_fields= saved_count_cuted_fields;
|
||||
thd->variables.sql_mode= orig_sql_mode;
|
||||
|
||||
return result;
|
||||
}
|
||||
|
||||
|
|
|
@ -3732,7 +3732,6 @@ static bool show_status_array(THD *thd, const char *wild,
|
|||
char name_buffer[NAME_CHAR_LEN];
|
||||
int len;
|
||||
SHOW_VAR tmp, *var;
|
||||
enum_check_fields save_count_cuted_fields= thd->count_cuted_fields;
|
||||
bool res= FALSE;
|
||||
CHARSET_INFO *charset= system_charset_info;
|
||||
DBUG_ENTER("show_status_array");
|
||||
|
@ -3855,7 +3854,6 @@ static bool show_status_array(THD *thd, const char *wild,
|
|||
}
|
||||
}
|
||||
end:
|
||||
thd->count_cuted_fields= save_count_cuted_fields;
|
||||
DBUG_RETURN(res);
|
||||
}
|
||||
|
||||
|
@ -4583,8 +4581,7 @@ fill_schema_table_by_open(THD *thd, MEM_ROOT *mem_root,
|
|||
Open_tables_backup *open_tables_state_backup,
|
||||
bool can_deadlock)
|
||||
{
|
||||
Query_arena i_s_arena(mem_root,
|
||||
Query_arena::STMT_CONVENTIONAL_EXECUTION),
|
||||
Query_arena i_s_arena(mem_root, Query_arena::STMT_CONVENTIONAL_EXECUTION),
|
||||
backup_arena, *old_arena;
|
||||
LEX *old_lex= thd->lex, temp_lex, *lex;
|
||||
LEX_CSTRING db_name, table_name;
|
||||
|
@ -5098,12 +5095,9 @@ end:
|
|||
class Warnings_only_error_handler : public Internal_error_handler
|
||||
{
|
||||
public:
|
||||
bool handle_condition(THD *thd,
|
||||
uint sql_errno,
|
||||
const char* sqlstate,
|
||||
bool handle_condition(THD *thd, uint sql_errno, const char* sqlstate,
|
||||
Sql_condition::enum_warning_level *level,
|
||||
const char* msg,
|
||||
Sql_condition ** cond_hdl)
|
||||
const char* msg, Sql_condition ** cond_hdl)
|
||||
{
|
||||
if (sql_errno == ER_TRG_NO_DEFINER || sql_errno == ER_TRG_NO_CREATION_CTX)
|
||||
return true;
|
||||
|
@ -8721,13 +8715,6 @@ static int optimize_schema_tables_memory_usage(TABLE_LIST *table_list)
|
|||
DBUG_ASSERT(table->s->keys == 0);
|
||||
DBUG_ASSERT(table->s->uniques == 0);
|
||||
|
||||
// XXX HACK HACK HACK: in a stored function, RETURN (SELECT ...)
|
||||
// enables warnings (in THD::sp_eval_expr) for the whole val_xxx/store pair,
|
||||
// while the intention is to warn only for store(). Until this is
|
||||
// fixed let's avoid data truncation warnings in I_S->fill_table()
|
||||
if (thd->count_cuted_fields == CHECK_FIELD_IGNORE)
|
||||
{
|
||||
|
||||
uchar *cur= table->field[0]->ptr;
|
||||
/* first recinfo could be a NULL bitmap, not an actual Field */
|
||||
from_recinfo= to_recinfo= p->start_recinfo + (cur != table->record[0]);
|
||||
|
@ -8761,7 +8748,6 @@ static int optimize_schema_tables_memory_usage(TABLE_LIST *table_list)
|
|||
to_recinfo++;
|
||||
}
|
||||
p->recinfo= to_recinfo;
|
||||
} // XXX end of HACK HACK HACK
|
||||
|
||||
// TODO switch from Aria to Memory if all blobs were optimized away?
|
||||
if (instantiate_tmp_table(table, p->keyinfo, p->start_recinfo, &p->recinfo,
|
||||
|
@ -8924,6 +8910,7 @@ bool get_schema_tables_result(JOIN *join,
|
|||
}
|
||||
|
||||
Switch_to_definer_security_ctx backup_ctx(thd, table_list);
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
if (table_list->schema_table->fill_table(thd, table_list, cond))
|
||||
{
|
||||
result= 1;
|
||||
|
|
|
@ -2893,7 +2893,6 @@ int read_statistics_for_table(THD *thd, TABLE *table, TABLE_LIST *stat_tables)
|
|||
Field **field_ptr;
|
||||
KEY *key_info, *key_info_end;
|
||||
TABLE_SHARE *table_share= table->s;
|
||||
enum_check_fields old_check_level= thd->count_cuted_fields;
|
||||
|
||||
DBUG_ENTER("read_statistics_for_table");
|
||||
DEBUG_SYNC(thd, "statistics_mem_alloc_start1");
|
||||
|
@ -2909,7 +2908,7 @@ int read_statistics_for_table(THD *thd, TABLE *table, TABLE_LIST *stat_tables)
|
|||
}
|
||||
|
||||
/* Don't write warnings for internal field conversions */
|
||||
thd->count_cuted_fields= CHECK_FIELD_IGNORE;
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_IGNORE);
|
||||
|
||||
/* Read statistics from the statistical table table_stats */
|
||||
Table_statistics *read_stats= table_share->stats_cb.table_stats;
|
||||
|
@ -2991,7 +2990,6 @@ int read_statistics_for_table(THD *thd, TABLE *table, TABLE_LIST *stat_tables)
|
|||
}
|
||||
}
|
||||
|
||||
thd->count_cuted_fields= old_check_level;
|
||||
table_share->stats_cb.end_stats_load();
|
||||
DBUG_RETURN(0);
|
||||
}
|
||||
|
|
|
@ -10149,16 +10149,14 @@ do_continue:;
|
|||
if (use_inplace)
|
||||
{
|
||||
table->s->frm_image= &frm;
|
||||
enum_check_fields save_count_cuted_fields= thd->count_cuted_fields;
|
||||
/*
|
||||
Set the truncated column values of thd as warning
|
||||
for alter table.
|
||||
*/
|
||||
thd->count_cuted_fields = CHECK_FIELD_WARN;
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_WARN);
|
||||
int res= mysql_inplace_alter_table(thd, table_list, table, &altered_table,
|
||||
&ha_alter_info,
|
||||
&target_mdl_request, &alter_ctx);
|
||||
thd->count_cuted_fields= save_count_cuted_fields;
|
||||
my_free(const_cast<uchar*>(frm.str));
|
||||
|
||||
if (res)
|
||||
|
|
|
@ -828,7 +828,8 @@ static bool cmp_row_types(Item* item1, Item* item2)
|
|||
Item *inner= item1->element_index(i);
|
||||
Item *outer= item2->element_index(i);
|
||||
if (!inner->type_handler()->subquery_type_allows_materialization(inner,
|
||||
outer))
|
||||
outer,
|
||||
true))
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
|
@ -894,7 +895,7 @@ Item *Item_func_in::in_predicate_to_in_subs_transformer(THD *thd,
|
|||
|
||||
for (uint i=1; i < arg_count; i++)
|
||||
{
|
||||
if (!args[i]->const_item() || cmp_row_types(args[0], args[i]))
|
||||
if (!args[i]->const_item() || cmp_row_types(args[i], args[0]))
|
||||
return this;
|
||||
}
|
||||
|
||||
|
@ -974,6 +975,7 @@ Item *Item_func_in::in_predicate_to_in_subs_transformer(THD *thd,
|
|||
if (!(in_subs=
|
||||
new (thd->mem_root) Item_in_subselect(thd, args[0], sq_select)))
|
||||
goto err;
|
||||
in_subs->converted_from_in_predicate= TRUE;
|
||||
sq= in_subs;
|
||||
if (negated)
|
||||
sq= negate_expression(thd, in_subs);
|
||||
|
|
|
@ -19,6 +19,7 @@
|
|||
#include "sql_const.h"
|
||||
#include "sql_class.h"
|
||||
#include "sql_time.h"
|
||||
#include "sql_string.h"
|
||||
#include "item.h"
|
||||
#include "log.h"
|
||||
#include "tztime.h"
|
||||
|
@ -6683,7 +6684,8 @@ uint Type_handler_timestamp_common::Item_decimal_precision(const Item *item) con
|
|||
|
||||
bool Type_handler_real_result::
|
||||
subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
DBUG_ASSERT(inner->cmp_type() == REAL_RESULT);
|
||||
return outer->cmp_type() == REAL_RESULT;
|
||||
|
@ -6692,7 +6694,8 @@ bool Type_handler_real_result::
|
|||
|
||||
bool Type_handler_int_result::
|
||||
subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
DBUG_ASSERT(inner->cmp_type() == INT_RESULT);
|
||||
return outer->cmp_type() == INT_RESULT;
|
||||
|
@ -6701,7 +6704,8 @@ bool Type_handler_int_result::
|
|||
|
||||
bool Type_handler_decimal_result::
|
||||
subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
DBUG_ASSERT(inner->cmp_type() == DECIMAL_RESULT);
|
||||
return outer->cmp_type() == DECIMAL_RESULT;
|
||||
|
@ -6710,23 +6714,37 @@ bool Type_handler_decimal_result::
|
|||
|
||||
bool Type_handler_string_result::
|
||||
subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
DBUG_ASSERT(inner->cmp_type() == STRING_RESULT);
|
||||
return outer->cmp_type() == STRING_RESULT &&
|
||||
outer->collation.collation == inner->collation.collation &&
|
||||
/*
|
||||
Materialization also is unable to work when create_tmp_table() will
|
||||
create a blob column because item->max_length is too big.
|
||||
The following test is copied from varstring_type_handler().
|
||||
*/
|
||||
!inner->too_big_for_varchar();
|
||||
if (outer->cmp_type() == STRING_RESULT &&
|
||||
/*
|
||||
Materialization also is unable to work when create_tmp_table() will
|
||||
create a blob column because item->max_length is too big.
|
||||
The following test is copied from varstring_type_handler().
|
||||
*/
|
||||
!inner->too_big_for_varchar())
|
||||
{
|
||||
if (outer->collation.collation == inner->collation.collation)
|
||||
return true;
|
||||
if (is_in_predicate)
|
||||
{
|
||||
Charset inner_col(inner->collation.collation);
|
||||
if (inner_col.encoding_allows_reinterpret_as(outer->
|
||||
collation.collation) &&
|
||||
inner_col.eq_collation_specific_names(outer->collation.collation))
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
bool Type_handler_temporal_result::
|
||||
subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
DBUG_ASSERT(inner->cmp_type() == TIME_RESULT);
|
||||
return mysql_timestamp_type() ==
|
||||
|
|
|
@ -3718,9 +3718,21 @@ public:
|
|||
Item *target_expr, Item *target_value,
|
||||
Item_bool_func2 *source,
|
||||
Item *source_expr, Item *source_const) const= 0;
|
||||
|
||||
/*
|
||||
@brief
|
||||
Check if an IN subquery allows materialization or not
|
||||
@param
|
||||
inner expression on the inner side of the IN subquery
|
||||
outer expression on the outer side of the IN subquery
|
||||
is_in_predicate SET to true if IN subquery was converted from an
|
||||
IN predicate or we are checking if materialization
|
||||
strategy can be used for an IN predicate
|
||||
*/
|
||||
virtual bool
|
||||
subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const= 0;
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const= 0;
|
||||
/**
|
||||
Make a simple constant replacement item for a constant "src",
|
||||
so the new item can futher be used for comparison with "cmp", e.g.:
|
||||
|
@ -3966,7 +3978,8 @@ public:
|
|||
return 0;
|
||||
}
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
DBUG_ASSERT(0);
|
||||
return false;
|
||||
|
@ -4313,7 +4326,8 @@ public:
|
|||
const Field *field) const;
|
||||
int stored_field_cmp_to_item(THD *thd, Field *field, Item *item) const;
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const;
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const;
|
||||
void make_sort_key(uchar *to, Item *item, const SORT_FIELD_ATTR *sort_field,
|
||||
Sort_param *param) const;
|
||||
void sortlength(THD *thd,
|
||||
|
@ -4399,7 +4413,8 @@ public:
|
|||
return item_val.is_null() ? 0 : my_decimal(field).cmp(item_val.ptr());
|
||||
}
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const;
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const;
|
||||
Field *make_num_distinct_aggregator_field(MEM_ROOT *, const Item *) const;
|
||||
void make_sort_key(uchar *to, Item *item, const SORT_FIELD_ATTR *sort_field,
|
||||
Sort_param *param) const;
|
||||
|
@ -4624,7 +4639,8 @@ public:
|
|||
const Type_handler *type_handler_for_comparison() const;
|
||||
int stored_field_cmp_to_item(THD *thd, Field *field, Item *item) const;
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const;
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const;
|
||||
Field *make_num_distinct_aggregator_field(MEM_ROOT *, const Item *) const;
|
||||
void make_sort_key(uchar *to, Item *item, const SORT_FIELD_ATTR *sort_field,
|
||||
Sort_param *param) const;
|
||||
|
@ -4733,7 +4749,8 @@ public:
|
|||
Item_bool_func2 *source,
|
||||
Item *source_expr, Item *source_const) const;
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const;
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const;
|
||||
bool Item_func_min_max_fix_attributes(THD *thd, Item_func_min_max *func,
|
||||
Item **items, uint nitems) const;
|
||||
bool Item_sum_hybrid_fix_length_and_dec(Item_sum_hybrid *func) const;
|
||||
|
@ -4855,7 +4872,8 @@ public:
|
|||
Item_bool_func2 *source,
|
||||
Item *source_expr, Item *source_const) const;
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const;
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const;
|
||||
Item *make_const_item_for_comparison(THD *, Item *src, const Item *cmp) const;
|
||||
Item_cache *Item_get_cache(THD *thd, const Item *item) const;
|
||||
bool set_comparator_func(Arg_comparator *cmp) const;
|
||||
|
@ -6185,7 +6203,8 @@ public:
|
|||
return blob_type_handler(item);
|
||||
}
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
return false; // Materialization does not work with BLOB columns
|
||||
}
|
||||
|
@ -6304,7 +6323,8 @@ public:
|
|||
return true;
|
||||
}
|
||||
bool subquery_type_allows_materialization(const Item *inner,
|
||||
const Item *outer) const
|
||||
const Item *outer,
|
||||
bool is_in_predicate) const
|
||||
{
|
||||
return false; // Materialization does not work with GEOMETRY columns
|
||||
}
|
||||
|
|
|
@ -2102,6 +2102,7 @@ bool st_select_lex::cleanup()
|
|||
delete join;
|
||||
join= 0;
|
||||
}
|
||||
leaf_tables.empty();
|
||||
for (SELECT_LEX_UNIT *lex_unit= first_inner_unit(); lex_unit ;
|
||||
lex_unit= lex_unit->next_unit())
|
||||
{
|
||||
|
|
|
@ -4123,7 +4123,7 @@ static bool fix_autocommit(sys_var *self, THD *thd, enum_var_type type)
|
|||
if (trans_commit_stmt(thd) || trans_commit(thd))
|
||||
{
|
||||
thd->variables.option_bits&= ~OPTION_AUTOCOMMIT;
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
WSREP_DEBUG("autocommit, MDL TRX lock released: %lld",
|
||||
(longlong) thd->thread_id);
|
||||
return true;
|
||||
|
|
|
@ -8499,7 +8499,7 @@ int TABLE::period_make_insert(Item *src, Field *dst)
|
|||
{
|
||||
THD *thd= in_use;
|
||||
|
||||
file->store_auto_increment();
|
||||
ulonglong prev_insert_id= file->next_insert_id;
|
||||
store_record(this, record[1]);
|
||||
int res= src->save_in_field(dst, true);
|
||||
|
||||
|
@ -8519,7 +8519,7 @@ int TABLE::period_make_insert(Item *src, Field *dst)
|
|||
|
||||
restore_record(this, record[1]);
|
||||
if (res)
|
||||
file->restore_auto_increment();
|
||||
file->restore_auto_increment(prev_insert_id);
|
||||
return res;
|
||||
}
|
||||
|
||||
|
|
|
@ -209,12 +209,11 @@ void tp_callback(TP_connection *c)
|
|||
|
||||
error:
|
||||
c->thd= 0;
|
||||
delete c;
|
||||
|
||||
if (thd)
|
||||
{
|
||||
threadpool_remove_connection(thd);
|
||||
}
|
||||
delete c;
|
||||
worker_context.restore();
|
||||
}
|
||||
|
||||
|
|
|
@ -138,7 +138,7 @@ bool trans_begin(THD *thd, uint flags)
|
|||
Release transactional metadata locks only after the
|
||||
transaction has been committed.
|
||||
*/
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
thd->release_transactional_locks();
|
||||
|
||||
// The RO/RW options are mutually exclusive.
|
||||
DBUG_ASSERT(!((flags & MYSQL_START_TRANS_OPT_READ_ONLY) &&
|
||||
|
|
|
@ -1028,7 +1028,6 @@ static bool make_empty_rec(THD *thd, uchar *buff, uint table_options,
|
|||
TABLE table;
|
||||
TABLE_SHARE share;
|
||||
Create_field *field;
|
||||
enum_check_fields old_count_cuted_fields= thd->count_cuted_fields;
|
||||
DBUG_ENTER("make_empty_rec");
|
||||
|
||||
/* We need a table to generate columns for default values */
|
||||
|
@ -1047,7 +1046,7 @@ static bool make_empty_rec(THD *thd, uchar *buff, uint table_options,
|
|||
null_pos= buff;
|
||||
|
||||
List_iterator<Create_field> it(create_fields);
|
||||
thd->count_cuted_fields= CHECK_FIELD_WARN; // To find wrong default values
|
||||
Check_level_instant_set check_level_save(thd, CHECK_FIELD_WARN);
|
||||
while ((field=it++))
|
||||
{
|
||||
Record_addr addr(buff + field->offset + data_offset,
|
||||
|
@ -1094,6 +1093,5 @@ static bool make_empty_rec(THD *thd, uchar *buff, uint table_options,
|
|||
*(null_pos + null_count / 8)|= ~(((uchar) 1 << (null_count & 7)) - 1);
|
||||
|
||||
err:
|
||||
thd->count_cuted_fields= old_count_cuted_fields;
|
||||
DBUG_RETURN(error);
|
||||
} /* make_empty_rec */
|
||||
|
|
|
@ -513,11 +513,13 @@ bool trans_xa_prepare(THD *thd)
|
|||
|
||||
/**
|
||||
Commit and terminate the a XA transaction.
|
||||
Transactional locks are released if transaction ended
|
||||
|
||||
@param thd Current thread
|
||||
|
||||
@retval FALSE Success
|
||||
@retval TRUE Failure
|
||||
|
||||
*/
|
||||
|
||||
bool trans_xa_commit(THD *thd)
|
||||
|
@ -601,6 +603,7 @@ bool trans_xa_commit(THD *thd)
|
|||
xid_cache_delete(thd, &thd->transaction.xid_state);
|
||||
|
||||
trans_track_end_trx(thd);
|
||||
thd->mdl_context.release_transactional_locks();
|
||||
|
||||
DBUG_RETURN(res);
|
||||
}
|
||||
|
@ -608,6 +611,7 @@ bool trans_xa_commit(THD *thd)
|
|||
|
||||
/**
|
||||
Roll back and terminate a XA transaction.
|
||||
Transactional locks are released if transaction ended
|
||||
|
||||
@param thd Current thread
|
||||
|
||||
|
|
|
@ -1986,6 +1986,7 @@ dict_index_remove_from_cache_low(
|
|||
if (index->online_log) {
|
||||
ut_ad(index->online_status == ONLINE_INDEX_CREATION);
|
||||
row_log_free(index->online_log);
|
||||
index->online_log = NULL;
|
||||
}
|
||||
|
||||
/* Remove the index from the list of indexes of the table */
|
||||
|
|
|
@ -1980,7 +1980,7 @@ static index_stats_t dict_stats_analyze_index(dict_index_t* index)
|
|||
since it will be faster and will give better results. */
|
||||
|
||||
if (root_level == 0
|
||||
|| N_SAMPLE_PAGES(index) * n_uniq > index->stat_n_leaf_pages) {
|
||||
|| N_SAMPLE_PAGES(index) * n_uniq > result.n_leaf_pages) {
|
||||
|
||||
if (root_level == 0) {
|
||||
DEBUG_PRINTF(" %s(): just one page,"
|
||||
|
|
|
@ -7047,6 +7047,7 @@ error_handling_drop_uncached:
|
|||
if (ok && a == 1) {
|
||||
row_log_free(
|
||||
index->online_log);
|
||||
index->online_log = NULL;
|
||||
ok = false;
|
||||
});
|
||||
|
||||
|
@ -8619,6 +8620,7 @@ innobase_online_rebuild_log_free(
|
|||
== ONLINE_INDEX_CREATION);
|
||||
clust_index->online_status = ONLINE_INDEX_COMPLETE;
|
||||
row_log_free(clust_index->online_log);
|
||||
clust_index->online_log = NULL;
|
||||
DEBUG_SYNC_C("innodb_online_rebuild_log_free_aborted");
|
||||
}
|
||||
|
||||
|
|
|
@ -1,7 +1,7 @@
|
|||
/*****************************************************************************
|
||||
|
||||
Copyright (c) 2011, 2016, Oracle and/or its affiliates. All Rights Reserved.
|
||||
Copyright (c) 2017, MariaDB Corporation.
|
||||
Copyright (c) 2017, 2020, MariaDB Corporation.
|
||||
|
||||
This program is free software; you can redistribute it and/or modify it under
|
||||
the terms of the GNU General Public License as published by the Free Software
|
||||
|
@ -69,7 +69,7 @@ Free the row log for an index that was being created online. */
|
|||
void
|
||||
row_log_free(
|
||||
/*=========*/
|
||||
row_log_t*& log) /*!< in,own: row log */
|
||||
row_log_t* log) /*!< in,own: row log */
|
||||
MY_ATTRIBUTE((nonnull));
|
||||
|
||||
/******************************************************//**
|
||||
|
|
|
@ -1,6 +1,7 @@
|
|||
/*****************************************************************************
|
||||
|
||||
Copyright (c) 2011, 2015, Oracle and/or its affiliates. All Rights Reserved.
|
||||
Copyright (c) 2020, MariaDB Corporation.
|
||||
|
||||
This program is free software; you can redistribute it and/or modify it under
|
||||
the terms of the GNU General Public License as published by the Free Software
|
||||
|
@ -38,6 +39,7 @@ row_log_abort_sec(
|
|||
ut_ad(!dict_index_is_clust(index));
|
||||
dict_index_set_online_status(index, ONLINE_INDEX_ABORTED);
|
||||
row_log_free(index->online_log);
|
||||
index->online_log = NULL;
|
||||
}
|
||||
|
||||
/******************************************************//**
|
||||
|
|
|
@ -735,7 +735,7 @@ inline lsn_t mtr_t::finish_write(ulint len)
|
|||
return start_lsn;
|
||||
}
|
||||
|
||||
/** Find out whether a block was X-latched by the mini-transaction */
|
||||
/** Find out whether a block was not X-latched by the mini-transaction */
|
||||
struct FindBlockX
|
||||
{
|
||||
const buf_block_t █
|
||||
|
@ -745,7 +745,7 @@ struct FindBlockX
|
|||
/** @return whether the block was not found x-latched */
|
||||
bool operator()(const mtr_memo_slot_t *slot) const
|
||||
{
|
||||
return slot->object != &block || slot->type == MTR_MEMO_PAGE_X_FIX;
|
||||
return slot->object != &block || slot->type != MTR_MEMO_PAGE_X_FIX;
|
||||
}
|
||||
};
|
||||
|
||||
|
|
|
@ -3240,7 +3240,6 @@ row_log_allocate(
|
|||
}
|
||||
|
||||
dict_index_set_online_status(index, ONLINE_INDEX_CREATION);
|
||||
index->online_log = log;
|
||||
|
||||
if (log_tmp_is_encrypted()) {
|
||||
ulint size = srv_sort_buf_size;
|
||||
|
@ -3253,6 +3252,7 @@ row_log_allocate(
|
|||
}
|
||||
}
|
||||
|
||||
index->online_log = log;
|
||||
/* While we might be holding an exclusive data dictionary lock
|
||||
here, in row_log_abort_sec() we will not always be holding it. Use
|
||||
atomic operations in both cases. */
|
||||
|
@ -3266,7 +3266,7 @@ Free the row log for an index that was being created online. */
|
|||
void
|
||||
row_log_free(
|
||||
/*=========*/
|
||||
row_log_t*& log) /*!< in,own: row log */
|
||||
row_log_t* log) /*!< in,own: row log */
|
||||
{
|
||||
MONITOR_ATOMIC_DEC(MONITOR_ONLINE_CREATE_INDEX);
|
||||
|
||||
|
@ -3286,7 +3286,6 @@ row_log_free(
|
|||
|
||||
mutex_free(&log->mutex);
|
||||
ut_free(log);
|
||||
log = NULL;
|
||||
}
|
||||
|
||||
/******************************************************//**
|
||||
|
|
|
@ -448,6 +448,13 @@ close_table:
|
|||
node->heap);
|
||||
} else {
|
||||
node->ref = &trx_undo_metadata;
|
||||
if (!row_undo_search_clust_to_pcur(node)) {
|
||||
/* An error probably occurred during
|
||||
an insert into the clustered index,
|
||||
after we wrote the undo log record. */
|
||||
goto close_table;
|
||||
}
|
||||
return true;
|
||||
}
|
||||
|
||||
if (!row_undo_search_clust_to_pcur(node)) {
|
||||
|
|
|
@ -1337,7 +1337,7 @@ close_table:
|
|||
}
|
||||
|
||||
/* Extract indexed virtual columns from undo log */
|
||||
if (node->table->n_v_cols) {
|
||||
if (node->ref != &trx_undo_metadata && node->table->n_v_cols) {
|
||||
row_upd_replace_vcol(node->row, node->table,
|
||||
node->update, false, node->undo_row,
|
||||
(node->cmpl_info & UPD_NODE_NO_ORD_CHANGE)
|
||||
|
|
|
@ -571,11 +571,6 @@ public:
|
|||
void set_next_insert_id(ulonglong id);
|
||||
void get_auto_increment(ulonglong offset, ulonglong increment, ulonglong nb_desired_values,
|
||||
ulonglong *first_value, ulonglong *nb_reserved_values) mrn_override;
|
||||
/** Fix spurious -Werror=overloaded-virtual in GCC 9 */
|
||||
void restore_auto_increment() mrn_override
|
||||
{
|
||||
handler::restore_auto_increment();
|
||||
}
|
||||
void restore_auto_increment(ulonglong prev_insert_id) mrn_override;
|
||||
void release_auto_increment() mrn_override;
|
||||
int check_for_upgrade(HA_CHECK_OPT *check_opt) mrn_override;
|
||||
|
|
|
@ -20,6 +20,10 @@ INCLUDE_DIRECTORIES(${Judy_INCLUDE_DIR})
|
|||
SET(OQGRAPH_OK 1)
|
||||
ENDMACRO()
|
||||
|
||||
IF(PLUGIN_OQGRAPH STREQUAL "NO")
|
||||
RETURN()
|
||||
ENDIF()
|
||||
|
||||
IF(NOT DEFINED OQGRAPH_OK)
|
||||
CHECK_OQGRAPH()
|
||||
IF (NOT OQGRAPH_OK)
|
||||
|
|
|
@ -87,7 +87,7 @@ let $checkpoint = $MYSQL_TMP_DIR/already-existing-directory;
|
|||
--mkdir $checkpoint
|
||||
let $succeeds = 0;
|
||||
--source set_checkpoint.inc
|
||||
--exec rm -rf $checkpoint
|
||||
rmdir $checkpoint;
|
||||
|
||||
--disable_result_log
|
||||
truncate table t1;
|
||||
|
|
Loading…
Reference in a new issue