mariadb/storage/innobase/buf/buf0dblwr.cc

1247 lines
35 KiB
C++

/*****************************************************************************
Copyright (c) 1995, 2017, Oracle and/or its affiliates. All Rights Reserved.
Copyright (c) 2013, 2019, MariaDB Corporation.
This program is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free Software
Foundation; version 2 of the License.
This program is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with
this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA
*****************************************************************************/
/**************************************************//**
@file buf/buf0dblwr.cc
Doublwrite buffer module
Created 2011/12/19
*******************************************************/
#include "buf0dblwr.h"
#include "buf0buf.h"
#include "buf0checksum.h"
#include "srv0start.h"
#include "srv0srv.h"
#include "page0zip.h"
#include "trx0sys.h"
#include "fil0crypt.h"
#include "fil0pagecompress.h"
/** The doublewrite buffer */
buf_dblwr_t* buf_dblwr = NULL;
/** Set to TRUE when the doublewrite buffer is being created */
ibool buf_dblwr_being_created = FALSE;
#define TRX_SYS_DOUBLEWRITE_BLOCKS 2
/****************************************************************//**
Determines if a page number is located inside the doublewrite buffer.
@return TRUE if the location is inside the two blocks of the
doublewrite buffer */
ibool
buf_dblwr_page_inside(
/*==================*/
ulint page_no) /*!< in: page number */
{
if (buf_dblwr == NULL) {
return(FALSE);
}
if (page_no >= buf_dblwr->block1
&& page_no < buf_dblwr->block1
+ TRX_SYS_DOUBLEWRITE_BLOCK_SIZE) {
return(TRUE);
}
if (page_no >= buf_dblwr->block2
&& page_no < buf_dblwr->block2
+ TRX_SYS_DOUBLEWRITE_BLOCK_SIZE) {
return(TRUE);
}
return(FALSE);
}
/** @return the TRX_SYS page */
inline buf_block_t *buf_dblwr_trx_sys_get(mtr_t *mtr)
{
buf_block_t *block= buf_page_get(page_id_t(TRX_SYS_SPACE, TRX_SYS_PAGE_NO),
0, RW_X_LATCH, mtr);
buf_block_dbg_add_level(block, SYNC_NO_ORDER_CHECK);
return block;
}
/********************************************************************//**
Flush a batch of writes to the datafiles that have already been
written to the dblwr buffer on disk. */
void
buf_dblwr_sync_datafiles()
/*======================*/
{
/* Wait that all async writes to tablespaces have been posted to
the OS */
os_aio_wait_until_no_pending_writes();
}
/****************************************************************//**
Creates or initialializes the doublewrite buffer at a database start. */
static void buf_dblwr_init(const byte *doublewrite)
{
ulint buf_size;
buf_dblwr = static_cast<buf_dblwr_t*>(
ut_zalloc_nokey(sizeof(buf_dblwr_t)));
/* There are two blocks of same size in the doublewrite
buffer. */
buf_size = TRX_SYS_DOUBLEWRITE_BLOCKS * TRX_SYS_DOUBLEWRITE_BLOCK_SIZE;
/* There must be atleast one buffer for single page writes
and one buffer for batch writes. */
ut_a(srv_doublewrite_batch_size > 0
&& srv_doublewrite_batch_size < buf_size);
mutex_create(LATCH_ID_BUF_DBLWR, &buf_dblwr->mutex);
buf_dblwr->b_event = os_event_create("dblwr_batch_event");
buf_dblwr->s_event = os_event_create("dblwr_single_event");
buf_dblwr->first_free = 0;
buf_dblwr->s_reserved = 0;
buf_dblwr->b_reserved = 0;
buf_dblwr->block1 = mach_read_from_4(
doublewrite + TRX_SYS_DOUBLEWRITE_BLOCK1);
buf_dblwr->block2 = mach_read_from_4(
doublewrite + TRX_SYS_DOUBLEWRITE_BLOCK2);
buf_dblwr->in_use = static_cast<bool*>(
ut_zalloc_nokey(buf_size * sizeof(bool)));
buf_dblwr->write_buf = static_cast<byte*>(
aligned_malloc(buf_size << srv_page_size_shift,
srv_page_size));
buf_dblwr->buf_block_arr = static_cast<buf_page_t**>(
ut_zalloc_nokey(buf_size * sizeof(void*)));
}
/** Create the doublewrite buffer if the doublewrite buffer header
is not present in the TRX_SYS page.
@return whether the operation succeeded
@retval true if the doublewrite buffer exists or was created
@retval false if the creation failed (too small first data file) */
bool
buf_dblwr_create()
{
buf_block_t* block2;
buf_block_t* new_block;
byte* fseg_header;
ulint page_no;
ulint prev_page_no;
ulint i;
mtr_t mtr;
if (buf_dblwr) {
/* Already inited */
return(true);
}
start_again:
mtr.start();
buf_dblwr_being_created = TRUE;
buf_block_t *trx_sys_block = buf_dblwr_trx_sys_get(&mtr);
if (mach_read_from_4(TRX_SYS_DOUBLEWRITE + TRX_SYS_DOUBLEWRITE_MAGIC
+ trx_sys_block->frame)
== TRX_SYS_DOUBLEWRITE_MAGIC_N) {
/* The doublewrite buffer has already been created:
just read in some numbers */
buf_dblwr_init(TRX_SYS_DOUBLEWRITE + trx_sys_block->frame);
mtr.commit();
buf_dblwr_being_created = FALSE;
return(true);
} else {
if (UT_LIST_GET_FIRST(fil_system.sys_space->chain)->size
< 3 * FSP_EXTENT_SIZE) {
goto too_small;
}
}
block2 = fseg_create(fil_system.sys_space, TRX_SYS_PAGE_NO,
TRX_SYS_DOUBLEWRITE
+ TRX_SYS_DOUBLEWRITE_FSEG, &mtr);
if (block2 == NULL) {
too_small:
ib::error()
<< "Cannot create doublewrite buffer: "
"the first file in innodb_data_file_path"
" must be at least "
<< (3 * (FSP_EXTENT_SIZE
>> (20U - srv_page_size_shift)))
<< "M.";
mtr.commit();
return(false);
}
ib::info() << "Doublewrite buffer not found: creating new";
/* FIXME: After this point, the doublewrite buffer creation
is not atomic. The doublewrite buffer should not exist in
the InnoDB system tablespace file in the first place.
It could be located in separate optional file(s) in a
user-specified location. */
/* fseg_create acquires a second latch on the page,
therefore we must declare it: */
buf_block_dbg_add_level(block2, SYNC_NO_ORDER_CHECK);
fseg_header = TRX_SYS_DOUBLEWRITE + TRX_SYS_DOUBLEWRITE_FSEG
+ trx_sys_block->frame;
prev_page_no = 0;
for (i = 0; i < TRX_SYS_DOUBLEWRITE_BLOCKS * TRX_SYS_DOUBLEWRITE_BLOCK_SIZE
+ FSP_EXTENT_SIZE / 2; i++) {
new_block = fseg_alloc_free_page(
fseg_header, prev_page_no + 1, FSP_UP, &mtr);
if (new_block == NULL) {
ib::error() << "Cannot create doublewrite buffer: "
" you must increase your tablespace size."
" Cannot continue operation.";
/* This may essentially corrupt the doublewrite
buffer. However, usually the doublewrite buffer
is created at database initialization, and it
should not matter (just remove all newly created
InnoDB files and restart). */
mtr.commit();
return(false);
}
/* We read the allocated pages to the buffer pool;
when they are written to disk in a flush, the space
id and page number fields are also written to the
pages. When we at database startup read pages
from the doublewrite buffer, we know that if the
space id and page number in them are the same as
the page position in the tablespace, then the page
has not been written to in doublewrite. */
ut_ad(rw_lock_get_x_lock_count(&new_block->lock) == 1);
page_no = new_block->page.id.page_no();
/* We only do this in the debug build, to ensure that
both the check in buf_flush_init_for_writing() and
recv_parse_or_apply_log_rec_body() will see a valid
page type. The flushes of new_block are actually
unnecessary here. */
ut_d(mtr.write<2>(*new_block,
FIL_PAGE_TYPE + new_block->frame,
FIL_PAGE_TYPE_SYS));
if (i == FSP_EXTENT_SIZE / 2) {
ut_a(page_no == FSP_EXTENT_SIZE);
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE
+ TRX_SYS_DOUBLEWRITE_BLOCK1
+ trx_sys_block->frame,
page_no);
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE
+ TRX_SYS_DOUBLEWRITE_REPEAT
+ TRX_SYS_DOUBLEWRITE_BLOCK1
+ trx_sys_block->frame,
page_no);
} else if (i == FSP_EXTENT_SIZE / 2
+ TRX_SYS_DOUBLEWRITE_BLOCK_SIZE) {
ut_a(page_no == 2 * FSP_EXTENT_SIZE);
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE
+ TRX_SYS_DOUBLEWRITE_BLOCK2
+ trx_sys_block->frame,
page_no);
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE
+ TRX_SYS_DOUBLEWRITE_REPEAT
+ TRX_SYS_DOUBLEWRITE_BLOCK2
+ trx_sys_block->frame,
page_no);
} else if (i > FSP_EXTENT_SIZE / 2) {
ut_a(page_no == prev_page_no + 1);
}
if (((i + 1) & 15) == 0) {
/* rw_locks can only be recursively x-locked
2048 times. (on 32 bit platforms,
(lint) 0 - (X_LOCK_DECR * 2049)
is no longer a negative number, and thus
lock_word becomes like a shared lock).
For 4k page size this loop will
lock the fseg header too many times. Since
this code is not done while any other threads
are active, restart the MTR occasionally. */
mtr.commit();
mtr.start();
trx_sys_block = buf_dblwr_trx_sys_get(&mtr);
fseg_header = TRX_SYS_DOUBLEWRITE
+ TRX_SYS_DOUBLEWRITE_FSEG
+ trx_sys_block->frame;
}
prev_page_no = page_no;
}
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE + TRX_SYS_DOUBLEWRITE_MAGIC
+ trx_sys_block->frame,
TRX_SYS_DOUBLEWRITE_MAGIC_N);
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE + TRX_SYS_DOUBLEWRITE_MAGIC
+ TRX_SYS_DOUBLEWRITE_REPEAT
+ trx_sys_block->frame,
TRX_SYS_DOUBLEWRITE_MAGIC_N);
mtr.write<4>(*trx_sys_block,
TRX_SYS_DOUBLEWRITE + TRX_SYS_DOUBLEWRITE_SPACE_ID_STORED
+ trx_sys_block->frame,
TRX_SYS_DOUBLEWRITE_SPACE_ID_STORED_N);
mtr.commit();
/* Flush the modified pages to disk and make a checkpoint */
log_make_checkpoint();
buf_dblwr_being_created = FALSE;
/* Remove doublewrite pages from LRU */
buf_pool_invalidate();
ib::info() << "Doublewrite buffer created";
goto start_again;
}
/**
At database startup initializes the doublewrite buffer memory structure if
we already have a doublewrite buffer created in the data files. If we are
upgrading to an InnoDB version which supports multiple tablespaces, then this
function performs the necessary update operations. If we are in a crash
recovery, this function loads the pages from double write buffer into memory.
@param[in] file File handle
@param[in] path Path name of file
@return DB_SUCCESS or error code */
dberr_t
buf_dblwr_init_or_load_pages(
pfs_os_file_t file,
const char* path)
{
byte* buf;
byte* page;
ulint block1;
ulint block2;
ulint space_id;
byte* read_buf;
byte* doublewrite;
ibool reset_space_ids = FALSE;
recv_dblwr_t& recv_dblwr = recv_sys.dblwr;
/* We do the file i/o past the buffer pool */
read_buf = static_cast<byte*>(
aligned_malloc(2 * srv_page_size, srv_page_size));
/* Read the trx sys header to check if we are using the doublewrite
buffer */
dberr_t err;
IORequest read_request(IORequest::READ);
err = os_file_read(
read_request,
file, read_buf, TRX_SYS_PAGE_NO << srv_page_size_shift,
srv_page_size);
if (err != DB_SUCCESS) {
ib::error()
<< "Failed to read the system tablespace header page";
func_exit:
aligned_free(read_buf);
return(err);
}
doublewrite = read_buf + TRX_SYS_DOUBLEWRITE;
/* TRX_SYS_PAGE_NO is not encrypted see fil_crypt_rotate_page() */
if (mach_read_from_4(doublewrite + TRX_SYS_DOUBLEWRITE_MAGIC)
== TRX_SYS_DOUBLEWRITE_MAGIC_N) {
/* The doublewrite buffer has been created */
buf_dblwr_init(doublewrite);
block1 = buf_dblwr->block1;
block2 = buf_dblwr->block2;
buf = buf_dblwr->write_buf;
} else {
err = DB_SUCCESS;
goto func_exit;
}
if (mach_read_from_4(doublewrite + TRX_SYS_DOUBLEWRITE_SPACE_ID_STORED)
!= TRX_SYS_DOUBLEWRITE_SPACE_ID_STORED_N) {
/* We are upgrading from a version < 4.1.x to a version where
multiple tablespaces are supported. We must reset the space id
field in the pages in the doublewrite buffer because starting
from this version the space id is stored to
FIL_PAGE_ARCH_LOG_NO_OR_SPACE_ID. */
reset_space_ids = TRUE;
ib::info() << "Resetting space id's in the doublewrite buffer";
}
/* Read the pages from the doublewrite buffer to memory */
err = os_file_read(
read_request,
file, buf, block1 << srv_page_size_shift,
TRX_SYS_DOUBLEWRITE_BLOCK_SIZE << srv_page_size_shift);
if (err != DB_SUCCESS) {
ib::error()
<< "Failed to read the first double write buffer "
"extent";
goto func_exit;
}
err = os_file_read(
read_request,
file,
buf + (TRX_SYS_DOUBLEWRITE_BLOCK_SIZE << srv_page_size_shift),
block2 << srv_page_size_shift,
TRX_SYS_DOUBLEWRITE_BLOCK_SIZE << srv_page_size_shift);
if (err != DB_SUCCESS) {
ib::error()
<< "Failed to read the second double write buffer "
"extent";
goto func_exit;
}
/* Check if any of these pages is half-written in data files, in the
intended position */
page = buf;
for (ulint i = 0; i < TRX_SYS_DOUBLEWRITE_BLOCK_SIZE * 2; i++) {
if (reset_space_ids) {
ulint source_page_no;
space_id = 0;
mach_write_to_4(page + FIL_PAGE_ARCH_LOG_NO_OR_SPACE_ID,
space_id);
/* We do not need to calculate new checksums for the
pages because the field .._SPACE_ID does not affect
them. Write the page back to where we read it from. */
if (i < TRX_SYS_DOUBLEWRITE_BLOCK_SIZE) {
source_page_no = block1 + i;
} else {
source_page_no = block2
+ i - TRX_SYS_DOUBLEWRITE_BLOCK_SIZE;
}
err = os_file_write(
IORequestWrite, path, file, page,
source_page_no << srv_page_size_shift,
srv_page_size);
if (err != DB_SUCCESS) {
ib::error()
<< "Failed to write to the double write"
" buffer";
goto func_exit;
}
} else if (mach_read_from_8(page + FIL_PAGE_LSN)) {
/* Each valid page header must contain
a nonzero FIL_PAGE_LSN field. */
recv_dblwr.add(page);
}
page += srv_page_size;
}
if (reset_space_ids) {
os_file_flush(file);
}
err = DB_SUCCESS;
goto func_exit;
}
/** Process and remove the double write buffer pages for all tablespaces. */
void
buf_dblwr_process()
{
ulint page_no_dblwr = 0;
byte* read_buf;
recv_dblwr_t& recv_dblwr = recv_sys.dblwr;
if (!buf_dblwr) {
return;
}
read_buf = static_cast<byte*>(
aligned_malloc(2 * srv_page_size, srv_page_size));
byte* const buf = read_buf + srv_page_size;
for (recv_dblwr_t::list::iterator i = recv_dblwr.pages.begin();
i != recv_dblwr.pages.end();
++i, ++page_no_dblwr) {
byte* page = *i;
ulint space_id = page_get_space_id(page);
fil_space_t* space = fil_space_get(space_id);
if (space == NULL) {
/* Maybe we have dropped the tablespace
and this page once belonged to it: do nothing */
continue;
}
fil_space_open_if_needed(space);
const ulint page_no = page_get_page_no(page);
const page_id_t page_id(space_id, page_no);
if (page_no >= space->size) {
/* Do not report the warning for undo
tablespaces, because they can be truncated in place. */
if (!srv_is_undo_tablespace(space_id)) {
ib::warn() << "A copy of page " << page_id
<< " in the doublewrite buffer slot "
<< page_no_dblwr
<< " is not within space bounds";
}
continue;
}
const ulint physical_size = space->physical_size();
const ulint zip_size = space->zip_size();
ut_ad(!buf_page_is_zeroes(page, physical_size));
/* We want to ensure that for partial reads the
unread portion of the page is NUL. */
memset(read_buf, 0x0, physical_size);
IORequest request;
request.dblwr_recover();
/* Read in the actual page from the file */
dberr_t err = fil_io(
request, true,
page_id, zip_size,
0, physical_size, read_buf, NULL);
if (err != DB_SUCCESS) {
ib::warn()
<< "Double write buffer recovery: "
<< page_id << " read failed with "
<< "error: " << ut_strerr(err);
}
const bool is_all_zero = buf_page_is_zeroes(
read_buf, physical_size);
const bool expect_encrypted = space->crypt_data
&& space->crypt_data->type != CRYPT_SCHEME_UNENCRYPTED;
bool is_corrupted = false;
if (is_all_zero) {
/* We will check if the copy in the
doublewrite buffer is valid. If not, we will
ignore this page (there should be redo log
records to initialize it). */
} else {
/* Decompress the page before
validating the checksum. */
ulint decomp = fil_page_decompress(buf, read_buf,
space->flags);
if (!decomp || (zip_size && decomp != srv_page_size)) {
goto bad;
}
if (expect_encrypted
&& buf_page_get_key_version(read_buf, space->flags)) {
is_corrupted = !buf_page_verify_crypt_checksum(
read_buf, space->flags);
} else {
is_corrupted = buf_page_is_corrupted(
true, read_buf, space->flags);
}
if (!is_corrupted) {
/* The page is good; there is no need
to consult the doublewrite buffer. */
continue;
}
bad:
/* We intentionally skip this message for
is_all_zero pages. */
ib::info()
<< "Trying to recover page " << page_id
<< " from the doublewrite buffer.";
}
ulint decomp = fil_page_decompress(buf, page, space->flags);
if (!decomp || (zip_size && decomp != srv_page_size)) {
continue;
}
if (expect_encrypted
&& buf_page_get_key_version(read_buf, space->flags)) {
is_corrupted = !buf_page_verify_crypt_checksum(
page, space->flags);
} else {
is_corrupted = buf_page_is_corrupted(
true, page, space->flags);
}
if (is_corrupted) {
/* Theoretically we could have another good
copy for this page in the doublewrite
buffer. If not, we will report a fatal error
for a corrupted page somewhere else if that
page was truly needed. */
continue;
}
if (page_no == 0) {
/* Check the FSP_SPACE_FLAGS. */
ulint flags = fsp_header_get_flags(page);
if (!fil_space_t::is_valid_flags(flags, space_id)
&& fsp_flags_convert_from_101(flags)
== ULINT_UNDEFINED) {
ib::warn() << "Ignoring a doublewrite copy"
" of page " << page_id
<< " due to invalid flags "
<< ib::hex(flags);
continue;
}
/* The flags on the page should be converted later. */
}
/* Write the good page from the doublewrite buffer to
the intended position. */
IORequest write_request(IORequest::WRITE);
fil_io(write_request, true, page_id, zip_size,
0, physical_size,
const_cast<byte*>(page), NULL);
ib::info() << "Recovered page " << page_id
<< " from the doublewrite buffer.";
}
recv_dblwr.pages.clear();
fil_flush_file_spaces(FIL_TYPE_TABLESPACE);
aligned_free(read_buf);
}
/****************************************************************//**
Frees doublewrite buffer. */
void
buf_dblwr_free()
{
/* Free the double write data structures. */
ut_a(buf_dblwr != NULL);
ut_ad(buf_dblwr->s_reserved == 0);
ut_ad(buf_dblwr->b_reserved == 0);
os_event_destroy(buf_dblwr->b_event);
os_event_destroy(buf_dblwr->s_event);
aligned_free(buf_dblwr->write_buf);
ut_free(buf_dblwr->buf_block_arr);
ut_free(buf_dblwr->in_use);
mutex_free(&buf_dblwr->mutex);
ut_free(buf_dblwr);
buf_dblwr = NULL;
}
/********************************************************************//**
Updates the doublewrite buffer when an IO request is completed. */
void
buf_dblwr_update(
/*=============*/
const buf_page_t* bpage, /*!< in: buffer block descriptor */
buf_flush_t flush_type)/*!< in: flush type */
{
ut_ad(srv_use_doublewrite_buf);
ut_ad(buf_dblwr);
ut_ad(!fsp_is_system_temporary(bpage->id.space()));
ut_ad(!srv_read_only_mode);
switch (flush_type) {
case BUF_FLUSH_LIST:
case BUF_FLUSH_LRU:
mutex_enter(&buf_dblwr->mutex);
ut_ad(buf_dblwr->batch_running);
ut_ad(buf_dblwr->b_reserved > 0);
ut_ad(buf_dblwr->b_reserved <= buf_dblwr->first_free);
buf_dblwr->b_reserved--;
if (buf_dblwr->b_reserved == 0) {
mutex_exit(&buf_dblwr->mutex);
/* This will finish the batch. Sync data files
to the disk. */
fil_flush_file_spaces(FIL_TYPE_TABLESPACE);
mutex_enter(&buf_dblwr->mutex);
/* We can now reuse the doublewrite memory buffer: */
buf_dblwr->first_free = 0;
buf_dblwr->batch_running = false;
os_event_set(buf_dblwr->b_event);
}
mutex_exit(&buf_dblwr->mutex);
break;
case BUF_FLUSH_SINGLE_PAGE:
{
const ulint size = TRX_SYS_DOUBLEWRITE_BLOCKS * TRX_SYS_DOUBLEWRITE_BLOCK_SIZE;
ulint i;
mutex_enter(&buf_dblwr->mutex);
for (i = srv_doublewrite_batch_size; i < size; ++i) {
if (buf_dblwr->buf_block_arr[i] == bpage) {
buf_dblwr->s_reserved--;
buf_dblwr->buf_block_arr[i] = NULL;
buf_dblwr->in_use[i] = false;
break;
}
}
/* The block we are looking for must exist as a
reserved block. */
ut_a(i < size);
}
os_event_set(buf_dblwr->s_event);
mutex_exit(&buf_dblwr->mutex);
break;
case BUF_FLUSH_N_TYPES:
ut_error;
}
}
#ifdef UNIV_DEBUG
/** Check the LSN values on the page.
@param[in] page page to check
@param[in] s tablespace */
static void buf_dblwr_check_page_lsn(const page_t* page, const fil_space_t& s)
{
/* Ignore page compressed or encrypted pages */
if (s.is_compressed() || buf_page_get_key_version(page, s.flags))
return;
const byte* lsn_start= FIL_PAGE_LSN + 4 + page;
const byte* lsn_end= page +
srv_page_size - (s.full_crc32()
? FIL_PAGE_FCRC32_END_LSN
: FIL_PAGE_END_LSN_OLD_CHKSUM - 4);
static_assert(FIL_PAGE_FCRC32_END_LSN % 4 == 0, "alignment");
static_assert(FIL_PAGE_LSN % 4 == 0, "alignment");
ut_ad(!memcmp_aligned<4>(lsn_start, lsn_end, 4));
}
static void buf_dblwr_check_page_lsn(const buf_page_t& b, const byte* page)
{
if (fil_space_t* space = fil_space_acquire_for_io(b.id.space())) {
buf_dblwr_check_page_lsn(page, *space);
space->release_for_io();
}
}
#endif /* UNIV_DEBUG */
/********************************************************************//**
Asserts when a corrupt block is find during writing out data to the
disk. */
static
void
buf_dblwr_assert_on_corrupt_block(
/*==============================*/
const buf_block_t* block) /*!< in: block to check */
{
buf_page_print(block->frame);
ib::fatal() << "Apparent corruption of an index page "
<< block->page.id
<< " to be written to data file. We intentionally crash"
" the server to prevent corrupt data from ending up in"
" data files.";
}
/********************************************************************//**
Check the LSN values on the page with which this block is associated.
Also validate the page if the option is set. */
static
void
buf_dblwr_check_block(
/*==================*/
const buf_block_t* block) /*!< in: block to check */
{
ut_ad(buf_block_get_state(block) == BUF_BLOCK_FILE_PAGE);
if (block->skip_flush_check) {
return;
}
switch (fil_page_get_type(block->frame)) {
case FIL_PAGE_INDEX:
case FIL_PAGE_TYPE_INSTANT:
case FIL_PAGE_RTREE:
if (page_is_comp(block->frame)) {
if (page_simple_validate_new(block->frame)) {
return;
}
} else if (page_simple_validate_old(block->frame)) {
return;
}
/* While it is possible that this is not an index page
but just happens to have wrongly set FIL_PAGE_TYPE,
such pages should never be modified to without also
adjusting the page type during page allocation or
buf_flush_init_for_writing() or fil_block_reset_type(). */
break;
case FIL_PAGE_TYPE_FSP_HDR:
case FIL_PAGE_IBUF_BITMAP:
case FIL_PAGE_TYPE_UNKNOWN:
/* Do not complain again, we already reset this field. */
case FIL_PAGE_UNDO_LOG:
case FIL_PAGE_INODE:
case FIL_PAGE_IBUF_FREE_LIST:
case FIL_PAGE_TYPE_SYS:
case FIL_PAGE_TYPE_TRX_SYS:
case FIL_PAGE_TYPE_XDES:
case FIL_PAGE_TYPE_BLOB:
case FIL_PAGE_TYPE_ZBLOB:
case FIL_PAGE_TYPE_ZBLOB2:
/* TODO: validate also non-index pages */
return;
case FIL_PAGE_TYPE_ALLOCATED:
/* empty pages should never be flushed */
return;
}
buf_dblwr_assert_on_corrupt_block(block);
}
/********************************************************************//**
Writes a page that has already been written to the doublewrite buffer
to the datafile. It is the job of the caller to sync the datafile. */
static
void
buf_dblwr_write_block_to_datafile(
/*==============================*/
const buf_page_t* bpage, /*!< in: page to write */
bool sync) /*!< in: true if sync IO
is requested */
{
ut_a(buf_page_in_file(bpage));
ulint type = IORequest::WRITE;
IORequest request(type, const_cast<buf_page_t*>(bpage));
/* We request frame here to get correct buffer in case of
encryption and/or page compression */
void * frame = buf_page_get_frame(bpage);
if (bpage->zip.data != NULL) {
ut_ad(bpage->zip_size());
fil_io(request, sync, bpage->id, bpage->zip_size(), 0,
bpage->zip_size(),
(void*) frame,
(void*) bpage);
} else {
ut_ad(!bpage->zip_size());
/* Our IO API is common for both reads and writes and is
therefore geared towards a non-const parameter. */
buf_block_t* block = reinterpret_cast<buf_block_t*>(
const_cast<buf_page_t*>(bpage));
ut_a(buf_block_get_state(block) == BUF_BLOCK_FILE_PAGE);
ut_d(buf_dblwr_check_page_lsn(block->page, block->frame));
fil_io(request,
sync, bpage->id, bpage->zip_size(), 0, bpage->real_size,
frame, block);
}
}
/********************************************************************//**
Flushes possible buffered writes from the doublewrite memory buffer to disk.
It is very important to call this function after a batch of writes has been posted,
and also when we may have to wait for a page latch! Otherwise a deadlock
of threads can occur. */
void
buf_dblwr_flush_buffered_writes()
{
byte* write_buf;
ulint first_free;
ulint len;
if (!srv_use_doublewrite_buf || buf_dblwr == NULL) {
/* Sync the writes to the disk. */
buf_dblwr_sync_datafiles();
/* Now we flush the data to disk (for example, with fsync) */
fil_flush_file_spaces(FIL_TYPE_TABLESPACE);
return;
}
ut_ad(!srv_read_only_mode);
try_again:
mutex_enter(&buf_dblwr->mutex);
/* Write first to doublewrite buffer blocks. We use synchronous
aio and thus know that file write has been completed when the
control returns. */
if (buf_dblwr->first_free == 0) {
mutex_exit(&buf_dblwr->mutex);
return;
}
if (buf_dblwr->batch_running) {
/* Another thread is running the batch right now. Wait
for it to finish. */
int64_t sig_count = os_event_reset(buf_dblwr->b_event);
mutex_exit(&buf_dblwr->mutex);
os_event_wait_low(buf_dblwr->b_event, sig_count);
goto try_again;
}
ut_ad(buf_dblwr->first_free == buf_dblwr->b_reserved);
/* Disallow anyone else to post to doublewrite buffer or to
start another batch of flushing. */
buf_dblwr->batch_running = true;
first_free = buf_dblwr->first_free;
/* Now safe to release the mutex. Note that though no other
thread is allowed to post to the doublewrite batch flushing
but any threads working on single page flushes are allowed
to proceed. */
mutex_exit(&buf_dblwr->mutex);
write_buf = buf_dblwr->write_buf;
for (ulint len2 = 0, i = 0;
i < buf_dblwr->first_free;
len2 += srv_page_size, i++) {
const buf_block_t* block;
block = (buf_block_t*) buf_dblwr->buf_block_arr[i];
if (buf_block_get_state(block) != BUF_BLOCK_FILE_PAGE
|| block->page.zip.data) {
/* No simple validate for compressed
pages exists. */
continue;
}
/* Check that the actual page in the buffer pool is
not corrupt and the LSN values are sane. */
buf_dblwr_check_block(block);
ut_d(buf_dblwr_check_page_lsn(block->page, write_buf + len2));
}
/* Write out the first block of the doublewrite buffer */
len = std::min<ulint>(TRX_SYS_DOUBLEWRITE_BLOCK_SIZE,
buf_dblwr->first_free) << srv_page_size_shift;
fil_io(IORequestWrite, true,
page_id_t(TRX_SYS_SPACE, buf_dblwr->block1), 0,
0, len, (void*) write_buf, NULL);
if (buf_dblwr->first_free <= TRX_SYS_DOUBLEWRITE_BLOCK_SIZE) {
/* No unwritten pages in the second block. */
goto flush;
}
/* Write out the second block of the doublewrite buffer. */
len = (buf_dblwr->first_free - TRX_SYS_DOUBLEWRITE_BLOCK_SIZE)
<< srv_page_size_shift;
write_buf = buf_dblwr->write_buf
+ (TRX_SYS_DOUBLEWRITE_BLOCK_SIZE << srv_page_size_shift);
fil_io(IORequestWrite, true,
page_id_t(TRX_SYS_SPACE, buf_dblwr->block2), 0,
0, len, (void*) write_buf, NULL);
flush:
/* increment the doublewrite flushed pages counter */
srv_stats.dblwr_pages_written.add(buf_dblwr->first_free);
srv_stats.dblwr_writes.inc();
/* Now flush the doublewrite buffer data to disk */
fil_flush(TRX_SYS_SPACE);
/* We know that the writes have been flushed to disk now
and in recovery we will find them in the doublewrite buffer
blocks. Next do the writes to the intended positions. */
/* Up to this point first_free and buf_dblwr->first_free are
same because we have set the buf_dblwr->batch_running flag
disallowing any other thread to post any request but we
can't safely access buf_dblwr->first_free in the loop below.
This is so because it is possible that after we are done with
the last iteration and before we terminate the loop, the batch
gets finished in the IO helper thread and another thread posts
a new batch setting buf_dblwr->first_free to a higher value.
If this happens and we are using buf_dblwr->first_free in the
loop termination condition then we'll end up dispatching
the same block twice from two different threads. */
ut_ad(first_free == buf_dblwr->first_free);
for (ulint i = 0; i < first_free; i++) {
buf_dblwr_write_block_to_datafile(
buf_dblwr->buf_block_arr[i], false);
}
}
/********************************************************************//**
Posts a buffer page for writing. If the doublewrite memory buffer is
full, calls buf_dblwr_flush_buffered_writes and waits for for free
space to appear. */
void
buf_dblwr_add_to_batch(
/*====================*/
buf_page_t* bpage) /*!< in: buffer block to write */
{
ut_a(buf_page_in_file(bpage));
try_again:
mutex_enter(&buf_dblwr->mutex);
ut_a(buf_dblwr->first_free <= srv_doublewrite_batch_size);
if (buf_dblwr->batch_running) {
/* This not nearly as bad as it looks. There is only
page_cleaner thread which does background flushing
in batches therefore it is unlikely to be a contention
point. The only exception is when a user thread is
forced to do a flush batch because of a sync
checkpoint. */
int64_t sig_count = os_event_reset(buf_dblwr->b_event);
mutex_exit(&buf_dblwr->mutex);
os_event_wait_low(buf_dblwr->b_event, sig_count);
goto try_again;
}
if (buf_dblwr->first_free == srv_doublewrite_batch_size) {
mutex_exit(&(buf_dblwr->mutex));
buf_dblwr_flush_buffered_writes();
goto try_again;
}
byte* p = buf_dblwr->write_buf
+ srv_page_size * buf_dblwr->first_free;
/* We request frame here to get correct buffer in case of
encryption and/or page compression */
void * frame = buf_page_get_frame(bpage);
if (auto zip_size = bpage->zip_size()) {
UNIV_MEM_ASSERT_RW(bpage->zip.data, zip_size);
/* Copy the compressed page and clear the rest. */
memcpy(p, frame, zip_size);
memset(p + zip_size, 0x0, srv_page_size - zip_size);
} else {
ut_a(buf_page_get_state(bpage) == BUF_BLOCK_FILE_PAGE);
UNIV_MEM_ASSERT_RW(frame, srv_page_size);
memcpy_aligned<OS_FILE_LOG_BLOCK_SIZE>(p, frame,
srv_page_size);
}
buf_dblwr->buf_block_arr[buf_dblwr->first_free] = bpage;
buf_dblwr->first_free++;
buf_dblwr->b_reserved++;
ut_ad(!buf_dblwr->batch_running);
ut_ad(buf_dblwr->first_free == buf_dblwr->b_reserved);
ut_ad(buf_dblwr->b_reserved <= srv_doublewrite_batch_size);
if (buf_dblwr->first_free == srv_doublewrite_batch_size) {
mutex_exit(&(buf_dblwr->mutex));
buf_dblwr_flush_buffered_writes();
return;
}
mutex_exit(&(buf_dblwr->mutex));
}
/********************************************************************//**
Writes a page to the doublewrite buffer on disk, sync it, then write
the page to the datafile and sync the datafile. This function is used
for single page flushes. If all the buffers allocated for single page
flushes in the doublewrite buffer are in use we wait here for one to
become free. We are guaranteed that a slot will become free because any
thread that is using a slot must also release the slot before leaving
this function. */
void
buf_dblwr_write_single_page(
/*========================*/
buf_page_t* bpage, /*!< in: buffer block to write */
bool sync) /*!< in: true if sync IO requested */
{
ulint n_slots;
ulint size;
ulint offset;
ulint i;
ut_a(buf_page_in_file(bpage));
ut_a(srv_use_doublewrite_buf);
ut_a(buf_dblwr != NULL);
/* total number of slots available for single page flushes
starts from srv_doublewrite_batch_size to the end of the
buffer. */
size = TRX_SYS_DOUBLEWRITE_BLOCKS * TRX_SYS_DOUBLEWRITE_BLOCK_SIZE;
ut_a(size > srv_doublewrite_batch_size);
n_slots = size - srv_doublewrite_batch_size;
if (buf_page_get_state(bpage) == BUF_BLOCK_FILE_PAGE) {
/* Check that the actual page in the buffer pool is
not corrupt and the LSN values are sane. */
buf_dblwr_check_block((buf_block_t*) bpage);
/* Check that the page as written to the doublewrite
buffer has sane LSN values. */
if (!bpage->zip.data) {
ut_d(buf_dblwr_check_page_lsn(
*bpage, ((buf_block_t*) bpage)->frame));
}
}
retry:
mutex_enter(&buf_dblwr->mutex);
if (buf_dblwr->s_reserved == n_slots) {
/* All slots are reserved. */
int64_t sig_count = os_event_reset(buf_dblwr->s_event);
mutex_exit(&buf_dblwr->mutex);
os_event_wait_low(buf_dblwr->s_event, sig_count);
goto retry;
}
for (i = srv_doublewrite_batch_size; i < size; ++i) {
if (!buf_dblwr->in_use[i]) {
break;
}
}
/* We are guaranteed to find a slot. */
ut_a(i < size);
buf_dblwr->in_use[i] = true;
buf_dblwr->s_reserved++;
buf_dblwr->buf_block_arr[i] = bpage;
/* increment the doublewrite flushed pages counter */
srv_stats.dblwr_pages_written.inc();
srv_stats.dblwr_writes.inc();
mutex_exit(&buf_dblwr->mutex);
/* Lets see if we are going to write in the first or second
block of the doublewrite buffer. */
if (i < TRX_SYS_DOUBLEWRITE_BLOCK_SIZE) {
offset = buf_dblwr->block1 + i;
} else {
offset = buf_dblwr->block2 + i
- TRX_SYS_DOUBLEWRITE_BLOCK_SIZE;
}
/* We deal with compressed and uncompressed pages a little
differently here. In case of uncompressed pages we can
directly write the block to the allocated slot in the
doublewrite buffer in the system tablespace and then after
syncing the system table space we can proceed to write the page
in the datafile.
In case of compressed page we first do a memcpy of the block
to the in-memory buffer of doublewrite before proceeding to
write it. This is so because we want to pad the remaining
bytes in the doublewrite page with zeros. */
/* We request frame here to get correct buffer in case of
encryption and/or page compression */
void * frame = buf_page_get_frame(bpage);
if (auto zip_size = bpage->zip_size()) {
memcpy(buf_dblwr->write_buf + srv_page_size * i,
frame, zip_size);
memset(buf_dblwr->write_buf + srv_page_size * i
+ zip_size, 0x0,
srv_page_size - zip_size);
fil_io(IORequestWrite,
true,
page_id_t(TRX_SYS_SPACE, offset),
0,
0,
srv_page_size,
(void *)(buf_dblwr->write_buf + srv_page_size * i),
NULL);
} else {
/* It is a regular page. Write it directly to the
doublewrite buffer */
fil_io(IORequestWrite,
true,
page_id_t(TRX_SYS_SPACE, offset),
0,
0,
srv_page_size,
(void*) frame,
NULL);
}
/* Now flush the doublewrite buffer data to disk */
fil_flush(TRX_SYS_SPACE);
/* We know that the write has been flushed to disk now
and during recovery we will find it in the doublewrite buffer
blocks. Next do the write to the intended position. */
buf_dblwr_write_block_to_datafile(bpage, sync);
}