mirror of
https://github.com/MariaDB/server.git
synced 2025-01-16 12:02:42 +01:00
57c526ffb8
This patch also fixes some bugs detected by valgrind after this patch: - Not enough copy_func elements was allocated by Create_tmp_table() which causes an memory overwrite in Create_tmp_table::add_fields() I added an ASSERT() to be able to detect this also without valgrind. The bug was that TMP_TABLE_PARAM::copy_fields was not correctly set when calling create_tmp_table(). - Aria::empty_bits is not allocated if there is no varchar/char/blob fields in the table. Fixed code to take this into account. This cannot cause any issues as this is just a memory access into other Aria memory and the content of the memory would not be used. - Aria::last_key_buff was not allocated big enough. This may have caused issues with rtrees and ma_extra(HA_EXTRA_REMEMBER_POS) as they would use the same memory area. - Aria and MyISAM didn't take extended key parts into account, which caused problems when copying rec_per_key from engine to sql level. - Mark asan builds with 'asan' in version strihng to detect these in not_valgrind_build.inc. This is needed to not have main.sp-no-valgrind fail with asan.
2823 lines
90 KiB
C++
2823 lines
90 KiB
C++
/*
|
|
Copyright (c) 2000, 2018, Oracle and/or its affiliates.
|
|
Copyright (c) 2009, 2022, MariaDB Corporation.
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; version 2 of the License.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1335 USA */
|
|
|
|
|
|
#ifdef USE_PRAGMA_IMPLEMENTATION
|
|
#pragma implementation // gcc: Class implementation
|
|
#endif
|
|
|
|
#define MYSQL_SERVER 1
|
|
#include <my_global.h>
|
|
#include "sql_plugin.h"
|
|
#include "myisamdef.h"
|
|
#include "sql_priv.h"
|
|
#include "key.h" // key_copy
|
|
#include <m_ctype.h>
|
|
#include <my_bit.h>
|
|
#include "ha_myisam.h"
|
|
#include "rt_index.h"
|
|
#include "sql_table.h" // tablename_to_filename
|
|
#include "sql_class.h" // THD
|
|
#include "debug_sync.h"
|
|
#include "sql_debug.h"
|
|
|
|
ulonglong myisam_recover_options;
|
|
static ulong opt_myisam_block_size;
|
|
|
|
/* bits in myisam_recover_options */
|
|
const char *myisam_recover_names[] =
|
|
{ "DEFAULT", "BACKUP", "FORCE", "QUICK", "BACKUP_ALL", "OFF", NullS};
|
|
TYPELIB myisam_recover_typelib= {array_elements(myisam_recover_names)-1,"",
|
|
myisam_recover_names, NULL};
|
|
|
|
const char *myisam_stats_method_names[] = {"NULLS_UNEQUAL", "NULLS_EQUAL",
|
|
"NULLS_IGNORED", NullS};
|
|
TYPELIB myisam_stats_method_typelib= {
|
|
array_elements(myisam_stats_method_names) - 1, "",
|
|
myisam_stats_method_names, NULL};
|
|
|
|
static MYSQL_SYSVAR_ULONG(block_size, opt_myisam_block_size,
|
|
PLUGIN_VAR_READONLY | PLUGIN_VAR_RQCMDARG,
|
|
"Block size to be used for MyISAM index pages", NULL, NULL,
|
|
MI_KEY_BLOCK_LENGTH, MI_MIN_KEY_BLOCK_LENGTH, MI_MAX_KEY_BLOCK_LENGTH,
|
|
MI_MIN_KEY_BLOCK_LENGTH);
|
|
|
|
static MYSQL_SYSVAR_ULONG(data_pointer_size, myisam_data_pointer_size,
|
|
PLUGIN_VAR_RQCMDARG, "Default pointer size to be used for MyISAM tables",
|
|
NULL, NULL, 6, 2, 7, 1);
|
|
|
|
#define MB (1024*1024)
|
|
static MYSQL_SYSVAR_ULONGLONG(max_sort_file_size, myisam_max_temp_length,
|
|
PLUGIN_VAR_RQCMDARG, "Don't use the fast sort index method to created "
|
|
"index if the temporary file would get bigger than this", NULL, NULL,
|
|
LONG_MAX/MB*MB, 0, MAX_FILE_SIZE, MB);
|
|
|
|
static MYSQL_SYSVAR_SET(recover_options, myisam_recover_options,
|
|
PLUGIN_VAR_OPCMDARG|PLUGIN_VAR_READONLY,
|
|
"Specifies how corrupted tables should be automatically repaired",
|
|
NULL, NULL, HA_RECOVER_BACKUP|HA_RECOVER_QUICK, &myisam_recover_typelib);
|
|
|
|
static MYSQL_THDVAR_ULONG(repair_threads, PLUGIN_VAR_RQCMDARG,
|
|
"If larger than 1, when repairing a MyISAM table all indexes will be "
|
|
"created in parallel, with one thread per index. The value of 1 "
|
|
"disables parallel repair", NULL, NULL,
|
|
1, 1, ULONG_MAX, 1);
|
|
|
|
static MYSQL_THDVAR_ULONGLONG(sort_buffer_size, PLUGIN_VAR_RQCMDARG,
|
|
"The buffer that is allocated when sorting the index when doing "
|
|
"a REPAIR or when creating indexes with CREATE INDEX or ALTER TABLE", NULL, NULL,
|
|
SORT_BUFFER_INIT, MIN_SORT_BUFFER, SIZE_T_MAX, 1);
|
|
|
|
static MYSQL_SYSVAR_BOOL(use_mmap, opt_myisam_use_mmap, PLUGIN_VAR_NOCMDARG,
|
|
"Use memory mapping for reading and writing MyISAM tables", NULL, NULL, FALSE);
|
|
|
|
static MYSQL_SYSVAR_ULONGLONG(mmap_size, myisam_mmap_size,
|
|
PLUGIN_VAR_RQCMDARG|PLUGIN_VAR_READONLY, "Restricts the total memory "
|
|
"used for memory mapping of MySQL tables", NULL, NULL,
|
|
SIZE_T_MAX, MEMMAP_EXTRA_MARGIN, SIZE_T_MAX, 1);
|
|
|
|
static MYSQL_THDVAR_ENUM(stats_method, PLUGIN_VAR_RQCMDARG,
|
|
"Specifies how MyISAM index statistics collection code should "
|
|
"treat NULLs. Possible values of name are NULLS_UNEQUAL (default "
|
|
"behavior for 4.1 and later), NULLS_EQUAL (emulate 4.0 behavior), "
|
|
"and NULLS_IGNORED", NULL, NULL,
|
|
MI_STATS_METHOD_NULLS_NOT_EQUAL, &myisam_stats_method_typelib);
|
|
|
|
const char *MI_CHECK_INFO= "info";
|
|
const char *MI_CHECK_WARNING= "warning";
|
|
const char *MI_CHECK_ERROR= "error";
|
|
|
|
#ifndef DBUG_OFF
|
|
/**
|
|
Causes the thread to wait in a spin lock for a query kill signal.
|
|
This function is used by the test frame work to identify race conditions.
|
|
|
|
The signal is caught and ignored and the thread is not killed.
|
|
*/
|
|
|
|
static void debug_wait_for_kill(const char *info)
|
|
{
|
|
DBUG_ENTER("debug_wait_for_kill");
|
|
const char *prev_info;
|
|
THD *thd;
|
|
thd= current_thd;
|
|
prev_info= thd_proc_info(thd, info);
|
|
while(!thd->killed)
|
|
my_sleep(1000);
|
|
DBUG_PRINT("info", ("Exit debug_wait_for_kill"));
|
|
thd_proc_info(thd, prev_info);
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
|
|
class Debug_key_myisam: public Debug_key
|
|
{
|
|
public:
|
|
Debug_key_myisam() { }
|
|
|
|
static void print_keys_myisam(THD *thd, const char *where,
|
|
const TABLE *table,
|
|
const MI_KEYDEF *keydef, uint count)
|
|
{
|
|
for (uint i= 0; i < count; i++)
|
|
{
|
|
Debug_key_myisam tmp;
|
|
if (!tmp.append(where) &&
|
|
!tmp.append_key(table->s->key_info[i].name, keydef[i].flag))
|
|
tmp.print(thd);
|
|
print_keysegs(thd, keydef[i].seg, keydef[i].keysegs);
|
|
}
|
|
}
|
|
};
|
|
|
|
#endif
|
|
|
|
/*****************************************************************************
|
|
** MyISAM tables
|
|
*****************************************************************************/
|
|
|
|
static handler *myisam_create_handler(handlerton *hton,
|
|
TABLE_SHARE *table,
|
|
MEM_ROOT *mem_root)
|
|
{
|
|
return new (mem_root) ha_myisam(hton, table);
|
|
}
|
|
|
|
|
|
static void mi_check_print(HA_CHECK *param, const char* msg_type,
|
|
const char *msgbuf)
|
|
{
|
|
if (msg_type == MI_CHECK_INFO)
|
|
sql_print_information("%s.%s: %s", param->db_name, param->table_name,
|
|
msgbuf);
|
|
else if (msg_type == MI_CHECK_WARNING)
|
|
sql_print_warning("%s.%s: %s", param->db_name, param->table_name,
|
|
msgbuf);
|
|
else
|
|
sql_print_error("%s.%s: %s", param->db_name, param->table_name, msgbuf);
|
|
}
|
|
|
|
// collect errors printed by mi_check routines
|
|
|
|
static void mi_check_print_msg(HA_CHECK *param, const char* msg_type,
|
|
const char *fmt, va_list args)
|
|
{
|
|
THD* thd = (THD*)param->thd;
|
|
Protocol *protocol= thd->protocol;
|
|
size_t length, msg_length;
|
|
char msgbuf[MYSQL_ERRMSG_SIZE];
|
|
char name[NAME_LEN*2+2];
|
|
|
|
if (param->testflag & T_SUPPRESS_ERR_HANDLING)
|
|
return;
|
|
|
|
msg_length= my_vsnprintf(msgbuf, sizeof(msgbuf), fmt, args);
|
|
msgbuf[sizeof(msgbuf) - 1] = 0; // healthy paranoia
|
|
|
|
DBUG_PRINT(msg_type,("message: %s",msgbuf));
|
|
|
|
if (!thd->vio_ok())
|
|
{
|
|
mi_check_print(param, msg_type, msgbuf);
|
|
return;
|
|
}
|
|
|
|
if (param->testflag & (T_CREATE_MISSING_KEYS | T_SAFE_REPAIR |
|
|
T_AUTO_REPAIR))
|
|
{
|
|
myf flag= 0;
|
|
if (msg_type == MI_CHECK_INFO)
|
|
flag= ME_NOTE;
|
|
else if (msg_type == MI_CHECK_WARNING)
|
|
flag= ME_WARNING;
|
|
my_message(ER_NOT_KEYFILE, msgbuf, MYF(flag));
|
|
if (thd->variables.log_warnings > 2 && ! thd->log_all_errors)
|
|
mi_check_print(param, msg_type, msgbuf);
|
|
return;
|
|
}
|
|
length=(uint) (strxmov(name, param->db_name,".",param->table_name,NullS) -
|
|
name);
|
|
/*
|
|
TODO: switch from protocol to push_warning here. The main reason we didn't
|
|
it yet is parallel repair, which threads have no THD object accessible via
|
|
current_thd.
|
|
|
|
Also we likely need to lock mutex here (in both cases with protocol and
|
|
push_warning).
|
|
*/
|
|
if (param->need_print_msg_lock)
|
|
mysql_mutex_lock(¶m->print_msg_mutex);
|
|
|
|
protocol->prepare_for_resend();
|
|
protocol->store(name, length, system_charset_info);
|
|
protocol->store(param->op_name, system_charset_info);
|
|
protocol->store(msg_type, system_charset_info);
|
|
protocol->store(msgbuf, msg_length, system_charset_info);
|
|
if (protocol->write())
|
|
sql_print_error("Failed on my_net_write, writing to stderr instead: %s\n",
|
|
msgbuf);
|
|
else if (thd->variables.log_warnings > 2)
|
|
mi_check_print(param, msg_type, msgbuf);
|
|
|
|
if (param->need_print_msg_lock)
|
|
mysql_mutex_unlock(¶m->print_msg_mutex);
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
/*
|
|
Convert TABLE object to MyISAM key and column definition
|
|
|
|
SYNOPSIS
|
|
table2myisam()
|
|
table_arg in TABLE object.
|
|
keydef_out out MyISAM key definition.
|
|
recinfo_out out MyISAM column definition.
|
|
records_out out Number of fields.
|
|
|
|
DESCRIPTION
|
|
This function will allocate and initialize MyISAM key and column
|
|
definition for further use in mi_create or for a check for underlying
|
|
table conformance in merge engine.
|
|
|
|
The caller needs to free *recinfo_out after use. Since *recinfo_out
|
|
and *keydef_out are allocated with a my_multi_malloc, *keydef_out
|
|
is freed automatically when *recinfo_out is freed.
|
|
|
|
RETURN VALUE
|
|
0 OK
|
|
!0 error code
|
|
*/
|
|
|
|
int table2myisam(TABLE *table_arg, MI_KEYDEF **keydef_out,
|
|
MI_COLUMNDEF **recinfo_out, uint *records_out)
|
|
{
|
|
uint i, j, recpos, minpos, fieldpos, temp_length, length;
|
|
enum ha_base_keytype type= HA_KEYTYPE_BINARY;
|
|
uchar *record;
|
|
KEY *pos;
|
|
MI_KEYDEF *keydef;
|
|
MI_COLUMNDEF *recinfo, *recinfo_pos;
|
|
HA_KEYSEG *keyseg;
|
|
TABLE_SHARE *share= table_arg->s;
|
|
uint options= share->db_options_in_use;
|
|
DBUG_ENTER("table2myisam");
|
|
if (!(my_multi_malloc(PSI_INSTRUMENT_ME, MYF(MY_WME),
|
|
recinfo_out, (share->fields * 2 + 2) * sizeof(MI_COLUMNDEF),
|
|
keydef_out, share->keys * sizeof(MI_KEYDEF),
|
|
&keyseg,
|
|
(share->key_parts + share->keys) * sizeof(HA_KEYSEG),
|
|
NullS)))
|
|
DBUG_RETURN(HA_ERR_OUT_OF_MEM); /* purecov: inspected */
|
|
keydef= *keydef_out;
|
|
recinfo= *recinfo_out;
|
|
pos= table_arg->key_info;
|
|
for (i= 0; i < share->keys; i++, pos++)
|
|
{
|
|
keydef[i].flag= ((uint16) pos->flags & (HA_NOSAME | HA_FULLTEXT | HA_SPATIAL));
|
|
keydef[i].key_alg= pos->algorithm == HA_KEY_ALG_UNDEF ?
|
|
(pos->flags & HA_SPATIAL ? HA_KEY_ALG_RTREE : HA_KEY_ALG_BTREE) :
|
|
pos->algorithm;
|
|
keydef[i].block_length= pos->block_size;
|
|
keydef[i].seg= keyseg;
|
|
keydef[i].keysegs= pos->user_defined_key_parts;
|
|
for (j= 0; j < pos->user_defined_key_parts; j++)
|
|
{
|
|
Field *field= pos->key_part[j].field;
|
|
type= field->key_type();
|
|
keydef[i].seg[j].flag= pos->key_part[j].key_part_flag;
|
|
|
|
if (options & HA_OPTION_PACK_KEYS ||
|
|
(pos->flags & (HA_PACK_KEY | HA_BINARY_PACK_KEY |
|
|
HA_SPACE_PACK_USED)))
|
|
{
|
|
if (pos->key_part[j].length > 8 &&
|
|
(type == HA_KEYTYPE_TEXT ||
|
|
type == HA_KEYTYPE_NUM ||
|
|
(type == HA_KEYTYPE_BINARY && !field->zero_pack())))
|
|
{
|
|
/* No blobs here */
|
|
if (j == 0)
|
|
keydef[i].flag|= HA_PACK_KEY;
|
|
if (!(field->flags & ZEROFILL_FLAG) &&
|
|
(field->type() == MYSQL_TYPE_STRING ||
|
|
field->type() == MYSQL_TYPE_VAR_STRING ||
|
|
((int) (pos->key_part[j].length - field->decimals())) >= 4))
|
|
keydef[i].seg[j].flag|= HA_SPACE_PACK;
|
|
}
|
|
else if (j == 0 && (!(pos->flags & HA_NOSAME) || pos->key_length > 16))
|
|
keydef[i].flag|= HA_BINARY_PACK_KEY;
|
|
}
|
|
keydef[i].seg[j].type= (int) type;
|
|
keydef[i].seg[j].start= pos->key_part[j].offset;
|
|
keydef[i].seg[j].length= pos->key_part[j].length;
|
|
keydef[i].seg[j].bit_start= keydef[i].seg[j].bit_length= 0;
|
|
keydef[i].seg[j].bit_pos= 0;
|
|
keydef[i].seg[j].language= field->charset_for_protocol()->number;
|
|
|
|
if (field->null_ptr)
|
|
{
|
|
keydef[i].seg[j].null_bit= field->null_bit;
|
|
keydef[i].seg[j].null_pos= (uint) (field->null_ptr-
|
|
(uchar*) table_arg->record[0]);
|
|
}
|
|
else
|
|
{
|
|
keydef[i].seg[j].null_bit= 0;
|
|
keydef[i].seg[j].null_pos= 0;
|
|
}
|
|
if (field->type() == MYSQL_TYPE_BLOB ||
|
|
field->type() == MYSQL_TYPE_GEOMETRY)
|
|
{
|
|
keydef[i].seg[j].flag|= HA_BLOB_PART;
|
|
/* save number of bytes used to pack length */
|
|
keydef[i].seg[j].bit_start= (uint) (field->pack_length() -
|
|
portable_sizeof_char_ptr);
|
|
}
|
|
else if (field->type() == MYSQL_TYPE_BIT)
|
|
{
|
|
keydef[i].seg[j].bit_length= ((Field_bit *) field)->bit_len;
|
|
keydef[i].seg[j].bit_start= ((Field_bit *) field)->bit_ofs;
|
|
keydef[i].seg[j].bit_pos= (uint) (((Field_bit *) field)->bit_ptr -
|
|
(uchar*) table_arg->record[0]);
|
|
}
|
|
}
|
|
keyseg+= pos->user_defined_key_parts;
|
|
}
|
|
if (table_arg->found_next_number_field)
|
|
keydef[share->next_number_index].flag|= HA_AUTO_KEY;
|
|
record= table_arg->record[0];
|
|
recpos= 0;
|
|
recinfo_pos= recinfo;
|
|
while (recpos < (uint) share->stored_rec_length)
|
|
{
|
|
Field **field, *found= 0;
|
|
minpos= share->reclength;
|
|
length= 0;
|
|
|
|
for (field= table_arg->field; *field; field++)
|
|
{
|
|
if ((fieldpos= (*field)->offset(record)) >= recpos &&
|
|
fieldpos <= minpos)
|
|
{
|
|
/* skip null fields */
|
|
if (!(temp_length= (*field)->pack_length_in_rec()))
|
|
continue; /* Skip null-fields */
|
|
if (! found || fieldpos < minpos ||
|
|
(fieldpos == minpos && temp_length < length))
|
|
{
|
|
minpos= fieldpos;
|
|
found= *field;
|
|
length= temp_length;
|
|
}
|
|
}
|
|
}
|
|
DBUG_PRINT("loop", ("found: %p recpos: %d minpos: %d length: %d",
|
|
found, recpos, minpos, length));
|
|
if (recpos != minpos)
|
|
{
|
|
/* reserve space for null bits */
|
|
bzero((char*) recinfo_pos, sizeof(*recinfo_pos));
|
|
recinfo_pos->type= FIELD_NORMAL;
|
|
recinfo_pos++->length= (uint16) (minpos - recpos);
|
|
}
|
|
if (!found)
|
|
break;
|
|
|
|
if (found->flags & BLOB_FLAG)
|
|
recinfo_pos->type= FIELD_BLOB;
|
|
else if (found->real_type() == MYSQL_TYPE_TIMESTAMP)
|
|
{
|
|
/* pre-MySQL-5.6.4 TIMESTAMP, or MariaDB-5.3+ TIMESTAMP */
|
|
recinfo_pos->type= FIELD_NORMAL;
|
|
}
|
|
else if (found->type() == MYSQL_TYPE_VARCHAR)
|
|
recinfo_pos->type= FIELD_VARCHAR;
|
|
else if (!(options & HA_OPTION_PACK_RECORD))
|
|
recinfo_pos->type= FIELD_NORMAL;
|
|
else if (found->real_type() == MYSQL_TYPE_TIMESTAMP2)
|
|
{
|
|
/*
|
|
MySQL-5.6.4+ erroneously marks Field_timestampf as FIELD_SKIP_PRESPACE,
|
|
but only if HA_OPTION_PACK_RECORD is set.
|
|
*/
|
|
recinfo_pos->type= FIELD_SKIP_PRESPACE;
|
|
}
|
|
else if (found->zero_pack())
|
|
recinfo_pos->type= FIELD_SKIP_ZERO;
|
|
else
|
|
recinfo_pos->type= ((length <= 3 ||
|
|
(found->flags & ZEROFILL_FLAG)) ?
|
|
FIELD_NORMAL :
|
|
found->type() == MYSQL_TYPE_STRING ||
|
|
found->type() == MYSQL_TYPE_VAR_STRING ?
|
|
FIELD_SKIP_ENDSPACE :
|
|
FIELD_SKIP_PRESPACE);
|
|
if (found->null_ptr)
|
|
{
|
|
recinfo_pos->null_bit= found->null_bit;
|
|
recinfo_pos->null_pos= (uint) (found->null_ptr -
|
|
(uchar*) table_arg->record[0]);
|
|
}
|
|
else
|
|
{
|
|
recinfo_pos->null_bit= 0;
|
|
recinfo_pos->null_pos= 0;
|
|
}
|
|
(recinfo_pos++)->length= (uint16) length;
|
|
recpos= minpos + length;
|
|
DBUG_PRINT("loop", ("length: %d type: %d",
|
|
recinfo_pos[-1].length,recinfo_pos[-1].type));
|
|
}
|
|
*records_out= (uint) (recinfo_pos - recinfo);
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
|
|
/*
|
|
Check for underlying table conformance
|
|
|
|
SYNOPSIS
|
|
myisam_check_definition()
|
|
t1_keyinfo in First table key definition
|
|
t1_recinfo in First table record definition
|
|
t1_keys in Number of keys in first table
|
|
t1_recs in Number of records in first table
|
|
t2_keyinfo in Second table key definition
|
|
t2_recinfo in Second table record definition
|
|
t2_keys in Number of keys in second table
|
|
t2_recs in Number of records in second table
|
|
strict in Strict check switch
|
|
table in handle to the table object
|
|
|
|
DESCRIPTION
|
|
This function compares two MyISAM definitions. By intention it was done
|
|
to compare merge table definition against underlying table definition.
|
|
It may also be used to compare dot-frm and MYI definitions of MyISAM
|
|
table as well to compare different MyISAM table definitions.
|
|
|
|
For merge table it is not required that number of keys in merge table
|
|
must exactly match number of keys in underlying table. When calling this
|
|
function for underlying table conformance check, 'strict' flag must be
|
|
set to false, and converted merge definition must be passed as t1_*.
|
|
|
|
Otherwise 'strict' flag must be set to 1 and it is not required to pass
|
|
converted dot-frm definition as t1_*.
|
|
|
|
For compatibility reasons we relax some checks, specifically:
|
|
- 4.0 (and earlier versions) always set key_alg to 0.
|
|
- 4.0 (and earlier versions) have the same language for all keysegs.
|
|
|
|
RETURN VALUE
|
|
0 - Equal definitions.
|
|
1 - Different definitions.
|
|
|
|
TODO
|
|
- compare FULLTEXT keys;
|
|
- compare SPATIAL keys;
|
|
- compare FIELD_SKIP_ZERO which is converted to FIELD_NORMAL correctly
|
|
(should be corretly detected in table2myisam).
|
|
*/
|
|
|
|
int check_definition(MI_KEYDEF *t1_keyinfo, MI_COLUMNDEF *t1_recinfo,
|
|
uint t1_keys, uint t1_recs,
|
|
MI_KEYDEF *t2_keyinfo, MI_COLUMNDEF *t2_recinfo,
|
|
uint t2_keys, uint t2_recs, bool strict, TABLE *table_arg)
|
|
{
|
|
uint i, j;
|
|
DBUG_ENTER("check_definition");
|
|
my_bool mysql_40_compat= table_arg && table_arg->s->frm_version < FRM_VER_TRUE_VARCHAR;
|
|
if ((strict ? t1_keys != t2_keys : t1_keys > t2_keys))
|
|
{
|
|
DBUG_PRINT("error", ("Number of keys differs: t1_keys=%u, t2_keys=%u",
|
|
t1_keys, t2_keys));
|
|
DBUG_RETURN(1);
|
|
}
|
|
if (t1_recs != t2_recs)
|
|
{
|
|
DBUG_PRINT("error", ("Number of recs differs: t1_recs=%u, t2_recs=%u",
|
|
t1_recs, t2_recs));
|
|
DBUG_RETURN(1);
|
|
}
|
|
for (i= 0; i < t1_keys; i++)
|
|
{
|
|
HA_KEYSEG *t1_keysegs= t1_keyinfo[i].seg;
|
|
HA_KEYSEG *t2_keysegs= t2_keyinfo[i].seg;
|
|
if (t1_keyinfo[i].flag & HA_FULLTEXT && t2_keyinfo[i].flag & HA_FULLTEXT)
|
|
continue;
|
|
else if (t1_keyinfo[i].flag & HA_FULLTEXT ||
|
|
t2_keyinfo[i].flag & HA_FULLTEXT)
|
|
{
|
|
DBUG_PRINT("error", ("Key %d has different definition", i));
|
|
DBUG_PRINT("error", ("t1_fulltext= %d, t2_fulltext=%d",
|
|
MY_TEST(t1_keyinfo[i].flag & HA_FULLTEXT),
|
|
MY_TEST(t2_keyinfo[i].flag & HA_FULLTEXT)));
|
|
DBUG_RETURN(1);
|
|
}
|
|
if (t1_keyinfo[i].flag & HA_SPATIAL && t2_keyinfo[i].flag & HA_SPATIAL)
|
|
continue;
|
|
else if (t1_keyinfo[i].flag & HA_SPATIAL ||
|
|
t2_keyinfo[i].flag & HA_SPATIAL)
|
|
{
|
|
DBUG_PRINT("error", ("Key %d has different definition", i));
|
|
DBUG_PRINT("error", ("t1_spatial= %d, t2_spatial=%d",
|
|
MY_TEST(t1_keyinfo[i].flag & HA_SPATIAL),
|
|
MY_TEST(t2_keyinfo[i].flag & HA_SPATIAL)));
|
|
DBUG_RETURN(1);
|
|
}
|
|
if ((!mysql_40_compat &&
|
|
t1_keyinfo[i].key_alg != t2_keyinfo[i].key_alg) ||
|
|
t1_keyinfo[i].keysegs != t2_keyinfo[i].keysegs)
|
|
{
|
|
DBUG_PRINT("error", ("Key %d has different definition", i));
|
|
DBUG_PRINT("error", ("t1_keysegs=%d, t1_key_alg=%d",
|
|
t1_keyinfo[i].keysegs, t1_keyinfo[i].key_alg));
|
|
DBUG_PRINT("error", ("t2_keysegs=%d, t2_key_alg=%d",
|
|
t2_keyinfo[i].keysegs, t2_keyinfo[i].key_alg));
|
|
DBUG_RETURN(1);
|
|
}
|
|
for (j= t1_keyinfo[i].keysegs; j--;)
|
|
{
|
|
uint8 t1_keysegs_j__type= t1_keysegs[j].type;
|
|
|
|
/*
|
|
Table migration from 4.1 to 5.1. In 5.1 a *TEXT key part is
|
|
always HA_KEYTYPE_VARTEXT2. In 4.1 we had only the equivalent of
|
|
HA_KEYTYPE_VARTEXT1. Since we treat both the same on MyISAM
|
|
level, we can ignore a mismatch between these types.
|
|
*/
|
|
if ((t1_keysegs[j].flag & HA_BLOB_PART) &&
|
|
(t2_keysegs[j].flag & HA_BLOB_PART))
|
|
{
|
|
if ((t1_keysegs_j__type == HA_KEYTYPE_VARTEXT2) &&
|
|
(t2_keysegs[j].type == HA_KEYTYPE_VARTEXT1))
|
|
t1_keysegs_j__type= HA_KEYTYPE_VARTEXT1; /* purecov: tested */
|
|
else if ((t1_keysegs_j__type == HA_KEYTYPE_VARBINARY2) &&
|
|
(t2_keysegs[j].type == HA_KEYTYPE_VARBINARY1))
|
|
t1_keysegs_j__type= HA_KEYTYPE_VARBINARY1; /* purecov: inspected */
|
|
}
|
|
|
|
if ((!mysql_40_compat &&
|
|
t1_keysegs[j].language != t2_keysegs[j].language) ||
|
|
t1_keysegs_j__type != t2_keysegs[j].type ||
|
|
t1_keysegs[j].null_bit != t2_keysegs[j].null_bit ||
|
|
t1_keysegs[j].length != t2_keysegs[j].length ||
|
|
t1_keysegs[j].start != t2_keysegs[j].start)
|
|
{
|
|
DBUG_PRINT("error", ("Key segment %d (key %d) has different "
|
|
"definition", j, i));
|
|
DBUG_PRINT("error", ("t1_type=%d, t1_language=%d, t1_null_bit=%d, "
|
|
"t1_length=%d",
|
|
t1_keysegs[j].type, t1_keysegs[j].language,
|
|
t1_keysegs[j].null_bit, t1_keysegs[j].length));
|
|
DBUG_PRINT("error", ("t2_type=%d, t2_language=%d, t2_null_bit=%d, "
|
|
"t2_length=%d",
|
|
t2_keysegs[j].type, t2_keysegs[j].language,
|
|
t2_keysegs[j].null_bit, t2_keysegs[j].length));
|
|
|
|
DBUG_RETURN(1);
|
|
}
|
|
}
|
|
}
|
|
for (i= 0; i < t1_recs; i++)
|
|
{
|
|
MI_COLUMNDEF *t1_rec= &t1_recinfo[i];
|
|
MI_COLUMNDEF *t2_rec= &t2_recinfo[i];
|
|
/*
|
|
FIELD_SKIP_ZERO can be changed to FIELD_NORMAL in mi_create,
|
|
see NOTE1 in mi_create.c
|
|
*/
|
|
if ((t1_rec->type != t2_rec->type &&
|
|
!(t1_rec->type == (int) FIELD_SKIP_ZERO &&
|
|
t1_rec->length == 1 &&
|
|
t2_rec->type == (int) FIELD_NORMAL)) ||
|
|
t1_rec->length != t2_rec->length ||
|
|
t1_rec->null_bit != t2_rec->null_bit)
|
|
{
|
|
DBUG_PRINT("error", ("Field %d has different definition", i));
|
|
DBUG_PRINT("error", ("t1_type=%d, t1_length=%d, t1_null_bit=%d",
|
|
t1_rec->type, t1_rec->length, t1_rec->null_bit));
|
|
DBUG_PRINT("error", ("t2_type=%d, t2_length=%d, t2_null_bit=%d",
|
|
t2_rec->type, t2_rec->length, t2_rec->null_bit));
|
|
DBUG_RETURN(1);
|
|
}
|
|
}
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
extern "C" {
|
|
|
|
int killed_ptr(HA_CHECK *param)
|
|
{
|
|
if (likely(thd_killed((THD*)param->thd)) == 0)
|
|
return 0;
|
|
my_errno= HA_ERR_ABORTED_BY_USER;
|
|
return 1;
|
|
}
|
|
|
|
void mi_check_print_error(HA_CHECK *param, const char *fmt,...)
|
|
{
|
|
param->error_printed|=1;
|
|
param->out_flag|= O_DATA_LOST;
|
|
if (param->testflag & T_SUPPRESS_ERR_HANDLING)
|
|
return;
|
|
va_list args;
|
|
va_start(args, fmt);
|
|
mi_check_print_msg(param, MI_CHECK_ERROR, fmt, args);
|
|
va_end(args);
|
|
}
|
|
|
|
void mi_check_print_info(HA_CHECK *param, const char *fmt,...)
|
|
{
|
|
va_list args;
|
|
va_start(args, fmt);
|
|
mi_check_print_msg(param, MI_CHECK_INFO, fmt, args);
|
|
param->note_printed= 1;
|
|
va_end(args);
|
|
}
|
|
|
|
void mi_check_print_warning(HA_CHECK *param, const char *fmt,...)
|
|
{
|
|
param->warning_printed=1;
|
|
param->out_flag|= O_DATA_LOST;
|
|
va_list args;
|
|
va_start(args, fmt);
|
|
mi_check_print_msg(param, MI_CHECK_WARNING, fmt, args);
|
|
va_end(args);
|
|
}
|
|
|
|
|
|
/**
|
|
Report list of threads (and queries) accessing a table, thread_id of a
|
|
thread that detected corruption, ource file name and line number where
|
|
this corruption was detected, optional extra information (string).
|
|
|
|
This function is intended to be used when table corruption is detected.
|
|
|
|
@param[in] file MI_INFO object.
|
|
@param[in] message Optional error message.
|
|
@param[in] sfile Name of source file.
|
|
@param[in] sline Line number in source file.
|
|
|
|
@return void
|
|
*/
|
|
|
|
void _mi_report_crashed(MI_INFO *file, const char *message,
|
|
const char *sfile, uint sline)
|
|
{
|
|
THD *cur_thd;
|
|
LIST *element;
|
|
char buf[1024];
|
|
mysql_mutex_lock(&file->s->intern_lock);
|
|
if ((cur_thd= (THD*) file->in_use.data))
|
|
sql_print_error("Got an error from thread_id=%lld, %s:%d", cur_thd->thread_id,
|
|
sfile, sline);
|
|
else
|
|
sql_print_error("Got an error from unknown thread, %s:%d", sfile, sline);
|
|
if (message)
|
|
sql_print_error("%s", message);
|
|
for (element= file->s->in_use; element; element= list_rest(element))
|
|
{
|
|
THD *thd= (THD*) element->data;
|
|
sql_print_error("%s",
|
|
thd ?
|
|
thd_get_error_context_description(thd, buf, sizeof(buf), 0)
|
|
: "Unknown thread accessing table");
|
|
}
|
|
mysql_mutex_unlock(&file->s->intern_lock);
|
|
}
|
|
|
|
/* Return 1 if user have requested query to be killed */
|
|
|
|
my_bool mi_killed_in_mariadb(MI_INFO *info)
|
|
{
|
|
return (((TABLE*) (info->external_ref))->in_use->killed != 0);
|
|
}
|
|
|
|
static int compute_vcols(MI_INFO *info, uchar *record, int keynum)
|
|
{
|
|
/* This mutex is needed for parallel repair */
|
|
mysql_mutex_lock(&info->s->intern_lock);
|
|
TABLE *table= (TABLE*)(info->external_ref);
|
|
table->move_fields(table->field, record, table->field[0]->record_ptr());
|
|
if (keynum == -1) // update all vcols
|
|
{
|
|
int error= table->update_virtual_fields(table->file, VCOL_UPDATE_FOR_READ);
|
|
if (table->update_virtual_fields(table->file, VCOL_UPDATE_INDEXED))
|
|
error= 1;
|
|
mysql_mutex_unlock(&info->s->intern_lock);
|
|
return error;
|
|
}
|
|
// update only one key
|
|
KEY *key= table->key_info + keynum;
|
|
KEY_PART_INFO *kp= key->key_part, *end= kp + key->ext_key_parts;
|
|
for (; kp < end; kp++)
|
|
{
|
|
Field *f= table->field[kp->fieldnr - 1];
|
|
if (f->vcol_info && !f->vcol_info->stored_in_db)
|
|
table->update_virtual_field(f, false);
|
|
}
|
|
mysql_mutex_unlock(&info->s->intern_lock);
|
|
return 0;
|
|
}
|
|
|
|
}
|
|
|
|
ha_myisam::ha_myisam(handlerton *hton, TABLE_SHARE *table_arg)
|
|
:handler(hton, table_arg), file(0),
|
|
int_table_flags(HA_NULL_IN_KEY | HA_CAN_FULLTEXT | HA_CAN_SQL_HANDLER |
|
|
HA_BINLOG_ROW_CAPABLE | HA_BINLOG_STMT_CAPABLE |
|
|
HA_CAN_VIRTUAL_COLUMNS | HA_CAN_EXPORT |
|
|
HA_REQUIRES_KEY_COLUMNS_FOR_DELETE |
|
|
HA_DUPLICATE_POS | HA_CAN_INDEX_BLOBS | HA_AUTO_PART_KEY |
|
|
HA_FILE_BASED | HA_CAN_GEOMETRY | HA_NO_TRANSACTIONS |
|
|
HA_CAN_INSERT_DELAYED | HA_CAN_BIT_FIELD | HA_CAN_RTREEKEYS |
|
|
HA_HAS_RECORDS | HA_STATS_RECORDS_IS_EXACT | HA_CAN_REPAIR |
|
|
HA_CAN_TABLES_WITHOUT_ROLLBACK),
|
|
can_enable_indexes(0)
|
|
{}
|
|
|
|
handler *ha_myisam::clone(const char *name __attribute__((unused)),
|
|
MEM_ROOT *mem_root)
|
|
{
|
|
ha_myisam *new_handler=
|
|
static_cast <ha_myisam *>(handler::clone(file->filename, mem_root));
|
|
if (new_handler)
|
|
new_handler->file->state= file->state;
|
|
return new_handler;
|
|
}
|
|
|
|
|
|
static const char *ha_myisam_exts[] = {
|
|
".MYI",
|
|
".MYD",
|
|
NullS
|
|
};
|
|
|
|
const char *ha_myisam::index_type(uint key_number)
|
|
{
|
|
return ((table->key_info[key_number].flags & HA_FULLTEXT) ?
|
|
"FULLTEXT" :
|
|
(table->key_info[key_number].flags & HA_SPATIAL) ?
|
|
"SPATIAL" :
|
|
(table->key_info[key_number].algorithm == HA_KEY_ALG_RTREE) ?
|
|
"RTREE" :
|
|
"BTREE");
|
|
}
|
|
|
|
|
|
ulong ha_myisam::index_flags(uint inx, uint part, bool all_parts) const
|
|
{
|
|
ulong flags;
|
|
if (table_share->key_info[inx].algorithm == HA_KEY_ALG_FULLTEXT)
|
|
flags= 0;
|
|
else
|
|
if ((table_share->key_info[inx].flags & HA_SPATIAL ||
|
|
table_share->key_info[inx].algorithm == HA_KEY_ALG_RTREE))
|
|
{
|
|
/* All GIS scans are non-ROR scans. We also disable IndexConditionPushdown */
|
|
flags= HA_READ_NEXT | HA_READ_PREV | HA_READ_RANGE |
|
|
HA_READ_ORDER | HA_KEYREAD_ONLY | HA_KEY_SCAN_NOT_ROR;
|
|
}
|
|
else
|
|
{
|
|
flags= HA_READ_NEXT | HA_READ_PREV | HA_READ_RANGE |
|
|
HA_READ_ORDER | HA_KEYREAD_ONLY | HA_DO_INDEX_COND_PUSHDOWN |
|
|
HA_DO_RANGE_FILTER_PUSHDOWN;
|
|
}
|
|
return flags;
|
|
}
|
|
|
|
|
|
/* Name is here without an extension */
|
|
int ha_myisam::open(const char *name, int mode, uint test_if_locked)
|
|
{
|
|
MI_KEYDEF *keyinfo;
|
|
MI_COLUMNDEF *recinfo= 0;
|
|
char readlink_buf[FN_REFLEN], name_buff[FN_REFLEN];
|
|
uint recs;
|
|
uint i;
|
|
|
|
/*
|
|
If the user wants to have memory mapped data files, add an
|
|
open_flag. Do not memory map temporary tables because they are
|
|
expected to be inserted and thus extended a lot. Memory mapping is
|
|
efficient for files that keep their size, but very inefficient for
|
|
growing files. Using an open_flag instead of calling mi_extra(...
|
|
HA_EXTRA_MMAP ...) after mi_open() has the advantage that the
|
|
mapping is not repeated for every open, but just done on the initial
|
|
open, when the MyISAM share is created. Every time the server
|
|
requires opening a new instance of a table it calls this method. We
|
|
will always supply HA_OPEN_MMAP for a permanent table. However, the
|
|
MyISAM storage engine will ignore this flag if this is a secondary
|
|
open of a table that is in use by other threads already (if the
|
|
MyISAM share exists already).
|
|
*/
|
|
if (!(test_if_locked & HA_OPEN_TMP_TABLE) && opt_myisam_use_mmap)
|
|
test_if_locked|= HA_OPEN_MMAP;
|
|
|
|
if (!(file=mi_open(name, mode, test_if_locked | HA_OPEN_FROM_SQL_LAYER)))
|
|
return (my_errno ? my_errno : -1);
|
|
|
|
file->s->chst_invalidator= query_cache_invalidate_by_MyISAM_filename_ref;
|
|
/* Set external_ref, mainly for temporary tables */
|
|
file->external_ref= (void*) table; // For mi_killed()
|
|
|
|
/* No need to perform a check for tmp table or if it's already checked */
|
|
if (!table->s->tmp_table && file->s->reopen == 1)
|
|
{
|
|
if ((my_errno= table2myisam(table, &keyinfo, &recinfo, &recs)))
|
|
{
|
|
/* purecov: begin inspected */
|
|
DBUG_PRINT("error", ("Failed to convert TABLE object to MyISAM "
|
|
"key and column definition"));
|
|
goto err;
|
|
/* purecov: end */
|
|
}
|
|
if (check_definition(keyinfo, recinfo, table->s->keys, recs,
|
|
file->s->keyinfo, file->s->rec,
|
|
file->s->base.keys, file->s->base.fields,
|
|
true, table))
|
|
{
|
|
/* purecov: begin inspected */
|
|
my_errno= HA_ERR_INCOMPATIBLE_DEFINITION;
|
|
goto err;
|
|
/* purecov: end */
|
|
}
|
|
}
|
|
|
|
DBUG_EXECUTE_IF("key",
|
|
Debug_key_myisam::print_keys_myisam(table->in_use,
|
|
"ha_myisam::open: ",
|
|
table, file->s->keyinfo,
|
|
file->s->base.keys);
|
|
);
|
|
|
|
if (test_if_locked & (HA_OPEN_IGNORE_IF_LOCKED | HA_OPEN_TMP_TABLE))
|
|
(void) mi_extra(file, HA_EXTRA_NO_WAIT_LOCK, 0);
|
|
|
|
info(HA_STATUS_NO_LOCK | HA_STATUS_VARIABLE | HA_STATUS_CONST);
|
|
|
|
/*
|
|
Set data_file_name and index_file_name to point at the symlink value
|
|
if table is symlinked (Ie; Real name is not same as generated name)
|
|
*/
|
|
fn_format(name_buff, file->filename, "", MI_NAME_DEXT,
|
|
MY_APPEND_EXT | MY_UNPACK_FILENAME);
|
|
if (my_is_symlink(name_buff))
|
|
{
|
|
my_readlink(readlink_buf, name_buff, MYF(0));
|
|
data_file_name= strdup_root(&table->mem_root, readlink_buf);
|
|
}
|
|
else
|
|
data_file_name= 0;
|
|
fn_format(name_buff, file->filename, "", MI_NAME_IEXT,
|
|
MY_APPEND_EXT | MY_UNPACK_FILENAME);
|
|
if (my_is_symlink(name_buff))
|
|
{
|
|
my_readlink(readlink_buf, name_buff, MYF(0));
|
|
index_file_name= strdup_root(&table->mem_root, readlink_buf);
|
|
}
|
|
else
|
|
index_file_name= 0;
|
|
|
|
if (!(test_if_locked & HA_OPEN_WAIT_IF_LOCKED))
|
|
(void) mi_extra(file, HA_EXTRA_WAIT_LOCK, 0);
|
|
if (!table->s->db_record_offset)
|
|
int_table_flags|=HA_REC_NOT_IN_SEQ;
|
|
if (file->s->options & (HA_OPTION_CHECKSUM | HA_OPTION_COMPRESS_RECORD))
|
|
{
|
|
/*
|
|
Set which type of automatic checksum we have
|
|
The old checksum and new checksum are identical if there is no
|
|
null fields.
|
|
Files with new checksum has the HA_OPTION_NULL_FIELDS bit set.
|
|
*/
|
|
if ((file->s->options & HA_OPTION_NULL_FIELDS) ||
|
|
!file->s->has_null_fields)
|
|
int_table_flags|= HA_HAS_NEW_CHECKSUM;
|
|
if (!(file->s->options & HA_OPTION_NULL_FIELDS))
|
|
int_table_flags|= HA_HAS_OLD_CHECKSUM;
|
|
}
|
|
|
|
/*
|
|
For static size rows, tell MariaDB that we will access all bytes
|
|
in the record when writing it. This signals MariaDB to initialize
|
|
the full row to ensure we don't get any errors from valgrind and
|
|
that all bytes in the row is properly reset.
|
|
*/
|
|
if (!(file->s->options &
|
|
(HA_OPTION_PACK_RECORD | HA_OPTION_COMPRESS_RECORD)) &&
|
|
(file->s->has_varchar_fields || file->s->has_null_fields))
|
|
int_table_flags|= HA_RECORD_MUST_BE_CLEAN_ON_WRITE;
|
|
|
|
for (i= 0; i < table->s->keys; i++)
|
|
{
|
|
plugin_ref parser= table->key_info[i].parser;
|
|
if (table->key_info[i].flags & HA_USES_PARSER)
|
|
file->s->keyinfo[i].parser=
|
|
(struct st_mysql_ftparser *)plugin_decl(parser)->info;
|
|
table->key_info[i].block_size= file->s->keyinfo[i].block_length;
|
|
}
|
|
my_errno= 0;
|
|
|
|
/* Count statistics of usage for newly open normal files */
|
|
if (file->s->reopen == 1 && ! (test_if_locked & HA_OPEN_TMP_TABLE))
|
|
{
|
|
/* use delay_key_write from .frm, not .MYI */
|
|
file->s->delay_key_write= delay_key_write_options == DELAY_KEY_WRITE_ALL ||
|
|
(delay_key_write_options == DELAY_KEY_WRITE_ON &&
|
|
table->s->db_create_options & HA_OPTION_DELAY_KEY_WRITE);
|
|
if (file->s->delay_key_write)
|
|
feature_files_opened_with_delayed_keys++;
|
|
}
|
|
goto end;
|
|
|
|
err:
|
|
this->close();
|
|
end:
|
|
/*
|
|
Both recinfo and keydef are allocated by my_multi_malloc(), thus only
|
|
recinfo must be freed.
|
|
*/
|
|
if (recinfo)
|
|
my_free(recinfo);
|
|
return my_errno;
|
|
}
|
|
|
|
int ha_myisam::close(void)
|
|
{
|
|
MI_INFO *tmp=file;
|
|
if (!tmp)
|
|
return 0;
|
|
file=0;
|
|
return mi_close(tmp);
|
|
}
|
|
|
|
int ha_myisam::write_row(const uchar *buf)
|
|
{
|
|
/*
|
|
If we have an auto_increment column and we are writing a changed row
|
|
or a new row, then update the auto_increment value in the record.
|
|
*/
|
|
if (table->next_number_field && buf == table->record[0])
|
|
{
|
|
int error;
|
|
if ((error= update_auto_increment()))
|
|
return error;
|
|
}
|
|
return mi_write(file,buf);
|
|
}
|
|
|
|
void ha_myisam::setup_vcols_for_repair(HA_CHECK *param)
|
|
{
|
|
DBUG_ASSERT(file->s->base.reclength <= file->s->vreclength);
|
|
if (!table->vfield)
|
|
return;
|
|
|
|
if (file->s->base.reclength == file->s->vreclength)
|
|
{
|
|
bool indexed_vcols= false;
|
|
ulong new_vreclength= file->s->vreclength;
|
|
for (Field **vf= table->vfield; *vf; vf++)
|
|
{
|
|
if (!(*vf)->stored_in_db())
|
|
{
|
|
uint vf_end= ((*vf)->offset(table->record[0]) +
|
|
(*vf)->pack_length_in_rec());
|
|
set_if_bigger(new_vreclength, vf_end);
|
|
indexed_vcols|= ((*vf)->flags & PART_KEY_FLAG) != 0;
|
|
}
|
|
}
|
|
if (!indexed_vcols)
|
|
return;
|
|
file->s->vreclength= new_vreclength;
|
|
}
|
|
DBUG_ASSERT(file->s->base.reclength < file->s->vreclength ||
|
|
!table->s->stored_fields);
|
|
param->fix_record= compute_vcols;
|
|
table->use_all_columns();
|
|
}
|
|
|
|
void ha_myisam::restore_vcos_after_repair()
|
|
{
|
|
if (file->s->base.reclength < file->s->vreclength)
|
|
{
|
|
table->move_fields(table->field, table->record[0],
|
|
table->field[0]->record_ptr());
|
|
table->default_column_bitmaps();
|
|
}
|
|
}
|
|
|
|
int ha_myisam::check(THD* thd, HA_CHECK_OPT* check_opt)
|
|
{
|
|
if (!file) return HA_ADMIN_INTERNAL_ERROR;
|
|
int error;
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
MYISAM_SHARE* share = file->s;
|
|
const char *old_proc_info=thd->proc_info;
|
|
|
|
if (!param)
|
|
return HA_ADMIN_INTERNAL_ERROR;
|
|
|
|
thd_proc_info(thd, "Checking table");
|
|
myisamchk_init(param);
|
|
param->thd = thd;
|
|
param->op_name = "check";
|
|
param->db_name= table->s->db.str;
|
|
param->table_name= table->alias.c_ptr();
|
|
param->testflag = check_opt->flags | T_CHECK | T_SILENT;
|
|
param->stats_method= (enum_handler_stats_method)THDVAR(thd, stats_method);
|
|
|
|
if (!(table->db_stat & HA_READ_ONLY))
|
|
param->testflag|= T_STATISTICS;
|
|
param->using_global_keycache = 1;
|
|
|
|
if (!mi_is_crashed(file) &&
|
|
(((param->testflag & T_CHECK_ONLY_CHANGED) &&
|
|
!(share->state.changed & (STATE_CHANGED | STATE_CRASHED |
|
|
STATE_CRASHED_ON_REPAIR)) &&
|
|
share->state.open_count == 0) ||
|
|
((param->testflag & T_FAST) && (share->state.open_count ==
|
|
(uint) (share->global_changed ? 1 : 0)))))
|
|
return HA_ADMIN_ALREADY_DONE;
|
|
|
|
setup_vcols_for_repair(param);
|
|
|
|
error = chk_status(param, file); // Not fatal
|
|
error = chk_size(param, file);
|
|
if (!error)
|
|
error |= chk_del(param, file, param->testflag);
|
|
if (!error)
|
|
error = chk_key(param, file);
|
|
if (!error)
|
|
{
|
|
if ((!(param->testflag & T_QUICK) &&
|
|
((share->options &
|
|
(HA_OPTION_PACK_RECORD | HA_OPTION_COMPRESS_RECORD)) ||
|
|
(param->testflag & (T_EXTEND | T_MEDIUM)))) ||
|
|
mi_is_crashed(file))
|
|
{
|
|
ulonglong old_testflag= param->testflag;
|
|
param->testflag|=T_MEDIUM;
|
|
if (!(error= init_io_cache(¶m->read_cache, file->dfile,
|
|
my_default_record_cache_size, READ_CACHE,
|
|
share->pack.header_length, 1, MYF(MY_WME))))
|
|
{
|
|
error= chk_data_link(param, file, MY_TEST(param->testflag & T_EXTEND));
|
|
end_io_cache(¶m->read_cache);
|
|
}
|
|
param->testflag= old_testflag;
|
|
}
|
|
}
|
|
if (!error)
|
|
{
|
|
if ((share->state.changed & (STATE_CHANGED |
|
|
STATE_CRASHED_ON_REPAIR |
|
|
STATE_CRASHED | STATE_NOT_ANALYZED)) ||
|
|
(param->testflag & T_STATISTICS) ||
|
|
mi_is_crashed(file))
|
|
{
|
|
file->update|=HA_STATE_CHANGED | HA_STATE_ROW_CHANGED;
|
|
mysql_mutex_lock(&share->intern_lock);
|
|
share->state.changed&= ~(STATE_CHANGED | STATE_CRASHED |
|
|
STATE_CRASHED_ON_REPAIR);
|
|
if (!(table->db_stat & HA_READ_ONLY))
|
|
error=update_state_info(param,file,UPDATE_TIME | UPDATE_OPEN_COUNT |
|
|
UPDATE_STAT);
|
|
mysql_mutex_unlock(&share->intern_lock);
|
|
info(HA_STATUS_NO_LOCK | HA_STATUS_TIME | HA_STATUS_VARIABLE |
|
|
HA_STATUS_CONST);
|
|
/*
|
|
Write a 'table is ok' message to error log if table is ok and
|
|
we have written to error log that table was getting checked
|
|
*/
|
|
if (!error && !(table->db_stat & HA_READ_ONLY) &&
|
|
!mi_is_crashed(file) && thd->error_printed_to_log &&
|
|
(param->warning_printed || param->error_printed ||
|
|
param->note_printed))
|
|
mi_check_print_info(param, "Table is fixed");
|
|
}
|
|
}
|
|
else if (!mi_is_crashed(file) && !thd->killed)
|
|
{
|
|
mi_mark_crashed(file);
|
|
file->update |= HA_STATE_CHANGED | HA_STATE_ROW_CHANGED;
|
|
}
|
|
|
|
restore_vcos_after_repair();
|
|
|
|
thd_proc_info(thd, old_proc_info);
|
|
return error ? HA_ADMIN_CORRUPT : HA_ADMIN_OK;
|
|
}
|
|
|
|
|
|
/*
|
|
analyze the key distribution in the table
|
|
As the table may be only locked for read, we have to take into account that
|
|
two threads may do an analyze at the same time!
|
|
*/
|
|
|
|
int ha_myisam::analyze(THD *thd, HA_CHECK_OPT* check_opt)
|
|
{
|
|
int error=0;
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
MYISAM_SHARE* share = file->s;
|
|
|
|
if (!param)
|
|
return HA_ADMIN_INTERNAL_ERROR;
|
|
|
|
myisamchk_init(param);
|
|
param->thd = thd;
|
|
param->op_name= "analyze";
|
|
param->db_name= table->s->db.str;
|
|
param->table_name= table->alias.c_ptr();
|
|
param->testflag= (T_FAST | T_CHECK | T_SILENT | T_STATISTICS |
|
|
T_DONT_CHECK_CHECKSUM);
|
|
param->using_global_keycache = 1;
|
|
param->stats_method= (enum_handler_stats_method)THDVAR(thd, stats_method);
|
|
|
|
if (!(share->state.changed & STATE_NOT_ANALYZED))
|
|
return HA_ADMIN_ALREADY_DONE;
|
|
|
|
setup_vcols_for_repair(param);
|
|
|
|
error = chk_key(param, file);
|
|
if (!error)
|
|
{
|
|
mysql_mutex_lock(&share->intern_lock);
|
|
error=update_state_info(param,file,UPDATE_STAT);
|
|
mysql_mutex_unlock(&share->intern_lock);
|
|
}
|
|
else if (!mi_is_crashed(file) && !thd->killed)
|
|
mi_mark_crashed(file);
|
|
|
|
restore_vcos_after_repair();
|
|
|
|
return error ? HA_ADMIN_CORRUPT : HA_ADMIN_OK;
|
|
}
|
|
|
|
|
|
int ha_myisam::repair(THD* thd, HA_CHECK_OPT *check_opt)
|
|
{
|
|
int error;
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
ha_rows start_records;
|
|
|
|
if (!file || !param) return HA_ADMIN_INTERNAL_ERROR;
|
|
|
|
myisamchk_init(param);
|
|
param->thd = thd;
|
|
param->op_name= "repair";
|
|
param->testflag= ((check_opt->flags & ~(T_EXTEND)) |
|
|
T_SILENT | T_FORCE_CREATE | T_CALC_CHECKSUM |
|
|
(check_opt->flags & T_EXTEND ? T_REP : T_REP_BY_SORT));
|
|
param->tmpfile_createflag= O_RDWR | O_TRUNC;
|
|
param->sort_buffer_length= THDVAR(thd, sort_buffer_size);
|
|
param->backup_time= check_opt->start_time;
|
|
start_records=file->state->records;
|
|
|
|
setup_vcols_for_repair(param);
|
|
|
|
while ((error=repair(thd,*param,0)) && param->retry_repair)
|
|
{
|
|
param->retry_repair=0;
|
|
if (test_all_bits(param->testflag,
|
|
(uint) (T_RETRY_WITHOUT_QUICK | T_QUICK)))
|
|
{
|
|
param->testflag&= ~(T_RETRY_WITHOUT_QUICK | T_QUICK);
|
|
/* Ensure we don't loose any rows when retrying without quick */
|
|
param->testflag|= T_SAFE_REPAIR;
|
|
sql_print_information("Retrying repair of: '%s' including modifying data file",
|
|
table->s->path.str);
|
|
continue;
|
|
}
|
|
param->testflag&= ~T_QUICK;
|
|
if ((param->testflag & (T_REP_BY_SORT | T_REP_PARALLEL)))
|
|
{
|
|
param->testflag= (param->testflag & ~T_REP_ANY) | T_REP;
|
|
sql_print_information("Retrying repair of: '%s' with keycache",
|
|
table->s->path.str);
|
|
continue;
|
|
}
|
|
break;
|
|
}
|
|
|
|
restore_vcos_after_repair();
|
|
|
|
if (!error && start_records != file->state->records &&
|
|
!(check_opt->flags & T_VERY_SILENT))
|
|
{
|
|
char llbuff[22],llbuff2[22];
|
|
sql_print_information("Found %s of %s rows when repairing '%s'",
|
|
llstr(file->state->records, llbuff),
|
|
llstr(start_records, llbuff2),
|
|
table->s->path.str);
|
|
}
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::optimize(THD* thd, HA_CHECK_OPT *check_opt)
|
|
{
|
|
int error;
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
|
|
if (!file || !param) return HA_ADMIN_INTERNAL_ERROR;
|
|
|
|
myisamchk_init(param);
|
|
param->thd = thd;
|
|
param->op_name= "optimize";
|
|
param->testflag= (check_opt->flags | T_SILENT | T_FORCE_CREATE |
|
|
T_REP_BY_SORT | T_STATISTICS | T_SORT_INDEX);
|
|
param->tmpfile_createflag= O_RDWR | O_TRUNC;
|
|
param->sort_buffer_length= THDVAR(thd, sort_buffer_size);
|
|
|
|
setup_vcols_for_repair(param);
|
|
|
|
if ((error= repair(thd,*param,1)) && param->retry_repair)
|
|
{
|
|
sql_print_warning("Warning: Optimize table got errno %d on %s.%s, retrying",
|
|
my_errno, param->db_name, param->table_name);
|
|
param->testflag&= ~T_REP_BY_SORT;
|
|
error= repair(thd,*param,1);
|
|
}
|
|
|
|
restore_vcos_after_repair();
|
|
|
|
return error;
|
|
}
|
|
|
|
|
|
int ha_myisam::repair(THD *thd, HA_CHECK ¶m, bool do_optimize)
|
|
{
|
|
int error=0;
|
|
ulonglong local_testflag= param.testflag;
|
|
bool optimize_done= !do_optimize, statistics_done=0;
|
|
const char *old_proc_info=thd->proc_info;
|
|
char fixed_name[FN_REFLEN];
|
|
MYISAM_SHARE* share = file->s;
|
|
ha_rows rows= file->state->records;
|
|
my_bool locking= 0;
|
|
DBUG_ENTER("ha_myisam::repair");
|
|
|
|
param.db_name= table->s->db.str;
|
|
param.table_name= table->alias.c_ptr();
|
|
param.using_global_keycache = 1;
|
|
param.thd= thd;
|
|
param.tmpdir= &mysql_tmpdir_list;
|
|
param.out_flag= 0;
|
|
share->state.dupp_key= MI_MAX_KEY;
|
|
strmov(fixed_name,file->filename);
|
|
|
|
/*
|
|
Don't lock tables if we have used LOCK TABLE or if we come from
|
|
enable_index()
|
|
*/
|
|
if (!thd->locked_tables_mode && ! (param.testflag & T_NO_LOCKS))
|
|
{
|
|
locking= 1;
|
|
if (mi_lock_database(file, table->s->tmp_table ? F_EXTRA_LCK : F_WRLCK))
|
|
{
|
|
mi_check_print_error(¶m, ER_THD(thd, ER_CANT_LOCK), my_errno);
|
|
DBUG_RETURN(HA_ADMIN_FAILED);
|
|
}
|
|
}
|
|
|
|
if (!do_optimize ||
|
|
((file->state->del || share->state.split != file->state->records) &&
|
|
(!(param.testflag & T_QUICK) ||
|
|
!(share->state.changed & STATE_NOT_OPTIMIZED_KEYS))))
|
|
{
|
|
ulonglong tmp_key_map= ((local_testflag & T_CREATE_MISSING_KEYS) ?
|
|
mi_get_mask_all_keys_active(share->base.keys) :
|
|
share->state.key_map);
|
|
ulonglong testflag= param.testflag;
|
|
#ifdef HAVE_MMAP
|
|
bool remap= MY_TEST(share->file_map);
|
|
/*
|
|
mi_repair*() functions family use file I/O even if memory
|
|
mapping is available.
|
|
|
|
Since mixing mmap I/O and file I/O may cause various artifacts,
|
|
memory mapping must be disabled.
|
|
*/
|
|
if (remap)
|
|
mi_munmap_file(file);
|
|
#endif
|
|
/*
|
|
The following is to catch errors when my_errno is no set properly
|
|
during repairt
|
|
*/
|
|
my_errno= 0;
|
|
if (mi_test_if_sort_rep(file,file->state->records,tmp_key_map,0) &&
|
|
(local_testflag & T_REP_BY_SORT))
|
|
{
|
|
local_testflag|= T_STATISTICS;
|
|
param.testflag|= T_STATISTICS; // We get this for free
|
|
statistics_done=1;
|
|
if (THDVAR(thd, repair_threads)>1)
|
|
{
|
|
/* TODO: respect myisam_repair_threads variable */
|
|
thd_proc_info(thd, "Parallel repair");
|
|
error = mi_repair_parallel(¶m, file, fixed_name,
|
|
MY_TEST(param.testflag & T_QUICK));
|
|
}
|
|
else
|
|
{
|
|
thd_proc_info(thd, "Repair by sorting");
|
|
DEBUG_SYNC(thd, "myisam_before_repair_by_sort");
|
|
error = mi_repair_by_sort(¶m, file, fixed_name,
|
|
MY_TEST(param.testflag & T_QUICK));
|
|
}
|
|
if (error && file->create_unique_index_by_sort &&
|
|
share->state.dupp_key != MAX_KEY)
|
|
{
|
|
my_errno= HA_ERR_FOUND_DUPP_KEY;
|
|
print_keydup_error(table, &table->key_info[share->state.dupp_key],
|
|
MYF(0));
|
|
}
|
|
}
|
|
else
|
|
{
|
|
thd_proc_info(thd, "Repair with keycache");
|
|
param.testflag &= ~T_REP_BY_SORT;
|
|
error= mi_repair(¶m, file, fixed_name,
|
|
MY_TEST(param.testflag & T_QUICK));
|
|
}
|
|
param.testflag= testflag | (param.testflag & T_RETRY_WITHOUT_QUICK);
|
|
#ifdef HAVE_MMAP
|
|
if (remap)
|
|
mi_dynmap_file(file, file->state->data_file_length);
|
|
#endif
|
|
optimize_done=1;
|
|
}
|
|
if (!error)
|
|
{
|
|
if ((local_testflag & T_SORT_INDEX) &&
|
|
(share->state.changed & STATE_NOT_SORTED_PAGES))
|
|
{
|
|
optimize_done=1;
|
|
thd_proc_info(thd, "Sorting index");
|
|
error=mi_sort_index(¶m,file,fixed_name);
|
|
}
|
|
if (!error && !statistics_done && (local_testflag & T_STATISTICS))
|
|
{
|
|
if (share->state.changed & STATE_NOT_ANALYZED)
|
|
{
|
|
optimize_done=1;
|
|
thd_proc_info(thd, "Analyzing");
|
|
error = chk_key(¶m, file);
|
|
}
|
|
else
|
|
local_testflag&= ~T_STATISTICS; // Don't update statistics
|
|
}
|
|
}
|
|
thd_proc_info(thd, "Saving state");
|
|
if (!error)
|
|
{
|
|
if ((share->state.changed & STATE_CHANGED) || mi_is_crashed(file))
|
|
{
|
|
share->state.changed&= ~(STATE_CHANGED | STATE_CRASHED |
|
|
STATE_CRASHED_ON_REPAIR);
|
|
file->update|=HA_STATE_CHANGED | HA_STATE_ROW_CHANGED;
|
|
}
|
|
/*
|
|
the following 'if', thought conceptually wrong,
|
|
is a useful optimization nevertheless.
|
|
*/
|
|
if (file->state != &file->s->state.state)
|
|
file->s->state.state = *file->state;
|
|
if (file->s->base.auto_key)
|
|
update_auto_increment_key(¶m, file, 1);
|
|
if (optimize_done)
|
|
{
|
|
mysql_mutex_lock(&share->intern_lock);
|
|
error = update_state_info(¶m, file,
|
|
UPDATE_TIME | UPDATE_OPEN_COUNT |
|
|
(local_testflag &
|
|
T_STATISTICS ? UPDATE_STAT : 0));
|
|
mysql_mutex_unlock(&share->intern_lock);
|
|
}
|
|
info(HA_STATUS_NO_LOCK | HA_STATUS_TIME | HA_STATUS_VARIABLE |
|
|
HA_STATUS_CONST);
|
|
if (rows != file->state->records && ! (param.testflag & T_VERY_SILENT))
|
|
{
|
|
char llbuff[22],llbuff2[22];
|
|
mi_check_print_warning(¶m,"Number of rows changed from %s to %s",
|
|
llstr(rows,llbuff),
|
|
llstr(file->state->records,llbuff2));
|
|
}
|
|
}
|
|
else
|
|
{
|
|
mi_mark_crashed_on_repair(file);
|
|
file->update |= HA_STATE_CHANGED | HA_STATE_ROW_CHANGED;
|
|
update_state_info(¶m, file, 0);
|
|
}
|
|
thd_proc_info(thd, old_proc_info);
|
|
if (locking)
|
|
mi_lock_database(file,F_UNLCK);
|
|
DBUG_RETURN(error ? HA_ADMIN_FAILED :
|
|
!optimize_done ? HA_ADMIN_ALREADY_DONE : HA_ADMIN_OK);
|
|
}
|
|
|
|
|
|
/*
|
|
Assign table indexes to a specific key cache.
|
|
*/
|
|
|
|
int ha_myisam::assign_to_keycache(THD* thd, HA_CHECK_OPT *check_opt)
|
|
{
|
|
KEY_CACHE *new_key_cache= check_opt->key_cache;
|
|
const char *errmsg= 0;
|
|
char buf[STRING_BUFFER_USUAL_SIZE];
|
|
int error= HA_ADMIN_OK;
|
|
ulonglong map;
|
|
TABLE_LIST *table_list= table->pos_in_table_list;
|
|
DBUG_ENTER("ha_myisam::assign_to_keycache");
|
|
|
|
table->keys_in_use_for_query.clear_all();
|
|
|
|
if (table_list->process_index_hints(table))
|
|
DBUG_RETURN(HA_ADMIN_FAILED);
|
|
map= ~(ulonglong) 0;
|
|
if (!table->keys_in_use_for_query.is_clear_all())
|
|
/* use all keys if there's no list specified by the user through hints */
|
|
map= table->keys_in_use_for_query.to_ulonglong();
|
|
|
|
if ((error= mi_assign_to_key_cache(file, map, new_key_cache)))
|
|
{
|
|
my_snprintf(buf, sizeof(buf),
|
|
"Failed to flush to index file (errno: %d)", error);
|
|
errmsg= buf;
|
|
error= HA_ADMIN_CORRUPT;
|
|
}
|
|
|
|
if (error != HA_ADMIN_OK)
|
|
{
|
|
/* Send error to user */
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
if (!param)
|
|
return HA_ADMIN_INTERNAL_ERROR;
|
|
|
|
myisamchk_init(param);
|
|
param->thd= thd;
|
|
param->op_name= "assign_to_keycache";
|
|
param->db_name= table->s->db.str;
|
|
param->table_name= table->s->table_name.str;
|
|
param->testflag= 0;
|
|
mi_check_print_error(param, errmsg);
|
|
}
|
|
DBUG_RETURN(error);
|
|
}
|
|
|
|
|
|
/*
|
|
Preload pages of the index file for a table into the key cache.
|
|
*/
|
|
|
|
int ha_myisam::preload_keys(THD* thd, HA_CHECK_OPT *check_opt)
|
|
{
|
|
int error;
|
|
const char *errmsg;
|
|
ulonglong map;
|
|
TABLE_LIST *table_list= table->pos_in_table_list;
|
|
my_bool ignore_leaves= table_list->ignore_leaves;
|
|
char buf[MYSQL_ERRMSG_SIZE];
|
|
|
|
DBUG_ENTER("ha_myisam::preload_keys");
|
|
|
|
table->keys_in_use_for_query.clear_all();
|
|
|
|
if (table_list->process_index_hints(table))
|
|
DBUG_RETURN(HA_ADMIN_FAILED);
|
|
|
|
map= ~(ulonglong) 0;
|
|
/* Check validity of the index references */
|
|
if (!table->keys_in_use_for_query.is_clear_all())
|
|
/* use all keys if there's no list specified by the user through hints */
|
|
map= table->keys_in_use_for_query.to_ulonglong();
|
|
|
|
mi_extra(file, HA_EXTRA_PRELOAD_BUFFER_SIZE,
|
|
(void *) &thd->variables.preload_buff_size);
|
|
|
|
if ((error= mi_preload(file, map, ignore_leaves)))
|
|
{
|
|
switch (error) {
|
|
case HA_ERR_NON_UNIQUE_BLOCK_SIZE:
|
|
errmsg= "Indexes use different block sizes";
|
|
break;
|
|
case HA_ERR_OUT_OF_MEM:
|
|
errmsg= "Failed to allocate buffer";
|
|
break;
|
|
default:
|
|
my_snprintf(buf, sizeof(buf),
|
|
"Failed to read from index file (errno: %d)", my_errno);
|
|
errmsg= buf;
|
|
}
|
|
error= HA_ADMIN_FAILED;
|
|
goto err;
|
|
}
|
|
|
|
DBUG_RETURN(HA_ADMIN_OK);
|
|
|
|
err:
|
|
{
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
if (!param)
|
|
return HA_ADMIN_INTERNAL_ERROR;
|
|
myisamchk_init(param);
|
|
param->thd= thd;
|
|
param->op_name= "preload_keys";
|
|
param->db_name= table->s->db.str;
|
|
param->table_name= table->s->table_name.str;
|
|
param->testflag= 0;
|
|
mi_check_print_error(param, errmsg);
|
|
DBUG_RETURN(error);
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
Disable indexes, making it persistent if requested.
|
|
|
|
SYNOPSIS
|
|
disable_indexes()
|
|
mode mode of operation:
|
|
HA_KEY_SWITCH_NONUNIQ disable all non-unique keys
|
|
HA_KEY_SWITCH_ALL disable all keys
|
|
HA_KEY_SWITCH_NONUNIQ_SAVE dis. non-uni. and make persistent
|
|
HA_KEY_SWITCH_ALL_SAVE dis. all keys and make persistent
|
|
|
|
IMPLEMENTATION
|
|
HA_KEY_SWITCH_NONUNIQ is not implemented.
|
|
HA_KEY_SWITCH_ALL_SAVE is not implemented.
|
|
|
|
RETURN
|
|
0 ok
|
|
HA_ERR_WRONG_COMMAND mode not implemented.
|
|
*/
|
|
|
|
int ha_myisam::disable_indexes(uint mode)
|
|
{
|
|
int error;
|
|
|
|
if (mode == HA_KEY_SWITCH_ALL)
|
|
{
|
|
/* call a storage engine function to switch the key map */
|
|
error= mi_disable_indexes(file);
|
|
}
|
|
else if (mode == HA_KEY_SWITCH_NONUNIQ_SAVE)
|
|
{
|
|
mi_extra(file, HA_EXTRA_NO_KEYS, 0);
|
|
info(HA_STATUS_CONST); // Read new key info
|
|
error= 0;
|
|
}
|
|
else
|
|
{
|
|
/* mode not implemented */
|
|
error= HA_ERR_WRONG_COMMAND;
|
|
}
|
|
return error;
|
|
}
|
|
|
|
|
|
/*
|
|
Enable indexes, making it persistent if requested.
|
|
|
|
SYNOPSIS
|
|
enable_indexes()
|
|
mode mode of operation:
|
|
HA_KEY_SWITCH_NONUNIQ enable all non-unique keys
|
|
HA_KEY_SWITCH_ALL enable all keys
|
|
HA_KEY_SWITCH_NONUNIQ_SAVE en. non-uni. and make persistent
|
|
HA_KEY_SWITCH_ALL_SAVE en. all keys and make persistent
|
|
|
|
DESCRIPTION
|
|
Enable indexes, which might have been disabled by disable_index() before.
|
|
The modes without _SAVE work only if both data and indexes are empty,
|
|
since the MyISAM repair would enable them persistently.
|
|
To be sure in these cases, call handler::delete_all_rows() before.
|
|
|
|
IMPLEMENTATION
|
|
HA_KEY_SWITCH_NONUNIQ is not implemented.
|
|
HA_KEY_SWITCH_ALL_SAVE is not implemented.
|
|
|
|
RETURN
|
|
0 ok
|
|
!=0 Error, among others:
|
|
HA_ERR_CRASHED data or index is non-empty. Delete all rows and retry.
|
|
HA_ERR_WRONG_COMMAND mode not implemented.
|
|
*/
|
|
|
|
int ha_myisam::enable_indexes(uint mode)
|
|
{
|
|
int error;
|
|
DBUG_ENTER("ha_myisam::enable_indexes");
|
|
|
|
DBUG_EXECUTE_IF("wait_in_enable_indexes",
|
|
debug_wait_for_kill("wait_in_enable_indexes"); );
|
|
|
|
if (mi_is_all_keys_active(file->s->state.key_map, file->s->base.keys))
|
|
{
|
|
/* All indexes are enabled already. */
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
if (mode == HA_KEY_SWITCH_ALL)
|
|
{
|
|
error= mi_enable_indexes(file);
|
|
/*
|
|
Do not try to repair on error,
|
|
as this could make the enabled state persistent,
|
|
but mode==HA_KEY_SWITCH_ALL forbids it.
|
|
*/
|
|
}
|
|
else if (mode == HA_KEY_SWITCH_NONUNIQ_SAVE)
|
|
{
|
|
THD *thd= table->in_use;
|
|
int was_error= thd->is_error();
|
|
HA_CHECK *param= (HA_CHECK*) thd->alloc(sizeof *param);
|
|
const char *save_proc_info=thd->proc_info;
|
|
|
|
if (!param)
|
|
DBUG_RETURN(HA_ADMIN_INTERNAL_ERROR);
|
|
|
|
thd_proc_info(thd, "Creating index");
|
|
myisamchk_init(param);
|
|
param->op_name= "recreating_index";
|
|
param->testflag= (T_SILENT | T_REP_BY_SORT | T_QUICK |
|
|
T_CREATE_MISSING_KEYS);
|
|
/*
|
|
Don't lock and unlock table if it's locked.
|
|
Normally table should be locked. This test is mostly for safety.
|
|
*/
|
|
if (likely(file->lock_type != F_UNLCK))
|
|
param->testflag|= T_NO_LOCKS;
|
|
|
|
if (file->create_unique_index_by_sort)
|
|
param->testflag|= T_CREATE_UNIQUE_BY_SORT;
|
|
|
|
param->myf_rw&= ~MY_WAIT_IF_FULL;
|
|
param->sort_buffer_length= THDVAR(thd, sort_buffer_size);
|
|
param->stats_method= (enum_handler_stats_method)THDVAR(thd, stats_method);
|
|
param->tmpdir=&mysql_tmpdir_list;
|
|
|
|
setup_vcols_for_repair(param);
|
|
|
|
if ((error= (repair(thd,*param,0) != HA_ADMIN_OK)) && param->retry_repair)
|
|
{
|
|
sql_print_warning("Warning: Enabling keys got errno %d on %s.%s, retrying",
|
|
my_errno, param->db_name, param->table_name);
|
|
/*
|
|
Repairing by sort failed. Now try standard repair method.
|
|
Still we want to fix only index file. If data file corruption
|
|
was detected (T_RETRY_WITHOUT_QUICK), we shouldn't do much here.
|
|
Let implicit repair do this job.
|
|
*/
|
|
if (!(param->testflag & T_RETRY_WITHOUT_QUICK))
|
|
{
|
|
param->testflag&= ~T_REP_BY_SORT;
|
|
error= (repair(thd,*param,0) != HA_ADMIN_OK);
|
|
}
|
|
/*
|
|
If the standard repair succeeded, clear all error messages which
|
|
might have been set by the first repair. They can still be seen
|
|
with SHOW WARNINGS then.
|
|
*/
|
|
if (! error && ! was_error)
|
|
thd->clear_error();
|
|
}
|
|
info(HA_STATUS_CONST);
|
|
thd_proc_info(thd, save_proc_info);
|
|
|
|
restore_vcos_after_repair();
|
|
}
|
|
else
|
|
{
|
|
/* mode not implemented */
|
|
error= HA_ERR_WRONG_COMMAND;
|
|
}
|
|
DBUG_RETURN(error);
|
|
}
|
|
|
|
|
|
/*
|
|
Test if indexes are disabled.
|
|
|
|
|
|
SYNOPSIS
|
|
indexes_are_disabled()
|
|
no parameters
|
|
|
|
|
|
RETURN
|
|
0 indexes are not disabled
|
|
1 all indexes are disabled
|
|
[2 non-unique indexes are disabled - NOT YET IMPLEMENTED]
|
|
*/
|
|
|
|
int ha_myisam::indexes_are_disabled(void)
|
|
{
|
|
|
|
return mi_indexes_are_disabled(file);
|
|
}
|
|
|
|
|
|
/*
|
|
prepare for a many-rows insert operation
|
|
e.g. - disable indexes (if they can be recreated fast) or
|
|
activate special bulk-insert optimizations
|
|
|
|
SYNOPSIS
|
|
start_bulk_insert(rows, flags)
|
|
rows Rows to be inserted
|
|
0 if we don't know
|
|
flags Flags to control index creation
|
|
|
|
NOTICE
|
|
Do not forget to call end_bulk_insert() later!
|
|
*/
|
|
|
|
void ha_myisam::start_bulk_insert(ha_rows rows, uint flags)
|
|
{
|
|
DBUG_ENTER("ha_myisam::start_bulk_insert");
|
|
THD *thd= table->in_use;
|
|
ulong size= MY_MIN(thd->variables.read_buff_size,
|
|
(ulong) (table->s->avg_row_length*rows));
|
|
bool index_disabled= 0;
|
|
DBUG_PRINT("info",("start_bulk_insert: rows %lu size %lu",
|
|
(ulong) rows, size));
|
|
|
|
/* don't enable row cache if too few rows */
|
|
if ((!rows || rows > MI_MIN_ROWS_TO_USE_WRITE_CACHE) && !has_long_unique())
|
|
mi_extra(file, HA_EXTRA_WRITE_CACHE, (void*) &size);
|
|
|
|
can_enable_indexes= mi_is_all_keys_active(file->s->state.key_map,
|
|
file->s->base.keys);
|
|
|
|
/*
|
|
Only disable old index if the table was empty and we are inserting
|
|
a lot of rows.
|
|
Note that in end_bulk_insert() we may truncate the table if
|
|
enable_indexes() failed, thus it's essential that indexes are
|
|
disabled ONLY for an empty table.
|
|
*/
|
|
if (file->state->records == 0 && can_enable_indexes &&
|
|
(!rows || rows >= MI_MIN_ROWS_TO_DISABLE_INDEXES))
|
|
{
|
|
if (file->open_flag & HA_OPEN_INTERNAL_TABLE)
|
|
{
|
|
file->update|= HA_STATE_CHANGED;
|
|
index_disabled= file->s->base.keys > 0;
|
|
mi_clear_all_keys_active(file->s->state.key_map);
|
|
}
|
|
else
|
|
{
|
|
my_bool all_keys= MY_TEST(flags & HA_CREATE_UNIQUE_INDEX_BY_SORT);
|
|
MYISAM_SHARE *share=file->s;
|
|
MI_KEYDEF *key=share->keyinfo;
|
|
uint i;
|
|
/*
|
|
Deactivate all indexes that can be recreated fast.
|
|
These include packed keys on which sorting will use more temporary
|
|
space than the max allowed file length or for which the unpacked keys
|
|
will take much more space than packed keys.
|
|
Note that 'rows' may be zero for the case when we don't know how many
|
|
rows we will put into the file.
|
|
Long Unique Index (HA_KEY_ALG_LONG_HASH) will not be disabled because
|
|
there unique property is enforced at the time of ha_write_row
|
|
(check_duplicate_long_entries). So we need active index at the time of
|
|
insert.
|
|
*/
|
|
DBUG_ASSERT(file->state->records == 0 &&
|
|
(!rows || rows >= MI_MIN_ROWS_TO_DISABLE_INDEXES));
|
|
for (i=0 ; i < share->base.keys ; i++,key++)
|
|
{
|
|
if (!(key->flag & (HA_SPATIAL | HA_AUTO_KEY)) &&
|
|
! mi_too_big_key_for_sort(key,rows) && file->s->base.auto_key != i+1 &&
|
|
(all_keys || !(key->flag & HA_NOSAME)) &&
|
|
table->key_info[i].algorithm != HA_KEY_ALG_LONG_HASH)
|
|
{
|
|
mi_clear_key_active(share->state.key_map, i);
|
|
index_disabled= 1;
|
|
file->update|= HA_STATE_CHANGED;
|
|
file->create_unique_index_by_sort= all_keys;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (!file->bulk_insert &&
|
|
(!rows || rows >= MI_MIN_ROWS_TO_USE_BULK_INSERT))
|
|
{
|
|
mi_init_bulk_insert(file, (size_t) thd->variables.bulk_insert_buff_size,
|
|
rows);
|
|
}
|
|
}
|
|
can_enable_indexes= index_disabled;
|
|
DBUG_VOID_RETURN;
|
|
}
|
|
|
|
/*
|
|
end special bulk-insert optimizations,
|
|
which have been activated by start_bulk_insert().
|
|
|
|
SYNOPSIS
|
|
end_bulk_insert(fatal_error)
|
|
abort 0 normal end, store everything
|
|
1 abort quickly. No need to flush/write anything. Table will be deleted
|
|
|
|
RETURN
|
|
0 OK
|
|
!= 0 Error
|
|
*/
|
|
|
|
int ha_myisam::end_bulk_insert()
|
|
{
|
|
int first_error, error;
|
|
my_bool abort= file->s->deleting;
|
|
DBUG_ENTER("ha_myisam::end_bulk_insert");
|
|
|
|
if ((first_error= mi_end_bulk_insert(file, abort)))
|
|
abort= 1;
|
|
|
|
if ((error= mi_extra(file, HA_EXTRA_NO_CACHE, 0)))
|
|
{
|
|
first_error= first_error ? first_error : error;
|
|
abort= 1;
|
|
}
|
|
|
|
if (!abort)
|
|
{
|
|
if (can_enable_indexes)
|
|
{
|
|
/*
|
|
Truncate the table when enable index operation is killed.
|
|
After truncating the table we don't need to enable the
|
|
indexes, because the last repair operation is aborted after
|
|
setting the indexes as active and trying to recreate them.
|
|
*/
|
|
|
|
if (((first_error= enable_indexes(HA_KEY_SWITCH_NONUNIQ_SAVE)) != 0) &&
|
|
table->in_use->killed)
|
|
{
|
|
delete_all_rows();
|
|
/* not crashed, despite being killed during repair */
|
|
file->s->state.changed&= ~(STATE_CRASHED|STATE_CRASHED_ON_REPAIR);
|
|
}
|
|
}
|
|
can_enable_indexes= 0;
|
|
}
|
|
DBUG_PRINT("exit", ("first_error: %d", first_error));
|
|
DBUG_RETURN(first_error);
|
|
}
|
|
|
|
|
|
bool ha_myisam::check_and_repair(THD *thd)
|
|
{
|
|
int error=0;
|
|
int marked_crashed;
|
|
HA_CHECK_OPT check_opt;
|
|
DBUG_ENTER("ha_myisam::check_and_repair");
|
|
|
|
check_opt.init();
|
|
check_opt.flags= T_MEDIUM | T_AUTO_REPAIR;
|
|
// Don't use quick if deleted rows
|
|
if (!file->state->del && (myisam_recover_options & HA_RECOVER_QUICK))
|
|
check_opt.flags|=T_QUICK;
|
|
sql_print_warning("Checking table: '%s'",table->s->path.str);
|
|
|
|
const CSET_STRING query_backup= thd->query_string;
|
|
thd->set_query((char*) table->s->table_name.str,
|
|
(uint) table->s->table_name.length, system_charset_info);
|
|
|
|
if ((marked_crashed= mi_is_crashed(file)) || check(thd, &check_opt))
|
|
{
|
|
bool save_log_all_errors;
|
|
sql_print_warning("Recovering table: '%s'",table->s->path.str);
|
|
save_log_all_errors= thd->log_all_errors;
|
|
thd->log_all_errors|= (thd->variables.log_warnings > 2);
|
|
if (myisam_recover_options & HA_RECOVER_FULL_BACKUP)
|
|
{
|
|
char buff[MY_BACKUP_NAME_EXTRA_LENGTH+1];
|
|
my_create_backup_name(buff, "", check_opt.start_time);
|
|
sql_print_information("Making backup of index file %s with extension '%s'",
|
|
file->s->index_file_name, buff);
|
|
mi_make_backup_of_index(file, check_opt.start_time,
|
|
MYF(MY_WME | ME_WARNING));
|
|
}
|
|
check_opt.flags=
|
|
(((myisam_recover_options &
|
|
(HA_RECOVER_BACKUP | HA_RECOVER_FULL_BACKUP)) ? T_BACKUP_DATA : 0) |
|
|
(marked_crashed ? 0 : T_QUICK) |
|
|
(myisam_recover_options & HA_RECOVER_FORCE ? 0 : T_SAFE_REPAIR) |
|
|
T_AUTO_REPAIR);
|
|
if (repair(thd, &check_opt))
|
|
error=1;
|
|
thd->log_all_errors= save_log_all_errors;
|
|
}
|
|
thd->set_query(query_backup);
|
|
DBUG_RETURN(error);
|
|
}
|
|
|
|
bool ha_myisam::is_crashed() const
|
|
{
|
|
return (file->s->state.changed & STATE_CRASHED ||
|
|
(my_disable_locking && file->s->state.open_count));
|
|
}
|
|
|
|
int ha_myisam::update_row(const uchar *old_data, const uchar *new_data)
|
|
{
|
|
return mi_update(file,old_data,new_data);
|
|
}
|
|
|
|
int ha_myisam::delete_row(const uchar *buf)
|
|
{
|
|
return mi_delete(file,buf);
|
|
}
|
|
|
|
|
|
int ha_myisam::index_init(uint idx, bool sorted)
|
|
{
|
|
active_index=idx;
|
|
if (pushed_idx_cond_keyno == idx)
|
|
mi_set_index_cond_func(file, handler_index_cond_check, this);
|
|
if (pushed_rowid_filter)
|
|
mi_set_rowid_filter_func(file, handler_rowid_filter_check,
|
|
handler_rowid_filter_is_active, this);
|
|
return 0;
|
|
}
|
|
|
|
|
|
int ha_myisam::index_end()
|
|
{
|
|
DBUG_ENTER("ha_myisam::index_end");
|
|
active_index=MAX_KEY;
|
|
//pushed_idx_cond_keyno= MAX_KEY;
|
|
mi_set_index_cond_func(file, NULL, 0);
|
|
in_range_check_pushed_down= FALSE;
|
|
mi_set_rowid_filter_func(file, NULL, NULL, 0);
|
|
ds_mrr.dsmrr_close();
|
|
#if !defined(DBUG_OFF) && defined(SQL_SELECT_FIXED_FOR_UPDATE)
|
|
file->update&= ~HA_STATE_AKTIV; // Forget active row
|
|
#endif
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
int ha_myisam::rnd_end()
|
|
{
|
|
DBUG_ENTER("ha_myisam::rnd_end");
|
|
ds_mrr.dsmrr_close();
|
|
#if !defined(DBUG_OFF) && defined(SQL_SELECT_FIXED_FOR_UPDATE)
|
|
file->update&= ~HA_STATE_AKTIV; // Forget active row
|
|
#endif
|
|
DBUG_RETURN(0);
|
|
}
|
|
|
|
int ha_myisam::index_read_map(uchar *buf, const uchar *key,
|
|
key_part_map keypart_map,
|
|
enum ha_rkey_function find_flag)
|
|
{
|
|
DBUG_ASSERT(inited==INDEX);
|
|
int error=mi_rkey(file, buf, active_index, key, keypart_map, find_flag);
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::index_read_idx_map(uchar *buf, uint index, const uchar *key,
|
|
key_part_map keypart_map,
|
|
enum ha_rkey_function find_flag)
|
|
{
|
|
int res;
|
|
/* Use the pushed index condition if it matches the index we're scanning */
|
|
end_range= NULL;
|
|
if (index == pushed_idx_cond_keyno)
|
|
mi_set_index_cond_func(file, handler_index_cond_check, this);
|
|
if (pushed_rowid_filter)
|
|
mi_set_rowid_filter_func(file, handler_rowid_filter_check,
|
|
handler_rowid_filter_is_active, this);
|
|
res= mi_rkey(file, buf, index, key, keypart_map, find_flag);
|
|
mi_set_index_cond_func(file, NULL, 0);
|
|
return res;
|
|
}
|
|
|
|
int ha_myisam::index_next(uchar *buf)
|
|
{
|
|
DBUG_ASSERT(inited==INDEX);
|
|
int error=mi_rnext(file,buf,active_index);
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::index_prev(uchar *buf)
|
|
{
|
|
DBUG_ASSERT(inited==INDEX);
|
|
int error=mi_rprev(file,buf, active_index);
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::index_first(uchar *buf)
|
|
{
|
|
DBUG_ASSERT(inited==INDEX);
|
|
int error=mi_rfirst(file, buf, active_index);
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::index_last(uchar *buf)
|
|
{
|
|
DBUG_ASSERT(inited==INDEX);
|
|
int error=mi_rlast(file, buf, active_index);
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::index_next_same(uchar *buf,
|
|
const uchar *key __attribute__((unused)),
|
|
uint length __attribute__((unused)))
|
|
{
|
|
int error;
|
|
DBUG_ASSERT(inited==INDEX);
|
|
do
|
|
{
|
|
error= mi_rnext_same(file,buf);
|
|
} while (error == HA_ERR_RECORD_DELETED);
|
|
return error;
|
|
}
|
|
|
|
|
|
int ha_myisam::rnd_init(bool scan)
|
|
{
|
|
if (scan)
|
|
return mi_scan_init(file);
|
|
return mi_reset(file); // Free buffers
|
|
}
|
|
|
|
int ha_myisam::rnd_next(uchar *buf)
|
|
{
|
|
int error=mi_scan(file, buf);
|
|
return error;
|
|
}
|
|
|
|
int ha_myisam::remember_rnd_pos()
|
|
{
|
|
position((uchar*) 0);
|
|
return 0;
|
|
}
|
|
|
|
int ha_myisam::restart_rnd_next(uchar *buf)
|
|
{
|
|
return rnd_pos(buf, ref);
|
|
}
|
|
|
|
int ha_myisam::rnd_pos(uchar *buf, uchar *pos)
|
|
{
|
|
int error=mi_rrnd(file, buf, my_get_ptr(pos,ref_length));
|
|
return error;
|
|
}
|
|
|
|
void ha_myisam::position(const uchar *record)
|
|
{
|
|
my_off_t row_position= mi_position(file);
|
|
my_store_ptr(ref, ref_length, row_position);
|
|
file->update|= HA_STATE_AKTIV; // Row can be updated
|
|
}
|
|
|
|
int ha_myisam::info(uint flag)
|
|
{
|
|
MI_ISAMINFO misam_info;
|
|
|
|
if (!table)
|
|
return 1;
|
|
|
|
(void) mi_status(file,&misam_info,flag);
|
|
if (flag & HA_STATUS_VARIABLE)
|
|
{
|
|
stats.records= misam_info.records;
|
|
stats.deleted= misam_info.deleted;
|
|
stats.data_file_length= misam_info.data_file_length;
|
|
stats.index_file_length= misam_info.index_file_length;
|
|
stats.delete_length= misam_info.delete_length;
|
|
stats.check_time= (ulong) misam_info.check_time;
|
|
stats.mean_rec_length= misam_info.mean_reclength;
|
|
stats.checksum= file->state->checksum;
|
|
}
|
|
if (flag & HA_STATUS_CONST)
|
|
{
|
|
TABLE_SHARE *share= table->s;
|
|
stats.max_data_file_length= misam_info.max_data_file_length;
|
|
stats.max_index_file_length= misam_info.max_index_file_length;
|
|
stats.create_time= (ulong) misam_info.create_time;
|
|
/*
|
|
We want the value of stats.mrr_length_per_rec to be platform independent.
|
|
The size of the chunk at the end of the join buffer used for MRR needs
|
|
is calculated now basing on the values passed in the stats structure.
|
|
The remaining part of the join buffer is used for records. A different
|
|
number of records in the buffer results in a different number of buffer
|
|
refills and in a different order of records in the result set.
|
|
*/
|
|
stats.mrr_length_per_rec= misam_info.reflength + 8; // 8=MY_MAX(sizeof(void *))
|
|
|
|
ref_length= misam_info.reflength;
|
|
share->db_options_in_use= misam_info.options;
|
|
/* record block size. We adjust with IO_SIZE to not make it too small */
|
|
stats.block_size= MY_MAX(myisam_block_size, IO_SIZE);
|
|
|
|
if (table_share->tmp_table == NO_TMP_TABLE)
|
|
mysql_mutex_lock(&table_share->LOCK_share);
|
|
share->keys_in_use.set_prefix(share->keys);
|
|
share->keys_in_use.intersect_extended(misam_info.key_map);
|
|
share->keys_for_keyread.intersect(share->keys_in_use);
|
|
share->db_record_offset= misam_info.record_offset;
|
|
if (share->key_parts)
|
|
{
|
|
ulong *from= misam_info.rec_per_key;
|
|
KEY *key, *key_end;
|
|
for (key= table->key_info, key_end= key + share->keys;
|
|
key < key_end ; key++)
|
|
{
|
|
memcpy(key->rec_per_key, from,
|
|
key->user_defined_key_parts * sizeof(*from));
|
|
from+= key->user_defined_key_parts;
|
|
}
|
|
}
|
|
if (table_share->tmp_table == NO_TMP_TABLE)
|
|
mysql_mutex_unlock(&table_share->LOCK_share);
|
|
}
|
|
if (flag & HA_STATUS_ERRKEY)
|
|
{
|
|
errkey = misam_info.errkey;
|
|
my_store_ptr(dup_ref, ref_length, misam_info.dupp_key_pos);
|
|
}
|
|
if (flag & HA_STATUS_TIME)
|
|
stats.update_time = (ulong) misam_info.update_time;
|
|
if (flag & HA_STATUS_AUTO)
|
|
stats.auto_increment_value= misam_info.auto_increment;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
int ha_myisam::extra(enum ha_extra_function operation)
|
|
{
|
|
if ((operation == HA_EXTRA_MMAP && !opt_myisam_use_mmap) ||
|
|
(operation == HA_EXTRA_WRITE_CACHE && has_long_unique()))
|
|
return 0;
|
|
return mi_extra(file, operation, 0);
|
|
}
|
|
|
|
|
|
int ha_myisam::reset(void)
|
|
{
|
|
mi_set_index_cond_func(file, NULL, 0);
|
|
ds_mrr.dsmrr_close();
|
|
return mi_reset(file);
|
|
}
|
|
|
|
/* To be used with WRITE_CACHE and EXTRA_CACHE */
|
|
|
|
int ha_myisam::extra_opt(enum ha_extra_function operation, ulong cache_size)
|
|
{
|
|
return mi_extra(file, operation, (void*) &cache_size);
|
|
}
|
|
|
|
int ha_myisam::delete_all_rows()
|
|
{
|
|
return mi_delete_all_rows(file);
|
|
}
|
|
|
|
|
|
int ha_myisam::reset_auto_increment(ulonglong value)
|
|
{
|
|
file->s->state.auto_increment= value;
|
|
return 0;
|
|
}
|
|
|
|
int ha_myisam::delete_table(const char *name)
|
|
{
|
|
return mi_delete_table(name);
|
|
}
|
|
|
|
void ha_myisam::change_table_ptr(TABLE *table_arg, TABLE_SHARE *share)
|
|
{
|
|
handler::change_table_ptr(table_arg, share);
|
|
if (file)
|
|
file->external_ref= table_arg;
|
|
}
|
|
|
|
|
|
int ha_myisam::external_lock(THD *thd, int lock_type)
|
|
{
|
|
file->in_use.data= thd;
|
|
file->external_ref= (void*) table; // For mi_killed()
|
|
return mi_lock_database(file, !table->s->tmp_table ?
|
|
lock_type : ((lock_type == F_UNLCK) ?
|
|
F_UNLCK : F_EXTRA_LCK));
|
|
}
|
|
|
|
THR_LOCK_DATA **ha_myisam::store_lock(THD *thd,
|
|
THR_LOCK_DATA **to,
|
|
enum thr_lock_type lock_type)
|
|
{
|
|
if (lock_type != TL_IGNORE && file->lock.type == TL_UNLOCK)
|
|
file->lock.type=lock_type;
|
|
*to++= &file->lock;
|
|
return to;
|
|
}
|
|
|
|
void ha_myisam::update_create_info(HA_CREATE_INFO *create_info)
|
|
{
|
|
ha_myisam::info(HA_STATUS_AUTO | HA_STATUS_CONST);
|
|
if (!(create_info->used_fields & HA_CREATE_USED_AUTO))
|
|
{
|
|
create_info->auto_increment_value= stats.auto_increment_value;
|
|
}
|
|
create_info->data_file_name=data_file_name;
|
|
create_info->index_file_name=index_file_name;
|
|
}
|
|
|
|
|
|
int ha_myisam::create(const char *name, TABLE *table_arg,
|
|
HA_CREATE_INFO *ha_create_info)
|
|
{
|
|
int error;
|
|
uint create_flags= 0, record_count, i;
|
|
char buff[FN_REFLEN];
|
|
MI_KEYDEF *keydef;
|
|
MI_COLUMNDEF *recinfo;
|
|
MI_CREATE_INFO create_info;
|
|
TABLE_SHARE *share= table_arg->s;
|
|
uint options= share->db_options_in_use;
|
|
DBUG_ENTER("ha_myisam::create");
|
|
|
|
for (i= 0; i < share->virtual_fields && !create_flags; i++)
|
|
if (table_arg->vfield[i]->flags & PART_KEY_FLAG)
|
|
create_flags|= HA_CREATE_RELIES_ON_SQL_LAYER;
|
|
for (i= 0; i < share->keys && !create_flags; i++)
|
|
if (table_arg->key_info[i].flags & HA_USES_PARSER)
|
|
create_flags|= HA_CREATE_RELIES_ON_SQL_LAYER;
|
|
|
|
if ((error= table2myisam(table_arg, &keydef, &recinfo, &record_count)))
|
|
DBUG_RETURN(error); /* purecov: inspected */
|
|
|
|
#ifndef DBUG_OFF
|
|
DBUG_EXECUTE_IF("key",
|
|
Debug_key_myisam::print_keys_myisam(table_arg->in_use,
|
|
"ha_myisam::create: ",
|
|
table_arg, keydef, share->keys);
|
|
);
|
|
#endif
|
|
|
|
bzero((char*) &create_info, sizeof(create_info));
|
|
create_info.max_rows= share->max_rows;
|
|
create_info.reloc_rows= share->min_rows;
|
|
create_info.with_auto_increment= share->next_number_key_offset == 0;
|
|
create_info.auto_increment= (ha_create_info->auto_increment_value ?
|
|
ha_create_info->auto_increment_value -1 :
|
|
(ulonglong) 0);
|
|
create_info.data_file_length= ((ulonglong) share->max_rows *
|
|
share->avg_row_length);
|
|
create_info.language= share->table_charset->number;
|
|
|
|
#ifdef HAVE_READLINK
|
|
if (my_use_symdir)
|
|
{
|
|
create_info.data_file_name= ha_create_info->data_file_name;
|
|
create_info.index_file_name= ha_create_info->index_file_name;
|
|
}
|
|
else
|
|
#endif /* HAVE_READLINK */
|
|
{
|
|
THD *thd= table_arg->in_use;
|
|
if (ha_create_info->data_file_name)
|
|
push_warning_printf(thd, Sql_condition::WARN_LEVEL_WARN,
|
|
WARN_OPTION_IGNORED,
|
|
ER_THD(thd, WARN_OPTION_IGNORED),
|
|
"DATA DIRECTORY");
|
|
if (ha_create_info->index_file_name)
|
|
push_warning_printf(thd, Sql_condition::WARN_LEVEL_WARN,
|
|
WARN_OPTION_IGNORED,
|
|
ER_THD(thd, WARN_OPTION_IGNORED),
|
|
"INDEX DIRECTORY");
|
|
}
|
|
|
|
if (ha_create_info->tmp_table())
|
|
create_flags|= HA_CREATE_TMP_TABLE | HA_CREATE_DELAY_KEY_WRITE;
|
|
if (ha_create_info->options & HA_CREATE_KEEP_FILES)
|
|
create_flags|= HA_CREATE_KEEP_FILES;
|
|
if (options & HA_OPTION_PACK_RECORD)
|
|
create_flags|= HA_PACK_RECORD;
|
|
if (options & HA_OPTION_CHECKSUM)
|
|
create_flags|= HA_CREATE_CHECKSUM;
|
|
if (options & HA_OPTION_DELAY_KEY_WRITE)
|
|
create_flags|= HA_CREATE_DELAY_KEY_WRITE;
|
|
|
|
/* TODO: Check that the following fn_format is really needed */
|
|
error= mi_create(fn_format(buff, name, "", "",
|
|
MY_UNPACK_FILENAME|MY_APPEND_EXT),
|
|
share->keys, keydef,
|
|
record_count, recinfo,
|
|
0, (MI_UNIQUEDEF*) 0,
|
|
&create_info, create_flags);
|
|
my_free(recinfo);
|
|
DBUG_RETURN(error);
|
|
}
|
|
|
|
|
|
int ha_myisam::rename_table(const char * from, const char * to)
|
|
{
|
|
return mi_rename(from,to);
|
|
}
|
|
|
|
|
|
void ha_myisam::get_auto_increment(ulonglong offset, ulonglong increment,
|
|
ulonglong nb_desired_values,
|
|
ulonglong *first_value,
|
|
ulonglong *nb_reserved_values)
|
|
{
|
|
ulonglong nr;
|
|
int error;
|
|
uchar key[HA_MAX_KEY_LENGTH];
|
|
|
|
if (!table->s->next_number_key_offset)
|
|
{ // Autoincrement at key-start
|
|
ha_myisam::info(HA_STATUS_AUTO);
|
|
*first_value= stats.auto_increment_value;
|
|
/* MyISAM has only table-level lock, so reserves to +inf */
|
|
*nb_reserved_values= ULONGLONG_MAX;
|
|
return;
|
|
}
|
|
|
|
/* it's safe to call the following if bulk_insert isn't on */
|
|
mi_flush_bulk_insert(file, table->s->next_number_index);
|
|
|
|
(void) extra(HA_EXTRA_KEYREAD);
|
|
key_copy(key, table->record[0],
|
|
table->key_info + table->s->next_number_index,
|
|
table->s->next_number_key_offset);
|
|
error= mi_rkey(file, table->record[1], (int) table->s->next_number_index,
|
|
key, make_prev_keypart_map(table->s->next_number_keypart),
|
|
HA_READ_PREFIX_LAST);
|
|
if (error)
|
|
nr= 1;
|
|
else
|
|
{
|
|
/* Get data from record[1] */
|
|
nr= ((ulonglong) table->next_number_field->
|
|
val_int_offset(table->s->rec_buff_length)+1);
|
|
}
|
|
extra(HA_EXTRA_NO_KEYREAD);
|
|
*first_value= nr;
|
|
/*
|
|
MySQL needs to call us for next row: assume we are inserting ("a",null)
|
|
here, we return 3, and next this statement will want to insert ("b",null):
|
|
there is no reason why ("b",3+1) would be the good row to insert: maybe it
|
|
already exists, maybe 3+1 is too large...
|
|
*/
|
|
*nb_reserved_values= 1;
|
|
}
|
|
|
|
|
|
/*
|
|
Find out how many rows there is in the given range
|
|
|
|
SYNOPSIS
|
|
records_in_range()
|
|
inx Index to use
|
|
min_key Start of range. Null pointer if from first key
|
|
max_key End of range. Null pointer if to last key
|
|
pages Store first and last page for the range in case of
|
|
b-trees. In other cases it's not touched.
|
|
|
|
NOTES
|
|
min_key.flag can have one of the following values:
|
|
HA_READ_KEY_EXACT Include the key in the range
|
|
HA_READ_AFTER_KEY Don't include key in range
|
|
|
|
max_key.flag can have one of the following values:
|
|
HA_READ_BEFORE_KEY Don't include key in range
|
|
HA_READ_AFTER_KEY Include all 'end_key' values in the range
|
|
|
|
RETURN
|
|
HA_POS_ERROR Something is wrong with the index tree.
|
|
0 There is no matching keys in the given range
|
|
number > 0 There is approximately 'number' matching rows in
|
|
the range.
|
|
*/
|
|
|
|
ha_rows ha_myisam::records_in_range(uint inx, const key_range *min_key,
|
|
const key_range *max_key,
|
|
page_range *pages)
|
|
{
|
|
return (ha_rows) mi_records_in_range(file, (int) inx, min_key, max_key,
|
|
pages);
|
|
}
|
|
|
|
|
|
int ha_myisam::ft_read(uchar *buf)
|
|
{
|
|
int error;
|
|
|
|
if (!ft_handler)
|
|
return -1;
|
|
|
|
thread_safe_increment(table->in_use->status_var.ha_read_next_count,
|
|
&LOCK_status); // why ?
|
|
|
|
error=ft_handler->please->read_next(ft_handler,(char*) buf);
|
|
return error;
|
|
}
|
|
|
|
enum_alter_inplace_result
|
|
ha_myisam::check_if_supported_inplace_alter(TABLE *new_table,
|
|
Alter_inplace_info *alter_info)
|
|
{
|
|
DBUG_ENTER("ha_myisam::check_if_supported_inplace_alter");
|
|
|
|
const alter_table_operations readd_index=
|
|
ALTER_ADD_NON_UNIQUE_NON_PRIM_INDEX |
|
|
ALTER_DROP_NON_UNIQUE_NON_PRIM_INDEX;
|
|
const alter_table_operations readd_unique=
|
|
ALTER_ADD_UNIQUE_INDEX |
|
|
ALTER_DROP_UNIQUE_INDEX;
|
|
const alter_table_operations readd_pk=
|
|
ALTER_ADD_PK_INDEX |
|
|
ALTER_DROP_PK_INDEX;
|
|
|
|
const alter_table_operations op= alter_info->handler_flags;
|
|
|
|
if (op & ALTER_COLUMN_VCOL)
|
|
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
|
|
|
|
/*
|
|
ha_myisam::open() updates table->key_info->block_size to be the actual
|
|
MYI index block size, overwriting user-specified value (if any).
|
|
So, the server can not reliably detect whether ALTER TABLE changes
|
|
key_block_size or not, it might think the block size was changed,
|
|
when it wasn't, and in this case the server will recreate (drop+add)
|
|
the index unnecessary. Fix it.
|
|
*/
|
|
|
|
if (table->s->keys == new_table->s->keys &&
|
|
((op & readd_pk) == readd_pk ||
|
|
(op & readd_unique) == readd_unique ||
|
|
(op & readd_index) == readd_index))
|
|
{
|
|
for (uint i=0; i < table->s->keys; i++)
|
|
{
|
|
KEY *old_key= table->key_info + i;
|
|
KEY *new_key= new_table->key_info + i;
|
|
|
|
if (old_key->block_size == new_key->block_size)
|
|
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED); // must differ somewhere else
|
|
|
|
if (new_key->block_size && new_key->block_size != old_key->block_size)
|
|
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED); // really changed
|
|
|
|
/* any difference besides the block_size, and we give up */
|
|
if (old_key->key_length != new_key->key_length ||
|
|
old_key->flags != new_key->flags ||
|
|
old_key->user_defined_key_parts != new_key->user_defined_key_parts ||
|
|
old_key->algorithm != new_key->algorithm ||
|
|
strcmp(old_key->name.str, new_key->name.str))
|
|
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
|
|
|
|
for (uint j= 0; j < old_key->user_defined_key_parts; j++)
|
|
{
|
|
KEY_PART_INFO *old_kp= old_key->key_part + j;
|
|
KEY_PART_INFO *new_kp= new_key->key_part + j;
|
|
if (old_kp->offset != new_kp->offset ||
|
|
old_kp->null_offset != new_kp->null_offset ||
|
|
old_kp->length != new_kp->length ||
|
|
old_kp->fieldnr != new_kp->fieldnr ||
|
|
old_kp->key_part_flag != new_kp->key_part_flag ||
|
|
old_kp->type != new_kp->type ||
|
|
old_kp->null_bit != new_kp->null_bit)
|
|
DBUG_RETURN(HA_ALTER_INPLACE_NOT_SUPPORTED);
|
|
}
|
|
}
|
|
alter_info->handler_flags &= ~(readd_pk | readd_unique | readd_index);
|
|
}
|
|
DBUG_RETURN(handler::check_if_supported_inplace_alter(new_table, alter_info));
|
|
}
|
|
|
|
|
|
static bool directories_differ(const char *d1, const char *d2)
|
|
{
|
|
if (!d1 && !d2)
|
|
return false;
|
|
if (!d1 || !d2)
|
|
return true;
|
|
size_t l1= dirname_length(d1), l2= dirname_length(d2);
|
|
return l1 != l2 || strncmp(d1, d2, l1);
|
|
}
|
|
|
|
|
|
bool ha_myisam::check_if_incompatible_data(HA_CREATE_INFO *create_info,
|
|
uint table_changes)
|
|
{
|
|
uint options= table->s->db_options_in_use;
|
|
|
|
if ((create_info->used_fields & HA_CREATE_USED_AUTO &&
|
|
create_info->auto_increment_value != stats.auto_increment_value) ||
|
|
directories_differ(create_info->data_file_name, data_file_name) ||
|
|
directories_differ(create_info->index_file_name, index_file_name) ||
|
|
table_changes == IS_EQUAL_NO ||
|
|
table_changes & IS_EQUAL_PACK_LENGTH) // Not implemented yet
|
|
return COMPATIBLE_DATA_NO;
|
|
|
|
if ((options & (HA_OPTION_PACK_RECORD | HA_OPTION_CHECKSUM)) !=
|
|
(create_info->table_options & (HA_OPTION_PACK_RECORD | HA_OPTION_CHECKSUM)))
|
|
return COMPATIBLE_DATA_NO;
|
|
return COMPATIBLE_DATA_YES;
|
|
}
|
|
|
|
|
|
/**
|
|
Check if a table is incompatible with the current version.
|
|
|
|
The cases are:
|
|
- Table has checksum, varchars and are not of dynamic record type
|
|
*/
|
|
|
|
int ha_myisam::check_for_upgrade(HA_CHECK_OPT *check_opt)
|
|
{
|
|
if ((file->s->options & HA_OPTION_CHECKSUM) &&
|
|
!(file->s->options & HA_OPTION_NULL_FIELDS) &&
|
|
!(file->s->options & HA_OPTION_PACK_RECORD) &&
|
|
file->s->has_varchar_fields)
|
|
{
|
|
/* We need alter there to get the HA_OPTION_NULL_FIELDS flag to be set */
|
|
return HA_ADMIN_NEEDS_ALTER;
|
|
}
|
|
return HA_ADMIN_OK;
|
|
}
|
|
|
|
|
|
extern int mi_panic(enum ha_panic_function flag);
|
|
int myisam_panic(handlerton *hton, ha_panic_function flag)
|
|
{
|
|
return mi_panic(flag);
|
|
}
|
|
|
|
static int myisam_drop_table(handlerton *hton, const char *path)
|
|
{
|
|
return mi_delete_table(path);
|
|
}
|
|
|
|
static int myisam_init(void *p)
|
|
{
|
|
handlerton *hton;
|
|
|
|
init_myisam_psi_keys();
|
|
|
|
/* Set global variables based on startup options */
|
|
if (myisam_recover_options && myisam_recover_options != HA_RECOVER_OFF)
|
|
ha_open_options|=HA_OPEN_ABORT_IF_CRASHED;
|
|
else
|
|
myisam_recover_options= HA_RECOVER_OFF;
|
|
|
|
myisam_block_size=(uint) 1 << my_bit_log2_uint64(opt_myisam_block_size);
|
|
|
|
hton= (handlerton *)p;
|
|
hton->db_type= DB_TYPE_MYISAM;
|
|
hton->create= myisam_create_handler;
|
|
hton->drop_table= myisam_drop_table;
|
|
hton->panic= myisam_panic;
|
|
hton->flags= HTON_CAN_RECREATE | HTON_SUPPORT_LOG_TABLES;
|
|
hton->tablefile_extensions= ha_myisam_exts;
|
|
mi_killed= mi_killed_in_mariadb;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct st_mysql_sys_var* myisam_sysvars[]= {
|
|
MYSQL_SYSVAR(block_size),
|
|
MYSQL_SYSVAR(data_pointer_size),
|
|
MYSQL_SYSVAR(max_sort_file_size),
|
|
MYSQL_SYSVAR(recover_options),
|
|
MYSQL_SYSVAR(repair_threads),
|
|
MYSQL_SYSVAR(sort_buffer_size),
|
|
MYSQL_SYSVAR(use_mmap),
|
|
MYSQL_SYSVAR(mmap_size),
|
|
MYSQL_SYSVAR(stats_method),
|
|
0
|
|
};
|
|
|
|
/****************************************************************************
|
|
* MyISAM MRR implementation: use DS-MRR
|
|
***************************************************************************/
|
|
|
|
int ha_myisam::multi_range_read_init(RANGE_SEQ_IF *seq, void *seq_init_param,
|
|
uint n_ranges, uint mode,
|
|
HANDLER_BUFFER *buf)
|
|
{
|
|
return ds_mrr.dsmrr_init(this, seq, seq_init_param, n_ranges, mode, buf);
|
|
}
|
|
|
|
int ha_myisam::multi_range_read_next(range_id_t *range_info)
|
|
{
|
|
return ds_mrr.dsmrr_next(range_info);
|
|
}
|
|
|
|
ha_rows ha_myisam::multi_range_read_info_const(uint keyno, RANGE_SEQ_IF *seq,
|
|
void *seq_init_param,
|
|
uint n_ranges, uint *bufsz,
|
|
uint *flags, Cost_estimate *cost)
|
|
{
|
|
/*
|
|
This call is here because there is no location where this->table would
|
|
already be known.
|
|
TODO: consider moving it into some per-query initialization call.
|
|
*/
|
|
ds_mrr.init(this, table);
|
|
return ds_mrr.dsmrr_info_const(keyno, seq, seq_init_param, n_ranges, bufsz,
|
|
flags, cost);
|
|
}
|
|
|
|
ha_rows ha_myisam::multi_range_read_info(uint keyno, uint n_ranges, uint keys,
|
|
uint key_parts, uint *bufsz,
|
|
uint *flags, Cost_estimate *cost)
|
|
{
|
|
ds_mrr.init(this, table);
|
|
return ds_mrr.dsmrr_info(keyno, n_ranges, keys, key_parts, bufsz, flags, cost);
|
|
}
|
|
|
|
|
|
int ha_myisam::multi_range_read_explain_info(uint mrr_mode, char *str,
|
|
size_t size)
|
|
{
|
|
return ds_mrr.dsmrr_explain_info(mrr_mode, str, size);
|
|
}
|
|
|
|
/* MyISAM MRR implementation ends */
|
|
|
|
|
|
/* Index condition pushdown implementation*/
|
|
|
|
|
|
Item *ha_myisam::idx_cond_push(uint keyno_arg, Item* idx_cond_arg)
|
|
{
|
|
/*
|
|
Check if the key contains a blob field. If it does then MyISAM
|
|
should not accept the pushed index condition since MyISAM will not
|
|
read the blob field from the index entry during evaluation of the
|
|
pushed index condition and the BLOB field might be part of the
|
|
range evaluation done by the ICP code.
|
|
*/
|
|
const KEY *key= &table_share->key_info[keyno_arg];
|
|
|
|
for (uint k= 0; k < key->user_defined_key_parts; ++k)
|
|
{
|
|
const KEY_PART_INFO *key_part= &key->key_part[k];
|
|
if (key_part->key_part_flag & HA_BLOB_PART)
|
|
{
|
|
/* Let the server handle the index condition */
|
|
return idx_cond_arg;
|
|
}
|
|
}
|
|
|
|
pushed_idx_cond_keyno= keyno_arg;
|
|
pushed_idx_cond= idx_cond_arg;
|
|
in_range_check_pushed_down= TRUE;
|
|
if (active_index == pushed_idx_cond_keyno)
|
|
mi_set_index_cond_func(file, handler_index_cond_check, this);
|
|
return NULL;
|
|
}
|
|
|
|
bool ha_myisam::rowid_filter_push(Rowid_filter* rowid_filter)
|
|
{
|
|
pushed_rowid_filter= rowid_filter;
|
|
mi_set_rowid_filter_func(file, handler_rowid_filter_check,
|
|
handler_rowid_filter_is_active, this);
|
|
return false;
|
|
}
|
|
|
|
struct st_mysql_storage_engine myisam_storage_engine=
|
|
{ MYSQL_HANDLERTON_INTERFACE_VERSION };
|
|
|
|
maria_declare_plugin(myisam)
|
|
{
|
|
MYSQL_STORAGE_ENGINE_PLUGIN,
|
|
&myisam_storage_engine,
|
|
"MyISAM",
|
|
"MySQL AB",
|
|
"Non-transactional engine with good performance and small data footprint",
|
|
PLUGIN_LICENSE_GPL,
|
|
myisam_init, /* Plugin Init */
|
|
NULL, /* Plugin Deinit */
|
|
0x0100, /* 1.0 */
|
|
NULL, /* status variables */
|
|
myisam_sysvars, /* system variables */
|
|
"1.0", /* string version */
|
|
MariaDB_PLUGIN_MATURITY_STABLE /* maturity */
|
|
}
|
|
maria_declare_plugin_end;
|
|
|
|
|
|
#ifdef HAVE_QUERY_CACHE
|
|
/**
|
|
@brief Register a named table with a call back function to the query cache.
|
|
|
|
@param thd The thread handle
|
|
@param table_key A pointer to the table name in the table cache
|
|
@param key_length The length of the table name
|
|
@param[out] engine_callback The pointer to the storage engine call back
|
|
function, currently 0
|
|
@param[out] engine_data Engine data will be set to 0.
|
|
|
|
@note Despite the name of this function, it is used to check each statement
|
|
before it is cached and not to register a table or callback function.
|
|
|
|
@see handler::register_query_cache_table
|
|
|
|
@return The error code. The engine_data and engine_callback will be set to 0.
|
|
@retval TRUE Success
|
|
@retval FALSE An error occurred
|
|
*/
|
|
|
|
my_bool ha_myisam::register_query_cache_table(THD *thd, const char *table_name,
|
|
uint table_name_len,
|
|
qc_engine_callback
|
|
*engine_callback,
|
|
ulonglong *engine_data)
|
|
{
|
|
DBUG_ENTER("ha_myisam::register_query_cache_table");
|
|
/*
|
|
No call back function is needed to determine if a cached statement
|
|
is valid or not.
|
|
*/
|
|
*engine_callback= 0;
|
|
|
|
/*
|
|
No engine data is needed.
|
|
*/
|
|
*engine_data= 0;
|
|
|
|
if (file->s->concurrent_insert)
|
|
{
|
|
/*
|
|
If a concurrent INSERT has happened just before the currently
|
|
processed SELECT statement, the total size of the table is
|
|
unknown.
|
|
|
|
To determine if the table size is known, the current thread's snap
|
|
shot of the table size with the actual table size are compared.
|
|
|
|
If the table size is unknown the SELECT statement can't be cached.
|
|
|
|
When concurrent inserts are disabled at table open, mi_ondopen()
|
|
does not assign a get_status() function. In this case the local
|
|
("current") status is never updated. We would wrongly think that
|
|
we cannot cache the statement.
|
|
*/
|
|
ulonglong actual_data_file_length;
|
|
ulonglong current_data_file_length;
|
|
|
|
/*
|
|
POSIX visibility rules specify that "2. Whatever memory values a
|
|
thread can see when it unlocks a mutex <...> can also be seen by any
|
|
thread that later locks the same mutex". In this particular case,
|
|
concurrent insert thread had modified the data_file_length in
|
|
MYISAM_SHARE before it has unlocked (or even locked)
|
|
structure_guard_mutex. So, here we're guaranteed to see at least that
|
|
value after we've locked the same mutex. We can see a later value
|
|
(modified by some other thread) though, but it's ok, as we only want
|
|
to know if the variable was changed, the actual new value doesn't matter
|
|
*/
|
|
actual_data_file_length= file->s->state.state.data_file_length;
|
|
current_data_file_length= file->save_state.data_file_length;
|
|
|
|
if (current_data_file_length != actual_data_file_length)
|
|
{
|
|
/* Don't cache current statement. */
|
|
DBUG_RETURN(FALSE);
|
|
}
|
|
}
|
|
|
|
/*
|
|
This query execution might have started after the query cache was flushed
|
|
by a concurrent INSERT. In this case, don't cache this statement as the
|
|
data file length difference might not be visible yet if the tables haven't
|
|
been unlocked by the concurrent insert thread.
|
|
*/
|
|
if (file->state->uncacheable)
|
|
DBUG_RETURN(FALSE);
|
|
|
|
/* It is ok to try to cache current statement. */
|
|
DBUG_RETURN(TRUE);
|
|
}
|
|
#endif
|