mirror of
https://github.com/MariaDB/server.git
synced 2025-01-16 12:02:42 +01:00
Merge quadxeon.mysql.com:/benchmarks/ext3/TOSAVE/tsmith/bk/50
into quadxeon.mysql.com:/benchmarks/ext3/TOSAVE/tsmith/bk/maint/50 strings/strtod.c: Auto merged
This commit is contained in:
commit
7d06ed8281
30 changed files with 357 additions and 310 deletions
|
@ -28,6 +28,7 @@ EXTRA_DIST = FINISH.sh \
|
|||
compile-alpha-debug \
|
||||
compile-amd64-debug-max \
|
||||
compile-amd64-max \
|
||||
compile-amd64-max-sci \
|
||||
compile-darwin-mwcc \
|
||||
compile-dist \
|
||||
compile-hpux11-parisc2-aCC \
|
||||
|
@ -53,6 +54,7 @@ EXTRA_DIST = FINISH.sh \
|
|||
compile-pentium-valgrind-max \
|
||||
compile-pentium64-debug \
|
||||
compile-pentium64-debug-max \
|
||||
compile-pentium64-max-sci \
|
||||
compile-pentium64-valgrind-max \
|
||||
compile-ppc \
|
||||
compile-ppc-debug \
|
||||
|
|
8
BUILD/compile-amd64-max-sci
Normal file
8
BUILD/compile-amd64-max-sci
Normal file
|
@ -0,0 +1,8 @@
|
|||
#! /bin/sh
|
||||
|
||||
path=`dirname $0`
|
||||
. "$path/SETUP.sh"
|
||||
extra_flags="$amd64_cflags $fast_cflags -g"
|
||||
extra_configs="$amd64_configs $max_configs --with-ndb-sci=/opt/DIS"
|
||||
|
||||
. "$path/FINISH.sh"
|
9
BUILD/compile-pentium64-max-sci
Normal file
9
BUILD/compile-pentium64-max-sci
Normal file
|
@ -0,0 +1,9 @@
|
|||
#! /bin/sh
|
||||
|
||||
path=`dirname $0`
|
||||
. "$path/SETUP.sh"
|
||||
|
||||
extra_flags="$pentium64_cflags $fast_cflags -g"
|
||||
extra_configs="$pentium_configs $max_configs --with-ndb-sci=/opt/DIS"
|
||||
|
||||
. "$path/FINISH.sh"
|
|
@ -157,6 +157,8 @@ test-bt:
|
|||
@PERL@ ./mysql-test-run.pl --force --comment=funcs1_ps --ps-protocol --suite=funcs_1
|
||||
-cd mysql-test ; MTR_BUILD_THREAD=auto \
|
||||
@PERL@ ./mysql-test-run.pl --force --comment=funcs2 --suite=funcs_2
|
||||
-cd mysql-test ; MTR_BUILD_THREAD=auto \
|
||||
@PERL@ ./mysql-test-run.pl --force --comment=rowlock --suite=row_lock
|
||||
|
||||
test-bt-debug:
|
||||
-cd mysql-test ; MTR_BUILD_THREAD=auto \
|
||||
|
|
|
@ -22,7 +22,7 @@ AC_DEFUN([MYSQL_CHECK_NDB_OPTIONS], [
|
|||
if test -f "$mysql_sci_dir/lib/libsisci.a" -a \
|
||||
-f "$mysql_sci_dir/include/sisci_api.h"; then
|
||||
NDB_SCI_INCLUDES="-I$mysql_sci_dir/include"
|
||||
NDB_SCI_LIBS="-L$mysql_sci_dir/lib -lsisci"
|
||||
NDB_SCI_LIBS="$mysql_sci_dir/lib/libsisci.a"
|
||||
AC_MSG_RESULT([-- including sci transporter])
|
||||
AC_DEFINE([NDB_SCI_TRANSPORTER], [1],
|
||||
[Including Ndb Cluster DB sci transporter])
|
||||
|
|
|
@ -66,8 +66,8 @@ extern int NEAR my_errno; /* Last error in mysys */
|
|||
#define MY_ALLOW_ZERO_PTR 64 /* my_realloc() ; zero ptr -> malloc */
|
||||
#define MY_FREE_ON_ERROR 128 /* my_realloc() ; Free old ptr on error */
|
||||
#define MY_HOLD_ON_ERROR 256 /* my_realloc() ; Return old ptr on error */
|
||||
#define MY_THREADSAFE 128 /* pread/pwrite: Don't allow interrupts */
|
||||
#define MY_DONT_OVERWRITE_FILE 1024 /* my_copy: Don't overwrite file */
|
||||
#define MY_THREADSAFE 2048 /* my_seek(): lock fd mutex */
|
||||
|
||||
#define MY_CHECK_ERROR 1 /* Params to my_end; Check open-close */
|
||||
#define MY_GIVE_INFO 2 /* Give time info about process*/
|
||||
|
|
|
@ -18,6 +18,10 @@ INCLUDE("${PROJECT_SOURCE_DIR}/win/mysql_manifest.cmake")
|
|||
# storage does not work properly in DLLs.
|
||||
SET(CMAKE_CXX_FLAGS_DEBUG "${CMAKE_CXX_FLAGS_DEBUG} -DSAFEMALLOC -DSAFE_MUTEX -DUSE_TLS")
|
||||
SET(CMAKE_C_FLAGS_DEBUG "${CMAKE_C_FLAGS_DEBUG} -DSAFEMALLOC -DSAFE_MUTEX -DUSE_TLS")
|
||||
SET(CMAKE_CXX_FLAGS_RELEASE "${CMAKE_CXX_FLAGS_RELEASE} -DSAFEMALLOC -DSAFE_MUTEX -DUSE_TLS")
|
||||
SET(CMAKE_C_FLAGS_RELEASE "${CMAKE_C_FLAGS_RELEASE} -DSAFEMALLOC -DSAFE_MUTEX -DUSE_TLS")
|
||||
SET(CMAKE_CXX_FLAGS_RELWITHDEBINFO "${CMAKE_CXX_FLAGS_RELWITHDEBINFO} -DSAFEMALLOC -DSAFE_MUTEX -DUSE_TLS")
|
||||
SET(CMAKE_C_FLAGS_RELWITHDEBINFO "${CMAKE_C_FLAGS_RELWITHDEBINFO} -DSAFEMALLOC -DSAFE_MUTEX -DUSE_TLS")
|
||||
|
||||
INCLUDE_DIRECTORIES(${CMAKE_SOURCE_DIR}/include
|
||||
${CMAKE_SOURCE_DIR}/zlib
|
||||
|
|
|
@ -81,7 +81,7 @@ INC_LIB= $(top_builddir)/regex/libregex.a \
|
|||
$(top_builddir)/mysys/libmysys.a \
|
||||
$(top_builddir)/strings/libmystrings.a \
|
||||
$(top_builddir)/dbug/libdbug.a \
|
||||
$(top_builddir)/vio/libvio.a
|
||||
$(top_builddir)/vio/libvio.a @NDB_SCI_LIBS@
|
||||
|
||||
|
||||
#
|
||||
|
|
|
@ -35,7 +35,8 @@ INCLUDES = -I$(top_builddir)/include -I$(top_srcdir)/include -I$(srcdir) \
|
|||
-I$(top_srcdir) -I$(top_srcdir)/client -I$(top_srcdir)/regex \
|
||||
$(openssl_includes)
|
||||
LIBS = @LIBS@ @WRAPLIBS@ @CLIENT_LIBS@ $(yassl_libs)
|
||||
LDADD = @CLIENT_EXTRA_LDFLAGS@ ../libmysqld.a @innodb_system_libs@ @LIBDL@ $(CXXLDFLAGS)
|
||||
LDADD = @CLIENT_EXTRA_LDFLAGS@ ../libmysqld.a @innodb_system_libs@ @LIBDL@ $(CXXLDFLAGS) \
|
||||
@NDB_SCI_LIBS@
|
||||
|
||||
mysqltest_embedded_LINK = $(CXXLINK)
|
||||
mysqltest_embedded_SOURCES = mysqltest.c
|
||||
|
|
|
@ -335,7 +335,7 @@ int chk_size(MI_CHECK *param, register MI_INFO *info)
|
|||
flush_key_blocks(info->s->key_cache,
|
||||
info->s->kfile, FLUSH_FORCE_WRITE);
|
||||
|
||||
size=my_seek(info->s->kfile,0L,MY_SEEK_END,MYF(0));
|
||||
size= my_seek(info->s->kfile, 0L, MY_SEEK_END, MYF(MY_THREADSAFE));
|
||||
if ((skr=(my_off_t) info->state->key_file_length) != size)
|
||||
{
|
||||
/* Don't give error if file generated by myisampack */
|
||||
|
@ -595,7 +595,8 @@ static int chk_index_down(MI_CHECK *param, MI_INFO *info, MI_KEYDEF *keyinfo,
|
|||
{
|
||||
/* purecov: begin tested */
|
||||
/* Give it a chance to fit in the real file size. */
|
||||
my_off_t max_length= my_seek(info->s->kfile, 0L, MY_SEEK_END, MYF(0));
|
||||
my_off_t max_length= my_seek(info->s->kfile, 0L, MY_SEEK_END,
|
||||
MYF(MY_THREADSAFE));
|
||||
mi_check_print_error(param, "Invalid key block position: %s "
|
||||
"key block size: %u file_length: %s",
|
||||
llstr(page, llbuff), keyinfo->block_length,
|
||||
|
@ -4052,10 +4053,10 @@ int test_if_almost_full(MI_INFO *info)
|
|||
{
|
||||
if (info->s->options & HA_OPTION_COMPRESS_RECORD)
|
||||
return 0;
|
||||
return (my_seek(info->s->kfile,0L,MY_SEEK_END,MYF(0))/10*9 >
|
||||
(my_off_t) (info->s->base.max_key_file_length) ||
|
||||
my_seek(info->dfile,0L,MY_SEEK_END,MYF(0))/10*9 >
|
||||
(my_off_t) info->s->base.max_data_file_length);
|
||||
return my_seek(info->s->kfile, 0L, MY_SEEK_END, MYF(MY_THREADSAFE)) / 10 * 9 >
|
||||
(my_off_t) info->s->base.max_key_file_length ||
|
||||
my_seek(info->dfile, 0L, MY_SEEK_END, MYF(0)) / 10 * 9 >
|
||||
(my_off_t) info->s->base.max_data_file_length;
|
||||
}
|
||||
|
||||
/* Recreate table with bigger more alloced record-data */
|
||||
|
|
4
mysql-test/include/have_log_bin.inc
Normal file
4
mysql-test/include/have_log_bin.inc
Normal file
|
@ -0,0 +1,4 @@
|
|||
-- require r/have_log_bin.require
|
||||
disable_query_log;
|
||||
show variables like "log_bin";
|
||||
enable_query_log;
|
|
@ -123,6 +123,10 @@ master-bin.000001 # Query 1 # use `test`; create table t3 like t1
|
|||
master-bin.000001 # Query 1 # use `test`; insert into t1 select * from t3
|
||||
master-bin.000001 # Query 1 # use `test`; replace into t1 select * from t3
|
||||
drop table t1,t2,t3;
|
||||
CREATE TABLE t1(a INT) ENGINE=BLACKHOLE;
|
||||
INSERT DELAYED INTO t1 VALUES(1);
|
||||
ERROR HY000: Table storage engine for 't1' doesn't have this option
|
||||
DROP TABLE t1;
|
||||
CREATE TABLE t1(a INT, b INT) ENGINE=BLACKHOLE;
|
||||
DELETE FROM t1 WHERE a=10;
|
||||
ALTER TABLE t1 ADD INDEX(a);
|
||||
|
|
2
mysql-test/r/have_log_bin.require
Normal file
2
mysql-test/r/have_log_bin.require
Normal file
|
@ -0,0 +1,2 @@
|
|||
Variable_name Value
|
||||
have_log_bin ON
|
|
@ -127,6 +127,17 @@ show binlog events;
|
|||
|
||||
drop table t1,t2,t3;
|
||||
|
||||
#
|
||||
# BUG#27998 - mysqld crashed when executing INSERT DELAYED on a BLACKHOLE
|
||||
# table
|
||||
#
|
||||
CREATE TABLE t1(a INT) ENGINE=BLACKHOLE;
|
||||
--error 1031
|
||||
INSERT DELAYED INTO t1 VALUES(1);
|
||||
DROP TABLE t1;
|
||||
|
||||
# End of 4.1 tests
|
||||
|
||||
#
|
||||
#Bug#19717: DELETE Query Error on BLACKHOLE when using WHERE on column with UNIQUE INDEX
|
||||
#
|
||||
|
@ -142,4 +153,4 @@ ALTER TABLE t1 ADD PRIMARY KEY(a);
|
|||
DELETE FROM t1 WHERE a=10;
|
||||
DROP TABLE t1;
|
||||
|
||||
# End of 4.1 tests
|
||||
# End of 5.0 tests
|
||||
|
|
|
@ -23,7 +23,9 @@
|
|||
my_off_t pos The expected position (absolute or relative)
|
||||
int whence A direction parameter and one of
|
||||
{SEEK_SET, SEEK_CUR, SEEK_END}
|
||||
myf MyFlags Not used.
|
||||
myf MyFlags MY_THREADSAFE must be set in case my_seek may be mixed
|
||||
with my_pread/my_pwrite calls and fd is shared among
|
||||
threads.
|
||||
|
||||
DESCRIPTION
|
||||
The my_seek function is a wrapper around the system call lseek and
|
||||
|
@ -54,9 +56,16 @@ my_off_t my_seek(File fd, my_off_t pos, int whence,
|
|||
Make sure we are using a valid file descriptor!
|
||||
*/
|
||||
DBUG_ASSERT(fd != -1);
|
||||
|
||||
newpos= lseek(fd, pos, whence);
|
||||
|
||||
#if defined(THREAD) && !defined(HAVE_PREAD)
|
||||
if (MyFlags & MY_THREADSAFE)
|
||||
{
|
||||
pthread_mutex_lock(&my_file_info[fd].mutex);
|
||||
newpos= lseek(fd, pos, whence);
|
||||
pthread_mutex_unlock(&my_file_info[fd].mutex);
|
||||
}
|
||||
else
|
||||
#endif
|
||||
newpos= lseek(fd, pos, whence);
|
||||
if (newpos == (os_off_t) -1)
|
||||
{
|
||||
my_errno=errno;
|
||||
|
|
|
@ -79,6 +79,8 @@ typedef ndbd_exit_classification_enum ndbd_exit_classification;
|
|||
#define NDBD_EXIT_NO_MORE_UNDOLOG 2312
|
||||
#define NDBD_EXIT_SR_UNDOLOG 2313
|
||||
#define NDBD_EXIT_SINGLE_USER_MODE 2314
|
||||
#define NDBD_EXIT_NODE_DECLARED_DEAD 2315
|
||||
#define NDBD_EXIT_SR_SCHEMAFILE 2316
|
||||
#define NDBD_EXIT_MEMALLOC 2327
|
||||
#define NDBD_EXIT_BLOCK_JBUFCONGESTION 2334
|
||||
#define NDBD_EXIT_TIME_QUEUE_SHORT 2335
|
||||
|
|
|
@ -20,7 +20,12 @@
|
|||
#include <TransporterCallback.hpp>
|
||||
#include <RefConvert.hpp>
|
||||
|
||||
#ifdef ERROR_INSERT
|
||||
Uint32 MAX_RECEIVED_SIGNALS = 1024;
|
||||
#else
|
||||
#define MAX_RECEIVED_SIGNALS 1024
|
||||
#endif
|
||||
|
||||
Uint32
|
||||
TransporterRegistry::unpack(Uint32 * readPtr,
|
||||
Uint32 sizeOfData,
|
||||
|
|
|
@ -65,13 +65,10 @@ SCI_Transporter::SCI_Transporter(TransporterRegistry &t_reg,
|
|||
|
||||
|
||||
m_initLocal=false;
|
||||
m_swapCounter=0;
|
||||
m_failCounter=0;
|
||||
m_remoteNodes[0]=remoteSciNodeId0;
|
||||
m_remoteNodes[1]=remoteSciNodeId1;
|
||||
m_adapters = nAdapters;
|
||||
// The maximum number of times to try and create,
|
||||
// start and destroy a sequence
|
||||
m_ActiveAdapterId=0;
|
||||
m_StandbyAdapterId=1;
|
||||
|
||||
|
@ -102,8 +99,6 @@ SCI_Transporter::SCI_Transporter(TransporterRegistry &t_reg,
|
|||
DBUG_VOID_RETURN;
|
||||
}
|
||||
|
||||
|
||||
|
||||
void SCI_Transporter::disconnectImpl()
|
||||
{
|
||||
DBUG_ENTER("SCI_Transporter::disconnectImpl");
|
||||
|
@ -129,7 +124,8 @@ void SCI_Transporter::disconnectImpl()
|
|||
|
||||
if(err != SCI_ERR_OK) {
|
||||
report_error(TE_SCI_UNABLE_TO_CLOSE_CHANNEL);
|
||||
DBUG_PRINT("error", ("Cannot close channel to the driver. Error code 0x%x",
|
||||
DBUG_PRINT("error",
|
||||
("Cannot close channel to the driver. Error code 0x%x",
|
||||
err));
|
||||
}
|
||||
}
|
||||
|
@ -164,19 +160,18 @@ bool SCI_Transporter::initTransporter() {
|
|||
m_sendBuffer.m_buffer = new Uint32[m_sendBuffer.m_sendBufferSize / 4];
|
||||
m_sendBuffer.m_dataSize = 0;
|
||||
|
||||
DBUG_PRINT("info", ("Created SCI Send Buffer with buffer size %d and packet size %d",
|
||||
DBUG_PRINT("info",
|
||||
("Created SCI Send Buffer with buffer size %d and packet size %d",
|
||||
m_sendBuffer.m_sendBufferSize, m_PacketSize * 4));
|
||||
if(!getLinkStatus(m_ActiveAdapterId) ||
|
||||
(m_adapters > 1 &&
|
||||
!getLinkStatus(m_StandbyAdapterId))) {
|
||||
DBUG_PRINT("error", ("The link is not fully operational. Check the cables and the switches"));
|
||||
//reportDisconnect(remoteNodeId, 0);
|
||||
//doDisconnect();
|
||||
DBUG_PRINT("error",
|
||||
("The link is not fully operational. Check the cables and the switches"));
|
||||
//NDB should terminate
|
||||
report_error(TE_SCI_LINK_ERROR);
|
||||
DBUG_RETURN(false);
|
||||
}
|
||||
|
||||
DBUG_RETURN(true);
|
||||
} // initTransporter()
|
||||
|
||||
|
@ -235,7 +230,8 @@ sci_error_t SCI_Transporter::initLocalSegment() {
|
|||
DBUG_PRINT("info", ("SCInode iD %d adapter %d\n",
|
||||
sciAdapters[i].localSciNodeId, i));
|
||||
if(err != SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Cannot open an SCI virtual device. Error code 0x%x",
|
||||
DBUG_PRINT("error",
|
||||
("Cannot open an SCI virtual device. Error code 0x%x",
|
||||
err));
|
||||
DBUG_RETURN(err);
|
||||
}
|
||||
|
@ -269,7 +265,8 @@ sci_error_t SCI_Transporter::initLocalSegment() {
|
|||
&err);
|
||||
|
||||
if(err != SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Local Segment is not accessible by an SCI adapter. Error code 0x%x\n",
|
||||
DBUG_PRINT("error",
|
||||
("Local Segment is not accessible by an SCI adapter. Error code 0x%x\n",
|
||||
err));
|
||||
DBUG_RETURN(err);
|
||||
}
|
||||
|
@ -303,15 +300,13 @@ sci_error_t SCI_Transporter::initLocalSegment() {
|
|||
&err);
|
||||
|
||||
if(err != SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Local Segment is not available for remote connections. Error code 0x%x\n",
|
||||
DBUG_PRINT("error",
|
||||
("Local Segment is not available for remote connections. Error code 0x%x\n",
|
||||
err));
|
||||
DBUG_RETURN(err);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
setupLocalSegment();
|
||||
|
||||
DBUG_RETURN(err);
|
||||
|
||||
} // initLocalSegment()
|
||||
|
@ -343,12 +338,6 @@ bool SCI_Transporter::doSend() {
|
|||
if(sizeToSend==4097)
|
||||
i4097++;
|
||||
#endif
|
||||
if(startSequence(m_ActiveAdapterId)!=SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Start sequence failed"));
|
||||
report_error(TE_SCI_UNABLE_TO_START_SEQUENCE);
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
tryagain:
|
||||
retry++;
|
||||
|
@ -374,119 +363,36 @@ bool SCI_Transporter::doSend() {
|
|||
SCI_FLAG_ERROR_CHECK,
|
||||
&err);
|
||||
|
||||
|
||||
if (err != SCI_ERR_OK) {
|
||||
if(err == SCI_ERR_OUT_OF_RANGE) {
|
||||
DBUG_PRINT("error", ("Data transfer : out of range error"));
|
||||
goto tryagain;
|
||||
}
|
||||
if(err == SCI_ERR_SIZE_ALIGNMENT) {
|
||||
DBUG_PRINT("error", ("Data transfer : alignment error"));
|
||||
DBUG_PRINT("info", ("sendPtr 0x%x, sizeToSend = %d", sendPtr, sizeToSend));
|
||||
goto tryagain;
|
||||
}
|
||||
if(err == SCI_ERR_OFFSET_ALIGNMENT) {
|
||||
DBUG_PRINT("error", ("Data transfer : offset alignment"));
|
||||
goto tryagain;
|
||||
}
|
||||
if(err == SCI_ERR_TRANSFER_FAILED) {
|
||||
//(m_TargetSegm[m_StandbyAdapterId].writer)->heavyLock();
|
||||
if(getLinkStatus(m_ActiveAdapterId)) {
|
||||
goto tryagain;
|
||||
}
|
||||
if (m_adapters == 1) {
|
||||
DBUG_PRINT("error", ("SCI Transfer failed"));
|
||||
if (err == SCI_ERR_OUT_OF_RANGE ||
|
||||
err == SCI_ERR_SIZE_ALIGNMENT ||
|
||||
err == SCI_ERR_OFFSET_ALIGNMENT) {
|
||||
DBUG_PRINT("error", ("Data transfer error = %d", err));
|
||||
report_error(TE_SCI_UNRECOVERABLE_DATA_TFX_ERROR);
|
||||
return false;
|
||||
}
|
||||
m_failCounter++;
|
||||
Uint32 temp=m_ActiveAdapterId;
|
||||
switch(m_swapCounter) {
|
||||
case 0:
|
||||
/**swap from active (0) to standby (1)*/
|
||||
if(getLinkStatus(m_StandbyAdapterId)) {
|
||||
DBUG_PRINT("error", ("Swapping from adapter 0 to 1"));
|
||||
}
|
||||
if(err == SCI_ERR_TRANSFER_FAILED) {
|
||||
if(getLinkStatus(m_ActiveAdapterId))
|
||||
goto tryagain;
|
||||
if (m_adapters == 1) {
|
||||
DBUG_PRINT("error", ("SCI Transfer failed"));
|
||||
report_error(TE_SCI_UNRECOVERABLE_DATA_TFX_ERROR);
|
||||
return false;
|
||||
}
|
||||
m_failCounter++;
|
||||
Uint32 temp=m_ActiveAdapterId;
|
||||
if (getLinkStatus(m_StandbyAdapterId)) {
|
||||
failoverShmWriter();
|
||||
SCIStoreBarrier(m_TargetSegm[m_StandbyAdapterId].sequence,0);
|
||||
m_ActiveAdapterId=m_StandbyAdapterId;
|
||||
m_StandbyAdapterId=temp;
|
||||
SCIRemoveSequence((m_TargetSegm[m_StandbyAdapterId].sequence),
|
||||
FLAGS,
|
||||
&err);
|
||||
if(err!=SCI_ERR_OK) {
|
||||
report_error(TE_SCI_UNABLE_TO_REMOVE_SEQUENCE);
|
||||
DBUG_PRINT("error", ("Unable to remove sequence"));
|
||||
return false;
|
||||
}
|
||||
if(startSequence(m_ActiveAdapterId)!=SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Start sequence failed"));
|
||||
report_error(TE_SCI_UNABLE_TO_START_SEQUENCE);
|
||||
return false;
|
||||
}
|
||||
m_swapCounter++;
|
||||
DBUG_PRINT("info", ("failover complete"));
|
||||
goto tryagain;
|
||||
} else {
|
||||
DBUG_PRINT("error", ("Swapping from adapter %u to %u",
|
||||
m_StandbyAdapterId, m_ActiveAdapterId));
|
||||
} else {
|
||||
report_error(TE_SCI_UNRECOVERABLE_DATA_TFX_ERROR);
|
||||
DBUG_PRINT("error", ("SCI Transfer failed"));
|
||||
return false;
|
||||
}
|
||||
return false;
|
||||
break;
|
||||
case 1:
|
||||
/** swap back from 1 to 0
|
||||
must check that the link is up */
|
||||
|
||||
if(getLinkStatus(m_StandbyAdapterId)) {
|
||||
failoverShmWriter();
|
||||
m_ActiveAdapterId=m_StandbyAdapterId;
|
||||
m_StandbyAdapterId=temp;
|
||||
DBUG_PRINT("info", ("Swapping from 1 to 0"));
|
||||
if(createSequence(m_ActiveAdapterId)!=SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Unable to create sequence"));
|
||||
report_error(TE_SCI_UNABLE_TO_CREATE_SEQUENCE);
|
||||
return false;
|
||||
}
|
||||
if(startSequence(m_ActiveAdapterId)!=SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("startSequence failed... disconnecting"));
|
||||
report_error(TE_SCI_UNABLE_TO_START_SEQUENCE);
|
||||
return false;
|
||||
}
|
||||
|
||||
SCIRemoveSequence((m_TargetSegm[m_StandbyAdapterId].sequence)
|
||||
, FLAGS,
|
||||
&err);
|
||||
if(err!=SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Unable to remove sequence"));
|
||||
report_error(TE_SCI_UNABLE_TO_REMOVE_SEQUENCE);
|
||||
return false;
|
||||
}
|
||||
|
||||
if(createSequence(m_StandbyAdapterId)!=SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Unable to create sequence on standby"));
|
||||
report_error(TE_SCI_UNABLE_TO_CREATE_SEQUENCE);
|
||||
return false;
|
||||
}
|
||||
|
||||
m_swapCounter=0;
|
||||
|
||||
DBUG_PRINT("info", ("failover complete.."));
|
||||
goto tryagain;
|
||||
|
||||
} else {
|
||||
DBUG_PRINT("error", ("Unrecoverable data transfer error"));
|
||||
report_error(TE_SCI_UNRECOVERABLE_DATA_TFX_ERROR);
|
||||
return false;
|
||||
}
|
||||
|
||||
break;
|
||||
default:
|
||||
DBUG_PRINT("error", ("Unrecoverable data transfer error"));
|
||||
report_error(TE_SCI_UNRECOVERABLE_DATA_TFX_ERROR);
|
||||
return false;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
} else {
|
||||
SHM_Writer * writer = (m_TargetSegm[m_ActiveAdapterId].writer);
|
||||
writer->updateWritePtr(sizeToSend);
|
||||
|
@ -497,7 +403,6 @@ bool SCI_Transporter::doSend() {
|
|||
m_sendBuffer.m_dataSize = 0;
|
||||
m_sendBuffer.m_forceSendLimit = sendLimit;
|
||||
}
|
||||
|
||||
} else {
|
||||
/**
|
||||
* If we end up here, the SCI segment is full.
|
||||
|
@ -552,15 +457,12 @@ void SCI_Transporter::setupLocalSegment()
|
|||
DBUG_VOID_RETURN;
|
||||
} //setupLocalSegment
|
||||
|
||||
|
||||
|
||||
void SCI_Transporter::setupRemoteSegment()
|
||||
{
|
||||
DBUG_ENTER("SCI_Transporter::setupRemoteSegment");
|
||||
Uint32 sharedSize = 0;
|
||||
sharedSize =4096; //start of the buffer is page aligned
|
||||
|
||||
|
||||
Uint32 sizeOfBuffer = m_BufferSize;
|
||||
const Uint32 slack = MAX_MESSAGE_SIZE;
|
||||
sizeOfBuffer -= sharedSize;
|
||||
|
@ -666,7 +568,6 @@ SCI_Transporter::init_remote()
|
|||
DBUG_PRINT("error", ("Error connecting segment, err 0x%x", err));
|
||||
DBUG_RETURN(false);
|
||||
}
|
||||
|
||||
}
|
||||
// Map the remote memory segment into program space
|
||||
for(Uint32 i=0; i < m_adapters ; i++) {
|
||||
|
@ -679,13 +580,14 @@ SCI_Transporter::init_remote()
|
|||
FLAGS,
|
||||
&err);
|
||||
|
||||
|
||||
if(err!= SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Cannot map a segment to the remote node %d. Error code 0x%x",m_RemoteSciNodeId, err));
|
||||
//NDB SHOULD TERMINATE AND COMPUTER REBOOTED!
|
||||
report_error(TE_SCI_CANNOT_MAP_REMOTESEGMENT);
|
||||
DBUG_RETURN(false);
|
||||
}
|
||||
if(err!= SCI_ERR_OK) {
|
||||
DBUG_PRINT("error",
|
||||
("Cannot map a segment to the remote node %d. Error code 0x%x",
|
||||
m_RemoteSciNodeId, err));
|
||||
//NDB SHOULD TERMINATE AND COMPUTER REBOOTED!
|
||||
report_error(TE_SCI_CANNOT_MAP_REMOTESEGMENT);
|
||||
DBUG_RETURN(false);
|
||||
}
|
||||
}
|
||||
m_mapped=true;
|
||||
setupRemoteSegment();
|
||||
|
@ -713,7 +615,6 @@ SCI_Transporter::connect_client_impl(NDB_SOCKET_TYPE sockfd)
|
|||
NDB_CLOSE_SOCKET(sockfd);
|
||||
DBUG_RETURN(false);
|
||||
}
|
||||
|
||||
if (!init_local()) {
|
||||
NDB_CLOSE_SOCKET(sockfd);
|
||||
DBUG_RETURN(false);
|
||||
|
@ -788,29 +689,9 @@ sci_error_t SCI_Transporter::createSequence(Uint32 adapterid) {
|
|||
&(m_TargetSegm[adapterid].sequence),
|
||||
SCI_FLAG_FAST_BARRIER,
|
||||
&err);
|
||||
|
||||
|
||||
return err;
|
||||
} // createSequence()
|
||||
|
||||
|
||||
sci_error_t SCI_Transporter::startSequence(Uint32 adapterid) {
|
||||
|
||||
sci_error_t err;
|
||||
/** Perform preliminary error check on an SCI adapter before starting a
|
||||
* sequence of read and write operations on the mapped segment.
|
||||
*/
|
||||
m_SequenceStatus = SCIStartSequence(
|
||||
(m_TargetSegm[adapterid].sequence),
|
||||
FLAGS, &err);
|
||||
|
||||
|
||||
// If there still is an error then data cannot be safely send
|
||||
return err;
|
||||
} // startSequence()
|
||||
|
||||
|
||||
|
||||
bool SCI_Transporter::disconnectLocal()
|
||||
{
|
||||
DBUG_ENTER("SCI_Transporter::disconnectLocal");
|
||||
|
@ -878,9 +759,6 @@ SCI_Transporter::~SCI_Transporter() {
|
|||
DBUG_VOID_RETURN;
|
||||
} // ~SCI_Transporter()
|
||||
|
||||
|
||||
|
||||
|
||||
void SCI_Transporter::closeSCI() {
|
||||
// Termination of SCI
|
||||
sci_error_t err;
|
||||
|
@ -897,8 +775,9 @@ void SCI_Transporter::closeSCI() {
|
|||
SCIClose(activeSCIDescriptor, FLAGS, &err);
|
||||
|
||||
if(err != SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Cannot close SCI channel to the driver. Error code 0x%x",
|
||||
err));
|
||||
DBUG_PRINT("error",
|
||||
("Cannot close SCI channel to the driver. Error code 0x%x",
|
||||
err));
|
||||
}
|
||||
SCITerminate();
|
||||
DBUG_VOID_RETURN;
|
||||
|
@ -973,7 +852,6 @@ SCI_Transporter::getConnectionStatus() {
|
|||
return false;
|
||||
}
|
||||
|
||||
|
||||
void
|
||||
SCI_Transporter::setConnected() {
|
||||
*m_remoteStatusFlag = SCICONNECTED;
|
||||
|
@ -983,7 +861,6 @@ SCI_Transporter::setConnected() {
|
|||
*m_localStatusFlag = SCICONNECTED;
|
||||
}
|
||||
|
||||
|
||||
void
|
||||
SCI_Transporter::setDisconnect() {
|
||||
if(getLinkStatus(m_ActiveAdapterId))
|
||||
|
@ -994,7 +871,6 @@ SCI_Transporter::setDisconnect() {
|
|||
}
|
||||
}
|
||||
|
||||
|
||||
bool
|
||||
SCI_Transporter::checkConnected() {
|
||||
if (*m_localStatusFlag == SCIDISCONNECT) {
|
||||
|
@ -1015,8 +891,9 @@ SCI_Transporter::initSCI() {
|
|||
SCIInitialize(0, &error);
|
||||
if(error != SCI_ERR_OK) {
|
||||
DBUG_PRINT("error", ("Cannot initialize SISCI library."));
|
||||
DBUG_PRINT("error", ("Inconsistency between SISCI library and SISCI driver. Error code 0x%x",
|
||||
error));
|
||||
DBUG_PRINT("error",
|
||||
("Inconsistency between SISCI library and SISCI driver. Error code 0x%x",
|
||||
error));
|
||||
DBUG_RETURN(false);
|
||||
}
|
||||
init = true;
|
||||
|
@ -1029,3 +906,4 @@ SCI_Transporter::get_free_buffer() const
|
|||
{
|
||||
return (m_TargetSegm[m_ActiveAdapterId].writer)->get_free_buffer();
|
||||
}
|
||||
|
||||
|
|
|
@ -54,12 +54,12 @@
|
|||
* local segment, the SCI transporter connects to a segment created by another
|
||||
* transporter at a remote node, and the maps the remote segment into its
|
||||
* virtual address space. However, since NDB Cluster relies on redundancy
|
||||
* at the network level, by using dual SCI adapters communica
|
||||
*
|
||||
* at the network level, by using dual SCI adapters communication can be
|
||||
* maintained even if one of the adapter cards fails (or anything on the
|
||||
* network this adapter card exists in e.g. an SCI switch failure).
|
||||
*
|
||||
*/
|
||||
|
||||
|
||||
/**
|
||||
* class SCITransporter
|
||||
* @brief - main class for the SCI transporter.
|
||||
|
@ -84,16 +84,6 @@ public:
|
|||
sci_error_t createSequence(Uint32 adapterid);
|
||||
|
||||
|
||||
/**
|
||||
* starts a sequence for error checking.
|
||||
* The actual checking that a sequence is correct is done implicitly
|
||||
* in SCIMemCpy (in doSend).
|
||||
* @param adapterid the adapter on which to start the sequence.
|
||||
* @return SCI_ERR_OK if ok, otherwize something else.
|
||||
*/
|
||||
sci_error_t startSequence(Uint32 adapterid);
|
||||
|
||||
|
||||
/** Initiate Local Segment: create a memory segment,
|
||||
* prepare a memory segment, map the local segment
|
||||
* into memory space and make segment available.
|
||||
|
@ -159,7 +149,6 @@ private:
|
|||
bool m_mapped;
|
||||
bool m_initLocal;
|
||||
bool m_sciinit;
|
||||
Uint32 m_swapCounter;
|
||||
Uint32 m_failCounter;
|
||||
/**
|
||||
* For statistics on transfered packets
|
||||
|
@ -195,7 +184,6 @@ private:
|
|||
*/
|
||||
Uint32 m_reportFreq;
|
||||
|
||||
|
||||
Uint32 m_adapters;
|
||||
Uint32 m_numberOfRemoteNodes;
|
||||
|
||||
|
|
|
@ -102,6 +102,10 @@ private:
|
|||
virtual void updateReceiveDataPtr(Uint32 bytesRead);
|
||||
|
||||
virtual Uint32 get_free_buffer() const;
|
||||
|
||||
inline bool hasReceiveData () const {
|
||||
return receiveBuffer.sizeOfData > 0;
|
||||
}
|
||||
protected:
|
||||
/**
|
||||
* Setup client/server and perform connect/accept
|
||||
|
|
|
@ -841,28 +841,13 @@ TransporterRegistry::poll_OSE(Uint32 timeOutMillis)
|
|||
Uint32
|
||||
TransporterRegistry::poll_TCP(Uint32 timeOutMillis)
|
||||
{
|
||||
bool hasdata = false;
|
||||
if (false && nTCPTransporters == 0)
|
||||
{
|
||||
tcpReadSelectReply = 0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
struct timeval timeout;
|
||||
#ifdef NDB_OSE
|
||||
// Return directly if there are no TCP transporters configured
|
||||
|
||||
if(timeOutMillis <= 1){
|
||||
timeout.tv_sec = 0;
|
||||
timeout.tv_usec = 1025;
|
||||
} else {
|
||||
timeout.tv_sec = timeOutMillis / 1000;
|
||||
timeout.tv_usec = (timeOutMillis % 1000) * 1000;
|
||||
}
|
||||
#else
|
||||
timeout.tv_sec = timeOutMillis / 1000;
|
||||
timeout.tv_usec = (timeOutMillis % 1000) * 1000;
|
||||
#endif
|
||||
|
||||
NDB_SOCKET_TYPE maxSocketValue = -1;
|
||||
|
||||
// Needed for TCP/IP connections
|
||||
|
@ -885,8 +870,27 @@ TransporterRegistry::poll_TCP(Uint32 timeOutMillis)
|
|||
// Put the connected transporters in the socket read-set
|
||||
FD_SET(socket, &tcpReadset);
|
||||
}
|
||||
hasdata |= t->hasReceiveData();
|
||||
}
|
||||
|
||||
timeOutMillis = hasdata ? 0 : timeOutMillis;
|
||||
|
||||
struct timeval timeout;
|
||||
#ifdef NDB_OSE
|
||||
// Return directly if there are no TCP transporters configured
|
||||
|
||||
if(timeOutMillis <= 1){
|
||||
timeout.tv_sec = 0;
|
||||
timeout.tv_usec = 1025;
|
||||
} else {
|
||||
timeout.tv_sec = timeOutMillis / 1000;
|
||||
timeout.tv_usec = (timeOutMillis % 1000) * 1000;
|
||||
}
|
||||
#else
|
||||
timeout.tv_sec = timeOutMillis / 1000;
|
||||
timeout.tv_usec = (timeOutMillis % 1000) * 1000;
|
||||
#endif
|
||||
|
||||
// The highest socket value plus one
|
||||
maxSocketValue++;
|
||||
|
||||
|
@ -901,7 +905,7 @@ TransporterRegistry::poll_TCP(Uint32 timeOutMillis)
|
|||
}
|
||||
#endif
|
||||
|
||||
return tcpReadSelectReply;
|
||||
return tcpReadSelectReply || hasdata;
|
||||
}
|
||||
#endif
|
||||
|
||||
|
@ -937,26 +941,26 @@ TransporterRegistry::performReceive()
|
|||
#endif
|
||||
|
||||
#ifdef NDB_TCP_TRANSPORTER
|
||||
if(tcpReadSelectReply > 0)
|
||||
for (int i=0; i<nTCPTransporters; i++)
|
||||
{
|
||||
for (int i=0; i<nTCPTransporters; i++)
|
||||
{
|
||||
checkJobBuffer();
|
||||
TCP_Transporter *t = theTCPTransporters[i];
|
||||
const NodeId nodeId = t->getRemoteNodeId();
|
||||
const NDB_SOCKET_TYPE socket = t->getSocket();
|
||||
if(is_connected(nodeId)){
|
||||
if(t->isConnected() && FD_ISSET(socket, &tcpReadset))
|
||||
checkJobBuffer();
|
||||
TCP_Transporter *t = theTCPTransporters[i];
|
||||
const NodeId nodeId = t->getRemoteNodeId();
|
||||
const NDB_SOCKET_TYPE socket = t->getSocket();
|
||||
if(is_connected(nodeId)){
|
||||
if(t->isConnected())
|
||||
{
|
||||
if (FD_ISSET(socket, &tcpReadset))
|
||||
{
|
||||
const int receiveSize = t->doReceive();
|
||||
if(receiveSize > 0)
|
||||
{
|
||||
Uint32 * ptr;
|
||||
Uint32 sz = t->getReceiveData(&ptr);
|
||||
transporter_recv_from(callbackObj, nodeId);
|
||||
Uint32 szUsed = unpack(ptr, sz, nodeId, ioStates[nodeId]);
|
||||
t->updateReceiveDataPtr(szUsed);
|
||||
}
|
||||
t->doReceive();
|
||||
}
|
||||
|
||||
if (t->hasReceiveData())
|
||||
{
|
||||
Uint32 * ptr;
|
||||
Uint32 sz = t->getReceiveData(&ptr);
|
||||
Uint32 szUsed = unpack(ptr, sz, nodeId, ioStates[nodeId]);
|
||||
t->updateReceiveDataPtr(szUsed);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -132,6 +132,7 @@ Cmvmi::~Cmvmi()
|
|||
|
||||
#ifdef ERROR_INSERT
|
||||
NodeBitmask c_error_9000_nodes_mask;
|
||||
extern Uint32 MAX_RECEIVED_SIGNALS;
|
||||
#endif
|
||||
|
||||
void Cmvmi::execNDB_TAMPER(Signal* signal)
|
||||
|
@ -161,6 +162,22 @@ void Cmvmi::execNDB_TAMPER(Signal* signal)
|
|||
kill(getpid(), SIGABRT);
|
||||
}
|
||||
#endif
|
||||
|
||||
#ifdef ERROR_INSERT
|
||||
if (signal->theData[0] == 9003)
|
||||
{
|
||||
if (MAX_RECEIVED_SIGNALS < 1024)
|
||||
{
|
||||
MAX_RECEIVED_SIGNALS = 1024;
|
||||
}
|
||||
else
|
||||
{
|
||||
MAX_RECEIVED_SIGNALS = 1 + (rand() % 128);
|
||||
}
|
||||
ndbout_c("MAX_RECEIVED_SIGNALS: %d", MAX_RECEIVED_SIGNALS);
|
||||
CLEAR_ERROR_INSERT_VALUE;
|
||||
}
|
||||
#endif
|
||||
}//execNDB_TAMPER()
|
||||
|
||||
void Cmvmi::execSET_LOGLEVELORD(Signal* signal)
|
||||
|
|
|
@ -79,6 +79,9 @@
|
|||
#include <NdbSleep.h>
|
||||
#include <signaldata/ApiBroadcast.hpp>
|
||||
|
||||
#include <EventLogger.hpp>
|
||||
extern EventLogger g_eventLogger;
|
||||
|
||||
#define ZNOT_FOUND 626
|
||||
#define ZALREADYEXIST 630
|
||||
|
||||
|
@ -1069,17 +1072,36 @@ void Dbdict::readSchemaConf(Signal* signal,
|
|||
|
||||
for (Uint32 n = 0; n < xsf->noOfPages; n++) {
|
||||
SchemaFile * sf = &xsf->schemaPage[n];
|
||||
bool ok =
|
||||
memcmp(sf->Magic, NDB_SF_MAGIC, sizeof(sf->Magic)) == 0 &&
|
||||
sf->FileSize != 0 &&
|
||||
sf->FileSize % NDB_SF_PAGE_SIZE == 0 &&
|
||||
sf->FileSize == sf0->FileSize &&
|
||||
sf->PageNumber == n &&
|
||||
computeChecksum((Uint32*)sf, NDB_SF_PAGE_SIZE_IN_WORDS) == 0;
|
||||
ndbrequire(ok || !crashInd);
|
||||
if (! ok) {
|
||||
bool ok = false;
|
||||
const char *reason;
|
||||
if (memcmp(sf->Magic, NDB_SF_MAGIC, sizeof(sf->Magic)) != 0)
|
||||
{ jam(); reason = "magic code"; }
|
||||
else if (sf->FileSize == 0)
|
||||
{ jam(); reason = "file size == 0"; }
|
||||
else if (sf->FileSize % NDB_SF_PAGE_SIZE != 0)
|
||||
{ jam(); reason = "invalid size multiple"; }
|
||||
else if (sf->FileSize != sf0->FileSize)
|
||||
{ jam(); reason = "invalid size"; }
|
||||
else if (sf->PageNumber != n)
|
||||
{ jam(); reason = "invalid page number"; }
|
||||
else if (computeChecksum((Uint32*)sf, NDB_SF_PAGE_SIZE_IN_WORDS) != 0)
|
||||
{ jam(); reason = "invalid checksum"; }
|
||||
else
|
||||
ok = true;
|
||||
|
||||
if (!ok)
|
||||
{
|
||||
char reason_msg[128];
|
||||
snprintf(reason_msg, sizeof(reason_msg),
|
||||
"schema file corrupt, page %u (%s, "
|
||||
"sz=%u sz0=%u pn=%u)",
|
||||
n, reason, sf->FileSize, sf0->FileSize, sf->PageNumber);
|
||||
if (crashInd)
|
||||
progError(__LINE__, NDBD_EXIT_SR_SCHEMAFILE, reason_msg);
|
||||
ndbrequireErr(fsPtr.p->fsState == FsConnectRecord::READ_SCHEMA1,
|
||||
NDBD_EXIT_SR_SCHEMAFILE);
|
||||
jam();
|
||||
ndbrequire(fsPtr.p->fsState == FsConnectRecord::READ_SCHEMA1);
|
||||
infoEvent("primary %s, trying backup", reason_msg);
|
||||
readSchemaRef(signal, fsPtr);
|
||||
return;
|
||||
}
|
||||
|
|
|
@ -2816,7 +2816,7 @@ void Qmgr::failReportLab(Signal* signal, Uint16 aFailedNode,
|
|||
if (failedNodePtr.i == getOwnNodeId()) {
|
||||
jam();
|
||||
|
||||
Uint32 code = 0;
|
||||
Uint32 code = NDBD_EXIT_NODE_DECLARED_DEAD;
|
||||
const char * msg = 0;
|
||||
char extra[100];
|
||||
switch(aFailCause){
|
||||
|
|
|
@ -57,12 +57,15 @@ static const ErrStruct errArray[] =
|
|||
"error(s) on other node(s)"},
|
||||
{NDBD_EXIT_PARTITIONED_SHUTDOWN, XAE, "Partitioned cluster detected. "
|
||||
"Please check if cluster is already running"},
|
||||
{NDBD_EXIT_NODE_DECLARED_DEAD, XAE,
|
||||
"Node declared dead. See error log for details"},
|
||||
{NDBD_EXIT_POINTER_NOTINRANGE, XIE, "Pointer too large"},
|
||||
{NDBD_EXIT_SR_OTHERNODEFAILED, XRE, "Another node failed during system "
|
||||
"restart, please investigate error(s) on other node(s)"},
|
||||
{NDBD_EXIT_NODE_NOT_DEAD, XRE, "Internal node state conflict, "
|
||||
"most probably resolved by restarting node again"},
|
||||
{NDBD_EXIT_SR_REDOLOG, XFI, "Error while reading the REDO log"},
|
||||
{NDBD_EXIT_SR_SCHEMAFILE, XFI, "Error while reading the schema file"},
|
||||
/* Currently unused? */
|
||||
{2311, XIE, "Conflict when selecting restart type"},
|
||||
{NDBD_EXIT_NO_MORE_UNDOLOG, XCR,
|
||||
|
|
|
@ -1306,6 +1306,36 @@ int runTestExecuteAsynch(NDBT_Context* ctx, NDBT_Step* step){
|
|||
|
||||
template class Vector<NdbScanOperation*>;
|
||||
|
||||
int
|
||||
runBug28443(NDBT_Context* ctx, NDBT_Step* step)
|
||||
{
|
||||
int result = NDBT_OK;
|
||||
int records = ctx->getNumRecords();
|
||||
|
||||
NdbRestarter restarter;
|
||||
|
||||
restarter.insertErrorInAllNodes(9003);
|
||||
|
||||
for (Uint32 i = 0; i<ctx->getNumLoops(); i++)
|
||||
{
|
||||
HugoTransactions hugoTrans(*ctx->getTab());
|
||||
if (hugoTrans.loadTable(GETNDB(step), records, 2048) != 0)
|
||||
{
|
||||
result = NDBT_FAILED;
|
||||
goto done;
|
||||
}
|
||||
if (runClearTable(ctx, step) != 0)
|
||||
{
|
||||
result = NDBT_FAILED;
|
||||
goto done;
|
||||
}
|
||||
}
|
||||
|
||||
done:
|
||||
restarter.insertErrorInAllNodes(9003);
|
||||
|
||||
return result;
|
||||
}
|
||||
|
||||
NDBT_TESTSUITE(testNdbApi);
|
||||
TESTCASE("MaxNdb",
|
||||
|
@ -1396,6 +1426,10 @@ TESTCASE("ExecuteAsynch",
|
|||
"Check that executeAsync() works (BUG#27495)\n"){
|
||||
INITIALIZER(runTestExecuteAsynch);
|
||||
}
|
||||
TESTCASE("Bug28443",
|
||||
""){
|
||||
INITIALIZER(runBug28443);
|
||||
}
|
||||
NDBT_TESTSUITE_END(testNdbApi);
|
||||
|
||||
int main(int argc, const char** argv){
|
||||
|
|
|
@ -629,6 +629,10 @@ max-time: 500
|
|||
cmd: testNdbApi
|
||||
args: -n ExecuteAsynch T1
|
||||
|
||||
max-time: 1000
|
||||
cmd: testNdbApi
|
||||
args: -n BugBug28443
|
||||
|
||||
#max-time: 500
|
||||
#cmd: testInterpreter
|
||||
#args: T1
|
||||
|
|
|
@ -102,11 +102,41 @@ case $system in
|
|||
;;
|
||||
esac
|
||||
|
||||
# This is needed to prefer GNU tar over platform tar because that can't
|
||||
# always handle long filenames
|
||||
|
||||
PATH_DIRS=`echo $PATH | \
|
||||
sed -e 's/^:/. /' -e 's/:$/ ./' -e 's/::/ . /g' -e 's/:/ /g' `
|
||||
|
||||
which_1 ()
|
||||
{
|
||||
for cmd
|
||||
do
|
||||
for d in $PATH_DIRS
|
||||
do
|
||||
for file in $d/$cmd
|
||||
do
|
||||
if [ -x $file -a ! -d $file ] ; then
|
||||
echo $file
|
||||
exit 0
|
||||
fi
|
||||
done
|
||||
done
|
||||
done
|
||||
exit 1
|
||||
}
|
||||
|
||||
tar=`which_1 gnutar gtar`
|
||||
if [ "$?" = "1" -o x"$tar" = x"" ] ; then
|
||||
tar=tar
|
||||
fi
|
||||
|
||||
|
||||
mkdir $BASE $BASE/bin $BASE/docs \
|
||||
$BASE/include $BASE/lib $BASE/support-files $BASE/share $BASE/scripts \
|
||||
$BASE/mysql-test $BASE/mysql-test/t $BASE/mysql-test/r \
|
||||
$BASE/mysql-test/include $BASE/mysql-test/std_data $BASE/mysql-test/lib
|
||||
$BASE/mysql-test/include $BASE/mysql-test/std_data $BASE/mysql-test/lib \
|
||||
$BASE/mysql-test/suite
|
||||
|
||||
if [ $BASE_SYSTEM != "netware" ] ; then
|
||||
mkdir $BASE/share/mysql $BASE/tests $BASE/sql-bench $BASE/man \
|
||||
|
@ -117,8 +147,8 @@ fi
|
|||
|
||||
# Copy files if they exists, warn for those that don't.
|
||||
# Note that when listing files to copy, we might list the file name
|
||||
# twice, once in the directory location where it is build, and a
|
||||
# second time in the ".libs" location. In the case the firs one
|
||||
# twice, once in the directory location where it is built, and a
|
||||
# second time in the ".libs" location. In the case the first one
|
||||
# is a wrapper script, the second one will overwrite it with the
|
||||
# binary file.
|
||||
copyfileto()
|
||||
|
@ -274,6 +304,13 @@ $CP mysql-test/t/*.test mysql-test/t/*.imtest \
|
|||
$CP mysql-test/r/*.result mysql-test/r/*.require \
|
||||
$BASE/mysql-test/r
|
||||
|
||||
# Copy the additional suites "as is", they are in flux
|
||||
$tar cf - mysql-test/suite | ( cd $BASE ; $tar xf - )
|
||||
# Clean up if we did this from a bk tree
|
||||
if [ -d mysql-test/SCCS ] ; then
|
||||
find $BASE/mysql-test -name SCCS -print | xargs rm -rf
|
||||
fi
|
||||
|
||||
if [ $BASE_SYSTEM != "netware" ] ; then
|
||||
chmod a+x $BASE/bin/*
|
||||
copyfileto $BASE/bin scripts/*
|
||||
|
@ -374,41 +411,12 @@ if [ x$DEBUG = x1 ] ; then
|
|||
exit
|
||||
fi
|
||||
|
||||
# This is needed to prefere gnu tar instead of tar because tar can't
|
||||
# always handle long filenames
|
||||
|
||||
PATH_DIRS=`echo $PATH | \
|
||||
sed -e 's/^:/. /' -e 's/:$/ ./' -e 's/::/ . /g' -e 's/:/ /g' `
|
||||
|
||||
which_1 ()
|
||||
{
|
||||
for cmd
|
||||
do
|
||||
for d in $PATH_DIRS
|
||||
do
|
||||
for file in $d/$cmd
|
||||
do
|
||||
if [ -x $file -a ! -d $file ] ; then
|
||||
echo $file
|
||||
exit 0
|
||||
fi
|
||||
done
|
||||
done
|
||||
done
|
||||
exit 1
|
||||
}
|
||||
|
||||
if [ $BASE_SYSTEM != "netware" ] ; then
|
||||
|
||||
#
|
||||
# Create the result tar file
|
||||
#
|
||||
|
||||
tar=`which_1 gnutar gtar`
|
||||
if [ "$?" = "1" -o x"$tar" = x"" ] ; then
|
||||
tar=tar
|
||||
fi
|
||||
|
||||
echo "Using $tar to create archive"
|
||||
|
||||
OPT=cvf
|
||||
|
|
|
@ -43,8 +43,7 @@ public:
|
|||
{
|
||||
return(HA_NULL_IN_KEY | HA_CAN_FULLTEXT | HA_CAN_SQL_HANDLER |
|
||||
HA_DUPP_POS | HA_CAN_INDEX_BLOBS | HA_AUTO_PART_KEY |
|
||||
HA_FILE_BASED | HA_CAN_GEOMETRY | HA_READ_RND_SAME |
|
||||
HA_CAN_INSERT_DELAYED);
|
||||
HA_FILE_BASED | HA_CAN_GEOMETRY | HA_READ_RND_SAME);
|
||||
}
|
||||
ulong index_flags(uint inx, uint part, bool all_parts) const
|
||||
{
|
||||
|
|
|
@ -2309,16 +2309,24 @@ int ha_ndbcluster::write_row(byte *record)
|
|||
{
|
||||
// Table has hidden primary key
|
||||
Ndb *ndb= get_ndb();
|
||||
int ret;
|
||||
Uint64 auto_value;
|
||||
uint retries= NDB_AUTO_INCREMENT_RETRIES;
|
||||
do {
|
||||
ret= ndb->getAutoIncrementValue((const NDBTAB *) m_table, auto_value, 1);
|
||||
} while (ret == -1 &&
|
||||
--retries &&
|
||||
ndb->getNdbError().status == NdbError::TemporaryError);
|
||||
if (ret == -1)
|
||||
ERR_RETURN(ndb->getNdbError());
|
||||
int retry_sleep= 30; /* 30 milliseconds, transaction */
|
||||
for (;;)
|
||||
{
|
||||
if (ndb->getAutoIncrementValue((const NDBTAB *) m_table,
|
||||
auto_value, 1) == -1)
|
||||
{
|
||||
if (--retries &&
|
||||
ndb->getNdbError().status == NdbError::TemporaryError);
|
||||
{
|
||||
my_sleep(retry_sleep);
|
||||
continue;
|
||||
}
|
||||
ERR_RETURN(ndb->getNdbError());
|
||||
}
|
||||
break;
|
||||
}
|
||||
if (set_hidden_key(op, table->s->fields, (const byte*)&auto_value))
|
||||
ERR_RETURN(op->getNdbError());
|
||||
}
|
||||
|
@ -4855,22 +4863,27 @@ ulonglong ha_ndbcluster::get_auto_increment()
|
|||
m_rows_to_insert - m_rows_inserted :
|
||||
((m_rows_to_insert > m_autoincrement_prefetch) ?
|
||||
m_rows_to_insert : m_autoincrement_prefetch));
|
||||
int ret;
|
||||
uint retries= NDB_AUTO_INCREMENT_RETRIES;
|
||||
do {
|
||||
ret=
|
||||
m_skip_auto_increment ?
|
||||
ndb->readAutoIncrementValue((const NDBTAB *) m_table, auto_value) :
|
||||
ndb->getAutoIncrementValue((const NDBTAB *) m_table, auto_value, cache_size);
|
||||
} while (ret == -1 &&
|
||||
--retries &&
|
||||
ndb->getNdbError().status == NdbError::TemporaryError);
|
||||
if (ret == -1)
|
||||
int retry_sleep= 30; /* 30 milliseconds, transaction */
|
||||
for (;;)
|
||||
{
|
||||
const NdbError err= ndb->getNdbError();
|
||||
sql_print_error("Error %lu in ::get_auto_increment(): %s",
|
||||
(ulong) err.code, err.message);
|
||||
DBUG_RETURN(~(ulonglong) 0);
|
||||
if (m_skip_auto_increment &&
|
||||
ndb->readAutoIncrementValue((const NDBTAB *) m_table, auto_value) ||
|
||||
ndb->getAutoIncrementValue((const NDBTAB *) m_table,
|
||||
auto_value, cache_size))
|
||||
{
|
||||
if (--retries &&
|
||||
ndb->getNdbError().status == NdbError::TemporaryError);
|
||||
{
|
||||
my_sleep(retry_sleep);
|
||||
continue;
|
||||
}
|
||||
const NdbError err= ndb->getNdbError();
|
||||
sql_print_error("Error %lu in ::get_auto_increment(): %s",
|
||||
(ulong) err.code, err.message);
|
||||
DBUG_RETURN(~(ulonglong) 0);
|
||||
}
|
||||
break;
|
||||
}
|
||||
DBUG_RETURN((longlong)auto_value);
|
||||
}
|
||||
|
@ -5011,27 +5024,36 @@ int ha_ndbcluster::open(const char *name, int mode, uint test_if_locked)
|
|||
set_dbname(name);
|
||||
set_tabname(name);
|
||||
|
||||
if (check_ndb_connection()) {
|
||||
free_share(m_share); m_share= 0;
|
||||
DBUG_RETURN(HA_ERR_NO_CONNECTION);
|
||||
if ((res= check_ndb_connection()) ||
|
||||
(res= get_metadata(name)))
|
||||
{
|
||||
free_share(m_share);
|
||||
m_share= 0;
|
||||
DBUG_RETURN(res);
|
||||
}
|
||||
|
||||
res= get_metadata(name);
|
||||
if (!res)
|
||||
while (1)
|
||||
{
|
||||
Ndb *ndb= get_ndb();
|
||||
if (ndb->setDatabaseName(m_dbname))
|
||||
{
|
||||
ERR_RETURN(ndb->getNdbError());
|
||||
res= ndb_to_mysql_error(&ndb->getNdbError());
|
||||
break;
|
||||
}
|
||||
struct Ndb_statistics stat;
|
||||
res= ndb_get_table_statistics(NULL, false, ndb, m_tabname, &stat);
|
||||
records= stat.row_count;
|
||||
if(!res)
|
||||
res= info(HA_STATUS_CONST);
|
||||
break;
|
||||
}
|
||||
|
||||
DBUG_RETURN(res);
|
||||
if (res)
|
||||
{
|
||||
free_share(m_share);
|
||||
m_share= 0;
|
||||
release_metadata();
|
||||
DBUG_RETURN(res);
|
||||
}
|
||||
DBUG_RETURN(0);
|
||||
}
|
||||
|
||||
|
||||
|
|
Loading…
Reference in a new issue