into perch.ndb.mysql.com:/home/jonas/src/mysql-5.0
ndb/src/kernel/blocks/backup/Backup.cpp:
Auto merged
scripts/make_binary_distribution.sh:
Auto merged
into neptunus.(none):/home/msvensson/mysql/bug11835/my50-bug11835
client/mysqltest.c:
Auto merged
configure.in:
Auto merged
mysql-test/mysql-test-run.pl:
Auto merged
sql/item_func.cc:
Auto merged
sql/sql_udf.cc:
Auto merged
Remove useless and tricky state fiddleing in TC
to syncronize NF_CompleteRep as code is already present in DIH aswell
Keep broadcast of TAKEOVER_TCCONF for online upgrade
ndb/src/kernel/blocks/dblqh/DblqhMain.cpp:
Add clever dump for showing active operations
ndb/src/kernel/blocks/dbtc/Dbtc.hpp:
Remove useless and tricky state fiddleing in TC
to syncronize NF_CompleteRep as code is already present in DIH aswell
Keep broadcast of TAKEOVER_TCCONF for online upgrade
ndb/src/kernel/blocks/dbtc/DbtcMain.cpp:
Remove useless and tricky state fiddleing in TC
to syncronize NF_CompleteRep as code is already present in DIH aswell
Keep broadcast of TAKEOVER_TCCONF for online upgrade
Use variable waitfor_response_timeout (depending on TransactionDeadLockTimeout)
When getting 4012, set NeedAbort and ReleaseOnClose
ndb/src/ndbapi/NdbConnection.cpp:
Use variable for WAITFOR_RESPONSE_TIMEOUT
ndb/src/ndbapi/Ndbif.cpp:
Use variable timeout for waitfor,
when receiving 4012, set NeedAbort and ReleaseOnClose
ndb/src/ndbapi/TransporterFacade.cpp:
Init wait_for_response_timoue as max TRANSACTION_DEADLOCK_TIMEOUT
ndb/src/ndbapi/TransporterFacade.hpp:
Init wait_for_response_timoue as max TRANSACTION_DEADLOCK_TIMEOUT
ndb/test/ndbapi/testTimeout.cpp:
Add testcase for 4012
ndb/test/run-test/daily-basic-tests.txt:
Add testcase for 4012
Fixed that LIKE worked case insensitively for latin2_czech_cs,
which was wrong for a case sensitive collation.
include/m_ctype.h:
Making my_wildcmp_bin public instead of static
strings/ctype-bin.c:
Making my_wildcmp_bin public instead of static
strings/ctype-czech.c:
Use my_wildcmp_bin instead of case insensitive my_wildcmp_8bit
mysql-test/include/have_latin2_ch.inc:
New BitKeeper file ``mysql-test/include/have_latin2_ch.inc''
mysql-test/r/ctype_latin2_ch.result:
New BitKeeper file ``mysql-test/r/ctype_latin2_ch.result''
mysql-test/r/have_latin2_ch.require:
New BitKeeper file ``mysql-test/r/have_latin2_ch.require''
mysql-test/t/ctype_latin2_ch.test:
New BitKeeper file ``mysql-test/t/ctype_latin2_ch.test''
table.cc:
Fixing to use system_charset_info instead of default_charset_info.
Crash happened because the "ctype" array is empty in UCS2,
and thus cannot be used with my_isspace().
The reason why UCS2 appeared in this context was because of
of default_charset_info variable incorrectly substituted to my_isspace().
As functions check_db_name(), check_table_name() and check_column_name()
always get values in utf8, system_charset_info must be used instead.
ctype_ucs2_def.test, ctype_ucs2_def-master.opt, ctype_ucs2_def.result:
new file
sql/table.cc:
Bug#18004 Connecting crashes server when default charset is UCS2
Use of default_charset_info was wrong.
Functions check_db_name, check_table_name and check_column_name
get values of system_charset_info character set (utf8).
Activly abort transactions (that's affected) during NF
This removes a lot of bugs that can occur otherwise is using
high value for TransactionDeadLockTimout
ndb/include/kernel/signaldata/TcContinueB.hpp:
New continueb for active transaction abort on nf
ndb/src/kernel/blocks/dbtc/Dbtc.hpp:
Add bitmask of participating nodes to transaction record
Add bitmask of node fail steps, so that NF_CompleteRep is not sent until all steps has completed
ndb/src/kernel/blocks/dbtc/DbtcMain.cpp:
Active transaction baortion
additional check for subselect
mysql-test/r/information_schema.result:
Fix for bug#18224 VIEW on information_schema crashes the server
test case
mysql-test/t/information_schema.test:
Fix for bug#18224 VIEW on information_schema crashes the server
test case
Crash happened when one selected data from one of INFORMATION_SCHEMA
tables and in order to build its contents server had to open view which
used stored function and table or view on which one had not global or
database-level privileges (e.g. had only table-level or had no
privileges at all).
The crash was caused by usage of check_grant() function, which assumes
that either number of tables to be inspected by it is limited explicitly
or table list used and thd->lex->query_tables_own_last value correspond
to each other (the latter should be either 0 or point to next_global
member of one of elements of this table list), in conditions when
above assumptions were not true. This fix just explicitly limits
number of tables to be inspected. Other negative effects which are
caused by the fact that thd->lex->query_tables_own_last might not
be set properly during processing of I_S tables are less disastrous
and will be reported and fixed separetely.
mysql-test/r/information_schema_db.result:
Fix for bug #18113 "SELECT * FROM information_schema.xxx crashes server"
test case
mysql-test/t/information_schema_db.test:
Fix for bug #18113 "SELECT * FROM information_schema.xxx crashes server"
test case
sql/sql_acl.cc:
added note
vio/viosocket.c:
ORs were wrong to begin with, but more importantly, we mustn't use
shortable-circuit logic.
Also newly reap the connection-closed event handle.
Changed release name to 5.0.19a
ha_innodb.cc:
InnoDB used table locks (not row locks) within stored functions. (Bug #18077)
sql/ha_innodb.cc:
InnoDB used table locks (not row locks) within stored functions. (Bug #18077)
configure.in:
Changed release name to 5.0.19a
dont't allow node to join cluster until all nodes has completed failure handling
ndb/src/kernel/blocks/qmgr/QmgrMain.cpp:
When getting CM_ADD for node that I haven't completed failure handling for do _not_ just override.
But instead set state...and send CM_ACK_ADD on execCONNECT_REP (much...later)
ndb/test/ndbapi/testNodeRestart.cpp:
testcase for bug#16772
ndb/test/run-test/daily-basic-tests.txt:
Run test in basic suite
to binlog too much.
When InnoDB has to rollback a transaction because the lock table has
filled up, it also needs to inform the upper layer that the transaction
was rolled back so that the cached transaction is not written to the
binary log.
sql/ha_innodb.cc:
When InnoDB rolls back a transaction in HA_ERR_LOCK_TABLE_FULL, it
needs to inform the upper layer to rollback the transaction also.
8 repeated nr with table wo/ logging cause crash
Dont create crashed replica for temporary tables
ndb/src/kernel/blocks/dbdih/Dbdih.hpp:
Dont create crashed replica for temporary tables
ndb/src/kernel/blocks/dbdih/DbdihMain.cpp:
Dont create crashed replica for temporary tables
into mysql.com:/usr/local/mysql/mysql-5.0
mysql-test/lib/mtr_timer.pl:
Auto merged
ndb/src/kernel/blocks/backup/Backup.cpp:
Auto merged
scripts/make_binary_distribution.sh:
Auto merged
mysql-test/lib/mtr_timer.pl:
Fix bug where ^C would trigger cleanup handler in both parent and
timeout child processes, causing duplicated messages and potential
conflicts.
¨MySQL server crashes if you try to access to InnoDB table¨
crash caused by schizophrenic mysqld - 2 memory locations for logically same function
with conflicting values.
Fixed by backporting from 5.1 changes to have_xyz_db declarations.
sql/mysql_priv.h:
Backport have_xyz_db changes from 5.1
sql/mysqld.cc:
Backport have_xyz_db changes from 5.1
because it can couse problem with Query cache (BUG#14767)
mysql-test/r/query_cache_notembedded.result:
BUG#14767 test suite.
mysql-test/t/query_cache_notembedded.test:
BUG#14767 test suite.
sql/sql_insert.cc:
Query cache invalidation table we was inserting in just after
unlocking table added to avoid the race condition as we had
with SP.
sql/sql_parse.cc:
Query cache invalidation table we was inserting in just after
unlocking table added to avoid the race condition as we had
with SP.
sql/sql_yacc.yy:
We should prohibit concurent read of unserting file in SP
because it can cause problem with query cache.
into perch.ndb.mysql.com:/home/jonas/src/mysql-5.0
ndb/src/kernel/blocks/backup/Backup.cpp:
Auto merged
scripts/make_binary_distribution.sh:
Auto merged
Afterfix for bug#17366: Unchecked Item_int results in server crash
sql/sql_select.cc:
Afterfix for bug#17366: Unchecked Item_int results in server crash
- Add HAVE_OPENSSL and HAVE_YASSL to config-win.h
include/config-win.h:
Define HAVE_OPENSSL and HAVE_YASSL to make the server and client enable SSL support
When there is conjunction of conds, the substitute_for_best_equal_field()
will call the eliminate_item_equal() function in loop to build final
expression. But if eliminate_item_equal() finds that some cond will always
evaluate to 0, then that cond will be substituted by Item_int with value ==
0. In this case on the next iteration eliminate_item_equal() will get that
Item_int and treat it as Item_cond. This is leads to memory corruption and
server crash on cleanup phase.
To the eliminate_item_equal() function was added DBUG_ASSERT for checking
that all items treaten as Item_cond are really Item_cond.
The substitute_for_best_equal_field() now checks that if
eliminate_item_equal() returns Item_int and it's value is 0 then this
value is returned as the result of whole conjunction.
mysql-test/t/subselect.test:
Added test for bug#17366: Unchecked Item_int results in server crash
mysql-test/r/subselect.result:
Added test for bug#17366: Unchecked Item_int results in server crash
sql/sql_select.cc:
Fixed bug#17366: Unchecked Item_int results in server crash
To the eliminate_item_equal() function was added DBUG_ASSERT for checking
that all items treaten as Item_cond are really Item_cond.
The substitute_for_best_equal_field() now checks that if
eliminate_item_equal() returns something other than Item_cond and if it is
then this value is returned as the result of whole conjunction.
ndb crash if trigger record get to big, fix incorrect max size of trigger record
ndb/src/kernel/blocks/backup/Backup.cpp:
Increase size of max trigger record
ndb crash if trigger record get to big, fix incorrect max size of trigger record
ndb/src/kernel/blocks/backup/Backup.cpp:
Increase size of max trigger record
mysql-test/mysql-test-run.pl:
Take a snapshot of the data dirs just after all db's have been installed and usedthat snasphot to restore the db(s) to a known state after a test case has failed.
Thus avoiding subsequent test cases to fail because of previous test failures.