Logging to logging@openlogging.org accepted
sql_show.cc, type_enum.test, type_enum.result:
Bug #5628 German characters in field-defs will be '?' with some table definitions
mysql-test/r/type_enum.result:
Bug #5628 German characters in field-defs will be '?' with some table definitions
mysql-test/t/type_enum.test:
Bug #5628 German characters in field-defs will be '?' with some table definitions
sql/sql_show.cc:
Bug #5628 German characters in field-defs will be '?' with some table definitions
BitKeeper/etc/logging_ok:
Logging to logging@openlogging.org accepted
mysql-test/r/ps.result:
Test results fixed: the test case for bug#5688 "Upgraded 4.1.5 Server
seg faults"
mysql-test/t/ps.test:
Test case for bug#5688 "Upgraded 4.1.5 Server seg faults"
sql/item_cmpfunc.cc:
A fix for bug#5688 "Upgraded 4.1.5 Server seg faults":
fix just another place where we use wrong memory root for an Item
in statement prepare.
In addition, make the check for charsets in Item_bool_func2
more generic (fixes the test case when we use LIKE to compare BLOBs
with TEXT data).
we do not increment rli->group_master_log_pos if we are just after a SET ONE_SHOT (it's not a standalone event)
mysql-test/r/rpl_charset.result:
testing interruption of slave SQL thread between SET CHARACTER_SET_SERVER... and the companion INSERT.
mysql-test/t/rpl_charset.test:
testing interruption of slave SQL thread between SET CHARACTER_SET_SERVER... and the companion INSERT.
sql/log_event.cc:
we do not increment rli->group_master_log_pos if we are just after a SET
ONE_SHOT, because SET ONE_SHOT should not be separated from its following
updating query.
CREATE TABLE t1 SELECT POINT(1,2); fixed
mysql-test/r/gis.result:
Appropriate test result
mysql-test/t/gis.test:
test case
sql/item_geofunc.cc:
Item_geometry_func::fix_lengths_and_dec implementation
several fix_length_and_dec's not needed now
sql/item_geofunc.h:
Item_geometry_func class presented
mysql-test/r/ndb_autodiscover.result:
Update test result, number of rows is 1
mysql-test/t/ndb_autodiscover.test:
Dont run the test where table is dropped in NDb with ndb_drop_table
sql/ha_ndbcluster.cc:
Implement function ndbcluster_find_files which will discover new tables and delete old tables
sql/ha_ndbcluster.h:
Implement function ndbcluster_find_files
Remove function ndbcluster_list_tables and ndbcluster_can_discover
sql/handler.cc:
Add ha_find_files called from mysql_find_files
Remove ha_can_discover and ha_list_tables
sql/handler.h:
Add ha_find_files called from mysql_find_files
Remove ha_can_discover and ha_list_tables
sql/sql_show.cc:
Revert to original version of sql_show.cc
Only changes to this version is that ha_find_files is called from mysql_find_files in order to give the handlers a possibility to find new files in engine
OPTIMIZE TABLE <archive table>
This recompresses the table, thus removing any additional gzip headers caused by opening/closing or flushing the table.
mysql-test/r/archive.result:
Added optimize test case for archive engine.
mysql-test/t/archive.test:
Added test case for OPTIMIZE table <archive table>
sql/examples/ha_archive.cc:
The big change was the addition of optimize() call to allow tables to be recompressed (so if you have been reading/writing/reading/writing and ending up with larger files then you should, this will solve it). Though adding this feature is going to make it a real headache to add row level locking.
Also fixed bug reported by JD where storage engine code was not functioning (this of course was because I didn't check for the propper return value for hash_init). Removed BROKEN_GZIP ifdef since there was no way to enable it.
sql/examples/ha_archive.h:
Added optimize() method.
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Auto merged
sql/ha_ndbcluster.h:
Auto merged
sql/handler.cc:
Auto merged
sql/sql_base.cc:
Auto merged
sql/sql_table.cc:
Auto merged
They are separated from the other WL#1856 stuff, because they improve the behaviour of the current tests.
Make the result sets (order of rows) more predictable by using ORDER BY.
mysql-test/include/ps_modify.inc:
Make the result sets more predictable by using ORDER BY
mysql-test/include/ps_modify1.inc:
Make the result sets more predictable by using ORDER BY
mysql-test/r/ps_2myisam.result:
updated results
mysql-test/r/ps_3innodb.result:
updated results
mysql-test/r/ps_4heap.result:
updated results
mysql-test/r/ps_5merge.result:
updated results
mysql-test/r/ps_6bdb.result:
updated results
BitKeeper/etc/logging_ok:
Logging to logging@openlogging.org accepted
Merge fixes
mysql-test/r/ndb_insert.result:
Correct test cases and result after merge
mysql-test/t/ndb_insert.test:
Correct test cases and result after merge
sql/ha_ndbcluster.cc:
Correct get_error_message after merge
Use buf pointer in get_ndb_value
sql/ha_ndbcluster.cc:
Auto merged
sql/ha_ndbcluster.h:
Auto merged
mysql-test/r/ndb_insert.result:
Merge new test cases
mysql-test/t/ndb_insert.test:
Merge new test cases
key Column Fails".
mysql-test/r/ps.result:
Test results updated: a test case for Bug#5510 "inserting Null in
AutoIncrement primary key Column Fails".
mysql-test/r/ps_2myisam.result:
Bug#5510: a different warning in case of NULL->default truncation.This is
OK, the new warning is the same as produced by conventional execution.
mysql-test/r/ps_3innodb.result:
Bug#5510: a different warning in case of NULL->default truncation.This is
OK, the new warning is the same as produced by conventional execution.
mysql-test/r/ps_4heap.result:
Bug#5510: a different warning in case of NULL->default truncation.This is
OK, the new warning is the same as produced by conventional execution.
mysql-test/r/ps_5merge.result:
Bug#5510: a different warning in case of NULL->default truncation.This is
OK, the new warning is the same as produced by conventional execution.
mysql-test/r/ps_6bdb.result:
Bug#5510: a different warning in case of NULL->default truncation.This is
OK, the new warning is the same as produced by conventional execution.
mysql-test/t/ps.test:
A test case for Bug#5510 "inserting Null in AutoIncrement primary key Column
Fails".
sql/item.cc:
A fix for bug#5510 "inserting Null in AutoIncrement primary key Column
Fails": use proper Field API function for NULL placholeders.
single row subquery always can return NULL (no rows found) (BUG#5590)
mysql-test/r/subselect.result:
maybe_null flag returning by subquwery for temporary table creation
mysql-test/t/subselect.test:
maybe_null flag returning by subquwery for temporary table creation
sql/item.cc:
storing maybe_null in type holder
sql/item_subselect.cc:
single row subquery always can return NULL (no rows found)
sql/sql_derived.cc:
fixed error handling if creating derived table failed
ndb/src/kernel/blocks/dbtup/DbtupMeta.cpp:
oops
ndb/src/kernel/blocks/dbtux/DbtuxScan.cpp:
jamEntry
ndb/src/kernel/blocks/dblqh/DblqhMain.cpp:
jamEntry
mysql-test/r/ndb_index.result:
ndb charsets: minimal fix to make test pass
mysql-test/t/ndb_index.test:
ndb charsets: minimal fix to make test pass
We don't want the update to abort when IGNORE is specified
mysql-test/r/update.result:
Bug#5553 - UPDATE IGNORE fails on dup key
New test
mysql-test/t/update.test:
Bug#5553 - UPDATE IGNORE fails on dup key
New test
BitKeeper/etc/logging_ok:
Logging to logging@openlogging.org accepted
into hundin.mysql.fi:/home/jan/mysql-4.1
mysql-test/r/ctype_utf8.result:
Auto merged
mysql-test/t/ctype_utf8.test:
Auto merged
sql/ha_innodb.cc:
Auto merged
sql/ha_innodb.h:
Auto merged
mysql-test/r/func_compress.result:
clean-up. The length of compress() may depend on zlib version, etc.
mysql-test/t/func_compress.test:
clean-up. The length of compress() may depend on zlib version, etc.
into mysql.com:/home/wax/mysql/mysql-4.1mysqltest
client/mysqltest.c:
Auto merged
mysql-test/Makefile.am:
Auto merged
mysql-test/r/fulltext.result:
Auto merged
mysql-test/r/type_float.result:
Auto merged
mysql-test/t/innodb.test:
Auto merged
mysql-test/t/type_float.test:
Auto merged
mysql-test/t/variables.test:
Auto merged
into mysql.com:/usr/local/home/marty/MySQL/test/mysql-4.1
ndb/src/common/transporter/TransporterRegistry.cpp:
Auto merged
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
Auto merged
This bug is the result of weird error happening with mi_float8get and double
arythmetic. I described that in 'Bug that looks potentially dangerous' email
myisam/rt_mbr.c:
*ab_area was changed with the local variable just to make the function
working without errors
mysql-test/r/gis.result:
appropriate test result
mysql-test/t/gis.test:
Test case
BUG#4312 "wrong behaviour on insert .. on duplicate key" functionality disabled
mysql-test/r/ndb_insert.result:
New tests for fduplicate inserts in combination with transaction
New tests for INSERT IGNORE and REPLACE
mysql-test/t/ndb_insert.test:
New tests for fduplicate inserts in combination with transaction
New tests for INSERT IGNORE and REPLACE
ndb/src/ndbapi/NdbConnection.cpp:
Return error 4350 "Transaction already aborted" if execute(Commit) is called when theCommitStatus==Aborted
Add DBUG_PRINT's
ndb/src/ndbapi/ndberror.c:
Add new error message indicating that the transaction already has been aborted.
sql/ha_ndbcluster.cc:
Map all error code 0 to 1 in order to catch errors caused by NdbApi returning -1 without having set an error code.
Use ndb object in THD in get_error_message
BUG# 4312 Return HA_ERR_WRONG_COMMAND if extra(HA_EXTRA_IGNORE_DUP_KEY) is called
Only use writeTuple if command is REPLACE
sql/ha_ndbcluster.h:
Added member variable to keep track of when HA_EXTRA_IGNORE_DUP_KEY is used, but NDB can't support it.
on intelxeon3 (Solaris x86))
mysql-test/r/ps_2myisam.result:
Test results fixed (order by for a couple of statements in the PS test).
mysql-test/r/ps_3innodb.result:
Test results fixed (order by for a couple of statements in the PS test).
mysql-test/r/ps_4heap.result:
Test results fixed (order by for a couple of statements in the PS test).
mysql-test/r/ps_5merge.result:
Test results fixed (order by for a couple of statements in the PS test).
mysql-test/r/ps_6bdb.result:
Test results fixed (order by for a couple of statements in the PS test).
added fix for keeping "records" up to date when execute() fails
mysql-test/r/ndb_insert.result:
Added testcase for select count() during transaction with failures
mysql-test/t/ndb_insert.test:
Added testcase for select count() during transaction with failures
sql/ha_ndbcluster.cc:
added fix for keeping "records" up to date when execute() fails
sql/ha_ndbcluster.h:
added fix for keeping "records" up to date when execute() fails
Added a check to recover from IGNORE_SPACE in this situation:
<ident-character(s)><space><dot><ident-character(s)>
The ignored space led to the false identification of the dot
as an ident separator (like "db.table").
mysql-test/r/sql_mode.result:
BUG#5318 - failure: 'IGNORE_SPACE' affects numeric values after DEFAULT.
Added the test results.
mysql-test/t/sql_mode.test:
BUG#5318 - failure: 'IGNORE_SPACE' affects numeric values after DEFAULT.
Added new tests for the bug.
sql/sql_lex.cc:
BUG#5318 - failure: 'IGNORE_SPACE' affects numeric values after DEFAULT.
Added code to recover from skipped spaces in mode IGNORE_SPACES,
when testing for an ident separator (which happens to be a dot).
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
new methods to keep "records" up to datecorrect record field in ndbcluster handler
new method for ndbcluster handler to store/retrieve table and thread specific data
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
hanged deleteKey to return ponter to deleted object
moved heavy global cache fetch from inline to separate method
mysql-test/r/ndb_alter_table.result:
correct record field in ndbcluster handler
mysql-test/r/ndb_blob.result:
correct record field in ndbcluster handler
ndb/include/ndbapi/NdbDictionary.hpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/DictCache.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/DictCache.hpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
ndb/src/ndbapi/Ndb.cpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbDictionary.cpp:
new method for ndbcluster handler to store/retrieve table and thread specific data
ndb/src/ndbapi/NdbDictionaryImpl.cpp:
changed local hash to store new table_info object, with placeholders for local data, instead of TableImpl
moved heavy global cache fetch from inline to separate method
ndb/src/ndbapi/NdbDictionaryImpl.hpp:
replaced method DictionaryImpl::getTable with DictionaryImpl::get_local_table_info
ndb/src/ndbapi/NdbLinHash.hpp:
changed deleteKey to return ponter to deleted object
sql/ha_ndbcluster.cc:
moved all ndb thread specific data into new placeholder
new methods to keep "records" up to date
unset flag HA_NOT_EXACT_COUNT to make handler read "records" field, for count() optim and join optimization
sql/ha_ndbcluster.h:
new methods to keep "records" up to date
sql/sql_class.h:
moved all ndb thread specific data into new placeholder