into perch.ndb.mysql.com:/home/jonas/src/mysql-5.1-new-ndb
storage/ndb/src/kernel/blocks/ERROR_codes.txt:
Auto merged
storage/ndb/src/kernel/blocks/dbdih/Dbdih.hpp:
Auto merged
storage/ndb/src/kernel/blocks/dbdih/DbdihMain.cpp:
Auto merged
storage/ndb/src/kernel/blocks/dbtc/DbtcMain.cpp:
Auto merged
storage/ndb/src/kernel/blocks/suma/Suma.cpp:
Auto merged
storage/ndb/test/run-test/daily-basic-tests.txt:
Auto merged
storage/ndb/test/ndbapi/testNodeRestart.cpp:
merge
Fix race condition between NODE_FAILREP and local INCL_NODEREQ loop
Also retry on ZNODE_START_DISALLOWED_ERROR
storage/ndb/include/kernel/signaldata/StartPerm.hpp:
Move error code
storage/ndb/src/kernel/blocks/ERROR_codes.txt:
new error code
storage/ndb/src/kernel/blocks/dbdih/Dbdih.hpp:
Move error code
storage/ndb/src/kernel/blocks/dbdih/DbdihMain.cpp:
1) retry also on ZNODE_START_DISALLOWED_ERROR
2) Change if() else in INCL_NODECONF to for-loop instead
3) (last but not least) fix bug, that could cause different block withing same node to have different opinion about node status
solution is to check if node is still alive before sening next local INCL_NODEREQ
storage/ndb/src/kernel/blocks/dbtc/DbtcMain.cpp:
Add error insert to allow node to die during INCL_NODEREQ
storage/ndb/src/kernel/blocks/suma/Suma.cpp:
1) let suma be well behaved (i.e reply to INCL_NODEREQ)
2) Add dump to print c_connceted_nodes/c_subscriber_nodes (8010)
storage/ndb/test/ndbapi/testNodeRestart.cpp:
new testcase
storage/ndb/test/run-test/daily-basic-tests.txt:
new testcase
into chilla.local:/home/mydev/mysql-5.1-bug24985
mysql-test/r/heap_btree.result:
Auto merged
mysql-test/t/heap_btree.test:
Auto merged
storage/heap/ha_heap.cc:
Auto merged
storage/myisam/ha_myisam.cc:
Auto merged
rows event):
Adding a after image COLS bitmap to Update_rows_log_event (for telling
what columns that are present in the after image of each row update).
Also fixing case where Rows_log_event length was not correctly computed
(happened when the number of columns in a table was more than 251).
mysql-test/r/rpl_row_inexist_tbl.result:
Result change.
sql/log_event.cc:
Extending Rows_log_event with two new fields: m_bitbuf_ai and m_cols_ai. These fields are only used for the Update_rows_log_event.
Adding implementation of Update_rows_log_event destructor.
Using new after image fields inside the Update_rows_log_event.
Factoring out common constructor bodies into Update_rows_log_event::init()
function.
Fixing case where length of Rows_log_event was not correctly computed (for
tables with more than 251 columns).
sql/log_event.h:
Moving implementation of Rows_log_event::get_data_size() into .cc file.
Adding Update_rows_log_event constructor accepting both before image
and after image COLS vector.
Adding Update_rows_log_vector destructor.
Adding fields m_bitbuf_ai and m_cols_ai to Rows_log_event.
Fixing is_valid() to look at m_cols_ai as well.
- Fixing utf8_general_cs according to recent changes.
- Compiling utf8_general_cs in pentium-debug-max configuration
to avoid these problems in the future.
BUILD/compile-pentium-debug-max:
Enable compiling of experimental collations in compile-pentium-debug-max
config/ac-macros/character_sets.m4:
Adding hidden flag --with-experimental-collations,
not seen in "configure --help".
strings/ctype-utf8.c:
Compilation failure changes:
catching up with previous character set changes:
- uni_plane is now not a global variables
- adding new parameter into my_strnncollsp_utf8_cs
- adding my_strnxfrm_len into MY_COLLATION_HANDLER
for utf8_general_cs
into chilla.local:/home/mydev/mysql-5.0-bug24985
mysql-test/r/heap_btree.result:
Auto merged
sql/ha_heap.cc:
Auto merged
mysql-test/t/heap_btree.test:
Bug#24985 - UTF8 ENUM primary key on MEMORY using BTREE
causes incorrect duplicate entries
Manual merge from 4.1
causes incorrect duplicate entries
Keys for BTREE indexes on ENUM and SET columns of MEMORY tables
with character set UTF8 were computed incorrectly. Many
different column values got the same key value.
Apart of possible performance problems, it made unique indexes
of this type unusable because it rejected many different
values as duplicates.
The problem was that multibyte character detection was tried
on the internal numeric column value. Many values were not
identified as characters. Their key value became blank filled.
Thanks to Alexander Barkov and Ramil Kalimullin for the patch,
which sets the character set of ENUM and SET key segments to
the pseudo binary character set.
mysql-test/r/heap_btree.result:
Bug#24985 - UTF8 ENUM primary key on MEMORY using BTREE
causes incorrect duplicate entries
Added test result.
mysql-test/t/heap_btree.test:
Bug#24985 - UTF8 ENUM primary key on MEMORY using BTREE
causes incorrect duplicate entries
Added test.
sql/ha_heap.cc:
Bug#24985 - UTF8 ENUM primary key on MEMORY using BTREE
causes incorrect duplicate entries
Set key segment charset to my_charset_bin for ENUM and SET
columns.
Fix: adding namespace reference into "mysql --xml" output,
to make it work similary to "mysqldump --xml".
client/mysql.cc:
Adding namespace reference.
mysql-test/r/client_xml.result:
Fixing test results
Problem: GROUP BY on empty ucs2 strings crashed server.
Reason: sometimes mi_unique_hash() is executed with
ptr=null and length=0, which means "empty string".
The branch of code handling UCS2 character set
was not safe against ptr=null and fell into and
endless loop even if length=0 because of poiter
arithmetic overflow.
Fix: adding special check for length=0 to avoid pointer arithmetic
overflow.
mysql-test/r/ctype_uca.result:
Adding test case
mysql-test/t/ctype_uca.test:
Adding test case
strings/ctype-uca.c:
Fix my_uca_scanner_init_ucs2 to be safe against
strings with length=0 and ptr=0.
into whalegate.ndb.mysql.com:/home/tomas/mysql-5.1-single-user
storage/ndb/include/ndbapi/NdbRecAttr.hpp:
Auto merged
storage/ndb/src/ndbapi/NdbRecAttr.cpp:
Auto merged
sql/ha_ndbcluster.cc:
manual merge
storage/ndb/test/src/NDBT_Table.cpp:
manual merge
renamed "--old-mode" to "--old" to prevent
ambiguity.
"old" now appears in SHOW VARIABLES as a
read-only option.
mysql-test/r/group_by.result:
WL3527: 5.1 test case
mysql-test/t/group_by.test:
WL3527: 5.1 test case
sql/mysqld.cc:
WL3527: 5.1
renamed the "old-mode" to "old" to fit
the options naming scheme
sql/set_var.cc:
WL3527: 5.1
added "-old" as read-only system variable
sql/set_var.h:
WL3527: 5.1
added class for boolean pointer read-only
option to support the "--old" option.
sql/sql_base.cc:
fixed 5.0->5.1 merge problems.
correct cit printout
correct bit store retrieve
ndb/include/ndbapi/NdbRecAttr.hpp:
correct medium int printout
ndb/src/ndbapi/NdbRecAttr.cpp:
correct cit printout
sql/ha_ndbcluster.cc:
correct bit store retrieve
into magare.gmz:/home/kgeorge/mysql/work/WL3527-5.1-opt
mysql-test/r/gis.result:
Auto merged
mysql-test/t/gis.test:
Auto merged
sql/field.h:
Auto merged
sql/item.cc:
Auto merged
sql/item_sum.cc:
Auto merged
sql/item_sum.h:
Auto merged
sql/sql_class.h:
Auto merged
sql/sql_delete.cc:
Auto merged
sql/sql_insert.cc:
Auto merged
sql/sql_select.cc:
Auto merged
mysql-test/r/subselect.result:
merge 5.0-opt -> 5.1-opt
mysql-test/t/subselect.test:
merge 5.0-opt -> 5.1-opt
sql/sql_base.cc:
merge 5.0-opt -> 5.1-opt
into mysql_cab_desk.:C:/source/c++/mysql-5.1_WL_3629
BitKeeper/deleted/.del-init_db.sql~a77d572c39d5a1f8:
Auto merged
BitKeeper/deleted/.del-mysql_create_system_tables.sh:
Auto merged
scripts/mysql_system_tables_fix.sql:
Auto merged
sql/event_data_objects.cc:
Auto merged
sql/slave.cc:
Auto merged
sql/sql_parse.cc:
Auto merged
sql/sql_show.cc:
Auto merged
sql/sql_yacc.yy:
Auto merged
sql/event_data_objects.h:
Manual merge.
sql/event_queue.cc:
Manual merge.
sql/events.cc:
Manual merge.
enabled the optional FOR JOIN to all the three
clauses : USE, FORCE and IGNORE
mysql-test/r/select.result:
WL3527: 5.0 part: test cases
mysql-test/t/select.test:
WL3527: 5.0 part: test cases
to 0 causes wrong (large) length to be read
from the row in _mi_calc_blob_length() when
storing NULL values in (e.g) POINT columns.
This large length is then used to allocate
a block of memory that (on some OSes) causes
trouble.
Fixed by calling the base class's
Field_blob::reset() from Field_geom::reset()
that is called when storing a NULL value into
the column.
mysql-test/r/gis.result:
Bug #27164: test case
mysql-test/t/gis.test:
Bug #27164: test case
sql/field.h:
Bug #27164: not reseting the data pointer
to 0 causes wrong (large) length to be read
from the row in _mi_calc_blob_length() when
storing NULL values in (e.g) POINT columns.
This large length is then used to allocate
a block of memory that (on some OSes) causes
trouble.
This may lead to buffer overflow.
The String::qs_append() function will append a string
without checking if there's enough space.
So qs_append() must be called beforehand to ensure
there's enough space in the buffer for the subsequent
qs_append() calls.
Fixed Item_case_expr::print() to make sure there's
enough space before appending data by adding a call to
String::reserve() to make sure qs_append() will have
enough space.
mysql-test/r/sp-code.result:
Bug #26303: test case
mysql-test/t/sp-code.test:
Bug #26303: test case
sql/item.cc:
Bug #26303: added a call to String::reserve() to
make sure qs_append will have enough space
sql/item.h:
Bug #26303: m_case_expr_id made unsigned
because it's offset in an array.
Kinda obvious that this was going to come up ;)
client/client_priv.h:
New options
client/mysqlslap.c:
Added option for pre/post run
mysql-test/r/mysqlslap.result:
Updated tests
mysql-test/t/mysqlslap.test:
Updated test
into whalegate.ndb.mysql.com:/home/tomas/mysql-5.1-new-ndb
storage/ndb/src/kernel/blocks/dbdih/DbdihMain.cpp:
Auto merged
storage/ndb/test/ndbapi/testNodeRestart.cpp:
Auto merged
into whalegate.ndb.mysql.com:/home/tomas/mysql-5.1-new-ndb
storage/ndb/src/kernel/blocks/dbdih/Dbdih.hpp:
Auto merged
storage/ndb/src/kernel/blocks/dbdih/DbdihMain.cpp:
Auto merged
storage/ndb/src/kernel/blocks/dbtc/Dbtc.hpp:
Auto merged
storage/ndb/src/kernel/blocks/dbtc/DbtcMain.cpp:
Auto merged
storage/ndb/test/ndbapi/testNodeRestart.cpp:
Auto merged
from func_group.test after the patch for bug #27229 had been applied.
The memory corruption happened because in some rare cases the function
count_field_types underestimated the number of elements in
in the array param->items_to_copy.
sql/item_sum.cc:
The return value of the Item_sum::update_used_tables method
should not depend on the place of aggregation of the set
function for which the Item_sum object has been created.
sql/sql_select.cc:
This is a fix for the memory corruption occurred in one of test cases
from func_group.test after the patch for bug #27229 had been applied.
The memory corruption happened because in some rare cases the function
count_field_types underestimated the number of elements in
in the array param->items_to_copy.
Currently it's not guaranteed that after JOIN::prepare() the
used_tables attribute is calculated for all items. For example
for the expression SUM(outer_ref)+1 used_tables() must return
OUTER_REF_TABLE_BIT. Yet by the moment when the used_tables
attribute is calculated in JOIN::prepare SUM(outer_ref) has
not been substituted for Item_aggregate_ref yet.
By this reason additional calls of the method update_used_tables
are needed for some items passed as parameters to the function
create_tmp_table.