mirror of
https://github.com/MariaDB/server.git
synced 2025-01-29 10:14:19 +01:00
Merge siva.hindu.god:/home/tsmith/m/bk/50
into siva.hindu.god:/home/tsmith/m/bk/maint/50
This commit is contained in:
commit
686be32eca
8 changed files with 110 additions and 15 deletions
|
@ -381,3 +381,27 @@ SELECT id, tag, doc, type FROM t1 WHERE id IN ('flipper','sakila');
|
|||
id tag doc type
|
||||
sakila 1 Some text goes here text
|
||||
DROP TABLE t1;
|
||||
CREATE TABLE t1 (
|
||||
var1 int(2) NOT NULL,
|
||||
var2 int(2) NOT NULL,
|
||||
PRIMARY KEY (var1)
|
||||
) ENGINE=ndbcluster DEFAULT CHARSET=ascii CHECKSUM=1;
|
||||
CREATE TABLE t2 (
|
||||
var1 int(2) NOT NULL,
|
||||
var2 int(2) NOT NULL,
|
||||
PRIMARY KEY (var1)
|
||||
) ENGINE=MyISAM DEFAULT CHARSET=ascii CHECKSUM=1;
|
||||
CREATE TRIGGER testtrigger
|
||||
AFTER UPDATE ON t1 FOR EACH ROW BEGIN
|
||||
REPLACE INTO t2 SELECT * FROM t1 WHERE t1.var1 = NEW.var1;END|
|
||||
INSERT INTO t1 VALUES (1,1),(2,2),(3,3);
|
||||
UPDATE t1 SET var2 = 9 WHERE var1 IN(1,2,3);
|
||||
DROP TRIGGER testtrigger;
|
||||
DROP TABLE t1, t2;
|
||||
create table t2 (a int, b int, primary key (a), key ab (a,b)) engine=ndbcluster;
|
||||
insert into t2 values (1,1), (10,10);
|
||||
select * from t2 use index (ab) where a in(1,10) order by a;
|
||||
a b
|
||||
1 1
|
||||
10 10
|
||||
drop table t2;
|
||||
|
|
|
@ -253,3 +253,41 @@ SELECT id, tag, doc, type FROM t1 WHERE id IN ('flipper','orka');
|
|||
SELECT id, tag, doc, type FROM t1 WHERE id IN ('flipper','sakila');
|
||||
|
||||
DROP TABLE t1;
|
||||
|
||||
#bug#25522
|
||||
CREATE TABLE t1 (
|
||||
var1 int(2) NOT NULL,
|
||||
var2 int(2) NOT NULL,
|
||||
PRIMARY KEY (var1)
|
||||
) ENGINE=ndbcluster DEFAULT CHARSET=ascii CHECKSUM=1;
|
||||
|
||||
|
||||
CREATE TABLE t2 (
|
||||
var1 int(2) NOT NULL,
|
||||
var2 int(2) NOT NULL,
|
||||
PRIMARY KEY (var1)
|
||||
) ENGINE=MyISAM DEFAULT CHARSET=ascii CHECKSUM=1;
|
||||
|
||||
|
||||
DELIMITER |;
|
||||
CREATE TRIGGER testtrigger
|
||||
AFTER UPDATE ON t1 FOR EACH ROW BEGIN
|
||||
REPLACE INTO t2 SELECT * FROM t1 WHERE t1.var1 = NEW.var1;END|
|
||||
DELIMITER ;|
|
||||
|
||||
INSERT INTO t1 VALUES (1,1),(2,2),(3,3);
|
||||
|
||||
UPDATE t1 SET var2 = 9 WHERE var1 IN(1,2,3);
|
||||
|
||||
DROP TRIGGER testtrigger;
|
||||
|
||||
DROP TABLE t1, t2;
|
||||
|
||||
#bug#25821
|
||||
create table t2 (a int, b int, primary key (a), key ab (a,b)) engine=ndbcluster;
|
||||
|
||||
insert into t2 values (1,1), (10,10);
|
||||
|
||||
select * from t2 use index (ab) where a in(1,10) order by a;
|
||||
|
||||
drop table t2;
|
||||
|
|
|
@ -63,12 +63,14 @@ public:
|
|||
bool order_by,
|
||||
bool order_desc = false,
|
||||
bool read_range_no = false,
|
||||
bool keyinfo = false) {
|
||||
bool keyinfo = false,
|
||||
bool multi_range = false) {
|
||||
Uint32 scan_flags =
|
||||
(SF_OrderBy & -(Int32)order_by) |
|
||||
(SF_Descending & -(Int32)order_desc) |
|
||||
(SF_ReadRangeNo & -(Int32)read_range_no) |
|
||||
(SF_KeyInfo & -(Int32)keyinfo);
|
||||
(SF_KeyInfo & -(Int32)keyinfo) |
|
||||
(SF_MultiRange & -(Int32)multi_range);
|
||||
|
||||
return readTuples(lock_mode, scan_flags, parallel, batch);
|
||||
}
|
||||
|
|
|
@ -37,13 +37,15 @@ class NdbScanOperation : public NdbOperation {
|
|||
public:
|
||||
/**
|
||||
* Scan flags. OR-ed together and passed as second argument to
|
||||
* readTuples.
|
||||
* readTuples. Note that SF_MultiRange has to be set if several
|
||||
* ranges (bounds) are to be passed.
|
||||
*/
|
||||
enum ScanFlag {
|
||||
SF_TupScan = (1 << 16), // scan TUP
|
||||
SF_OrderBy = (1 << 24), // index scan in order
|
||||
SF_Descending = (2 << 24), // index scan in descending order
|
||||
SF_ReadRangeNo = (4 << 24), // enable @ref get_range_no
|
||||
SF_MultiRange = (8 << 24), // scan is part of multi-range scan
|
||||
SF_KeyInfo = 1 // request KeyInfo to be sent back
|
||||
};
|
||||
|
||||
|
@ -70,7 +72,8 @@ public:
|
|||
*/
|
||||
#ifdef ndb_readtuples_impossible_overload
|
||||
int readTuples(LockMode lock_mode = LM_Read,
|
||||
Uint32 batch = 0, Uint32 parallel = 0, bool keyinfo = false);
|
||||
Uint32 batch = 0, Uint32 parallel = 0,
|
||||
bool keyinfo = false, bool multi_range = false);
|
||||
#endif
|
||||
|
||||
inline int readTuples(int parallell){
|
||||
|
@ -262,6 +265,7 @@ protected:
|
|||
bool m_descending;
|
||||
Uint32 m_read_range_no;
|
||||
NdbRecAttr *m_curr_row; // Pointer to last returned row
|
||||
bool m_multi_range; // Mark if operation is part of multi-range scan
|
||||
};
|
||||
|
||||
inline
|
||||
|
|
|
@ -2120,6 +2120,18 @@ CommandInterpreter::executeStatus(int processId,
|
|||
ndbout << processId << ": Node not found" << endl;
|
||||
return -1;
|
||||
}
|
||||
if (cl->node_states[i].node_type != NDB_MGM_NODE_TYPE_NDB){
|
||||
if (cl->node_states[i].version != 0){
|
||||
ndbout << "Node "<< cl->node_states[i].node_id <<": connected" ;
|
||||
ndbout_c(" (Version %d.%d.%d)",
|
||||
getMajor(version) ,
|
||||
getMinor(version),
|
||||
getBuild(version));
|
||||
|
||||
}else
|
||||
ndbout << "Node "<< cl->node_states[i].node_id <<": not connected" << endl;
|
||||
return 0;
|
||||
}
|
||||
status = cl->node_states[i].node_status;
|
||||
startPhase = cl->node_states[i].start_phase;
|
||||
version = cl->node_states[i].version;
|
||||
|
|
|
@ -983,6 +983,8 @@ Ndb::opTupleIdOnNdb(Ndb_local_table_info* info, Uint64 & opValue, Uint32 op)
|
|||
Uint64 tValue;
|
||||
NdbRecAttr* tRecAttrResult;
|
||||
|
||||
NdbError savedError;
|
||||
|
||||
CHECK_STATUS_MACRO_ZERO;
|
||||
|
||||
BaseString currentDb(getDatabaseName());
|
||||
|
@ -1077,7 +1079,12 @@ Ndb::opTupleIdOnNdb(Ndb_local_table_info* info, Uint64 & opValue, Uint32 op)
|
|||
|
||||
error_handler:
|
||||
theError.code = tConnection->theError.code;
|
||||
|
||||
savedError = theError;
|
||||
|
||||
this->closeTransaction(tConnection);
|
||||
theError = savedError;
|
||||
|
||||
error_return:
|
||||
// Restore current name space
|
||||
setDatabaseName(currentDb.c_str());
|
||||
|
|
|
@ -1188,7 +1188,7 @@ NdbIndexScanOperation::setBound(const NdbColumnImpl* tAttrInfo,
|
|||
* so it's safe to use [tIndexAttrId]
|
||||
* (instead of looping as is NdbOperation::equal_impl)
|
||||
*/
|
||||
if(type == BoundEQ && tDistrKey)
|
||||
if(type == BoundEQ && tDistrKey && !m_multi_range)
|
||||
{
|
||||
theNoOfTupKeyLeft--;
|
||||
return handle_distribution_key((Uint64*)aValue, sizeInWords);
|
||||
|
@ -1249,7 +1249,8 @@ NdbIndexScanOperation::readTuples(LockMode lm,
|
|||
const bool order_by = scan_flags & SF_OrderBy;
|
||||
const bool order_desc = scan_flags & SF_Descending;
|
||||
const bool read_range_no = scan_flags & SF_ReadRangeNo;
|
||||
|
||||
m_multi_range = scan_flags & SF_MultiRange;
|
||||
|
||||
int res = NdbScanOperation::readTuples(lm, scan_flags, parallel, batch);
|
||||
if(!res && read_range_no)
|
||||
{
|
||||
|
@ -1716,6 +1717,12 @@ NdbIndexScanOperation::reset_bounds(bool forceSend){
|
|||
int
|
||||
NdbIndexScanOperation::end_of_bound(Uint32 no)
|
||||
{
|
||||
DBUG_ENTER("end_of_bound");
|
||||
DBUG_PRINT("info", ("Range number %u", no));
|
||||
/* Check that SF_MultiRange has been specified if more
|
||||
than one range is specified */
|
||||
if (no > 0 && !m_multi_range)
|
||||
DBUG_RETURN(-1);
|
||||
if(no < (1 << 13)) // Only 12-bits no of ranges
|
||||
{
|
||||
Uint32 bound_head = * m_first_bound_word;
|
||||
|
@ -1724,9 +1731,9 @@ NdbIndexScanOperation::end_of_bound(Uint32 no)
|
|||
|
||||
m_first_bound_word = theKEYINFOptr + theTotalNrOfKeyWordInSignal;;
|
||||
m_this_bound_start = theTupKeyLen;
|
||||
return 0;
|
||||
DBUG_RETURN(0);
|
||||
}
|
||||
return -1;
|
||||
DBUG_RETURN(-1);
|
||||
}
|
||||
|
||||
int
|
||||
|
|
|
@ -1995,8 +1995,7 @@ int ha_ndbcluster::set_bounds(NdbIndexScanOperation *op,
|
|||
DBUG_PRINT("error", ("key %d unknown flag %d", j, p.key->flag));
|
||||
DBUG_ASSERT(FALSE);
|
||||
// Stop setting bounds but continue with what we have
|
||||
op->end_of_bound(range_no);
|
||||
DBUG_RETURN(0);
|
||||
DBUG_RETURN(op->end_of_bound(range_no));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -2043,8 +2042,7 @@ int ha_ndbcluster::set_bounds(NdbIndexScanOperation *op,
|
|||
|
||||
tot_len+= part_store_len;
|
||||
}
|
||||
op->end_of_bound(range_no);
|
||||
DBUG_RETURN(0);
|
||||
DBUG_RETURN(op->end_of_bound(range_no));
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -3882,11 +3880,10 @@ int ha_ndbcluster::start_stmt(THD *thd, thr_lock_type lock_type)
|
|||
ERR_RETURN(ndb->getNdbError());
|
||||
no_uncommitted_rows_reset(thd);
|
||||
thd_ndb->stmt= trans;
|
||||
thd_ndb->query_state&= NDB_QUERY_NORMAL;
|
||||
trans_register_ha(thd, FALSE, &ndbcluster_hton);
|
||||
}
|
||||
thd_ndb->query_state&= NDB_QUERY_NORMAL;
|
||||
m_active_trans= trans;
|
||||
|
||||
// Start of statement
|
||||
m_retrieve_all_fields= FALSE;
|
||||
m_retrieve_primary_key= FALSE;
|
||||
|
@ -6365,7 +6362,7 @@ ha_ndbcluster::read_multi_range_first(KEY_MULTI_RANGE **found_range_p,
|
|||
}
|
||||
else if ((scanOp= m_active_trans->getNdbIndexScanOperation(idx, tab))
|
||||
&&!scanOp->readTuples(lm, 0, parallelism, sorted,
|
||||
FALSE, TRUE, need_pk)
|
||||
FALSE, TRUE, need_pk, TRUE)
|
||||
&&!generate_scan_filter(m_cond_stack, scanOp)
|
||||
&&!define_read_attrs(end_of_buffer-reclength, scanOp))
|
||||
{
|
||||
|
@ -6528,7 +6525,11 @@ close_scan:
|
|||
}
|
||||
|
||||
if (multi_range_curr == multi_range_end)
|
||||
{
|
||||
Thd_ndb *thd_ndb= get_thd_ndb(current_thd);
|
||||
thd_ndb->query_state&= NDB_QUERY_NORMAL;
|
||||
DBUG_RETURN(HA_ERR_END_OF_FILE);
|
||||
}
|
||||
|
||||
/**
|
||||
* Read remaining ranges
|
||||
|
|
Loading…
Add table
Reference in a new issue