mirror of
https://github.com/MariaDB/server.git
synced 2025-01-22 14:54:20 +01:00
ee1c5226b9
Fixed memory leak in _db_set() (Bug#24497 Valgrind warning: get_one_option) Don't call net_clear() on COM_QUIT. This avoids a warning from net_clear() after shutdown: "skipped ## bytes from file" BUG#21428: skipped 9 bytes from file: socket (3)" on "mysqladmin shutdown" sql-bench/test-wisconsin.sh: Rename: BitKeeper/deleted/.del-test-wisconsin.sh~c0b86821b5f95f26 -> sql-bench/test-wisconsin.sh sql-bench/test-transactions.sh: Rename: BitKeeper/deleted/.del-test-transactions.sh~c1c892f10c40caf -> sql-bench/test-transactions.sh sql-bench/test-insert.sh: Rename: BitKeeper/deleted/.del-test-insert.sh~893bfac9dedb79a7 -> sql-bench/test-insert.sh sql-bench/test-select.sh: Rename: BitKeeper/deleted/.del-test-select.sh~6d5cc770acf11be6 -> sql-bench/test-select.sh sql-bench/test-create.sh: Rename: BitKeeper/deleted/.del-test-create.sh~f9a1ea38c191b17a -> sql-bench/test-create.sh sql-bench/test-big-tables.sh: Rename: BitKeeper/deleted/.del-test-big-tables.sh~5b1b0c5fb623565a -> sql-bench/test-big-tables.sh sql-bench/test-connect.sh: Rename: BitKeeper/deleted/.del-test-connect.sh~382a728c949ee075 -> sql-bench/test-connect.sh sql-bench/test-ATIS.sh: Rename: BitKeeper/deleted/.del-test-ATIS.sh~e8ebff7086c95773 -> sql-bench/test-ATIS.sh sql-bench/test-alter-table.sh: Rename: BitKeeper/deleted/.del-test-alter-table.sh~eba6cfa9972fcced -> sql-bench/test-alter-table.sh sql-bench/innotest2b.sh: Rename: BitKeeper/deleted/.del-innotest2b.sh~7b99ece835e8dff3 -> sql-bench/innotest2b.sh sql-bench/run-all-tests.sh: Rename: BitKeeper/deleted/.del-run-all-tests.sh~41d6da1cf211ee95 -> sql-bench/run-all-tests.sh sql-bench/innotest2.sh: Rename: BitKeeper/deleted/.del-innotest2.sh~9c14df528285603 -> sql-bench/innotest2.sh sql-bench/innotest2a.sh: Rename: BitKeeper/deleted/.del-innotest2a.sh~e01d016a7cafdc0b -> sql-bench/innotest2a.sh sql-bench/innotest1a.sh: Rename: BitKeeper/deleted/.del-innotest1a.sh~c64f4610ae1e26fe -> sql-bench/innotest1a.sh sql-bench/innotest1b.sh: Rename: BitKeeper/deleted/.del-innotest1b.sh~aafd0819ae84da7b -> sql-bench/innotest1b.sh sql-bench/innotest1.sh: Rename: BitKeeper/deleted/.del-innotest1.sh~aa36bce09ca783c7 -> sql-bench/innotest1.sh sql-bench/crash-me.sh: Rename: BitKeeper/deleted/.del-crash-me.sh~2fa881d0b40339c8 -> sql-bench/crash-me.sh sql-bench/graph-compare-results.sh: Rename: BitKeeper/deleted/.del-graph-compare-results.sh~7e4e28b3591b4542 -> sql-bench/graph-compare-results.sh sql-bench/copy-db.sh: Rename: BitKeeper/deleted/.del-copy-db.sh~e8116afb93144ccd -> sql-bench/copy-db.sh sql-bench/compare-results.sh: Rename: BitKeeper/deleted/.del-compare-results.sh~a9e26e2644c694b3 -> sql-bench/compare-results.sh sql-bench/bench-count-distinct.sh: Rename: BitKeeper/deleted/.del-bench-count-distinct.sh~a92f174271a831d7 -> sql-bench/bench-count-distinct.sh sql-bench/as3ap.sh: Rename: BitKeeper/deleted/.del-as3ap.sh~f54eebbd8d34c9b6 -> sql-bench/as3ap.sh sql-bench/Comments/Informix.crash-me: Rename: BitKeeper/deleted/.del-Informix.crash-me~51ab5b717cefe74 -> sql-bench/Comments/Informix.crash-me sql-bench/Comments/postgres.crash-me: Rename: BitKeeper/deleted/.del-postgres.crash-me~eacac145c3e30f17 -> sql-bench/Comments/postgres.crash-me sql-bench/Comments/Empress.crash-me: Rename: BitKeeper/deleted/.del-Empress.crash-me~bdaff0c68ce10f02 -> sql-bench/Comments/Empress.crash-me sql-bench/Comments/Adabas.crash-me: Rename: BitKeeper/deleted/.del-Adabas.crash-me~ce88ba1a540971ac -> sql-bench/Comments/Adabas.crash-me sql-bench/Comments/Access.crash-me: Rename: BitKeeper/deleted/.del-Access.crash-me~bb457ec282d939b6 -> sql-bench/Comments/Access.crash-me sql-bench/Comments/postgres.benchmark: Rename: BitKeeper/deleted/.del-postgres.benchmark~4d30890732b784a -> sql-bench/Comments/postgres.benchmark sql-bench/Comments/mysql.benchmark: Rename: BitKeeper/deleted/.del-mysql.benchmark~4d8729c0937456fc -> sql-bench/Comments/mysql.benchmark sql-bench/Comments/FrontBase.benchmark: Rename: BitKeeper/deleted/.del-FrontBase.benchmark~217041ef18274c2e -> sql-bench/Comments/FrontBase.benchmark sql-bench/Comments/interbase: Rename: BitKeeper/deleted/.del-interbase~cdad59622b4d6f3 -> sql-bench/Comments/interbase sql-bench/uname.bat: Rename: BitKeeper/deleted/.del-uname.bat~a6d933d2ee9314c -> sql-bench/uname.bat sql-bench/pwd.bat: Rename: BitKeeper/deleted/.del-pwd.bat~9b64050849abf51 -> sql-bench/pwd.bat sql-bench/example.bat: Rename: BitKeeper/deleted/.del-example.bat~22d0170bccf0f030 -> sql-bench/example.bat sql-bench/Makefile.am: Rename: BitKeeper/deleted/.del-Makefile.am~7b07da85b2e9375 -> sql-bench/Makefile.am sql-bench/Data/ATIS/transport.txt: Rename: BitKeeper/deleted/.del-transport.txt~fa4ca40735f8354c -> sql-bench/Data/ATIS/transport.txt sql-bench/Data/ATIS/time_zone.txt: Rename: BitKeeper/deleted/.del-time_zone.txt~4171f9ca732f65c0 -> sql-bench/Data/ATIS/time_zone.txt sql-bench/Data/Wisconsin/tenk.data: Rename: BitKeeper/deleted/.del-tenk.data~6aeaebdd534e458e -> sql-bench/Data/Wisconsin/tenk.data sql-bench/Data/Wisconsin/onek.data: Rename: BitKeeper/deleted/.del-onek.data~6cd1edaf596a7f7 -> sql-bench/Data/Wisconsin/onek.data sql-bench/Data/ATIS/stop1.txt: Rename: BitKeeper/deleted/.del-stop1.txt~f09ba164ad44a288 -> sql-bench/Data/ATIS/stop1.txt sql-bench/Data/ATIS/time_interval.txt: Rename: BitKeeper/deleted/.del-time_interval.txt~a1def62e267a59b2 -> sql-bench/Data/ATIS/time_interval.txt sql-bench/Data/ATIS/stop.txt: Rename: BitKeeper/deleted/.del-stop.txt~31fb564e1f415e34 -> sql-bench/Data/ATIS/stop.txt sql-bench/Data/ATIS/restriction.txt: Rename: BitKeeper/deleted/.del-restriction.txt~6ae208924617784a -> sql-bench/Data/ATIS/restriction.txt sql-bench/Data/ATIS/state.txt: Rename: BitKeeper/deleted/.del-state.txt~9dd470ce14075b90 -> sql-bench/Data/ATIS/state.txt sql-bench/Data/ATIS/restrict_class.txt: Rename: BitKeeper/deleted/.del-restrict_class.txt~2f741bf0ea498f84 -> sql-bench/Data/ATIS/restrict_class.txt sql-bench/Data/ATIS/month_name.txt: Rename: BitKeeper/deleted/.del-month_name.txt~4c44f7a323d57d92 -> sql-bench/Data/ATIS/month_name.txt sql-bench/Data/ATIS/restrict_carrier.txt: Rename: BitKeeper/deleted/.del-restrict_carrier.txt~925b5492f3f9cba3 -> sql-bench/Data/ATIS/restrict_carrier.txt sql-bench/Data/ATIS/ground_service.txt: Rename: BitKeeper/deleted/.del-ground_service.txt~1087e477e86e84c -> sql-bench/Data/ATIS/ground_service.txt sql-bench/Data/ATIS/food_service.txt: Rename: BitKeeper/deleted/.del-food_service.txt~66d95a150c28458 -> sql-bench/Data/ATIS/food_service.txt sql-bench/Data/ATIS/flight_day.txt: Rename: BitKeeper/deleted/.del-flight_day.txt~76868d6d265d441e -> sql-bench/Data/ATIS/flight_day.txt sql-bench/Data/ATIS/flight_fare.txt: Rename: BitKeeper/deleted/.del-flight_fare.txt~d7322593c8530487 -> sql-bench/Data/ATIS/flight_fare.txt sql-bench/Data/ATIS/flight_class.txt: Rename: BitKeeper/deleted/.del-flight_class.txt~1801101474c29098 -> sql-bench/Data/ATIS/flight_class.txt sql-bench/Data/ATIS/fconnection.txt: Rename: BitKeeper/deleted/.del-fconnection.txt~e0ef6a8b5560a713 -> sql-bench/Data/ATIS/fconnection.txt sql-bench/Data/ATIS/flight.txt: Rename: BitKeeper/deleted/.del-flight.txt~e5065423760e99eb -> sql-bench/Data/ATIS/flight.txt sql-bench/Data/ATIS/fare.txt: Rename: BitKeeper/deleted/.del-fare.txt~ea0652f490bc24a6 -> sql-bench/Data/ATIS/fare.txt sql-bench/Data/ATIS/day_name.txt: Rename: BitKeeper/deleted/.del-day_name.txt~f813b215955d894c -> sql-bench/Data/ATIS/day_name.txt sql-bench/Data/ATIS/dual_carrier.txt: Rename: BitKeeper/deleted/.del-dual_carrier.txt~a7dd776224fbd92b -> sql-bench/Data/ATIS/dual_carrier.txt sql-bench/Data/ATIS/date_day.txt: Rename: BitKeeper/deleted/.del-date_day.txt~4e9a282fcf54cfd8 -> sql-bench/Data/ATIS/date_day.txt sql-bench/Data/ATIS/compound_class.txt: Rename: BitKeeper/deleted/.del-compound_class.txt~d4a2f1b7f96340b9 -> sql-bench/Data/ATIS/compound_class.txt sql-bench/Data/ATIS/connect_leg.txt: Rename: BitKeeper/deleted/.del-connect_leg.txt~f97b6e94e108bb36 -> sql-bench/Data/ATIS/connect_leg.txt sql-bench/Data/ATIS/code_description.txt: Rename: BitKeeper/deleted/.del-code_description.txt~f9117373e438b0e2 -> sql-bench/Data/ATIS/code_description.txt sql-bench/Data/ATIS/city.txt: Rename: BitKeeper/deleted/.del-city.txt~d96dd6d073344d2e -> sql-bench/Data/ATIS/city.txt sql-bench/Data/ATIS/class_of_service.txt: Rename: BitKeeper/deleted/.del-class_of_service.txt~21f6b9848b8c76d -> sql-bench/Data/ATIS/class_of_service.txt sql-bench/Data/ATIS/airport_service.txt: Rename: BitKeeper/deleted/.del-airport_service.txt~6ee6d5b852b3e38 -> sql-bench/Data/ATIS/airport_service.txt sql-bench/Data/ATIS/airline.txt: Rename: BitKeeper/deleted/.del-airline.txt~a79f8eadf853f2c8 -> sql-bench/Data/ATIS/airline.txt sql-bench/Data/ATIS/airport.txt: Rename: BitKeeper/deleted/.del-airport.txt~59c78514130e1f45 -> sql-bench/Data/ATIS/airport.txt sql-bench/Data/ATIS/aircraft.txt: Rename: BitKeeper/deleted/.del-aircraft.txt~15e4de7ab37c92d3 -> sql-bench/Data/ATIS/aircraft.txt sql-bench/TODO: Rename: BitKeeper/deleted/.del-TODO~cac6d7a63c426ae5 -> sql-bench/TODO sql-bench/limits/sybase.cfg: Rename: BitKeeper/deleted/.del-sybase.cfg~c4636b12767b3f14 -> sql-bench/limits/sybase.cfg sql-bench/limits/solid-nt4.cfg: Rename: BitKeeper/deleted/.del-solid-nt4.cfg~cca779f0c9e29d31 -> sql-bench/limits/solid-nt4.cfg sql-bench/limits/solid.cfg: Rename: BitKeeper/deleted/.del-solid.cfg~5ae0e4342eadb0fb -> sql-bench/limits/solid.cfg sql-bench/limits/pg.cfg: Rename: BitKeeper/deleted/.del-pg.cfg~db59cf39a5d417be -> sql-bench/limits/pg.cfg sql-bench/limits/mysql.cfg: Rename: BitKeeper/deleted/.del-mysql.cfg~9cab20a8771b93cf -> sql-bench/limits/mysql.cfg sql-bench/limits/oracle.cfg: Rename: BitKeeper/deleted/.del-oracle.cfg~affab21af8f438fd -> sql-bench/limits/oracle.cfg sql-bench/limits/mysql-4.1.cfg: Rename: BitKeeper/deleted/.del-mysql-4.1.cfg~b6c5e74aefc99e3c -> sql-bench/limits/mysql-4.1.cfg sql-bench/limits/mysql-3.23.cfg: Rename: BitKeeper/deleted/.del-mysql-3.23.cfg~3cf9d6be54b77a3c -> sql-bench/limits/mysql-3.23.cfg sql-bench/limits/mysql-4.0.cfg: Rename: BitKeeper/deleted/.del-mysql-4.0.cfg~6c9d63c85b5ef574 -> sql-bench/limits/mysql-4.0.cfg sql-bench/limits/mysql-3.22.cfg: Rename: BitKeeper/deleted/.del-mysql-3.22.cfg~e706f26a161175cd -> sql-bench/limits/mysql-3.22.cfg sql-bench/limits/msql.cfg: Rename: BitKeeper/deleted/.del-msql.cfg~52710b12932cceb9 -> sql-bench/limits/msql.cfg sql-bench/limits/ms-sql65.cfg: Rename: BitKeeper/deleted/.del-ms-sql65.cfg~6b9bc3c460dbee05 -> sql-bench/limits/ms-sql65.cfg sql-bench/limits/mimer.cfg: Rename: BitKeeper/deleted/.del-mimer.cfg~234e6c1c3c47b612 -> sql-bench/limits/mimer.cfg sql-bench/limits/ms-sql.cfg: Rename: BitKeeper/deleted/.del-ms-sql.cfg~1907964264d2786a -> sql-bench/limits/ms-sql.cfg sql-bench/limits/interbase.cfg: Rename: BitKeeper/deleted/.del-interbase.cfg~c6951a0376cc6ff3 -> sql-bench/limits/interbase.cfg sql-bench/limits/interbase-dialect3.cfg: Rename: BitKeeper/deleted/.del-interbase-dialect3.cfg~46277bdfc74c667a -> sql-bench/limits/interbase-dialect3.cfg sql-bench/limits/interbase-superserver.cfg: Rename: BitKeeper/deleted/.del-interbase-superserver.cfg~22501198689243b0 -> sql-bench/limits/interbase-superserver.cfg sql-bench/limits/interbase-dialect1.cfg: Rename: BitKeeper/deleted/.del-interbase-dialect1.cfg~659206b5b9a11036 -> sql-bench/limits/interbase-dialect1.cfg sql-bench/limits/empress.cfg: Rename: BitKeeper/deleted/.del-empress.cfg~1f97f34d6560a499 -> sql-bench/limits/empress.cfg sql-bench/limits/frontbase.cfg: Rename: BitKeeper/deleted/.del-frontbase.cfg~71369e9c002696e -> sql-bench/limits/frontbase.cfg sql-bench/limits/db2.cfg: Rename: BitKeeper/deleted/.del-db2.cfg~711099b4d7906959 -> sql-bench/limits/db2.cfg sql-bench/limits/access.cfg: Rename: BitKeeper/deleted/.del-access.cfg~5239ea3655b7bba0 -> sql-bench/limits/access.cfg sql-bench/limits/access_odbc.cfg: Rename: BitKeeper/deleted/.del-access_odbc.cfg~34b4cf0eda56c4b1 -> sql-bench/limits/access_odbc.cfg sql-bench/limits/Informix.cfg: Rename: BitKeeper/deleted/.del-Informix.cfg~b94188e1ececb51b -> sql-bench/limits/Informix.cfg sql-bench/limits/Adabas.cfg: Rename: BitKeeper/deleted/.del-Adabas.cfg~343ed4e7f8b02111 -> sql-bench/limits/Adabas.cfg sql-bench/README: Rename: BitKeeper/deleted/.del-README~b1aa0c1bf9f5eb5e -> sql-bench/README BitKeeper/deleted/.del-create-pg_fast-Linux_2.2.14_my_SMP_i686-crashed: Delete: sql-bench/Results/create-pg_fast-Linux_2.2.14_my_SMP_i686-crashed Makefile.am: Added back sql-bench configure.in: Added back sql-bench dbug/dbug.c: Fixed memory leak in _db_set() This is not a complete fix as we can still get memory leaks if we do any of the following: mysqld --debug --debug CAll DBUG_SET or DBUG_SET_INITIAL more than once in a program Use SET @@global.debug=xxxx At some point we should fix the above, but this is not critical for normal operation (only for debugging) Bug #24497 Valgrind warning: get_one_option Note that all changes to _db_set_ is only a variable access change, and can be ignored while reviewing the bug fix. mysql-test/t/flush_read_lock_kill-master.opt: Added '+' to option, to be able to use this together with --debug mysql-test/valgrind.supp: Removed warning that is now fixed mysys/ptr_cmp.c: Fixed wrong pointer read (probably never used) scripts/make_binary_distribution.sh: Added back sql_bench sql/field.h: Removed class function that was identical in parent class sql/mysqld.cc: Removed some calls to my_thread_init() / my_thread_end() that was not needed. Only call DBUG_SET_INITAL, not DBUG_SET (Fixes memory leak in dbug) Mark some code as deadcode sql/net_serv.cc: Add information about from where error comes sql-bench/bench-init.pl.sh: Cleaned up help text sql-bench/server-cfg.sh: Don't write '/' after server name of no ssl version sql-common/client.c: Don't call net_clear() on COM_QUIT. This avoids a warning from net_clear() after shutdown: "skipped ## bytes from file" BUG#21428: skipped 9 bytes from file: socket (3)" on "mysqladmin shutdown" sql/sql_parse.cc: Added comment sql/table.h: Removed compiler warning
3620 lines
103 KiB
Bash
3620 lines
103 KiB
Bash
#!@PERL@
|
||
# -*- perl -*-
|
||
# Copyright (C) 2000 MySQL AB & MySQL Finland AB & TCX DataKonsult AB
|
||
#
|
||
# This library is free software; you can redistribute it and/or
|
||
# modify it under the terms of the GNU Library General Public
|
||
# License as published by the Free Software Foundation; either
|
||
# version 2 of the License, or (at your option) any later version.
|
||
#
|
||
# This library is distributed in the hope that it will be useful,
|
||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||
# Library General Public License for more details.
|
||
#
|
||
# You should have received a copy of the GNU Library General Public
|
||
# License along with this library; if not, write to the Free
|
||
# Software Foundation, Inc., 59 Temple Place - Suite 330, Boston,
|
||
# MA 02111-1307, USA
|
||
#
|
||
# The configuration file for the DBI/DBD tests on different databases ....
|
||
# You will need the DBD module for the database you are running.
|
||
# Monty made this bench script and I (Luuk de Boer) rewrote it to DBI/DBD.
|
||
# Monty rewrote this again to use packages.
|
||
#
|
||
# Each database has a different package that has 3 functions:
|
||
# new Creates a object with some standard slot
|
||
# version Version number of the server
|
||
# create Generates commands to create a table
|
||
#
|
||
|
||
#
|
||
# First some global functions that help use the packages:
|
||
#
|
||
|
||
sub get_server
|
||
{
|
||
my ($name,$host,$database,$odbc,$machine,$socket,$connect_options)=@_;
|
||
my ($server);
|
||
if ($name =~ /mysql/i)
|
||
{ $server=new db_MySQL($host, $database, $machine, $socket,$connect_options); }
|
||
elsif ($name =~ /pg/i)
|
||
{ $server= new db_Pg($host,$database); }
|
||
elsif ($name =~ /msql/i)
|
||
{ $server= new db_mSQL($host,$database); }
|
||
elsif ($name =~ /solid/i)
|
||
{ $server= new db_Solid($host,$database); }
|
||
elsif ($name =~ /Empress/i)
|
||
{ $server= new db_Empress($host,$database); }
|
||
elsif ($name =~ /FrontBase/i)
|
||
{ $server= new db_FrontBase($host,$database); }
|
||
elsif ($name =~ /Oracle/i)
|
||
{ $server= new db_Oracle($host,$database); }
|
||
elsif ($name =~ /Access/i)
|
||
{ $server= new db_access($host,$database); }
|
||
elsif ($name =~ /Informix/i)
|
||
{ $server= new db_Informix($host,$database); }
|
||
elsif ($name =~ /ms-sql/i)
|
||
{ $server= new db_ms_sql($host,$database); }
|
||
elsif ($name =~ /sybase/i)
|
||
{ $server= new db_sybase($host,$database); }
|
||
elsif ($name =~ /Adabas/i) # Adabas has two drivers
|
||
{
|
||
$server= new db_Adabas($host,$database);
|
||
if ($name =~ /AdabasD/i)
|
||
{
|
||
$server->{'data_source'} =~ s/:Adabas:/:AdabasD:/;
|
||
}
|
||
}
|
||
elsif ($name =~ /DB2/i)
|
||
{ $server= new db_db2($host,$database); }
|
||
elsif ($name =~ /Mimer/i)
|
||
{ $server= new db_Mimer($host,$database); }
|
||
elsif ($name =~ /Sapdb/i)
|
||
{ $server= new db_sapdb($host,$database); }
|
||
elsif ($name =~ /interBase/i)
|
||
{ $server= new db_interbase($host,$database); }
|
||
else
|
||
{
|
||
die "Unknown sql server name used: $name\nUse one of: Access, Adabas, AdabasD, Empress, FrontBase, Oracle, Informix, InterBase, DB2, mSQL, Mimer, MS-SQL, MySQL, Pg, Solid, SAPDB or Sybase.\nIf the connection is done trough ODBC the name must end with _ODBC\n";
|
||
}
|
||
if ($name =~ /_ODBC$/i || defined($odbc) && $odbc)
|
||
{
|
||
if (! ($server->{'data_source'} =~ /^([^:]*):([^:]+):([^:]*)/ ))
|
||
{
|
||
die "Can't find databasename in data_source: '" .
|
||
$server->{'data_source'}. "'\n";
|
||
}
|
||
if ($3) {
|
||
$server->{'data_source'} = "$1:ODBC:$3";
|
||
} else {
|
||
$server->{'data_source'} = "$1:ODBC:$database";
|
||
}
|
||
}
|
||
return $server;
|
||
}
|
||
|
||
sub all_servers
|
||
{
|
||
return ["Access", "Adabas", "DB2", "Empress", "FrontBase", "Oracle",
|
||
"Informix", "InterBase", "Mimer", "mSQL", "MS-SQL", "MySQL", "Pg","SAPDB",
|
||
"Solid", "Sybase"];
|
||
}
|
||
|
||
#############################################################################
|
||
# First the configuration for MySQL off course :-)
|
||
#############################################################################
|
||
|
||
package db_MySQL;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database,$machine,$socket,$connect_options)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "mysql";
|
||
$self->{'data_source'} = "DBI:mysql:database=$database;host=$host";
|
||
$self->{'data_source'} .= ";mysql_socket=$socket" if($socket);
|
||
$self->{'data_source'} .= ";$connect_options" if($connect_options);
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "blob";
|
||
$self->{'text'} = "text";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'vacuum'} = 1; # When using with --fast
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 0; # Transactions disabled by default
|
||
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'alter_add_multi_col'}= 1; #Have ALTER TABLE t add a int,add b int;
|
||
$limits{'alter_table'} = 1; # Have ALTER TABLE
|
||
$limits{'alter_table_dropcol'}= 1; # Have ALTER TABLE DROP column
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'func_extra_%'} = 1; # Has % as alias for mod()
|
||
$limits{'func_extra_if'} = 1; # Have function if.
|
||
$limits{'func_extra_in_num'} = 1; # Has function in
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_odbc_mod'} = 1; # Have function mod.
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_by_position'} = 1; # Can use 'GROUP BY 1'
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'group_func_extra_std'} = 1; # Have group function std().
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'having_with_alias'} = 1; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can use group functions in HAVING
|
||
$limits{'insert_multi_value'} = 1; # Have INSERT ... values (1,2),(3,4)
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'limit'} = 1; # supports the limit attribute
|
||
$limits{'truncate_table'} = 1;
|
||
$limits{'load_data_infile'} = 1; # Has load data infile
|
||
$limits{'lock_tables'} = 1; # Has lock tables
|
||
$limits{'max_column_name'} = 64; # max table and column name
|
||
$limits{'max_columns'} = 2000; # Max number of columns in table
|
||
$limits{'max_conditions'} = 9999; # (Actually not a limit)
|
||
$limits{'max_index'} = 16; # Max number of keys
|
||
$limits{'max_index_parts'} = 16; # Max segments/key
|
||
$limits{'max_tables'} = (($machine || '') =~ "^win") ? 5000 : 65000;
|
||
$limits{'max_text_size'} = 1000000; # Good enough for tests
|
||
$limits{'multi_drop'} = 1; # Drop table can take many tables
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'query_size'} = 1000000; # Max size with default buffers.
|
||
$limits{'select_without_from'}= 1; # Can do 'select 1';
|
||
$limits{'subqueries'} = 0; # Doesn't support sub-queries.
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
# Some fixes that depends on the environment
|
||
if (defined($main::opt_create_options) &&
|
||
$main::opt_create_options =~ /engine=heap/i)
|
||
{
|
||
$limits{'working_blobs'} = 0; # HEAP tables can't handle BLOB's
|
||
}
|
||
if (defined($main::opt_create_options) &&
|
||
$main::opt_create_options =~ /engine=innodb/i)
|
||
{
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
}
|
||
if (defined($main::opt_create_options) &&
|
||
$main::opt_create_options =~ /engine=ndb/i)
|
||
{
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$limits{'max_columns'} = 90; # Max number of columns in table
|
||
$limits{'max_tables'} = 32; # No comments
|
||
}
|
||
if (defined($main::opt_create_options) &&
|
||
$main::opt_create_options =~ /engine=bdb/i)
|
||
{
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
}
|
||
if (defined($main::opt_create_options) &&
|
||
$main::opt_create_options =~ /engine=gemini/i)
|
||
{
|
||
$limits{'working_blobs'} = 0; # Blobs not implemented yet
|
||
$limits{'max_tables'} = 500;
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
}
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare("select VERSION()") or die $DBI::errstr;
|
||
$version="MySQL 3.20.?";
|
||
if ($sth->execute && (@row = $sth->fetchrow_array))
|
||
{
|
||
$row[0] =~ s/-/ /g; # To get better tables with long names
|
||
$version="MySQL $row[0]";
|
||
}
|
||
$sth->finish;
|
||
|
||
$sth = $dbh->prepare("show status like 'ssl_version'") or die $DBI::errstr;
|
||
if ($sth->execute && (@row = $sth->fetchrow_array) && $row[1])
|
||
{
|
||
$version .= "/$row[1]";
|
||
}
|
||
$sth->finish;
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
#
|
||
# Connection with optional disabling of logging
|
||
#
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
|
||
$dbh->do("SET OPTION LOG_OFF=1,UPDATE_LOG=0");
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
# If one uses $main::opt_fast then one is allowed to use
|
||
# non standard types to get better speed.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index,$options) = @_;
|
||
my($query,@queries);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
# $field =~ s/ decimal/ double(10,2)/i;
|
||
$field =~ s/ big_decimal/ double(10,2)/i;
|
||
$query.= $field . ',';
|
||
}
|
||
foreach $index (@$index)
|
||
{
|
||
$query.= $index . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
$query.=" $options" if (defined($options));
|
||
$query.=" $main::opt_create_options" if (defined($main::opt_create_options));
|
||
push(@queries,$query);
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my ($self,$dbname, $file, $dbh) = @_;
|
||
my ($command, $sth);
|
||
|
||
$file =~ s|\\|/|g; # Change Win32 names to Unix syntax
|
||
$command = "load data infile '$file' into table $dbname columns optionally enclosed by '\\'' terminated by ','";
|
||
# print "$command\n";
|
||
$sth = $dbh->do($command) or die $DBI::errstr;
|
||
return $sth; # Contains number of rows
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index {
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index ON $table";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# This should return 1 if we to do disconnect / connect when doing
|
||
# big batches
|
||
#
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# reconnect on errors (needed mainly be crash-me)
|
||
#
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#
|
||
# Optimize tables for better performance
|
||
#
|
||
|
||
sub vacuum
|
||
{
|
||
my ($self,$full_vacuum,$dbh_ref,@tables)=@_;
|
||
my ($loop_time,$end_time,$dbh);
|
||
if ($#tables >= 0)
|
||
{
|
||
$dbh=$$dbh_ref;
|
||
$loop_time=new Benchmark;
|
||
$dbh->do("OPTIMIZE TABLE " . join(',',@tables)) || die "Got error: $DBI::errstr when executing 'OPTIMIZE TABLE'\n";
|
||
$end_time=new Benchmark;
|
||
print "Time for book-keeping (1): " .
|
||
Benchmark::timestr(Benchmark::timediff($end_time, $loop_time),"all") . "\n\n";
|
||
}
|
||
}
|
||
|
||
#############################################################################
|
||
# Definitions for mSQL
|
||
#############################################################################
|
||
|
||
package db_mSQL;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "msql";
|
||
$self->{'data_source'} = "DBI:mSQL:$database:$host";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'double_quotes'} = 0;
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 0; # No transactions
|
||
$self->{'blob'} = "text(" . $limits{'max_text_size'} .")";
|
||
$self->{'text'} = "text(" . $limits{'max_text_size'} .")";
|
||
|
||
$limits{'max_conditions'} = 74;
|
||
$limits{'max_columns'} = 75;
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 32000;
|
||
$limits{'query_size'} = 65535;
|
||
$limits{'max_index'} = 5;
|
||
$limits{'max_index_parts'} = 10;
|
||
$limits{'max_column_name'} = 35;
|
||
|
||
$limits{'join_optimizer'} = 0; # Can't optimize FROM tables
|
||
$limits{'load_data_infile'} = 0;
|
||
$limits{'lock_tables'} = 0;
|
||
$limits{'functions'} = 0;
|
||
$limits{'group_functions'} = 0;
|
||
$limits{'group_distinct_functions'}= 0; # Have count(distinct)
|
||
$limits{'multi_drop'} = 0;
|
||
$limits{'select_without_from'}= 0;
|
||
$limits{'subqueries'} = 0;
|
||
$limits{'left_outer_join'} = 0;
|
||
$limits{'table_wildcard'} = 0;
|
||
$limits{'having_with_alias'} = 0;
|
||
$limits{'having_with_group'} = 0;
|
||
$limits{'like_with_column'} = 1;
|
||
$limits{'order_by_position'} = 1;
|
||
$limits{'group_by_position'} = 1;
|
||
$limits{'alter_table'} = 0;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
$limits{'group_func_extra_std'} = 0;
|
||
$limits{'limit'} = 1; # supports the limit attribute
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 0;
|
||
|
||
$limits{'func_odbc_mod'} = 0;
|
||
$limits{'func_extra_%'} = 0;
|
||
$limits{'func_odbc_floor'} = 0;
|
||
$limits{'func_extra_if'} = 0;
|
||
$limits{'column_alias'} = 0;
|
||
$limits{'NEG'} = 0;
|
||
$limits{'func_extra_in_num'} = 0;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($tmp,$dir);
|
||
foreach $dir ("/usr/local/Hughes", "/usr/local/mSQL","/my/local/mSQL",
|
||
"/usr/local")
|
||
{
|
||
if (-x "$dir/bin/msqladmin")
|
||
{
|
||
$tmp=`$dir/bin/msqladmin version | grep server`;
|
||
if ($tmp =~ /^\s*(.*\w)\s*$/)
|
||
{ # Strip pre- and endspace
|
||
$tmp=$1;
|
||
$tmp =~ s/\s+/ /g; # Remove unnecessary spaces
|
||
$tmp .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
|
||
return $tmp;
|
||
}
|
||
}
|
||
}
|
||
return "mSQL version ???";
|
||
}
|
||
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Can't handle many field types, so we map everything to int and real.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$name,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/varchar/char/i; # mSQL doesn't have VARCHAR()
|
||
# mSQL can't handle more than the real basic int types
|
||
$field =~ s/tinyint|smallint|mediumint|integer/int/i;
|
||
# mSQL can't handle different visual lengths
|
||
$field =~ s/int\(\d*\)/int/i;
|
||
# mSQL doesn't have float, change it to real
|
||
$field =~ s/float(\(\d*,\d*\)){0,1}/real/i;
|
||
$field =~ s/double(\(\d*,\d*\)){0,1}/real/i;
|
||
# mSQL doesn't have blob, it has text instead
|
||
if ($field =~ / blob/i)
|
||
{
|
||
$name=$self->{'blob'};
|
||
$field =~ s/ blob/ $name/;
|
||
}
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
|
||
# Prepend table_name to index name because the the name may clash with
|
||
# a field name. (Should be diffent name space, but this is mSQL...)
|
||
|
||
foreach $index (@$index)
|
||
{
|
||
# Primary key is unique index in mSQL
|
||
$index =~ s/primary key/unique index primary/i;
|
||
if ($index =~ /^unique\s*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
push(@queries,"create unique index ${table_name}_$nr on $table_name ($1)");
|
||
}
|
||
else
|
||
{
|
||
if (!($index =~ /^(.*index)\s+(\w*)\s+(\(.*\))$/i))
|
||
{
|
||
die "Can't parse index information in '$index'\n";
|
||
}
|
||
push(@queries,"create $1 ${table_name}_$2 on $table_name $3");
|
||
}
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert an ascii file isn't supported by mSQL\n";
|
||
return 0;
|
||
}
|
||
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index FROM $table";
|
||
}
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Definitions for PostgreSQL #
|
||
#############################################################################
|
||
|
||
package db_Pg;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "pg";
|
||
$self->{'data_source'} = "DBI:Pg:dbname=$database";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "text";
|
||
$self->{'text'} = "text";
|
||
$self->{'double_quotes'} = 1;
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{"vacuum"} = 1;
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0;
|
||
|
||
$limits{'NEG'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0; # alter_add_multi_col ?
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
$limits{'column_alias'} = 1;
|
||
$limits{'func_extra_%'} = 1;
|
||
$limits{'func_extra_if'} = 0;
|
||
$limits{'func_extra_in_num'} = 1;
|
||
$limits{'func_odbc_floor'} = 1;
|
||
$limits{'func_odbc_mod'} = 1; # Has %
|
||
$limits{'functions'} = 1;
|
||
$limits{'group_by_position'} = 1;
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'group_func_extra_std'} = 0;
|
||
$limits{'group_func_sql_min_str'}= 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_functions'} = 1;
|
||
$limits{'having_with_alias'} = 0;
|
||
$limits{'having_with_group'} = 1;
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'left_outer_join'} = 1;
|
||
$limits{'like_with_column'} = 1;
|
||
$limits{'lock_tables'} = 0; # in ATIS gives this a problem
|
||
$limits{'max_column_name'} = 128;
|
||
$limits{'max_columns'} = 1000; # 500 crashes pg 6.3
|
||
$limits{'max_conditions'} = 9999; # This makes Pg real slow
|
||
$limits{'max_index'} = 64; # Big enough
|
||
$limits{'max_index_parts'} = 16;
|
||
$limits{'max_tables'} = 5000; # 10000 crashes pg 7.0.2
|
||
$limits{'max_text_size'} = 65000; # Good enough for test
|
||
$limits{'multi_drop'} = 1;
|
||
$limits{'order_by_position'} = 1;
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'query_size'} = 16777216;
|
||
$limits{'select_without_from'}= 1;
|
||
$limits{'subqueries'} = 1;
|
||
$limits{'table_wildcard'} = 1;
|
||
$limits{'truncate_table'} = 1;
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
return $self;
|
||
}
|
||
|
||
# couldn't find the option to get the version number
|
||
|
||
sub version
|
||
{
|
||
my ($version,$dir);
|
||
$version = "PostgreSQL version ???";
|
||
foreach $dir ($ENV{'PGDATA'},"/usr/local/pgsql/data", "/usr/local/pg/data")
|
||
{
|
||
if ($dir && -e "$dir/PG_VERSION")
|
||
{
|
||
$version= `cat $dir/PG_VERSION`;
|
||
if ($? == 0)
|
||
{
|
||
chomp($version);
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return "PostgreSQL $version";
|
||
}
|
||
}
|
||
}
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$name,$in,$indfield,$table,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
if ($main::opt_fast)
|
||
{
|
||
# Allow use of char2, char4, char8 or char16
|
||
$field =~ s/char(2|4|8|16)/char$1/;
|
||
}
|
||
# Pg can't handle more than the real basic int types
|
||
$field =~ s/tinyint|smallint|mediumint|integer/int/;
|
||
# Pg can't handle different visual lengths
|
||
$field =~ s/int\(\d*\)/int/;
|
||
$field =~ s/float\(\d*,\d*\)/float/;
|
||
$field =~ s/ double/ float/;
|
||
# $field =~ s/ decimal/ float/i;
|
||
# $field =~ s/ big_decimal/ float/i;
|
||
# $field =~ s/ date/ int/i;
|
||
# Pg doesn't have blob, it has text instead
|
||
$field =~ s/ blob/ text/;
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
foreach $index (@$index)
|
||
{
|
||
$index =~ s/primary key/unique index primary_key/i;
|
||
if ($index =~ /^unique.*\(([^\(]*)\)$/i)
|
||
{
|
||
# original: $indfield="using btree (" .$1.")";
|
||
# using btree doesn<73>t seem to work with Postgres anymore; it creates
|
||
# the table and adds the index, but it isn<73>t unique
|
||
$indfield=" (" .$1.")";
|
||
$in="unique index";
|
||
$table="index_$nr"; $nr++;
|
||
}
|
||
elsif ($index =~ /^(.*index)\s+(\w*)\s+(\(.*\))$/i)
|
||
{
|
||
# original: $indfield="using btree (" .$1.")";
|
||
$indfield=" " .$3;
|
||
$in="index";
|
||
$table="index_$nr"; $nr++;
|
||
}
|
||
else
|
||
{
|
||
die "Can't parse index information in '$index'\n";
|
||
}
|
||
push(@queries,"create $in ${table_name}_$table on $table_name $indfield");
|
||
}
|
||
$queries[0]=$query;
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my ($self,$dbname, $file, $dbh) = @_;
|
||
my ($command, $sth);
|
||
|
||
# Syntax:
|
||
# copy [binary] <class_name> [with oids]
|
||
# {to|from} {<filename>|stdin|stdout} [using delimiters <delim>]
|
||
print "The ascii files aren't correct for postgres ....!!!\n";
|
||
$command = "copy $dbname from '$file' using delimiters ','";
|
||
print "$command\n";
|
||
$sth = $dbh->do($command) or die $DBI::errstr;
|
||
return $sth;
|
||
}
|
||
|
||
#
|
||
# As postgreSQL wants A % B instead of standard mod(A,B) we have to map
|
||
# This will not handle all cases, but as the benchmarks doesn't use functions
|
||
# inside MOD() the following should work
|
||
#
|
||
# PostgreSQL cant handle count(*) or even count(1), but it can handle
|
||
# count(1+1) sometimes. ==> this is solved in PostgreSQL 6.3
|
||
#
|
||
# PostgreSQL 6.5 is supporting MOD.
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
my(@select,$change);
|
||
# if you use PostgreSQL 6.x and x is lower as 5 then uncomment the line below.
|
||
# $sql =~ s/mod\(([^,]*),([^\)]*)\)/\($1 % $2\)/gi;
|
||
#
|
||
# if you use PostgreSQL 6.1.x uncomment the lines below
|
||
# if ($sql =~ /select\s+count\(\*\)\s+from/i) {
|
||
# }
|
||
# elsif ($sql =~ /count\(\*\)/i)
|
||
# {
|
||
# if ($sql =~ /select\s+(.*)\s+from/i)
|
||
# {
|
||
# @select = split(/,/,$1);
|
||
# if ($select[0] =~ /(.*)\s+as\s+\w+$/i)
|
||
# {
|
||
# $change = $1;
|
||
# }
|
||
# else
|
||
# {
|
||
# $change = $select[0];
|
||
# }
|
||
# }
|
||
# if (($change =~ /count/i) || ($change eq "")) {
|
||
# $change = "1+1";
|
||
# }
|
||
# $sql =~ s/count\(\*\)/count($change)/gi;
|
||
# }
|
||
# till here.
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 1 if ($DBI::errstr =~ /sent to backend, but backend closed/i);
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
sub vacuum
|
||
{
|
||
my ($self,$full_vacuum,$dbh_ref,@tables)=@_;
|
||
my ($loop_time,$end_time,$dbh,$table);
|
||
if (defined($full_vacuum))
|
||
{
|
||
$$dbh_ref->disconnect; $$dbh_ref= $self->connect();
|
||
}
|
||
$dbh=$$dbh_ref;
|
||
$loop_time=new Benchmark;
|
||
if ($#tables >= 0)
|
||
{
|
||
foreach $table (@tables)
|
||
{
|
||
$dbh->do("vacuum analyze $table") || die "Got error: $DBI::errstr when executing 'vacuum analyze $table'\n";
|
||
$dbh->do("vacuum $table") || die "Got error: $DBI::errstr when executing 'vacuum'\n";
|
||
}
|
||
}
|
||
else
|
||
{
|
||
# $dbh->do("vacuum pg_attributes") || die "Got error: $DBI::errstr when executing 'vacuum'\n";
|
||
# $dbh->do("vacuum pg_index") || die "Got error: $DBI::errstr when executing 'vacuum'\n";
|
||
$dbh->do("vacuum analyze") || die "Got error: $DBI::errstr when executing 'vacuum analyze'\n";
|
||
$dbh->do("vacuum") || die "Got error: $DBI::errstr when executing 'vacuum'\n";
|
||
}
|
||
$end_time=new Benchmark;
|
||
print "Time for book-keeping (1): " .
|
||
Benchmark::timestr(Benchmark::timediff($end_time, $loop_time),"all") . "\n\n";
|
||
$dbh->disconnect; $$dbh_ref= $self->connect();
|
||
}
|
||
|
||
|
||
#############################################################################
|
||
# Definitions for Solid
|
||
#############################################################################
|
||
|
||
package db_Solid;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "solid";
|
||
$self->{'data_source'} = "DBI:Solid:";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "long varchar";
|
||
$self->{'text'} = "long varchar";
|
||
$self->{'double_quotes'} = 1;
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
|
||
$limits{'max_conditions'} = 9999; # Probably big enough
|
||
$limits{'max_columns'} = 2000; # From crash-me
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 65492; # According to tests
|
||
$limits{'query_size'} = 65535; # Probably a limit
|
||
$limits{'max_index'} = 64; # Probably big enough
|
||
$limits{'max_index_parts'} = 64;
|
||
$limits{'max_column_name'} = 80;
|
||
|
||
$limits{'join_optimizer'} = 1;
|
||
$limits{'load_data_infile'} = 0;
|
||
$limits{'lock_tables'} = 0;
|
||
$limits{'functions'} = 1;
|
||
$limits{'group_functions'} = 1;
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0; # Can do 'select 1' ?;
|
||
$limits{'multi_drop'} = 0;
|
||
$limits{'subqueries'} = 1;
|
||
$limits{'left_outer_join'} = 1;
|
||
$limits{'table_wildcard'} = 1;
|
||
$limits{'having_with_alias'} = 0;
|
||
$limits{'having_with_group'} = 1;
|
||
$limits{'like_with_column'} = 1;
|
||
$limits{'order_by_position'} = 0; # 2.30.0018 can this
|
||
$limits{'group_by_position'} = 0;
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 1;
|
||
$limits{'func_extra_%'} = 0;
|
||
$limits{'func_odbc_floor'} = 1;
|
||
$limits{'column_alias'} = 1;
|
||
$limits{'NEG'} = 1;
|
||
$limits{'func_extra_in_num'} = 1;
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($version,$dir);
|
||
$version="Solid version ??";
|
||
foreach $dir ($ENV{'SOLIDDIR'},"/usr/local/solid", "/my/local/solid")
|
||
{
|
||
if ($dir && -e "$dir/bin/solcon")
|
||
{
|
||
$version=`$dir/bin/solcon -e"ver" $main::opt_user $main::opt_password | grep Server | sed q`;
|
||
if ($? == 0)
|
||
{
|
||
chomp($version);
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
}
|
||
}
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/mediumint/integer/i;
|
||
$field =~ s/ double/ float/i;
|
||
# Solid doesn't have blob, it has long varchar
|
||
$field =~ s/ blob/ long varchar/;
|
||
# $field =~ s/ decimal/ float/i;
|
||
# $field =~ s/ big_decimal/ float/i;
|
||
# $field =~ s/ date/ int/i;
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
foreach $index (@$index)
|
||
{
|
||
if ($index =~ /^primary key/i || $index =~ /^unique/i)
|
||
{ # Add to create statement
|
||
substr($queries[0],-1,0)="," . $index;
|
||
}
|
||
else
|
||
{
|
||
$index =~ /^(.*)\s+(\(.*\))$/;
|
||
push(@queries,"create ${1}$nr on $table_name $2");
|
||
$nr++;
|
||
}
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
# there is no sql statement in solid which can do the load from
|
||
# an ascii file in the db ... but there is the speedloader program
|
||
# an external program which can load the ascii file in the db ...
|
||
# the server must be down before using speedloader !!!!
|
||
# (in the standalone version)
|
||
# it works also with a control file ... that one must be made ....
|
||
sub insert_file {
|
||
my ($self, $dbname, $file) = @_;
|
||
my ($speedcmd);
|
||
$speedcmd = '/usr/local/solid/bin/solload';
|
||
print "At this moment not supported - solid server must go down \n";
|
||
return 0;
|
||
}
|
||
|
||
# solid can't handle an alias in a having statement so
|
||
# select test as foo from tmp group by foo having foor > 2
|
||
# becomes
|
||
# select test as foo from tmp group by foo having test > 2
|
||
#
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
my(@select,$tmp,$newhaving,$key,%change);
|
||
|
||
if ($sql =~ /having\s+/i)
|
||
{
|
||
if ($sql =~ /select (.*) from/i)
|
||
{
|
||
(@select) = split(/,\s*/, $1);
|
||
foreach $tmp (@select)
|
||
{
|
||
if ($tmp =~ /(.*)\s+as\s+(\w+)/)
|
||
{
|
||
$change{$2} = $1;
|
||
}
|
||
}
|
||
}
|
||
if ($sql =~ /having\s+(\w+)/i)
|
||
{
|
||
$newhaving = $1;
|
||
foreach $key (sort {$a cmp $b} keys %change)
|
||
{
|
||
if ($newhaving eq $key)
|
||
{
|
||
$newhaving =~ s/$key/$change{$key}/g;
|
||
}
|
||
}
|
||
}
|
||
$sql =~ s/(having)\s+(\w+)/$1 $newhaving/i;
|
||
}
|
||
return $sql;
|
||
}
|
||
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
#############################################################################
|
||
# Definitions for Empress
|
||
#
|
||
# at this moment DBI:Empress can only handle 200 prepare statements ...
|
||
# so Empress can't be tested with the benchmark test :(
|
||
#############################################################################
|
||
|
||
package db_Empress;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "empress";
|
||
$self->{'data_source'} = "DBI:EmpressNet:SERVER=$host;Database=/usr/local/empress/rdbms/bin/$database";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "text";
|
||
$self->{'text'} = "text";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
|
||
$limits{'max_conditions'} = 1258;
|
||
$limits{'max_columns'} = 226; # server is disconnecting????
|
||
# above this value .... but can handle 2419 columns
|
||
# maybe something for crash-me ... but how to check ???
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 4095; # max returned ....
|
||
$limits{'query_size'} = 65535; # Not a limit, big enough
|
||
$limits{'max_index'} = 64; # Big enough
|
||
$limits{'max_index_parts'} = 64; # Big enough
|
||
$limits{'max_column_name'} = 31;
|
||
|
||
$limits{'join_optimizer'} = 1;
|
||
$limits{'load_data_infile'} = 0;
|
||
$limits{'lock_tables'} = 1;
|
||
$limits{'functions'} = 1;
|
||
$limits{'group_functions'} = 1;
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0;
|
||
$limits{'multi_drop'} = 0;
|
||
$limits{'subqueries'} = 1;
|
||
$limits{'table_wildcard'} = 0;
|
||
$limits{'having_with_alias'} = 0; # AS isn't supported in a select
|
||
$limits{'having_with_group'} = 1;
|
||
$limits{'like_with_column'} = 1;
|
||
$limits{'order_by_position'} = 1;
|
||
$limits{'group_by_position'} = 0;
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
|
||
$limits{'group_func_extra_std'}= 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0;
|
||
$limits{'func_extra_%'} = 1;
|
||
$limits{'func_odbc_floor'} = 1;
|
||
$limits{'func_extra_if'} = 0;
|
||
$limits{'column_alias'} = 0;
|
||
$limits{'NEG'} = 1;
|
||
$limits{'func_extra_in_num'} = 0;
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self,$dbh)=@_;
|
||
my ($version);
|
||
$version="";
|
||
if (-x "/usr/local/empress/rdbms/bin/empvers")
|
||
{
|
||
$version=`/usr/local/empress/rdbms/bin/empvers | grep Version`;
|
||
}
|
||
if ($version)
|
||
{
|
||
chomp($version);
|
||
}
|
||
else
|
||
{
|
||
$version="Empress version ???";
|
||
}
|
||
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
my($command,$sth);
|
||
$command = "insert into $dbname from '$file'";
|
||
print "$command\n" if ($opt_debug);
|
||
$sth = $dbh->do($command) or die $DBI::errstr;
|
||
|
||
return $sth;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/mediumint/int/i;
|
||
$field =~ s/tinyint/int/i;
|
||
$field =~ s/smallint/int/i;
|
||
$field =~ s/longint/int/i;
|
||
$field =~ s/integer/int/i;
|
||
$field =~ s/ double/ longfloat/i;
|
||
# Solid doesn't have blob, it has long varchar
|
||
# $field =~ s/ blob/ text(65535,65535,65535,65535)/;
|
||
$field =~ s/ blob/ text/;
|
||
$field =~ s/ varchar\((\d+)\)/ char($1,3)/;
|
||
$field =~ s/ char\((\d+)\)/ char($1,3)/;
|
||
# $field =~ s/ decimal/ float/i;
|
||
# $field =~ s/ big_decimal/ longfloat/i;
|
||
# $field =~ s/ date/ int/i;
|
||
$field =~ s/ float(.*)/ float/i;
|
||
if ($field =~ / int\((\d+)\)/) {
|
||
if ($1 > 4) {
|
||
$field =~ s/ int\(\d+\)/ longinteger/i;
|
||
} else {
|
||
$field =~ s/ int\(\d+\)/ longinteger/i;
|
||
}
|
||
} else {
|
||
$field =~ s/ int/ longinteger/i;
|
||
}
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=1;
|
||
foreach $index (@$index)
|
||
{
|
||
# Primary key is unique index in Empress
|
||
$index =~ s/primary key/unique index/i;
|
||
if ($index =~ /^unique.*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
push(@queries,"create unique index ${table_name}_$nr on $table_name ($1)");
|
||
}
|
||
else
|
||
{
|
||
if (!($index =~ /^(.*index)\s+(\w*)\s+(\(.*\))$/i))
|
||
{
|
||
die "Can't parse index information in '$index'\n";
|
||
}
|
||
push(@queries,"create $1 ${table_name}_$2 on $table_name $3");
|
||
}
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
# empress can't handle an alias and but can handle the number of the
|
||
# columname - so
|
||
# select test as foo from tmp order by foo
|
||
# becomes
|
||
# select test from tmp order by 1
|
||
#
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
my(@select,$i,$tmp,$newselect,$neworder,@order,$key,%change);
|
||
my($tmp1,$otmp,$tmp2);
|
||
|
||
if ($sql =~ /\s+as\s+/i)
|
||
{
|
||
if ($sql =~ /select\s+(.*)\s+from/i) {
|
||
$newselect = $1;
|
||
(@select) = split(/,\s*/, $1);
|
||
$i = 1;
|
||
foreach $tmp (@select) {
|
||
if ($tmp =~ /\s+as\s+(\w+)/) {
|
||
$change{$1} = $i;
|
||
}
|
||
$i++;
|
||
}
|
||
}
|
||
$newselect =~ s/\s+as\s+(\w+)//gi;
|
||
$tmp2 = 0;
|
||
if ($sql =~ /order\s+by\s+(.*)$/i) {
|
||
(@order) = split(/,\s*/, $1);
|
||
foreach $otmp (@order) {
|
||
foreach $key (sort {$a cmp $b} keys %change) {
|
||
if ($otmp eq $key) {
|
||
$neworder .= "$tmp1"."$change{$key}";
|
||
$tmp1 = ", ";
|
||
$tmp2 = 1;
|
||
} elsif ($otmp =~ /(\w+)\s+(.+)$/) {
|
||
if ($key eq $1) {
|
||
$neworder .= "$tmp1"."$change{$key} $2";
|
||
$tmp2 = 1;
|
||
}
|
||
}
|
||
}
|
||
if ($tmp2 == 0) {
|
||
$neworder .= "$tmp1"."$otmp";
|
||
}
|
||
$tmp2 = 0;
|
||
$tmp1 = ", ";
|
||
}
|
||
}
|
||
$sql =~ s/(select)\s+(.*)\s+(from)/$1 $newselect $3/i;
|
||
$sql =~ s/(order\s+by)\s+(.*)$/$1 $neworder/i;
|
||
}
|
||
return $sql;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
$cmd =~ s/\'\'/\' \'/g;
|
||
return $cmd;
|
||
}
|
||
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
# This is a because of the 200 statement problem with DBI-Empress
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
if ($DBI::errstr =~ /Overflow of table of prepared statements/i)
|
||
{
|
||
print "Overflow of prepared statements ... killing the process\n";
|
||
exit 1;
|
||
}
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
#############################################################################
|
||
# Definitions for Oracle
|
||
#############################################################################
|
||
|
||
package db_Oracle;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "Oracle";
|
||
$self->{'data_source'} = "DBI:Oracle:$database";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "long";
|
||
$self->{'text'} = "long";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{"vacuum"} = 1;
|
||
|
||
$limits{'max_conditions'} = 9999; # (Actually not a limit)
|
||
$limits{'max_columns'} = 254; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 2000; # Limit for blob test-connect
|
||
$limits{'query_size'} = 65525; # Max size with default buffers.
|
||
$limits{'max_index'} = 16; # Max number of keys
|
||
$limits{'max_index_parts'} = 16; # Max segments/key
|
||
$limits{'max_column_name'} = 32; # max table and column name
|
||
|
||
$limits{'truncate_table'} = 1;
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0;
|
||
$limits{'multi_drop'} = 0;
|
||
$limits{'subqueries'} = 1;
|
||
$limits{'left_outer_join'} = 0; # This may be fixed in the query module
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can't use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0;
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Oracle has problem with mod()
|
||
$limits{'func_extra_%'} = 0; # Has % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 1; # Has function in
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare("select VERSION from product_component_version WHERE PRODUCT like 'Oracle%'") or die $DBI::errstr;
|
||
$version="Oracle 7.x";
|
||
if ($sth->execute && (@row = $sth->fetchrow_array))
|
||
{
|
||
$version="Oracle $row[0]";
|
||
}
|
||
$sth->finish;
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
# If one uses $main::opt_fast then one is allowed to use
|
||
# non standard types to get better speed.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$ind,@keys);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/ character\((\d+)\)/ char\($1\)/i;
|
||
$field =~ s/ character varying\((\d+)\)/ varchar\($1\)/i;
|
||
$field =~ s/ char varying\((\d+)\)/ varchar\($1\)/i;
|
||
$field =~ s/ integer/ number\(38\)/i;
|
||
$field =~ s/ int/ number\(38\)/i;
|
||
$field =~ s/ tinyint/ number\(38\)/i;
|
||
$field =~ s/ smallint/ number\(38\)/i;
|
||
$field =~ s/ mediumint/ number\(38\)/i;
|
||
$field =~ s/ tinynumber\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ smallnumber\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ mediumnumber\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ number\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ numeric\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ decimal\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ dec\((\d+)\)\((\d+)\)/ number\($1,$2\)/i;
|
||
$field =~ s/ float/ number/;
|
||
$field =~ s/ real/ number/;
|
||
$field =~ s/ double precision/ number/;
|
||
$field =~ s/ double/ number/;
|
||
$field =~ s/ blob/ long/;
|
||
$query.= $field . ',';
|
||
}
|
||
|
||
foreach $ind (@$index)
|
||
{
|
||
my @index;
|
||
if ( $ind =~ /\bKEY\b/i ){
|
||
push(@keys,"ALTER TABLE $table_name ADD $ind");
|
||
}else{
|
||
my @fields = split(' ',$index);
|
||
my $query="CREATE INDEX $fields[1] ON $table_name $fields[2]";
|
||
push(@index,$query);
|
||
}
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query,@keys,@index);
|
||
#print "query:$query\n";
|
||
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert an ascii file isn't supported by Oracle (?)\n";
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
$cmd =~ s/\'\'/\' \'/g;
|
||
return $cmd;
|
||
}
|
||
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 1;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# optimize the tables ....
|
||
#
|
||
sub vacuum
|
||
{
|
||
my ($self,$full_vacuum,$dbh_ref)=@_;
|
||
my ($loop_time,$end_time,$sth,$dbh);
|
||
|
||
if (defined($full_vacuum))
|
||
{
|
||
$$dbh_ref->disconnect; $$dbh_ref= $self->connect();
|
||
}
|
||
$dbh=$$dbh_ref;
|
||
$loop_time=new Benchmark;
|
||
# first analyze all tables
|
||
$sth = $dbh->prepare("select table_name from user_tables") || die "Got error: $DBI::errstr";
|
||
$sth->execute || die "Got error: $DBI::errstr when select user_tables";
|
||
while (my @r = $sth->fetchrow_array)
|
||
{
|
||
$dbh->do("analyze table $r[0] compute statistics") || die "Got error: $DBI::errstr when executing 'analyze table'\n";
|
||
}
|
||
# now analyze all indexes ...
|
||
$sth = $dbh->prepare("select index_name from user_indexes") || die "Got error: $DBI::errstr";
|
||
$sth->execute || die "Got error: $DBI::errstr when select user_indexes";
|
||
while (my @r1 = $sth->fetchrow_array)
|
||
{
|
||
$dbh->do("analyze index $r1[0] compute statistics") || die "Got error: $DBI::errstr when executing 'analyze index $r1[0]'\n";
|
||
}
|
||
$end_time=new Benchmark;
|
||
print "Time for book-keeping (1): " .
|
||
Benchmark::timestr(Benchmark::timediff($end_time, $loop_time),"all") . "\n\n";
|
||
$dbh->disconnect; $$dbh_ref= $self->connect();
|
||
}
|
||
|
||
|
||
#############################################################################
|
||
# Definitions for Informix
|
||
#############################################################################
|
||
|
||
package db_Informix;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "Informix";
|
||
$self->{'data_source'} = "DBI:Informix:$database";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "byte in table";
|
||
$self->{'text'} = "byte in table";
|
||
$self->{'double_quotes'} = 0; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{'host'} = $host;
|
||
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 1;
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'func_extra_%'} = 0; # Has % as alias for mod()
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'func_extra_in_num'}= 0; # Has function in
|
||
$limits{'func_odbc_floor'} = 0; # Has func_odbc_floor function
|
||
$limits{'func_odbc_mod'} = 1; # Have function mod.
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_by_position'} = 1; # Can use 'GROUP BY 1'
|
||
$limits{'group_by_alias'} = 0; # Can use 'select a as ab from x GROUP BY ab'
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'}= 1; # Can't use group functions in HAVING
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables (always 1 only for msql)
|
||
$limits{'left_outer_join'} = 0; # Supports left outer joins (ANSI)
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 1; # Has lock tables
|
||
$limits{'max_conditions'} = 1214; # (Actually not a limit)
|
||
$limits{'max_column_name'} = 18; # max table and column name
|
||
$limits{'max_columns'} = 994; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_index'} = 64; # Max number of keys
|
||
$limits{'max_index_parts'} = 15; # Max segments/key
|
||
$limits{'max_text_size'} = 65535; # Max size with default buffers. ??
|
||
$limits{'multi_drop'} = 0; # Drop table can take many tables
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'query_size'} = 32766; # Max size with default buffers.
|
||
$limits{'select_without_from'}= 0; # Can do 'select 1';
|
||
$limits{'subqueries'} = 1; # Doesn't support sub-queries.
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$ENV{'INFORMIXSERVER'} = $self->{'host'};
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare("SELECT owner FROM systables WHERE tabname = ' VERSION'")
|
||
or die $DBI::errstr;
|
||
$version='Informix unknown';
|
||
if ($sth->execute && (@row = $sth->fetchrow_array))
|
||
{
|
||
$version="Informix $row[0]";
|
||
}
|
||
$sth->finish;
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
|
||
#
|
||
# Create table
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$name,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
# $field =~ s/\btransport_description\b/transport_desc/;
|
||
# to overcome limit 18 chars
|
||
$field =~ s/tinyint/smallint/i;
|
||
$field =~ s/tinyint\(\d+\)/smallint/i;
|
||
$field =~ s/mediumint/integer/i;
|
||
$field =~ s/mediumint\(\d+\)/integer/i;
|
||
$field =~ s/smallint\(\d+\)/smallint/i;
|
||
$field =~ s/integer\(\d+\)/integer/i;
|
||
$field =~ s/int\(\d+\)/integer/i;
|
||
# $field =~ s/\b(?:small)?int(?:eger)?\((\d+)\)/decimal($1)/i;
|
||
# $field =~ s/float(\(\d*,\d*\)){0,1}/real/i;
|
||
$field =~ s/(float|double)(\(.*?\))?/float/i;
|
||
|
||
if ($field =~ / blob/i)
|
||
{
|
||
$name=$self->{'blob'};
|
||
$field =~ s/ blob/ $name/;
|
||
}
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
|
||
foreach $index (@$index)
|
||
{
|
||
# Primary key is unique index in Informix
|
||
$index =~ s/primary key/unique index primary/i;
|
||
if ($index =~ /^unique\s*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
push(@queries,"create unique index ${table_name}_$nr on $table_name ($1)");
|
||
}
|
||
else
|
||
{
|
||
if (!($index =~ /^(.*index)\s+(\w*)\s+(\(.*\))$/i))
|
||
{
|
||
die "Can't parse index information in '$index'\n";
|
||
}
|
||
### push(@queries,"create $1 ${table_name}_$2 on $table_name $3");
|
||
$nr++;
|
||
push(@queries,"create $1 ${table_name}_$nr on $table_name $3");
|
||
}
|
||
}
|
||
return @queries;
|
||
}
|
||
#
|
||
# Some test needed this
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
$cmd =~ s/\\\'//g;
|
||
return $cmd;
|
||
}
|
||
|
||
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
|
||
#############################################################################
|
||
# Configuration for Access
|
||
#############################################################################
|
||
|
||
package db_access;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "access";
|
||
$self->{'data_source'} = "DBI:ODBC:$database";
|
||
if (defined($host) && $host ne "")
|
||
{
|
||
$self->{'data_source'} .= ":$host";
|
||
}
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "blob";
|
||
$self->{'text'} = "blob"; # text ?
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
|
||
$limits{'max_conditions'} = 97; # We get 'Query is too complex'
|
||
$limits{'max_columns'} = 255; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 255; # Max size with default buffers.
|
||
$limits{'query_size'} = 65535; # Not a limit, big enough
|
||
$limits{'max_index'} = 32; # Max number of keys
|
||
$limits{'max_index_parts'} = 10; # Max segments/key
|
||
$limits{'max_column_name'} = 64; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 0; # Have count(distinct)
|
||
$limits{'select_without_from'}= 1; # Can do 'select 1';
|
||
$limits{'multi_drop'} = 0; # Drop table can take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Can use 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 2; #Have ALTER TABLE t add a int, b int;
|
||
$limits{'alter_table_dropcol'}= 1;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Have function mod.
|
||
$limits{'func_extra_%'} = 0; # Has % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 0; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 1; # Has function in
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my $version="Access 2000";
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version; #DBI/ODBC can't return the server version
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/mediumint/integer/i;
|
||
$field =~ s/tinyint/smallint/i;
|
||
$field =~ s/float\(\d+,\d+\)/float/i;
|
||
$field =~ s/integer\(\d+\)/integer/i;
|
||
$field =~ s/smallint\(\d+\)/smallint/i;
|
||
$field =~ s/int\(\d+\)/integer/i;
|
||
$field =~ s/blob/text/i;
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
foreach $index (@$index)
|
||
{
|
||
$ext="WITH DISALLOW NULL";
|
||
if (($index =~ s/primary key/unique index primary_key/i))
|
||
{
|
||
$ext="WITH PRIMARY;"
|
||
}
|
||
if ($index =~ /^unique.*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
$index="unique index ${table_name}_$nr ($1)";
|
||
}
|
||
$index =~ /^(.*)\s+(\(.*\))$/;
|
||
push(@queries,"create ${1} on $table_name $2");
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index ON $table";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 1 if (($DBI::errstr =~ /The database engine couldn\'t lock table/i) ||
|
||
($DBI::errstr =~ /niet vergrendelen. De tabel is momenteel in gebruik /i) ||
|
||
($DBI::errstr =~ /Den anv.* redan av en annan/i) ||
|
||
($DBI::errstr =~ /non-exclusive access/));
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 1;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for Microsoft SQL server
|
||
#############################################################################
|
||
|
||
package db_ms_sql;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "ms-sql";
|
||
$self->{'data_source'} = "DBI:ODBC:$database";
|
||
if (defined($host) && $host ne "")
|
||
{
|
||
$self->{'data_source'} .= ":$host";
|
||
}
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "text";
|
||
$self->{'text'} = "text";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
|
||
$limits{'max_conditions'} = 1030; # We get 'Query is too complex'
|
||
$limits{'max_columns'} = 250; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 9830; # Max size with default buffers.
|
||
$limits{'query_size'} = 9830; # Max size with default buffers.
|
||
$limits{'max_index'} = 64; # Max number of keys
|
||
$limits{'max_index_parts'} = 15; # Max segments/key
|
||
$limits{'max_column_name'} = 30; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 1; # Can do 'select 1';
|
||
$limits{'multi_drop'} = 1; # Drop table can take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can't use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Can use 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Have function mod.
|
||
$limits{'func_extra_%'} = 1; # Has % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 0; # Has function in
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my($sth,@row, $version);
|
||
$version='MS SQL server ?';
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare("SELECT \@\@VERSION") or die $DBI::errstr;
|
||
$sth->execute or die $DBI::errstr;
|
||
@row = $sth->fetchrow_array;
|
||
if ($row[0]) {
|
||
@server = split(/\n/,$row[0]);
|
||
chomp(@server);
|
||
$version= "$server[0]";
|
||
}
|
||
$sth->finish;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/mediumint/integer/i;
|
||
$field =~ s/float\(\d+,\d+\)/float/i;
|
||
$field =~ s/double\(\d+,\d+\)/float/i;
|
||
$field =~ s/double/float/i;
|
||
$field =~ s/integer\(\d+\)/integer/i;
|
||
$field =~ s/int\(\d+\)/integer/i;
|
||
$field =~ s/smallint\(\d+\)/smallint/i;
|
||
$field =~ s/smallinteger/smallint/i;
|
||
$field =~ s/tinyint\(\d+\)/tinyint/i;
|
||
$field =~ s/tinyinteger/tinyint/i;
|
||
$field =~ s/blob/text/i;
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
foreach $index (@$index)
|
||
{
|
||
$ext="WITH DISALLOW NULL";
|
||
if (($index =~ s/primary key/unique index primary_key/i))
|
||
{
|
||
$ext="WITH PRIMARY;"
|
||
}
|
||
if ($index =~ /^unique.*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
$index="unique index ${table_name}_$nr ($1)";
|
||
}
|
||
$index =~ /^(.*)\s+(\(.*\))$/;
|
||
push(@queries,"create ${1} on $table_name $2");
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $table.$index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for Sybase
|
||
#############################################################################
|
||
package db_sybase;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "sybase";
|
||
$self->{'data_source'} = "DBI:Sybase:database=$database";
|
||
if (defined($host) && $host ne "")
|
||
{
|
||
$self->{'data_source'} .= ";hostname=$host";
|
||
}
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "text";
|
||
$self->{'text'} = "text";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{"vacuum"} = 1;
|
||
|
||
$limits{'max_conditions'} = 1030; # We get 'Query is too complex'
|
||
$limits{'max_columns'} = 250; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 9830; # Max size with default buffers.
|
||
$limits{'query_size'} = 9830; # Max size with default buffers.
|
||
$limits{'max_index'} = 64; # Max number of keys
|
||
$limits{'max_index_parts'} = 15; # Max segments/key
|
||
$limits{'max_column_name'} = 30; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 1; # Can do 'select 1';
|
||
$limits{'multi_drop'} = 1; # Drop table can take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can't use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Can use 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Have function mod.
|
||
$limits{'func_extra_%'} = 1; # Has % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 0; # Has function in
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare('SELECT @@version') or die $DBI::errstr;
|
||
$version="Sybase (unknown)";
|
||
if ($sth->execute && (@row = $sth->fetchrow_array))
|
||
{
|
||
$version=$row[0];
|
||
}
|
||
$sth->finish;
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0 , AutoCommit => 1}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/mediumint/integer/i;
|
||
$field =~ s/float\(\d+,\d+\)/float/i;
|
||
$field =~ s/int\(\d+\)/int/i;
|
||
$field =~ s/double/float/i;
|
||
$field =~ s/integer\(\d+\)/integer/i;
|
||
$field =~ s/smallint\(\d+\)/smallint/i;
|
||
$field =~ s/tinyint\(\d+\)/tinyint/i;
|
||
$field =~ s/blob/text/i;
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
foreach $index (@$index)
|
||
{
|
||
# $ext="WITH DISALLOW NULL";
|
||
if (($index =~ s/primary key/unique index primary_key/i))
|
||
{
|
||
# $ext="WITH PRIMARY;"
|
||
}
|
||
if ($index =~ /^unique.*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
$index="unique index ${table_name}_$nr ($1)";
|
||
}
|
||
$index =~ /^(.*)\s+(\(.*\))$/;
|
||
push(@queries,"create ${1} on $table_name $2");
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $table.$index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#
|
||
# optimize the tables ....
|
||
# WARNING (from walrus)! This sub will work only from DBD:sybase
|
||
# driver. Because if we use ODBC we don't know actual database name
|
||
# (but DSN name only)
|
||
sub vacuum
|
||
{
|
||
my ($self,$full_vacuum,$dbh_ref)=@_;
|
||
my ($loop_time,$end_time,$dbh);
|
||
|
||
if (defined($full_vacuum))
|
||
{
|
||
$$dbh_ref->disconnect; $$dbh_ref= $self->connect();
|
||
}
|
||
$dbh=$$dbh_ref;
|
||
$loop_time=new Benchmark;
|
||
my (@tables,$sth,$current_table,$current_base);
|
||
$dbh->do("dump tran $database with truncate_only");
|
||
$sth=$dbh->prepare("sp_tables" ) or die "prepere";
|
||
$sth->execute() or die "execute";
|
||
while (@row = $sth->fetchrow_array()) {
|
||
$current_table = $row[2];
|
||
$current_base = $row[0];
|
||
next if ($current_table =~ /^sys/);
|
||
push(@tables,$current_table) if ($database == $current_base);
|
||
}
|
||
|
||
$sth->finish();
|
||
|
||
foreach $table (@tables) {
|
||
# print "$table: \n";
|
||
$dbh->do("update statistics $table") or print "Oops!";
|
||
}
|
||
|
||
# $dbh->do("analyze table ?? compute statistics") || die "Got error: $DBI::errstr when executing 'vacuum'\n";
|
||
$end_time=new Benchmark;
|
||
print "Time for book-keeping (1): " .
|
||
Benchmark::timestr(Benchmark::timediff($end_time, $loop_time),"all") . "\n\n";
|
||
$dbh->disconnect; $$dbh_ref= $self->connect();
|
||
}
|
||
|
||
|
||
|
||
|
||
#############################################################################
|
||
# Definitions for Adabas
|
||
#############################################################################
|
||
|
||
package db_Adabas;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "Adabas";
|
||
$self->{'data_source'} = "DBI:Adabas:$database";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "long";
|
||
$self->{'text'} = "long";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
|
||
$limits{'max_conditions'} = 50; # (Actually not a limit)
|
||
$limits{'max_columns'} = 254; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 2000; # Limit for blob test-connect
|
||
$limits{'query_size'} = 65525; # Max size with default buffers.
|
||
$limits{'max_index'} = 16; # Max number of keys
|
||
$limits{'max_index_parts'} = 16; # Max segments/key
|
||
$limits{'max_column_name'} = 32; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0;
|
||
$limits{'multi_drop'} = 0;
|
||
$limits{'subqueries'} = 1;
|
||
$limits{'left_outer_join'} = 0; # This may be fixed in the query module
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can't use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 1;
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 2; #Have ALTER TABLE t add a int, b int;
|
||
$limits{'alter_table_dropcol'}= 1;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Oracle has problem with mod()
|
||
$limits{'func_extra_%'} = 0; # Has % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 1; # Has function in
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare("SELECT KERNEL FROM VERSIONS") or die $DBI::errstr;
|
||
$version="Adabas (unknown)";
|
||
if ($sth->execute && (@row = $sth->fetchrow_array)
|
||
&& $row[0] =~ /([\d\.]+)/)
|
||
{
|
||
$version="Adabas $1";
|
||
}
|
||
$sth->finish;
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
# If one uses $main::opt_fast then one is allowed to use
|
||
# non standard types to get better speed.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$ind,@keys);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/CHARACTER\s+VARYING/VARCHAR/i;
|
||
$field =~ s/TINYINT/SMALLINT/i;
|
||
$field =~ s/MEDIUMINT/INT/i;
|
||
$field =~ s/SMALLINT\s*\(\d+\)/SMALLINT/i;
|
||
$field =~ s/INT\s*\(\d+\)/INT/i;
|
||
$field =~ s/BLOB/LONG/i;
|
||
$field =~ s/INTEGER\s*\(\d+\)/INTEGER/i;
|
||
$field =~ s/FLOAT\s*\((\d+),\d+\)/FLOAT\($1\)/i;
|
||
$field =~ s/DOUBLE/FLOAT\(38\)/i;
|
||
$field =~ s/DOUBLE\s+PRECISION/FLOAT\(38\)/i;
|
||
$query.= $field . ',';
|
||
}
|
||
|
||
foreach $ind (@$index)
|
||
{
|
||
my @index;
|
||
if ( $ind =~ /\bKEY\b/i ){
|
||
push(@keys,"ALTER TABLE $table_name ADD $ind");
|
||
}else{
|
||
my @fields = split(' ',$index);
|
||
my $query="CREATE INDEX $fields[1] ON $table_name $fields[2]";
|
||
push(@index,$query);
|
||
}
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query,@keys,@index);
|
||
#print "query:$query\n";
|
||
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert an ascii file isn't supported by Oracle (?)\n";
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for IBM DB2
|
||
#############################################################################
|
||
|
||
package db_db2;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "DB2";
|
||
$self->{'data_source'} = "DBI:ODBC:$database";
|
||
if (defined($host) && $host ne "")
|
||
{
|
||
$self->{'data_source'} .= ":$host";
|
||
}
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "varchar(255)";
|
||
$self->{'text'} = "varchar(255)";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
|
||
$limits{'max_conditions'} = 418; # We get 'Query is too complex'
|
||
$limits{'max_columns'} = 500; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 254; # Max size with default buffers.
|
||
$limits{'query_size'} = 254; # Max size with default buffers.
|
||
$limits{'max_index'} = 48; # Max number of keys
|
||
$limits{'max_index_parts'} = 15; # Max segments/key
|
||
$limits{'max_column_name'} = 18; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'}= 1;
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0; # Can do 'select 1';
|
||
$limits{'multi_drop'} = 0; # Drop table can take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can't use group functions in HAVING
|
||
$limits{'like_with_column'} = 0; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Can use 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1;
|
||
$limits{'alter_add_multi_col'}= 0;
|
||
$limits{'alter_table_dropcol'}= 0;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 1; # Have function mod.
|
||
$limits{'func_extra_%'} = 0; # Has % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 0; # Has function in
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
return "IBM DB2 5"; #DBI/ODBC can't return the server version
|
||
}
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user, $main::opt_password) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index) = @_;
|
||
my($query,@queries,$nr);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/mediumint/integer/i;
|
||
$field =~ s/float\(\d+,\d+\)/float/i;
|
||
$field =~ s/integer\(\d+\)/integer/i;
|
||
$field =~ s/int\(\d+\)/integer/i;
|
||
$field =~ s/tinyint\(\d+\)/smallint/i;
|
||
$field =~ s/tinyint/smallint/i;
|
||
$field =~ s/smallint\(\d+\)/smallint/i;
|
||
$field =~ s/smallinteger/smallint/i;
|
||
$field =~ s/blob/varchar(256)/i;
|
||
$query.= $field . ',';
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
push(@queries,$query);
|
||
$nr=0;
|
||
foreach $index (@$index)
|
||
{
|
||
$ext="WITH DISALLOW NULL";
|
||
if (($index =~ s/primary key/unique index primary_key/i))
|
||
{
|
||
$ext="WITH PRIMARY;"
|
||
}
|
||
if ($index =~ /^unique.*\(([^\(]*)\)$/i)
|
||
{
|
||
$nr++;
|
||
$index="unique index ${table_name}_$nr ($1)";
|
||
}
|
||
$index =~ /^(.*)\s+(\(.*\))$/;
|
||
push(@queries,"create ${1} on $table_name $2");
|
||
}
|
||
return @queries;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index
|
||
{
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $table.$index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 1;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for MIMER
|
||
#############################################################################
|
||
|
||
package db_Mimer;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "mimer";
|
||
$self->{'data_source'} = "DBI:mimer:$database:$host";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "binary varying(15000)";
|
||
$self->{'text'} = "character varying(15000)";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{'char_null'} = "cast(NULL as char(1))";
|
||
$self->{'numeric_null'} = "cast(NULL as int)";
|
||
|
||
$limits{'max_conditions'} = 9999; # (Actually not a limit)
|
||
$limits{'max_columns'} = 252; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 15000; # Max size with default buffers.
|
||
$limits{'query_size'} = 1000000; # Max size with default buffers.
|
||
$limits{'max_index'} = 32; # Max number of keys
|
||
$limits{'max_index_parts'} = 16; # Max segments/key
|
||
$limits{'max_column_name'} = 128; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 1; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0; # Cannot do 'select 1';
|
||
$limits{'multi_drop'} = 0; # Drop table cannot take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Cannot use 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1; # Have ALTER TABLE
|
||
$limits{'alter_add_multi_col'}= 0; # Have ALTER TABLE t add a int,add b int;
|
||
$limits{'alter_table_dropcol'}= 1; # Have ALTER TABLE DROP column
|
||
$limits{'insert_multi_value'} = 0; # Does not have INSERT ... values (1,2),(3,4)
|
||
$limits{'multi_distinct'} = 0; # Does not allow select count(distinct a),count(distinct b)..
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Does not have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 1; # Have function mod.
|
||
$limits{'func_extra_%'} = 0; # Does not have % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Does not have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 1; # Has function in
|
||
$limits{'limit'} = 0; # Does not support the limit attribute
|
||
$limits{'unique_index'} = 1; # Unique index works or not
|
||
$limits{'insert_select'} = 1;
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 0;
|
||
$limits{'working_all_fields'} = 1;
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$dbh=$self->connect();
|
||
#
|
||
# Pick up SQLGetInfo option SQL_DBMS_VER (18)
|
||
#
|
||
$version = $dbh->func(18, GetInfo);
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
#
|
||
# Connection with optional disabling of logging
|
||
#
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
|
||
$dbh->do("SET OPTION LOG_OFF=1,UPDATE_LOG=0");
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
# If one uses $main::opt_fast then one is allowed to use
|
||
# non standard types to get better speed.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index,$options) = @_;
|
||
my($query,@queries,@indexes);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
# $field =~ s/ decimal/ double(10,2)/i;
|
||
# $field =~ s/ big_decimal/ double(10,2)/i;
|
||
$field =~ s/ double/ double precision/i;
|
||
$field =~ s/ tinyint\(.*\)/ smallint/i;
|
||
$field =~ s/ smallint\(.*\)/ smallint/i;
|
||
$field =~ s/ mediumint/ integer/i;
|
||
$field =~ s/ float\(.*\)/ float/i;
|
||
# $field =~ s/ date/ int/i; # Because of tcp ?
|
||
$query.= $field . ',';
|
||
}
|
||
foreach $index (@$index)
|
||
{
|
||
if ( $index =~ /\bINDEX\b/i )
|
||
{
|
||
my @fields = split(' ',$index);
|
||
my $query="CREATE INDEX $fields[1] ON $table_name $fields[2]";
|
||
push(@indexes,$query);
|
||
|
||
} else {
|
||
$query.= $index . ',';
|
||
}
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
$query.=" $options" if (defined($options));
|
||
push(@queries,$query,@indexes);
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert of an ascii file isn't supported by Mimer\n";
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index {
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 1 if ($DBI::errstr =~ /Table locked by another cursor/);
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for InterBase
|
||
#############################################################################
|
||
|
||
package db_interbase;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "interbase";
|
||
$self->{'data_source'} = "DBI:InterBase:database=$database;ib_dialect=3";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "blob";
|
||
$self->{'text'} = "";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{'char_null'} = "";
|
||
$self->{'numeric_null'} = "";
|
||
|
||
$limits{'max_conditions'} = 9999; # (Actually not a limit)
|
||
$limits{'max_columns'} = 252; # Max number of columns in table
|
||
$limits{'max_tables'} = 65000; # Should be big enough
|
||
$limits{'max_text_size'} = 15000; # Max size with default buffers.
|
||
$limits{'query_size'} = 1000000; # Max size with default buffers.
|
||
$limits{'max_index'} = 65000; # Max number of keys
|
||
$limits{'max_index_parts'} = 8; # Max segments/key
|
||
$limits{'max_column_name'} = 128; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 0; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0; # Cannot do 'select 1';
|
||
$limits{'multi_drop'} = 0; # Drop table cannot take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 1; # Can use group functions in HAVING
|
||
$limits{'like_with_column'} = 0; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Cannot use 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1; # Have ALTER TABLE
|
||
$limits{'alter_add_multi_col'}= 1; # Have ALTER TABLE t add a int,add b int;
|
||
$limits{'alter_table_dropcol'}= 1; # Have ALTER TABLE DROP column
|
||
$limits{'insert_multi_value'} = 0; # Does not have INSERT ... values (1,2),(3,4)
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Does not have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Have function mod.
|
||
$limits{'func_extra_%'} = 0; # Does not have % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 0; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Does not have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 0; # Supports -id
|
||
$limits{'func_extra_in_num'} = 0; # Has function in
|
||
$limits{'limit'} = 0; # Does not support the limit attribute
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works
|
||
$limits{'order_by_unused'} = 1;
|
||
$limits{'working_all_fields'} = 1;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$version);
|
||
|
||
$version='Interbase ?';
|
||
|
||
$dbh=$self->connect();
|
||
eval { $version = $dbh->func('version','ib_database_info')->{'version'}; };
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
#
|
||
# Connection with optional disabling of logging
|
||
#
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0, AutoCommit => 1}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
# If one uses $main::opt_fast then one is allowed to use
|
||
# non standard types to get better speed.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index,$options) = @_;
|
||
my($query,@queries,@keys,@indexes);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
# $field =~ s/ big_decimal/ decimal/i;
|
||
$field =~ s/ double/ double precision/i;
|
||
$field =~ s/ tinyint/ smallint/i;
|
||
$field =~ s/ mediumint/ integer/i;
|
||
$field =~ s/\bint\b/integer/i;
|
||
$field =~ s/ float\(\d,\d\)/ float/i;
|
||
$field =~ s/ smallint\(\d\)/ smallint/i;
|
||
$field =~ s/ integer\(\d\)/ integer/i;
|
||
$query.= $field . ',';
|
||
}
|
||
foreach $ind (@$index)
|
||
{
|
||
if ( $ind =~ /(\bKEY\b)|(\bUNIQUE\b)/i ){
|
||
push(@keys,"ALTER TABLE $table_name ADD $ind");
|
||
}else{
|
||
my @fields = split(' ',$ind);
|
||
my $query="CREATE INDEX $fields[1] ON $table_name $fields[2]";
|
||
push(@indexes,$query);
|
||
}
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
$query.=" $options" if (defined($options));
|
||
push(@queries,$query,@keys,@indexes);
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert of an ascii file isn't supported by InterBase\n";
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index {
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 1 if ($DBI::errstr =~ /Table locked by another cursor/);
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 1;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 1;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for FrontBase
|
||
#############################################################################
|
||
|
||
package db_FrontBase;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "FrontBase";
|
||
$self->{'data_source'} = "DBI:FB:dbname=$database;host=$host";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "varchar(8000000)";
|
||
$self->{'text'} = "varchar(8000000)";
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = ' restrict';
|
||
$self->{'transactions'} = 1; # Transactions enabled
|
||
$self->{'error_on_execute_means_zero_rows'}=1;
|
||
|
||
$limits{'max_conditions'} = 5427; # (Actually not a limit)
|
||
# The following should be 8192, but is smaller because Frontbase crashes..
|
||
$limits{'max_columns'} = 150; # Max number of columns in table
|
||
$limits{'max_tables'} = 5000; # 10000 crashed FrontBase
|
||
$limits{'max_text_size'} = 65000; # Max size with default buffers.
|
||
$limits{'query_size'} = 8000000; # Max size with default buffers.
|
||
$limits{'max_index'} = 38; # Max number of keys
|
||
$limits{'max_index_parts'} = 20; # Max segments/key
|
||
$limits{'max_column_name'} = 128; # max table and column name
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables
|
||
$limits{'load_data_infile'} = 1; # Has load data infile
|
||
$limits{'lock_tables'} = 0; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-)
|
||
$limits{'group_functions'} = 1; # Have group functions
|
||
$limits{'group_distinct_functions'}= 0; # Have count(distinct)
|
||
$limits{'select_without_from'}= 0;
|
||
$limits{'multi_drop'} = 0; # Drop table cannot take many tables
|
||
$limits{'subqueries'} = 1; # Supports sub-queries.
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING
|
||
$limits{'having_with_group'} = 0; # Can use group functions in HAVING
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1'
|
||
$limits{'group_by_position'} = 0; # Use of 'GROUP BY 1'
|
||
$limits{'alter_table'} = 1; # Have ALTER TABLE
|
||
$limits{'alter_add_multi_col'}= 0; # Have ALTER TABLE t add a int,add b int;
|
||
$limits{'alter_table_dropcol'}= 0; # Have ALTER TABLE DROP column
|
||
$limits{'insert_multi_value'} = 1;
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Does not have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Have function mod.
|
||
$limits{'func_extra_%'} = 0; # Does not have % as alias for mod()
|
||
$limits{'func_odbc_floor'} = 0; # Has func_odbc_floor function
|
||
$limits{'func_extra_if'} = 0; # Does not have function if.
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement.
|
||
$limits{'NEG'} = 1; # Supports -id
|
||
$limits{'func_extra_in_num'} = 0; # Has function in
|
||
$limits{'limit'} = 0; # Does not support the limit attribute
|
||
$limits{'insert_select'} = 0;
|
||
$limits{'order_by_unused'} = 0;
|
||
|
||
# We don't get an error for duplicate row in 'test-insert'
|
||
$limits{'unique_index'} = 0; # Unique index works or not
|
||
# We can't use a blob as a normal string (we got a wierd error)
|
||
$limits{'working_blobs'} = 0;
|
||
# 'select min(region),max(region) from bench1' kills the server after a while
|
||
$limits{'group_func_sql_min_str'} = 0;
|
||
# If you do select f1,f2,f3...f200 from table, Frontbase dies.
|
||
$limits{'working_all_fields'} = 0;
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
# $dbh=$self->connect();
|
||
#
|
||
# Pick up SQLGetInfo option SQL_DBMS_VER (18)
|
||
#
|
||
#$version = $dbh->func(18, GetInfo);
|
||
$version="FrontBase 3.3";
|
||
# $dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
#
|
||
# Connection with optional disabling of logging
|
||
#
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'},
|
||
$main::opt_user,
|
||
$main::opt_password,
|
||
{ PrintError => 0 ,
|
||
'fb_host'=>$main::opt_host
|
||
}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
$db->{AutoCommit}=1;
|
||
# $dbh->do("SET OPTION LOG_OFF=1,UPDATE_LOG=0");
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
# If one uses $main::opt_fast then one is allowed to use
|
||
# non standard types to get better speed.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index,$options) = @_;
|
||
my($query,@queries,@indexes,@keys);
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/ blob/ varchar(32000)/i;
|
||
$field =~ s/ big_decimal/ float/i;
|
||
$field =~ s/ double/ float/i;
|
||
$field =~ s/ tinyint/ smallint/i;
|
||
$field =~ s/ mediumint/ int/i;
|
||
$field =~ s/ integer/ int/i;
|
||
$field =~ s/ float\(\d,\d\)/ float/i;
|
||
$field =~ s/ smallint\(\d\)/ smallint/i;
|
||
$field =~ s/ int\(\d\)/ int/i;
|
||
$query.= $field . ',';
|
||
}
|
||
foreach $ind (@$index)
|
||
{
|
||
# my @index;
|
||
if ( $ind =~ /(\bKEY\b)|(\bUNIQUE\b)/i ){
|
||
push(@keys,"ALTER TABLE $table_name ADD $ind");
|
||
}else{
|
||
my @fields = split(' ',$ind);
|
||
my $query="CREATE INDEX $fields[1] ON $table_name $fields[2]";
|
||
push(@indexes,$query);
|
||
}
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
$query.=" $options" if (defined($options));
|
||
push(@queries,$query,@keys,@indexes);
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert of an ascii file isn't supported by InterBase\n";
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index {
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0 if ($DBI::errstr =~ /No raw data handle/);
|
||
return 1;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 1;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
#############################################################################
|
||
# Configuration for SAPDB
|
||
#############################################################################
|
||
|
||
package db_sapdb;
|
||
|
||
sub new
|
||
{
|
||
my ($type,$host,$database)= @_;
|
||
my $self= {};
|
||
my %limits;
|
||
bless $self;
|
||
|
||
$self->{'cmp_name'} = "sapdb";
|
||
$self->{'data_source'} = "DBI:SAP_DB:$database";
|
||
$self->{'limits'} = \%limits;
|
||
$self->{'blob'} = "LONG"; # *
|
||
$self->{'text'} = "LONG"; # *
|
||
$self->{'double_quotes'} = 1; # Can handle: 'Walker''s'
|
||
$self->{'drop_attr'} = "";
|
||
$self->{'transactions'} = 1; # Transactions enabled *
|
||
$self->{'char_null'} = "";
|
||
$self->{'numeric_null'} = "";
|
||
|
||
$limits{'max_conditions'} = 9999; # (Actually not a limit) *
|
||
$limits{'max_columns'} = 1023; # Max number of columns in table *
|
||
$limits{'max_tables'} = 65000; # Should be big enough * unlimited actually
|
||
$limits{'max_text_size'} = 15000; # Max size with default buffers.
|
||
$limits{'query_size'} = 64*1024; # Max size with default buffers. *64 kb by default. May be set by system variable
|
||
$limits{'max_index'} = 510; # Max number of keys *
|
||
$limits{'max_index_parts'} = 16; # Max segments/key *
|
||
$limits{'max_column_name'} = 32; # max table and column name *
|
||
|
||
$limits{'join_optimizer'} = 1; # Can optimize FROM tables *
|
||
$limits{'load_data_infile'} = 0; # Has load data infile *
|
||
$limits{'lock_tables'} = 1; # Has lock tables
|
||
$limits{'functions'} = 1; # Has simple functions (+/-) *
|
||
$limits{'group_functions'} = 1; # Have group functions *
|
||
$limits{'group_func_sql_min_str'} = 1; # Can execute MIN() and MAX() on strings *
|
||
$limits{'group_distinct_functions'}= 1; # Have count(distinct) *
|
||
$limits{'select_without_from'}= 0; # Cannot do 'select 1'; *
|
||
$limits{'multi_drop'} = 0; # Drop table cannot take many tables *
|
||
$limits{'subqueries'} = 1; # Supports sub-queries. *
|
||
$limits{'left_outer_join'} = 1; # Supports left outer joins *
|
||
$limits{'table_wildcard'} = 1; # Has SELECT table_name.*
|
||
$limits{'having_with_alias'} = 0; # Can use aliases in HAVING *
|
||
$limits{'having_with_group'} = 1; # Can use group functions in HAVING *
|
||
$limits{'like_with_column'} = 1; # Can use column1 LIKE column2 *
|
||
$limits{'order_by_position'} = 1; # Can use 'ORDER BY 1' *
|
||
$limits{'group_by_position'} = 0; # Cannot use 'GROUP BY 1' *
|
||
$limits{'alter_table'} = 1; # Have ALTER TABLE *
|
||
$limits{'alter_add_multi_col'}= 1; # Have ALTER TABLE t add a int,add b int; *
|
||
$limits{'alter_table_dropcol'}= 1; # Have ALTER TABLE DROP column *
|
||
$limits{'insert_multi_value'} = 0; # INSERT ... values (1,2),(3,4) *
|
||
|
||
$limits{'group_func_extra_std'} = 0; # Does not have group function std().
|
||
|
||
$limits{'func_odbc_mod'} = 0; # Have function mod. *
|
||
$limits{'func_extra_%'} = 0; # Does not have % as alias for mod() *
|
||
$limits{'func_odbc_floor'} = 1; # Has func_odbc_floor function *
|
||
$limits{'func_extra_if'} = 0; # Does not have function if. *
|
||
$limits{'column_alias'} = 1; # Alias for fields in select statement. *
|
||
$limits{'NEG'} = 1; # Supports -id *
|
||
$limits{'func_extra_in_num'} = 0; # Has function in *
|
||
$limits{'limit'} = 0; # Does not support the limit attribute *
|
||
$limits{'working_blobs'} = 1; # If big varchar/blobs works *
|
||
$limits{'order_by_unused'} = 1; #
|
||
$limits{'working_all_fields'} = 1; #
|
||
$limits{'multi_distinct'} = 1; # allows select count(distinct a),count(distinct b)..
|
||
|
||
|
||
return $self;
|
||
}
|
||
|
||
#
|
||
# Get the version number of the database
|
||
#
|
||
|
||
sub version
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh,$sth,$version,@row);
|
||
|
||
$dbh=$self->connect();
|
||
$sth = $dbh->prepare("SELECT KERNEL FROM VERSIONS") or die $DBI::errstr;
|
||
$version="SAP DB (unknown)";
|
||
if ($sth->execute && (@row = $sth->fetchrow_array)
|
||
&& $row[0] =~ /([\d\.]+)/)
|
||
{
|
||
$version=$row[0];
|
||
$version =~ s/KERNEL/SAP DB/i;
|
||
}
|
||
$sth->finish;
|
||
$dbh->disconnect;
|
||
$version .= "/ODBC" if ($self->{'data_source'} =~ /:ODBC:/);
|
||
return $version;
|
||
}
|
||
|
||
#
|
||
# Connection with optional disabling of logging
|
||
#
|
||
|
||
sub connect
|
||
{
|
||
my ($self)=@_;
|
||
my ($dbh);
|
||
$dbh=DBI->connect($self->{'data_source'}, $main::opt_user,
|
||
$main::opt_password,{ PrintError => 0, AutoCommit => 1}) ||
|
||
die "Got error: '$DBI::errstr' when connecting to " . $self->{'data_source'} ." with user: '$main::opt_user' password: '$main::opt_password'\n";
|
||
|
||
return $dbh;
|
||
}
|
||
|
||
#
|
||
# Returns a list of statements to create a table
|
||
# The field types are in ANSI SQL format.
|
||
#
|
||
|
||
sub create
|
||
{
|
||
my($self,$table_name,$fields,$index,$options) = @_;
|
||
my($query,@queries,$nr);
|
||
my @index;
|
||
my @keys;
|
||
|
||
$query="create table $table_name (";
|
||
foreach $field (@$fields)
|
||
{
|
||
$field =~ s/\bmediumint\b/int/i;
|
||
$field =~ s/\btinyint\b/int/i;
|
||
$field =~ s/ int\(\d\)/ int/i;
|
||
$field =~ s/BLOB/LONG/i;
|
||
$field =~ s/INTEGER\s*\(\d+\)/INTEGER/i;
|
||
$field =~ s/SMALLINT\s*\(\d+\)/SMALLINT/i;
|
||
$field =~ s/FLOAT\s*\((\d+),\d+\)/FLOAT\($1\)/i;
|
||
$field =~ s/DOUBLE/FLOAT\(38\)/i;
|
||
$field =~ s/DOUBLE\s+PRECISION/FLOAT\(38\)/i;
|
||
$query.= $field . ',';
|
||
}
|
||
$nr=0;
|
||
foreach $ind (@$index)
|
||
{
|
||
if ( $ind =~ /\bKEY\b/i ){
|
||
push(@keys,"ALTER TABLE $table_name ADD $ind");
|
||
} elsif ($ind =~ /^unique.*\(([^\(]*)\)$/i) {
|
||
$nr++;
|
||
my $query="create unique index ${table_name}_$nr on $table_name ($1)";
|
||
push(@index,$query);
|
||
}else{
|
||
my @fields = split(' ',$ind);
|
||
my $query="CREATE INDEX $fields[1] ON $table_name $fields[2]";
|
||
push(@index,$query);
|
||
}
|
||
}
|
||
substr($query,-1)=")"; # Remove last ',';
|
||
$query.=" $options" if (defined($options));
|
||
push(@queries,$query);
|
||
push(@queries,@keys);
|
||
push(@queries,@index);
|
||
return @queries;
|
||
}
|
||
|
||
sub insert_file {
|
||
my($self,$dbname, $file) = @_;
|
||
print "insert of an ascii file isn't supported by SAPDB\n";
|
||
return 0;
|
||
}
|
||
|
||
#
|
||
# Do any conversions to the ANSI SQL query so that the database can handle it
|
||
#
|
||
|
||
sub query {
|
||
my($self,$sql) = @_;
|
||
return $sql;
|
||
}
|
||
|
||
sub drop_index {
|
||
my ($self,$table,$index) = @_;
|
||
return "DROP INDEX $index";
|
||
}
|
||
|
||
#
|
||
# Abort if the server has crashed
|
||
# return: 0 if ok
|
||
# 1 question should be retried
|
||
#
|
||
|
||
sub abort_if_fatal_error
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub small_rollback_segment
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub reconnect_on_errors
|
||
{
|
||
return 0;
|
||
}
|
||
|
||
sub fix_for_insert
|
||
{
|
||
my ($self,$cmd) = @_;
|
||
return $cmd;
|
||
}
|
||
|
||
1;
|