Linux ip-172-26-2-223 5.4.0-1018-aws #18-Ubuntu SMP Wed Jun 24 01:15:00 UTC 2020 x86_64
Apache
: 172.26.2.223 | : 3.16.70.193
Cant Read [ /etc/named.conf ]
8.1.13
www
www.github.com/MadExploits
Terminal
AUTO ROOT
Adminer
Backdoor Destroyer
Linux Exploit
Lock Shell
Lock File
Create User
CREATE RDP
PHP Mailer
BACKCONNECT
UNLOCK SHELL
HASH IDENTIFIER
CPANEL RESET
CREATE WP USER
BLACK DEFEND!
README
+ Create Folder
+ Create File
/
www /
server /
mysql /
mysql-test /
suite /
ndb_big /
[ HOME SHELL ]
Name
Size
Permission
Action
bug13637411-master.opt
22
B
-rw-r--r--
bug13637411.cnf
490
B
-rw-r--r--
bug13637411.test
3.38
KB
-rw-r--r--
bug14000373.cnf
79
B
-rw-r--r--
bug14000373.result
939
B
-rw-r--r--
bug14000373.test
1.1
KB
-rw-r--r--
bug37983-master.opt
22
B
-rw-r--r--
bug37983.test
1.46
KB
-rw-r--r--
disabled.def
44
B
-rw-r--r--
end_timer.inc
138
B
-rw-r--r--
my.cnf
1.64
KB
-rw-r--r--
ndb_big_addnode.cnf
1.45
KB
-rw-r--r--
ndb_big_addnode.result
5.66
KB
-rw-r--r--
ndb_big_addnode.test
8.29
KB
-rw-r--r--
ndb_multi_tc_takeover.cnf
1.16
KB
-rw-r--r--
ndb_multi_tc_takeover.test
411
B
-rw-r--r--
ndb_verify_redo_log_queue.test
405
B
-rw-r--r--
rqg_spj-master.opt
22
B
-rw-r--r--
rqg_spj.test
1.7
KB
-rw-r--r--
run_query_with_retry.inc
435
B
-rw-r--r--
smoke.test
10.16
KB
-rw-r--r--
start_timer.inc
83
B
-rw-r--r--
suite.inc
129
B
-rw-r--r--
Delete
Unzip
Zip
${this.title}
Close
Code Editor : bug14000373.test
# The include statement below is a temp one for tests that are yet to #be ported to run with InnoDB, #but needs to be kept for tests that would need MyISAM in future. --source include/force_myisam_default.inc -- source include/have_ndb.inc # Test is using error insert, check that binaries support it -- source suite/ndb/t/have_ndb_error_insert.inc # Use small LoadFactors to force sparse hash table --exec $NDB_MGM --no-defaults --ndb-connectstring="$NDB_CONNECTSTRING" -e "all error 3003" >> $NDB_TOOLS_OUTPUT set max_heap_table_size = 286720000; create table t1 (a int key) engine=memory; load data local infile 'suite/ndb/data/table_data10000.dat' into table t1 columns terminated by ' ' (a, @col2); let $i = 9; let $b = 10000; while ($i) { --eval insert into t1 select a + $b from t1; let $b = $b * 2; dec $i; } select count(*) from t1; alter table t1 engine=ndbcluster comment='NDB_TABLE=NOLOGGING' partition by key() partitions 1; --exec $NDB_MGM --no-defaults --ndb-connectstring="$NDB_CONNECTSTRING" -e "all report memory" >> $NDB_TOOLS_OUTPUT alter table t1 engine=memory; select count(*) from t1; drop table t1;
Close