mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-15 12:14:18 +00:00
be831d09f7
* [WIP] * Update skip-list * Update ci_config.json * Do not sync inserts for test * Fix more tests * Fix another test * Enable one more test * More fixed tests * More test fixes * Do not absolutize server path for now * More test fixes * Unset CLICKHOUSE_LOG_COMMENT where necessary * Remove debugging set -e * Fix more tests * Fix test reference * Fix style check
162 lines
6.3 KiB
SQL
162 lines
6.3 KiB
SQL
-- set insert_distributed_sync = 1; -- see https://github.com/ClickHouse/ClickHouse/issues/18971
|
|
|
|
DROP TABLE IF EXISTS local_01099_a;
|
|
DROP TABLE IF EXISTS local_01099_b;
|
|
DROP TABLE IF EXISTS distributed_01099_a;
|
|
DROP TABLE IF EXISTS distributed_01099_b;
|
|
|
|
SET parallel_distributed_insert_select=1;
|
|
SELECT 'parallel_distributed_insert_select=1';
|
|
|
|
--
|
|
-- test_shard_localhost
|
|
--
|
|
|
|
SELECT 'test_shard_localhost';
|
|
|
|
CREATE TABLE local_01099_a (number UInt64) ENGINE = Log;
|
|
CREATE TABLE local_01099_b (number UInt64) ENGINE = Log;
|
|
CREATE TABLE distributed_01099_a AS local_01099_a ENGINE = Distributed('test_shard_localhost', currentDatabase(), local_01099_a, rand());
|
|
CREATE TABLE distributed_01099_b AS local_01099_b ENGINE = Distributed('test_shard_localhost', currentDatabase(), local_01099_b, rand());
|
|
|
|
INSERT INTO local_01099_a SELECT number from system.numbers limit 3;
|
|
INSERT INTO distributed_01099_b SELECT * from distributed_01099_a;
|
|
|
|
SELECT * FROM distributed_01099_b;
|
|
|
|
DROP TABLE local_01099_a;
|
|
DROP TABLE local_01099_b;
|
|
DROP TABLE distributed_01099_a;
|
|
DROP TABLE distributed_01099_b;
|
|
|
|
--
|
|
-- test_cluster_two_shards_localhost
|
|
--
|
|
|
|
SELECT 'test_cluster_two_shards_localhost';
|
|
|
|
CREATE TABLE local_01099_a (number UInt64) ENGINE = Log;
|
|
CREATE TABLE local_01099_b (number UInt64) ENGINE = Log;
|
|
CREATE TABLE distributed_01099_a AS local_01099_a ENGINE = Distributed('test_cluster_two_shards_localhost', currentDatabase(), local_01099_a, rand());
|
|
CREATE TABLE distributed_01099_b AS local_01099_b ENGINE = Distributed('test_cluster_two_shards_localhost', currentDatabase(), local_01099_b, rand());
|
|
|
|
INSERT INTO local_01099_a SELECT number from system.numbers limit 3;
|
|
INSERT INTO distributed_01099_b SELECT * from distributed_01099_a;
|
|
|
|
SELECT number, count(number) FROM local_01099_b group by number order by number;
|
|
|
|
DROP TABLE local_01099_a;
|
|
DROP TABLE local_01099_b;
|
|
DROP TABLE distributed_01099_a;
|
|
DROP TABLE distributed_01099_b;
|
|
|
|
--
|
|
-- test_cluster_two_shards
|
|
--
|
|
|
|
SELECT 'test_cluster_two_shards';
|
|
|
|
CREATE TABLE local_01099_a (number UInt64) ENGINE = Log;
|
|
CREATE TABLE local_01099_b (number UInt64) ENGINE = Log;
|
|
CREATE TABLE distributed_01099_a AS local_01099_a ENGINE = Distributed('test_cluster_two_shards', currentDatabase(), local_01099_a, rand());
|
|
CREATE TABLE distributed_01099_b AS local_01099_b ENGINE = Distributed('test_cluster_two_shards', currentDatabase(), local_01099_b, rand());
|
|
|
|
SYSTEM STOP DISTRIBUTED SENDS distributed_01099_b;
|
|
SET prefer_localhost_replica=0; -- to require distributed send for local replica too
|
|
INSERT INTO local_01099_a SELECT number from system.numbers limit 3;
|
|
INSERT INTO distributed_01099_b SELECT * from distributed_01099_a;
|
|
SET prefer_localhost_replica=1;
|
|
|
|
-- distributed sends disabled, 0 rows (since parallel_distributed_insert_select=1)
|
|
SELECT 'distributed';
|
|
SELECT number, count(number) FROM distributed_01099_b group by number order by number;
|
|
SYSTEM FLUSH DISTRIBUTED distributed_01099_b;
|
|
|
|
SELECT 'local';
|
|
SELECT number, count(number) FROM local_01099_b group by number order by number;
|
|
SELECT 'distributed';
|
|
SELECT number, count(number) FROM distributed_01099_b group by number order by number;
|
|
|
|
DROP TABLE local_01099_a;
|
|
DROP TABLE local_01099_b;
|
|
DROP TABLE distributed_01099_a;
|
|
DROP TABLE distributed_01099_b;
|
|
|
|
|
|
SET parallel_distributed_insert_select=2;
|
|
SELECT 'parallel_distributed_insert_select=2';
|
|
|
|
--
|
|
-- test_shard_localhost
|
|
--
|
|
|
|
SELECT 'test_shard_localhost';
|
|
|
|
CREATE TABLE local_01099_a (number UInt64) ENGINE = Log;
|
|
CREATE TABLE local_01099_b (number UInt64) ENGINE = Log;
|
|
CREATE TABLE distributed_01099_a AS local_01099_a ENGINE = Distributed('test_shard_localhost', currentDatabase(), local_01099_a, rand());
|
|
CREATE TABLE distributed_01099_b AS local_01099_b ENGINE = Distributed('test_shard_localhost', currentDatabase(), local_01099_b, rand());
|
|
|
|
INSERT INTO local_01099_a SELECT number from system.numbers limit 3;
|
|
INSERT INTO distributed_01099_b SELECT * from distributed_01099_a;
|
|
|
|
SELECT * FROM distributed_01099_b;
|
|
|
|
DROP TABLE local_01099_a;
|
|
DROP TABLE local_01099_b;
|
|
DROP TABLE distributed_01099_a;
|
|
DROP TABLE distributed_01099_b;
|
|
|
|
--
|
|
-- test_cluster_two_shards_localhost
|
|
--
|
|
|
|
SELECT 'test_cluster_two_shards_localhost';
|
|
|
|
-- Log engine will lead to deadlock:
|
|
-- DB::Exception: std::system_error: Resource deadlock avoided.
|
|
-- So use MergeTree instead.
|
|
CREATE TABLE local_01099_a (number UInt64) ENGINE = MergeTree() ORDER BY number;
|
|
CREATE TABLE local_01099_b (number UInt64) ENGINE = MergeTree() ORDER BY number;
|
|
CREATE TABLE distributed_01099_a AS local_01099_a ENGINE = Distributed('test_cluster_two_shards_localhost', currentDatabase(), local_01099_a, rand());
|
|
CREATE TABLE distributed_01099_b AS local_01099_b ENGINE = Distributed('test_cluster_two_shards_localhost', currentDatabase(), local_01099_b, rand());
|
|
|
|
INSERT INTO local_01099_a SELECT number from system.numbers limit 3;
|
|
INSERT INTO distributed_01099_b SELECT * from distributed_01099_a;
|
|
|
|
SELECT number, count(number) FROM local_01099_b group by number order by number;
|
|
|
|
DROP TABLE local_01099_a;
|
|
DROP TABLE local_01099_b;
|
|
DROP TABLE distributed_01099_a;
|
|
DROP TABLE distributed_01099_b;
|
|
|
|
--
|
|
-- test_cluster_two_shards
|
|
--
|
|
|
|
SELECT 'test_cluster_two_shards';
|
|
|
|
CREATE TABLE local_01099_a (number UInt64) ENGINE = MergeTree() ORDER BY number;
|
|
CREATE TABLE local_01099_b (number UInt64) ENGINE = MergeTree() ORDER BY number;
|
|
CREATE TABLE distributed_01099_a AS local_01099_a ENGINE = Distributed('test_cluster_two_shards', currentDatabase(), local_01099_a, rand());
|
|
CREATE TABLE distributed_01099_b AS local_01099_b ENGINE = Distributed('test_cluster_two_shards', currentDatabase(), local_01099_b, rand());
|
|
|
|
SYSTEM STOP DISTRIBUTED SENDS distributed_01099_b;
|
|
SET prefer_localhost_replica=0; -- to require distributed send for local replica too
|
|
INSERT INTO local_01099_a SELECT number from system.numbers limit 3;
|
|
INSERT INTO distributed_01099_b SELECT * from distributed_01099_a;
|
|
SET prefer_localhost_replica=1;
|
|
|
|
-- distributed sends disabled, but they are not required, since insert is done into local table.
|
|
-- (since parallel_distributed_insert_select=2)
|
|
SELECT 'distributed';
|
|
SELECT number, count(number) FROM distributed_01099_b group by number order by number;
|
|
SELECT 'local';
|
|
SELECT number, count(number) FROM local_01099_b group by number order by number;
|
|
|
|
DROP TABLE local_01099_a;
|
|
DROP TABLE local_01099_b;
|
|
DROP TABLE distributed_01099_a;
|
|
DROP TABLE distributed_01099_b;
|