mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-21 15:12:02 +00:00
Copy some tests with granulairty
This commit is contained in:
parent
b027d745a7
commit
3f29e24be1
@ -0,0 +1,32 @@
|
||||
----00489----
|
||||
1 11 1235
|
||||
2 11 4395
|
||||
3 22 3545
|
||||
4 22 6984
|
||||
5 33 4596
|
||||
123 33 1235
|
||||
121 33 2791
|
||||
122 33 2791
|
||||
125 44 4578
|
||||
124 44 4935
|
||||
128 55 1235
|
||||
127 55 2791
|
||||
126 55 5786
|
||||
1 11 1235
|
||||
63 11 3572
|
||||
61 11 4563
|
||||
62 11 4578
|
||||
63 11 3572
|
||||
61 11 4563
|
||||
62 11 4578
|
||||
----00607----
|
||||
1
|
||||
1
|
||||
1
|
||||
1
|
||||
1
|
||||
1
|
||||
----00804----
|
||||
2
|
||||
1
|
||||
|
@ -0,0 +1,85 @@
|
||||
SET send_logs_level = 'none';
|
||||
SELECT '----00489----';
|
||||
DROP TABLE IF EXISTS test.pk;
|
||||
|
||||
CREATE TABLE test.pk (d Date DEFAULT '2000-01-01', x DateTime, y UInt64, z UInt64) ENGINE = MergeTree() PARTITION BY d ORDER BY (toStartOfMinute(x), y, z) SETTINGS index_granularity_bytes=19; -- one row granule
|
||||
|
||||
INSERT INTO test.pk (x, y, z) VALUES (1, 11, 1235), (2, 11, 4395), (3, 22, 3545), (4, 22, 6984), (5, 33, 4596), (61, 11, 4563), (62, 11, 4578), (63, 11, 3572), (64, 22, 5786), (65, 22, 5786), (66, 22, 2791), (67, 22, 2791), (121, 33, 2791), (122, 33, 2791), (123, 33, 1235), (124, 44, 4935), (125, 44, 4578), (126, 55, 5786), (127, 55, 2791), (128, 55, 1235);
|
||||
|
||||
SET min_insert_block_size_rows = 0, min_insert_block_size_bytes = 0;
|
||||
SET max_block_size = 1;
|
||||
|
||||
-- Test inferred limit
|
||||
SET max_rows_to_read = 5;
|
||||
SELECT toUInt32(x), y, z FROM test.pk WHERE x BETWEEN toDateTime(0) AND toDateTime(59);
|
||||
|
||||
SET max_rows_to_read = 9;
|
||||
SELECT toUInt32(x), y, z FROM test.pk WHERE x BETWEEN toDateTime(120) AND toDateTime(240);
|
||||
|
||||
-- Index is coarse, cannot read single row
|
||||
SET max_rows_to_read = 5;
|
||||
SELECT toUInt32(x), y, z FROM test.pk WHERE x = toDateTime(1);
|
||||
|
||||
-- Index works on interval 00:01:00 - 00:01:59
|
||||
SET max_rows_to_read = 4;
|
||||
SELECT toUInt32(x), y, z FROM test.pk WHERE x BETWEEN toDateTime(60) AND toDateTime(119) AND y = 11;
|
||||
|
||||
-- Cannot read less rows as PK is coarser on interval 00:01:00 - 00:02:00
|
||||
SET max_rows_to_read = 5;
|
||||
SELECT toUInt32(x), y, z FROM test.pk WHERE x BETWEEN toDateTime(60) AND toDateTime(120) AND y = 11;
|
||||
|
||||
DROP TABLE test.pk;
|
||||
|
||||
SELECT '----00607----';
|
||||
|
||||
SET max_rows_to_read = 0;
|
||||
DROP TABLE IF EXISTS test.merge_tree;
|
||||
CREATE TABLE test.merge_tree (x UInt32) ENGINE = MergeTree ORDER BY x SETTINGS index_granularity_bytes = 4;
|
||||
INSERT INTO test.merge_tree VALUES (0), (1);
|
||||
|
||||
SET force_primary_key = 1;
|
||||
SET max_rows_to_read = 1;
|
||||
|
||||
SELECT count() FROM test.merge_tree WHERE x = 0;
|
||||
SELECT count() FROM test.merge_tree WHERE toUInt32(x) = 0;
|
||||
SELECT count() FROM test.merge_tree WHERE toUInt64(x) = 0;
|
||||
|
||||
SELECT count() FROM test.merge_tree WHERE x IN (0, 0);
|
||||
SELECT count() FROM test.merge_tree WHERE toUInt32(x) IN (0, 0);
|
||||
SELECT count() FROM test.merge_tree WHERE toUInt64(x) IN (0, 0);
|
||||
|
||||
DROP TABLE test.merge_tree;
|
||||
|
||||
SELECT '----00804----';
|
||||
|
||||
SET max_rows_to_read = 0;
|
||||
DROP TABLE IF EXISTS test.large_alter_table;
|
||||
DROP TABLE IF EXISTS test.store_of_hash;
|
||||
|
||||
CREATE TABLE test.large_alter_table (
|
||||
somedate Date CODEC(ZSTD, ZSTD, ZSTD(12), LZ4HC(12)),
|
||||
id UInt64 CODEC(LZ4, ZSTD, NONE, LZ4HC),
|
||||
data String CODEC(ZSTD(2), LZ4HC, NONE, LZ4, LZ4)
|
||||
) ENGINE = MergeTree() PARTITION BY somedate ORDER BY id SETTINGS index_granularity_bytes=19;
|
||||
|
||||
INSERT INTO test.large_alter_table SELECT toDate('2019-01-01'), number, toString(number + rand()) FROM system.numbers LIMIT 300000;
|
||||
|
||||
CREATE TABLE test.store_of_hash (hash UInt64) ENGINE = Memory();
|
||||
|
||||
INSERT INTO test.store_of_hash SELECT sum(cityHash64(*)) FROM test.large_alter_table;
|
||||
|
||||
ALTER TABLE test.large_alter_table MODIFY COLUMN data CODEC(NONE, LZ4, LZ4HC, ZSTD);
|
||||
|
||||
OPTIMIZE TABLE test.large_alter_table;
|
||||
|
||||
DETACH TABLE test.large_alter_table;
|
||||
ATTACH TABLE test.large_alter_table;
|
||||
|
||||
INSERT INTO test.store_of_hash SELECT sum(cityHash64(*)) FROM test.large_alter_table;
|
||||
|
||||
SELECT COUNT(hash) FROM test.store_of_hash;
|
||||
SELECT COUNT(DISTINCT hash) FROM test.store_of_hash;
|
||||
|
||||
DROP TABLE IF EXISTS test.large_alter_table;
|
||||
DROP TABLE IF EXISTS test.store_of_hash;
|
||||
|
Loading…
Reference in New Issue
Block a user