Remove unnecessary test diffs

This commit is contained in:
Raúl Marín 2024-10-29 16:41:18 +00:00
parent 1a259af626
commit e6969f5419
42 changed files with 80 additions and 40 deletions

View File

@ -16,4 +16,3 @@ no_implicit_reexport = True
strict_equality = True
extra_checks = True
ignore_missing_imports = True
logging-fstring-interpolation = False

View File

@ -22,7 +22,13 @@ $CLICKHOUSE_CLIENT --query "ALTER TABLE table_for_bad_alters MODIFY COLUMN value
sleep 2
while [[ $($CLICKHOUSE_CLIENT --query "KILL MUTATION WHERE mutation_id='0000000000' and database = '$CLICKHOUSE_DATABASE'" 2>&1) ]]; do
counter=0 retries=60
while [[ $counter -lt $retries ]]; do
output=$($CLICKHOUSE_CLIENT --query "KILL MUTATION WHERE mutation_id='0000000000' and database = '$CLICKHOUSE_DATABASE'" 2>&1)
if [[ "$output" == *"finished"* ]]; then
break
fi
((++counter))
sleep 1
done

View File

@ -1,6 +1,7 @@
#!/usr/bin/env bash
# Tags: no-replicated-database
# Tags: no-replicated-database, no-shared-merge-tree
# Tag no-replicated-database: CREATE AS SELECT is disabled
# Tag no-shared-merge-tree -- implemented separate test, just bad substituion here
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,4 +1,5 @@
-- Tags: zookeeper, no-parallel
-- Tags: zookeeper, no-parallel, no-shared-merge-tree
-- no-shared-merge-tree: doesn't support databases without UUID
drop database if exists test_1164_memory;
create database test_1164_memory engine=Memory;

View File

@ -1,4 +1,5 @@
-- Tags: zookeeper
-- Tags: zookeeper, no-shared-merge-tree
-- no-shared-merge-tree: shared merge tree doesn't loose data parts
SET max_rows_to_read = 0; -- system.text_log can be really big

View File

@ -1,3 +1,6 @@
-- Tags: no-shared-catalog
-- no-shared-catalog: standard MergeTree is not supported
drop table if exists trunc;
set default_table_engine='ReplicatedMergeTree';

View File

@ -1,5 +1,9 @@
#!/usr/bin/env bash
# Tags: zookeeper, no-parallel, no-fasttest
# Tags: zookeeper, no-parallel, no-fasttest, no-shared-merge-tree
# no-shared-merge-tree: database ordinary not supported
# Creation of a database with Ordinary engine emits a warning.
CLICKHOUSE_CLIENT_SERVER_LOGS_LEVEL=fatal
# Creation of a database with Ordinary engine emits a warning.
CLICKHOUSE_CLIENT_SERVER_LOGS_LEVEL=fatal

View File

@ -1,5 +1,5 @@
-- Tags: long, no-random-merge-tree-settings
-- FIXME no-random-merge-tree-settings requires investigation
--- FIXME no-random-merge-tree-settings requires investigation
drop table if exists data_01513;
create table data_01513 (key String) engine=MergeTree() order by key;

View File

@ -1,4 +1,4 @@
-- Tags: long
-- Tags: long, no-shared-merge-tree
DROP TABLE IF EXISTS test_01640;
DROP TABLE IF EXISTS restore_01640;
@ -32,5 +32,3 @@ SELECT _part, * FROM restore_01640;
DROP TABLE test_01640;
DROP TABLE restore_01640;

View File

@ -1,5 +1,6 @@
#!/usr/bin/env bash
# Tags: long, no-random-merge-tree-settings
# Tags: long, no-random-merge-tree-settings, no-random-settings
# no sanitizers -- bad idea to check memory usage with sanitizers
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,4 +1,5 @@
-- Tags: zookeeper
-- Tags: zookeeper, no-shared-merge-tree
-- no-shared-merge-tree: depend on replicated merge tree zookeeper structure
DROP TABLE IF EXISTS sample_table;

View File

@ -1,4 +1,5 @@
-- Tags: long, zookeeper
-- Tags: long, zookeeper, no-shared-merge-tree
-- no-shared-merge-tree: no replication queue
DROP TABLE IF EXISTS i20203_1 SYNC;
DROP TABLE IF EXISTS i20203_2 SYNC;

View File

@ -89,7 +89,7 @@ SELECT dictGetOrDefault('01754_dictionary_db.direct_dictionary_complex_key_compl
SELECT 'dictHas';
SELECT dictHas('01754_dictionary_db.direct_dictionary_complex_key_complex_attributes', (number, concat('id_key_', toString(number)))) FROM system.numbers LIMIT 4;
SELECT 'select all values as input stream';
SELECT * FROM 01754_dictionary_db.direct_dictionary_complex_key_complex_attributes;
SELECT * FROM 01754_dictionary_db.direct_dictionary_complex_key_complex_attributes ORDER BY ALL;
DROP DICTIONARY 01754_dictionary_db.direct_dictionary_complex_key_complex_attributes;
DROP TABLE 01754_dictionary_db.complex_key_complex_attributes_source_table;

View File

@ -1,5 +1,6 @@
#!/usr/bin/env bash
# Tags: long, no-tsan, no-debug, no-asan, no-msan, no-ubsan
# Tags: long, no-tsan, no-debug, no-asan, no-msan, no-ubsan, no-shared-merge-tree
# no-shared-merge-tree -- this test is too slow
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,5 +1,6 @@
#!/usr/bin/env bash
# Tags: long, no-tsan, no-debug, no-asan, no-msan, no-ubsan, no-parallel
# Tags: long, no-tsan, no-debug, no-asan, no-msan, no-ubsan, no-parallel, no-shared-merge-tree
# no-shared-merge-tree -- this test is too slow
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,3 +1,5 @@
SET mutations_sync=2;
DROP TABLE IF EXISTS t_ephemeral_02205_1;
CREATE TABLE t_ephemeral_02205_1 (x UInt32 DEFAULT y, y UInt32 EPHEMERAL 17, z UInt32 DEFAULT 5) ENGINE = Memory;

View File

@ -1,5 +1,6 @@
#!/usr/bin/env bash
# Tags: no-replicated-database, zookeeper
# Tags: no-replicated-database, zookeeper, no-shared-merge-tree
# no-shared-merge-tree: depend on specific paths created by replicated tables
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,6 +1,7 @@
#!/usr/bin/env bash
# Tags: zookeeper, no-replicated-database
# Tags: zookeeper, no-replicated-database, no-shared-merge-tree
# no-replicated-database because it adds extra replicas
# no-shared-merge-tree do something with parts on local fs
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,5 +1,5 @@
#!/usr/bin/env bash
# Tags: long, zookeeper
# Tags: long, zookeeper, no-shared-merge-tree
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,5 +1,5 @@
#!/usr/bin/env bash
# Tags: long, zookeeper
# Tags: long, zookeeper, no-shared-merge-tree
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,5 +1,6 @@
#!/usr/bin/env bash
# Tags: long, zookeeper
# Tags: zookeeper, no-shared-merge-tree, long
# no-shared-merge-tree: depend on local fs
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,5 +1,6 @@
#!/usr/bin/env bash
# Tags: long, zookeeper
# Tags: long, zookeeper, no-shared-merge-tree
# no-shared-merge-tree: depend on local fs (remove parts)
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -11,6 +11,8 @@ CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=./parts.lib
. "$CURDIR"/parts.lib
CLICKHOUSE_CLIENT="$CLICKHOUSE_CLIENT --apply_mutations_on_fly 0"
function reset_table()
{
table=${1:-"tt"}

View File

@ -1,10 +1,13 @@
-- Tags: no-fasttest
-- Tags: no-fasttest, no-shared-merge-tree
-- no-shared-merge-tree: depend on custom storage policy
SET send_logs_level = 'fatal';
drop table if exists rmt;
drop table if exists rmt2;
set apply_mutations_on_fly = 0;
-- Disable compact parts, because we need hardlinks in mutations.
create table rmt (n int, m int, k int) engine=ReplicatedMergeTree('/test/02432/{database}', '1') order by tuple()
settings storage_policy = 's3_cache', allow_remote_fs_zero_copy_replication=1,

View File

@ -1,4 +1,5 @@
-- Tags: long
-- Tags: long, no-shared-merge-tree
-- no-shared-merge-tree: depend on replication queue/fetches
SET insert_keeper_fault_injection_probability=0; -- disable fault injection; part ids are non-deterministic in case of insert retries

View File

@ -1,4 +1,5 @@
-- Tags: no-object-storage, no-replicated-database
-- Tags: no-object-storage, no-replicated-database, no-shared-merge-tree
-- no-shared-merge-tree: custom disk
DROP TABLE IF EXISTS test;

View File

@ -1,7 +1,7 @@
#!/usr/bin/env bash
# Tags: replica, no-fasttest
# Tags: replica, no-fasttest, no-shared-merge-tree
# no-fasttest: Mutation load can be slow
# no-shared-merge-tree -- have separate test for it
CUR_DIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh
. "$CUR_DIR"/../shell_config.sh

View File

@ -1,4 +1,4 @@
-- Tags: no-object-storage, no-random-merge-tree-settings
-- Tags: no-object-storage, no-random-merge-tree-settings, no-fasttest
-- We allocate a lot of memory for buffers when reading or writing to S3
DROP TABLE IF EXISTS 02725_memory_for_merges SYNC;

View File

@ -1,4 +1,5 @@
-- Tags: no-fasttest, no-replicated-database
-- Tags: no-fasttest, no-replicated-database, no-shared-merge-tree
-- no-shared-merge-tree -- smt doesn't support aux zookeepers
DROP TABLE IF EXISTS test_zk_connection_table;

View File

@ -1,3 +1,5 @@
-- Tags: no-shared-merge-tree
DROP TABLE IF EXISTS checksums_r3;
DROP TABLE IF EXISTS checksums_r2;
DROP TABLE IF EXISTS checksums_r1;

View File

@ -1,4 +1,5 @@
-- Tags: zookeeper, no-replicated-database
-- Tags: zookeeper, no-replicated-database, no-shared-merge-tree
CREATE TABLE t
(
`id` UInt64,

View File

@ -1,4 +1,4 @@
-- Tags: no-replicated-database, no-fasttest
-- Tags: no-replicated-database, no-fasttest, no-shared-merge-tree
-- Tag no-replicated-database: different number of replicas
create table tableIn (n int)

View File

@ -1,4 +1,5 @@
-- Tags: zookeeper, no-parallel
-- Tags: zookeeper, no-parallel, no-shared-merge-tree
-- no-shared-merge-tree: This failure injection is only RMT specific
DROP TABLE IF EXISTS t_hardware_error NO DELAY;

View File

@ -1,5 +1,5 @@
#!/usr/bin/env bash
# Tags: long, no-replicated-database, no-fasttest
# Tags: long, no-replicated-database, no-fasttest, no-shared-merge-tree
set -e

View File

@ -1,6 +1,7 @@
#!/usr/bin/env bash
# Tags: no-parallel
# Tags: no-parallel, no-shared-merge-tree
# Tag no-parallel: failpoint is in use
# Tag no-shared-merge-tree: looks like it tests a specific behaviour of ReplicatedMergeTree with failpoints
CUR_DIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh

View File

@ -1,5 +1,5 @@
#!/usr/bin/env bash
# Tags: long, no-random-settings, no-random-merge-tree-settings, no-tsan, no-msan, no-ubsan, no-asan
# Tags: long, no-random-settings, no-random-merge-tree-settings, no-tsan, no-msan, no-ubsan, no-asan, no-debug
# no sanitizers: too slow
CURDIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)

View File

@ -16,4 +16,4 @@ using id;
INSERT INTO Null SELECT number AS id FROM numbers(2);
select * from Example; -- should return 4 rows
select * from Example order by all; -- should return 4 rows

View File

@ -1,5 +1,6 @@
-- Tags: no-fasttest
-- Tags: no-fasttest, no-shared-merge-tree
-- Tag no-fasttest -- due to EmbeddedRocksDB
-- Tag no-shared-merge-tree -- due to system.replication_queue
drop table if exists null;
drop table if exists dist;

View File

@ -11,7 +11,7 @@ function tx()
session="${CLICKHOUSE_TEST_ZOOKEEPER_PREFIX}_tx$tx_num"
query_id="${session}_${RANDOM}"
url_without_session="https://${CLICKHOUSE_HOST}:${CLICKHOUSE_PORT_HTTPS}/?"
url="${url_without_session}session_id=$session&query_id=$query_id&database=$CLICKHOUSE_DATABASE"
url="${url_without_session}session_id=$session&query_id=$query_id&database=$CLICKHOUSE_DATABASE&apply_mutations_on_fly=0"
${CLICKHOUSE_CURL} -m 90 -sSk "$url" --data "$query" | sed "s/^/tx$tx_num\t/"
}
@ -56,7 +56,7 @@ function tx_async()
session="${CLICKHOUSE_TEST_ZOOKEEPER_PREFIX}_tx$tx_num"
query_id="${session}_${RANDOM}"
url_without_session="https://${CLICKHOUSE_HOST}:${CLICKHOUSE_PORT_HTTPS}/?"
url="${url_without_session}session_id=$session&query_id=$query_id&database=$CLICKHOUSE_DATABASE"
url="${url_without_session}session_id=$session&query_id=$query_id&database=$CLICKHOUSE_DATABASE&apply_mutations_on_fly=0"
# We cannot be sure that query will actually start execution and appear in system.processes before the next call to tx_wait
# Also we cannot use global map in bash to store last query_id for each tx_num, so we use tmp file...

View File

@ -3,6 +3,7 @@
SET allow_prefetched_read_pool_for_remote_filesystem=0;
SET enable_filesystem_cache_on_write_operations=0;
SET max_memory_usage='20G';
SET read_through_distributed_cache = 1;
SYSTEM DROP FILESYSTEM CACHE;
SELECT count() FROM test.hits_s3;
8873898

View File

@ -5,6 +5,7 @@
SET allow_prefetched_read_pool_for_remote_filesystem=0;
SET enable_filesystem_cache_on_write_operations=0;
SET max_memory_usage='20G';
SET read_through_distributed_cache = 1;
SYSTEM DROP FILESYSTEM CACHE;
SELECT count() FROM test.hits_s3;
SELECT count() FROM test.hits_s3 WHERE AdvEngineID != 0;