Avoid using s3_truncate_on_insert in 02700_s3_part_INT_MAX

Seems that version of MinIO that is used on CI is too slow for this [1],
I've tried the same locally with RELEASE.2023-02-10T18-48-39Z and
everythings works OK, but I have NVME locally and more recent MinIO.

  [1]: https://s3.amazonaws.com/clickhouse-test-reports/47693/5c701b090c3ec10cb2a4a708b60d364eb02192fe/stateless_tests_flaky_check__asan_.html

Signed-off-by: Azat Khuzhin <a.khuzhin@semrush.com>
This commit is contained in:
Azat Khuzhin 2023-03-21 08:54:05 +02:00
parent 8dcee4c482
commit f55f33b64d
2 changed files with 19 additions and 8 deletions

View File

@ -0,0 +1,19 @@
#!/usr/bin/env bash
# Tags: no-parallel, long
CUR_DIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh
. "$CUR_DIR"/../shell_config.sh
# Regression test for crash in case of part exceeds INT_MAX
#
# NOTE: .sh test is used over .sql because it needs $CLICKHOUSE_DATABASE to
# avoid truncation, since seems that the version of MinIO that is used on CI
# too slow with this.
$CLICKHOUSE_CLIENT -nm -q "
INSERT INTO FUNCTION s3('http://localhost:11111/test/$CLICKHOUSE_DATABASE/test_INT_MAX.tsv', '', '', 'TSV')
SELECT repeat('a', 1024) FROM numbers((pow(2, 30) * 2) / 1024)
SETTINGS s3_max_single_part_upload_size = '10Gi';
SELECT count() FROM s3('http://localhost:11111/test/$CLICKHOUSE_DATABASE/test_INT_MAX.tsv');
"

View File

@ -1,8 +0,0 @@
-- Tags: no-parallel, long
-- Regression test for crash in case of part exceeds INT_MAX
INSERT INTO FUNCTION s3('http://localhost:11111/test/test_INT_MAX.tsv', '', '', 'TSV')
SELECT repeat('a', 1024) FROM numbers((pow(2, 30) * 2) / 1024)
SETTINGS s3_truncate_on_insert = 1, s3_max_single_part_upload_size = '10Gi';
SELECT count() FROM s3('http://localhost:11111/test/test_INT_MAX.tsv');