ClickHouse/tests/queries/0_stateless/02700_s3_part_INT_MAX.sh
Azat Khuzhin f55f33b64d Avoid using s3_truncate_on_insert in 02700_s3_part_INT_MAX
Seems that version of MinIO that is used on CI is too slow for this [1],
I've tried the same locally with RELEASE.2023-02-10T18-48-39Z and
everythings works OK, but I have NVME locally and more recent MinIO.

  [1]: https://s3.amazonaws.com/clickhouse-test-reports/47693/5c701b090c3ec10cb2a4a708b60d364eb02192fe/stateless_tests_flaky_check__asan_.html

Signed-off-by: Azat Khuzhin <a.khuzhin@semrush.com>
2023-04-17 10:47:42 +02:00

20 lines
771 B
Bash
Executable File

#!/usr/bin/env bash
# Tags: no-parallel, long
CUR_DIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)
# shellcheck source=../shell_config.sh
. "$CUR_DIR"/../shell_config.sh
# Regression test for crash in case of part exceeds INT_MAX
#
# NOTE: .sh test is used over .sql because it needs $CLICKHOUSE_DATABASE to
# avoid truncation, since seems that the version of MinIO that is used on CI
# too slow with this.
$CLICKHOUSE_CLIENT -nm -q "
INSERT INTO FUNCTION s3('http://localhost:11111/test/$CLICKHOUSE_DATABASE/test_INT_MAX.tsv', '', '', 'TSV')
SELECT repeat('a', 1024) FROM numbers((pow(2, 30) * 2) / 1024)
SETTINGS s3_max_single_part_upload_size = '10Gi';
SELECT count() FROM s3('http://localhost:11111/test/$CLICKHOUSE_DATABASE/test_INT_MAX.tsv');
"