fix settings in tests

This commit is contained in:
Yarik Briukhovetskyi 2024-07-05 19:36:59 +02:00 committed by GitHub
parent 6c4c17f119
commit 9064bb1b83
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 15 additions and 15 deletions

View File

@ -1484,7 +1484,7 @@ def test_hive_partitioning_with_one_parameter(cluster):
f"blob_path='{path}', format='CSV', structure='{table_format}')"
)
assert azure_query(
node, query, settings={"azure_blob_storage_hive_partitioning": 1}
node, query, settings={"use_hive_partitioning": 1}
).splitlines() == [
"Elizabeth\tGordon\tsample.csv\t{bucket}/{max_path}\tElizabeth".format(
bucket="cont", max_path=path
@ -1497,7 +1497,7 @@ def test_hive_partitioning_with_one_parameter(cluster):
f"blob_path='{path}', format='CSV', structure='{table_format}') WHERE column1=_column1;"
)
assert azure_query(
node, query, settings={"azure_blob_storage_hive_partitioning": 1}
node, query, settings={"use_hive_partitioning": 1}
).splitlines() == ["Gordon"]
@ -1521,7 +1521,7 @@ def test_hive_partitioning_with_two_parameters(cluster):
f"blob_path='{path}', format='CSV', structure='{table_format}') WHERE column1=_column1;"
)
assert azure_query(
node, query, settings={"azure_blob_storage_hive_partitioning": 1}
node, query, settings={"use_hive_partitioning": 1}
).splitlines() == [
"Elizabeth\tGordon\tsample.csv\t{bucket}/{max_path}\tElizabeth\tGordon".format(
bucket="cont", max_path=path
@ -1534,7 +1534,7 @@ def test_hive_partitioning_with_two_parameters(cluster):
f"blob_path='{path}', format='CSV', structure='{table_format}') WHERE column2=_column2;"
)
assert azure_query(
node, query, settings={"azure_blob_storage_hive_partitioning": 1}
node, query, settings={"use_hive_partitioning": 1}
).splitlines() == ["Elizabeth"]
query = (
@ -1543,7 +1543,7 @@ def test_hive_partitioning_with_two_parameters(cluster):
f"blob_path='{path}', format='CSV', structure='{table_format}') WHERE column2=_column2 AND column1=_column1;"
)
assert azure_query(
node, query, settings={"azure_blob_storage_hive_partitioning": 1}
node, query, settings={"use_hive_partitioning": 1}
).splitlines() == ["Elizabeth"]
@ -1571,4 +1571,4 @@ def test_hive_partitioning_without_setting(cluster):
)
with pytest.raises(Exception, match=pattern):
azure_query(node, query, settings={"azure_blob_storage_hive_partitioning": 0})
azure_query(node, query, settings={"use_hive_partitioning": 0})

View File

@ -1188,7 +1188,7 @@ def test_hive_partitioning_with_one_parameter(started_cluster):
r = node1.query(
"SELECT _column0 FROM hdfs('hdfs://hdfs1:9000/column0=Elizabeth/parquet_1', 'TSV')",
settings={"hdfs_hive_partitioning": 1},
settings={"use_hive_partitioning": 1},
)
assert r == f"Elizabeth\n"
@ -1205,7 +1205,7 @@ def test_hive_partitioning_with_two_parameters(started_cluster):
r = node1.query(
"SELECT _column1 FROM hdfs('hdfs://hdfs1:9000/column0=Elizabeth/column1=Gordon/parquet_2', 'TSV');",
settings={"hdfs_hive_partitioning": 1},
settings={"use_hive_partitioning": 1},
)
assert r == f"Gordon\n"
@ -1226,7 +1226,7 @@ def test_hive_partitioning_without_setting(started_cluster):
with pytest.raises(QueryRuntimeException, match=pattern):
node1.query(
f"SELECT _column1 FROM hdfs('hdfs://hdfs1:9000/column0=Elizabeth/column1=Gordon/parquet_2', 'TSV');",
settings={"hdfs_hive_partitioning": 0},
settings={"use_hive_partitioning": 0},
)

View File

@ -9,7 +9,7 @@ $CLICKHOUSE_LOCAL -q "SELECT 'TESTING THE FILE HIVE PARTITIONING'"
$CLICKHOUSE_LOCAL -n -q """
set file_hive_partitioning = 1;
set use_hive_partitioning = 1;
SELECT *, _column0 FROM file('$CURDIR/data_hive/partitioning/column0=Elizabeth/sample.parquet') LIMIT 10;
@ -31,7 +31,7 @@ SELECT *, _non_existing_column FROM file('$CURDIR/data_hive/partitioning/non_exi
SELECT *, _column0 FROM file('$CURDIR/data_hive/partitioning/column0=*/sample.parquet') WHERE column0 = _column0;"""
$CLICKHOUSE_LOCAL -n -q """
set file_hive_partitioning = 0;
set use_hive_partitioning = 0;
SELECT *, _column0 FROM file('$CURDIR/data_hive/partitioning/column0=Elizabeth/sample.parquet') LIMIT 10;
""" 2>&1 | grep -c "UNKNOWN_IDENTIFIER"
@ -41,7 +41,7 @@ $CLICKHOUSE_LOCAL -q "SELECT 'TESTING THE URL PARTITIONING'"
$CLICKHOUSE_LOCAL -n -q """
set url_hive_partitioning = 1;
set use_hive_partitioning = 1;
SELECT *, _column0 FROM url('http://localhost:11111/test/partitioning/column0=Elizabeth/sample.parquet') LIMIT 10;
@ -63,7 +63,7 @@ SELECT *, _non_existing_column FROM url('http://localhost:11111/test/partitionin
SELECT *, _column0 FROM url('http://localhost:11111/test/partitioning/column0=*/sample.parquet') WHERE column0 = _column0;"""
$CLICKHOUSE_LOCAL -n -q """
set url_hive_partitioning = 0;
set use_hive_partitioning = 0;
SELECT *, _column0 FROM url('http://localhost:11111/test/partitioning/column0=Elizabeth/sample.parquet') LIMIT 10;
""" 2>&1 | grep -c "UNKNOWN_IDENTIFIER"
@ -73,7 +73,7 @@ $CLICKHOUSE_LOCAL -q "SELECT 'TESTING THE S3 PARTITIONING'"
$CLICKHOUSE_LOCAL -n -q """
set s3_hive_partitioning = 1;
set use_hive_partitioning = 1;
SELECT *, _column0 FROM s3('http://localhost:11111/test/partitioning/column0=Elizabeth/sample.parquet') LIMIT 10;
@ -96,7 +96,7 @@ SELECT *, _column0 FROM s3('http://localhost:11111/test/partitioning/column0=*/s
"""
$CLICKHOUSE_LOCAL -n -q """
set s3_hive_partitioning = 0;
set use_hive_partitioning = 0;
SELECT *, _column0 FROM s3('http://localhost:11111/test/partitioning/column0=Elizabeth/sample.parquet') LIMIT 10;
""" 2>&1 | grep -c "UNKNOWN_IDENTIFIER"