mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-01 12:01:58 +00:00
rm dirs in test_storage_delta
This commit is contained in:
parent
1d85f9b1cb
commit
9c05a0ad5a
@ -8,6 +8,7 @@ import os
|
|||||||
import json
|
import json
|
||||||
import time
|
import time
|
||||||
import glob
|
import glob
|
||||||
|
import shutil
|
||||||
|
|
||||||
import pyspark
|
import pyspark
|
||||||
import delta
|
import delta
|
||||||
@ -52,15 +53,6 @@ def get_spark():
|
|||||||
return builder.master("local").getOrCreate()
|
return builder.master("local").getOrCreate()
|
||||||
|
|
||||||
|
|
||||||
def remove_local_directory_contents(full_path):
|
|
||||||
for path in glob.glob(f"{full_path}/**"):
|
|
||||||
if os.path.isfile(path):
|
|
||||||
os.unlink(path)
|
|
||||||
else:
|
|
||||||
remove_local_directory_contents(path)
|
|
||||||
os.rmdir(path)
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="module")
|
@pytest.fixture(scope="module")
|
||||||
def started_cluster():
|
def started_cluster():
|
||||||
try:
|
try:
|
||||||
@ -179,7 +171,7 @@ def test_single_log_file(started_cluster):
|
|||||||
)
|
)
|
||||||
|
|
||||||
os.unlink(parquet_data_path)
|
os.unlink(parquet_data_path)
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
def test_partition_by(started_cluster):
|
def test_partition_by(started_cluster):
|
||||||
@ -203,7 +195,7 @@ def test_partition_by(started_cluster):
|
|||||||
create_delta_table(instance, TABLE_NAME)
|
create_delta_table(instance, TABLE_NAME)
|
||||||
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 10
|
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 10
|
||||||
|
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
def test_checkpoint(started_cluster):
|
def test_checkpoint(started_cluster):
|
||||||
@ -280,7 +272,7 @@ def test_checkpoint(started_cluster):
|
|||||||
).strip()
|
).strip()
|
||||||
)
|
)
|
||||||
|
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
spark.sql(f"DROP TABLE {TABLE_NAME}")
|
spark.sql(f"DROP TABLE {TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
@ -321,7 +313,7 @@ def test_multiple_log_files(started_cluster):
|
|||||||
"SELECT number, toString(number + 1) FROM numbers(200)"
|
"SELECT number, toString(number + 1) FROM numbers(200)"
|
||||||
)
|
)
|
||||||
|
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
def test_metadata(started_cluster):
|
def test_metadata(started_cluster):
|
||||||
@ -357,7 +349,7 @@ def test_metadata(started_cluster):
|
|||||||
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 100
|
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 100
|
||||||
|
|
||||||
os.unlink(parquet_data_path)
|
os.unlink(parquet_data_path)
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
def test_types(started_cluster):
|
def test_types(started_cluster):
|
||||||
@ -431,7 +423,7 @@ def test_types(started_cluster):
|
|||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
remove_local_directory_contents(f"/{result_file}")
|
shutil.rmtree(f"/{result_file}")
|
||||||
spark.sql(f"DROP TABLE {TABLE_NAME}")
|
spark.sql(f"DROP TABLE {TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
@ -496,7 +488,7 @@ def test_restart_broken(started_cluster):
|
|||||||
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 100
|
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 100
|
||||||
|
|
||||||
os.unlink(parquet_data_path)
|
os.unlink(parquet_data_path)
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
def test_restart_broken_table_function(started_cluster):
|
def test_restart_broken_table_function(started_cluster):
|
||||||
@ -553,7 +545,7 @@ def test_restart_broken_table_function(started_cluster):
|
|||||||
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 100
|
assert int(instance.query(f"SELECT count() FROM {TABLE_NAME}")) == 100
|
||||||
|
|
||||||
os.unlink(parquet_data_path)
|
os.unlink(parquet_data_path)
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
|
|
||||||
|
|
||||||
def test_partition_columns(started_cluster):
|
def test_partition_columns(started_cluster):
|
||||||
@ -753,5 +745,5 @@ SELECT * FROM deltaLake('http://{started_cluster.minio_ip}:{started_cluster.mini
|
|||||||
== 1
|
== 1
|
||||||
)
|
)
|
||||||
|
|
||||||
remove_local_directory_contents(f"/{TABLE_NAME}")
|
shutil.rmtree(f"/{TABLE_NAME}")
|
||||||
spark.sql(f"DROP TABLE {TABLE_NAME}")
|
spark.sql(f"DROP TABLE {TABLE_NAME}")
|
||||||
|
Loading…
Reference in New Issue
Block a user