import logging import os import time import pytest from helpers.client import QueryRuntimeException from helpers.cluster import ClickHouseCluster # Runs custom python-based S3 endpoint. def run_endpoint(cluster): logging.info("Starting custom S3 endpoint") container_id = cluster.get_container_id("resolver") current_dir = os.path.dirname(__file__) cluster.copy_file_to_container( container_id, os.path.join(current_dir, "s3_endpoint", "endpoint.py"), "endpoint.py", ) cluster.exec_in_container(container_id, ["python", "endpoint.py"], detach=True) # Wait for S3 endpoint start num_attempts = 100 for attempt in range(num_attempts): ping_response = cluster.exec_in_container( cluster.get_container_id("resolver"), ["curl", "-s", "http://resolver:8080/"], nothrow=True, ) if ping_response != "OK": if attempt == num_attempts - 1: assert ping_response == "OK", 'Expected "OK", but got "{}"'.format( ping_response ) else: time.sleep(1) else: break logging.info("S3 endpoint started") def fail_request(cluster, request): response = cluster.exec_in_container( cluster.get_container_id("resolver"), ["curl", "-s", "http://resolver:8080/fail_request/{}".format(request)], ) assert response == "OK", 'Expected "OK", but got "{}"'.format(response) def throttle_request(cluster, request): response = cluster.exec_in_container( cluster.get_container_id("resolver"), ["curl", "-s", "http://resolver:8080/throttle_request/{}".format(request)], ) assert response == "OK", 'Expected "OK", but got "{}"'.format(response) @pytest.fixture(scope="module") def cluster(): try: cluster = ClickHouseCluster(__file__) cluster.add_instance( "node", main_configs=[ "configs/config.d/storage_conf.xml", "configs/config.d/instant_moves.xml", "configs/config.d/part_log.xml", ], with_minio=True, ) logging.info("Starting cluster...") cluster.start() logging.info("Cluster started") run_endpoint(cluster) yield cluster finally: cluster.shutdown() @pytest.fixture(autouse=True) def drop_table(cluster): yield node = cluster.instances["node"] node.query("DROP TABLE IF EXISTS s3_failover_test NO DELAY") # S3 request will be failed for an appropriate part file write. FILES_PER_PART_BASE = 5 # partition.dat, default_compression_codec.txt, count.txt, columns.txt, checksums.txt FILES_PER_PART_WIDE = ( FILES_PER_PART_BASE + 1 + 1 + 3 * 2 ) # Primary index, MinMax, Mark and data file for column(s) # In debug build there are additional requests (from MergeTreeDataPartWriterWide.cpp:554 due to additional validation). FILES_PER_PART_WIDE_DEBUG = 2 # Additional requests to S3 in debug build FILES_PER_PART_COMPACT = FILES_PER_PART_BASE + 1 + 1 + 2 FILES_PER_PART_COMPACT_DEBUG = 0 @pytest.mark.parametrize( "min_bytes_for_wide_part,request_count,debug_request_count", [ (0, FILES_PER_PART_WIDE, FILES_PER_PART_WIDE_DEBUG), (1024 * 1024, FILES_PER_PART_COMPACT, FILES_PER_PART_COMPACT_DEBUG), ], ) def test_write_failover( cluster, min_bytes_for_wide_part, request_count, debug_request_count ): node = cluster.instances["node"] node.query( """ CREATE TABLE s3_failover_test ( dt Date, id Int64, data String ) ENGINE=MergeTree() ORDER BY id PARTITION BY dt SETTINGS storage_policy='s3', min_bytes_for_wide_part={} """.format( min_bytes_for_wide_part ) ) is_debug_mode = False success_count = 0 for request in range(request_count + debug_request_count + 1): # Fail N-th request to S3. fail_request(cluster, request + 1) data = "('2020-03-01',0,'data'),('2020-03-01',1,'data')" positive = request >= ( request_count + debug_request_count if is_debug_mode else request_count ) try: node.query("INSERT INTO s3_failover_test VALUES {}".format(data)) assert positive, "Insert query should be failed, request {}".format(request) success_count += 1 except QueryRuntimeException as e: if not is_debug_mode and positive: is_debug_mode = True positive = False assert not positive, "Insert query shouldn't be failed, request {}".format( request ) assert str(e).find("Expected Error") != -1, "Unexpected error {}".format( str(e) ) if positive: # Disable request failing. fail_request(cluster, 0) assert node.query("CHECK TABLE s3_failover_test") == "1\n" assert ( success_count > 1 or node.query("SELECT * FROM s3_failover_test FORMAT Values") == data ) assert success_count == ( 1 if is_debug_mode else debug_request_count + 1 ), "Insert query should be successful at least once" # Check that second data part move is ended successfully if first attempt was failed. def test_move_failover(cluster): node = cluster.instances["node"] node.query( """ CREATE TABLE s3_failover_test ( dt DateTime, id Int64, data String ) ENGINE=MergeTree() ORDER BY id TTL dt + INTERVAL 4 SECOND TO VOLUME 'external' SETTINGS storage_policy='s3_cold' """ ) # Fail a request to S3 to break first TTL move. fail_request(cluster, 1) node.query( "INSERT INTO s3_failover_test VALUES (now() - 1, 0, 'data'), (now() - 1, 1, 'data')" ) # Wait for part move to S3. max_attempts = 10 for attempt in range(max_attempts + 1): disk = node.query( "SELECT disk_name FROM system.parts WHERE table='s3_failover_test' LIMIT 1" ) if disk != "s3\n": if attempt == max_attempts: assert disk == "s3\n", ( "Expected move to S3 while part still on disk " + disk ) else: time.sleep(1) else: break # Ensure part_log is created. node.query("SYSTEM FLUSH LOGS") # There should be 2 attempts to move part. assert ( node.query( """ SELECT count(*) FROM system.part_log WHERE event_type='MovePart' AND table='s3_failover_test' """ ) == "2\n" ) # First attempt should be failed with expected error. exception = node.query( """ SELECT exception FROM system.part_log WHERE event_type='MovePart' AND table='s3_failover_test' AND notEmpty(exception) ORDER BY event_time LIMIT 1 """ ) assert exception.find("Expected Error") != -1, exception # Ensure data is not corrupted. assert node.query("CHECK TABLE s3_failover_test") == "1\n" assert ( node.query("SELECT id,data FROM s3_failover_test FORMAT Values") == "(0,'data'),(1,'data')" ) # Check that throttled request retries and does not cause an error on disk with default `retry_attempts` (>0) def test_throttle_retry(cluster): node = cluster.instances["node"] node.query( """ CREATE TABLE s3_throttle_retry_test ( id Int64 ) ENGINE=MergeTree() ORDER BY id SETTINGS storage_policy='s3_retryable' """ ) data = "(42)" node.query("INSERT INTO s3_throttle_retry_test VALUES {}".format(data)) throttle_request(cluster, 1) assert ( node.query( """ SELECT * FROM s3_throttle_retry_test """ ) == "42\n" )