mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-18 20:32:43 +00:00
b75963d370
This PR formats all the `*.py` files found under the `tests/integration` folder. It also reorders the imports and cleans up a bunch of unused imports. The formatting also takes care of other things like wrapping lines and fixing spaces and indents such that the tests look more readable.
117 lines
7.7 KiB
Python
117 lines
7.7 KiB
Python
import pytest
|
|
from helpers.cluster import ClickHouseCluster
|
|
from helpers.hdfs_api import HDFSApi
|
|
|
|
cluster = ClickHouseCluster(__file__)
|
|
node1 = cluster.add_instance('node1', main_configs=['configs/config_with_hosts.xml'])
|
|
node2 = cluster.add_instance('node2', main_configs=['configs/config_with_only_primary_hosts.xml'])
|
|
node3 = cluster.add_instance('node3', main_configs=['configs/config_with_only_regexp_hosts.xml'])
|
|
node4 = cluster.add_instance('node4', main_configs=['configs/config_without_allowed_hosts.xml'])
|
|
node6 = cluster.add_instance('node6', main_configs=['configs/config_for_remote.xml'])
|
|
node7 = cluster.add_instance('node7', main_configs=['configs/config_for_redirect.xml'], with_hdfs=True)
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
def start_cluster():
|
|
try:
|
|
cluster.start()
|
|
yield cluster
|
|
finally:
|
|
cluster.shutdown()
|
|
|
|
|
|
def test_config_with_hosts(start_cluster):
|
|
assert node1.query("CREATE TABLE table_test_1_1 (word String) Engine=URL('http://host:80', HDFS)") == ""
|
|
assert node1.query("CREATE TABLE table_test_1_2 (word String) Engine=URL('https://yandex.ru', CSV)") == ""
|
|
assert "not allowed" in node1.query_and_get_error(
|
|
"CREATE TABLE table_test_1_4 (word String) Engine=URL('https://host:123', S3)")
|
|
assert "not allowed" in node1.query_and_get_error(
|
|
"CREATE TABLE table_test_1_4 (word String) Engine=URL('https://yandex2.ru', CSV)")
|
|
|
|
|
|
def test_config_with_only_primary_hosts(start_cluster):
|
|
assert node2.query("CREATE TABLE table_test_2_1 (word String) Engine=URL('https://host:80', CSV)") == ""
|
|
assert node2.query("CREATE TABLE table_test_2_2 (word String) Engine=URL('https://host:123', S3)") == ""
|
|
assert node2.query("CREATE TABLE table_test_2_3 (word String) Engine=URL('https://yandex.ru', CSV)") == ""
|
|
assert node2.query("CREATE TABLE table_test_2_4 (word String) Engine=URL('https://yandex.ru:87', HDFS)") == ""
|
|
assert "not allowed" in node2.query_and_get_error(
|
|
"CREATE TABLE table_test_2_5 (word String) Engine=URL('https://host', HDFS)")
|
|
assert "not allowed" in node2.query_and_get_error(
|
|
"CREATE TABLE table_test_2_5 (word String) Engine=URL('https://host:234', CSV)")
|
|
assert "not allowed" in node2.query_and_get_error(
|
|
"CREATE TABLE table_test_2_6 (word String) Engine=URL('https://yandex2.ru', S3)")
|
|
|
|
|
|
def test_config_with_only_regexp_hosts(start_cluster):
|
|
assert node3.query("CREATE TABLE table_test_3_1 (word String) Engine=URL('https://host:80', HDFS)") == ""
|
|
assert node3.query("CREATE TABLE table_test_3_2 (word String) Engine=URL('https://yandex.ru', CSV)") == ""
|
|
assert "not allowed" in node3.query_and_get_error(
|
|
"CREATE TABLE table_test_3_3 (word String) Engine=URL('https://host', CSV)")
|
|
assert "not allowed" in node3.query_and_get_error(
|
|
"CREATE TABLE table_test_3_4 (word String) Engine=URL('https://yandex2.ru', S3)")
|
|
|
|
|
|
def test_config_without_allowed_hosts(start_cluster):
|
|
assert node4.query("CREATE TABLE table_test_4_1 (word String) Engine=URL('https://host:80', CSV)") == ""
|
|
assert node4.query("CREATE TABLE table_test_4_2 (word String) Engine=URL('https://host', HDFS)") == ""
|
|
assert node4.query("CREATE TABLE table_test_4_3 (word String) Engine=URL('https://yandex.ru', CSV)") == ""
|
|
assert node4.query("CREATE TABLE table_test_4_4 (word String) Engine=URL('ftp://something.com', S3)") == ""
|
|
|
|
|
|
def test_table_function_remote(start_cluster):
|
|
assert "not allowed in config.xml" not in node6.query_and_get_error(
|
|
"SELECT * FROM remoteSecure('example01-01-{1|2}', system, events)",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed in config.xml" not in node6.query_and_get_error(
|
|
"SELECT * FROM remoteSecure('example01-01-1,example01-02-1', system, events)",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed in config.xml" not in node6.query_and_get_error(
|
|
"SELECT * FROM remote('example01-0{1,2}-1', system, events",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed in config.xml" not in node6.query_and_get_error(
|
|
"SELECT * FROM remote('example01-0{1,2}-{1|2}', system, events)",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed in config.xml" not in node6.query_and_get_error(
|
|
"SELECT * FROM remoteSecure('example01-{01..02}-{1|2}', system, events)",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed" in node6.query_and_get_error(
|
|
"SELECT * FROM remoteSecure('example01-01-1,example01-03-1', system, events)",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed" in node6.query_and_get_error("SELECT * FROM remote('example01-01-{1|3}', system, events)",
|
|
settings={"connections_with_failover_max_tries": 1,
|
|
"connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000,
|
|
"connect_timeout": 1, "send_timeout": 1})
|
|
assert "not allowed" in node6.query_and_get_error(
|
|
"SELECT * FROM remoteSecure('example01-0{1,3}-1', system, metrics)",
|
|
settings={"connections_with_failover_max_tries": 1, "connect_timeout_with_failover_ms": 1000,
|
|
"connect_timeout_with_failover_secure_ms": 1000, "connect_timeout": 1, "send_timeout": 1})
|
|
assert node6.query("SELECT * FROM remote('localhost', system, events)") != ""
|
|
assert node6.query("SELECT * FROM remoteSecure('localhost', system, metrics)") != ""
|
|
assert "URL \"localhost:800\" is not allowed in config.xml" in node6.query_and_get_error(
|
|
"SELECT * FROM remoteSecure('localhost:800', system, events)")
|
|
assert "URL \"localhost:800\" is not allowed in config.xml" in node6.query_and_get_error(
|
|
"SELECT * FROM remote('localhost:800', system, metrics)")
|
|
|
|
|
|
def test_redirect(start_cluster):
|
|
hdfs_api = HDFSApi("root")
|
|
hdfs_api.write_data("/simple_storage", "1\t\n")
|
|
assert hdfs_api.read_data("/simple_storage") == "1\t\n"
|
|
node7.query(
|
|
"CREATE TABLE table_test_7_1 (word String) ENGINE=URL('http://hdfs1:50070/webhdfs/v1/simple_storage?op=OPEN&namenoderpcaddress=hdfs1:9000&offset=0', CSV)")
|
|
assert "not allowed" in node7.query_and_get_error("SET max_http_get_redirects=1; SELECT * from table_test_7_1")
|
|
|
|
|
|
def test_HDFS(start_cluster):
|
|
assert "not allowed" in node7.query_and_get_error(
|
|
"CREATE TABLE table_test_7_2 (word String) ENGINE=HDFS('http://hdfs1:50075/webhdfs/v1/simple_storage?op=OPEN&namenoderpcaddress=hdfs1:9000&offset=0', 'CSV')")
|
|
assert "not allowed" in node7.query_and_get_error(
|
|
"SELECT * FROM hdfs('http://hdfs1:50075/webhdfs/v1/simple_storage?op=OPEN&namenoderpcaddress=hdfs1:9000&offset=0', 'TSV', 'word String')")
|