mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-10-19 15:01:03 +00:00
92 lines
3.1 KiB
Python
92 lines
3.1 KiB
Python
|
import time
|
||
|
import pytest
|
||
|
|
||
|
from helpers.cluster import ClickHouseCluster
|
||
|
from multiprocessing.dummy import Pool
|
||
|
|
||
|
from helpers.test_tools import assert_eq_with_retry
|
||
|
|
||
|
def _fill_nodes(nodes, shard, connections_count):
|
||
|
for node in nodes:
|
||
|
node.query(
|
||
|
'''
|
||
|
CREATE DATABASE test;
|
||
|
|
||
|
CREATE TABLE test_table(date Date, id UInt32, dummy UInt32)
|
||
|
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test{shard}/replicated', '{replica}')
|
||
|
PARTITION BY date
|
||
|
ORDER BY id
|
||
|
SETTINGS
|
||
|
replicated_max_parallel_fetches_for_host={connections},
|
||
|
index_granularity=8192;
|
||
|
'''.format(shard=shard, replica=node.name, connections=connections_count))
|
||
|
|
||
|
cluster = ClickHouseCluster(__file__)
|
||
|
node1 = cluster.add_instance('node1', config_dir="configs", main_configs=['configs/remote_servers.xml'], with_zookeeper=True)
|
||
|
node2 = cluster.add_instance('node2', config_dir="configs", main_configs=['configs/remote_servers.xml'], with_zookeeper=True)
|
||
|
|
||
|
@pytest.fixture(scope="module")
|
||
|
def start_small_cluster():
|
||
|
try:
|
||
|
cluster.start()
|
||
|
|
||
|
_fill_nodes([node1, node2], 1, 1)
|
||
|
|
||
|
yield cluster
|
||
|
|
||
|
finally:
|
||
|
cluster.shutdown()
|
||
|
|
||
|
def test_single_endpoint_connections_count(start_small_cluster):
|
||
|
|
||
|
def task(count):
|
||
|
print("Inserting ten times from {}".format(count))
|
||
|
for i in xrange(count, count + 10):
|
||
|
node1.query("insert into test_table values ('2017-06-16', {}, 0)".format(i))
|
||
|
|
||
|
p = Pool(10)
|
||
|
p.map(task, xrange(0, 100, 10))
|
||
|
|
||
|
assert_eq_with_retry(node1, "select count() from test_table", "100")
|
||
|
assert_eq_with_retry(node2, "select count() from test_table", "100")
|
||
|
|
||
|
assert node2.query("SELECT value FROM system.events where event='CreatedHTTPConnections'") == '1\n'
|
||
|
|
||
|
node3 = cluster.add_instance('node3', config_dir="configs", main_configs=['configs/remote_servers.xml'], with_zookeeper=True)
|
||
|
node4 = cluster.add_instance('node4', config_dir="configs", main_configs=['configs/remote_servers.xml'], with_zookeeper=True)
|
||
|
node5 = cluster.add_instance('node5', config_dir="configs", main_configs=['configs/remote_servers.xml'], with_zookeeper=True)
|
||
|
|
||
|
@pytest.fixture(scope="module")
|
||
|
def start_big_cluster():
|
||
|
try:
|
||
|
cluster.start()
|
||
|
|
||
|
_fill_nodes([node3, node4, node5], 2, 2)
|
||
|
|
||
|
yield cluster
|
||
|
|
||
|
finally:
|
||
|
cluster.shutdown()
|
||
|
|
||
|
def test_multiple_endpoint_connections_count(start_big_cluster):
|
||
|
|
||
|
def task(count):
|
||
|
print("Inserting ten times from {}".format(count))
|
||
|
if (count / 10) % 2 == 1:
|
||
|
node = node3
|
||
|
else:
|
||
|
node = node4
|
||
|
|
||
|
for i in xrange(count, count + 10):
|
||
|
node.query("insert into test_table values ('2017-06-16', {}, 0)".format(i))
|
||
|
|
||
|
p = Pool(10)
|
||
|
p.map(task, xrange(0, 100, 10))
|
||
|
|
||
|
assert_eq_with_retry(node3, "select count() from test_table", "100")
|
||
|
assert_eq_with_retry(node4, "select count() from test_table", "100")
|
||
|
assert_eq_with_retry(node5, "select count() from test_table", "100")
|
||
|
|
||
|
# two per each host
|
||
|
assert node5.query("SELECT value FROM system.events where event='CreatedHTTPConnections'") == '4\n'
|