ClickHouse/tests/integration/test_storage_rabbitmq/test.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

3485 lines
112 KiB
Python
Raw Normal View History

2023-01-11 11:17:44 +00:00
import pytest
import json
import os.path as p
import random
import subprocess
import threading
2021-04-30 09:18:12 +00:00
import logging
import time
from random import randrange
import math
import pika
from google.protobuf.internal.encoder import _VarintBytes
from helpers.client import QueryRuntimeException
from helpers.cluster import ClickHouseCluster, check_rabbitmq_is_available
from helpers.test_tools import TSV
2020-10-02 16:54:07 +00:00
from . import rabbitmq_pb2
cluster = ClickHouseCluster(__file__)
instance = cluster.add_instance(
"instance",
main_configs=[
"configs/rabbitmq.xml",
"configs/macros.xml",
"configs/named_collection.xml",
],
user_configs=["configs/users.xml"],
with_rabbitmq=True,
2022-05-31 12:38:11 +00:00
stay_alive=True,
)
2022-06-24 00:07:16 +00:00
instance2 = cluster.add_instance(
"instance2",
user_configs=["configs/users.xml"],
with_rabbitmq=True,
)
# Helpers
def rabbitmq_check_result(result, check=False, ref_file="test_rabbitmq_json.reference"):
fpath = p.join(p.dirname(__file__), ref_file)
with open(fpath) as reference:
if check:
assert TSV(result) == TSV(reference)
else:
return TSV(result) == TSV(reference)
2021-12-28 19:16:16 +00:00
def wait_rabbitmq_to_start(rabbitmq_docker_id, timeout=180):
2022-12-20 18:32:59 +00:00
logging.getLogger("pika").propagate = False
2021-12-28 19:16:16 +00:00
start = time.time()
while time.time() - start < timeout:
try:
if check_rabbitmq_is_available(rabbitmq_docker_id):
2021-12-28 19:16:16 +00:00
logging.debug("RabbitMQ is available")
return
time.sleep(0.5)
except Exception as ex:
logging.debug("Can't connect to RabbitMQ " + str(ex))
time.sleep(0.5)
2021-04-30 09:18:12 +00:00
def kill_rabbitmq(rabbitmq_id):
p = subprocess.Popen(("docker", "stop", rabbitmq_id), stdout=subprocess.PIPE)
2020-08-08 16:45:52 +00:00
p.communicate()
return p.returncode == 0
2021-04-30 09:18:12 +00:00
def revive_rabbitmq(rabbitmq_id):
p = subprocess.Popen(("docker", "start", rabbitmq_id), stdout=subprocess.PIPE)
2020-08-08 16:45:52 +00:00
p.communicate()
2021-12-28 19:16:16 +00:00
wait_rabbitmq_to_start(rabbitmq_id)
2020-08-08 16:45:52 +00:00
# Fixtures
@pytest.fixture(scope="module")
def rabbitmq_cluster():
try:
cluster.start()
2021-04-30 09:18:12 +00:00
logging.debug("rabbitmq_id is {}".format(instance.cluster.rabbitmq_docker_id))
instance.query("CREATE DATABASE test")
yield cluster
finally:
cluster.shutdown()
@pytest.fixture(autouse=True)
def rabbitmq_setup_teardown():
print("RabbitMQ is available - running test")
yield # run test
instance.query("DROP DATABASE test NO DELAY")
instance.query("CREATE DATABASE test")
# Tests
2020-08-31 09:12:36 +00:00
def test_rabbitmq_select(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
2021-02-16 14:16:15 +00:00
SETTINGS rabbitmq_host_port = '{}:5672',
2020-08-31 09:12:36 +00:00
rabbitmq_exchange_name = 'select',
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
""".format(
rabbitmq_cluster.rabbitmq_host
)
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for i in range(50):
messages.append(json.dumps({"key": i, "value": i}))
2020-05-20 09:42:56 +00:00
for message in messages:
channel.basic_publish(exchange="select", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
connection.close()
2020-09-07 10:21:29 +00:00
# The order of messages in select * from test.rabbitmq is not guaranteed, so sleep to collect everything in one select
time.sleep(1)
2020-06-08 01:11:48 +00:00
result = ""
2020-06-08 01:11:48 +00:00
while True:
result += instance.query(
"SELECT * FROM test.rabbitmq ORDER BY key", ignore_error=True
)
2020-06-08 01:11:48 +00:00
if rabbitmq_check_result(result):
break
2020-05-20 09:42:56 +00:00
rabbitmq_check_result(result, True)
def test_rabbitmq_select_empty(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
2021-02-16 14:16:15 +00:00
SETTINGS rabbitmq_host_port = '{}:5672',
2020-08-31 09:12:36 +00:00
rabbitmq_exchange_name = 'empty',
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'TSV',
rabbitmq_flush_interval_ms=1000,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = '\\n';
""".format(
rabbitmq_cluster.rabbitmq_host
)
)
2020-05-20 09:42:56 +00:00
assert int(instance.query("SELECT count() FROM test.rabbitmq")) == 0
2020-05-20 09:42:56 +00:00
def test_rabbitmq_json_without_delimiter(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
2021-02-16 14:16:15 +00:00
SETTINGS rabbitmq_host_port = '{}:5672',
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
rabbitmq_exchange_name = 'json',
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow'
""".format(
rabbitmq_cluster.rabbitmq_host
)
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = ""
2020-05-20 09:42:56 +00:00
for i in range(25):
messages += json.dumps({"key": i, "value": i}) + "\n"
2020-05-20 09:42:56 +00:00
all_messages = [messages]
for message in all_messages:
channel.basic_publish(exchange="json", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
messages = ""
2020-05-20 09:42:56 +00:00
for i in range(25, 50):
messages += json.dumps({"key": i, "value": i}) + "\n"
2020-05-20 09:42:56 +00:00
all_messages = [messages]
for message in all_messages:
channel.basic_publish(exchange="json", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
2020-09-07 10:21:29 +00:00
connection.close()
time.sleep(1)
result = ""
2020-05-20 09:42:56 +00:00
while True:
result += instance.query(
"SELECT * FROM test.rabbitmq ORDER BY key", ignore_error=True
)
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result(result):
break
rabbitmq_check_result(result, True)
def test_rabbitmq_csv_with_delimiter(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'csv',
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'CSV',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = '\\n';
"""
)
2020-05-20 09:42:56 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for i in range(50):
messages.append("{i}, {i}".format(i=i))
2020-05-20 09:42:56 +00:00
for message in messages:
channel.basic_publish(exchange="csv", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
2020-09-07 10:21:29 +00:00
connection.close()
time.sleep(1)
result = ""
2020-05-20 09:42:56 +00:00
while True:
result += instance.query(
"SELECT * FROM test.rabbitmq ORDER BY key", ignore_error=True
)
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result(result):
break
rabbitmq_check_result(result, True)
def test_rabbitmq_tsv_with_delimiter(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'tsv',
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'TSV',
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2021-03-30 18:48:33 +00:00
rabbitmq_queue_base = 'tsv',
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = '\\n';
2021-03-31 07:37:34 +00:00
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
2020-05-20 09:42:56 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for i in range(50):
messages.append("{i}\t{i}".format(i=i))
2020-05-20 09:42:56 +00:00
for message in messages:
channel.basic_publish(exchange="tsv", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
connection.close()
result = ""
while True:
result = instance.query("SELECT * FROM test.view ORDER BY key")
if rabbitmq_check_result(result):
break
rabbitmq_check_result(result, True)
def test_rabbitmq_macros(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = '{rabbitmq_host}:{rabbitmq_port}',
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
rabbitmq_exchange_name = '{rabbitmq_exchange_name}',
rabbitmq_format = '{rabbitmq_format}'
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
message = ""
for i in range(50):
message += json.dumps({"key": i, "value": i}) + "\n"
channel.basic_publish(exchange="macro", routing_key="", body=message)
connection.close()
time.sleep(1)
result = ""
while True:
result += instance.query(
"SELECT * FROM test.rabbitmq ORDER BY key", ignore_error=True
)
if rabbitmq_check_result(result):
break
rabbitmq_check_result(result, True)
2020-05-20 09:42:56 +00:00
def test_rabbitmq_materialized_view(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'mv',
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
CREATE TABLE test.view2 (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer2 TO test.view2 AS
SELECT * FROM test.rabbitmq group by (key, value);
"""
)
2020-05-20 09:42:56 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
2023-02-02 11:15:29 +00:00
instance.wait_for_log_line("Started streaming to 2 attached views")
2020-05-20 09:42:56 +00:00
messages = []
for i in range(50):
2023-02-02 11:15:29 +00:00
message = json.dumps({"key": i, "value": i})
channel.basic_publish(exchange="mv", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
time_limit_sec = 60
deadline = time.monotonic() + time_limit_sec
while time.monotonic() < deadline:
result = instance.query("SELECT * FROM test.view ORDER BY key")
if rabbitmq_check_result(result):
2020-09-07 11:08:53 +00:00
break
2020-05-20 09:42:56 +00:00
rabbitmq_check_result(result, True)
deadline = time.monotonic() + time_limit_sec
while time.monotonic() < deadline:
result = instance.query("SELECT * FROM test.view2 ORDER BY key")
2023-02-02 11:15:29 +00:00
print(f"Result: {result}")
if rabbitmq_check_result(result):
break
2023-02-02 11:15:29 +00:00
time.sleep(1)
rabbitmq_check_result(result, True)
connection.close()
2020-05-20 09:42:56 +00:00
def test_rabbitmq_materialized_view_with_subquery(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'mvsq',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM (SELECT * FROM test.rabbitmq);
"""
)
2020-05-20 09:42:56 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for i in range(50):
messages.append(json.dumps({"key": i, "value": i}))
2020-05-20 09:42:56 +00:00
for message in messages:
channel.basic_publish(exchange="mvsq", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
while True:
result = instance.query("SELECT * FROM test.view ORDER BY key")
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result(result):
break
2020-09-07 11:08:53 +00:00
connection.close()
2020-05-20 09:42:56 +00:00
rabbitmq_check_result(result, True)
def test_rabbitmq_many_materialized_views(rabbitmq_cluster):
instance.query(
"""
2020-05-20 09:42:56 +00:00
DROP TABLE IF EXISTS test.view1;
DROP TABLE IF EXISTS test.view2;
DROP TABLE IF EXISTS test.consumer1;
DROP TABLE IF EXISTS test.consumer2;
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'mmv',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.view1 (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE TABLE test.view2 (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer1 TO test.view1 AS
SELECT * FROM test.rabbitmq;
CREATE MATERIALIZED VIEW test.consumer2 TO test.view2 AS
SELECT * FROM test.rabbitmq;
"""
)
2020-05-20 09:42:56 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
2023-02-02 11:15:29 +00:00
instance.wait_for_log_line("Started streaming to 2 attached views")
2020-05-20 09:42:56 +00:00
messages = []
for i in range(50):
messages.append(json.dumps({"key": i, "value": i}))
2020-05-20 09:42:56 +00:00
for message in messages:
channel.basic_publish(exchange="mmv", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
while True:
result1 = instance.query("SELECT * FROM test.view1 ORDER BY key")
result2 = instance.query("SELECT * FROM test.view2 ORDER BY key")
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result(result1) and rabbitmq_check_result(result2):
break
instance.query(
"""
2020-05-20 09:42:56 +00:00
DROP TABLE test.consumer1;
DROP TABLE test.consumer2;
DROP TABLE test.view1;
DROP TABLE test.view2;
"""
)
2020-05-20 09:42:56 +00:00
connection.close()
2020-05-20 09:42:56 +00:00
rabbitmq_check_result(result1, True)
rabbitmq_check_result(result2, True)
2020-09-11 16:16:24 +00:00
@pytest.mark.skip(reason="clichouse_path with rabbitmq.proto fails to be exported")
2020-09-07 10:21:29 +00:00
def test_rabbitmq_protobuf(rabbitmq_cluster):
instance.query(
"""
2020-09-07 10:21:29 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value String)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'pb',
rabbitmq_format = 'Protobuf',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-09-07 10:21:29 +00:00
rabbitmq_schema = 'rabbitmq.proto:KeyValueProto';
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
2020-09-07 10:21:29 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-09-07 10:21:29 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
data = ""
2020-09-07 10:21:29 +00:00
for i in range(0, 20):
msg = rabbitmq_pb2.KeyValueProto()
msg.key = i
msg.value = str(i)
serialized_msg = msg.SerializeToString()
data = data + _VarintBytes(len(serialized_msg)) + serialized_msg
channel.basic_publish(exchange="pb", routing_key="", body=data)
data = ""
2020-09-07 10:21:29 +00:00
for i in range(20, 21):
msg = rabbitmq_pb2.KeyValueProto()
msg.key = i
msg.value = str(i)
serialized_msg = msg.SerializeToString()
data = data + _VarintBytes(len(serialized_msg)) + serialized_msg
channel.basic_publish(exchange="pb", routing_key="", body=data)
data = ""
2020-09-07 10:21:29 +00:00
for i in range(21, 50):
msg = rabbitmq_pb2.KeyValueProto()
msg.key = i
msg.value = str(i)
serialized_msg = msg.SerializeToString()
data = data + _VarintBytes(len(serialized_msg)) + serialized_msg
channel.basic_publish(exchange="pb", routing_key="", body=data)
2020-09-07 10:21:29 +00:00
connection.close()
result = ""
2020-09-07 10:21:29 +00:00
while True:
result = instance.query("SELECT * FROM test.view ORDER BY key")
2020-09-07 10:21:29 +00:00
if rabbitmq_check_result(result):
break
rabbitmq_check_result(result, True)
2020-05-20 09:42:56 +00:00
def test_rabbitmq_big_message(rabbitmq_cluster):
# Create batchs of messages of size ~100Kb
rabbitmq_messages = 1000
batch_messages = 1000
messages = [
json.dumps({"key": i, "value": "x" * 100}) * batch_messages
for i in range(rabbitmq_messages)
]
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value String)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'big',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow';
CREATE TABLE test.view (key UInt64, value String)
ENGINE = MergeTree
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
2020-05-20 09:42:56 +00:00
for message in messages:
channel.basic_publish(exchange="big", routing_key="", body=message)
2020-05-20 09:42:56 +00:00
while True:
result = instance.query("SELECT count() FROM test.view")
2020-05-20 09:42:56 +00:00
if int(result) == batch_messages * rabbitmq_messages:
break
connection.close()
assert (
int(result) == rabbitmq_messages * batch_messages
), "ClickHouse lost some messages: {}".format(result)
2020-05-20 09:42:56 +00:00
def test_rabbitmq_sharding_between_queues_publish(rabbitmq_cluster):
NUM_CONSUMERS = 10
2020-10-27 07:14:38 +00:00
NUM_QUEUES = 10
2022-12-20 18:32:59 +00:00
logging.getLogger("pika").propagate = False
2020-05-20 09:42:56 +00:00
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'test_sharding',
2022-12-23 10:47:51 +00:00
rabbitmq_num_queues = 5,
2020-05-20 09:42:56 +00:00
rabbitmq_num_consumers = 10,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100,
2023-01-19 17:06:43 +00:00
rabbitmq_flush_interval_ms=500,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
2020-08-15 06:50:53 +00:00
CREATE TABLE test.view (key UInt64, value UInt64, channel_id String)
2020-05-20 09:42:56 +00:00
ENGINE = MergeTree
ORDER BY key
SETTINGS old_parts_lifetime=5, cleanup_delay_period=2, cleanup_delay_period_random_add=3;
2020-05-20 09:42:56 +00:00
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
2020-08-15 06:50:53 +00:00
SELECT *, _channel_id AS channel_id FROM test.rabbitmq;
"""
)
2020-05-20 09:42:56 +00:00
i = [0]
messages_num = 10000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
def produce():
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-05-20 09:42:56 +00:00
i[0] += 1
2020-07-24 12:33:07 +00:00
current = 0
2020-05-20 09:42:56 +00:00
for message in messages:
2020-07-24 12:33:07 +00:00
current += 1
mes_id = str(current)
channel.basic_publish(
exchange="test_sharding",
routing_key="",
properties=pika.BasicProperties(message_id=mes_id),
body=message,
)
2020-05-20 09:42:56 +00:00
connection.close()
threads = []
2022-12-23 10:47:51 +00:00
threads_num = 10
2020-05-20 09:42:56 +00:00
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
result1 = ""
2020-05-20 09:42:56 +00:00
while True:
result1 = instance.query("SELECT count() FROM test.view")
2020-05-20 09:42:56 +00:00
time.sleep(1)
2022-12-20 18:32:59 +00:00
expected = messages_num * threads_num
if int(result1) == expected:
2020-05-20 09:42:56 +00:00
break
2022-12-20 18:32:59 +00:00
print(f"Result {result1} / {expected}")
2020-05-20 09:42:56 +00:00
2020-08-15 06:50:53 +00:00
result2 = instance.query("SELECT count(DISTINCT channel_id) FROM test.view")
2020-07-24 12:33:07 +00:00
2020-05-20 09:42:56 +00:00
for thread in threads:
thread.join()
assert (
int(result1) == messages_num * threads_num
), "ClickHouse lost some messages: {}".format(result)
2023-01-27 16:22:24 +00:00
assert int(result2) == 10
2020-05-20 09:42:56 +00:00
2020-09-01 14:11:34 +00:00
def test_rabbitmq_mv_combo(rabbitmq_cluster):
2020-09-07 11:08:53 +00:00
NUM_MV = 5
2020-05-20 09:42:56 +00:00
NUM_CONSUMERS = 4
2022-12-20 18:32:59 +00:00
logging.getLogger("pika").propagate = False
2020-05-20 09:42:56 +00:00
instance.query(
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'combo',
2020-09-01 14:11:34 +00:00
rabbitmq_queue_base = 'combo',
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100,
rabbitmq_flush_interval_ms=1000,
rabbitmq_num_consumers = 2,
2020-10-27 07:14:38 +00:00
rabbitmq_num_queues = 5,
2020-05-20 09:42:56 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
"""
)
2020-05-20 09:42:56 +00:00
for mv_id in range(NUM_MV):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE IF EXISTS test.combo_{0};
DROP TABLE IF EXISTS test.combo_{0}_mv;
CREATE TABLE test.combo_{0} (key UInt64, value UInt64)
2020-05-20 09:42:56 +00:00
ENGINE = MergeTree()
ORDER BY key;
2020-08-26 08:54:29 +00:00
CREATE MATERIALIZED VIEW test.combo_{0}_mv TO test.combo_{0} AS
2020-05-20 09:42:56 +00:00
SELECT * FROM test.rabbitmq;
""".format(
mv_id
)
)
2020-05-20 09:42:56 +00:00
time.sleep(2)
i = [0]
messages_num = 10000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-05-20 09:42:56 +00:00
def produce():
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-05-20 09:42:56 +00:00
i[0] += 1
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="combo",
routing_key="",
properties=pika.BasicProperties(message_id=str(msg_id)),
body=messages[msg_id],
)
2020-05-20 09:42:56 +00:00
connection.close()
threads = []
threads_num = 20
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
while True:
result = 0
2020-08-26 08:54:29 +00:00
for mv_id in range(NUM_MV):
result += int(
instance.query("SELECT count() FROM test.combo_{0}".format(mv_id))
)
2022-12-20 18:32:59 +00:00
expected = messages_num * threads_num * NUM_MV
if int(result) == expected:
2020-05-20 09:42:56 +00:00
break
2022-12-20 18:32:59 +00:00
print(f"Result: {result} / {expected}")
2020-05-20 09:42:56 +00:00
time.sleep(1)
for thread in threads:
thread.join()
for mv_id in range(NUM_MV):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.combo_{0}_mv;
2020-09-07 10:21:29 +00:00
DROP TABLE test.combo_{0};
""".format(
mv_id
)
)
2020-05-20 09:42:56 +00:00
assert (
int(result) == messages_num * threads_num * NUM_MV
), "ClickHouse lost some messages: {}".format(result)
2020-06-01 16:19:59 +00:00
def test_rabbitmq_insert(rabbitmq_cluster):
instance.query(
"""
2020-06-01 16:19:59 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
2020-06-13 18:44:17 +00:00
rabbitmq_exchange_name = 'insert',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
rabbitmq_exchange_type = 'direct',
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = 'insert1',
2020-06-01 16:19:59 +00:00
rabbitmq_format = 'TSV',
rabbitmq_row_delimiter = '\\n';
"""
)
2020-06-01 16:19:59 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-01 16:19:59 +00:00
consumer_connection = pika.BlockingConnection(parameters)
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
2020-06-01 16:19:59 +00:00
queue_name = result.method.queue
consumer.queue_bind(exchange="insert", queue=queue_name, routing_key="insert1")
2020-06-01 16:19:59 +00:00
values = []
for i in range(50):
values.append("({i}, {i})".format(i=i))
values = ",".join(values)
2020-06-01 16:19:59 +00:00
while True:
try:
instance.query("INSERT INTO test.rabbitmq VALUES {}".format(values))
break
except QueryRuntimeException as e:
if "Local: Timed out." in str(e):
2020-06-01 16:19:59 +00:00
continue
else:
raise
insert_messages = []
2020-06-01 16:19:59 +00:00
def onReceived(channel, method, properties, body):
i = 0
insert_messages.append(body.decode())
if len(insert_messages) == 50:
2020-06-01 16:19:59 +00:00
channel.stop_consuming()
2020-07-02 14:38:09 +00:00
consumer.basic_consume(onReceived, queue_name)
2020-06-01 16:19:59 +00:00
consumer.start_consuming()
consumer_connection.close()
result = "\n".join(insert_messages)
2020-06-01 16:19:59 +00:00
rabbitmq_check_result(result, True)
def test_rabbitmq_insert_headers_exchange(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
2020-06-01 16:19:59 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'insert_headers',
rabbitmq_exchange_type = 'headers',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
rabbitmq_routing_key_list = 'test=insert,topic=headers',
2020-06-01 16:19:59 +00:00
rabbitmq_format = 'TSV',
rabbitmq_row_delimiter = '\\n';
"""
)
2020-06-01 16:19:59 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
consumer_connection = pika.BlockingConnection(parameters)
2020-06-01 16:19:59 +00:00
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
queue_name = result.method.queue
consumer.queue_bind(
exchange="insert_headers",
queue=queue_name,
routing_key="",
arguments={"x-match": "all", "test": "insert", "topic": "headers"},
)
2020-06-01 16:19:59 +00:00
values = []
for i in range(50):
values.append("({i}, {i})".format(i=i))
values = ",".join(values)
2020-06-01 16:19:59 +00:00
while True:
try:
instance.query("INSERT INTO test.rabbitmq VALUES {}".format(values))
2020-06-01 16:19:59 +00:00
break
except QueryRuntimeException as e:
if "Local: Timed out." in str(e):
continue
else:
raise
2020-06-01 16:19:59 +00:00
insert_messages = []
def onReceived(channel, method, properties, body):
i = 0
insert_messages.append(body.decode())
if len(insert_messages) == 50:
channel.stop_consuming()
2020-06-01 16:19:59 +00:00
consumer.basic_consume(onReceived, queue_name)
consumer.start_consuming()
consumer_connection.close()
2020-06-01 16:19:59 +00:00
result = "\n".join(insert_messages)
rabbitmq_check_result(result, True)
2020-06-01 16:19:59 +00:00
def test_rabbitmq_many_inserts(rabbitmq_cluster):
instance.query(
"""
DROP TABLE IF EXISTS test.rabbitmq_many;
2020-07-23 11:45:01 +00:00
DROP TABLE IF EXISTS test.rabbitmq_consume;
DROP TABLE IF EXISTS test.view_many;
DROP TABLE IF EXISTS test.consumer_many;
CREATE TABLE test.rabbitmq_many (key UInt64, value UInt64)
2020-06-01 16:19:59 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'many_inserts',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'insert2',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-01 16:19:59 +00:00
rabbitmq_format = 'TSV',
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.rabbitmq_consume (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'many_inserts',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'insert2',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
rabbitmq_format = 'TSV',
rabbitmq_row_delimiter = '\\n';
"""
)
2020-06-01 16:19:59 +00:00
2021-06-27 19:05:20 +00:00
messages_num = 10000
values = []
for i in range(messages_num):
values.append("({i}, {i})".format(i=i))
values = ",".join(values)
2020-06-01 16:19:59 +00:00
def insert():
while True:
try:
instance.query(
"INSERT INTO test.rabbitmq_many VALUES {}".format(values)
)
2020-06-01 16:19:59 +00:00
break
except QueryRuntimeException as e:
if "Local: Timed out." in str(e):
2020-06-01 16:19:59 +00:00
continue
else:
raise
threads = []
2021-06-27 19:05:20 +00:00
threads_num = 10
2020-06-01 16:19:59 +00:00
for _ in range(threads_num):
threads.append(threading.Thread(target=insert))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
instance.query(
"""
2021-06-27 19:05:20 +00:00
CREATE TABLE test.view_many (key UInt64, value UInt64)
ENGINE = MergeTree
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer_many TO test.view_many AS
SELECT * FROM test.rabbitmq_consume;
"""
)
2021-06-27 19:05:20 +00:00
for thread in threads:
thread.join()
2020-06-01 16:19:59 +00:00
while True:
result = instance.query("SELECT count() FROM test.view_many")
2021-09-16 10:46:43 +00:00
print(result, messages_num * threads_num)
2020-06-01 16:19:59 +00:00
if int(result) == messages_num * threads_num:
break
2021-06-27 19:05:20 +00:00
time.sleep(1)
2020-06-01 16:19:59 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.rabbitmq_consume;
DROP TABLE test.rabbitmq_many;
DROP TABLE test.consumer_many;
DROP TABLE test.view_many;
"""
)
2020-06-01 16:19:59 +00:00
assert (
int(result) == messages_num * threads_num
), "ClickHouse lost some messages: {}".format(result)
2020-06-01 16:19:59 +00:00
def test_rabbitmq_overloaded_insert(rabbitmq_cluster):
instance.query(
"""
2020-06-01 16:19:59 +00:00
DROP TABLE IF EXISTS test.view_overload;
DROP TABLE IF EXISTS test.consumer_overload;
2020-07-23 11:45:01 +00:00
DROP TABLE IF EXISTS test.rabbitmq_consume;
CREATE TABLE test.rabbitmq_consume (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'over',
2020-09-01 14:11:34 +00:00
rabbitmq_queue_base = 'over',
rabbitmq_exchange_type = 'direct',
2022-12-23 10:47:51 +00:00
rabbitmq_num_consumers = 3,
rabbitmq_flush_interval_ms=1000,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100,
2022-12-23 10:47:51 +00:00
rabbitmq_num_queues = 2,
rabbitmq_routing_key_list = 'over',
rabbitmq_format = 'TSV',
rabbitmq_row_delimiter = '\\n';
2020-06-01 16:19:59 +00:00
CREATE TABLE test.rabbitmq_overload (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'over',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'over',
rabbitmq_flush_interval_ms=1000,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100,
2020-06-01 16:19:59 +00:00
rabbitmq_format = 'TSV',
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.view_overload (key UInt64, value UInt64)
ENGINE = MergeTree
2020-06-25 09:44:39 +00:00
ORDER BY key
SETTINGS old_parts_lifetime=5, cleanup_delay_period=2, cleanup_delay_period_random_add=3;
2020-06-01 16:19:59 +00:00
CREATE MATERIALIZED VIEW test.consumer_overload TO test.view_overload AS
SELECT * FROM test.rabbitmq_consume;
"""
)
2020-06-01 16:19:59 +00:00
2023-02-02 19:16:53 +00:00
instance.wait_for_log_line("Started streaming to 1 attached views")
2020-06-01 16:19:59 +00:00
messages_num = 100000
2020-06-01 16:19:59 +00:00
def insert():
values = []
for i in range(messages_num):
values.append("({i}, {i})".format(i=i))
values = ",".join(values)
2020-06-01 16:19:59 +00:00
while True:
try:
instance.query(
"INSERT INTO test.rabbitmq_overload VALUES {}".format(values)
)
2020-06-01 16:19:59 +00:00
break
except QueryRuntimeException as e:
if "Local: Timed out." in str(e):
2020-06-01 16:19:59 +00:00
continue
else:
raise
threads = []
2023-01-31 19:25:08 +00:00
threads_num = 3
2020-06-01 16:19:59 +00:00
for _ in range(threads_num):
threads.append(threading.Thread(target=insert))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
while True:
result = instance.query("SELECT count() FROM test.view_overload")
2020-06-01 16:19:59 +00:00
time.sleep(1)
2022-12-20 18:32:59 +00:00
expected = messages_num * threads_num
if int(result) == expected:
2020-06-01 16:19:59 +00:00
break
2022-12-20 18:32:59 +00:00
print(f"Result: {result} / {expected}")
2020-06-01 16:19:59 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.consumer_overload;
DROP TABLE test.view_overload;
DROP TABLE test.rabbitmq_consume;
DROP TABLE test.rabbitmq_overload;
"""
)
2020-06-01 16:19:59 +00:00
for thread in threads:
thread.join()
assert (
int(result) == messages_num * threads_num
), "ClickHouse lost some messages: {}".format(result)
2020-06-01 16:19:59 +00:00
2020-06-10 23:01:47 +00:00
def test_rabbitmq_direct_exchange(rabbitmq_cluster):
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.destination;
CREATE TABLE test.destination(key UInt64, value UInt64)
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree()
2020-06-25 09:44:39 +00:00
ORDER BY key
SETTINGS old_parts_lifetime=5, cleanup_delay_period=2, cleanup_delay_period_random_add=3;
"""
)
2020-06-10 23:01:47 +00:00
num_tables = 5
for consumer_id in range(num_tables):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(consumer_id)))
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.direct_exchange_{0};
DROP TABLE IF EXISTS test.direct_exchange_{0}_mv;
CREATE TABLE test.direct_exchange_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_num_consumers = 2,
rabbitmq_num_queues = 2,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = 'direct_exchange_testing',
rabbitmq_exchange_type = 'direct',
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = 'direct_{0}',
2020-06-10 23:01:47 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.direct_exchange_{0}_mv TO test.destination AS
SELECT key, value FROM test.direct_exchange_{0};
""".format(
consumer_id
)
)
2020-06-10 23:01:47 +00:00
i = [0]
messages_num = 1000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-10 23:01:47 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-06-10 23:01:47 +00:00
i[0] += 1
key_num = 0
for num in range(num_tables):
key = "direct_" + str(key_num)
key_num += 1
for message in messages:
mes_id = str(randrange(10))
channel.basic_publish(
exchange="direct_exchange_testing",
routing_key=key,
properties=pika.BasicProperties(message_id=mes_id),
body=message,
)
2020-06-10 23:01:47 +00:00
connection.close()
while True:
result = instance.query("SELECT count() FROM test.destination")
2020-06-10 23:01:47 +00:00
time.sleep(1)
if int(result) == messages_num * num_tables:
break
2020-06-13 18:44:17 +00:00
for consumer_id in range(num_tables):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.direct_exchange_{0}_mv;
2020-09-07 10:21:29 +00:00
DROP TABLE test.direct_exchange_{0};
""".format(
consumer_id
)
)
2020-06-13 18:44:17 +00:00
instance.query(
"""
2020-06-13 18:44:17 +00:00
DROP TABLE IF EXISTS test.destination;
"""
)
2020-06-13 18:44:17 +00:00
assert (
int(result) == messages_num * num_tables
), "ClickHouse lost some messages: {}".format(result)
2020-06-10 23:01:47 +00:00
def test_rabbitmq_fanout_exchange(rabbitmq_cluster):
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.destination;
CREATE TABLE test.destination(key UInt64, value UInt64)
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree()
ORDER BY key;
"""
)
2020-06-10 23:01:47 +00:00
num_tables = 5
for consumer_id in range(num_tables):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(consumer_id)))
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.fanout_exchange_{0};
DROP TABLE IF EXISTS test.fanout_exchange_{0}_mv;
CREATE TABLE test.fanout_exchange_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_num_consumers = 2,
rabbitmq_num_queues = 2,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = 'key_{0}',
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = 'fanout_exchange_testing',
rabbitmq_exchange_type = 'fanout',
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.fanout_exchange_{0}_mv TO test.destination AS
SELECT key, value FROM test.fanout_exchange_{0};
""".format(
consumer_id
)
)
2020-06-10 23:01:47 +00:00
i = [0]
messages_num = 1000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-10 23:01:47 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-06-10 23:01:47 +00:00
i[0] += 1
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="fanout_exchange_testing",
routing_key="",
properties=pika.BasicProperties(message_id=str(msg_id)),
body=messages[msg_id],
)
2020-06-10 23:01:47 +00:00
connection.close()
while True:
result = instance.query("SELECT count() FROM test.destination")
2020-06-10 23:01:47 +00:00
time.sleep(1)
if int(result) == messages_num * num_tables:
break
2020-06-13 18:44:17 +00:00
for consumer_id in range(num_tables):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.fanout_exchange_{0}_mv;
2020-09-07 10:21:29 +00:00
DROP TABLE test.fanout_exchange_{0};
""".format(
consumer_id
)
)
2020-06-13 18:44:17 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.destination;
"""
)
2020-06-13 18:44:17 +00:00
assert (
int(result) == messages_num * num_tables
), "ClickHouse lost some messages: {}".format(result)
2020-06-10 23:01:47 +00:00
def test_rabbitmq_topic_exchange(rabbitmq_cluster):
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.destination;
CREATE TABLE test.destination(key UInt64, value UInt64)
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree()
ORDER BY key;
"""
)
2020-06-10 23:01:47 +00:00
num_tables = 5
for consumer_id in range(num_tables):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(consumer_id)))
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.topic_exchange_{0};
DROP TABLE IF EXISTS test.topic_exchange_{0}_mv;
CREATE TABLE test.topic_exchange_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_num_consumers = 2,
rabbitmq_num_queues = 2,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = 'topic_exchange_testing',
rabbitmq_exchange_type = 'topic',
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = '*.{0}',
2020-06-10 23:01:47 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.topic_exchange_{0}_mv TO test.destination AS
SELECT key, value FROM test.topic_exchange_{0};
""".format(
consumer_id
)
)
2020-06-10 23:01:47 +00:00
for consumer_id in range(num_tables):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(num_tables + consumer_id)))
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.topic_exchange_{0};
DROP TABLE IF EXISTS test.topic_exchange_{0}_mv;
CREATE TABLE test.topic_exchange_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_num_consumers = 2,
rabbitmq_num_queues = 2,
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = 'topic_exchange_testing',
rabbitmq_exchange_type = 'topic',
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = '*.logs',
2020-06-10 23:01:47 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.topic_exchange_{0}_mv TO test.destination AS
SELECT key, value FROM test.topic_exchange_{0};
""".format(
num_tables + consumer_id
)
)
2020-06-10 23:01:47 +00:00
i = [0]
messages_num = 1000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-10 23:01:47 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-06-10 23:01:47 +00:00
i[0] += 1
key_num = 0
for num in range(num_tables):
key = "topic." + str(key_num)
key_num += 1
for message in messages:
channel.basic_publish(
exchange="topic_exchange_testing", routing_key=key, body=message
)
2020-06-10 23:01:47 +00:00
key = "random.logs"
2020-07-24 12:33:07 +00:00
current = 0
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="topic_exchange_testing",
routing_key=key,
properties=pika.BasicProperties(message_id=str(msg_id)),
body=messages[msg_id],
)
2020-06-10 23:01:47 +00:00
connection.close()
while True:
result = instance.query("SELECT count() FROM test.destination")
2020-06-10 23:01:47 +00:00
time.sleep(1)
if int(result) == messages_num * num_tables + messages_num * num_tables:
break
for consumer_id in range(num_tables * 2):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.topic_exchange_{0}_mv;
2020-09-07 10:21:29 +00:00
DROP TABLE test.topic_exchange_{0};
""".format(
consumer_id
)
)
2020-06-13 18:44:17 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.destination;
"""
)
2020-06-13 18:44:17 +00:00
assert (
int(result) == messages_num * num_tables + messages_num * num_tables
), "ClickHouse lost some messages: {}".format(result)
2020-06-10 23:01:47 +00:00
def test_rabbitmq_hash_exchange(rabbitmq_cluster):
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.destination;
2020-08-15 06:50:53 +00:00
CREATE TABLE test.destination(key UInt64, value UInt64, channel_id String)
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree()
ORDER BY key;
"""
)
2020-06-10 23:01:47 +00:00
num_tables = 4
for consumer_id in range(num_tables):
table_name = "rabbitmq_consumer{}".format(consumer_id)
2020-10-02 16:54:07 +00:00
print(("Setting up {}".format(table_name)))
instance.query(
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test.{0};
DROP TABLE IF EXISTS test.{0}_mv;
CREATE TABLE test.{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_num_consumers = 4,
rabbitmq_num_queues = 2,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_type = 'consistent_hash',
rabbitmq_exchange_name = 'hash_exchange_testing',
rabbitmq_format = 'JSONEachRow',
rabbitmq_flush_interval_ms=1000,
2020-06-10 23:01:47 +00:00
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.{0}_mv TO test.destination AS
2020-08-15 06:50:53 +00:00
SELECT key, value, _channel_id AS channel_id FROM test.{0};
""".format(
table_name
)
)
2020-06-10 23:01:47 +00:00
i = [0]
messages_num = 500
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-10 23:01:47 +00:00
def produce():
# init connection here because otherwise python rabbitmq client might fail
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-06-10 23:01:47 +00:00
i[0] += 1
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="hash_exchange_testing",
routing_key=str(msg_id),
properties=pika.BasicProperties(message_id=str(msg_id)),
body=messages[msg_id],
)
2020-06-10 23:01:47 +00:00
connection.close()
threads = []
threads_num = 10
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
result1 = ""
2020-06-10 23:01:47 +00:00
while True:
result1 = instance.query("SELECT count() FROM test.destination")
2020-06-10 23:01:47 +00:00
time.sleep(1)
if int(result1) == messages_num * threads_num:
2020-06-10 23:01:47 +00:00
break
2020-08-15 06:50:53 +00:00
result2 = instance.query("SELECT count(DISTINCT channel_id) FROM test.destination")
2020-06-10 23:01:47 +00:00
for consumer_id in range(num_tables):
table_name = "rabbitmq_consumer{}".format(consumer_id)
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.{0}_mv;
2020-09-07 10:21:29 +00:00
DROP TABLE test.{0};
""".format(
table_name
)
)
2020-06-10 23:01:47 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.destination;
"""
)
2020-06-10 23:01:47 +00:00
for thread in threads:
thread.join()
assert (
int(result1) == messages_num * threads_num
), "ClickHouse lost some messages: {}".format(result)
assert int(result2) == 4 * num_tables
2020-06-10 23:01:47 +00:00
2020-06-11 09:23:23 +00:00
def test_rabbitmq_multiple_bindings(rabbitmq_cluster):
instance.query(
"""
DROP TABLE IF EXISTS test.destination;
CREATE TABLE test.destination(key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
"""
)
instance.query(
"""
DROP TABLE IF EXISTS test.bindings;
DROP TABLE IF EXISTS test.bindings_mv;
CREATE TABLE test.bindings (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'multiple_bindings_testing',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'key1,key2,key3,key4,key5',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.bindings_mv TO test.destination AS
SELECT * FROM test.bindings;
"""
)
2020-06-11 09:23:23 +00:00
i = [0]
messages_num = 500
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-11 09:23:23 +00:00
def produce():
# init connection here because otherwise python rabbitmq client might fail
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-06-11 09:23:23 +00:00
i[0] += 1
keys = ["key1", "key2", "key3", "key4", "key5"]
2020-06-11 09:23:23 +00:00
for key in keys:
for message in messages:
channel.basic_publish(
exchange="multiple_bindings_testing", routing_key=key, body=message
)
2020-06-11 09:23:23 +00:00
connection.close()
threads = []
threads_num = 10
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
while True:
result = instance.query("SELECT count() FROM test.destination")
2020-06-11 09:23:23 +00:00
time.sleep(1)
if int(result) == messages_num * threads_num * 5:
2020-06-11 09:23:23 +00:00
break
for thread in threads:
thread.join()
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.bindings;
DROP TABLE test.bindings_mv;
DROP TABLE test.destination;
"""
)
2020-06-13 18:44:17 +00:00
assert (
int(result) == messages_num * threads_num * 5
), "ClickHouse lost some messages: {}".format(result)
2020-06-11 09:23:23 +00:00
2020-06-13 21:37:37 +00:00
def test_rabbitmq_headers_exchange(rabbitmq_cluster):
instance.query(
"""
2020-06-13 21:37:37 +00:00
DROP TABLE IF EXISTS test.destination;
CREATE TABLE test.destination(key UInt64, value UInt64)
2020-06-13 21:37:37 +00:00
ENGINE = MergeTree()
ORDER BY key;
"""
)
2020-06-13 21:37:37 +00:00
num_tables_to_receive = 2
2020-06-13 21:37:37 +00:00
for consumer_id in range(num_tables_to_receive):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(consumer_id)))
instance.query(
"""
2020-06-13 21:37:37 +00:00
DROP TABLE IF EXISTS test.headers_exchange_{0};
DROP TABLE IF EXISTS test.headers_exchange_{0}_mv;
CREATE TABLE test.headers_exchange_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_num_consumers = 2,
2020-06-13 21:37:37 +00:00
rabbitmq_exchange_name = 'headers_exchange_testing',
rabbitmq_exchange_type = 'headers',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-13 21:37:37 +00:00
rabbitmq_routing_key_list = 'x-match=all,format=logs,type=report,year=2020',
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.headers_exchange_{0}_mv TO test.destination AS
SELECT key, value FROM test.headers_exchange_{0};
""".format(
consumer_id
)
)
2020-06-13 21:37:37 +00:00
num_tables_to_ignore = 2
for consumer_id in range(num_tables_to_ignore):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(consumer_id + num_tables_to_receive)))
instance.query(
"""
2020-06-13 21:37:37 +00:00
DROP TABLE IF EXISTS test.headers_exchange_{0};
DROP TABLE IF EXISTS test.headers_exchange_{0}_mv;
CREATE TABLE test.headers_exchange_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'headers_exchange_testing',
rabbitmq_exchange_type = 'headers',
rabbitmq_routing_key_list = 'x-match=all,format=logs,type=report,year=2019',
rabbitmq_format = 'JSONEachRow',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-06-13 21:37:37 +00:00
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.headers_exchange_{0}_mv TO test.destination AS
SELECT key, value FROM test.headers_exchange_{0};
""".format(
consumer_id + num_tables_to_receive
)
)
2020-06-13 21:37:37 +00:00
i = [0]
messages_num = 1000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-06-13 21:37:37 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-06-13 21:37:37 +00:00
i[0] += 1
fields = {}
fields["format"] = "logs"
fields["type"] = "report"
fields["year"] = "2020"
2020-06-13 21:37:37 +00:00
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="headers_exchange_testing",
routing_key="",
properties=pika.BasicProperties(headers=fields, message_id=str(msg_id)),
body=messages[msg_id],
)
2020-06-13 21:37:37 +00:00
connection.close()
while True:
result = instance.query("SELECT count() FROM test.destination")
2020-06-13 21:37:37 +00:00
time.sleep(1)
if int(result) == messages_num * num_tables_to_receive:
break
for consumer_id in range(num_tables_to_receive + num_tables_to_ignore):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.headers_exchange_{0}_mv;
2020-09-07 10:21:29 +00:00
DROP TABLE test.headers_exchange_{0};
""".format(
consumer_id
)
)
2020-06-13 21:37:37 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.destination;
"""
)
2020-06-13 21:37:37 +00:00
assert (
int(result) == messages_num * num_tables_to_receive
), "ClickHouse lost some messages: {}".format(result)
2020-06-13 21:37:37 +00:00
2020-07-20 10:05:00 +00:00
def test_rabbitmq_virtual_columns(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq_virtuals (key UInt64, value UInt64)
2020-07-20 10:05:00 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'virtuals',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-07-20 10:05:00 +00:00
rabbitmq_format = 'JSONEachRow';
CREATE MATERIALIZED VIEW test.view Engine=Log AS
2020-08-15 06:50:53 +00:00
SELECT value, key, _exchange_name, _channel_id, _delivery_tag, _redelivered FROM test.rabbitmq_virtuals;
"""
)
2020-07-20 10:05:00 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-07-20 10:05:00 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
message_num = 10
2020-08-15 06:50:53 +00:00
i = 0
2020-07-20 10:05:00 +00:00
messages = []
for _ in range(message_num):
messages.append(json.dumps({"key": i, "value": i}))
2020-08-15 06:50:53 +00:00
i += 1
2020-07-20 10:05:00 +00:00
for message in messages:
channel.basic_publish(exchange="virtuals", routing_key="", body=message)
2020-07-20 10:05:00 +00:00
while True:
result = instance.query("SELECT count() FROM test.view")
2020-07-20 10:05:00 +00:00
time.sleep(1)
if int(result) == message_num:
break
connection.close()
result = instance.query(
"""
SELECT key, value, _exchange_name, SUBSTRING(_channel_id, 1, 3), _delivery_tag, _redelivered
2020-08-15 06:50:53 +00:00
FROM test.view ORDER BY key
"""
)
2020-07-20 10:05:00 +00:00
expected = """\
2020-08-15 06:50:53 +00:00
0 0 virtuals 1_0 1 0
1 1 virtuals 1_0 2 0
2 2 virtuals 1_0 3 0
3 3 virtuals 1_0 4 0
4 4 virtuals 1_0 5 0
5 5 virtuals 1_0 6 0
6 6 virtuals 1_0 7 0
7 7 virtuals 1_0 8 0
8 8 virtuals 1_0 9 0
9 9 virtuals 1_0 10 0
"""
2020-08-15 06:50:53 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.rabbitmq_virtuals;
DROP TABLE test.view;
"""
)
2020-08-15 06:50:53 +00:00
2020-07-20 10:05:00 +00:00
assert TSV(result) == TSV(expected)
def test_rabbitmq_virtual_columns_with_materialized_view(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq_virtuals_mv (key UInt64, value UInt64)
2020-07-20 10:05:00 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'virtuals_mv',
rabbitmq_flush_interval_ms=1000,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size=100,
2020-07-20 10:05:00 +00:00
rabbitmq_format = 'JSONEachRow';
CREATE TABLE test.view (key UInt64, value UInt64,
2020-08-15 06:50:53 +00:00
exchange_name String, channel_id String, delivery_tag UInt64, redelivered UInt8) ENGINE = MergeTree()
2020-07-20 10:05:00 +00:00
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
2020-08-15 06:50:53 +00:00
SELECT *, _exchange_name as exchange_name, _channel_id as channel_id, _delivery_tag as delivery_tag, _redelivered as redelivered
FROM test.rabbitmq_virtuals_mv;
"""
)
2020-07-20 10:05:00 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-07-20 10:05:00 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
message_num = 10
2020-08-15 06:50:53 +00:00
i = 0
2020-07-20 10:05:00 +00:00
messages = []
for _ in range(message_num):
messages.append(json.dumps({"key": i, "value": i}))
2020-08-15 06:50:53 +00:00
i += 1
2020-07-20 10:05:00 +00:00
for message in messages:
channel.basic_publish(exchange="virtuals_mv", routing_key="", body=message)
2020-07-20 10:05:00 +00:00
while True:
result = instance.query("SELECT count() FROM test.view")
2020-07-20 10:05:00 +00:00
time.sleep(1)
if int(result) == message_num:
break
connection.close()
result = instance.query(
"SELECT key, value, exchange_name, SUBSTRING(channel_id, 1, 3), delivery_tag, redelivered FROM test.view ORDER BY delivery_tag"
)
expected = """\
2020-08-15 06:50:53 +00:00
0 0 virtuals_mv 1_0 1 0
1 1 virtuals_mv 1_0 2 0
2 2 virtuals_mv 1_0 3 0
3 3 virtuals_mv 1_0 4 0
4 4 virtuals_mv 1_0 5 0
5 5 virtuals_mv 1_0 6 0
6 6 virtuals_mv 1_0 7 0
7 7 virtuals_mv 1_0 8 0
8 8 virtuals_mv 1_0 9 0
9 9 virtuals_mv 1_0 10 0
"""
2020-07-20 10:05:00 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.consumer;
DROP TABLE test.view;
DROP TABLE test.rabbitmq_virtuals_mv
"""
)
2020-07-20 10:05:00 +00:00
assert TSV(result) == TSV(expected)
def test_rabbitmq_many_consumers_to_each_queue(rabbitmq_cluster):
instance.query(
"""
DROP TABLE IF EXISTS test.destination;
2020-08-15 06:50:53 +00:00
CREATE TABLE test.destination(key UInt64, value UInt64, channel_id String)
ENGINE = MergeTree()
ORDER BY key;
"""
)
num_tables = 4
for table_id in range(num_tables):
2020-10-02 16:54:07 +00:00
print(("Setting up table {}".format(table_id)))
instance.query(
"""
DROP TABLE IF EXISTS test.many_consumers_{0};
DROP TABLE IF EXISTS test.many_consumers_{0}_mv;
CREATE TABLE test.many_consumers_{0} (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'many_consumers',
rabbitmq_num_queues = 2,
rabbitmq_num_consumers = 2,
rabbitmq_flush_interval_ms=1000,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size=100,
rabbitmq_queue_base = 'many_consumers',
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
CREATE MATERIALIZED VIEW test.many_consumers_{0}_mv TO test.destination AS
2020-08-15 06:50:53 +00:00
SELECT key, value, _channel_id as channel_id FROM test.many_consumers_{0};
""".format(
table_id
)
)
i = [0]
messages_num = 1000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
def produce():
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
i[0] += 1
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="many_consumers",
routing_key="",
properties=pika.BasicProperties(message_id=str(msg_id)),
body=messages[msg_id],
)
connection.close()
threads = []
threads_num = 20
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
result1 = ""
while True:
result1 = instance.query("SELECT count() FROM test.destination")
time.sleep(1)
if int(result1) == messages_num * threads_num:
break
2020-08-15 06:50:53 +00:00
result2 = instance.query("SELECT count(DISTINCT channel_id) FROM test.destination")
for thread in threads:
thread.join()
for consumer_id in range(num_tables):
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.many_consumers_{0};
DROP TABLE test.many_consumers_{0}_mv;
""".format(
consumer_id
)
)
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.destination;
"""
)
assert (
int(result1) == messages_num * threads_num
), "ClickHouse lost some messages: {}".format(result)
# 4 tables, 2 consumers for each table => 8 consumer tags
assert int(result2) == 8
2020-08-26 08:54:29 +00:00
def test_rabbitmq_restore_failed_connection_without_losses_1(rabbitmq_cluster):
instance.query(
"""
2020-08-08 16:45:52 +00:00
DROP TABLE IF EXISTS test.consume;
2020-08-15 06:50:53 +00:00
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree
ORDER BY key;
2020-08-08 16:45:52 +00:00
CREATE TABLE test.consume (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
2023-01-19 17:06:43 +00:00
rabbitmq_flush_interval_ms=500,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100,
2020-08-08 16:45:52 +00:00
rabbitmq_exchange_name = 'producer_reconnect',
rabbitmq_format = 'JSONEachRow',
2020-09-01 14:11:34 +00:00
rabbitmq_num_consumers = 2,
2020-08-08 16:45:52 +00:00
rabbitmq_row_delimiter = '\\n';
2020-08-15 06:50:53 +00:00
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.consume;
2020-08-08 16:45:52 +00:00
DROP TABLE IF EXISTS test.producer_reconnect;
CREATE TABLE test.producer_reconnect (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'producer_reconnect',
2020-08-31 09:12:36 +00:00
rabbitmq_persistent = '1',
rabbitmq_flush_interval_ms=1000,
2020-08-08 16:45:52 +00:00
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
"""
)
2020-08-08 16:45:52 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-08-08 16:45:52 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages_num = 100000
values = []
for i in range(messages_num):
values.append("({i}, {i})".format(i=i))
values = ",".join(values)
2020-08-08 16:45:52 +00:00
while True:
try:
instance.query(
"INSERT INTO test.producer_reconnect VALUES {}".format(values)
)
2020-08-08 16:45:52 +00:00
break
except QueryRuntimeException as e:
if "Local: Timed out." in str(e):
2020-08-08 16:45:52 +00:00
continue
else:
raise
while int(instance.query("SELECT count() FROM test.view")) == 0:
2020-08-08 16:45:52 +00:00
time.sleep(0.1)
2021-04-30 09:18:12 +00:00
kill_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
2020-09-07 11:08:53 +00:00
time.sleep(4)
2021-04-30 09:18:12 +00:00
revive_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
2020-08-08 16:45:52 +00:00
while True:
result = instance.query("SELECT count(DISTINCT key) FROM test.view")
2020-08-08 16:45:52 +00:00
time.sleep(1)
2020-08-15 06:50:53 +00:00
if int(result) == messages_num:
2020-08-08 16:45:52 +00:00
break
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.consume;
DROP TABLE test.producer_reconnect;
"""
)
2020-08-15 06:50:53 +00:00
assert int(result) == messages_num, "ClickHouse lost some messages: {}".format(
result
)
2020-08-15 06:50:53 +00:00
2023-01-27 16:22:24 +00:00
@pytest.mark.skip(reason="Timeout: FIXME")
2020-08-26 08:54:29 +00:00
def test_rabbitmq_restore_failed_connection_without_losses_2(rabbitmq_cluster):
2022-12-20 18:32:59 +00:00
logging.getLogger("pika").propagate = False
instance.query(
"""
CREATE TABLE test.consumer_reconnect (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'consumer_reconnect',
rabbitmq_num_consumers = 10,
2022-12-20 18:32:59 +00:00
rabbitmq_flush_interval_ms = 100,
rabbitmq_max_block_size = 100,
2020-10-27 07:14:38 +00:00
rabbitmq_num_queues = 10,
rabbitmq_format = 'JSONEachRow',
rabbitmq_row_delimiter = '\\n';
"""
)
2020-08-15 06:50:53 +00:00
i = 0
messages_num = 150000
2020-08-15 06:50:53 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-08-15 06:50:53 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
messages.append(json.dumps({"key": i, "value": i}))
2020-08-15 06:50:53 +00:00
i += 1
2020-09-07 10:21:29 +00:00
for msg_id in range(messages_num):
channel.basic_publish(
exchange="consumer_reconnect",
routing_key="",
body=messages[msg_id],
properties=pika.BasicProperties(delivery_mode=2, message_id=str(msg_id)),
)
connection.close()
instance.query(
"""
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.consumer_reconnect;
"""
)
while int(instance.query("SELECT count() FROM test.view")) == 0:
2021-09-16 10:46:43 +00:00
print(3)
time.sleep(0.1)
2021-04-30 09:18:12 +00:00
kill_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
2020-09-07 11:08:53 +00:00
time.sleep(8)
2021-04-30 09:18:12 +00:00
revive_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
# while int(instance.query('SELECT count() FROM test.view')) == 0:
2020-08-26 08:54:29 +00:00
# time.sleep(0.1)
2020-08-15 06:50:53 +00:00
# kill_rabbitmq()
# time.sleep(2)
# revive_rabbitmq()
2020-08-15 06:50:53 +00:00
while True:
2022-12-20 18:32:59 +00:00
result = instance.query("SELECT count(DISTINCT key) FROM test.view").strip()
if int(result) == messages_num:
2020-08-15 06:50:53 +00:00
break
2022-12-20 18:32:59 +00:00
print(f"Result: {result} / {messages_num}")
time.sleep(1)
2020-08-15 06:50:53 +00:00
instance.query(
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test.consumer;
DROP TABLE test.consumer_reconnect;
"""
)
2020-08-08 16:45:52 +00:00
assert int(result) == messages_num, "ClickHouse lost some messages: {}".format(
result
)
2020-08-08 16:45:52 +00:00
2020-08-31 16:34:16 +00:00
def test_rabbitmq_commit_on_block_write(rabbitmq_cluster):
2022-12-20 18:32:59 +00:00
logging.getLogger("pika").propagate = False
instance.query(
"""
2020-08-31 16:34:16 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'block',
rabbitmq_format = 'JSONEachRow',
rabbitmq_queue_base = 'block',
rabbitmq_flush_interval_ms=1000,
2020-08-31 16:34:16 +00:00
rabbitmq_max_block_size = 100,
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
2020-08-31 16:34:16 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-08-31 16:34:16 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
cancel = threading.Event()
i = [0]
2020-08-31 16:34:16 +00:00
def produce():
while not cancel.is_set():
messages = []
for _ in range(101):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2020-08-31 16:34:16 +00:00
i[0] += 1
for message in messages:
channel.basic_publish(exchange="block", routing_key="", body=message)
2020-08-31 16:34:16 +00:00
rabbitmq_thread = threading.Thread(target=produce)
rabbitmq_thread.start()
while int(instance.query("SELECT count() FROM test.view")) == 0:
2020-08-31 16:34:16 +00:00
time.sleep(1)
cancel.set()
instance.query("DETACH TABLE test.rabbitmq;")
2020-08-31 16:34:16 +00:00
while (
int(
instance.query(
"SELECT count() FROM system.tables WHERE database='test' AND name='rabbitmq'"
)
)
== 1
):
2020-08-31 16:34:16 +00:00
time.sleep(1)
instance.query("ATTACH TABLE test.rabbitmq;")
2020-08-31 16:34:16 +00:00
while int(instance.query("SELECT uniqExact(key) FROM test.view")) < i[0]:
2020-08-31 16:34:16 +00:00
time.sleep(1)
result = int(instance.query("SELECT count() == uniqExact(key) FROM test.view"))
2020-08-31 16:34:16 +00:00
instance.query(
"""
2020-08-31 16:34:16 +00:00
DROP TABLE test.consumer;
DROP TABLE test.view;
"""
)
2020-08-31 16:34:16 +00:00
rabbitmq_thread.join()
connection.close()
assert result == 1, "Messages from RabbitMQ get duplicated!"
2020-08-31 16:34:16 +00:00
2021-09-16 10:46:43 +00:00
def test_rabbitmq_no_connection_at_startup_1(rabbitmq_cluster):
2020-12-02 18:34:01 +00:00
# no connection when table is initialized
rabbitmq_cluster.pause_container("rabbitmq1")
instance.query_and_get_error(
"""
2021-09-16 10:46:43 +00:00
CREATE TABLE test.cs (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'cs',
rabbitmq_format = 'JSONEachRow',
rabbitmq_flush_interval_ms=1000,
2021-09-16 10:46:43 +00:00
rabbitmq_num_consumers = '5',
rabbitmq_row_delimiter = '\\n';
"""
)
rabbitmq_cluster.unpause_container("rabbitmq1")
2021-09-16 10:46:43 +00:00
def test_rabbitmq_no_connection_at_startup_2(rabbitmq_cluster):
instance.query(
"""
2020-12-02 18:34:01 +00:00
CREATE TABLE test.cs (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'cs',
rabbitmq_format = 'JSONEachRow',
rabbitmq_num_consumers = '5',
rabbitmq_flush_interval_ms=1000,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size=100,
2020-12-02 18:34:01 +00:00
rabbitmq_row_delimiter = '\\n';
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.cs;
"""
)
2021-09-16 10:46:43 +00:00
instance.query("DETACH TABLE test.cs")
rabbitmq_cluster.pause_container("rabbitmq1")
2021-09-16 10:46:43 +00:00
instance.query("ATTACH TABLE test.cs")
rabbitmq_cluster.unpause_container("rabbitmq1")
2020-12-02 18:34:01 +00:00
messages_num = 1000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2020-12-02 18:34:01 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
for i in range(messages_num):
message = json.dumps({"key": i, "value": i})
channel.basic_publish(
exchange="cs",
routing_key="",
body=message,
properties=pika.BasicProperties(delivery_mode=2, message_id=str(i)),
)
2020-12-02 18:34:01 +00:00
connection.close()
while True:
result = instance.query("SELECT count() FROM test.view")
2020-12-02 18:34:01 +00:00
time.sleep(1)
if int(result) == messages_num:
break
instance.query(
"""
2020-12-02 18:34:01 +00:00
DROP TABLE test.consumer;
DROP TABLE test.cs;
"""
)
2020-12-02 18:34:01 +00:00
assert int(result) == messages_num, "ClickHouse lost some messages: {}".format(
result
)
2020-12-02 18:34:01 +00:00
2021-02-15 21:56:51 +00:00
def test_rabbitmq_format_factory_settings(rabbitmq_cluster):
instance.query(
"""
2021-02-15 21:56:51 +00:00
CREATE TABLE test.format_settings (
id String, date DateTime
) ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'format_settings',
rabbitmq_flush_interval_ms=1000,
2021-02-15 21:56:51 +00:00
rabbitmq_format = 'JSONEachRow',
date_time_input_format = 'best_effort';
"""
)
2021-02-15 21:56:51 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-02-15 21:56:51 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
message = json.dumps(
{"id": "format_settings_test", "date": "2021-01-19T14:42:33.1829214Z"}
)
expected = instance.query(
"""SELECT parseDateTimeBestEffort(CAST('2021-01-19T14:42:33.1829214Z', 'String'))"""
)
2021-02-15 21:56:51 +00:00
channel.basic_publish(exchange="format_settings", routing_key="", body=message)
result = ""
2021-02-15 21:56:51 +00:00
while True:
result = instance.query("SELECT date FROM test.format_settings")
2021-02-15 21:56:51 +00:00
if result == expected:
break
2021-02-15 21:56:51 +00:00
instance.query(
"""
2021-02-15 21:56:51 +00:00
CREATE TABLE test.view (
id String, date DateTime
) ENGINE = MergeTree ORDER BY id;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.format_settings;
"""
)
2021-02-15 21:56:51 +00:00
channel.basic_publish(exchange="format_settings", routing_key="", body=message)
result = ""
2021-02-15 21:56:51 +00:00
while True:
result = instance.query("SELECT date FROM test.view")
2021-02-15 21:56:51 +00:00
if result == expected:
break
2021-02-15 21:56:51 +00:00
connection.close()
instance.query(
"""
2021-02-15 21:56:51 +00:00
DROP TABLE test.consumer;
DROP TABLE test.format_settings;
"""
)
2021-02-15 21:56:51 +00:00
assert result == expected
2021-02-15 21:56:51 +00:00
2021-04-21 15:51:05 +00:00
def test_rabbitmq_vhost(rabbitmq_cluster):
instance.query(
"""
2021-04-21 15:51:05 +00:00
CREATE TABLE test.rabbitmq_vhost (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'vhost',
rabbitmq_format = 'JSONEachRow',
rabbitmq_flush_interval_ms=1000,
2021-04-21 15:51:05 +00:00
rabbitmq_vhost = '/'
"""
)
2021-04-21 15:51:05 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-04-21 15:51:05 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
channel.basic_publish(
exchange="vhost", routing_key="", body=json.dumps({"key": 1, "value": 2})
)
2021-04-21 15:51:05 +00:00
connection.close()
while True:
result = instance.query(
"SELECT * FROM test.rabbitmq_vhost ORDER BY key", ignore_error=True
)
2021-04-21 15:51:05 +00:00
if result == "1\t2\n":
break
2021-05-04 16:26:47 +00:00
def test_rabbitmq_drop_table_properly(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq_drop (key UInt64, value UInt64)
2021-05-04 16:26:47 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_flush_interval_ms=1000,
2021-05-04 16:26:47 +00:00
rabbitmq_exchange_name = 'drop',
rabbitmq_format = 'JSONEachRow',
2021-05-13 09:39:57 +00:00
rabbitmq_queue_base = 'rabbit_queue_drop'
"""
)
2021-05-04 16:26:47 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-05-04 16:26:47 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
channel.basic_publish(
exchange="drop", routing_key="", body=json.dumps({"key": 1, "value": 2})
)
2021-05-04 16:26:47 +00:00
while True:
result = instance.query(
"SELECT * FROM test.rabbitmq_drop ORDER BY key", ignore_error=True
)
2021-05-04 16:26:47 +00:00
if result == "1\t2\n":
break
exists = channel.queue_declare(queue="rabbit_queue_drop", passive=True)
assert exists
2021-05-04 16:26:47 +00:00
instance.query("DROP TABLE test.rabbitmq_drop")
2021-05-04 16:26:47 +00:00
time.sleep(30)
try:
exists = channel.queue_declare(
callback, queue="rabbit_queue_drop", passive=True
)
2021-05-04 16:26:47 +00:00
except Exception as e:
exists = False
assert not exists
2021-05-04 16:26:47 +00:00
def test_rabbitmq_queue_settings(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq_settings (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'rabbit_exchange',
rabbitmq_flush_interval_ms=1000,
rabbitmq_format = 'JSONEachRow',
2021-05-13 09:39:57 +00:00
rabbitmq_queue_base = 'rabbit_queue_settings',
rabbitmq_queue_settings_list = 'x-max-length=10,x-overflow=reject-publish'
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
for i in range(50):
channel.basic_publish(
exchange="rabbit_exchange",
routing_key="",
body=json.dumps({"key": 1, "value": 2}),
)
connection.close()
instance.query(
"""
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq_settings;
"""
)
time.sleep(5)
2022-12-20 18:32:59 +00:00
while True:
result = instance.query("SELECT count() FROM test.view", ignore_error=True)
2022-12-20 18:32:59 +00:00
if int(result) == 10:
break
time.sleep(0.5)
instance.query("DROP TABLE test.rabbitmq_settings")
2021-05-04 19:57:45 +00:00
# queue size is 10, but 50 messages were sent, they will be dropped (setting x-overflow = reject-publish) and only 10 will remain.
assert int(result) == 10
def test_rabbitmq_queue_consume(rabbitmq_cluster):
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
channel.queue_declare(queue="rabbit_queue", durable=True)
i = [0]
messages_num = 1000
def produce():
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for _ in range(messages_num):
message = json.dumps({"key": i[0], "value": i[0]})
channel.basic_publish(exchange="", routing_key="rabbit_queue", body=message)
i[0] += 1
threads = []
threads_num = 10
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
instance.query(
"""
CREATE TABLE test.rabbitmq_queue (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_format = 'JSONEachRow',
rabbitmq_queue_base = 'rabbit_queue',
rabbitmq_flush_interval_ms=1000,
rabbitmq_queue_consume = 1;
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq_queue;
"""
)
result = ""
while True:
result = instance.query("SELECT count() FROM test.view")
if int(result) == messages_num * threads_num:
break
time.sleep(1)
for thread in threads:
thread.join()
instance.query("DROP TABLE test.rabbitmq_queue")
2021-05-04 19:57:45 +00:00
def test_rabbitmq_produce_consume_avro(rabbitmq_cluster):
num_rows = 75
instance.query(
"""
DROP TABLE IF EXISTS test.view;
DROP TABLE IF EXISTS test.rabbit;
DROP TABLE IF EXISTS test.rabbit_writer;
CREATE TABLE test.rabbit_writer (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_format = 'Avro',
rabbitmq_flush_interval_ms=1000,
rabbitmq_exchange_name = 'avro',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'avro';
CREATE TABLE test.rabbit (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_format = 'Avro',
rabbitmq_flush_interval_ms=1000,
rabbitmq_exchange_name = 'avro',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'avro';
CREATE MATERIALIZED VIEW test.view Engine=Log AS
SELECT key, value FROM test.rabbit;
"""
)
instance.query(
"INSERT INTO test.rabbit_writer select number*10 as key, number*100 as value from numbers({num_rows}) SETTINGS output_format_avro_rows_in_file = 7".format(
num_rows=num_rows
)
)
# Ideally we should wait for an event
time.sleep(3)
expected_num_rows = instance.query(
"SELECT COUNT(1) FROM test.view", ignore_error=True
)
assert int(expected_num_rows) == num_rows
expected_max_key = instance.query(
"SELECT max(key) FROM test.view", ignore_error=True
)
assert int(expected_max_key) == (num_rows - 1) * 10
2021-09-16 10:46:43 +00:00
def test_rabbitmq_bad_args(rabbitmq_cluster):
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-09-16 10:46:43 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
channel.exchange_declare(exchange="f", exchange_type="fanout")
instance.query_and_get_error(
"""
CREATE TABLE test.drop (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_flush_interval_ms=1000,
2021-09-16 10:46:43 +00:00
rabbitmq_exchange_name = 'f',
rabbitmq_format = 'JSONEachRow';
"""
)
2021-11-02 12:05:10 +00:00
def test_rabbitmq_issue_30691(rabbitmq_cluster):
instance.query(
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test.rabbitmq_drop (json String)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_flush_interval_ms=1000,
2021-11-02 12:05:10 +00:00
rabbitmq_exchange_name = '30691',
rabbitmq_row_delimiter = '\\n', -- Works only if adding this setting
rabbitmq_format = 'LineAsString',
rabbitmq_queue_base = '30691';
"""
)
2021-11-02 12:05:10 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-11-02 12:05:10 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
channel.basic_publish(
exchange="30691",
routing_key="",
body=json.dumps(
{
"event_type": "purge",
"as_src": 1234,
"as_dst": 0,
"as_path": "",
"local_pref": 100,
"med": 0,
"peer_as_dst": 0,
"ip_src": "<redacted ipv6>",
"ip_dst": "<redacted ipv6>",
"port_src": 443,
"port_dst": 41930,
"ip_proto": "tcp",
"tos": 0,
"stamp_inserted": "2021-10-26 15:20:00",
"stamp_updated": "2021-10-26 15:23:14",
"packets": 2,
"bytes": 1216,
"writer_id": "default_amqp/449206",
}
),
)
result = ""
2021-11-02 12:05:10 +00:00
while True:
result = instance.query("SELECT * FROM test.rabbitmq_drop", ignore_error=True)
2021-11-02 12:05:10 +00:00
print(result)
if result != "":
break
assert (
result.strip()
== """{"event_type": "purge", "as_src": 1234, "as_dst": 0, "as_path": "", "local_pref": 100, "med": 0, "peer_as_dst": 0, "ip_src": "<redacted ipv6>", "ip_dst": "<redacted ipv6>", "port_src": 443, "port_dst": 41930, "ip_proto": "tcp", "tos": 0, "stamp_inserted": "2021-10-26 15:20:00", "stamp_updated": "2021-10-26 15:23:14", "packets": 2, "bytes": 1216, "writer_id": "default_amqp/449206"}"""
)
2021-11-02 12:05:10 +00:00
def test_rabbitmq_drop_mv(rabbitmq_cluster):
instance.query(
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'mv',
rabbitmq_format = 'JSONEachRow',
rabbitmq_flush_interval_ms=1000,
2021-11-02 12:05:10 +00:00
rabbitmq_queue_base = 'drop_mv';
"""
)
instance.query(
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
"""
)
instance.query(
"""
2021-11-02 12:05:10 +00:00
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
2021-11-02 12:05:10 +00:00
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-11-02 12:05:10 +00:00
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for i in range(20):
channel.basic_publish(
exchange="mv", routing_key="", body=json.dumps({"key": i, "value": i})
)
2021-11-02 12:05:10 +00:00
start = time.time()
while time.time() - start < 30:
res = instance.query("SELECT COUNT(*) FROM test.view")
if "20" == res:
break
else:
logging.debug(f"Number of rows in test.view: {res}")
instance.query("DROP VIEW test.consumer")
2021-11-02 12:05:10 +00:00
for i in range(20, 40):
channel.basic_publish(
exchange="mv", routing_key="", body=json.dumps({"key": i, "value": i})
)
2021-11-02 12:05:10 +00:00
instance.query(
"""
2021-11-02 12:05:10 +00:00
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
2021-11-02 12:05:10 +00:00
for i in range(40, 50):
channel.basic_publish(
exchange="mv", routing_key="", body=json.dumps({"key": i, "value": i})
)
2021-11-02 12:05:10 +00:00
while True:
result = instance.query("SELECT * FROM test.view ORDER BY key")
if rabbitmq_check_result(result):
2021-11-02 12:05:10 +00:00
break
2023-02-01 17:22:05 +00:00
time.sleep(1)
2021-11-02 12:05:10 +00:00
rabbitmq_check_result(result, True)
2022-12-20 18:32:59 +00:00
instance.query("DROP VIEW test.consumer NO DELAY")
time.sleep(10)
2021-11-02 12:05:10 +00:00
for i in range(50, 60):
channel.basic_publish(
exchange="mv", routing_key="", body=json.dumps({"key": i, "value": i})
)
2021-11-02 12:05:10 +00:00
connection.close()
count = 0
start = time.time()
while time.time() - start < 30:
count = int(instance.query("SELECT count() FROM test.rabbitmq"))
if count:
2021-11-02 12:05:10 +00:00
break
assert count > 0
2021-11-02 12:05:10 +00:00
def test_rabbitmq_random_detach(rabbitmq_cluster):
NUM_CONSUMERS = 2
NUM_QUEUES = 2
instance.query(
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'random',
rabbitmq_queue_base = 'random',
rabbitmq_num_queues = 2,
rabbitmq_flush_interval_ms=1000,
2021-11-02 12:05:10 +00:00
rabbitmq_num_consumers = 2,
rabbitmq_format = 'JSONEachRow';
CREATE TABLE test.view (key UInt64, value UInt64, channel_id String)
ENGINE = MergeTree
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT *, _channel_id AS channel_id FROM test.rabbitmq;
"""
)
2021-11-02 12:05:10 +00:00
i = [0]
messages_num = 10000
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
2021-11-02 12:05:10 +00:00
def produce():
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
messages = []
for j in range(messages_num):
messages.append(json.dumps({"key": i[0], "value": i[0]}))
2021-11-02 12:05:10 +00:00
i[0] += 1
mes_id = str(i)
channel.basic_publish(
exchange="test_sharding",
routing_key="",
properties=pika.BasicProperties(message_id=mes_id),
body=message,
)
2021-11-02 12:05:10 +00:00
connection.close()
threads = []
threads_num = 20
for _ in range(threads_num):
threads.append(threading.Thread(target=produce))
for thread in threads:
time.sleep(random.uniform(0, 1))
thread.start()
# time.sleep(5)
# kill_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
# instance.query("detach table test.rabbitmq")
# revive_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
2021-11-02 12:05:10 +00:00
for thread in threads:
thread.join()
def test_rabbitmq_predefined_configuration(rabbitmq_cluster):
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
connection = pika.BlockingConnection(parameters)
channel = connection.channel()
instance.query(
"""
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ(rabbit1, rabbitmq_vhost = '/')
SETTINGS rabbitmq_flush_interval_ms=1000;
"""
)
channel.basic_publish(
exchange="named", routing_key="", body=json.dumps({"key": 1, "value": 2})
)
while True:
result = instance.query(
"SELECT * FROM test.rabbitmq ORDER BY key", ignore_error=True
)
if result == "1\t2\n":
break
2022-05-31 12:38:11 +00:00
instance.restart_clickhouse()
channel.basic_publish(
exchange="named", routing_key="", body=json.dumps({"key": 1, "value": 2})
)
while True:
result = instance.query(
"SELECT * FROM test.rabbitmq ORDER BY key", ignore_error=True
)
if result == "1\t2\n":
break
2022-07-01 17:06:57 +00:00
def test_rabbitmq_msgpack(rabbitmq_cluster):
instance.query(
"""
drop table if exists rabbit_in;
drop table if exists rabbit_out;
create table
rabbit_in (val String)
engine=RabbitMQ
settings rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'xhep',
rabbitmq_format = 'MsgPack',
rabbitmq_flush_interval_ms=1000,
2022-07-01 17:06:57 +00:00
rabbitmq_num_consumers = 1;
create table
rabbit_out (val String)
engine=RabbitMQ
settings rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'xhep',
rabbitmq_format = 'MsgPack',
rabbitmq_flush_interval_ms=1000,
2022-07-01 17:06:57 +00:00
rabbitmq_num_consumers = 1;
set stream_like_engine_allow_direct_select=1;
insert into rabbit_out select 'kek';
"""
)
result = ""
try_no = 0
while True:
result = instance.query("select * from rabbit_in;")
if result.strip() == "kek":
break
else:
try_no = try_no + 1
if try_no == 20:
break
time.sleep(1)
assert result.strip() == "kek"
2022-07-04 12:06:20 +00:00
instance.query("drop table rabbit_in sync")
instance.query("drop table rabbit_out sync")
2022-07-01 17:06:57 +00:00
2022-06-24 00:07:16 +00:00
def test_rabbitmq_address(rabbitmq_cluster):
2022-06-29 05:32:27 +00:00
instance2.query(
"""
2022-06-24 00:07:16 +00:00
drop table if exists rabbit_in;
drop table if exists rabbit_out;
create table
rabbit_in (val String)
engine=RabbitMQ
SETTINGS rabbitmq_exchange_name = 'rxhep',
rabbitmq_format = 'CSV',
rabbitmq_num_consumers = 1,
rabbitmq_flush_interval_ms=1000,
2022-06-24 00:07:16 +00:00
rabbitmq_address='amqp://root:clickhouse@rabbitmq1:5672/';
create table
rabbit_out (val String) engine=RabbitMQ
SETTINGS rabbitmq_exchange_name = 'rxhep',
rabbitmq_format = 'CSV',
rabbitmq_num_consumers = 1,
rabbitmq_flush_interval_ms=1000,
2022-06-24 00:07:16 +00:00
rabbitmq_address='amqp://root:clickhouse@rabbitmq1:5672/';
set stream_like_engine_allow_direct_select=1;
insert into rabbit_out select 'kek';
2022-06-29 05:32:27 +00:00
"""
)
2022-06-24 00:07:16 +00:00
result = ""
try_no = 0
while True:
result = instance2.query("select * from rabbit_in;")
if result.strip() == "kek":
break
else:
try_no = try_no + 1
if try_no == 20:
break
time.sleep(1)
assert result.strip() == "kek"
2022-07-04 12:06:20 +00:00
2022-07-04 14:08:10 +00:00
instance2.query("drop table rabbit_in sync")
instance2.query("drop table rabbit_out sync")
def test_format_with_prefix_and_suffix(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'insert',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'custom',
rabbitmq_format = 'CustomSeparated';
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
consumer_connection = pika.BlockingConnection(parameters)
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
queue_name = result.method.queue
consumer.queue_bind(exchange="insert", queue=queue_name, routing_key="custom")
instance.query(
"INSERT INTO test.rabbitmq select number*10 as key, number*100 as value from numbers(2) settings format_custom_result_before_delimiter='<prefix>\n', format_custom_result_after_delimiter='<suffix>\n'"
)
insert_messages = []
def onReceived(channel, method, properties, body):
2023-02-02 19:16:53 +00:00
message = body.decode()
insert_messages.append(message)
print(f"Received {len(insert_messages)} message: {message}")
if len(insert_messages) == 2:
channel.stop_consuming()
consumer.basic_consume(onReceived, queue_name)
2023-01-31 19:25:08 +00:00
consumer.start_consuming()
consumer_connection.close()
assert (
"".join(insert_messages)
== "<prefix>\n0\t0\n<suffix>\n<prefix>\n10\t100\n<suffix>\n"
)
def test_max_rows_per_message(rabbitmq_cluster):
num_rows = 5
instance.query(
"""
DROP TABLE IF EXISTS test.view;
DROP TABLE IF EXISTS test.rabbit;
CREATE TABLE test.rabbit (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_format = 'CustomSeparated',
rabbitmq_exchange_name = 'custom',
rabbitmq_exchange_type = 'direct',
rabbitmq_routing_key_list = 'custom1',
rabbitmq_max_rows_per_message = 3,
2023-01-31 19:25:08 +00:00
rabbitmq_flush_interval_ms = 1000,
format_custom_result_before_delimiter = '<prefix>\n',
format_custom_result_after_delimiter = '<suffix>\n';
CREATE MATERIALIZED VIEW test.view Engine=Log AS
SELECT key, value FROM test.rabbit;
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
consumer_connection = pika.BlockingConnection(parameters)
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
queue_name = result.method.queue
consumer.queue_bind(exchange="custom", queue=queue_name, routing_key="custom1")
instance.query(
f"INSERT INTO test.rabbit select number*10 as key, number*100 as value from numbers({num_rows}) settings format_custom_result_before_delimiter='<prefix>\n', format_custom_result_after_delimiter='<suffix>\n'"
)
insert_messages = []
def onReceived(channel, method, properties, body):
insert_messages.append(body.decode())
if len(insert_messages) == 2:
channel.stop_consuming()
consumer.basic_consume(onReceived, queue_name)
consumer.start_consuming()
consumer_connection.close()
assert len(insert_messages) == 2
assert (
"".join(insert_messages)
== "<prefix>\n0\t0\n10\t100\n20\t200\n<suffix>\n<prefix>\n30\t300\n40\t400\n<suffix>\n"
)
attempt = 0
rows = 0
while attempt < 100:
rows = int(instance.query("SELECT count() FROM test.view"))
if rows == num_rows:
break
attempt += 1
assert rows == num_rows
result = instance.query("SELECT * FROM test.view")
assert result == "0\t0\n10\t100\n20\t200\n30\t300\n40\t400\n"
def test_row_based_formats(rabbitmq_cluster):
num_rows = 10
for format_name in [
"TSV",
"TSVWithNamesAndTypes",
"TSKV",
"CSV",
"CSVWithNamesAndTypes",
"CustomSeparatedWithNamesAndTypes",
"Values",
"JSON",
"JSONEachRow",
"JSONCompactEachRow",
"JSONCompactEachRowWithNamesAndTypes",
"JSONObjectEachRow",
"Avro",
"RowBinary",
"RowBinaryWithNamesAndTypes",
"MsgPack",
]:
print(format_name)
instance.query(
f"""
DROP TABLE IF EXISTS test.view;
DROP TABLE IF EXISTS test.rabbit;
CREATE TABLE test.rabbit (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_format = '{format_name}',
rabbitmq_exchange_name = '{format_name}',
rabbitmq_exchange_type = 'direct',
2023-01-27 16:22:24 +00:00
rabbitmq_max_block_size = 100,
rabbitmq_flush_interval_ms = 1000,
rabbitmq_routing_key_list = '{format_name}',
rabbitmq_max_rows_per_message = 5;
CREATE MATERIALIZED VIEW test.view Engine=Log AS
SELECT key, value FROM test.rabbit;
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip,
rabbitmq_cluster.rabbitmq_port,
"/",
credentials,
)
consumer_connection = pika.BlockingConnection(parameters)
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
queue_name = result.method.queue
consumer.queue_bind(
exchange=format_name, queue=queue_name, routing_key=format_name
)
instance.query(
f"INSERT INTO test.rabbit SELECT number * 10 as key, number * 100 as value FROM numbers({num_rows});"
)
insert_messages = 0
def onReceived(channel, method, properties, body):
nonlocal insert_messages
insert_messages += 1
if insert_messages == 2:
channel.stop_consuming()
consumer.basic_consume(onReceived, queue_name)
consumer.start_consuming()
consumer_connection.close()
assert insert_messages == 2
attempt = 0
rows = 0
while attempt < 100:
rows = int(instance.query("SELECT count() FROM test.view"))
if rows == num_rows:
break
attempt += 1
assert rows == num_rows
expected = ""
for i in range(num_rows):
expected += str(i * 10) + "\t" + str(i * 100) + "\n"
result = instance.query("SELECT * FROM test.view")
assert result == expected
def test_block_based_formats_1(rabbitmq_cluster):
instance.query(
"""
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'PrettySpace',
rabbitmq_exchange_type = 'direct',
2023-01-27 16:22:24 +00:00
rabbitmq_max_block_size = 100,
rabbitmq_flush_interval_ms = 1000,
rabbitmq_routing_key_list = 'PrettySpace',
rabbitmq_format = 'PrettySpace';
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip, rabbitmq_cluster.rabbitmq_port, "/", credentials
)
consumer_connection = pika.BlockingConnection(parameters)
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
queue_name = result.method.queue
consumer.queue_bind(
exchange="PrettySpace", queue=queue_name, routing_key="PrettySpace"
)
instance.query(
"INSERT INTO test.rabbitmq SELECT number * 10 as key, number * 100 as value FROM numbers(5) settings max_block_size=2, optimize_trivial_insert_select=0;"
)
insert_messages = []
def onReceived(channel, method, properties, body):
insert_messages.append(body.decode())
if len(insert_messages) == 3:
channel.stop_consuming()
consumer.basic_consume(onReceived, queue_name)
consumer.start_consuming()
consumer_connection.close()
assert len(insert_messages) == 3
data = []
for message in insert_messages:
splitted = message.split("\n")
assert splitted[0] == " \x1b[1mkey\x1b[0m \x1b[1mvalue\x1b[0m"
assert splitted[1] == ""
assert splitted[-1] == ""
data += [line.split() for line in splitted[2:-1]]
assert data == [
["0", "0"],
["10", "100"],
["20", "200"],
["30", "300"],
["40", "400"],
]
def test_block_based_formats_2(rabbitmq_cluster):
num_rows = 100
for format_name in [
"JSONColumns",
"Native",
"Arrow",
"Parquet",
"ORC",
"JSONCompactColumns",
]:
print(format_name)
instance.query(
f"""
DROP TABLE IF EXISTS test.view;
DROP TABLE IF EXISTS test.rabbit;
CREATE TABLE test.rabbit (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_format = '{format_name}',
rabbitmq_exchange_name = '{format_name}',
rabbitmq_exchange_type = 'direct',
2023-01-27 16:22:24 +00:00
rabbitmq_max_block_size = 100,
rabbitmq_flush_interval_ms = 1000,
rabbitmq_routing_key_list = '{format_name}';
CREATE MATERIALIZED VIEW test.view Engine=Log AS
SELECT key, value FROM test.rabbit;
"""
)
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip,
rabbitmq_cluster.rabbitmq_port,
"/",
credentials,
)
consumer_connection = pika.BlockingConnection(parameters)
consumer = consumer_connection.channel()
result = consumer.queue_declare(queue="")
queue_name = result.method.queue
consumer.queue_bind(
exchange=format_name, queue=queue_name, routing_key=format_name
)
instance.query(
f"INSERT INTO test.rabbit SELECT number * 10 as key, number * 100 as value FROM numbers({num_rows}) settings max_block_size=12, optimize_trivial_insert_select=0;"
)
insert_messages = 0
def onReceived(channel, method, properties, body):
nonlocal insert_messages
insert_messages += 1
if insert_messages == 9:
channel.stop_consuming()
consumer.basic_consume(onReceived, queue_name)
consumer.start_consuming()
consumer_connection.close()
assert insert_messages == 9
attempt = 0
rows = 0
while attempt < 100:
rows = int(instance.query("SELECT count() FROM test.view"))
if rows == num_rows:
break
attempt += 1
assert rows == num_rows
result = instance.query("SELECT * FROM test.view ORDER by key")
expected = ""
for i in range(num_rows):
expected += str(i * 10) + "\t" + str(i * 100) + "\n"
assert result == expected
def test_rabbitmq_flush_by_block_size(rabbitmq_cluster):
2023-01-06 15:53:24 +00:00
instance.query(
"""
DROP TABLE IF EXISTS test.view;
DROP TABLE IF EXISTS test.consumer;
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'flush_by_block',
rabbitmq_queue_base = 'flush_by_block',
rabbitmq_max_block_size = 100,
rabbitmq_flush_interval_ms = 640000, /* should not flush by time during test */
rabbitmq_format = 'JSONEachRow';
CREATE TABLE test.view (key UInt64, value UInt64)
ENGINE = MergeTree()
ORDER BY key;
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
2023-01-31 19:25:08 +00:00
SYSTEM STOP MERGES;
"""
2023-01-06 15:53:24 +00:00
)
2023-01-06 15:53:24 +00:00
cancel = threading.Event()
2023-01-06 15:53:24 +00:00
def produce():
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip,
rabbitmq_cluster.rabbitmq_port,
"/",
credentials,
)
connection = pika.BlockingConnection(parameters)
2023-01-06 15:53:24 +00:00
while not cancel.is_set():
try:
channel = connection.channel()
channel.basic_publish(
exchange="flush_by_block",
routing_key="",
body=json.dumps({"key": 0, "value": 0}),
)
except e:
print(f"Got error: {str(e)}")
2023-01-06 15:53:24 +00:00
produce_thread = threading.Thread(target=produce)
produce_thread.start()
2023-01-06 15:53:24 +00:00
while 0 == int(
instance.query(
"SELECT count() FROM system.parts WHERE database = 'test' AND table = 'view' AND name = 'all_1_1_0'"
)
):
time.sleep(0.5)
cancel.set()
produce_thread.join()
# more flushes can happens during test, we need to check only result of first flush (part named all_1_1_0).
result = instance.query("SELECT count() FROM test.view WHERE _part='all_1_1_0'")
# logging.debug(result)
instance.query(
"""
DROP TABLE test.consumer;
DROP TABLE test.view;
DROP TABLE test.rabbitmq;
"""
2023-01-06 15:53:24 +00:00
)
# 100 = first poll should return 100 messages (and rows)
# not waiting for stream_flush_interval_ms
assert (
int(result) == 100
), "Messages from rabbitmq should be flushed when block of size rabbitmq_max_block_size is formed!"
def test_rabbitmq_flush_by_time(rabbitmq_cluster):
instance.query(
"""
DROP TABLE IF EXISTS test.view;
DROP TABLE IF EXISTS test.consumer;
CREATE TABLE test.rabbitmq (key UInt64, value UInt64)
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = 'rabbitmq1:5672',
rabbitmq_exchange_name = 'flush_by_time',
rabbitmq_queue_base = 'flush_by_time',
rabbitmq_max_block_size = 100,
rabbitmq_flush_interval_ms = 5000,
rabbitmq_format = 'JSONEachRow';
CREATE TABLE test.view (key UInt64, value UInt64, ts DateTime64(3) MATERIALIZED now64(3))
ENGINE = MergeTree()
ORDER BY key;
"""
)
cancel = threading.Event()
def produce():
credentials = pika.PlainCredentials("root", "clickhouse")
parameters = pika.ConnectionParameters(
rabbitmq_cluster.rabbitmq_ip,
rabbitmq_cluster.rabbitmq_port,
"/",
credentials,
)
connection = pika.BlockingConnection(parameters)
while not cancel.is_set():
try:
channel = connection.channel()
channel.basic_publish(
exchange="flush_by_time",
routing_key="",
body=json.dumps({"key": 0, "value": 0}),
)
print("Produced a message")
time.sleep(0.8)
except e:
print(f"Got error: {str(e)}")
produce_thread = threading.Thread(target=produce)
produce_thread.start()
instance.query(
"""
CREATE MATERIALIZED VIEW test.consumer TO test.view AS
SELECT * FROM test.rabbitmq;
"""
)
while True:
time.sleep(0.2)
count = instance.query(
"SELECT count() FROM system.parts WHERE database = 'test' AND table = 'view'"
)
print(f"kssenii total count: {count}")
count = int(
instance.query(
"SELECT count() FROM system.parts WHERE database = 'test' AND table = 'view' AND name = 'all_1_1_0'"
)
)
print(f"kssenii count: {count}")
if count > 0:
break
time.sleep(12)
result = instance.query("SELECT uniqExact(ts) FROM test.view")
cancel.set()
produce_thread.join()
instance.query(
"""
DROP TABLE test.consumer;
DROP TABLE test.view;
DROP TABLE test.rabbitmq;
"""
)
2023-01-06 15:53:24 +00:00
assert int(result) == 3