2020-09-16 04:26:10 +00:00
import json
2024-09-27 10:19:39 +00:00
import logging
import math
2020-05-20 06:22:12 +00:00
import os . path as p
import random
2020-09-16 04:26:10 +00:00
import subprocess
2020-05-20 06:22:12 +00:00
import threading
import time
from random import randrange
2020-09-16 04:26:10 +00:00
import pika
2024-09-27 10:19:39 +00:00
import pytest
2020-09-16 04:26:10 +00:00
from google . protobuf . internal . encoder import _VarintBytes
2024-09-27 10:19:39 +00:00
2020-09-16 04:26:10 +00:00
from helpers . client import QueryRuntimeException
2022-05-05 12:25:42 +00:00
from helpers . cluster import ClickHouseCluster , check_rabbitmq_is_available
2020-05-20 06:22:12 +00:00
from helpers . test_tools import TSV
2020-10-02 16:54:07 +00:00
from . import rabbitmq_pb2
2020-08-15 14:38:29 +00:00
2020-05-20 06:22:12 +00:00
cluster = ClickHouseCluster ( __file__ )
instance = cluster . add_instance (
" instance " ,
2021-11-23 14:52:25 +00:00
main_configs = [
" configs/rabbitmq.xml " ,
" configs/macros.xml " ,
" configs/named_collection.xml " ,
] ,
2021-11-29 13:13:20 +00:00
user_configs = [ " configs/users.xml " ] ,
2020-09-11 16:16:24 +00:00
with_rabbitmq = True ,
2022-05-31 12:38:11 +00:00
stay_alive = True ,
2020-09-11 16:16:24 +00:00
)
2020-05-20 06:22:12 +00:00
2022-06-24 00:07:16 +00:00
instance2 = cluster . add_instance (
" instance2 " ,
user_configs = [ " configs/users.xml " ] ,
with_rabbitmq = True ,
)
2020-05-20 06:22:12 +00:00
2024-02-14 14:09:41 +00:00
instance3 = cluster . add_instance (
" instance3 " ,
user_configs = [ " configs/users.xml " ] ,
main_configs = [
" configs/rabbitmq.xml " ,
" configs/macros.xml " ,
" configs/named_collection.xml " ,
" configs/mergetree.xml " ,
] ,
with_rabbitmq = True ,
2024-02-19 19:52:36 +00:00
stay_alive = True ,
2024-02-14 14:09:41 +00:00
)
2020-05-20 06:22:12 +00:00
# Helpers
2024-03-13 16:41:35 +00:00
def rabbitmq_check_result ( result , check = False , reference = None ) :
if reference is None :
reference = " \n " . join ( [ f " { i } \t { i } " for i in range ( 50 ) ] )
if check :
assert TSV ( result ) == TSV ( reference )
else :
return TSV ( result ) == TSV ( reference )
2020-05-20 06:22:12 +00:00
2022-03-22 16:39:58 +00:00
2023-10-09 18:33:03 +00:00
def wait_rabbitmq_to_start ( rabbitmq_docker_id , cookie , timeout = 180 ) :
2022-12-20 18:32:59 +00:00
logging . getLogger ( " pika " ) . propagate = False
2021-12-28 19:16:16 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2023-10-09 18:33:03 +00:00
if check_rabbitmq_is_available ( rabbitmq_docker_id , cookie ) :
2021-12-28 19:16:16 +00:00
logging . debug ( " RabbitMQ is available " )
return
time . sleep ( 0.5 )
except Exception as ex :
logging . debug ( " Can ' t connect to RabbitMQ " + str ( ex ) )
time . sleep ( 0.5 )
2020-05-20 06:22:12 +00:00
2022-03-22 16:39:58 +00:00
2021-04-30 09:18:12 +00:00
def kill_rabbitmq ( rabbitmq_id ) :
2020-08-08 16:45:52 +00:00
p = subprocess . Popen ( ( " docker " , " stop " , rabbitmq_id ) , stdout = subprocess . PIPE )
2024-07-11 13:59:34 +00:00
p . wait ( timeout = 60 )
2020-08-08 16:45:52 +00:00
return p . returncode == 0
2023-10-09 18:33:03 +00:00
def revive_rabbitmq ( rabbitmq_id , cookie ) :
2020-08-08 16:45:52 +00:00
p = subprocess . Popen ( ( " docker " , " start " , rabbitmq_id ) , stdout = subprocess . PIPE )
2024-07-11 13:59:34 +00:00
p . wait ( timeout = 60 )
2023-10-09 18:33:03 +00:00
wait_rabbitmq_to_start ( rabbitmq_id , cookie )
2020-08-08 16:45:52 +00:00
2020-05-20 06:22:12 +00:00
# Fixtures
2022-03-22 16:39:58 +00:00
2020-05-20 06:22:12 +00:00
@pytest.fixture ( scope = " module " )
def rabbitmq_cluster ( ) :
try :
cluster . start ( )
2021-04-30 09:18:12 +00:00
logging . debug ( " rabbitmq_id is {} " . format ( instance . cluster . rabbitmq_docker_id ) )
2020-05-20 06:22:12 +00:00
instance . query ( " CREATE DATABASE test " )
2024-02-14 14:09:41 +00:00
instance3 . query ( " CREATE DATABASE test " )
2020-05-20 06:22:12 +00:00
yield cluster
finally :
cluster . shutdown ( )
@pytest.fixture ( autouse = True )
def rabbitmq_setup_teardown ( ) :
2023-10-09 18:33:03 +00:00
logging . debug ( " RabbitMQ is available - running test " )
2020-05-20 06:22:12 +00:00
yield # run test
2023-05-03 18:06:46 +00:00
instance . query ( " DROP DATABASE test SYNC " )
2021-12-20 21:03:43 +00:00
instance . query ( " CREATE DATABASE test " )
2020-05-20 06:22:12 +00:00
# Tests
2022-03-22 16:39:58 +00:00
2023-11-14 20:13:09 +00:00
@pytest.mark.parametrize (
" secure " ,
[
pytest . param ( 0 ) ,
pytest . param ( 1 ) ,
] ,
)
def test_rabbitmq_select ( rabbitmq_cluster , secure ) :
2023-11-21 11:26:14 +00:00
if secure and instance . is_built_with_thread_sanitizer ( ) :
2023-11-16 16:39:57 +00:00
pytest . skip (
" Data races: see https://github.com/ClickHouse/ClickHouse/issues/56866 "
)
2023-11-16 16:24:09 +00:00
2023-11-16 12:09:13 +00:00
port = cluster . rabbitmq_port
2023-11-14 20:13:09 +00:00
if secure :
2023-11-16 12:09:13 +00:00
port = cluster . rabbitmq_secure_port
2023-11-14 20:13:09 +00:00
2024-03-13 16:41:35 +00:00
# MATERIALIZED and ALIAS columns are not supported in RabbitMQ engine, but we can test that it does not fail
2020-05-20 09:42:56 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2024-03-13 16:41:35 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 , value2 ALIAS value + 1 , value3 MATERIALIZED value + 1 )
2020-05-20 09:42:56 +00:00
ENGINE = RabbitMQ
2023-11-14 20:13:09 +00:00
SETTINGS rabbitmq_host_port = ' {} : {} ' ,
2020-08-31 09:12:36 +00:00
rabbitmq_exchange_name = ' select ' ,
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ,
2023-11-14 20:13:09 +00:00
rabbitmq_row_delimiter = ' \\ n ' ,
rabbitmq_secure = { } ;
2021-02-16 14:16:15 +00:00
""" .format(
2023-11-14 20:26:32 +00:00
rabbitmq_cluster . rabbitmq_host , port , secure
2021-02-16 14:16:15 +00:00
)
2022-03-22 16:39:58 +00:00
)
2024-03-13 16:41:35 +00:00
assert (
" RabbitMQ table engine doesn \\ ' t support ALIAS, DEFAULT or MATERIALIZED columns "
in instance . query ( " SELECT * FROM system.warnings " )
)
2020-05-20 09:42:56 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for i in range ( 50 ) :
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
for message in messages :
2020-08-31 09:12:36 +00:00
channel . basic_publish ( exchange = " select " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
connection . close ( )
2020-09-07 10:21:29 +00:00
# The order of messages in select * from test.rabbitmq is not guaranteed, so sleep to collect everything in one select
time . sleep ( 1 )
2020-06-08 01:11:48 +00:00
result = " "
while True :
2020-09-07 10:21:29 +00:00
result + = instance . query (
" SELECT * FROM test.rabbitmq ORDER BY key " , ignore_error = True
)
2020-06-08 01:11:48 +00:00
if rabbitmq_check_result ( result ) :
break
2020-05-20 09:42:56 +00:00
rabbitmq_check_result ( result , True )
def test_rabbitmq_select_empty ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
2021-02-16 14:16:15 +00:00
SETTINGS rabbitmq_host_port = ' {} :5672 ' ,
2020-08-31 09:12:36 +00:00
rabbitmq_exchange_name = ' empty ' ,
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' TSV ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
2021-02-16 14:16:15 +00:00
""" .format(
rabbitmq_cluster . rabbitmq_host
)
2022-03-22 16:39:58 +00:00
)
2020-05-20 09:42:56 +00:00
assert int ( instance . query ( " SELECT count() FROM test.rabbitmq " ) ) == 0
def test_rabbitmq_json_without_delimiter ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
2021-02-16 14:16:15 +00:00
SETTINGS rabbitmq_host_port = ' {} :5672 ' ,
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-28 08:22:45 +00:00
rabbitmq_exchange_name = ' json ' ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow '
2021-02-16 14:16:15 +00:00
""" .format(
rabbitmq_cluster . rabbitmq_host
)
2022-03-22 16:39:58 +00:00
)
2020-05-20 09:42:56 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = " "
for i in range ( 25 ) :
messages + = json . dumps ( { " key " : i , " value " : i } ) + " \n "
all_messages = [ messages ]
for message in all_messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " json " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
messages = " "
for i in range ( 25 , 50 ) :
messages + = json . dumps ( { " key " : i , " value " : i } ) + " \n "
all_messages = [ messages ]
for message in all_messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " json " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
2020-09-07 10:21:29 +00:00
connection . close ( )
time . sleep ( 1 )
2020-05-20 09:42:56 +00:00
result = " "
while True :
2020-09-07 10:21:29 +00:00
result + = instance . query (
" SELECT * FROM test.rabbitmq ORDER BY key " , ignore_error = True
)
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result ( result ) :
break
rabbitmq_check_result ( result , True )
def test_rabbitmq_csv_with_delimiter ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_exchange_name = ' csv ' ,
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' CSV ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for i in range ( 50 ) :
messages . append ( " {i} , {i} " . format ( i = i ) )
for message in messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " csv " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
2020-09-07 10:21:29 +00:00
connection . close ( )
time . sleep ( 1 )
2020-05-20 09:42:56 +00:00
result = " "
while True :
2020-09-07 10:21:29 +00:00
result + = instance . query (
" SELECT * FROM test.rabbitmq ORDER BY key " , ignore_error = True
)
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result ( result ) :
break
rabbitmq_check_result ( result , True )
def test_rabbitmq_tsv_with_delimiter ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_exchange_name = ' tsv ' ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' TSV ' ,
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2021-03-30 18:48:33 +00:00
rabbitmq_queue_base = ' tsv ' ,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
2021-03-31 07:37:34 +00:00
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for i in range ( 50 ) :
messages . append ( " {i} \t {i} " . format ( i = i ) )
for message in messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " tsv " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
connection . close ( )
2020-08-15 14:38:29 +00:00
result = " "
while True :
2021-03-31 07:37:34 +00:00
result = instance . query ( " SELECT * FROM test.view ORDER BY key " )
2020-08-15 14:38:29 +00:00
if rabbitmq_check_result ( result ) :
break
rabbitmq_check_result ( result , True )
2021-09-12 13:43:22 +00:00
def test_rabbitmq_macros ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-09-12 13:43:22 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' {rabbitmq_host} : {rabbitmq_port} ' ,
2021-11-03 12:43:23 +00:00
rabbitmq_commit_on_select = 1 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2021-09-12 13:43:22 +00:00
rabbitmq_exchange_name = ' {rabbitmq_exchange_name} ' ,
rabbitmq_format = ' {rabbitmq_format} '
2022-03-22 16:39:58 +00:00
"""
2021-09-12 13:43:22 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
message = " "
for i in range ( 50 ) :
message + = json . dumps ( { " key " : i , " value " : i } ) + " \n "
channel . basic_publish ( exchange = " macro " , routing_key = " " , body = message )
2021-10-19 21:40:14 +00:00
2021-09-12 13:43:22 +00:00
connection . close ( )
time . sleep ( 1 )
result = " "
while True :
result + = instance . query (
" SELECT * FROM test.rabbitmq ORDER BY key " , ignore_error = True
)
if rabbitmq_check_result ( result ) :
break
rabbitmq_check_result ( result , True )
2020-05-20 09:42:56 +00:00
def test_rabbitmq_materialized_view ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2024-03-13 16:41:35 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 , dt1 DateTime MATERIALIZED now ( ) , value2 ALIAS value + 1 )
2020-05-20 09:42:56 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_name = ' mv ' ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-05-20 09:42:56 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq ;
2021-12-14 14:19:18 +00:00
CREATE TABLE test . view2 ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer2 TO test . view2 AS
SELECT * FROM test . rabbitmq group by ( key , value ) ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
2023-02-02 11:15:29 +00:00
instance . wait_for_log_line ( " Started streaming to 2 attached views " )
2020-05-20 09:42:56 +00:00
messages = [ ]
for i in range ( 50 ) :
2023-02-02 11:15:29 +00:00
message = json . dumps ( { " key " : i , " value " : i } )
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " mv " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
2021-12-14 14:19:18 +00:00
time_limit_sec = 60
deadline = time . monotonic ( ) + time_limit_sec
while time . monotonic ( ) < deadline :
2020-09-07 10:21:29 +00:00
result = instance . query ( " SELECT * FROM test.view ORDER BY key " )
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result ( result ) :
2020-09-07 11:08:53 +00:00
break
2020-05-20 09:42:56 +00:00
rabbitmq_check_result ( result , True )
2021-12-14 14:19:18 +00:00
deadline = time . monotonic ( ) + time_limit_sec
while time . monotonic ( ) < deadline :
result = instance . query ( " SELECT * FROM test.view2 ORDER BY key " )
2023-10-09 18:33:03 +00:00
logging . debug ( f " Result: { result } " )
2021-12-14 14:19:18 +00:00
if rabbitmq_check_result ( result ) :
break
2023-02-02 11:15:29 +00:00
time . sleep ( 1 )
2021-12-14 14:19:18 +00:00
rabbitmq_check_result ( result , True )
connection . close ( )
2020-05-20 09:42:56 +00:00
def test_rabbitmq_materialized_view_with_subquery ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_name = ' mvsq ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM ( SELECT * FROM test . rabbitmq ) ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for i in range ( 50 ) :
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
for message in messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " mvsq " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
while True :
2020-09-07 10:21:29 +00:00
result = instance . query ( " SELECT * FROM test.view ORDER BY key " )
2020-05-20 09:42:56 +00:00
if rabbitmq_check_result ( result ) :
break
2020-09-07 11:08:53 +00:00
connection . close ( )
2020-05-20 09:42:56 +00:00
rabbitmq_check_result ( result , True )
def test_rabbitmq_many_materialized_views ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
DROP TABLE IF EXISTS test . view1 ;
DROP TABLE IF EXISTS test . view2 ;
2024-03-13 16:41:35 +00:00
DROP TABLE IF EXISTS test . view3 ;
2020-05-20 09:42:56 +00:00
DROP TABLE IF EXISTS test . consumer1 ;
DROP TABLE IF EXISTS test . consumer2 ;
2024-03-13 16:41:35 +00:00
DROP TABLE IF EXISTS test . consumer3 ;
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 , value2 ALIAS value + 1 , value3 MATERIALIZED value + 1 , value4 DEFAULT 1 )
2020-05-20 09:42:56 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_name = ' mmv ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE TABLE test . view1 ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
2024-03-13 16:41:35 +00:00
CREATE TABLE test . view2 ( key UInt64 , value UInt64 , value2 UInt64 , value3 UInt64 , value4 UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE TABLE test . view3 ( key UInt64 )
2020-05-20 09:42:56 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer1 TO test . view1 AS
SELECT * FROM test . rabbitmq ;
CREATE MATERIALIZED VIEW test . consumer2 TO test . view2 AS
SELECT * FROM test . rabbitmq ;
2024-03-13 16:41:35 +00:00
CREATE MATERIALIZED VIEW test . consumer3 TO test . view3 AS
SELECT * FROM test . rabbitmq ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
2024-03-13 16:41:35 +00:00
instance . wait_for_log_line ( " Started streaming to 3 attached views " )
2023-02-02 11:15:29 +00:00
2020-05-20 09:42:56 +00:00
messages = [ ]
for i in range ( 50 ) :
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
for message in messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " mmv " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
2024-03-13 16:41:35 +00:00
is_check_passed = False
deadline = time . monotonic ( ) + 60
while time . monotonic ( ) < deadline :
2020-09-07 10:21:29 +00:00
result1 = instance . query ( " SELECT * FROM test.view1 ORDER BY key " )
result2 = instance . query ( " SELECT * FROM test.view2 ORDER BY key " )
2024-03-13 16:41:35 +00:00
result3 = instance . query ( " SELECT * FROM test.view3 ORDER BY key " )
# Note that for view2 result is `i i 0 0 0`, but not `i i i+1 i+1 1` as expected, ALIAS/MATERIALIZED/DEFAULT columns are not supported in RabbitMQ engine
# We onlt check that at least it do not fail
if (
rabbitmq_check_result ( result1 )
and rabbitmq_check_result (
result2 , reference = " \n " . join ( [ f " { i } \t { i } \t 0 \t 0 \t 0 " for i in range ( 50 ) ] )
)
and rabbitmq_check_result (
result3 , reference = " \n " . join ( [ str ( i ) for i in range ( 50 ) ] )
)
) :
is_check_passed = True
2020-05-20 09:42:56 +00:00
break
2024-03-13 16:41:35 +00:00
time . sleep ( 0.1 )
assert (
is_check_passed
) , f " References are not equal to results, result1: { result1 } , result2: { result2 } , result3: { result3 } "
2020-05-20 09:42:56 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
DROP TABLE test . consumer1 ;
DROP TABLE test . consumer2 ;
2024-03-13 16:41:35 +00:00
DROP TABLE test . consumer3 ;
2020-05-20 09:42:56 +00:00
DROP TABLE test . view1 ;
DROP TABLE test . view2 ;
2024-03-13 16:41:35 +00:00
DROP TABLE test . view3 ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
2020-08-15 14:38:29 +00:00
connection . close ( )
2020-05-20 09:42:56 +00:00
def test_rabbitmq_big_message ( rabbitmq_cluster ) :
# Create batchs of messages of size ~100Kb
rabbitmq_messages = 1000
batch_messages = 1000
messages = [
json . dumps ( { " key " : i , " value " : " x " * 100 } ) * batch_messages
for i in range ( rabbitmq_messages )
]
2022-03-22 16:39:58 +00:00
2020-05-20 09:42:56 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-05-20 09:42:56 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value String )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_name = ' big ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ;
CREATE TABLE test . view ( key UInt64 , value String )
ENGINE = MergeTree
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
for message in messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish ( exchange = " big " , routing_key = " " , body = message )
2020-05-20 09:42:56 +00:00
while True :
result = instance . query ( " SELECT count() FROM test.view " )
if int ( result ) == batch_messages * rabbitmq_messages :
break
connection . close ( )
2020-08-15 06:50:53 +00:00
assert (
int ( result ) == rabbitmq_messages * batch_messages
) , " ClickHouse lost some messages: {} " . format ( result )
2020-05-20 09:42:56 +00:00
def test_rabbitmq_sharding_between_queues_publish ( rabbitmq_cluster ) :
NUM_CONSUMERS = 10
2020-10-27 07:14:38 +00:00
NUM_QUEUES = 10
2022-12-20 18:32:59 +00:00
logging . getLogger ( " pika " ) . propagate = False
2020-05-20 09:42:56 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_name = ' test_sharding ' ,
2022-12-23 10:47:51 +00:00
rabbitmq_num_queues = 5 ,
2020-05-20 09:42:56 +00:00
rabbitmq_num_consumers = 10 ,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100 ,
2023-01-19 17:06:43 +00:00
rabbitmq_flush_interval_ms = 500 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2020-08-15 06:50:53 +00:00
CREATE TABLE test . view ( key UInt64 , value UInt64 , channel_id String )
2020-05-20 09:42:56 +00:00
ENGINE = MergeTree
2020-06-24 21:14:49 +00:00
ORDER BY key
2023-05-22 17:07:18 +00:00
SETTINGS old_parts_lifetime = 5 , cleanup_delay_period = 2 , cleanup_delay_period_random_add = 3 ,
cleanup_thread_preferred_points_per_iteration = 0 ;
2020-05-20 09:42:56 +00:00
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
2020-08-15 06:50:53 +00:00
SELECT * , _channel_id AS channel_id FROM test . rabbitmq ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
i = [ 0 ]
messages_num = 10000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-09-16 04:26:10 +00:00
2020-05-20 09:42:56 +00:00
def produce ( ) :
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
2020-07-24 12:33:07 +00:00
current = 0
2020-05-20 09:42:56 +00:00
for message in messages :
2020-07-24 12:33:07 +00:00
current + = 1
mes_id = str ( current )
2020-09-16 04:26:10 +00:00
channel . basic_publish (
exchange = " test_sharding " ,
routing_key = " " ,
properties = pika . BasicProperties ( message_id = mes_id ) ,
body = message ,
)
2020-05-20 09:42:56 +00:00
connection . close ( )
threads = [ ]
2022-12-23 10:47:51 +00:00
threads_num = 10
2020-05-20 09:42:56 +00:00
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
2020-07-24 12:33:07 +00:00
result1 = " "
2020-05-20 09:42:56 +00:00
while True :
2020-07-24 12:33:07 +00:00
result1 = instance . query ( " SELECT count() FROM test.view " )
2020-05-20 09:42:56 +00:00
time . sleep ( 1 )
2022-12-20 18:32:59 +00:00
expected = messages_num * threads_num
if int ( result1 ) == expected :
2020-05-20 09:42:56 +00:00
break
2023-10-09 18:33:03 +00:00
logging . debug ( f " Result { result1 } / { expected } " )
2020-05-20 09:42:56 +00:00
2020-08-15 06:50:53 +00:00
result2 = instance . query ( " SELECT count(DISTINCT channel_id) FROM test.view " )
2020-07-24 12:33:07 +00:00
2020-05-20 09:42:56 +00:00
for thread in threads :
thread . join ( )
2020-07-24 12:33:07 +00:00
assert (
int ( result1 ) == messages_num * threads_num
2023-12-29 14:02:11 +00:00
) , " ClickHouse lost some messages: {} " . format ( result1 )
2023-01-27 16:22:24 +00:00
assert int ( result2 ) == 10
2020-05-20 09:42:56 +00:00
2020-09-01 14:11:34 +00:00
def test_rabbitmq_mv_combo ( rabbitmq_cluster ) :
2020-09-07 11:08:53 +00:00
NUM_MV = 5
2020-05-20 09:42:56 +00:00
NUM_CONSUMERS = 4
2022-12-20 18:32:59 +00:00
logging . getLogger ( " pika " ) . propagate = False
2020-05-20 09:42:56 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_name = ' combo ' ,
2020-09-01 14:11:34 +00:00
rabbitmq_queue_base = ' combo ' ,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 2 ,
2020-10-27 07:14:38 +00:00
rabbitmq_num_queues = 5 ,
2020-05-20 09:42:56 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-05-20 09:42:56 +00:00
)
for mv_id in range ( NUM_MV ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE IF EXISTS test . combo_ { 0 } ;
DROP TABLE IF EXISTS test . combo_ { 0 } _mv ;
CREATE TABLE test . combo_ { 0 } ( key UInt64 , value UInt64 )
2020-05-20 09:42:56 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2020-08-26 08:54:29 +00:00
CREATE MATERIALIZED VIEW test . combo_ { 0 } _mv TO test . combo_ { 0 } AS
2020-05-20 09:42:56 +00:00
SELECT * FROM test . rabbitmq ;
2020-08-26 08:54:29 +00:00
""" .format(
mv_id
)
2022-03-22 16:39:58 +00:00
)
2020-05-20 09:42:56 +00:00
time . sleep ( 2 )
i = [ 0 ]
messages_num = 10000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-09-16 04:26:10 +00:00
2020-05-20 09:42:56 +00:00
def produce ( ) :
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
2020-07-28 08:22:45 +00:00
channel . basic_publish (
exchange = " combo " ,
routing_key = " " ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( message_id = str ( msg_id ) ) ,
body = messages [ msg_id ] ,
)
2020-05-20 09:42:56 +00:00
connection . close ( )
threads = [ ]
threads_num = 20
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
while True :
result = 0
2020-08-26 08:54:29 +00:00
for mv_id in range ( NUM_MV ) :
result + = int (
instance . query ( " SELECT count() FROM test.combo_ {0} " . format ( mv_id ) )
2022-03-22 16:39:58 +00:00
)
2022-12-20 18:32:59 +00:00
expected = messages_num * threads_num * NUM_MV
if int ( result ) == expected :
2020-05-20 09:42:56 +00:00
break
2023-10-09 18:33:03 +00:00
logging . debug ( f " Result: { result } / { expected } " )
2020-05-20 09:42:56 +00:00
time . sleep ( 1 )
for thread in threads :
thread . join ( )
for mv_id in range ( NUM_MV ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . combo_ { 0 } _mv ;
2020-09-07 10:21:29 +00:00
DROP TABLE test . combo_ { 0 } ;
2020-08-26 08:54:29 +00:00
""" .format(
mv_id
)
2022-03-22 16:39:58 +00:00
)
2020-05-20 09:42:56 +00:00
assert (
int ( result ) == messages_num * threads_num * NUM_MV
) , " ClickHouse lost some messages: {} " . format ( result )
2020-05-20 06:22:12 +00:00
2020-06-01 16:19:59 +00:00
def test_rabbitmq_insert ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-06-13 18:44:17 +00:00
rabbitmq_exchange_name = ' insert ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-20 06:21:18 +00:00
rabbitmq_exchange_type = ' direct ' ,
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = ' insert1 ' ,
2020-06-01 16:19:59 +00:00
rabbitmq_format = ' TSV ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-01 16:19:59 +00:00
consumer_connection = pika . BlockingConnection ( parameters )
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
2020-07-20 06:21:18 +00:00
consumer . queue_bind ( exchange = " insert " , queue = queue_name , routing_key = " insert1 " )
2020-06-01 16:19:59 +00:00
values = [ ]
for i in range ( 50 ) :
values . append ( " ( {i} , {i} ) " . format ( i = i ) )
values = " , " . join ( values )
while True :
try :
instance . query ( " INSERT INTO test.rabbitmq VALUES {} " . format ( values ) )
break
except QueryRuntimeException as e :
if " Local: Timed out. " in str ( e ) :
continue
else :
raise
insert_messages = [ ]
2020-09-16 04:26:10 +00:00
2020-06-01 16:19:59 +00:00
def onReceived ( channel , method , properties , body ) :
i = 0
insert_messages . append ( body . decode ( ) )
if len ( insert_messages ) == 50 :
channel . stop_consuming ( )
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2020-06-01 16:19:59 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
result = " \n " . join ( insert_messages )
rabbitmq_check_result ( result , True )
2020-07-20 06:21:18 +00:00
def test_rabbitmq_insert_headers_exchange ( rabbitmq_cluster ) :
2020-06-01 16:19:59 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-07-20 06:21:18 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
2020-06-01 16:19:59 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-20 06:21:18 +00:00
rabbitmq_exchange_name = ' insert_headers ' ,
rabbitmq_exchange_type = ' headers ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-20 06:21:18 +00:00
rabbitmq_routing_key_list = ' test=insert,topic=headers ' ,
2020-06-01 16:19:59 +00:00
rabbitmq_format = ' TSV ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
)
2020-07-20 06:21:18 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-07-20 06:21:18 +00:00
consumer_connection = pika . BlockingConnection ( parameters )
2020-06-01 16:19:59 +00:00
2020-07-20 06:21:18 +00:00
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
consumer . queue_bind (
exchange = " insert_headers " ,
queue = queue_name ,
routing_key = " " ,
2020-09-16 04:26:10 +00:00
arguments = { " x-match " : " all " , " test " : " insert " , " topic " : " headers " } ,
)
2020-06-01 16:19:59 +00:00
2020-07-20 06:21:18 +00:00
values = [ ]
for i in range ( 50 ) :
values . append ( " ( {i} , {i} ) " . format ( i = i ) )
values = " , " . join ( values )
2020-06-01 16:19:59 +00:00
while True :
2020-07-20 06:21:18 +00:00
try :
instance . query ( " INSERT INTO test.rabbitmq VALUES {} " . format ( values ) )
2020-06-01 16:19:59 +00:00
break
2020-07-20 06:21:18 +00:00
except QueryRuntimeException as e :
if " Local: Timed out. " in str ( e ) :
continue
else :
raise
2020-06-01 16:19:59 +00:00
2020-07-20 06:21:18 +00:00
insert_messages = [ ]
2020-09-16 04:26:10 +00:00
2020-07-20 06:21:18 +00:00
def onReceived ( channel , method , properties , body ) :
i = 0
insert_messages . append ( body . decode ( ) )
if len ( insert_messages ) == 50 :
channel . stop_consuming ( )
2020-06-01 16:19:59 +00:00
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2020-07-20 06:21:18 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
2020-06-01 16:19:59 +00:00
2020-07-20 06:21:18 +00:00
result = " \n " . join ( insert_messages )
rabbitmq_check_result ( result , True )
2020-06-01 16:19:59 +00:00
2020-07-20 06:21:18 +00:00
def test_rabbitmq_many_inserts ( rabbitmq_cluster ) :
2020-06-01 16:19:59 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-07-20 06:21:18 +00:00
DROP TABLE IF EXISTS test . rabbitmq_many ;
2020-07-23 11:45:01 +00:00
DROP TABLE IF EXISTS test . rabbitmq_consume ;
2020-07-20 06:21:18 +00:00
DROP TABLE IF EXISTS test . view_many ;
DROP TABLE IF EXISTS test . consumer_many ;
CREATE TABLE test . rabbitmq_many ( key UInt64 , value UInt64 )
2020-06-01 16:19:59 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-20 06:21:18 +00:00
rabbitmq_exchange_name = ' many_inserts ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' insert2 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-01 16:19:59 +00:00
rabbitmq_format = ' TSV ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2020-07-21 15:47:39 +00:00
CREATE TABLE test . rabbitmq_consume ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' many_inserts ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' insert2 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-21 15:47:39 +00:00
rabbitmq_format = ' TSV ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
)
2021-06-27 19:05:20 +00:00
messages_num = 10000
values = [ ]
for i in range ( messages_num ) :
values . append ( " ( {i} , {i} ) " . format ( i = i ) )
values = " , " . join ( values )
2020-09-16 04:26:10 +00:00
2020-06-01 16:19:59 +00:00
def insert ( ) :
while True :
try :
2020-07-20 06:21:18 +00:00
instance . query (
" INSERT INTO test.rabbitmq_many VALUES {} " . format ( values )
2022-03-22 16:39:58 +00:00
)
2020-06-01 16:19:59 +00:00
break
except QueryRuntimeException as e :
if " Local: Timed out. " in str ( e ) :
continue
else :
raise
threads = [ ]
2021-06-27 19:05:20 +00:00
threads_num = 10
2020-06-01 16:19:59 +00:00
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = insert ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
2021-06-27 19:05:20 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-06-27 19:05:20 +00:00
CREATE TABLE test . view_many ( key UInt64 , value UInt64 )
ENGINE = MergeTree
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer_many TO test . view_many AS
SELECT * FROM test . rabbitmq_consume ;
2022-03-22 16:39:58 +00:00
"""
2021-06-27 19:05:20 +00:00
)
for thread in threads :
thread . join ( )
2020-06-01 16:19:59 +00:00
while True :
2020-07-20 06:21:18 +00:00
result = instance . query ( " SELECT count() FROM test.view_many " )
2023-10-09 18:33:03 +00:00
logging . debug ( result , messages_num * threads_num )
2020-06-01 16:19:59 +00:00
if int ( result ) == messages_num * threads_num :
break
2021-06-27 19:05:20 +00:00
time . sleep ( 1 )
2020-06-01 16:19:59 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . rabbitmq_consume ;
DROP TABLE test . rabbitmq_many ;
DROP TABLE test . consumer_many ;
DROP TABLE test . view_many ;
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
)
assert (
int ( result ) == messages_num * threads_num
) , " ClickHouse lost some messages: {} " . format ( result )
def test_rabbitmq_overloaded_insert ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
DROP TABLE IF EXISTS test . view_overload ;
DROP TABLE IF EXISTS test . consumer_overload ;
2020-07-23 11:45:01 +00:00
DROP TABLE IF EXISTS test . rabbitmq_consume ;
2020-07-21 15:47:39 +00:00
CREATE TABLE test . rabbitmq_consume ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' over ' ,
2020-09-01 14:11:34 +00:00
rabbitmq_queue_base = ' over ' ,
2020-07-21 15:47:39 +00:00
rabbitmq_exchange_type = ' direct ' ,
2023-02-07 11:36:25 +00:00
rabbitmq_num_consumers = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2023-02-20 15:53:06 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-21 15:47:39 +00:00
rabbitmq_routing_key_list = ' over ' ,
rabbitmq_format = ' TSV ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2020-06-01 16:19:59 +00:00
CREATE TABLE test . rabbitmq_overload ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-20 06:21:18 +00:00
rabbitmq_exchange_name = ' over ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' over ' ,
2020-06-01 16:19:59 +00:00
rabbitmq_format = ' TSV ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE TABLE test . view_overload ( key UInt64 , value UInt64 )
ENGINE = MergeTree
2023-02-07 11:36:25 +00:00
ORDER BY key ;
2020-06-01 16:19:59 +00:00
CREATE MATERIALIZED VIEW test . consumer_overload TO test . view_overload AS
2020-07-21 15:47:39 +00:00
SELECT * FROM test . rabbitmq_consume ;
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
)
2023-02-02 19:16:53 +00:00
instance . wait_for_log_line ( " Started streaming to 1 attached views " )
2020-06-01 16:19:59 +00:00
messages_num = 100000
2020-09-16 04:26:10 +00:00
2020-06-01 16:19:59 +00:00
def insert ( ) :
values = [ ]
for i in range ( messages_num ) :
values . append ( " ( {i} , {i} ) " . format ( i = i ) )
values = " , " . join ( values )
while True :
try :
instance . query (
" INSERT INTO test.rabbitmq_overload VALUES {} " . format ( values )
2022-03-22 16:39:58 +00:00
)
2020-06-01 16:19:59 +00:00
break
except QueryRuntimeException as e :
if " Local: Timed out. " in str ( e ) :
continue
else :
raise
threads = [ ]
2023-02-07 11:36:25 +00:00
threads_num = 2
2020-06-01 16:19:59 +00:00
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = insert ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
2023-02-20 12:55:27 +00:00
for thread in threads :
thread . join ( )
2020-06-01 16:19:59 +00:00
while True :
result = instance . query ( " SELECT count() FROM test.view_overload " )
2022-12-20 18:32:59 +00:00
expected = messages_num * threads_num
if int ( result ) == expected :
2020-06-01 16:19:59 +00:00
break
2023-10-09 18:33:03 +00:00
logging . debug ( f " Result: { result } / { expected } " )
2023-02-07 11:36:25 +00:00
time . sleep ( 1 )
2020-06-01 16:19:59 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2023-05-03 18:06:46 +00:00
DROP TABLE test . consumer_overload SYNC ;
DROP TABLE test . view_overload SYNC ;
DROP TABLE test . rabbitmq_consume SYNC ;
DROP TABLE test . rabbitmq_overload SYNC ;
2022-03-22 16:39:58 +00:00
"""
2020-06-01 16:19:59 +00:00
)
assert (
int ( result ) == messages_num * threads_num
) , " ClickHouse lost some messages: {} " . format ( result )
2020-06-10 23:01:47 +00:00
def test_rabbitmq_direct_exchange ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-07-28 08:22:45 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 )
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree ( )
2020-06-25 09:44:39 +00:00
ORDER BY key
2023-05-22 17:07:18 +00:00
SETTINGS old_parts_lifetime = 5 , cleanup_delay_period = 2 , cleanup_delay_period_random_add = 3 ,
cleanup_thread_preferred_points_per_iteration = 0 ;
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
)
num_tables = 5
for consumer_id in range ( num_tables ) :
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up table {} " . format ( consumer_id ) ) )
2020-06-10 23:01:47 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . direct_exchange_ { 0 } ;
DROP TABLE IF EXISTS test . direct_exchange_ { 0 } _mv ;
CREATE TABLE test . direct_exchange_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 2 ,
rabbitmq_num_queues = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = ' direct_exchange_testing ' ,
rabbitmq_exchange_type = ' direct ' ,
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = ' direct_ {0} ' ,
2020-06-10 23:01:47 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . direct_exchange_ { 0 } _mv TO test . destination AS
2020-07-28 08:22:45 +00:00
SELECT key , value FROM test . direct_exchange_ { 0 } ;
2020-06-10 23:01:47 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-10 23:01:47 +00:00
i = [ 0 ]
messages_num = 1000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-10 23:01:47 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
key_num = 0
for num in range ( num_tables ) :
key = " direct_ " + str ( key_num )
key_num + = 1
for message in messages :
2020-06-14 16:26:37 +00:00
mes_id = str ( randrange ( 10 ) )
channel . basic_publish (
2020-09-16 04:26:10 +00:00
exchange = " direct_exchange_testing " ,
routing_key = key ,
properties = pika . BasicProperties ( message_id = mes_id ) ,
body = message ,
)
2020-06-10 23:01:47 +00:00
connection . close ( )
while True :
result = instance . query ( " SELECT count() FROM test.destination " )
time . sleep ( 1 )
if int ( result ) == messages_num * num_tables :
break
2020-06-13 18:44:17 +00:00
for consumer_id in range ( num_tables ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . direct_exchange_ { 0 } _mv ;
2020-09-07 10:21:29 +00:00
DROP TABLE test . direct_exchange_ { 0 } ;
2020-06-13 18:44:17 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-13 18:44:17 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-13 18:44:17 +00:00
DROP TABLE IF EXISTS test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-06-13 18:44:17 +00:00
)
2020-06-10 23:01:47 +00:00
assert (
int ( result ) == messages_num * num_tables
) , " ClickHouse lost some messages: {} " . format ( result )
def test_rabbitmq_fanout_exchange ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-07-28 08:22:45 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 )
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
)
num_tables = 5
for consumer_id in range ( num_tables ) :
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up table {} " . format ( consumer_id ) ) )
2020-06-10 23:01:47 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . fanout_exchange_ { 0 } ;
DROP TABLE IF EXISTS test . fanout_exchange_ { 0 } _mv ;
CREATE TABLE test . fanout_exchange_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 2 ,
rabbitmq_num_queues = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = ' key_ {0} ' ,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = ' fanout_exchange_testing ' ,
rabbitmq_exchange_type = ' fanout ' ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . fanout_exchange_ { 0 } _mv TO test . destination AS
2020-07-28 08:22:45 +00:00
SELECT key , value FROM test . fanout_exchange_ { 0 } ;
2020-06-10 23:01:47 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-10 23:01:47 +00:00
i = [ 0 ]
messages_num = 1000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-10 23:01:47 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
2020-07-24 12:33:07 +00:00
channel . basic_publish (
exchange = " fanout_exchange_testing " ,
routing_key = " " ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( message_id = str ( msg_id ) ) ,
body = messages [ msg_id ] ,
)
2020-06-10 23:01:47 +00:00
connection . close ( )
while True :
result = instance . query ( " SELECT count() FROM test.destination " )
time . sleep ( 1 )
if int ( result ) == messages_num * num_tables :
break
2020-06-13 18:44:17 +00:00
for consumer_id in range ( num_tables ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . fanout_exchange_ { 0 } _mv ;
2020-09-07 10:21:29 +00:00
DROP TABLE test . fanout_exchange_ { 0 } ;
2020-06-13 18:44:17 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-13 18:44:17 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-06-13 18:44:17 +00:00
)
2020-06-10 23:01:47 +00:00
assert (
int ( result ) == messages_num * num_tables
) , " ClickHouse lost some messages: {} " . format ( result )
def test_rabbitmq_topic_exchange ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-07-28 08:22:45 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 )
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
)
num_tables = 5
for consumer_id in range ( num_tables ) :
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up table {} " . format ( consumer_id ) ) )
2020-06-10 23:01:47 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . topic_exchange_ { 0 } ;
DROP TABLE IF EXISTS test . topic_exchange_ { 0 } _mv ;
CREATE TABLE test . topic_exchange_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 2 ,
rabbitmq_num_queues = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = ' topic_exchange_testing ' ,
rabbitmq_exchange_type = ' topic ' ,
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = ' *. {0} ' ,
2020-06-10 23:01:47 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . topic_exchange_ { 0 } _mv TO test . destination AS
2020-07-28 08:22:45 +00:00
SELECT key , value FROM test . topic_exchange_ { 0 } ;
2020-06-10 23:01:47 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-10 23:01:47 +00:00
for consumer_id in range ( num_tables ) :
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up table {} " . format ( num_tables + consumer_id ) ) )
2020-06-10 23:01:47 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . topic_exchange_ { 0 } ;
DROP TABLE IF EXISTS test . topic_exchange_ { 0 } _mv ;
CREATE TABLE test . topic_exchange_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 2 ,
rabbitmq_num_queues = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_name = ' topic_exchange_testing ' ,
rabbitmq_exchange_type = ' topic ' ,
2020-06-11 09:23:23 +00:00
rabbitmq_routing_key_list = ' *.logs ' ,
2020-06-10 23:01:47 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . topic_exchange_ { 0 } _mv TO test . destination AS
2020-07-28 08:22:45 +00:00
SELECT key , value FROM test . topic_exchange_ { 0 } ;
2020-06-10 23:01:47 +00:00
""" .format(
num_tables + consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-10 23:01:47 +00:00
i = [ 0 ]
messages_num = 1000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-10 23:01:47 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
key_num = 0
for num in range ( num_tables ) :
key = " topic. " + str ( key_num )
key_num + = 1
for message in messages :
channel . basic_publish (
exchange = " topic_exchange_testing " , routing_key = key , body = message
)
key = " random.logs "
2020-07-24 12:33:07 +00:00
current = 0
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
2020-07-24 12:33:07 +00:00
channel . basic_publish (
exchange = " topic_exchange_testing " ,
routing_key = key ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( message_id = str ( msg_id ) ) ,
body = messages [ msg_id ] ,
)
2020-06-10 23:01:47 +00:00
connection . close ( )
while True :
result = instance . query ( " SELECT count() FROM test.destination " )
time . sleep ( 1 )
if int ( result ) == messages_num * num_tables + messages_num * num_tables :
break
2020-06-14 16:26:37 +00:00
for consumer_id in range ( num_tables * 2 ) :
2020-06-13 18:44:17 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . topic_exchange_ { 0 } _mv ;
2020-09-07 10:21:29 +00:00
DROP TABLE test . topic_exchange_ { 0 } ;
2020-06-13 18:44:17 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-13 18:44:17 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-06-13 18:44:17 +00:00
)
2020-09-16 04:26:10 +00:00
assert (
int ( result ) == messages_num * num_tables + messages_num * num_tables
) , " ClickHouse lost some messages: {} " . format ( result )
2020-06-10 23:01:47 +00:00
def test_rabbitmq_hash_exchange ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-08-15 06:50:53 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 , channel_id String )
2020-06-10 23:01:47 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
)
num_tables = 4
for consumer_id in range ( num_tables ) :
table_name = " rabbitmq_consumer {} " . format ( consumer_id )
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up {} " . format ( table_name ) ) )
2020-06-10 23:01:47 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
DROP TABLE IF EXISTS test . { 0 } ;
DROP TABLE IF EXISTS test . { 0 } _mv ;
CREATE TABLE test . { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 4 ,
rabbitmq_num_queues = 2 ,
2020-06-10 23:01:47 +00:00
rabbitmq_exchange_type = ' consistent_hash ' ,
rabbitmq_exchange_name = ' hash_exchange_testing ' ,
rabbitmq_format = ' JSONEachRow ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2020-06-10 23:01:47 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . { 0 } _mv TO test . destination AS
2020-08-15 06:50:53 +00:00
SELECT key , value , _channel_id AS channel_id FROM test . { 0 } ;
2020-06-10 23:01:47 +00:00
""" .format(
table_name
)
2022-03-22 16:39:58 +00:00
)
2020-06-10 23:01:47 +00:00
i = [ 0 ]
messages_num = 500
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-10 23:01:47 +00:00
def produce ( ) :
# init connection here because otherwise python rabbitmq client might fail
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
channel . basic_publish (
exchange = " hash_exchange_testing " ,
routing_key = str ( msg_id ) ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( message_id = str ( msg_id ) ) ,
body = messages [ msg_id ] ,
)
2020-06-10 23:01:47 +00:00
connection . close ( )
threads = [ ]
threads_num = 10
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
2020-07-21 15:47:39 +00:00
result1 = " "
2020-06-10 23:01:47 +00:00
while True :
2020-07-21 15:47:39 +00:00
result1 = instance . query ( " SELECT count() FROM test.destination " )
2020-06-10 23:01:47 +00:00
time . sleep ( 1 )
2020-07-21 15:47:39 +00:00
if int ( result1 ) == messages_num * threads_num :
2020-06-10 23:01:47 +00:00
break
2020-08-15 06:50:53 +00:00
result2 = instance . query ( " SELECT count(DISTINCT channel_id) FROM test.destination " )
2020-06-10 23:01:47 +00:00
for consumer_id in range ( num_tables ) :
table_name = " rabbitmq_consumer {} " . format ( consumer_id )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . { 0 } _mv ;
2020-09-07 10:21:29 +00:00
DROP TABLE test . { 0 } ;
2020-06-10 23:01:47 +00:00
""" .format(
table_name
)
2022-03-22 16:39:58 +00:00
)
2020-06-10 23:01:47 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-06-10 23:01:47 +00:00
)
for thread in threads :
thread . join ( )
2020-07-21 15:47:39 +00:00
assert (
int ( result1 ) == messages_num * threads_num
2023-12-29 14:02:11 +00:00
) , " ClickHouse lost some messages: {} " . format ( result1 )
2020-07-28 08:22:45 +00:00
assert int ( result2 ) == 4 * num_tables
2020-07-21 15:47:39 +00:00
2020-06-10 23:01:47 +00:00
2020-06-11 09:23:23 +00:00
def test_rabbitmq_multiple_bindings ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-11 10:56:40 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-07-28 08:22:45 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 )
2020-06-11 10:56:40 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2022-03-22 16:39:58 +00:00
"""
2020-06-11 10:56:40 +00:00
)
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-07-28 08:22:45 +00:00
DROP TABLE IF EXISTS test . bindings ;
DROP TABLE IF EXISTS test . bindings_mv ;
CREATE TABLE test . bindings ( key UInt64 , value UInt64 )
2020-06-11 10:56:40 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' multiple_bindings_testing ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' key1,key2,key3,key4,key5 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-11 10:56:40 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2020-07-28 08:22:45 +00:00
CREATE MATERIALIZED VIEW test . bindings_mv TO test . destination AS
SELECT * FROM test . bindings ;
2022-03-22 16:39:58 +00:00
"""
2020-06-11 10:56:40 +00:00
)
2020-06-11 09:23:23 +00:00
i = [ 0 ]
messages_num = 500
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-11 09:23:23 +00:00
def produce ( ) :
# init connection here because otherwise python rabbitmq client might fail
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
keys = [ " key1 " , " key2 " , " key3 " , " key4 " , " key5 " ]
for key in keys :
for message in messages :
2020-07-28 08:22:45 +00:00
channel . basic_publish (
exchange = " multiple_bindings_testing " , routing_key = key , body = message
)
2020-06-11 09:23:23 +00:00
connection . close ( )
threads = [ ]
threads_num = 10
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
while True :
2020-06-11 10:56:40 +00:00
result = instance . query ( " SELECT count() FROM test.destination " )
2020-06-11 09:23:23 +00:00
time . sleep ( 1 )
2020-07-28 08:22:45 +00:00
if int ( result ) == messages_num * threads_num * 5 :
2020-06-11 09:23:23 +00:00
break
for thread in threads :
thread . join ( )
2020-06-13 18:44:17 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . bindings ;
DROP TABLE test . bindings_mv ;
DROP TABLE test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-06-13 18:44:17 +00:00
)
2020-07-28 08:22:45 +00:00
assert (
int ( result ) == messages_num * threads_num * 5
) , " ClickHouse lost some messages: {} " . format ( result )
2020-06-11 09:23:23 +00:00
2020-06-13 21:37:37 +00:00
def test_rabbitmq_headers_exchange ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-13 21:37:37 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-07-28 08:22:45 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 )
2020-06-13 21:37:37 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2022-03-22 16:39:58 +00:00
"""
2020-06-13 21:37:37 +00:00
)
2020-07-28 08:22:45 +00:00
num_tables_to_receive = 2
2020-06-13 21:37:37 +00:00
for consumer_id in range ( num_tables_to_receive ) :
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up table {} " . format ( consumer_id ) ) )
2020-06-13 21:37:37 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-13 21:37:37 +00:00
DROP TABLE IF EXISTS test . headers_exchange_ { 0 } ;
DROP TABLE IF EXISTS test . headers_exchange_ { 0 } _mv ;
CREATE TABLE test . headers_exchange_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2020-07-28 08:22:45 +00:00
rabbitmq_num_consumers = 2 ,
2020-06-13 21:37:37 +00:00
rabbitmq_exchange_name = ' headers_exchange_testing ' ,
rabbitmq_exchange_type = ' headers ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-13 21:37:37 +00:00
rabbitmq_routing_key_list = ' x-match=all,format=logs,type=report,year=2020 ' ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . headers_exchange_ { 0 } _mv TO test . destination AS
2020-07-28 08:22:45 +00:00
SELECT key , value FROM test . headers_exchange_ { 0 } ;
2020-06-13 21:37:37 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-13 21:37:37 +00:00
num_tables_to_ignore = 2
for consumer_id in range ( num_tables_to_ignore ) :
2023-10-10 08:38:11 +00:00
logging . debug (
( " Setting up table {} " . format ( consumer_id + num_tables_to_receive ) )
)
2020-06-13 21:37:37 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-06-13 21:37:37 +00:00
DROP TABLE IF EXISTS test . headers_exchange_ { 0 } ;
DROP TABLE IF EXISTS test . headers_exchange_ { 0 } _mv ;
CREATE TABLE test . headers_exchange_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' headers_exchange_testing ' ,
rabbitmq_exchange_type = ' headers ' ,
rabbitmq_routing_key_list = ' x-match=all,format=logs,type=report,year=2019 ' ,
rabbitmq_format = ' JSONEachRow ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-06-13 21:37:37 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . headers_exchange_ { 0 } _mv TO test . destination AS
2020-07-28 08:22:45 +00:00
SELECT key , value FROM test . headers_exchange_ { 0 } ;
2020-06-13 21:37:37 +00:00
""" .format(
consumer_id + num_tables_to_receive
)
2022-03-22 16:39:58 +00:00
)
2020-06-13 21:37:37 +00:00
i = [ 0 ]
messages_num = 1000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-06-13 21:37:37 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
2020-09-16 04:26:10 +00:00
fields = { }
fields [ " format " ] = " logs "
fields [ " type " ] = " report "
fields [ " year " ] = " 2020 "
2020-06-13 21:37:37 +00:00
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
2020-06-13 21:37:37 +00:00
channel . basic_publish (
exchange = " headers_exchange_testing " ,
routing_key = " " ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( headers = fields , message_id = str ( msg_id ) ) ,
body = messages [ msg_id ] ,
)
2020-06-13 21:37:37 +00:00
connection . close ( )
while True :
result = instance . query ( " SELECT count() FROM test.destination " )
time . sleep ( 1 )
if int ( result ) == messages_num * num_tables_to_receive :
break
2020-06-14 16:26:37 +00:00
for consumer_id in range ( num_tables_to_receive + num_tables_to_ignore ) :
2020-06-13 21:37:37 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . headers_exchange_ { 0 } _mv ;
2020-09-07 10:21:29 +00:00
DROP TABLE test . headers_exchange_ { 0 } ;
2020-06-14 16:26:37 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-06-13 21:37:37 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-06-13 21:37:37 +00:00
)
assert (
int ( result ) == messages_num * num_tables_to_receive
) , " ClickHouse lost some messages: {} " . format ( result )
2020-07-20 10:05:00 +00:00
def test_rabbitmq_virtual_columns ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-07-21 15:47:39 +00:00
CREATE TABLE test . rabbitmq_virtuals ( key UInt64 , value UInt64 )
2020-07-20 10:05:00 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' virtuals ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-20 10:05:00 +00:00
rabbitmq_format = ' JSONEachRow ' ;
CREATE MATERIALIZED VIEW test . view Engine = Log AS
2020-08-15 06:50:53 +00:00
SELECT value , key , _exchange_name , _channel_id , _delivery_tag , _redelivered FROM test . rabbitmq_virtuals ;
2022-03-22 16:39:58 +00:00
"""
2020-07-20 10:05:00 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-07-20 10:05:00 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
message_num = 10
2020-08-15 06:50:53 +00:00
i = 0
2020-07-20 10:05:00 +00:00
messages = [ ]
for _ in range ( message_num ) :
2020-08-15 06:50:53 +00:00
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
i + = 1
2020-07-20 10:05:00 +00:00
for message in messages :
channel . basic_publish ( exchange = " virtuals " , routing_key = " " , body = message )
while True :
result = instance . query ( " SELECT count() FROM test.view " )
time . sleep ( 1 )
if int ( result ) == message_num :
break
connection . close ( )
result = instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-15 14:38:29 +00:00
SELECT key , value , _exchange_name , SUBSTRING ( _channel_id , 1 , 3 ) , _delivery_tag , _redelivered
2020-08-15 06:50:53 +00:00
FROM test . view ORDER BY key
2022-03-22 16:39:58 +00:00
"""
2020-07-20 10:05:00 +00:00
)
expected = """ \
2020-08-15 06:50:53 +00:00
0 0 virtuals 1_0 1 0
1 1 virtuals 1_0 2 0
2 2 virtuals 1_0 3 0
3 3 virtuals 1_0 4 0
4 4 virtuals 1_0 5 0
5 5 virtuals 1_0 6 0
6 6 virtuals 1_0 7 0
7 7 virtuals 1_0 8 0
8 8 virtuals 1_0 9 0
9 9 virtuals 1_0 10 0
2022-03-22 16:39:58 +00:00
"""
2020-08-15 06:50:53 +00:00
2020-07-21 15:47:39 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . rabbitmq_virtuals ;
DROP TABLE test . view ;
2022-03-22 16:39:58 +00:00
"""
2020-07-21 15:47:39 +00:00
)
2020-08-15 06:50:53 +00:00
2020-07-20 10:05:00 +00:00
assert TSV ( result ) == TSV ( expected )
def test_rabbitmq_virtual_columns_with_materialized_view ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-07-21 15:47:39 +00:00
CREATE TABLE test . rabbitmq_virtuals_mv ( key UInt64 , value UInt64 )
2020-07-20 10:05:00 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' virtuals_mv ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-22 11:24:20 +00:00
rabbitmq_max_block_size = 100 ,
2020-07-20 10:05:00 +00:00
rabbitmq_format = ' JSONEachRow ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 ,
2020-08-15 06:50:53 +00:00
exchange_name String , channel_id String , delivery_tag UInt64 , redelivered UInt8 ) ENGINE = MergeTree ( )
2020-07-20 10:05:00 +00:00
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
2020-08-15 06:50:53 +00:00
SELECT * , _exchange_name as exchange_name , _channel_id as channel_id , _delivery_tag as delivery_tag , _redelivered as redelivered
2020-07-21 15:47:39 +00:00
FROM test . rabbitmq_virtuals_mv ;
2022-03-22 16:39:58 +00:00
"""
2020-07-20 10:05:00 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-07-20 10:05:00 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
message_num = 10
2020-08-15 06:50:53 +00:00
i = 0
2020-07-20 10:05:00 +00:00
messages = [ ]
for _ in range ( message_num ) :
2020-08-15 06:50:53 +00:00
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
i + = 1
2020-07-20 10:05:00 +00:00
for message in messages :
channel . basic_publish ( exchange = " virtuals_mv " , routing_key = " " , body = message )
while True :
result = instance . query ( " SELECT count() FROM test.view " )
time . sleep ( 1 )
if int ( result ) == message_num :
break
connection . close ( )
2020-09-16 04:26:10 +00:00
result = instance . query (
" SELECT key, value, exchange_name, SUBSTRING(channel_id, 1, 3), delivery_tag, redelivered FROM test.view ORDER BY delivery_tag "
)
2020-07-20 10:05:00 +00:00
expected = """ \
2020-08-15 06:50:53 +00:00
0 0 virtuals_mv 1_0 1 0
1 1 virtuals_mv 1_0 2 0
2 2 virtuals_mv 1_0 3 0
3 3 virtuals_mv 1_0 4 0
4 4 virtuals_mv 1_0 5 0
5 5 virtuals_mv 1_0 6 0
6 6 virtuals_mv 1_0 7 0
7 7 virtuals_mv 1_0 8 0
8 8 virtuals_mv 1_0 9 0
9 9 virtuals_mv 1_0 10 0
2022-03-22 16:39:58 +00:00
"""
2020-07-20 10:05:00 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . consumer ;
DROP TABLE test . view ;
DROP TABLE test . rabbitmq_virtuals_mv
2022-03-22 16:39:58 +00:00
"""
2020-07-20 10:05:00 +00:00
)
assert TSV ( result ) == TSV ( expected )
2020-08-06 13:33:46 +00:00
def test_rabbitmq_many_consumers_to_each_queue ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-06 13:33:46 +00:00
DROP TABLE IF EXISTS test . destination ;
2020-08-15 06:50:53 +00:00
CREATE TABLE test . destination ( key UInt64 , value UInt64 , channel_id String )
2020-08-06 13:33:46 +00:00
ENGINE = MergeTree ( )
ORDER BY key ;
2022-03-22 16:39:58 +00:00
"""
2020-08-06 13:33:46 +00:00
)
num_tables = 4
for table_id in range ( num_tables ) :
2023-10-09 18:33:03 +00:00
logging . debug ( ( " Setting up table {} " . format ( table_id ) ) )
2020-08-06 13:33:46 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-06 13:33:46 +00:00
DROP TABLE IF EXISTS test . many_consumers_ { 0 } ;
DROP TABLE IF EXISTS test . many_consumers_ { 0 } _mv ;
CREATE TABLE test . many_consumers_ { 0 } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' many_consumers ' ,
rabbitmq_num_queues = 2 ,
rabbitmq_num_consumers = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100 ,
2020-08-06 13:33:46 +00:00
rabbitmq_queue_base = ' many_consumers ' ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE MATERIALIZED VIEW test . many_consumers_ { 0 } _mv TO test . destination AS
2020-08-15 06:50:53 +00:00
SELECT key , value , _channel_id as channel_id FROM test . many_consumers_ { 0 } ;
2020-08-06 13:33:46 +00:00
""" .format(
table_id
)
2022-03-22 16:39:58 +00:00
)
2020-08-06 13:33:46 +00:00
i = [ 0 ]
messages_num = 1000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-09-16 04:26:10 +00:00
2020-08-06 13:33:46 +00:00
def produce ( ) :
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
2020-08-06 13:33:46 +00:00
channel . basic_publish (
exchange = " many_consumers " ,
routing_key = " " ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( message_id = str ( msg_id ) ) ,
body = messages [ msg_id ] ,
)
2020-08-06 13:33:46 +00:00
connection . close ( )
threads = [ ]
threads_num = 20
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
result1 = " "
while True :
result1 = instance . query ( " SELECT count() FROM test.destination " )
time . sleep ( 1 )
if int ( result1 ) == messages_num * threads_num :
break
2020-08-15 06:50:53 +00:00
result2 = instance . query ( " SELECT count(DISTINCT channel_id) FROM test.destination " )
2020-08-06 13:33:46 +00:00
for thread in threads :
thread . join ( )
for consumer_id in range ( num_tables ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . many_consumers_ { 0 } ;
DROP TABLE test . many_consumers_ { 0 } _mv ;
2020-08-06 13:33:46 +00:00
""" .format(
consumer_id
)
2022-03-22 16:39:58 +00:00
)
2020-08-06 13:33:46 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . destination ;
2022-03-22 16:39:58 +00:00
"""
2020-08-06 13:33:46 +00:00
)
assert (
int ( result1 ) == messages_num * threads_num
2023-12-29 14:02:11 +00:00
) , " ClickHouse lost some messages: {} " . format ( result1 )
2020-08-06 13:33:46 +00:00
# 4 tables, 2 consumers for each table => 8 consumer tags
assert int ( result2 ) == 8
2020-08-26 08:54:29 +00:00
def test_rabbitmq_restore_failed_connection_without_losses_1 ( rabbitmq_cluster ) :
2020-08-08 16:45:52 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-08 16:45:52 +00:00
DROP TABLE IF EXISTS test . consume ;
2020-08-15 06:50:53 +00:00
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree
ORDER BY key ;
2020-08-08 16:45:52 +00:00
CREATE TABLE test . consume ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2023-01-19 17:06:43 +00:00
rabbitmq_flush_interval_ms = 500 ,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100 ,
2020-08-08 16:45:52 +00:00
rabbitmq_exchange_name = ' producer_reconnect ' ,
rabbitmq_format = ' JSONEachRow ' ,
2020-09-01 14:11:34 +00:00
rabbitmq_num_consumers = 2 ,
2020-08-08 16:45:52 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
2020-08-15 06:50:53 +00:00
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . consume ;
2020-08-08 16:45:52 +00:00
DROP TABLE IF EXISTS test . producer_reconnect ;
CREATE TABLE test . producer_reconnect ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' producer_reconnect ' ,
2020-08-31 09:12:36 +00:00
rabbitmq_persistent = ' 1 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2020-08-08 16:45:52 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-08-08 16:45:52 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-08-08 16:45:52 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages_num = 100000
values = [ ]
for i in range ( messages_num ) :
values . append ( " ( {i} , {i} ) " . format ( i = i ) )
values = " , " . join ( values )
while True :
try :
instance . query (
" INSERT INTO test.producer_reconnect VALUES {} " . format ( values )
2022-03-22 16:39:58 +00:00
)
2020-08-08 16:45:52 +00:00
break
except QueryRuntimeException as e :
if " Local: Timed out. " in str ( e ) :
continue
else :
raise
2020-08-15 06:50:53 +00:00
while int ( instance . query ( " SELECT count() FROM test.view " ) ) == 0 :
2020-08-08 16:45:52 +00:00
time . sleep ( 0.1 )
2021-04-30 09:18:12 +00:00
kill_rabbitmq ( rabbitmq_cluster . rabbitmq_docker_id )
2020-09-07 11:08:53 +00:00
time . sleep ( 4 )
2023-10-10 08:38:11 +00:00
revive_rabbitmq (
rabbitmq_cluster . rabbitmq_docker_id , rabbitmq_cluster . rabbitmq_cookie
)
2020-08-08 16:45:52 +00:00
while True :
2020-08-15 06:50:53 +00:00
result = instance . query ( " SELECT count(DISTINCT key) FROM test.view " )
2020-08-08 16:45:52 +00:00
time . sleep ( 1 )
2020-08-15 06:50:53 +00:00
if int ( result ) == messages_num :
2020-08-08 16:45:52 +00:00
break
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . consume ;
DROP TABLE test . producer_reconnect ;
2022-03-22 16:39:58 +00:00
"""
2020-08-15 06:50:53 +00:00
)
assert int ( result ) == messages_num , " ClickHouse lost some messages: {} " . format (
result
)
2020-08-26 08:54:29 +00:00
def test_rabbitmq_restore_failed_connection_without_losses_2 ( rabbitmq_cluster ) :
2022-12-20 18:32:59 +00:00
logging . getLogger ( " pika " ) . propagate = False
2020-08-15 06:50:53 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-15 14:38:29 +00:00
CREATE TABLE test . consumer_reconnect ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' consumer_reconnect ' ,
rabbitmq_num_consumers = 10 ,
2022-12-20 18:32:59 +00:00
rabbitmq_flush_interval_ms = 100 ,
rabbitmq_max_block_size = 100 ,
2020-10-27 07:14:38 +00:00
rabbitmq_num_queues = 10 ,
2020-08-15 14:38:29 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2020-08-15 06:50:53 +00:00
)
2020-08-15 14:38:29 +00:00
i = 0
messages_num = 150000
2020-08-15 06:50:53 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-08-15 14:38:29 +00:00
2020-08-15 06:50:53 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
2020-08-15 14:38:29 +00:00
for _ in range ( messages_num ) :
2020-08-15 06:50:53 +00:00
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
i + = 1
2020-09-07 10:21:29 +00:00
for msg_id in range ( messages_num ) :
channel . basic_publish (
exchange = " consumer_reconnect " ,
routing_key = " " ,
body = messages [ msg_id ] ,
2020-09-16 04:26:10 +00:00
properties = pika . BasicProperties ( delivery_mode = 2 , message_id = str ( msg_id ) ) ,
)
2020-08-15 14:38:29 +00:00
connection . close ( )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-15 14:38:29 +00:00
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . consumer_reconnect ;
2022-03-22 16:39:58 +00:00
"""
2020-08-15 14:38:29 +00:00
)
while int ( instance . query ( " SELECT count() FROM test.view " ) ) == 0 :
2023-10-09 18:33:03 +00:00
logging . debug ( 3 )
2020-08-15 14:38:29 +00:00
time . sleep ( 0.1 )
2021-04-30 09:18:12 +00:00
kill_rabbitmq ( rabbitmq_cluster . rabbitmq_docker_id )
2020-09-07 11:08:53 +00:00
time . sleep ( 8 )
2023-10-10 08:38:11 +00:00
revive_rabbitmq (
rabbitmq_cluster . rabbitmq_docker_id , rabbitmq_cluster . rabbitmq_cookie
)
2020-08-15 14:38:29 +00:00
2020-09-16 04:26:10 +00:00
# while int(instance.query('SELECT count() FROM test.view')) == 0:
2020-08-26 08:54:29 +00:00
# time.sleep(0.1)
2020-08-15 06:50:53 +00:00
2020-09-16 04:26:10 +00:00
# kill_rabbitmq()
# time.sleep(2)
# revive_rabbitmq()
2020-08-15 06:50:53 +00:00
while True :
2022-12-20 18:32:59 +00:00
result = instance . query ( " SELECT count(DISTINCT key) FROM test.view " ) . strip ( )
2020-08-15 14:38:29 +00:00
if int ( result ) == messages_num :
2020-08-15 06:50:53 +00:00
break
2023-10-09 18:33:03 +00:00
logging . debug ( f " Result: { result } / { messages_num } " )
2022-12-20 18:32:59 +00:00
time . sleep ( 1 )
2020-08-15 06:50:53 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-26 08:54:29 +00:00
DROP TABLE test . consumer ;
DROP TABLE test . consumer_reconnect ;
2022-03-22 16:39:58 +00:00
"""
2020-08-08 16:45:52 +00:00
)
2020-08-15 14:38:29 +00:00
assert int ( result ) == messages_num , " ClickHouse lost some messages: {} " . format (
result
)
2020-08-08 16:45:52 +00:00
2020-08-31 16:34:16 +00:00
def test_rabbitmq_commit_on_block_write ( rabbitmq_cluster ) :
2022-12-20 18:32:59 +00:00
logging . getLogger ( " pika " ) . propagate = False
2020-08-31 16:34:16 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-31 16:34:16 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' block ' ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_queue_base = ' block ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2020-08-31 16:34:16 +00:00
rabbitmq_max_block_size = 100 ,
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq ;
2022-03-22 16:39:58 +00:00
"""
2020-08-31 16:34:16 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-08-31 16:34:16 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
cancel = threading . Event ( )
i = [ 0 ]
2020-09-16 04:26:10 +00:00
2020-08-31 16:34:16 +00:00
def produce ( ) :
while not cancel . is_set ( ) :
messages = [ ]
for _ in range ( 101 ) :
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
for message in messages :
channel . basic_publish ( exchange = " block " , routing_key = " " , body = message )
rabbitmq_thread = threading . Thread ( target = produce )
rabbitmq_thread . start ( )
while int ( instance . query ( " SELECT count() FROM test.view " ) ) == 0 :
time . sleep ( 1 )
cancel . set ( )
2021-05-04 19:54:16 +00:00
instance . query ( " DETACH TABLE test.rabbitmq; " )
2020-08-31 16:34:16 +00:00
while (
int (
instance . query (
" SELECT count() FROM system.tables WHERE database= ' test ' AND name= ' rabbitmq ' "
)
2022-03-22 16:39:58 +00:00
)
2020-08-31 16:34:16 +00:00
== 1
) :
time . sleep ( 1 )
2021-05-04 19:54:16 +00:00
instance . query ( " ATTACH TABLE test.rabbitmq; " )
2020-08-31 16:34:16 +00:00
while int ( instance . query ( " SELECT uniqExact(key) FROM test.view " ) ) < i [ 0 ] :
time . sleep ( 1 )
result = int ( instance . query ( " SELECT count() == uniqExact(key) FROM test.view " ) )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-08-31 16:34:16 +00:00
DROP TABLE test . consumer ;
DROP TABLE test . view ;
2022-03-22 16:39:58 +00:00
"""
2020-08-31 16:34:16 +00:00
)
rabbitmq_thread . join ( )
connection . close ( )
assert result == 1 , " Messages from RabbitMQ get duplicated! "
2021-09-16 10:46:43 +00:00
def test_rabbitmq_no_connection_at_startup_1 ( rabbitmq_cluster ) :
2024-07-19 12:10:28 +00:00
error = instance . query_and_get_error (
2022-03-22 16:39:58 +00:00
"""
2021-09-16 10:46:43 +00:00
CREATE TABLE test . cs ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
2024-07-19 12:10:28 +00:00
SETTINGS rabbitmq_host_port = ' no_connection_at_startup:5672 ' ,
2021-09-16 10:46:43 +00:00
rabbitmq_exchange_name = ' cs ' ,
rabbitmq_format = ' JSONEachRow ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-09-16 10:46:43 +00:00
rabbitmq_num_consumers = ' 5 ' ,
rabbitmq_row_delimiter = ' \\ n ' ;
2022-03-22 16:39:58 +00:00
"""
2021-09-16 10:46:43 +00:00
)
2024-07-19 12:10:28 +00:00
assert " CANNOT_CONNECT_RABBITMQ " in error
2021-09-16 10:46:43 +00:00
def test_rabbitmq_no_connection_at_startup_2 ( rabbitmq_cluster ) :
2020-12-02 18:34:01 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-12-02 18:34:01 +00:00
CREATE TABLE test . cs ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' cs ' ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_num_consumers = ' 5 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-12-20 18:32:59 +00:00
rabbitmq_max_block_size = 100 ,
2020-12-02 18:34:01 +00:00
rabbitmq_row_delimiter = ' \\ n ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . cs ;
2022-03-22 16:39:58 +00:00
"""
2020-12-02 18:34:01 +00:00
)
2021-09-16 10:46:43 +00:00
instance . query ( " DETACH TABLE test.cs " )
rabbitmq_cluster . pause_container ( " rabbitmq1 " )
instance . query ( " ATTACH TABLE test.cs " )
2020-12-02 18:34:01 +00:00
rabbitmq_cluster . unpause_container ( " rabbitmq1 " )
messages_num = 1000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2020-12-02 18:34:01 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
for i in range ( messages_num ) :
message = json . dumps ( { " key " : i , " value " : i } )
channel . basic_publish (
exchange = " cs " ,
routing_key = " " ,
body = message ,
properties = pika . BasicProperties ( delivery_mode = 2 , message_id = str ( i ) ) ,
)
connection . close ( )
while True :
result = instance . query ( " SELECT count() FROM test.view " )
time . sleep ( 1 )
if int ( result ) == messages_num :
break
instance . query (
2022-03-22 16:39:58 +00:00
"""
2020-12-02 18:34:01 +00:00
DROP TABLE test . consumer ;
DROP TABLE test . cs ;
2022-03-22 16:39:58 +00:00
"""
2020-12-02 18:34:01 +00:00
)
assert int ( result ) == messages_num , " ClickHouse lost some messages: {} " . format (
result
)
2021-02-15 21:56:51 +00:00
def test_rabbitmq_format_factory_settings ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-02-15 21:56:51 +00:00
CREATE TABLE test . format_settings (
id String , date DateTime
) ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' format_settings ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-02-15 21:56:51 +00:00
rabbitmq_format = ' JSONEachRow ' ,
date_time_input_format = ' best_effort ' ;
2022-03-22 16:39:58 +00:00
"""
2021-02-15 21:56:51 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2021-02-15 21:56:51 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
message = json . dumps (
{ " id " : " format_settings_test " , " date " : " 2021-01-19T14:42:33.1829214Z " }
)
expected = instance . query (
""" SELECT parseDateTimeBestEffort(CAST( ' 2021-01-19T14:42:33.1829214Z ' , ' String ' )) """
)
channel . basic_publish ( exchange = " format_settings " , routing_key = " " , body = message )
result = " "
while True :
result = instance . query ( " SELECT date FROM test.format_settings " )
if result == expected :
break
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-02-15 21:56:51 +00:00
CREATE TABLE test . view (
id String , date DateTime
) ENGINE = MergeTree ORDER BY id ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . format_settings ;
2022-03-22 16:39:58 +00:00
"""
2021-02-15 21:56:51 +00:00
)
channel . basic_publish ( exchange = " format_settings " , routing_key = " " , body = message )
result = " "
while True :
result = instance . query ( " SELECT date FROM test.view " )
if result == expected :
break
connection . close ( )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-02-15 21:56:51 +00:00
DROP TABLE test . consumer ;
DROP TABLE test . format_settings ;
2022-03-22 16:39:58 +00:00
"""
2021-02-15 21:56:51 +00:00
)
assert result == expected
2021-04-21 15:51:05 +00:00
def test_rabbitmq_vhost ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-04-21 15:51:05 +00:00
CREATE TABLE test . rabbitmq_vhost ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' vhost ' ,
rabbitmq_format = ' JSONEachRow ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-04-21 15:51:05 +00:00
rabbitmq_vhost = ' / '
2022-03-22 16:39:58 +00:00
"""
2021-04-21 15:51:05 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-04-30 09:18:12 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2021-04-21 15:51:05 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
channel . basic_publish (
exchange = " vhost " , routing_key = " " , body = json . dumps ( { " key " : 1 , " value " : 2 } )
2022-03-22 16:39:58 +00:00
)
2021-04-21 15:51:05 +00:00
connection . close ( )
while True :
result = instance . query (
" SELECT * FROM test.rabbitmq_vhost ORDER BY key " , ignore_error = True
)
if result == " 1 \t 2 \n " :
break
2021-05-04 16:26:47 +00:00
def test_rabbitmq_drop_table_properly ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-05-04 18:57:49 +00:00
CREATE TABLE test . rabbitmq_drop ( key UInt64 , value UInt64 )
2021-05-04 16:26:47 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-05-04 16:26:47 +00:00
rabbitmq_exchange_name = ' drop ' ,
rabbitmq_format = ' JSONEachRow ' ,
2021-05-13 09:39:57 +00:00
rabbitmq_queue_base = ' rabbit_queue_drop '
2022-03-22 16:39:58 +00:00
"""
2021-05-04 16:26:47 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-05-18 11:02:23 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2021-05-04 16:26:47 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
channel . basic_publish (
exchange = " drop " , routing_key = " " , body = json . dumps ( { " key " : 1 , " value " : 2 } )
2022-03-22 16:39:58 +00:00
)
2021-05-04 16:26:47 +00:00
while True :
2021-05-04 18:57:49 +00:00
result = instance . query (
" SELECT * FROM test.rabbitmq_drop ORDER BY key " , ignore_error = True
)
2021-05-04 16:26:47 +00:00
if result == " 1 \t 2 \n " :
break
2021-05-13 09:39:57 +00:00
exists = channel . queue_declare ( queue = " rabbit_queue_drop " , passive = True )
2021-05-04 16:26:47 +00:00
assert exists
2021-05-04 18:57:49 +00:00
instance . query ( " DROP TABLE test.rabbitmq_drop " )
2021-05-04 16:26:47 +00:00
time . sleep ( 30 )
try :
2023-12-29 14:02:11 +00:00
exists = channel . queue_declare ( queue = " rabbit_queue_drop " , passive = True )
2021-05-04 16:26:47 +00:00
except Exception as e :
exists = False
assert not exists
2021-05-04 18:57:49 +00:00
def test_rabbitmq_queue_settings ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-05-04 18:57:49 +00:00
CREATE TABLE test . rabbitmq_settings ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' rabbit_exchange ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-05-04 18:57:49 +00:00
rabbitmq_format = ' JSONEachRow ' ,
2021-05-13 09:39:57 +00:00
rabbitmq_queue_base = ' rabbit_queue_settings ' ,
2021-05-04 18:57:49 +00:00
rabbitmq_queue_settings_list = ' x-max-length=10,x-overflow=reject-publish '
2022-03-22 16:39:58 +00:00
"""
2021-05-04 18:57:49 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-05-18 11:02:23 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2021-05-04 18:57:49 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
for i in range ( 50 ) :
channel . basic_publish (
exchange = " rabbit_exchange " ,
routing_key = " " ,
body = json . dumps ( { " key " : 1 , " value " : 2 } ) ,
)
connection . close ( )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-05-04 18:57:49 +00:00
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq_settings ;
2022-03-22 16:39:58 +00:00
"""
2021-05-04 18:57:49 +00:00
)
time . sleep ( 5 )
2022-12-20 18:32:59 +00:00
while True :
2021-05-04 18:57:49 +00:00
result = instance . query ( " SELECT count() FROM test.view " , ignore_error = True )
2022-12-20 18:32:59 +00:00
if int ( result ) == 10 :
break
time . sleep ( 0.5 )
2021-05-04 18:57:49 +00:00
2021-05-04 19:57:45 +00:00
instance . query ( " DROP TABLE test.rabbitmq_settings " )
2021-05-04 18:57:49 +00:00
# queue size is 10, but 50 messages were sent, they will be dropped (setting x-overflow = reject-publish) and only 10 will remain.
assert int ( result ) == 10
2021-05-04 19:54:16 +00:00
def test_rabbitmq_queue_consume ( rabbitmq_cluster ) :
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
2021-05-18 11:02:23 +00:00
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
2021-05-04 19:54:16 +00:00
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
channel . queue_declare ( queue = " rabbit_queue " , durable = True )
i = [ 0 ]
messages_num = 1000
2022-03-22 16:39:58 +00:00
2021-05-04 19:54:16 +00:00
def produce ( ) :
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for _ in range ( messages_num ) :
message = json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } )
channel . basic_publish ( exchange = " " , routing_key = " rabbit_queue " , body = message )
i [ 0 ] + = 1
threads = [ ]
threads_num = 10
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-05-04 19:54:16 +00:00
CREATE TABLE test . rabbitmq_queue ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_queue_base = ' rabbit_queue ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-05-04 19:54:16 +00:00
rabbitmq_queue_consume = 1 ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq_queue ;
2022-03-22 16:39:58 +00:00
"""
2021-05-04 19:54:16 +00:00
)
result = " "
while True :
result = instance . query ( " SELECT count() FROM test.view " )
if int ( result ) == messages_num * threads_num :
break
time . sleep ( 1 )
for thread in threads :
thread . join ( )
2021-05-04 19:57:45 +00:00
instance . query ( " DROP TABLE test.rabbitmq_queue " )
2021-05-04 19:54:16 +00:00
2021-10-19 21:40:14 +00:00
def test_rabbitmq_produce_consume_avro ( rabbitmq_cluster ) :
num_rows = 75
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-10-19 21:40:14 +00:00
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . rabbit ;
DROP TABLE IF EXISTS test . rabbit_writer ;
CREATE TABLE test . rabbit_writer ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_format = ' Avro ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-10-19 21:40:14 +00:00
rabbitmq_exchange_name = ' avro ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' avro ' ;
CREATE TABLE test . rabbit ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_format = ' Avro ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-10-19 21:40:14 +00:00
rabbitmq_exchange_name = ' avro ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' avro ' ;
CREATE MATERIALIZED VIEW test . view Engine = Log AS
SELECT key , value FROM test . rabbit ;
2022-03-22 16:39:58 +00:00
"""
2021-10-19 21:40:14 +00:00
)
2020-09-16 04:26:10 +00:00
instance . query (
2021-10-19 21:40:14 +00:00
" INSERT INTO test.rabbit_writer select number*10 as key, number*100 as value from numbers( {num_rows} ) SETTINGS output_format_avro_rows_in_file = 7 " . format (
num_rows = num_rows
2022-03-22 16:39:58 +00:00
)
)
2021-10-19 21:40:14 +00:00
# Ideally we should wait for an event
time . sleep ( 3 )
expected_num_rows = instance . query (
" SELECT COUNT(1) FROM test.view " , ignore_error = True
)
assert int ( expected_num_rows ) == num_rows
expected_max_key = instance . query (
" SELECT max(key) FROM test.view " , ignore_error = True
)
assert int ( expected_max_key ) == ( num_rows - 1 ) * 10
2021-09-16 10:46:43 +00:00
def test_rabbitmq_bad_args ( rabbitmq_cluster ) :
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
channel . exchange_declare ( exchange = " f " , exchange_type = " fanout " )
2024-05-07 18:46:17 +00:00
assert " Unable to declare exchange " in instance . query_and_get_error (
2022-03-22 16:39:58 +00:00
"""
2021-07-19 20:17:17 +00:00
CREATE TABLE test . drop ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-09-16 10:46:43 +00:00
rabbitmq_exchange_name = ' f ' ,
2021-07-19 20:17:17 +00:00
rabbitmq_format = ' JSONEachRow ' ;
2022-03-22 16:39:58 +00:00
"""
2021-07-19 20:17:17 +00:00
)
2021-11-02 12:05:10 +00:00
def test_rabbitmq_issue_30691 ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test . rabbitmq_drop ( json String )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-11-02 12:05:10 +00:00
rabbitmq_exchange_name = ' 30691 ' ,
rabbitmq_row_delimiter = ' \\ n ' , - - Works only if adding this setting
rabbitmq_format = ' LineAsString ' ,
rabbitmq_queue_base = ' 30691 ' ;
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
channel . basic_publish (
exchange = " 30691 " ,
routing_key = " " ,
body = json . dumps (
2022-03-22 16:39:58 +00:00
{
2021-11-02 12:05:10 +00:00
" event_type " : " purge " ,
" as_src " : 1234 ,
" as_dst " : 0 ,
" as_path " : " " ,
" local_pref " : 100 ,
" med " : 0 ,
" peer_as_dst " : 0 ,
" ip_src " : " <redacted ipv6> " ,
" ip_dst " : " <redacted ipv6> " ,
" port_src " : 443 ,
" port_dst " : 41930 ,
" ip_proto " : " tcp " ,
" tos " : 0 ,
" stamp_inserted " : " 2021-10-26 15:20:00 " ,
" stamp_updated " : " 2021-10-26 15:23:14 " ,
" packets " : 2 ,
" bytes " : 1216 ,
" writer_id " : " default_amqp/449206 " ,
2022-03-22 16:39:58 +00:00
}
) ,
2021-11-02 12:05:10 +00:00
)
result = " "
while True :
result = instance . query ( " SELECT * FROM test.rabbitmq_drop " , ignore_error = True )
2023-10-09 18:33:03 +00:00
logging . debug ( result )
2021-11-02 12:05:10 +00:00
if result != " " :
break
assert (
result . strip ( )
== """ { " event_type " : " purge " , " as_src " : 1234, " as_dst " : 0, " as_path " : " " , " local_pref " : 100, " med " : 0, " peer_as_dst " : 0, " ip_src " : " <redacted ipv6> " , " ip_dst " : " <redacted ipv6> " , " port_src " : 443, " port_dst " : 41930, " ip_proto " : " tcp " , " tos " : 0, " stamp_inserted " : " 2021-10-26 15:20:00 " , " stamp_updated " : " 2021-10-26 15:23:14 " , " packets " : 2, " bytes " : 1216, " writer_id " : " default_amqp/449206 " } """
)
def test_rabbitmq_drop_mv ( rabbitmq_cluster ) :
instance . query (
2022-03-22 16:39:58 +00:00
"""
2023-03-27 17:20:48 +00:00
CREATE TABLE test . drop_mv ( key UInt64 , value UInt64 )
2021-11-02 12:05:10 +00:00
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' mv ' ,
rabbitmq_format = ' JSONEachRow ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-11-02 12:05:10 +00:00
rabbitmq_queue_base = ' drop_mv ' ;
2022-07-18 09:30:43 +00:00
"""
)
instance . query (
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
2022-07-18 09:30:43 +00:00
"""
)
instance . query (
"""
2021-11-02 12:05:10 +00:00
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
2023-03-27 17:20:48 +00:00
SELECT * FROM test . drop_mv ;
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
for i in range ( 20 ) :
channel . basic_publish (
exchange = " mv " , routing_key = " " , body = json . dumps ( { " key " : i , " value " : i } )
2022-03-22 16:39:58 +00:00
)
2021-11-02 12:05:10 +00:00
2023-03-27 17:20:48 +00:00
while True :
2022-07-18 09:30:43 +00:00
res = instance . query ( " SELECT COUNT(*) FROM test.view " )
2023-10-09 18:33:03 +00:00
logging . debug ( f " Current count (1): { res } " )
2023-03-27 17:20:48 +00:00
if int ( res ) == 20 :
2022-07-18 09:30:43 +00:00
break
else :
logging . debug ( f " Number of rows in test.view: { res } " )
2023-03-24 17:45:27 +00:00
instance . query ( " DROP VIEW test.consumer SYNC " )
2021-11-02 12:05:10 +00:00
for i in range ( 20 , 40 ) :
channel . basic_publish (
exchange = " mv " , routing_key = " " , body = json . dumps ( { " key " : i , " value " : i } )
2022-03-22 16:39:58 +00:00
)
2021-11-02 12:05:10 +00:00
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
2023-03-27 17:20:48 +00:00
SELECT * FROM test . drop_mv ;
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
)
for i in range ( 40 , 50 ) :
channel . basic_publish (
exchange = " mv " , routing_key = " " , body = json . dumps ( { " key " : i , " value " : i } )
2022-03-22 16:39:58 +00:00
)
2021-11-02 12:05:10 +00:00
while True :
2023-03-27 17:20:48 +00:00
result = instance . query ( " SELECT count() FROM test.view " )
2023-10-09 18:33:03 +00:00
logging . debug ( f " Current count (2): { result } " )
2023-03-27 17:20:48 +00:00
if int ( result ) == 50 :
2021-11-02 12:05:10 +00:00
break
2023-02-01 17:22:05 +00:00
time . sleep ( 1 )
2021-11-02 12:05:10 +00:00
2023-03-27 17:20:48 +00:00
result = instance . query ( " SELECT * FROM test.view ORDER BY key " )
2021-11-02 12:05:10 +00:00
rabbitmq_check_result ( result , True )
2023-05-03 18:06:46 +00:00
instance . query ( " DROP VIEW test.consumer SYNC " )
2022-12-20 18:32:59 +00:00
time . sleep ( 10 )
2021-11-02 12:05:10 +00:00
for i in range ( 50 , 60 ) :
channel . basic_publish (
exchange = " mv " , routing_key = " " , body = json . dumps ( { " key " : i , " value " : i } )
2022-03-22 16:39:58 +00:00
)
2021-11-02 12:05:10 +00:00
connection . close ( )
count = 0
2022-07-18 09:30:43 +00:00
start = time . time ( )
while time . time ( ) - start < 30 :
2023-03-27 17:20:48 +00:00
count = int ( instance . query ( " SELECT count() FROM test.drop_mv " ) )
2021-11-02 12:05:10 +00:00
if count :
break
2023-03-27 17:20:48 +00:00
instance . query ( " DROP TABLE test.drop_mv " )
2021-11-02 12:05:10 +00:00
assert count > 0
def test_rabbitmq_random_detach ( rabbitmq_cluster ) :
NUM_CONSUMERS = 2
NUM_QUEUES = 2
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' random ' ,
rabbitmq_queue_base = ' random ' ,
rabbitmq_num_queues = 2 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2021-11-02 12:05:10 +00:00
rabbitmq_num_consumers = 2 ,
rabbitmq_format = ' JSONEachRow ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 , channel_id String )
ENGINE = MergeTree
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * , _channel_id AS channel_id FROM test . rabbitmq ;
2022-03-22 16:39:58 +00:00
"""
2021-11-02 12:05:10 +00:00
)
i = [ 0 ]
messages_num = 10000
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
def produce ( ) :
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
2022-07-18 09:30:43 +00:00
for j in range ( messages_num ) :
2021-11-02 12:05:10 +00:00
messages . append ( json . dumps ( { " key " : i [ 0 ] , " value " : i [ 0 ] } ) )
i [ 0 ] + = 1
mes_id = str ( i )
channel . basic_publish (
2023-10-10 08:25:04 +00:00
exchange = " random " ,
2021-11-02 12:05:10 +00:00
routing_key = " " ,
properties = pika . BasicProperties ( message_id = mes_id ) ,
2023-10-10 07:42:30 +00:00
body = messages [ - 1 ] ,
2021-11-02 12:05:10 +00:00
)
connection . close ( )
threads = [ ]
threads_num = 20
for _ in range ( threads_num ) :
threads . append ( threading . Thread ( target = produce ) )
for thread in threads :
time . sleep ( random . uniform ( 0 , 1 ) )
thread . start ( )
2021-11-11 20:33:50 +00:00
# time.sleep(5)
# kill_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
# instance.query("detach table test.rabbitmq")
# revive_rabbitmq(rabbitmq_cluster.rabbitmq_docker_id)
2021-11-02 12:05:10 +00:00
for thread in threads :
thread . join ( )
2021-11-23 14:52:25 +00:00
def test_rabbitmq_predefined_configuration ( rabbitmq_cluster ) :
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
instance . query (
2022-03-22 16:39:58 +00:00
"""
2021-11-23 14:52:25 +00:00
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
2022-12-19 20:07:44 +00:00
ENGINE = RabbitMQ ( rabbit1 , rabbitmq_vhost = ' / ' )
SETTINGS rabbitmq_flush_interval_ms = 1000 ;
"""
2021-11-23 14:52:25 +00:00
)
channel . basic_publish (
exchange = " named " , routing_key = " " , body = json . dumps ( { " key " : 1 , " value " : 2 } )
2022-03-22 16:39:58 +00:00
)
2021-11-23 14:52:25 +00:00
while True :
result = instance . query (
" SELECT * FROM test.rabbitmq ORDER BY key " , ignore_error = True
)
if result == " 1 \t 2 \n " :
break
2022-05-31 12:38:11 +00:00
instance . restart_clickhouse ( )
channel . basic_publish (
exchange = " named " , routing_key = " " , body = json . dumps ( { " key " : 1 , " value " : 2 } )
)
while True :
result = instance . query (
" SELECT * FROM test.rabbitmq ORDER BY key " , ignore_error = True
)
if result == " 1 \t 2 \n " :
break
2021-11-23 14:52:25 +00:00
2022-07-01 17:06:57 +00:00
def test_rabbitmq_msgpack ( rabbitmq_cluster ) :
instance . query (
"""
drop table if exists rabbit_in ;
drop table if exists rabbit_out ;
create table
rabbit_in ( val String )
engine = RabbitMQ
settings rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' xhep ' ,
rabbitmq_format = ' MsgPack ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-07-01 17:06:57 +00:00
rabbitmq_num_consumers = 1 ;
create table
rabbit_out ( val String )
engine = RabbitMQ
settings rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' xhep ' ,
rabbitmq_format = ' MsgPack ' ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-07-01 17:06:57 +00:00
rabbitmq_num_consumers = 1 ;
set stream_like_engine_allow_direct_select = 1 ;
insert into rabbit_out select ' kek ' ;
"""
)
result = " "
try_no = 0
while True :
result = instance . query ( " select * from rabbit_in; " )
if result . strip ( ) == " kek " :
break
else :
try_no = try_no + 1
if try_no == 20 :
break
time . sleep ( 1 )
assert result . strip ( ) == " kek "
2022-07-04 12:06:20 +00:00
instance . query ( " drop table rabbit_in sync " )
instance . query ( " drop table rabbit_out sync " )
2022-07-01 17:06:57 +00:00
2022-06-24 00:07:16 +00:00
def test_rabbitmq_address ( rabbitmq_cluster ) :
2022-06-29 05:32:27 +00:00
instance2 . query (
"""
2022-06-24 00:07:16 +00:00
drop table if exists rabbit_in ;
drop table if exists rabbit_out ;
create table
rabbit_in ( val String )
engine = RabbitMQ
SETTINGS rabbitmq_exchange_name = ' rxhep ' ,
rabbitmq_format = ' CSV ' ,
rabbitmq_num_consumers = 1 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-06-24 00:07:16 +00:00
rabbitmq_address = ' amqp://root:clickhouse@rabbitmq1:5672/ ' ;
create table
rabbit_out ( val String ) engine = RabbitMQ
SETTINGS rabbitmq_exchange_name = ' rxhep ' ,
rabbitmq_format = ' CSV ' ,
rabbitmq_num_consumers = 1 ,
2022-12-19 20:07:44 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-06-24 00:07:16 +00:00
rabbitmq_address = ' amqp://root:clickhouse@rabbitmq1:5672/ ' ;
set stream_like_engine_allow_direct_select = 1 ;
insert into rabbit_out select ' kek ' ;
2022-06-29 05:32:27 +00:00
"""
)
2022-06-24 00:07:16 +00:00
result = " "
try_no = 0
while True :
result = instance2 . query ( " select * from rabbit_in; " )
if result . strip ( ) == " kek " :
break
else :
try_no = try_no + 1
if try_no == 20 :
break
time . sleep ( 1 )
assert result . strip ( ) == " kek "
2022-07-04 12:06:20 +00:00
2022-07-04 14:08:10 +00:00
instance2 . query ( " drop table rabbit_in sync " )
instance2 . query ( " drop table rabbit_out sync " )
2022-12-19 20:07:44 +00:00
2023-02-06 13:20:15 +00:00
2022-10-28 16:41:10 +00:00
def test_format_with_prefix_and_suffix ( rabbitmq_cluster ) :
2022-12-19 20:07:44 +00:00
instance . query (
"""
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2022-10-28 16:41:10 +00:00
rabbitmq_exchange_name = ' insert ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' custom ' ,
rabbitmq_format = ' CustomSeparated ' ;
"""
)
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
consumer_connection = pika . BlockingConnection ( parameters )
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
consumer . queue_bind ( exchange = " insert " , queue = queue_name , routing_key = " custom " )
instance . query (
" INSERT INTO test.rabbitmq select number*10 as key, number*100 as value from numbers(2) settings format_custom_result_before_delimiter= ' <prefix> \n ' , format_custom_result_after_delimiter= ' <suffix> \n ' "
)
insert_messages = [ ]
def onReceived ( channel , method , properties , body ) :
2023-02-02 19:16:53 +00:00
message = body . decode ( )
insert_messages . append ( message )
2023-10-09 18:33:03 +00:00
logging . debug ( f " Received { len ( insert_messages ) } message: { message } " )
2022-10-28 16:41:10 +00:00
if len ( insert_messages ) == 2 :
channel . stop_consuming ( )
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2023-01-31 19:25:08 +00:00
2022-10-28 16:41:10 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
assert (
" " . join ( insert_messages )
== " <prefix> \n 0 \t 0 \n <suffix> \n <prefix> \n 10 \t 100 \n <suffix> \n "
)
2023-02-05 11:58:51 +00:00
2022-10-28 16:41:10 +00:00
def test_max_rows_per_message ( rabbitmq_cluster ) :
num_rows = 5
instance . query (
"""
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . rabbit ;
CREATE TABLE test . rabbit ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_format = ' CustomSeparated ' ,
rabbitmq_exchange_name = ' custom ' ,
rabbitmq_exchange_type = ' direct ' ,
rabbitmq_routing_key_list = ' custom1 ' ,
rabbitmq_max_rows_per_message = 3 ,
2023-01-31 19:25:08 +00:00
rabbitmq_flush_interval_ms = 1000 ,
2022-10-28 16:41:10 +00:00
format_custom_result_before_delimiter = ' <prefix> \n ' ,
format_custom_result_after_delimiter = ' <suffix> \n ' ;
CREATE MATERIALIZED VIEW test . view Engine = Log AS
SELECT key , value FROM test . rabbit ;
2022-12-19 20:07:44 +00:00
"""
)
2022-10-28 16:41:10 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
consumer_connection = pika . BlockingConnection ( parameters )
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
consumer . queue_bind ( exchange = " custom " , queue = queue_name , routing_key = " custom1 " )
instance . query (
f " INSERT INTO test.rabbit select number*10 as key, number*100 as value from numbers( { num_rows } ) settings format_custom_result_before_delimiter= ' <prefix> \n ' , format_custom_result_after_delimiter= ' <suffix> \n ' "
)
insert_messages = [ ]
def onReceived ( channel , method , properties , body ) :
insert_messages . append ( body . decode ( ) )
if len ( insert_messages ) == 2 :
channel . stop_consuming ( )
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2022-10-28 16:41:10 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
assert len ( insert_messages ) == 2
assert (
" " . join ( insert_messages )
== " <prefix> \n 0 \t 0 \n 10 \t 100 \n 20 \t 200 \n <suffix> \n <prefix> \n 30 \t 300 \n 40 \t 400 \n <suffix> \n "
)
attempt = 0
rows = 0
while attempt < 100 :
rows = int ( instance . query ( " SELECT count() FROM test.view " ) )
if rows == num_rows :
break
attempt + = 1
assert rows == num_rows
result = instance . query ( " SELECT * FROM test.view " )
assert result == " 0 \t 0 \n 10 \t 100 \n 20 \t 200 \n 30 \t 300 \n 40 \t 400 \n "
2023-02-06 13:20:15 +00:00
2022-10-28 16:41:10 +00:00
def test_row_based_formats ( rabbitmq_cluster ) :
num_rows = 10
for format_name in [
" TSV " ,
" TSVWithNamesAndTypes " ,
" TSKV " ,
" CSV " ,
" CSVWithNamesAndTypes " ,
" CustomSeparatedWithNamesAndTypes " ,
" Values " ,
" JSON " ,
" JSONEachRow " ,
" JSONCompactEachRow " ,
" JSONCompactEachRowWithNamesAndTypes " ,
" JSONObjectEachRow " ,
" Avro " ,
" RowBinary " ,
" RowBinaryWithNamesAndTypes " ,
" MsgPack " ,
] :
2023-10-09 18:33:03 +00:00
logging . debug ( format_name )
2022-10-28 16:41:10 +00:00
instance . query (
f """
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . rabbit ;
CREATE TABLE test . rabbit ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_format = ' {format_name} ' ,
rabbitmq_exchange_name = ' {format_name} ' ,
rabbitmq_exchange_type = ' direct ' ,
2023-01-27 16:22:24 +00:00
rabbitmq_max_block_size = 100 ,
rabbitmq_flush_interval_ms = 1000 ,
2022-10-28 16:41:10 +00:00
rabbitmq_routing_key_list = ' {format_name} ' ,
rabbitmq_max_rows_per_message = 5 ;
2023-01-06 14:50:20 +00:00
2022-10-28 16:41:10 +00:00
CREATE MATERIALIZED VIEW test . view Engine = Log AS
SELECT key , value FROM test . rabbit ;
"""
)
2022-12-19 20:07:44 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip ,
rabbitmq_cluster . rabbitmq_port ,
" / " ,
credentials ,
)
2022-10-28 16:41:10 +00:00
consumer_connection = pika . BlockingConnection ( parameters )
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
consumer . queue_bind (
exchange = format_name , queue = queue_name , routing_key = format_name
)
2022-12-19 20:07:44 +00:00
instance . query (
2022-10-28 16:41:10 +00:00
f " INSERT INTO test.rabbit SELECT number * 10 as key, number * 100 as value FROM numbers( { num_rows } ); "
2022-12-19 20:07:44 +00:00
)
2022-10-28 16:41:10 +00:00
insert_messages = 0
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
def onReceived ( channel , method , properties , body ) :
nonlocal insert_messages
insert_messages + = 1
if insert_messages == 2 :
channel . stop_consuming ( )
2022-12-19 20:07:44 +00:00
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2022-10-28 16:41:10 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
assert insert_messages == 2
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
attempt = 0
rows = 0
while attempt < 100 :
rows = int ( instance . query ( " SELECT count() FROM test.view " ) )
if rows == num_rows :
break
attempt + = 1
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
assert rows == num_rows
expected = " "
for i in range ( num_rows ) :
expected + = str ( i * 10 ) + " \t " + str ( i * 100 ) + " \n "
result = instance . query ( " SELECT * FROM test.view " )
assert result == expected
2023-02-06 13:20:15 +00:00
2022-10-28 16:41:10 +00:00
def test_block_based_formats_1 ( rabbitmq_cluster ) :
2022-12-19 20:07:44 +00:00
instance . query (
"""
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
2022-10-28 16:41:10 +00:00
rabbitmq_exchange_name = ' PrettySpace ' ,
rabbitmq_exchange_type = ' direct ' ,
2023-01-27 16:22:24 +00:00
rabbitmq_max_block_size = 100 ,
rabbitmq_flush_interval_ms = 1000 ,
2022-10-28 16:41:10 +00:00
rabbitmq_routing_key_list = ' PrettySpace ' ,
rabbitmq_format = ' PrettySpace ' ;
2022-12-19 20:07:44 +00:00
"""
)
2022-10-28 16:41:10 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
consumer_connection = pika . BlockingConnection ( parameters )
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
consumer . queue_bind (
exchange = " PrettySpace " , queue = queue_name , routing_key = " PrettySpace "
)
instance . query (
2024-03-24 22:23:13 +00:00
" INSERT INTO test.rabbitmq SELECT number * 10 as key, number * 100 as value FROM numbers(5) settings max_block_size=2, optimize_trivial_insert_select=0, output_format_pretty_color=1, output_format_pretty_row_numbers=0; "
2022-10-28 16:41:10 +00:00
)
insert_messages = [ ]
def onReceived ( channel , method , properties , body ) :
insert_messages . append ( body . decode ( ) )
if len ( insert_messages ) == 3 :
channel . stop_consuming ( )
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2022-10-28 16:41:10 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
assert len ( insert_messages ) == 3
data = [ ]
for message in insert_messages :
splitted = message . split ( " \n " )
assert splitted [ 0 ] == " \x1b [1mkey \x1b [0m \x1b [1mvalue \x1b [0m "
assert splitted [ 1 ] == " "
assert splitted [ - 1 ] == " "
data + = [ line . split ( ) for line in splitted [ 2 : - 1 ] ]
assert data == [
[ " 0 " , " 0 " ] ,
[ " 10 " , " 100 " ] ,
[ " 20 " , " 200 " ] ,
[ " 30 " , " 300 " ] ,
[ " 40 " , " 400 " ] ,
]
2023-02-06 13:20:15 +00:00
2022-10-28 16:41:10 +00:00
def test_block_based_formats_2 ( rabbitmq_cluster ) :
num_rows = 100
for format_name in [
" JSONColumns " ,
" Native " ,
" Arrow " ,
" Parquet " ,
" ORC " ,
" JSONCompactColumns " ,
] :
2023-10-09 18:33:03 +00:00
logging . debug ( format_name )
2022-10-28 16:41:10 +00:00
instance . query (
f """
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . rabbit ;
CREATE TABLE test . rabbit ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_format = ' {format_name} ' ,
rabbitmq_exchange_name = ' {format_name} ' ,
rabbitmq_exchange_type = ' direct ' ,
2023-01-27 16:22:24 +00:00
rabbitmq_max_block_size = 100 ,
rabbitmq_flush_interval_ms = 1000 ,
2022-10-28 16:41:10 +00:00
rabbitmq_routing_key_list = ' {format_name} ' ;
2023-01-06 14:50:20 +00:00
2022-10-28 16:41:10 +00:00
CREATE MATERIALIZED VIEW test . view Engine = Log AS
SELECT key , value FROM test . rabbit ;
"""
)
2022-12-19 20:07:44 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip ,
rabbitmq_cluster . rabbitmq_port ,
" / " ,
credentials ,
)
2022-10-28 16:41:10 +00:00
consumer_connection = pika . BlockingConnection ( parameters )
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
consumer = consumer_connection . channel ( )
result = consumer . queue_declare ( queue = " " )
queue_name = result . method . queue
consumer . queue_bind (
exchange = format_name , queue = queue_name , routing_key = format_name
)
instance . query (
f " INSERT INTO test.rabbit SELECT number * 10 as key, number * 100 as value FROM numbers( { num_rows } ) settings max_block_size=12, optimize_trivial_insert_select=0; "
)
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
insert_messages = 0
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
def onReceived ( channel , method , properties , body ) :
nonlocal insert_messages
insert_messages + = 1
if insert_messages == 9 :
channel . stop_consuming ( )
2022-12-19 20:07:44 +00:00
2023-06-23 00:51:18 +00:00
consumer . basic_consume ( queue_name , onReceived )
2022-10-28 16:41:10 +00:00
consumer . start_consuming ( )
consumer_connection . close ( )
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
assert insert_messages == 9
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
attempt = 0
rows = 0
while attempt < 100 :
rows = int ( instance . query ( " SELECT count() FROM test.view " ) )
if rows == num_rows :
break
attempt + = 1
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
assert rows == num_rows
2022-12-19 20:07:44 +00:00
2022-10-28 16:41:10 +00:00
result = instance . query ( " SELECT * FROM test.view ORDER by key " )
expected = " "
for i in range ( num_rows ) :
expected + = str ( i * 10 ) + " \t " + str ( i * 100 ) + " \n "
assert result == expected
2023-01-06 14:50:20 +00:00
def test_rabbitmq_flush_by_block_size ( rabbitmq_cluster ) :
2023-01-06 15:53:24 +00:00
instance . query (
"""
2023-01-06 14:50:20 +00:00
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . consumer ;
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' flush_by_block ' ,
rabbitmq_queue_base = ' flush_by_block ' ,
rabbitmq_max_block_size = 100 ,
rabbitmq_flush_interval_ms = 640000 , / * should not flush by time during test * /
rabbitmq_format = ' JSONEachRow ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq ;
2023-01-31 19:25:08 +00:00
SYSTEM STOP MERGES ;
2023-01-06 14:50:20 +00:00
"""
2023-01-06 15:53:24 +00:00
)
2023-01-06 14:50:20 +00:00
2023-01-06 15:53:24 +00:00
cancel = threading . Event ( )
2023-01-06 14:50:20 +00:00
2023-01-06 15:53:24 +00:00
def produce ( ) :
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip ,
rabbitmq_cluster . rabbitmq_port ,
" / " ,
credentials ,
)
connection = pika . BlockingConnection ( parameters )
2023-01-06 14:50:20 +00:00
2023-01-06 15:53:24 +00:00
while not cancel . is_set ( ) :
try :
channel = connection . channel ( )
channel . basic_publish (
exchange = " flush_by_block " ,
routing_key = " " ,
body = json . dumps ( { " key " : 0 , " value " : 0 } ) ,
)
2023-12-29 14:02:11 +00:00
except Exception as e :
2023-10-09 18:33:03 +00:00
logging . debug ( f " Got error: { str ( e ) } " )
2023-01-06 14:50:20 +00:00
2023-01-06 15:53:24 +00:00
produce_thread = threading . Thread ( target = produce )
produce_thread . start ( )
2023-01-06 14:50:20 +00:00
2023-01-06 15:53:24 +00:00
while 0 == int (
instance . query (
" SELECT count() FROM system.parts WHERE database = ' test ' AND table = ' view ' AND name = ' all_1_1_0 ' "
)
) :
time . sleep ( 0.5 )
cancel . set ( )
produce_thread . join ( )
# more flushes can happens during test, we need to check only result of first flush (part named all_1_1_0).
result = instance . query ( " SELECT count() FROM test.view WHERE _part= ' all_1_1_0 ' " )
# logging.debug(result)
instance . query (
"""
2023-01-06 14:50:20 +00:00
DROP TABLE test . consumer ;
DROP TABLE test . view ;
DROP TABLE test . rabbitmq ;
"""
2023-01-06 15:53:24 +00:00
)
# 100 = first poll should return 100 messages (and rows)
# not waiting for stream_flush_interval_ms
assert (
int ( result ) == 100
) , " Messages from rabbitmq should be flushed when block of size rabbitmq_max_block_size is formed! "
def test_rabbitmq_flush_by_time ( rabbitmq_cluster ) :
instance . query (
"""
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . consumer ;
CREATE TABLE test . rabbitmq ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' rabbitmq1:5672 ' ,
rabbitmq_exchange_name = ' flush_by_time ' ,
rabbitmq_queue_base = ' flush_by_time ' ,
rabbitmq_max_block_size = 100 ,
rabbitmq_flush_interval_ms = 5000 ,
rabbitmq_format = ' JSONEachRow ' ;
CREATE TABLE test . view ( key UInt64 , value UInt64 , ts DateTime64 ( 3 ) MATERIALIZED now64 ( 3 ) )
ENGINE = MergeTree ( )
ORDER BY key ;
"""
)
cancel = threading . Event ( )
def produce ( ) :
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip ,
rabbitmq_cluster . rabbitmq_port ,
" / " ,
credentials ,
)
connection = pika . BlockingConnection ( parameters )
while not cancel . is_set ( ) :
try :
channel = connection . channel ( )
channel . basic_publish (
exchange = " flush_by_time " ,
routing_key = " " ,
body = json . dumps ( { " key " : 0 , " value " : 0 } ) ,
)
2023-10-09 18:33:03 +00:00
logging . debug ( " Produced a message " )
2023-01-06 15:53:24 +00:00
time . sleep ( 0.8 )
2023-12-29 14:02:11 +00:00
except Exception as e :
2023-10-09 18:33:03 +00:00
logging . debug ( f " Got error: { str ( e ) } " )
2023-01-06 15:53:24 +00:00
produce_thread = threading . Thread ( target = produce )
produce_thread . start ( )
instance . query (
"""
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . rabbitmq ;
"""
)
while True :
time . sleep ( 0.2 )
count = instance . query (
" SELECT count() FROM system.parts WHERE database = ' test ' AND table = ' view ' "
)
2023-10-09 18:33:03 +00:00
logging . debug ( f " kssenii total count: { count } " )
2023-01-06 15:53:24 +00:00
count = int (
instance . query (
" SELECT count() FROM system.parts WHERE database = ' test ' AND table = ' view ' AND name = ' all_1_1_0 ' "
)
)
2023-10-09 18:33:03 +00:00
logging . debug ( f " kssenii count: { count } " )
2023-01-06 15:53:24 +00:00
if count > 0 :
break
time . sleep ( 12 )
result = instance . query ( " SELECT uniqExact(ts) FROM test.view " )
cancel . set ( )
produce_thread . join ( )
instance . query (
"""
DROP TABLE test . consumer ;
DROP TABLE test . view ;
DROP TABLE test . rabbitmq ;
"""
)
2023-01-06 14:50:20 +00:00
2023-01-06 15:53:24 +00:00
assert int ( result ) == 3
2023-10-10 16:21:33 +00:00
def test_rabbitmq_handle_error_mode_stream ( rabbitmq_cluster ) :
instance . query (
"""
DROP TABLE IF EXISTS test . rabbitmq ;
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . data ;
DROP TABLE IF EXISTS test . errors ;
DROP TABLE IF EXISTS test . errors_view ;
CREATE TABLE test . rabbit ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' {} :5672 ' ,
rabbitmq_exchange_name = ' select ' ,
rabbitmq_commit_on_select = 1 ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ,
rabbitmq_handle_error_mode = ' stream ' ;
2023-11-14 20:13:09 +00:00
2023-10-10 16:21:33 +00:00
CREATE TABLE test . errors ( error Nullable ( String ) , broken_message Nullable ( String ) )
ENGINE = MergeTree ( )
ORDER BY tuple ( ) ;
CREATE MATERIALIZED VIEW test . errors_view TO test . errors AS
SELECT _error as error , _raw_message as broken_message FROM test . rabbit where not isNull ( _error ) ;
2023-11-14 20:13:09 +00:00
2023-10-10 16:21:33 +00:00
CREATE TABLE test . data ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
2023-11-14 20:13:09 +00:00
2023-10-10 16:21:33 +00:00
CREATE MATERIALIZED VIEW test . view TO test . data AS
SELECT key , value FROM test . rabbit ;
""" .format(
rabbitmq_cluster . rabbitmq_host
)
)
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
messages = [ ]
num_rows = 50
for i in range ( num_rows ) :
if i % 2 == 0 :
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
else :
messages . append ( " Broken message " + str ( i ) )
for message in messages :
channel . basic_publish ( exchange = " select " , routing_key = " " , body = message )
connection . close ( )
# The order of messages in select * from test.rabbitmq is not guaranteed, so sleep to collect everything in one select
time . sleep ( 1 )
attempt = 0
rows = 0
2023-10-30 19:51:03 +00:00
while attempt < 500 :
2023-10-10 16:21:33 +00:00
rows = int ( instance . query ( " SELECT count() FROM test.data " ) )
if rows == num_rows :
break
attempt + = 1
assert rows == num_rows
result = instance . query ( " SELECT * FROM test.data ORDER by key " )
expected = " 0 \t 0 \n " * ( num_rows / / 2 )
for i in range ( num_rows ) :
if i % 2 == 0 :
expected + = str ( i ) + " \t " + str ( i ) + " \n "
assert result == expected
attempt = 0
errors_count = 0
2023-10-30 19:51:03 +00:00
while attempt < 500 :
2023-10-10 16:21:33 +00:00
errors_count = int ( instance . query ( " SELECT count() FROM test.errors " ) )
if errors_count == num_rows :
break
attempt + = 1
assert errors_count == num_rows / 2
2023-10-11 17:35:18 +00:00
broken_messages = instance . query (
" SELECT broken_message FROM test.errors order by broken_message "
)
2023-10-10 16:21:33 +00:00
expected = [ ]
for i in range ( num_rows ) :
if i % 2 != 0 :
expected . append ( " Broken message " + str ( i ) + " \n " )
expected = " " . join ( sorted ( expected ) )
assert broken_messages == expected
2024-01-29 20:09:09 +00:00
def test_attach_broken_table ( rabbitmq_cluster ) :
instance . query (
" ATTACH TABLE rabbit_queue UUID ' 2d1cdf1a-f060-4a61-a7c9-5b59e59992c6 ' (`payload` String) ENGINE = RabbitMQ SETTINGS rabbitmq_host_port = ' nonexisting:5671 ' , rabbitmq_format = ' JSONEachRow ' , rabbitmq_username = ' test ' , rabbitmq_password = ' test ' "
)
error = instance . query_and_get_error ( " SELECT * FROM rabbit_queue " )
assert " CANNOT_CONNECT_RABBITMQ " in error
error = instance . query_and_get_error ( " INSERT INTO rabbit_queue VALUES ( ' test ' ) " )
assert " CANNOT_CONNECT_RABBITMQ " in error
2024-02-08 16:40:41 +00:00
2024-02-14 14:09:41 +00:00
def test_rabbitmq_nack_failed_insert ( rabbitmq_cluster ) :
table_name = " nack_failed_insert "
exchange = f " { table_name } _exchange "
2024-02-19 19:52:36 +00:00
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
2024-02-19 20:01:58 +00:00
channel . exchange_declare ( exchange = " deadl " )
2024-02-19 19:52:36 +00:00
2024-02-19 20:01:58 +00:00
result = channel . queue_declare ( queue = " deadq " )
2024-02-19 19:52:36 +00:00
queue_name = result . method . queue
2024-02-19 20:01:58 +00:00
channel . queue_bind ( exchange = " deadl " , routing_key = " " , queue = queue_name )
2024-02-19 19:52:36 +00:00
2024-02-14 14:09:41 +00:00
instance3 . query (
f """
CREATE TABLE test . { table_name } ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' {rabbitmq_cluster.rabbitmq_host} :5672 ' ,
rabbitmq_flush_interval_ms = 1000 ,
rabbitmq_exchange_name = ' {exchange} ' ,
2024-02-19 19:52:36 +00:00
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_queue_settings_list = ' x-dead-letter-exchange=deadl ' ;
2024-02-14 14:09:41 +00:00
DROP TABLE IF EXISTS test . view ;
CREATE TABLE test . view ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
DROP TABLE IF EXISTS test . consumer ;
CREATE MATERIALIZED VIEW test . consumer TO test . view AS
SELECT * FROM test . { table_name } ;
"""
)
num_rows = 25
for i in range ( num_rows ) :
message = json . dumps ( { " key " : i , " value " : i } ) + " \n "
channel . basic_publish ( exchange = exchange , routing_key = " " , body = message )
2024-02-14 14:35:59 +00:00
instance3 . wait_for_log_line (
" Failed to push to views. Error: Code: 252. DB::Exception: Too many parts "
)
2024-02-14 14:09:41 +00:00
instance3 . replace_in_config (
" /etc/clickhouse-server/config.d/mergetree.xml " ,
2024-02-19 19:52:36 +00:00
" parts_to_throw_insert>0 " ,
2024-02-14 14:09:41 +00:00
" parts_to_throw_insert>10 " ,
)
2024-02-19 19:52:36 +00:00
instance3 . restart_clickhouse ( )
count = [ 0 ]
2024-02-19 20:01:58 +00:00
2024-02-19 19:52:36 +00:00
def on_consume ( channel , method , properties , body ) :
channel . basic_publish ( exchange = exchange , routing_key = " " , body = body )
count [ 0 ] + = 1
if count [ 0 ] == num_rows :
channel . stop_consuming ( )
channel . basic_consume ( queue_name , on_consume )
channel . start_consuming ( )
2024-02-14 14:09:41 +00:00
attempt = 0
count = 0
while attempt < 100 :
count = int ( instance3 . query ( " SELECT count() FROM test.view " ) )
if count == num_rows :
break
attempt + = 1
assert count == num_rows
instance3 . query (
f """
DROP TABLE test . consumer ;
DROP TABLE test . view ;
DROP TABLE test . { table_name } ;
"""
)
2024-02-19 19:52:36 +00:00
connection . close ( )
2024-02-22 15:56:33 +00:00
def test_rabbitmq_reject_broken_messages ( rabbitmq_cluster ) :
credentials = pika . PlainCredentials ( " root " , " clickhouse " )
parameters = pika . ConnectionParameters (
rabbitmq_cluster . rabbitmq_ip , rabbitmq_cluster . rabbitmq_port , " / " , credentials
)
connection = pika . BlockingConnection ( parameters )
channel = connection . channel ( )
deadletter_exchange = " deadletter_exchange_handle_error_mode_stream "
deadletter_queue = " deadletter_queue_handle_error_mode_stream "
channel . exchange_declare ( exchange = deadletter_exchange )
result = channel . queue_declare ( queue = deadletter_queue )
channel . queue_bind (
exchange = deadletter_exchange , routing_key = " " , queue = deadletter_queue
)
instance . query (
f """
DROP TABLE IF EXISTS test . rabbitmq ;
DROP TABLE IF EXISTS test . view ;
DROP TABLE IF EXISTS test . data ;
DROP TABLE IF EXISTS test . errors ;
DROP TABLE IF EXISTS test . errors_view ;
CREATE TABLE test . rabbit ( key UInt64 , value UInt64 )
ENGINE = RabbitMQ
SETTINGS rabbitmq_host_port = ' {rabbitmq_cluster.rabbitmq_host} :5672 ' ,
rabbitmq_exchange_name = ' select ' ,
rabbitmq_commit_on_select = 1 ,
rabbitmq_format = ' JSONEachRow ' ,
rabbitmq_row_delimiter = ' \\ n ' ,
rabbitmq_handle_error_mode = ' stream ' ,
rabbitmq_queue_settings_list = ' x-dead-letter-exchange= {deadletter_exchange} ' ;
CREATE TABLE test . errors ( error Nullable ( String ) , broken_message Nullable ( String ) )
ENGINE = MergeTree ( )
ORDER BY tuple ( ) ;
CREATE MATERIALIZED VIEW test . errors_view TO test . errors AS
SELECT _error as error , _raw_message as broken_message FROM test . rabbit where not isNull ( _error ) ;
CREATE TABLE test . data ( key UInt64 , value UInt64 )
ENGINE = MergeTree ( )
ORDER BY key ;
CREATE MATERIALIZED VIEW test . view TO test . data AS
SELECT key , value FROM test . rabbit ;
"""
)
messages = [ ]
num_rows = 50
for i in range ( num_rows ) :
if i % 2 == 0 :
messages . append ( json . dumps ( { " key " : i , " value " : i } ) )
else :
messages . append ( " Broken message " + str ( i ) )
for message in messages :
channel . basic_publish ( exchange = " select " , routing_key = " " , body = message )
time . sleep ( 1 )
attempt = 0
rows = 0
while attempt < 500 :
rows = int ( instance . query ( " SELECT count() FROM test.data " ) )
if rows == num_rows :
break
attempt + = 1
2024-02-22 17:48:06 +00:00
time . sleep ( 1 )
2024-02-22 15:56:33 +00:00
assert rows == num_rows
dead_letters = [ ]
def on_dead_letter ( channel , method , properties , body ) :
dead_letters . append ( body )
if len ( dead_letters ) == num_rows / 2 :
channel . stop_consuming ( )
channel . basic_consume ( deadletter_queue , on_dead_letter )
channel . start_consuming ( )
assert len ( dead_letters ) == num_rows / 2
i = 1
for letter in dead_letters :
assert f " Broken message { i } " in str ( letter )
i + = 2
connection . close ( )