9.1 KiB
toc_priority | toc_title |
---|---|
6 | RabbitMQ |
RabbitMQ Engine
This engine allows integrating ClickHouse with RabbitMQ.
RabbitMQ
lets you:
- Publish or subscribe to data flows.
- Process streams as they become available.
Creating a Table
CREATE TABLE [IF NOT EXISTS] [db.]table_name [ON CLUSTER cluster]
(
name1 [type1] [DEFAULT|MATERIALIZED|ALIAS expr1],
name2 [type2] [DEFAULT|MATERIALIZED|ALIAS expr2],
...
) ENGINE = RabbitMQ SETTINGS
rabbitmq_host_port = 'host:port',
rabbitmq_exchange_name = 'exchange_name',
rabbitmq_format = 'data_format'[,]
[rabbitmq_exchange_type = 'exchange_type',]
[rabbitmq_routing_key_list = 'key1,key2,...',]
[rabbitmq_row_delimiter = 'delimiter_symbol',]
[rabbitmq_schema = '',]
[rabbitmq_num_consumers = N,]
[rabbitmq_num_queues = N,]
[rabbitmq_queue_base = 'queue',]
[rabbitmq_deadletter_exchange = 'dl-exchange',]
[rabbitmq_persistent = 0,]
[rabbitmq_skip_broken_messages = N,]
[rabbitmq_max_block_size = N,]
[rabbitmq_flush_interval_ms = N]
Required parameters:
rabbitmq_host_port
– host:port (for example,localhost:5672
).rabbitmq_exchange_name
– RabbitMQ exchange name.rabbitmq_format
– Message format. Uses the same notation as the SQLFORMAT
function, such asJSONEachRow
. For more information, see the Formats section.
Optional parameters:
rabbitmq_exchange_type
– The type of RabbitMQ exchange:direct
,fanout
,topic
,headers
,consistent_hash
. Default:fanout
.rabbitmq_routing_key_list
– A comma-separated list of routing keys.rabbitmq_row_delimiter
– Delimiter character, which ends the message.rabbitmq_schema
– Parameter that must be used if the format requires a schema definition. For example, Cap’n Proto requires the path to the schema file and the name of the rootschema.capnp:Message
object.rabbitmq_num_consumers
– The number of consumers per table. Default:1
. Specify more consumers if the throughput of one consumer is insufficient.rabbitmq_num_queues
– Total number of queues. Default:1
. Increasing this number can significantly improve performance.rabbitmq_queue_base
- Specify a hint for queue names. Use cases of this setting are described below.rabbitmq_deadletter_exchange
- Specify name for a dead letter exchange. You can create another table with this exchange name and collect messages in cases when they are republished to dead letter exchange. By default dead letter exchange is not specified.rabbitmq_persistent
- If set to 1 (true), in insert query delivery mode will be set to 2 (marks messages as 'persistent'). Default:0
.rabbitmq_skip_broken_messages
– RabbitMQ message parser tolerance to schema-incompatible messages per block. Default:0
. Ifrabbitmq_skip_broken_messages = N
then the engine skips N RabbitMQ messages that cannot be parsed (a message equals a row of data).rabbitmq_max_block_size
rabbitmq_flush_interval_ms
Also format settings can be added along with rabbitmq-related settings.
Example:
CREATE TABLE queue (
key UInt64,
value UInt64,
date DateTime
) ENGINE = RabbitMQ SETTINGS rabbitmq_host_port = 'localhost:5672',
rabbitmq_exchange_name = 'exchange1',
rabbitmq_format = 'JSONEachRow',
rabbitmq_num_consumers = 5,
date_time_input_format = 'best_effort';
The RabbitMQ server configuration should be added using the ClickHouse config file.
Required configuration:
<rabbitmq>
<username>root</username>
<password>clickhouse</password>
</rabbitmq>
Additional configuration:
<rabbitmq>
<vhost>clickhouse</vhost>
</rabbitmq>
Description
SELECT
is not particularly useful for reading messages (except for debugging), because each message can be read only once. It is more practical to create real-time threads using materialized views. To do this:
- Use the engine to create a RabbitMQ consumer and consider it a data stream.
- Create a table with the desired structure.
- Create a materialized view that converts data from the engine and puts it into a previously created table.
When the MATERIALIZED VIEW
joins the engine, it starts collecting data in the background. This allows you to continually receive messages from RabbitMQ and convert them to the required format using SELECT
.
One RabbitMQ table can have as many materialized views as you like.
Data can be channeled based on rabbitmq_exchange_type
and the specified rabbitmq_routing_key_list
.
There can be no more than one exchange per table. One exchange can be shared between multiple tables - it enables routing into multiple tables at the same time.
Exchange type options:
direct
- Routing is based on the exact matching of keys. Example table key list:key1,key2,key3,key4,key5
, message key can equal any of them.fanout
- Routing to all tables (where exchange name is the same) regardless of the keys.topic
- Routing is based on patterns with dot-separated keys. Examples:*.logs
,records.*.*.2020
,*.2018,*.2019,*.2020
.headers
- Routing is based onkey=value
matches with a settingx-match=all
orx-match=any
. Example table key list:x-match=all,format=logs,type=report,year=2020
.consistent_hash
- Data is evenly distributed between all bound tables (where the exchange name is the same). Note that this exchange type must be enabled with RabbitMQ plugin:rabbitmq-plugins enable rabbitmq_consistent_hash_exchange
.
Setting rabbitmq_queue_base
may be used for the following cases:
- to let different tables share queues, so that multiple consumers could be registered for the same queues, which makes a better performance. If using
rabbitmq_num_consumers
and/orrabbitmq_num_queues
settings, the exact match of queues is achieved in case these parameters are the same. - to be able to restore reading from certain durable queues when not all messages were successfully consumed. To resume consumption from one specific queue - set its name in
rabbitmq_queue_base
setting and do not specifyrabbitmq_num_consumers
andrabbitmq_num_queues
(defaults to 1). To resume consumption from all queues, which were declared for a specific table - just specify the same settings:rabbitmq_queue_base
,rabbitmq_num_consumers
,rabbitmq_num_queues
. By default, queue names will be unique to tables. - to reuse queues as they are declared durable and not auto-deleted. (Can be deleted via any of RabbitMQ CLI tools.)
To improve performance, received messages are grouped into blocks the size of max_insert_block_size. If the block wasn’t formed within stream_flush_interval_ms milliseconds, the data will be flushed to the table regardless of the completeness of the block.
If rabbitmq_num_consumers
and/or rabbitmq_num_queues
settings are specified along with rabbitmq_exchange_type
, then:
rabbitmq-consistent-hash-exchange
plugin must be enabled.message_id
property of the published messages must be specified (unique for each message/batch).
For insert query there is message metadata, which is added for each published message: messageID
and republished
flag (true, if published more than once) - can be accessed via message headers.
Do not use the same table for inserts and materialized views.
Example:
CREATE TABLE queue (
key UInt64,
value UInt64
) ENGINE = RabbitMQ SETTINGS rabbitmq_host_port = 'localhost:5672',
rabbitmq_exchange_name = 'exchange1',
rabbitmq_exchange_type = 'headers',
rabbitmq_routing_key_list = 'format=logs,type=report,year=2020',
rabbitmq_format = 'JSONEachRow',
rabbitmq_num_consumers = 5;
CREATE TABLE daily (key UInt64, value UInt64)
ENGINE = MergeTree() ORDER BY key;
CREATE MATERIALIZED VIEW consumer TO daily
AS SELECT key, value FROM queue;
SELECT key, value FROM daily ORDER BY key;
Virtual Columns
_exchange_name
- RabbitMQ exchange name._channel_id
- ChannelID, on which consumer, who received the message, was declared._delivery_tag
- DeliveryTag of the received message. Scoped per channel._redelivered
-redelivered
flag of the message._message_id
- messageID of the received message; non-empty if was set, when message was published._timestamp
- timestamp of the received message; non-empty if was set, when message was published.