mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-18 13:42:02 +00:00
97f2a2213e
* Move some code outside dbms/src folder * Fix paths
36 lines
1.5 KiB
C++
36 lines
1.5 KiB
C++
#pragma once
|
|
|
|
#include <Core/SettingsCollection.h>
|
|
|
|
|
|
namespace DB
|
|
{
|
|
|
|
class ASTStorage;
|
|
|
|
/** Settings for the Kafka engine.
|
|
* Could be loaded from a CREATE TABLE query (SETTINGS clause).
|
|
*/
|
|
struct KafkaSettings : public SettingsCollection<KafkaSettings>
|
|
{
|
|
|
|
|
|
#define LIST_OF_KAFKA_SETTINGS(M) \
|
|
M(SettingString, kafka_broker_list, "", "A comma-separated list of brokers for Kafka engine.", 0) \
|
|
M(SettingString, kafka_topic_list, "", "A list of Kafka topics.", 0) \
|
|
M(SettingString, kafka_group_name, "", "A group of Kafka consumers.", 0) \
|
|
M(SettingString, kafka_format, "", "The message format for Kafka engine.", 0) \
|
|
M(SettingChar, kafka_row_delimiter, '\0', "The character to be considered as a delimiter in Kafka message.", 0) \
|
|
M(SettingString, kafka_schema, "", "Schema identifier (used by schema-based formats) for Kafka engine", 0) \
|
|
M(SettingUInt64, kafka_num_consumers, 1, "The number of consumers per table for Kafka engine.", 0) \
|
|
M(SettingUInt64, kafka_max_block_size, 0, "The maximum block size per table for Kafka engine.", 0) \
|
|
M(SettingUInt64, kafka_skip_broken_messages, 0, "Skip at least this number of broken messages from Kafka topic per block", 0) \
|
|
M(SettingUInt64, kafka_commit_every_batch, 0, "Commit every consumed and handled batch instead of a single commit after writing a whole block", 0)
|
|
|
|
DECLARE_SETTINGS_COLLECTION(LIST_OF_KAFKA_SETTINGS)
|
|
|
|
void loadFromQuery(ASTStorage & storage_def);
|
|
};
|
|
|
|
}
|