mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-09-21 09:10:48 +00:00
refactor
This commit is contained in:
parent
cb51558bdf
commit
40be74ac65
@ -41,7 +41,7 @@ struct Settings
|
||||
M(SettingUInt64, min_insert_block_size_rows, DEFAULT_INSERT_BLOCK_SIZE, "Squash blocks passed to INSERT query to specified size in rows, if blocks are not big enough.") \
|
||||
M(SettingUInt64, min_insert_block_size_bytes, (DEFAULT_INSERT_BLOCK_SIZE * 256), "Squash blocks passed to INSERT query to specified size in bytes, if blocks are not big enough.") \
|
||||
M(SettingMaxThreads, max_threads, 0, "The maximum number of threads to execute the request. By default, it is determined automatically.") \
|
||||
M(SettingMaxAlterThreads, max_alter_threads, 0, "The maximum number of threads to execute the ALTER requests. By default, it is determined automatically.") \
|
||||
M(SettingMaxThreads, max_alter_threads, 0, "The maximum number of threads to execute the ALTER requests. By default, it is determined automatically.") \
|
||||
M(SettingUInt64, max_read_buffer_size, DBMS_DEFAULT_BUFFER_SIZE, "The maximum size of the buffer to read from the filesystem.") \
|
||||
M(SettingUInt64, max_distributed_connections, 1024, "The maximum number of connections for distributed processing of one query (should be greater than max_threads).") \
|
||||
M(SettingUInt64, max_query_size, 262144, "Which part of the query can be read into RAM for parsing (the remaining data for INSERT, if any, is read later)") \
|
||||
|
@ -87,13 +87,6 @@ struct SettingMaxThreads
|
||||
UInt64 getAutoValueImpl() const;
|
||||
};
|
||||
|
||||
struct SettingMaxAlterThreads: public SettingMaxThreads
|
||||
{
|
||||
SettingMaxAlterThreads(UInt64 x = 0):
|
||||
SettingMaxThreads(x)
|
||||
{}
|
||||
};
|
||||
|
||||
struct SettingSeconds
|
||||
{
|
||||
Poco::Timespan value;
|
||||
|
Loading…
Reference in New Issue
Block a user