mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-27 18:12:02 +00:00
4052 lines
159 KiB
Markdown
4052 lines
159 KiB
Markdown
---
|
||
sidebar_label: Core Settings
|
||
sidebar_position: 2
|
||
slug: /en/operations/settings/settings
|
||
toc_max_heading_level: 2
|
||
---
|
||
|
||
# Core Settings
|
||
|
||
## additional_table_filters
|
||
|
||
An additional filter expression that is applied after reading
|
||
from the specified table.
|
||
|
||
Default value: 0.
|
||
|
||
**Example**
|
||
|
||
``` sql
|
||
insert into table_1 values (1, 'a'), (2, 'bb'), (3, 'ccc'), (4, 'dddd');
|
||
```
|
||
```response
|
||
┌─x─┬─y────┐
|
||
│ 1 │ a │
|
||
│ 2 │ bb │
|
||
│ 3 │ ccc │
|
||
│ 4 │ dddd │
|
||
└───┴──────┘
|
||
```
|
||
```sql
|
||
SELECT *
|
||
FROM table_1
|
||
SETTINGS additional_table_filters = (('table_1', 'x != 2'))
|
||
```
|
||
```response
|
||
┌─x─┬─y────┐
|
||
│ 1 │ a │
|
||
│ 3 │ ccc │
|
||
│ 4 │ dddd │
|
||
└───┴──────┘
|
||
```
|
||
|
||
## allow_nondeterministic_mutations {#allow_nondeterministic_mutations}
|
||
|
||
User-level setting that allows mutations on replicated tables to make use of non-deterministic functions such as `dictGet`.
|
||
|
||
Given that, for example, dictionaries, can be out of sync across nodes, mutations that pull values from them are disallowed on replicated tables by default. Enabling this setting allows this behavior, making it the user's responsibility to ensure that the data used is in sync across all nodes.
|
||
|
||
Default value: 0.
|
||
|
||
**Example**
|
||
|
||
``` xml
|
||
<profiles>
|
||
<default>
|
||
<allow_nondeterministic_mutations>1</allow_nondeterministic_mutations>
|
||
|
||
<!-- ... -->
|
||
</default>
|
||
|
||
<!-- ... -->
|
||
|
||
</profiles>
|
||
```
|
||
|
||
## distributed_product_mode {#distributed-product-mode}
|
||
|
||
Changes the behaviour of [distributed subqueries](../../sql-reference/operators/in.md).
|
||
|
||
ClickHouse applies this setting when the query contains the product of distributed tables, i.e. when the query for a distributed table contains a non-GLOBAL subquery for the distributed table.
|
||
|
||
Restrictions:
|
||
|
||
- Only applied for IN and JOIN subqueries.
|
||
- Only if the FROM section uses a distributed table containing more than one shard.
|
||
- If the subquery concerns a distributed table containing more than one shard.
|
||
- Not used for a table-valued [remote](../../sql-reference/table-functions/remote.md) function.
|
||
|
||
Possible values:
|
||
|
||
- `deny` — Default value. Prohibits using these types of subqueries (returns the “Double-distributed in/JOIN subqueries is denied” exception).
|
||
- `local` — Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal `IN`/`JOIN.`
|
||
- `global` — Replaces the `IN`/`JOIN` query with `GLOBAL IN`/`GLOBAL JOIN.`
|
||
- `allow` — Allows the use of these types of subqueries.
|
||
|
||
## prefer_global_in_and_join {#prefer-global-in-and-join}
|
||
|
||
Enables the replacement of `IN`/`JOIN` operators with `GLOBAL IN`/`GLOBAL JOIN`.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled. `IN`/`JOIN` operators are not replaced with `GLOBAL IN`/`GLOBAL JOIN`.
|
||
- 1 — Enabled. `IN`/`JOIN` operators are replaced with `GLOBAL IN`/`GLOBAL JOIN`.
|
||
|
||
Default value: `0`.
|
||
|
||
**Usage**
|
||
|
||
Although `SET distributed_product_mode=global` can change the queries behavior for the distributed tables, it's not suitable for local tables or tables from external resources. Here is when the `prefer_global_in_and_join` setting comes into play.
|
||
|
||
For example, we have query serving nodes that contain local tables, which are not suitable for distribution. We need to scatter their data on the fly during distributed processing with the `GLOBAL` keyword — `GLOBAL IN`/`GLOBAL JOIN`.
|
||
|
||
Another use case of `prefer_global_in_and_join` is accessing tables created by external engines. This setting helps to reduce the number of calls to external sources while joining such tables: only one call per query.
|
||
|
||
**See also:**
|
||
|
||
- [Distributed subqueries](../../sql-reference/operators/in.md/#select-distributed-subqueries) for more information on how to use `GLOBAL IN`/`GLOBAL JOIN`
|
||
|
||
## enable_optimize_predicate_expression {#enable-optimize-predicate-expression}
|
||
|
||
Turns on predicate pushdown in `SELECT` queries.
|
||
|
||
Predicate pushdown may significantly reduce network traffic for distributed queries.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1.
|
||
|
||
Usage
|
||
|
||
Consider the following queries:
|
||
|
||
1. `SELECT count() FROM test_table WHERE date = '2018-10-10'`
|
||
2. `SELECT count() FROM (SELECT * FROM test_table) WHERE date = '2018-10-10'`
|
||
|
||
If `enable_optimize_predicate_expression = 1`, then the execution time of these queries is equal because ClickHouse applies `WHERE` to the subquery when processing it.
|
||
|
||
If `enable_optimize_predicate_expression = 0`, then the execution time of the second query is much longer because the `WHERE` clause applies to all the data after the subquery finishes.
|
||
|
||
## fallback_to_stale_replicas_for_distributed_queries {#settings-fallback_to_stale_replicas_for_distributed_queries}
|
||
|
||
Forces a query to an out-of-date replica if updated data is not available. See [Replication](../../engines/table-engines/mergetree-family/replication.md).
|
||
|
||
ClickHouse selects the most relevant from the outdated replicas of the table.
|
||
|
||
Used when performing `SELECT` from a distributed table that points to replicated tables.
|
||
|
||
By default, 1 (enabled).
|
||
|
||
## force_index_by_date {#settings-force_index_by_date}
|
||
|
||
Disables query execution if the index can’t be used by date.
|
||
|
||
Works with tables in the MergeTree family.
|
||
|
||
If `force_index_by_date=1`, ClickHouse checks whether the query has a date key condition that can be used for restricting data ranges. If there is no suitable condition, it throws an exception. However, it does not check whether the condition reduces the amount of data to read. For example, the condition `Date != ' 2000-01-01 '` is acceptable even when it matches all the data in the table (i.e., running the query requires a full scan). For more information about ranges of data in MergeTree tables, see [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md).
|
||
|
||
## force_primary_key {#force-primary-key}
|
||
|
||
Disables query execution if indexing by the primary key is not possible.
|
||
|
||
Works with tables in the MergeTree family.
|
||
|
||
If `force_primary_key=1`, ClickHouse checks to see if the query has a primary key condition that can be used for restricting data ranges. If there is no suitable condition, it throws an exception. However, it does not check whether the condition reduces the amount of data to read. For more information about data ranges in MergeTree tables, see [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md).
|
||
|
||
## use_skip_indexes {#settings-use_skip_indexes}
|
||
|
||
Use data skipping indexes during query execution.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1.
|
||
|
||
## force_data_skipping_indices {#settings-force_data_skipping_indices}
|
||
|
||
Disables query execution if passed data skipping indices wasn't used.
|
||
|
||
Consider the following example:
|
||
|
||
```sql
|
||
CREATE TABLE data
|
||
(
|
||
key Int,
|
||
d1 Int,
|
||
d1_null Nullable(Int),
|
||
INDEX d1_idx d1 TYPE minmax GRANULARITY 1,
|
||
INDEX d1_null_idx assumeNotNull(d1_null) TYPE minmax GRANULARITY 1
|
||
)
|
||
Engine=MergeTree()
|
||
ORDER BY key;
|
||
|
||
SELECT * FROM data_01515;
|
||
SELECT * FROM data_01515 SETTINGS force_data_skipping_indices=''; -- query will produce CANNOT_PARSE_TEXT error.
|
||
SELECT * FROM data_01515 SETTINGS force_data_skipping_indices='d1_idx'; -- query will produce INDEX_NOT_USED error.
|
||
SELECT * FROM data_01515 WHERE d1 = 0 SETTINGS force_data_skipping_indices='d1_idx'; -- Ok.
|
||
SELECT * FROM data_01515 WHERE d1 = 0 SETTINGS force_data_skipping_indices='`d1_idx`'; -- Ok (example of full featured parser).
|
||
SELECT * FROM data_01515 WHERE d1 = 0 SETTINGS force_data_skipping_indices='`d1_idx`, d1_null_idx'; -- query will produce INDEX_NOT_USED error, since d1_null_idx is not used.
|
||
SELECT * FROM data_01515 WHERE d1 = 0 AND assumeNotNull(d1_null) = 0 SETTINGS force_data_skipping_indices='`d1_idx`, d1_null_idx'; -- Ok.
|
||
```
|
||
|
||
Works with tables in the MergeTree family.
|
||
|
||
## convert_query_to_cnf {#convert_query_to_cnf}
|
||
|
||
When set to `true`, a `SELECT` query will be converted to conjuctive normal form (CNF). There are scenarios where rewriting a query in CNF may execute faster (view this [Github issue](https://github.com/ClickHouse/ClickHouse/issues/11749) for an explanation).
|
||
|
||
For example, notice how the following `SELECT` query is not modified (the default behavior):
|
||
|
||
```sql
|
||
EXPLAIN SYNTAX
|
||
SELECT *
|
||
FROM
|
||
(
|
||
SELECT number AS x
|
||
FROM numbers(20)
|
||
) AS a
|
||
WHERE ((x >= 1) AND (x <= 5)) OR ((x >= 10) AND (x <= 15))
|
||
SETTINGS convert_query_to_cnf = false;
|
||
```
|
||
|
||
The result is:
|
||
|
||
```response
|
||
┌─explain────────────────────────────────────────────────────────┐
|
||
│ SELECT x │
|
||
│ FROM │
|
||
│ ( │
|
||
│ SELECT number AS x │
|
||
│ FROM numbers(20) │
|
||
│ WHERE ((x >= 1) AND (x <= 5)) OR ((x >= 10) AND (x <= 15)) │
|
||
│ ) AS a │
|
||
│ WHERE ((x >= 1) AND (x <= 5)) OR ((x >= 10) AND (x <= 15)) │
|
||
│ SETTINGS convert_query_to_cnf = 0 │
|
||
└────────────────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
Let's set `convert_query_to_cnf` to `true` and see what changes:
|
||
|
||
```sql
|
||
EXPLAIN SYNTAX
|
||
SELECT *
|
||
FROM
|
||
(
|
||
SELECT number AS x
|
||
FROM numbers(20)
|
||
) AS a
|
||
WHERE ((x >= 1) AND (x <= 5)) OR ((x >= 10) AND (x <= 15))
|
||
SETTINGS convert_query_to_cnf = true;
|
||
```
|
||
|
||
Notice the `WHERE` clause is rewritten in CNF, but the result set is the identical - the Boolean logic is unchanged:
|
||
|
||
```response
|
||
┌─explain───────────────────────────────────────────────────────────────────────────────────────────────────────────────┐
|
||
│ SELECT x │
|
||
│ FROM │
|
||
│ ( │
|
||
│ SELECT number AS x │
|
||
│ FROM numbers(20) │
|
||
│ WHERE ((x <= 15) OR (x <= 5)) AND ((x <= 15) OR (x >= 1)) AND ((x >= 10) OR (x <= 5)) AND ((x >= 10) OR (x >= 1)) │
|
||
│ ) AS a │
|
||
│ WHERE ((x >= 10) OR (x >= 1)) AND ((x >= 10) OR (x <= 5)) AND ((x <= 15) OR (x >= 1)) AND ((x <= 15) OR (x <= 5)) │
|
||
│ SETTINGS convert_query_to_cnf = 1 │
|
||
└───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
Possible values: true, false
|
||
|
||
Default value: false
|
||
|
||
|
||
## fsync_metadata {#fsync-metadata}
|
||
|
||
Enables or disables [fsync](http://pubs.opengroup.org/onlinepubs/9699919799/functions/fsync.html) when writing `.sql` files. Enabled by default.
|
||
|
||
It makes sense to disable it if the server has millions of tiny tables that are constantly being created and destroyed.
|
||
|
||
## function_range_max_elements_in_block {#settings-function_range_max_elements_in_block}
|
||
|
||
Sets the safety threshold for data volume generated by function [range](../../sql-reference/functions/array-functions.md/#range). Defines the maximum number of values generated by function per block of data (sum of array sizes for every row in a block).
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `500,000,000`.
|
||
|
||
**See Also**
|
||
|
||
- [max_block_size](#setting-max_block_size)
|
||
- [min_insert_block_size_rows](#min-insert-block-size-rows)
|
||
|
||
## enable_http_compression {#settings-enable_http_compression}
|
||
|
||
Enables or disables data compression in the response to an HTTP request.
|
||
|
||
For more information, read the [HTTP interface description](../../interfaces/http.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0.
|
||
|
||
## http_zlib_compression_level {#settings-http_zlib_compression_level}
|
||
|
||
Sets the level of data compression in the response to an HTTP request if [enable_http_compression = 1](#settings-enable_http_compression).
|
||
|
||
Possible values: Numbers from 1 to 9.
|
||
|
||
Default value: 3.
|
||
|
||
## http_native_compression_disable_checksumming_on_decompress {#settings-http_native_compression_disable_checksumming_on_decompress}
|
||
|
||
Enables or disables checksum verification when decompressing the HTTP POST data from the client. Used only for ClickHouse native compression format (not used with `gzip` or `deflate`).
|
||
|
||
For more information, read the [HTTP interface description](../../interfaces/http.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0.
|
||
|
||
## http_max_uri_size {#http-max-uri-size}
|
||
|
||
Sets the maximum URI length of an HTTP request.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: 1048576.
|
||
|
||
## table_function_remote_max_addresses {#table_function_remote_max_addresses}
|
||
|
||
Sets the maximum number of addresses generated from patterns for the [remote](../../sql-reference/table-functions/remote.md) function.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `1000`.
|
||
|
||
## glob_expansion_max_elements {#glob_expansion_max_elements}
|
||
|
||
Sets the maximum number of addresses generated from patterns for external storages and table functions (like [url](../../sql-reference/table-functions/url.md)) except the `remote` function.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `1000`.
|
||
|
||
## send_progress_in_http_headers {#settings-send_progress_in_http_headers}
|
||
|
||
Enables or disables `X-ClickHouse-Progress` HTTP response headers in `clickhouse-server` responses.
|
||
|
||
For more information, read the [HTTP interface description](../../interfaces/http.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0.
|
||
|
||
## max_http_get_redirects {#setting-max_http_get_redirects}
|
||
|
||
Limits the maximum number of HTTP GET redirect hops for [URL](../../engines/table-engines/special/url.md)-engine tables. The setting applies to both types of tables: those created by the [CREATE TABLE](../../sql-reference/statements/create/table.md) query and by the [url](../../sql-reference/table-functions/url.md) table function.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer number of hops.
|
||
- 0 — No hops allowed.
|
||
|
||
Default value: 0.
|
||
|
||
## insert_null_as_default {#insert_null_as_default}
|
||
|
||
Enables or disables the insertion of [default values](../../sql-reference/statements/create/table.md/#create-default-values) instead of [NULL](../../sql-reference/syntax.md/#null-literal) into columns with not [nullable](../../sql-reference/data-types/nullable.md/#data_type-nullable) data type.
|
||
If column type is not nullable and this setting is disabled, then inserting `NULL` causes an exception. If column type is nullable, then `NULL` values are inserted as is, regardless of this setting.
|
||
|
||
This setting is applicable to [INSERT ... SELECT](../../sql-reference/statements/insert-into.md/#inserting-the-results-of-select) queries. Note that `SELECT` subqueries may be concatenated with `UNION ALL` clause.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Inserting `NULL` into a not nullable column causes an exception.
|
||
- 1 — Default column value is inserted instead of `NULL`.
|
||
|
||
Default value: `1`.
|
||
|
||
## join_default_strictness {#settings-join_default_strictness}
|
||
|
||
Sets default strictness for [JOIN clauses](../../sql-reference/statements/select/join.md/#select-join).
|
||
|
||
Possible values:
|
||
|
||
- `ALL` — If the right table has several matching rows, ClickHouse creates a [Cartesian product](https://en.wikipedia.org/wiki/Cartesian_product) from matching rows. This is the normal `JOIN` behaviour from standard SQL.
|
||
- `ANY` — If the right table has several matching rows, only the first one found is joined. If the right table has only one matching row, the results of `ANY` and `ALL` are the same.
|
||
- `ASOF` — For joining sequences with an uncertain match.
|
||
- `Empty string` — If `ALL` or `ANY` is not specified in the query, ClickHouse throws an exception.
|
||
|
||
Default value: `ALL`.
|
||
|
||
## join_algorithm {#settings-join_algorithm}
|
||
|
||
Specifies which [JOIN](../../sql-reference/statements/select/join.md) algorithm is used.
|
||
|
||
Several algorithms can be specified, and an available one would be chosen for a particular query based on kind/strictness and table engine.
|
||
|
||
Possible values:
|
||
|
||
- default
|
||
|
||
This is the equivalent of `hash` or `direct`, if possible (same as `direct,hash`)
|
||
|
||
- grace_hash
|
||
|
||
[Grace hash join](https://en.wikipedia.org/wiki/Hash_join#Grace_hash_join) is used. Grace hash provides an algorithm option that provides performant complex joins while limiting memory use.
|
||
|
||
The first phase of a grace join reads the right table and splits it into N buckets depending on the hash value of key columns (initially, N is `grace_hash_join_initial_buckets`). This is done in a way to ensure that each bucket can be processed independently. Rows from the first bucket are added to an in-memory hash table while the others are saved to disk. If the hash table grows beyond the memory limit (e.g., as set by [`max_bytes_in_join`](/docs/en/operations/settings/query-complexity.md/#settings-max_bytes_in_join)), the number of buckets is increased and the assigned bucket for each row. Any rows which don’t belong to the current bucket are flushed and reassigned.
|
||
|
||
- hash
|
||
|
||
[Hash join algorithm](https://en.wikipedia.org/wiki/Hash_join) is used. The most generic implementation that supports all combinations of kind and strictness and multiple join keys that are combined with `OR` in the `JOIN ON` section.
|
||
|
||
- parallel_hash
|
||
|
||
A variation of `hash` join that splits the data into buckets and builds several hashtables instead of one concurrently to speed up this process.
|
||
|
||
When using the `hash` algorithm, the right part of `JOIN` is uploaded into RAM.
|
||
|
||
- partial_merge
|
||
|
||
A variation of the [sort-merge algorithm](https://en.wikipedia.org/wiki/Sort-merge_join), where only the right table is fully sorted.
|
||
|
||
The `RIGHT JOIN` and `FULL JOIN` are supported only with `ALL` strictness (`SEMI`, `ANTI`, `ANY`, and `ASOF` are not supported).
|
||
|
||
When using the `partial_merge` algorithm, ClickHouse sorts the data and dumps it to the disk. The `partial_merge` algorithm in ClickHouse differs slightly from the classic realization. First, ClickHouse sorts the right table by joining keys in blocks and creates a min-max index for sorted blocks. Then it sorts parts of the left table by the `join key` and joins them over the right table. The min-max index is also used to skip unneeded right table blocks.
|
||
|
||
- direct
|
||
|
||
This algorithm can be applied when the storage for the right table supports key-value requests.
|
||
|
||
The `direct` algorithm performs a lookup in the right table using rows from the left table as keys. It's supported only by special storage such as [Dictionary](../../engines/table-engines/special/dictionary.md/#dictionary) or [EmbeddedRocksDB](../../engines/table-engines/integrations/embedded-rocksdb.md) and only the `LEFT` and `INNER` JOINs.
|
||
|
||
- auto
|
||
|
||
When set to `auto`, `hash` join is tried first, and the algorithm is switched on the fly to another algorithm if the memory limit is violated.
|
||
|
||
- full_sorting_merge
|
||
|
||
[Sort-merge algorithm](https://en.wikipedia.org/wiki/Sort-merge_join) with full sorting joined tables before joining.
|
||
|
||
- prefer_partial_merge
|
||
|
||
ClickHouse always tries to use `partial_merge` join if possible, otherwise, it uses `hash`. *Deprecated*, same as `partial_merge,hash`.
|
||
|
||
|
||
## join_any_take_last_row {#settings-join_any_take_last_row}
|
||
|
||
Changes the behaviour of join operations with `ANY` strictness.
|
||
|
||
:::warning
|
||
This setting applies only for `JOIN` operations with [Join](../../engines/table-engines/special/join.md) engine tables.
|
||
:::
|
||
|
||
Possible values:
|
||
|
||
- 0 — If the right table has more than one matching row, only the first one found is joined.
|
||
- 1 — If the right table has more than one matching row, only the last one found is joined.
|
||
|
||
Default value: 0.
|
||
|
||
See also:
|
||
|
||
- [JOIN clause](../../sql-reference/statements/select/join.md/#select-join)
|
||
- [Join table engine](../../engines/table-engines/special/join.md)
|
||
- [join_default_strictness](#settings-join_default_strictness)
|
||
|
||
## join_use_nulls {#join_use_nulls}
|
||
|
||
Sets the type of [JOIN](../../sql-reference/statements/select/join.md) behaviour. When merging tables, empty cells may appear. ClickHouse fills them differently based on this setting.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The empty cells are filled with the default value of the corresponding field type.
|
||
- 1 — `JOIN` behaves the same way as in standard SQL. The type of the corresponding field is converted to [Nullable](../../sql-reference/data-types/nullable.md/#data_type-nullable), and empty cells are filled with [NULL](../../sql-reference/syntax.md).
|
||
|
||
Default value: 0.
|
||
|
||
## group_by_use_nulls {#group_by_use_nulls}
|
||
|
||
Changes the way the [GROUP BY clause](/docs/en/sql-reference/statements/select/group-by.md) treats the types of aggregation keys.
|
||
When the `ROLLUP`, `CUBE`, or `GROUPING SETS` specifiers are used, some aggregation keys may not be used to produce some result rows.
|
||
Columns for these keys are filled with either default value or `NULL` in corresponding rows depending on this setting.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The default value for the aggregation key type is used to produce missing values.
|
||
- 1 — ClickHouse executes `GROUP BY` the same way as the SQL standard says. The types of aggregation keys are converted to [Nullable](/docs/en/sql-reference/data-types/nullable.md/#data_type-nullable). Columns for corresponding aggregation keys are filled with [NULL](/docs/en/sql-reference/syntax.md) for rows that didn't use it.
|
||
|
||
Default value: 0.
|
||
|
||
See also:
|
||
|
||
- [GROUP BY clause](/docs/en/sql-reference/statements/select/group-by.md)
|
||
|
||
## partial_merge_join_optimizations {#partial_merge_join_optimizations}
|
||
|
||
Disables optimizations in partial merge join algorithm for [JOIN](../../sql-reference/statements/select/join.md) queries.
|
||
|
||
By default, this setting enables improvements that could lead to wrong results. If you see suspicious results in your queries, disable optimizations by this setting. Optimizations can be different in different versions of the ClickHouse server.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Optimizations disabled.
|
||
- 1 — Optimizations enabled.
|
||
|
||
Default value: 1.
|
||
|
||
## partial_merge_join_rows_in_right_blocks {#partial_merge_join_rows_in_right_blocks}
|
||
|
||
Limits sizes of right-hand join data blocks in partial merge join algorithm for [JOIN](../../sql-reference/statements/select/join.md) queries.
|
||
|
||
ClickHouse server:
|
||
|
||
1. Splits right-hand join data into blocks with up to the specified number of rows.
|
||
2. Indexes each block with its minimum and maximum values.
|
||
3. Unloads prepared blocks to disk if it is possible.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer. Recommended range of values: \[1000, 100000\].
|
||
|
||
Default value: 65536.
|
||
|
||
## join_on_disk_max_files_to_merge {#join_on_disk_max_files_to_merge}
|
||
|
||
Limits the number of files allowed for parallel sorting in MergeJoin operations when they are executed on disk.
|
||
|
||
The bigger the value of the setting, the more RAM is used and the less disk I/O is needed.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer, starting from 2.
|
||
|
||
Default value: 64.
|
||
|
||
## any_join_distinct_right_table_keys {#any_join_distinct_right_table_keys}
|
||
|
||
Enables legacy ClickHouse server behaviour in `ANY INNER|LEFT JOIN` operations.
|
||
|
||
:::warning
|
||
Use this setting only for backward compatibility if your use cases depend on legacy `JOIN` behaviour.
|
||
:::
|
||
|
||
When the legacy behaviour is enabled:
|
||
|
||
- Results of `t1 ANY LEFT JOIN t2` and `t2 ANY RIGHT JOIN t1` operations are not equal because ClickHouse uses the logic with many-to-one left-to-right table keys mapping.
|
||
- Results of `ANY INNER JOIN` operations contain all rows from the left table like the `SEMI LEFT JOIN` operations do.
|
||
|
||
When the legacy behaviour is disabled:
|
||
|
||
- Results of `t1 ANY LEFT JOIN t2` and `t2 ANY RIGHT JOIN t1` operations are equal because ClickHouse uses the logic which provides one-to-many keys mapping in `ANY RIGHT JOIN` operations.
|
||
- Results of `ANY INNER JOIN` operations contain one row per key from both the left and right tables.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Legacy behaviour is disabled.
|
||
- 1 — Legacy behaviour is enabled.
|
||
|
||
Default value: 0.
|
||
|
||
See also:
|
||
|
||
- [JOIN strictness](../../sql-reference/statements/select/join.md/#join-settings)
|
||
|
||
## temporary_files_codec {#temporary_files_codec}
|
||
|
||
Sets compression codec for temporary files used in sorting and joining operations on disk.
|
||
|
||
Possible values:
|
||
|
||
- LZ4 — [LZ4](https://en.wikipedia.org/wiki/LZ4_(compression_algorithm)) compression is applied.
|
||
- NONE — No compression is applied.
|
||
|
||
Default value: LZ4.
|
||
|
||
## max_block_size {#setting-max_block_size}
|
||
|
||
In ClickHouse, data is processed by blocks (sets of column parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block. The `max_block_size` setting is a recommendation for what size of the block (in a count of rows) to load from tables. The block size shouldn’t be too small, so that the expenditures on each block are still noticeable, but not too large so that the query with LIMIT that is completed after the first block is processed quickly. The goal is to avoid consuming too much memory when extracting a large number of columns in multiple threads and to preserve at least some cache locality.
|
||
|
||
Default value: 65,536.
|
||
|
||
Blocks the size of `max_block_size` are not always loaded from the table. If it is obvious that less data needs to be retrieved, a smaller block is processed.
|
||
|
||
## preferred_block_size_bytes {#preferred-block-size-bytes}
|
||
|
||
Used for the same purpose as `max_block_size`, but it sets the recommended block size in bytes by adapting it to the number of rows in the block.
|
||
However, the block size cannot be more than `max_block_size` rows.
|
||
By default: 1,000,000. It only works when reading from MergeTree engines.
|
||
|
||
## merge_tree_min_rows_for_concurrent_read {#setting-merge-tree-min-rows-for-concurrent-read}
|
||
|
||
If the number of rows to be read from a file of a [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) table exceeds `merge_tree_min_rows_for_concurrent_read` then ClickHouse tries to perform a concurrent reading from this file on several threads.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `163840`.
|
||
|
||
## merge_tree_min_rows_for_concurrent_read_for_remote_filesystem {#merge-tree-min-rows-for-concurrent-read-for-remote-filesystem}
|
||
|
||
The minimum number of lines to read from one file before the [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) engine can parallelize reading, when reading from remote filesystem.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `163840`.
|
||
|
||
## merge_tree_min_bytes_for_concurrent_read {#setting-merge-tree-min-bytes-for-concurrent-read}
|
||
|
||
If the number of bytes to read from one file of a [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md)-engine table exceeds `merge_tree_min_bytes_for_concurrent_read`, then ClickHouse tries to concurrently read from this file in several threads.
|
||
|
||
Possible value:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `251658240`.
|
||
|
||
## merge_tree_min_bytes_for_concurrent_read_for_remote_filesystem {#merge-tree-min-bytes-for-concurrent-read-for-remote-filesystem}
|
||
|
||
The minimum number of bytes to read from one file before [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) engine can parallelize reading, when reading from remote filesystem.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `251658240`.
|
||
|
||
## merge_tree_min_rows_for_seek {#setting-merge-tree-min-rows-for-seek}
|
||
|
||
If the distance between two data blocks to be read in one file is less than `merge_tree_min_rows_for_seek` rows, then ClickHouse does not seek through the file but reads the data sequentially.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 0.
|
||
|
||
## merge_tree_min_bytes_for_seek {#setting-merge-tree-min-bytes-for-seek}
|
||
|
||
If the distance between two data blocks to be read in one file is less than `merge_tree_min_bytes_for_seek` bytes, then ClickHouse sequentially reads a range of file that contains both blocks, thus avoiding extra seek.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 0.
|
||
|
||
## merge_tree_coarse_index_granularity {#setting-merge-tree-coarse-index-granularity}
|
||
|
||
When searching for data, ClickHouse checks the data marks in the index file. If ClickHouse finds that required keys are in some range, it divides this range into `merge_tree_coarse_index_granularity` subranges and searches the required keys there recursively.
|
||
|
||
Possible values:
|
||
|
||
- Any positive even integer.
|
||
|
||
Default value: 8.
|
||
|
||
## merge_tree_max_rows_to_use_cache {#setting-merge-tree-max-rows-to-use-cache}
|
||
|
||
If ClickHouse should read more than `merge_tree_max_rows_to_use_cache` rows in one query, it does not use the cache of uncompressed blocks.
|
||
|
||
The cache of uncompressed blocks stores data extracted for queries. ClickHouse uses this cache to speed up responses to repeated small queries. This setting protects the cache from trashing by queries that read a large amount of data. The [uncompressed_cache_size](../../operations/server-configuration-parameters/settings.md/#server-settings-uncompressed_cache_size) server setting defines the size of the cache of uncompressed blocks.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 128 ✕ 8192.
|
||
|
||
## merge_tree_max_bytes_to_use_cache {#setting-merge-tree-max-bytes-to-use-cache}
|
||
|
||
If ClickHouse should read more than `merge_tree_max_bytes_to_use_cache` bytes in one query, it does not use the cache of uncompressed blocks.
|
||
|
||
The cache of uncompressed blocks stores data extracted for queries. ClickHouse uses this cache to speed up responses to repeated small queries. This setting protects the cache from trashing by queries that read a large amount of data. The [uncompressed_cache_size](../../operations/server-configuration-parameters/settings.md/#server-settings-uncompressed_cache_size) server setting defines the size of the cache of uncompressed blocks.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 2013265920.
|
||
|
||
## min_bytes_to_use_direct_io {#settings-min-bytes-to-use-direct-io}
|
||
|
||
The minimum data volume required for using direct I/O access to the storage disk.
|
||
|
||
ClickHouse uses this setting when reading data from tables. If the total storage volume of all the data to be read exceeds `min_bytes_to_use_direct_io` bytes, then ClickHouse reads the data from the storage disk with the `O_DIRECT` option.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Direct I/O is disabled.
|
||
- Positive integer.
|
||
|
||
Default value: 0.
|
||
|
||
## network_compression_method {#network_compression_method}
|
||
|
||
Sets the method of data compression that is used for communication between servers and between server and [clickhouse-client](../../interfaces/cli.md).
|
||
|
||
Possible values:
|
||
|
||
- `LZ4` — sets LZ4 compression method.
|
||
- `ZSTD` — sets ZSTD compression method.
|
||
|
||
Default value: `LZ4`.
|
||
|
||
**See Also**
|
||
|
||
- [network_zstd_compression_level](#network_zstd_compression_level)
|
||
|
||
## network_zstd_compression_level {#network_zstd_compression_level}
|
||
|
||
Adjusts the level of ZSTD compression. Used only when [network_compression_method](#network_compression_method) is set to `ZSTD`.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer from 1 to 15.
|
||
|
||
Default value: `1`.
|
||
|
||
## log_queries {#settings-log-queries}
|
||
|
||
Setting up query logging.
|
||
|
||
Queries sent to ClickHouse with this setup are logged according to the rules in the [query_log](../../operations/server-configuration-parameters/settings.md/#server_configuration_parameters-query-log) server configuration parameter.
|
||
|
||
Example:
|
||
|
||
``` text
|
||
log_queries=1
|
||
```
|
||
|
||
## log_queries_min_query_duration_ms {#settings-log-queries-min-query-duration-ms}
|
||
|
||
If enabled (non-zero), queries faster than the value of this setting will not be logged (you can think about this as a `long_query_time` for [MySQL Slow Query Log](https://dev.mysql.com/doc/refman/5.7/en/slow-query-log.html)), and this basically means that you will not find them in the following tables:
|
||
|
||
- `system.query_log`
|
||
- `system.query_thread_log`
|
||
|
||
Only the queries with the following type will get to the log:
|
||
|
||
- `QUERY_FINISH`
|
||
- `EXCEPTION_WHILE_PROCESSING`
|
||
|
||
- Type: milliseconds
|
||
- Default value: 0 (any query)
|
||
|
||
## log_queries_min_type {#settings-log-queries-min-type}
|
||
|
||
`query_log` minimal type to log.
|
||
|
||
Possible values:
|
||
- `QUERY_START` (`=1`)
|
||
- `QUERY_FINISH` (`=2`)
|
||
- `EXCEPTION_BEFORE_START` (`=3`)
|
||
- `EXCEPTION_WHILE_PROCESSING` (`=4`)
|
||
|
||
Default value: `QUERY_START`.
|
||
|
||
Can be used to limit which entities will go to `query_log`, say you are interested only in errors, then you can use `EXCEPTION_WHILE_PROCESSING`:
|
||
|
||
``` text
|
||
log_queries_min_type='EXCEPTION_WHILE_PROCESSING'
|
||
```
|
||
|
||
## log_query_threads {#settings-log-query-threads}
|
||
|
||
Setting up query threads logging.
|
||
|
||
Query threads log into the [system.query_thread_log](../../operations/system-tables/query_thread_log.md) table. This setting has effect only when [log_queries](#settings-log-queries) is true. Queries’ threads run by ClickHouse with this setup are logged according to the rules in the [query_thread_log](../../operations/server-configuration-parameters/settings.md/#server_configuration_parameters-query_thread_log) server configuration parameter.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
**Example**
|
||
|
||
``` text
|
||
log_query_threads=1
|
||
```
|
||
|
||
## log_query_views {#settings-log-query-views}
|
||
|
||
Setting up query views logging.
|
||
|
||
When a query run by ClickHouse with this setting enabled has associated views (materialized or live views), they are logged in the [query_views_log](../../operations/server-configuration-parameters/settings.md/#server_configuration_parameters-query_views_log) server configuration parameter.
|
||
|
||
Example:
|
||
|
||
``` text
|
||
log_query_views=1
|
||
```
|
||
|
||
## log_formatted_queries {#settings-log-formatted-queries}
|
||
|
||
Allows to log formatted queries to the [system.query_log](../../operations/system-tables/query_log.md) system table (populates `formatted_query` column in the [system.query_log](../../operations/system-tables/query_log.md)).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Formatted queries are not logged in the system table.
|
||
- 1 — Formatted queries are logged in the system table.
|
||
|
||
Default value: `0`.
|
||
|
||
## log_comment {#settings-log-comment}
|
||
|
||
Specifies the value for the `log_comment` field of the [system.query_log](../system-tables/query_log.md) table and comment text for the server log.
|
||
|
||
It can be used to improve the readability of server logs. Additionally, it helps to select queries related to the test from the `system.query_log` after running [clickhouse-test](../../development/tests.md).
|
||
|
||
Possible values:
|
||
|
||
- Any string no longer than [max_query_size](#settings-max_query_size). If the max_query_size is exceeded, the server throws an exception.
|
||
|
||
Default value: empty string.
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
SET log_comment = 'log_comment test', log_queries = 1;
|
||
SELECT 1;
|
||
SYSTEM FLUSH LOGS;
|
||
SELECT type, query FROM system.query_log WHERE log_comment = 'log_comment test' AND event_date >= yesterday() ORDER BY event_time DESC LIMIT 2;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌─type────────┬─query─────┐
|
||
│ QueryStart │ SELECT 1; │
|
||
│ QueryFinish │ SELECT 1; │
|
||
└─────────────┴───────────┘
|
||
```
|
||
|
||
## log_processors_profiles {#settings-log_processors_profiles}
|
||
|
||
Write time that processor spent during execution/waiting for data to `system.processors_profile_log` table.
|
||
|
||
See also:
|
||
|
||
- [`system.processors_profile_log`](../../operations/system-tables/processors_profile_log.md#system-processors_profile_log)
|
||
- [`EXPLAIN PIPELINE`](../../sql-reference/statements/explain.md#explain-pipeline)
|
||
|
||
## max_insert_block_size {#settings-max_insert_block_size}
|
||
|
||
The size of blocks (in a count of rows) to form for insertion into a table.
|
||
This setting only applies in cases when the server forms the blocks.
|
||
For example, for an INSERT via the HTTP interface, the server parses the data format and forms blocks of the specified size.
|
||
But when using clickhouse-client, the client parses the data itself, and the ‘max_insert_block_size’ setting on the server does not affect the size of the inserted blocks.
|
||
The setting also does not have a purpose when using INSERT SELECT, since data is inserted using the same blocks that are formed after SELECT.
|
||
|
||
Default value: 1,048,576.
|
||
|
||
The default is slightly more than `max_block_size`. The reason for this is that certain table engines (`*MergeTree`) form a data part on the disk for each inserted block, which is a fairly large entity. Similarly, `*MergeTree` tables sort data during insertion, and a large enough block size allow sorting more data in RAM.
|
||
|
||
## min_insert_block_size_rows {#min-insert-block-size-rows}
|
||
|
||
Sets the minimum number of rows in the block that can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Squashing disabled.
|
||
|
||
Default value: 1048576.
|
||
|
||
## min_insert_block_size_bytes {#min-insert-block-size-bytes}
|
||
|
||
Sets the minimum number of bytes in the block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Squashing disabled.
|
||
|
||
Default value: 268435456.
|
||
|
||
## max_replica_delay_for_distributed_queries {#settings-max_replica_delay_for_distributed_queries}
|
||
|
||
Disables lagging replicas for distributed queries. See [Replication](../../engines/table-engines/mergetree-family/replication.md).
|
||
|
||
Sets the time in seconds. If a replica's lag is greater than or equal to the set value, this replica is not used.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Replica lags are not checked.
|
||
|
||
To prevent the use of any replica with a non-zero lag, set this parameter to 1.
|
||
|
||
Default value: 300.
|
||
|
||
Used when performing `SELECT` from a distributed table that points to replicated tables.
|
||
|
||
## max_threads {#settings-max_threads}
|
||
|
||
The maximum number of query processing threads, excluding threads for retrieving data from remote servers (see the ‘max_distributed_connections’ parameter).
|
||
|
||
This parameter applies to threads that perform the same stages of the query processing pipeline in parallel.
|
||
For example, when reading from a table, if it is possible to evaluate expressions with functions, filter with WHERE and pre-aggregate for GROUP BY in parallel using at least ‘max_threads’ number of threads, then ‘max_threads’ are used.
|
||
|
||
Default value: the number of physical CPU cores.
|
||
|
||
For queries that are completed quickly because of a LIMIT, you can set a lower ‘max_threads’. For example, if the necessary number of entries are located in every block and max_threads = 8, then 8 blocks are retrieved, although it would have been enough to read just one.
|
||
|
||
The smaller the `max_threads` value, the less memory is consumed.
|
||
|
||
## max_insert_threads {#settings-max-insert-threads}
|
||
|
||
The maximum number of threads to execute the `INSERT SELECT` query.
|
||
|
||
Possible values:
|
||
|
||
- 0 (or 1) — `INSERT SELECT` no parallel execution.
|
||
- Positive integer. Bigger than 1.
|
||
|
||
Default value: 0.
|
||
|
||
Parallel `INSERT SELECT` has effect only if the `SELECT` part is executed in parallel, see [max_threads](#settings-max_threads) setting.
|
||
Higher values will lead to higher memory usage.
|
||
|
||
## max_compress_block_size {#max-compress-block-size}
|
||
|
||
The maximum size of blocks of uncompressed data before compressing for writing to a table. By default, 1,048,576 (1 MiB). Specifying a smaller block size generally leads to slightly reduced compression ratio, the compression and decompression speed increases slightly due to cache locality, and memory consumption is reduced.
|
||
|
||
:::warning
|
||
This is an expert-level setting, and you shouldn't change it if you're just getting started with ClickHouse.
|
||
:::
|
||
|
||
Don’t confuse blocks for compression (a chunk of memory consisting of bytes) with blocks for query processing (a set of rows from a table).
|
||
|
||
## min_compress_block_size {#min-compress-block-size}
|
||
|
||
For [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) tables. In order to reduce latency when processing queries, a block is compressed when writing the next mark if its size is at least `min_compress_block_size`. By default, 65,536.
|
||
|
||
The actual size of the block, if the uncompressed data is less than `max_compress_block_size`, is no less than this value and no less than the volume of data for one mark.
|
||
|
||
Let’s look at an example. Assume that `index_granularity` was set to 8192 during table creation.
|
||
|
||
We are writing a UInt32-type column (4 bytes per value). When writing 8192 rows, the total will be 32 KB of data. Since min_compress_block_size = 65,536, a compressed block will be formed for every two marks.
|
||
|
||
We are writing a URL column with the String type (average size of 60 bytes per value). When writing 8192 rows, the average will be slightly less than 500 KB of data. Since this is more than 65,536, a compressed block will be formed for each mark. In this case, when reading data from the disk in the range of a single mark, extra data won’t be decompressed.
|
||
|
||
:::warning
|
||
This is an expert-level setting, and you shouldn't change it if you're just getting started with ClickHouse.
|
||
:::
|
||
|
||
## max_query_size {#settings-max_query_size}
|
||
|
||
The maximum part of a query that can be taken to RAM for parsing with the SQL parser.
|
||
The INSERT query also contains data for INSERT that is processed by a separate stream parser (that consumes O(1) RAM), which is not included in this restriction.
|
||
|
||
Default value: 256 KiB.
|
||
|
||
## max_parser_depth {#max_parser_depth}
|
||
|
||
Limits maximum recursion depth in the recursive descent parser. Allows controlling the stack size.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Recursion depth is unlimited.
|
||
|
||
Default value: 1000.
|
||
|
||
## interactive_delay {#interactive-delay}
|
||
|
||
The interval in microseconds for checking whether request execution has been canceled and sending the progress.
|
||
|
||
Default value: 100,000 (checks for cancelling and sends the progress ten times per second).
|
||
|
||
## connect_timeout, receive_timeout, send_timeout {#connect-timeout-receive-timeout-send-timeout}
|
||
|
||
Timeouts in seconds on the socket used for communicating with the client.
|
||
|
||
Default value: 10, 300, 300.
|
||
|
||
## cancel_http_readonly_queries_on_client_close {#cancel-http-readonly-queries-on-client-close}
|
||
|
||
Cancels HTTP read-only queries (e.g. SELECT) when a client closes the connection without waiting for the response.
|
||
|
||
Default value: 0
|
||
|
||
## poll_interval {#poll-interval}
|
||
|
||
Lock in a wait loop for the specified number of seconds.
|
||
|
||
Default value: 10.
|
||
|
||
## max_distributed_connections {#max-distributed-connections}
|
||
|
||
The maximum number of simultaneous connections with remote servers for distributed processing of a single query to a single Distributed table. We recommend setting a value no less than the number of servers in the cluster.
|
||
|
||
Default value: 1024.
|
||
|
||
The following parameters are only used when creating Distributed tables (and when launching a server), so there is no reason to change them at runtime.
|
||
|
||
## distributed_connections_pool_size {#distributed-connections-pool-size}
|
||
|
||
The maximum number of simultaneous connections with remote servers for distributed processing of all queries to a single Distributed table. We recommend setting a value no less than the number of servers in the cluster.
|
||
|
||
Default value: 1024.
|
||
|
||
## max_distributed_depth {#max-distributed-depth}
|
||
|
||
Limits the maximum depth of recursive queries for [Distributed](../../engines/table-engines/special/distributed.md) tables.
|
||
|
||
If the value is exceeded, the server throws an exception.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Unlimited depth.
|
||
|
||
Default value: `5`.
|
||
|
||
## max_replicated_fetches_network_bandwidth_for_server {#max_replicated_fetches_network_bandwidth_for_server}
|
||
|
||
Limits the maximum speed of data exchange over the network in bytes per second for [replicated](../../engines/table-engines/mergetree-family/replication.md) fetches for the server. Only has meaning at server startup. You can also limit the speed for a particular table with [max_replicated_fetches_network_bandwidth](../../operations/settings/merge-tree-settings.md/#max_replicated_fetches_network_bandwidth) setting.
|
||
|
||
The setting isn't followed perfectly accurately.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Unlimited.
|
||
|
||
Default value: `0`.
|
||
|
||
**Usage**
|
||
|
||
Could be used for throttling speed when replicating the data to add or replace new nodes.
|
||
|
||
:::note
|
||
60000000 bytes/s approximatly corresponds to 457 Mbps (60000000 / 1024 / 1024 * 8).
|
||
:::
|
||
|
||
## max_replicated_sends_network_bandwidth_for_server {#max_replicated_sends_network_bandwidth_for_server}
|
||
|
||
Limits the maximum speed of data exchange over the network in bytes per second for [replicated](../../engines/table-engines/mergetree-family/replication.md) sends for the server. Only has meaning at server startup. You can also limit the speed for a particular table with [max_replicated_sends_network_bandwidth](../../operations/settings/merge-tree-settings.md/#max_replicated_sends_network_bandwidth) setting.
|
||
|
||
The setting isn't followed perfectly accurately.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Unlimited.
|
||
|
||
Default value: `0`.
|
||
|
||
**Usage**
|
||
|
||
Could be used for throttling speed when replicating the data to add or replace new nodes.
|
||
|
||
:::note
|
||
60000000 bytes/s approximatly corresponds to 457 Mbps (60000000 / 1024 / 1024 * 8).
|
||
:::
|
||
|
||
## connect_timeout_with_failover_ms {#connect-timeout-with-failover-ms}
|
||
|
||
The timeout in milliseconds for connecting to a remote server for a Distributed table engine, if the ‘shard’ and ‘replica’ sections are used in the cluster definition.
|
||
If unsuccessful, several attempts are made to connect to various replicas.
|
||
|
||
Default value: 50.
|
||
|
||
## connection_pool_max_wait_ms {#connection-pool-max-wait-ms}
|
||
|
||
The wait time in milliseconds for a connection when the connection pool is full.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Infinite timeout.
|
||
|
||
Default value: 0.
|
||
|
||
## connections_with_failover_max_tries {#connections-with-failover-max-tries}
|
||
|
||
The maximum number of connection attempts with each replica for the Distributed table engine.
|
||
|
||
Default value: 3.
|
||
|
||
## extremes {#extremes}
|
||
|
||
Whether to count extreme values (the minimums and maximums in columns of a query result). Accepts 0 or 1. By default, 0 (disabled).
|
||
For more information, see the section “Extreme values”.
|
||
|
||
## kafka_max_wait_ms {#kafka-max-wait-ms}
|
||
|
||
The wait time in milliseconds for reading messages from [Kafka](../../engines/table-engines/integrations/kafka.md/#kafka) before retry.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Infinite timeout.
|
||
|
||
Default value: 5000.
|
||
|
||
See also:
|
||
|
||
- [Apache Kafka](https://kafka.apache.org/)
|
||
|
||
## kafka_disable_num_consumers_limit {#kafka-disable-num-consumers-limit}
|
||
|
||
Disable limit on kafka_num_consumers that depends on the number of available CPU cores.
|
||
|
||
Default value: false.
|
||
|
||
## use_uncompressed_cache {#setting-use_uncompressed_cache}
|
||
|
||
Whether to use a cache of uncompressed blocks. Accepts 0 or 1. By default, 0 (disabled).
|
||
Using the uncompressed cache (only for tables in the MergeTree family) can significantly reduce latency and increase throughput when working with a large number of short queries. Enable this setting for users who send frequent short requests. Also pay attention to the [uncompressed_cache_size](../../operations/server-configuration-parameters/settings.md/#server-settings-uncompressed_cache_size) configuration parameter (only set in the config file) – the size of uncompressed cache blocks. By default, it is 8 GiB. The uncompressed cache is filled in as needed and the least-used data is automatically deleted.
|
||
|
||
For queries that read at least a somewhat large volume of data (one million rows or more), the uncompressed cache is disabled automatically to save space for truly small queries. This means that you can keep the ‘use_uncompressed_cache’ setting always set to 1.
|
||
|
||
## replace_running_query {#replace-running-query}
|
||
|
||
When using the HTTP interface, the ‘query_id’ parameter can be passed. This is any string that serves as the query identifier.
|
||
If a query from the same user with the same ‘query_id’ already exists at this time, the behaviour depends on the ‘replace_running_query’ parameter.
|
||
|
||
`0` (default) – Throw an exception (do not allow the query to run if a query with the same ‘query_id’ is already running).
|
||
|
||
`1` – Cancel the old query and start running the new one.
|
||
|
||
Set this parameter to 1 for implementing suggestions for segmentation conditions. After entering the next character, if the old query hasn’t finished yet, it should be cancelled.
|
||
|
||
## replace_running_query_max_wait_ms {#replace-running-query-max-wait-ms}
|
||
|
||
The wait time for running the query with the same `query_id` to finish, when the [replace_running_query](#replace-running-query) setting is active.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Throwing an exception that does not allow to run a new query if the server already executes a query with the same `query_id`.
|
||
|
||
Default value: 5000.
|
||
|
||
## stream_flush_interval_ms {#stream-flush-interval-ms}
|
||
|
||
Works for tables with streaming in the case of a timeout, or when a thread generates [max_insert_block_size](#settings-max_insert_block_size) rows.
|
||
|
||
The default value is 7500.
|
||
|
||
The smaller the value, the more often data is flushed into the table. Setting the value too low leads to poor performance.
|
||
|
||
## stream_poll_timeout_ms {#stream_poll_timeout_ms}
|
||
|
||
Timeout for polling data from/to streaming storages.
|
||
|
||
Default value: 500.
|
||
|
||
## load_balancing {#settings-load_balancing}
|
||
|
||
Specifies the algorithm of replicas selection that is used for distributed query processing.
|
||
|
||
ClickHouse supports the following algorithms of choosing replicas:
|
||
|
||
- [Random](#load_balancing-random) (by default)
|
||
- [Nearest hostname](#load_balancing-nearest_hostname)
|
||
- [In order](#load_balancing-in_order)
|
||
- [First or random](#load_balancing-first_or_random)
|
||
- [Round robin](#load_balancing-round_robin)
|
||
|
||
See also:
|
||
|
||
- [distributed_replica_max_ignored_errors](#settings-distributed_replica_max_ignored_errors)
|
||
|
||
### Random (by Default) {#load_balancing-random}
|
||
|
||
``` sql
|
||
load_balancing = random
|
||
```
|
||
|
||
The number of errors is counted for each replica. The query is sent to the replica with the fewest errors, and if there are several of these, to anyone of them.
|
||
Disadvantages: Server proximity is not accounted for; if the replicas have different data, you will also get different data.
|
||
|
||
### Nearest Hostname {#load_balancing-nearest_hostname}
|
||
|
||
``` sql
|
||
load_balancing = nearest_hostname
|
||
```
|
||
|
||
The number of errors is counted for each replica. Every 5 minutes, the number of errors is integrally divided by 2. Thus, the number of errors is calculated for a recent time with exponential smoothing. If there is one replica with a minimal number of errors (i.e. errors occurred recently on the other replicas), the query is sent to it. If there are multiple replicas with the same minimal number of errors, the query is sent to the replica with a hostname that is most similar to the server’s hostname in the config file (for the number of different characters in identical positions, up to the minimum length of both hostnames).
|
||
|
||
For instance, example01-01-1 and example01-01-2 are different in one position, while example01-01-1 and example01-02-2 differ in two places.
|
||
This method might seem primitive, but it does not require external data about network topology, and it does not compare IP addresses, which would be complicated for our IPv6 addresses.
|
||
|
||
Thus, if there are equivalent replicas, the closest one by name is preferred.
|
||
We can also assume that when sending a query to the same server, in the absence of failures, a distributed query will also go to the same servers. So even if different data is placed on the replicas, the query will return mostly the same results.
|
||
|
||
### In Order {#load_balancing-in_order}
|
||
|
||
``` sql
|
||
load_balancing = in_order
|
||
```
|
||
|
||
Replicas with the same number of errors are accessed in the same order as they are specified in the configuration.
|
||
This method is appropriate when you know exactly which replica is preferable.
|
||
|
||
### First or Random {#load_balancing-first_or_random}
|
||
|
||
``` sql
|
||
load_balancing = first_or_random
|
||
```
|
||
|
||
This algorithm chooses the first replica in the set or a random replica if the first is unavailable. It’s effective in cross-replication topology setups, but useless in other configurations.
|
||
|
||
The `first_or_random` algorithm solves the problem of the `in_order` algorithm. With `in_order`, if one replica goes down, the next one gets a double load while the remaining replicas handle the usual amount of traffic. When using the `first_or_random` algorithm, the load is evenly distributed among replicas that are still available.
|
||
|
||
It's possible to explicitly define what the first replica is by using the setting `load_balancing_first_offset`. This gives more control to rebalance query workloads among replicas.
|
||
|
||
### Round Robin {#load_balancing-round_robin}
|
||
|
||
``` sql
|
||
load_balancing = round_robin
|
||
```
|
||
|
||
This algorithm uses a round-robin policy across replicas with the same number of errors (only the queries with `round_robin` policy is accounted).
|
||
|
||
## prefer_localhost_replica {#settings-prefer-localhost-replica}
|
||
|
||
Enables/disables preferable using the localhost replica when processing distributed queries.
|
||
|
||
Possible values:
|
||
|
||
- 1 — ClickHouse always sends a query to the localhost replica if it exists.
|
||
- 0 — ClickHouse uses the balancing strategy specified by the [load_balancing](#settings-load_balancing) setting.
|
||
|
||
Default value: 1.
|
||
|
||
:::warning
|
||
Disable this setting if you use [max_parallel_replicas](#settings-max_parallel_replicas) without [parallel_replicas_custom_key](#settings-parallel_replicas_custom_key).
|
||
If [parallel_replicas_custom_key](#settings-parallel_replicas_custom_key) is set, disable this setting only if it's used on a cluster with multiple shards containing multiple replicas.
|
||
If it's used on a cluster with a single shard and multiple replicas, disabling this setting will have negative effects.
|
||
:::
|
||
|
||
## totals_mode {#totals-mode}
|
||
|
||
How to calculate TOTALS when HAVING is present, as well as when max_rows_to_group_by and group_by_overflow_mode = ‘any’ are present.
|
||
See the section “WITH TOTALS modifier”.
|
||
|
||
## totals_auto_threshold {#totals-auto-threshold}
|
||
|
||
The threshold for `totals_mode = 'auto'`.
|
||
See the section “WITH TOTALS modifier”.
|
||
|
||
## max_parallel_replicas {#settings-max_parallel_replicas}
|
||
|
||
The maximum number of replicas for each shard when executing a query.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `1`.
|
||
|
||
**Additional Info**
|
||
|
||
This options will produce different results depending on the settings used.
|
||
|
||
:::warning
|
||
This setting will produce incorrect results when joins or subqueries are involved, and all tables don't meet certain requirements. See [Distributed Subqueries and max_parallel_replicas](../../sql-reference/operators/in.md/#max_parallel_replica-subqueries) for more details.
|
||
:::
|
||
|
||
### Parallel processing using `SAMPLE` key
|
||
|
||
A query may be processed faster if it is executed on several servers in parallel. But the query performance may degrade in the following cases:
|
||
|
||
- The position of the sampling key in the partitioning key does not allow efficient range scans.
|
||
- Adding a sampling key to the table makes filtering by other columns less efficient.
|
||
- The sampling key is an expression that is expensive to calculate.
|
||
- The cluster latency distribution has a long tail, so that querying more servers increases the query overall latency.
|
||
|
||
### Parallel processing using [parallel_replicas_custom_key](#settings-parallel_replicas_custom_key)
|
||
|
||
This setting is useful for any replicated table.
|
||
|
||
## parallel_replicas_custom_key {#settings-parallel_replicas_custom_key}
|
||
|
||
An arbitrary integer expression that can be used to split work between replicas for a specific table.
|
||
The value can be any integer expression.
|
||
A query may be processed faster if it is executed on several servers in parallel but it depends on the used [parallel_replicas_custom_key](#settings-parallel_replicas_custom_key)
|
||
and [parallel_replicas_custom_key_filter_type](#settings-parallel_replicas_custom_key_filter_type).
|
||
|
||
Simple expressions using primary keys are preferred.
|
||
|
||
If the setting is used on a cluster that consists of a single shard with multiple replicas, those replicas will be converted into virtual shards.
|
||
Otherwise, it will behave same as for `SAMPLE` key, it will use multiple replicas of each shard.
|
||
|
||
## parallel_replicas_custom_key_filter_type {#settings-parallel_replicas_custom_key_filter_type}
|
||
|
||
How to use `parallel_replicas_custom_key` expression for splitting work between replicas.
|
||
|
||
Possible values:
|
||
|
||
- `default` — Use the default implementation using modulo operation on the `parallel_replicas_custom_key`.
|
||
- `range` — Split the entire value space of the expression in the ranges. This type of filtering is useful if values of `parallel_replicas_custom_key` are uniformly spread across the entire integer space, e.g. hash values.
|
||
|
||
Default value: `default`.
|
||
|
||
## compile_expressions {#compile-expressions}
|
||
|
||
Enables or disables compilation of frequently used simple functions and operators to native code with LLVM at runtime.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
## min_count_to_compile_expression {#min-count-to-compile-expression}
|
||
|
||
Minimum count of executing same expression before it is get compiled.
|
||
|
||
Default value: `3`.
|
||
|
||
## compile_aggregate_expressions {#compile_aggregate_expressions}
|
||
|
||
Enables or disables JIT-compilation of aggregate functions to native code. Enabling this setting can improve the performance.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Aggregation is done without JIT compilation.
|
||
- 1 — Aggregation is done using JIT compilation.
|
||
|
||
Default value: `1`.
|
||
|
||
**See Also**
|
||
|
||
- [min_count_to_compile_aggregate_expression](#min_count_to_compile_aggregate_expression)
|
||
|
||
## min_count_to_compile_aggregate_expression {#min_count_to_compile_aggregate_expression}
|
||
|
||
The minimum number of identical aggregate expressions to start JIT-compilation. Works only if the [compile_aggregate_expressions](#compile_aggregate_expressions) setting is enabled.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Identical aggregate expressions are always JIT-compiled.
|
||
|
||
Default value: `3`.
|
||
|
||
## use_query_cache {#use-query-cache}
|
||
|
||
If turned on, `SELECT` queries may utilize the [query cache](../query-cache.md). Parameters [enable_reads_from_query_cache](#enable-reads-from-query-cache)
|
||
and [enable_writes_to_query_cache](#enable-writes-to-query-cache) control in more detail how the cache is used.
|
||
|
||
Possible values:
|
||
|
||
- 0 - Yes
|
||
- 1 - No
|
||
|
||
Default value: `0`.
|
||
|
||
## enable_reads_from_query_cache {#enable-reads-from-query-cache}
|
||
|
||
If turned on, results of `SELECT` queries are retrieved from the [query cache](../query-cache.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 - Disabled
|
||
- 1 - Enabled
|
||
|
||
Default value: `1`.
|
||
|
||
## enable_writes_to_query_cache {#enable-writes-to-query-cache}
|
||
|
||
If turned on, results of `SELECT` queries are stored in the [query cache](../query-cache.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 - Disabled
|
||
- 1 - Enabled
|
||
|
||
Default value: `1`.
|
||
|
||
## query_cache_store_results_of_queries_with_nondeterministic_functions {#query--store-results-of-queries-with-nondeterministic-functions}
|
||
|
||
If turned on, then results of `SELECT` queries with non-deterministic functions (e.g. `rand()`, `now()`) can be cached in the [query cache](../query-cache.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 - Disabled
|
||
- 1 - Enabled
|
||
|
||
Default value: `0`.
|
||
|
||
## query_cache_min_query_runs {#query-cache-min-query-runs}
|
||
|
||
Minimum number of times a `SELECT` query must run before its result is stored in the [query cache](../query-cache.md).
|
||
|
||
Possible values:
|
||
|
||
- Positive integer >= 0.
|
||
|
||
Default value: `0`
|
||
|
||
## query_cache_min_query_duration {#query-cache-min-query-duration}
|
||
|
||
Minimum duration in milliseconds a query needs to run for its result to be stored in the [query cache](../query-cache.md).
|
||
|
||
Possible values:
|
||
|
||
- Positive integer >= 0.
|
||
|
||
Default value: `0`
|
||
|
||
## query_cache_ttl {#query-cache-ttl}
|
||
|
||
After this time in seconds entries in the [query cache](../query-cache.md) become stale.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer >= 0.
|
||
|
||
Default value: `60`
|
||
|
||
## query_cache_share_between_users {#query-cache-share-between-users}
|
||
|
||
If turned on, the result of `SELECT` queries cached in the [query cache](../query-cache.md) can be read by other users.
|
||
It is not recommended to enable this setting due to security reasons.
|
||
|
||
Possible values:
|
||
|
||
- 0 - Disabled
|
||
- 1 - Enabled
|
||
|
||
Default value: `0`.
|
||
|
||
## insert_quorum {#settings-insert_quorum}
|
||
|
||
Enables the quorum writes.
|
||
|
||
- If `insert_quorum < 2`, the quorum writes are disabled.
|
||
- If `insert_quorum >= 2`, the quorum writes are enabled.
|
||
- If `insert_quorum = 'auto'`, use majority number (`number_of_replicas / 2 + 1`) as quorum number.
|
||
|
||
Default value: 0 - disabled.
|
||
|
||
Quorum writes
|
||
|
||
`INSERT` succeeds only when ClickHouse manages to correctly write data to the `insert_quorum` of replicas during the `insert_quorum_timeout`. If for any reason the number of replicas with successful writes does not reach the `insert_quorum`, the write is considered failed and ClickHouse will delete the inserted block from all the replicas where data has already been written.
|
||
|
||
When `insert_quorum_parallel` is disabled, all replicas in the quorum are consistent, i.e. they contain data from all previous `INSERT` queries (the `INSERT` sequence is linearized). When reading data written using `insert_quorum` and `insert_quorum_parallel` is disabled, you can turn on sequential consistency for `SELECT` queries using [select_sequential_consistency](#settings-select_sequential_consistency).
|
||
|
||
ClickHouse generates an exception:
|
||
|
||
- If the number of available replicas at the time of the query is less than the `insert_quorum`.
|
||
- When `insert_quorum_parallel` is disabled and an attempt to write data is made when the previous block has not yet been inserted in `insert_quorum` of replicas. This situation may occur if the user tries to perform another `INSERT` query to the same table before the previous one with `insert_quorum` is completed.
|
||
|
||
See also:
|
||
|
||
- [insert_quorum_timeout](#settings-insert_quorum_timeout)
|
||
- [insert_quorum_parallel](#settings-insert_quorum_parallel)
|
||
- [select_sequential_consistency](#settings-select_sequential_consistency)
|
||
|
||
## insert_quorum_timeout {#settings-insert_quorum_timeout}
|
||
|
||
Write to a quorum timeout in milliseconds. If the timeout has passed and no write has taken place yet, ClickHouse will generate an exception and the client must repeat the query to write the same block to the same or any other replica.
|
||
|
||
Default value: 600 000 milliseconds (ten minutes).
|
||
|
||
See also:
|
||
|
||
- [insert_quorum](#settings-insert_quorum)
|
||
- [insert_quorum_parallel](#settings-insert_quorum_parallel)
|
||
- [select_sequential_consistency](#settings-select_sequential_consistency)
|
||
|
||
## insert_quorum_parallel {#settings-insert_quorum_parallel}
|
||
|
||
Enables or disables parallelism for quorum `INSERT` queries. If enabled, additional `INSERT` queries can be sent while previous queries have not yet finished. If disabled, additional writes to the same table will be rejected.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1.
|
||
|
||
See also:
|
||
|
||
- [insert_quorum](#settings-insert_quorum)
|
||
- [insert_quorum_timeout](#settings-insert_quorum_timeout)
|
||
- [select_sequential_consistency](#settings-select_sequential_consistency)
|
||
|
||
## select_sequential_consistency {#settings-select_sequential_consistency}
|
||
|
||
Enables or disables sequential consistency for `SELECT` queries. Requires `insert_quorum_parallel` to be disabled (enabled by default).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0.
|
||
|
||
Usage
|
||
|
||
When sequential consistency is enabled, ClickHouse allows the client to execute the `SELECT` query only for those replicas that contain data from all previous `INSERT` queries executed with `insert_quorum`. If the client refers to a partial replica, ClickHouse will generate an exception. The SELECT query will not include data that has not yet been written to the quorum of replicas.
|
||
|
||
When `insert_quorum_parallel` is enabled (the default), then `select_sequential_consistency` does not work. This is because parallel `INSERT` queries can be written to different sets of quorum replicas so there is no guarantee a single replica will have received all writes.
|
||
|
||
See also:
|
||
|
||
- [insert_quorum](#settings-insert_quorum)
|
||
- [insert_quorum_timeout](#settings-insert_quorum_timeout)
|
||
- [insert_quorum_parallel](#settings-insert_quorum_parallel)
|
||
|
||
## insert_deduplicate {#settings-insert-deduplicate}
|
||
|
||
Enables or disables block deduplication of `INSERT` (for Replicated\* tables).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1.
|
||
|
||
By default, blocks inserted into replicated tables by the `INSERT` statement are deduplicated (see [Data Replication](../../engines/table-engines/mergetree-family/replication.md)).
|
||
For the replicated tables by default the only 100 of the most recent blocks for each partition are deduplicated (see [replicated_deduplication_window](merge-tree-settings.md/#replicated-deduplication-window), [replicated_deduplication_window_seconds](merge-tree-settings.md/#replicated-deduplication-window-seconds)).
|
||
For not replicated tables see [non_replicated_deduplication_window](merge-tree-settings.md/#non-replicated-deduplication-window).
|
||
|
||
## Asynchronous Insert settings
|
||
### async_insert {#async-insert}
|
||
|
||
Enables or disables asynchronous inserts. This makes sense only for insertion over HTTP protocol. Note that deduplication isn't working for such inserts.
|
||
|
||
If enabled, the data is combined into batches before the insertion into tables, so it is possible to do small and frequent insertions into ClickHouse (up to 15000 queries per second) without buffer tables.
|
||
|
||
The data is inserted either after the [async_insert_max_data_size](#async-insert-max-data-size) is exceeded or after [async_insert_busy_timeout_ms](#async-insert-busy-timeout-ms) milliseconds since the first `INSERT` query. If the [async_insert_stale_timeout_ms](#async-insert-stale-timeout-ms) is set to a non-zero value, the data is inserted after `async_insert_stale_timeout_ms` milliseconds since the last query.
|
||
|
||
If [wait_for_async_insert](#wait-for-async-insert) is enabled, every client will wait for the data to be processed and flushed to the table. Otherwise, the query would be processed almost instantly, even if the data is not inserted.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Insertions are made synchronously, one after another.
|
||
- 1 — Multiple asynchronous insertions enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
### async_insert_threads {#async-insert-threads}
|
||
|
||
The maximum number of threads for background data parsing and insertion.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Asynchronous insertions are disabled.
|
||
|
||
Default value: `16`.
|
||
|
||
### wait_for_async_insert {#wait-for-async-insert}
|
||
|
||
Enables or disables waiting for processing of asynchronous insertion. If enabled, server will return `OK` only after the data is inserted. Otherwise, it will return `OK` even if the data wasn't inserted.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Server returns `OK` even if the data is not yet inserted.
|
||
- 1 — Server returns `OK` only after the data is inserted.
|
||
|
||
Default value: `1`.
|
||
|
||
### wait_for_async_insert_timeout {#wait-for-async-insert-timeout}
|
||
|
||
The timeout in seconds for waiting for processing of asynchronous insertion.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Disabled.
|
||
|
||
Default value: [lock_acquire_timeout](#lock_acquire_timeout).
|
||
|
||
### async_insert_max_data_size {#async-insert-max-data-size}
|
||
|
||
The maximum size of the unparsed data in bytes collected per query before being inserted.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Asynchronous insertions are disabled.
|
||
|
||
Default value: `100000`.
|
||
|
||
### async_insert_max_query_number {#async-insert-max-query-number}
|
||
|
||
The maximum number of insert queries per block before being inserted. This setting takes effect only if [async_insert_deduplicate](#settings-async-insert-deduplicate) is enabled.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Asynchronous insertions are disabled.
|
||
|
||
Default value: `450`.
|
||
|
||
### async_insert_busy_timeout_ms {#async-insert-busy-timeout-ms}
|
||
|
||
The maximum timeout in milliseconds since the first `INSERT` query before inserting collected data.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Timeout disabled.
|
||
|
||
Default value: `200`.
|
||
|
||
### async_insert_stale_timeout_ms {#async-insert-stale-timeout-ms}
|
||
|
||
The maximum timeout in milliseconds since the last `INSERT` query before dumping collected data. If enabled, the settings prolongs the [async_insert_busy_timeout_ms](#async-insert-busy-timeout-ms) with every `INSERT` query as long as [async_insert_max_data_size](#async-insert-max-data-size) is not exceeded.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Timeout disabled.
|
||
|
||
Default value: `0`.
|
||
### async_insert_deduplicate {#settings-async-insert-deduplicate}
|
||
|
||
Enables or disables insert deduplication of `ASYNC INSERT` (for Replicated\* tables).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1.
|
||
|
||
By default, async inserts are inserted into replicated tables by the `INSERT` statement enabling [async_isnert](#async-insert) are deduplicated (see [Data Replication](../../engines/table-engines/mergetree-family/replication.md)).
|
||
For the replicated tables, by default, only 10000 of the most recent inserts for each partition are deduplicated (see [replicated_deduplication_window_for_async_inserts](merge-tree-settings.md/#replicated-deduplication-window-async-inserts), [replicated_deduplication_window_seconds_for_async_inserts](merge-tree-settings.md/#replicated-deduplication-window-seconds-async-inserts)).
|
||
We recommend enabling the [async_block_ids_cache](merge-tree-settings.md/#use-async-block-ids-cache) to increase the efficiency of deduplication.
|
||
This function does not work for non-replicated tables.
|
||
|
||
## deduplicate_blocks_in_dependent_materialized_views {#settings-deduplicate-blocks-in-dependent-materialized-views}
|
||
|
||
Enables or disables the deduplication check for materialized views that receive data from Replicated\* tables.
|
||
|
||
Possible values:
|
||
|
||
0 — Disabled.
|
||
1 — Enabled.
|
||
|
||
Default value: 0.
|
||
|
||
Usage
|
||
|
||
By default, deduplication is not performed for materialized views but is done upstream, in the source table.
|
||
If an INSERTed block is skipped due to deduplication in the source table, there will be no insertion into attached materialized views. This behaviour exists to enable the insertion of highly aggregated data into materialized views, for cases where inserted blocks are the same after materialized view aggregation but derived from different INSERTs into the source table.
|
||
At the same time, this behaviour “breaks” `INSERT` idempotency. If an `INSERT` into the main table was successful and `INSERT` into a materialized view failed (e.g. because of communication failure with ClickHouse Keeper) a client will get an error and can retry the operation. However, the materialized view won’t receive the second insert because it will be discarded by deduplication in the main (source) table. The setting `deduplicate_blocks_in_dependent_materialized_views` allows for changing this behaviour. On retry, a materialized view will receive the repeat insert and will perform a deduplication check by itself,
|
||
ignoring check result for the source table, and will insert rows lost because of the first failure.
|
||
|
||
## insert_deduplication_token {#insert_deduplication_token}
|
||
|
||
The setting allows a user to provide own deduplication semantic in MergeTree/ReplicatedMergeTree
|
||
For example, by providing a unique value for the setting in each INSERT statement,
|
||
user can avoid the same inserted data being deduplicated.
|
||
|
||
Possible values:
|
||
|
||
- Any string
|
||
|
||
Default value: empty string (disabled)
|
||
|
||
`insert_deduplication_token` is used for deduplication _only_ when not empty.
|
||
|
||
For the replicated tables by default the only 100 of the most recent inserts for each partition are deduplicated (see [replicated_deduplication_window](merge-tree-settings.md/#replicated-deduplication-window), [replicated_deduplication_window_seconds](merge-tree-settings.md/#replicated-deduplication-window-seconds)).
|
||
For not replicated tables see [non_replicated_deduplication_window](merge-tree-settings.md/#non-replicated-deduplication-window).
|
||
|
||
Example:
|
||
|
||
```sql
|
||
CREATE TABLE test_table
|
||
( A Int64 )
|
||
ENGINE = MergeTree
|
||
ORDER BY A
|
||
SETTINGS non_replicated_deduplication_window = 100;
|
||
|
||
INSERT INTO test_table SETTINGS insert_deduplication_token = 'test' VALUES (1);
|
||
|
||
-- the next insert won't be deduplicated because insert_deduplication_token is different
|
||
INSERT INTO test_table SETTINGS insert_deduplication_token = 'test1' VALUES (1);
|
||
|
||
-- the next insert will be deduplicated because insert_deduplication_token
|
||
-- is the same as one of the previous
|
||
INSERT INTO test_table SETTINGS insert_deduplication_token = 'test' VALUES (2);
|
||
|
||
SELECT * FROM test_table
|
||
|
||
┌─A─┐
|
||
│ 1 │
|
||
└───┘
|
||
┌─A─┐
|
||
│ 1 │
|
||
└───┘
|
||
```
|
||
|
||
## insert_keeper_max_retries
|
||
|
||
The setting sets the maximum number of retries for ClickHouse Keeper (or ZooKeeper) requests during insert into replicated MergeTree. Only Keeper requests which failed due to network error, Keeper session timeout, or request timeout are considered for retries.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Retries are disabled
|
||
|
||
Default value: 0
|
||
|
||
Keeper request retries are done after some timeout. The timeout is controlled by the following settings: `insert_keeper_retry_initial_backoff_ms`, `insert_keeper_retry_max_backoff_ms`.
|
||
The first retry is done after `insert_keeper_retry_initial_backoff_ms` timeout. The consequent timeouts will be calculated as follows:
|
||
```
|
||
timeout = min(insert_keeper_retry_max_backoff_ms, latest_timeout * 2)
|
||
```
|
||
|
||
For example, if `insert_keeper_retry_initial_backoff_ms=100`, `insert_keeper_retry_max_backoff_ms=10000` and `insert_keeper_max_retries=8` then timeouts will be `100, 200, 400, 800, 1600, 3200, 6400, 10000`.
|
||
|
||
Apart from fault tolerance, the retries aim to provide a better user experience - they allow to avoid returning an error during INSERT execution if Keeper is restarted, for example, due to an upgrade.
|
||
|
||
## insert_keeper_retry_initial_backoff_ms {#insert_keeper_retry_initial_backoff_ms}
|
||
|
||
Initial timeout(in milliseconds) to retry a failed Keeper request during INSERT query execution
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — No timeout
|
||
|
||
Default value: 100
|
||
|
||
## insert_keeper_retry_max_backoff_ms {#insert_keeper_retry_max_backoff_ms}
|
||
|
||
Maximum timeout (in milliseconds) to retry a failed Keeper request during INSERT query execution
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Maximum timeout is not limited
|
||
|
||
Default value: 10000
|
||
|
||
## max_network_bytes {#settings-max-network-bytes}
|
||
|
||
Limits the data volume (in bytes) that is received or transmitted over the network when executing a query. This setting applies to every individual query.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Data volume control is disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## max_network_bandwidth {#settings-max-network-bandwidth}
|
||
|
||
Limits the speed of the data exchange over the network in bytes per second. This setting applies to every query.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Bandwidth control is disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## max_network_bandwidth_for_user {#settings-max-network-bandwidth-for-user}
|
||
|
||
Limits the speed of the data exchange over the network in bytes per second. This setting applies to all concurrently running queries performed by a single user.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Control of the data speed is disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## max_network_bandwidth_for_all_users {#settings-max-network-bandwidth-for-all-users}
|
||
|
||
Limits the speed that data is exchanged at over the network in bytes per second. This setting applies to all concurrently running queries on the server.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Control of the data speed is disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## count_distinct_implementation {#settings-count_distinct_implementation}
|
||
|
||
Specifies which of the `uniq*` functions should be used to perform the [COUNT(DISTINCT …)](../../sql-reference/aggregate-functions/reference/count.md/#agg_function-count) construction.
|
||
|
||
Possible values:
|
||
|
||
- [uniq](../../sql-reference/aggregate-functions/reference/uniq.md/#agg_function-uniq)
|
||
- [uniqCombined](../../sql-reference/aggregate-functions/reference/uniqcombined.md/#agg_function-uniqcombined)
|
||
- [uniqCombined64](../../sql-reference/aggregate-functions/reference/uniqcombined64.md/#agg_function-uniqcombined64)
|
||
- [uniqHLL12](../../sql-reference/aggregate-functions/reference/uniqhll12.md/#agg_function-uniqhll12)
|
||
- [uniqExact](../../sql-reference/aggregate-functions/reference/uniqexact.md/#agg_function-uniqexact)
|
||
|
||
Default value: `uniqExact`.
|
||
|
||
## skip_unavailable_shards {#settings-skip_unavailable_shards}
|
||
|
||
Enables or disables silently skipping of unavailable shards.
|
||
|
||
Shard is considered unavailable if all its replicas are unavailable. A replica is unavailable in the following cases:
|
||
|
||
- ClickHouse can’t connect to replica for any reason.
|
||
|
||
When connecting to a replica, ClickHouse performs several attempts. If all these attempts fail, the replica is considered unavailable.
|
||
|
||
- Replica can’t be resolved through DNS.
|
||
|
||
If replica’s hostname can’t be resolved through DNS, it can indicate the following situations:
|
||
|
||
- Replica’s host has no DNS record. It can occur in systems with dynamic DNS, for example, [Kubernetes](https://kubernetes.io), where nodes can be unresolvable during downtime, and this is not an error.
|
||
|
||
- Configuration error. ClickHouse configuration file contains a wrong hostname.
|
||
|
||
Possible values:
|
||
|
||
- 1 — skipping enabled.
|
||
|
||
If a shard is unavailable, ClickHouse returns a result based on partial data and does not report node availability issues.
|
||
|
||
- 0 — skipping disabled.
|
||
|
||
If a shard is unavailable, ClickHouse throws an exception.
|
||
|
||
Default value: 0.
|
||
|
||
## distributed_group_by_no_merge {#distributed-group-by-no-merge}
|
||
|
||
Do not merge aggregation states from different servers for distributed query processing, you can use this in case it is for certain that there are different keys on different shards
|
||
|
||
Possible values:
|
||
|
||
- `0` — Disabled (final query processing is done on the initiator node).
|
||
- `1` - Do not merge aggregation states from different servers for distributed query processing (query completelly processed on the shard, initiator only proxy the data), can be used in case it is for certain that there are different keys on different shards.
|
||
- `2` - Same as `1` but applies `ORDER BY` and `LIMIT` (it is not possible when the query processed completelly on the remote node, like for `distributed_group_by_no_merge=1`) on the initiator (can be used for queries with `ORDER BY` and/or `LIMIT`).
|
||
|
||
Default value: `0`
|
||
|
||
**Example**
|
||
|
||
```sql
|
||
SELECT *
|
||
FROM remote('127.0.0.{2,3}', system.one)
|
||
GROUP BY dummy
|
||
LIMIT 1
|
||
SETTINGS distributed_group_by_no_merge = 1
|
||
FORMAT PrettyCompactMonoBlock
|
||
|
||
┌─dummy─┐
|
||
│ 0 │
|
||
│ 0 │
|
||
└───────┘
|
||
```
|
||
|
||
```sql
|
||
SELECT *
|
||
FROM remote('127.0.0.{2,3}', system.one)
|
||
GROUP BY dummy
|
||
LIMIT 1
|
||
SETTINGS distributed_group_by_no_merge = 2
|
||
FORMAT PrettyCompactMonoBlock
|
||
|
||
┌─dummy─┐
|
||
│ 0 │
|
||
└───────┘
|
||
```
|
||
|
||
## distributed_push_down_limit {#distributed-push-down-limit}
|
||
|
||
Enables or disables [LIMIT](#limit) applying on each shard separatelly.
|
||
|
||
This will allow to avoid:
|
||
- Sending extra rows over network;
|
||
- Processing rows behind the limit on the initiator.
|
||
|
||
Starting from 21.9 version you cannot get inaccurate results anymore, since `distributed_push_down_limit` changes query execution only if at least one of the conditions met:
|
||
- [distributed_group_by_no_merge](#distributed-group-by-no-merge) > 0.
|
||
- Query **does not have** `GROUP BY`/`DISTINCT`/`LIMIT BY`, but it has `ORDER BY`/`LIMIT`.
|
||
- Query **has** `GROUP BY`/`DISTINCT`/`LIMIT BY` with `ORDER BY`/`LIMIT` and:
|
||
- [optimize_skip_unused_shards](#optimize-skip-unused-shards) is enabled.
|
||
- [optimize_distributed_group_by_sharding_key](#optimize-distributed-group-by-sharding-key) is enabled.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
See also:
|
||
|
||
- [distributed_group_by_no_merge](#distributed-group-by-no-merge)
|
||
- [optimize_skip_unused_shards](#optimize-skip-unused-shards)
|
||
- [optimize_distributed_group_by_sharding_key](#optimize-distributed-group-by-sharding-key)
|
||
|
||
## optimize_skip_unused_shards_limit {#optimize-skip-unused-shards-limit}
|
||
|
||
Limit for number of sharding key values, turns off `optimize_skip_unused_shards` if the limit is reached.
|
||
|
||
Too many values may require significant amount for processing, while the benefit is doubtful, since if you have huge number of values in `IN (...)`, then most likely the query will be sent to all shards anyway.
|
||
|
||
Default value: 1000
|
||
|
||
## optimize_skip_unused_shards {#optimize-skip-unused-shards}
|
||
|
||
Enables or disables skipping of unused shards for [SELECT](../../sql-reference/statements/select/index.md) queries that have sharding key condition in `WHERE/PREWHERE` (assuming that the data is distributed by sharding key, otherwise a query yields incorrect result).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0
|
||
|
||
## optimize_skip_unused_shards_rewrite_in {#optimize-skip-unused-shards-rewrite-in}
|
||
|
||
Rewrite IN in query for remote shards to exclude values that does not belong to the shard (requires optimize_skip_unused_shards).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1 (since it requires `optimize_skip_unused_shards` anyway, which `0` by default)
|
||
|
||
## allow_nondeterministic_optimize_skip_unused_shards {#allow-nondeterministic-optimize-skip-unused-shards}
|
||
|
||
Allow nondeterministic (like `rand` or `dictGet`, since later has some caveats with updates) functions in sharding key.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disallowed.
|
||
- 1 — Allowed.
|
||
|
||
Default value: 0
|
||
|
||
## optimize_skip_unused_shards_nesting {#optimize-skip-unused-shards-nesting}
|
||
|
||
Controls [`optimize_skip_unused_shards`](#optimize-skip-unused-shards) (hence still requires [`optimize_skip_unused_shards`](#optimize-skip-unused-shards)) depends on the nesting level of the distributed query (case when you have `Distributed` table that look into another `Distributed` table).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled, `optimize_skip_unused_shards` works always.
|
||
- 1 — Enables `optimize_skip_unused_shards` only for the first level.
|
||
- 2 — Enables `optimize_skip_unused_shards` up to the second level.
|
||
|
||
Default value: 0
|
||
|
||
## force_optimize_skip_unused_shards {#force-optimize-skip-unused-shards}
|
||
|
||
Enables or disables query execution if [optimize_skip_unused_shards](#optimize-skip-unused-shards) is enabled and skipping of unused shards is not possible. If the skipping is not possible and the setting is enabled, an exception will be thrown.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled. ClickHouse does not throw an exception.
|
||
- 1 — Enabled. Query execution is disabled only if the table has a sharding key.
|
||
- 2 — Enabled. Query execution is disabled regardless of whether a sharding key is defined for the table.
|
||
|
||
Default value: 0
|
||
|
||
## force_optimize_skip_unused_shards_nesting {#settings-force_optimize_skip_unused_shards_nesting}
|
||
|
||
Controls [`force_optimize_skip_unused_shards`](#force-optimize-skip-unused-shards) (hence still requires [`force_optimize_skip_unused_shards`](#force-optimize-skip-unused-shards)) depends on the nesting level of the distributed query (case when you have `Distributed` table that look into another `Distributed` table).
|
||
|
||
Possible values:
|
||
|
||
- 0 - Disabled, `force_optimize_skip_unused_shards` works always.
|
||
- 1 — Enables `force_optimize_skip_unused_shards` only for the first level.
|
||
- 2 — Enables `force_optimize_skip_unused_shards` up to the second level.
|
||
|
||
Default value: 0
|
||
|
||
## optimize_distributed_group_by_sharding_key {#optimize-distributed-group-by-sharding-key}
|
||
|
||
Optimize `GROUP BY sharding_key` queries, by avoiding costly aggregation on the initiator server (which will reduce memory usage for the query on the initiator server).
|
||
|
||
The following types of queries are supported (and all combinations of them):
|
||
|
||
- `SELECT DISTINCT [..., ]sharding_key[, ...] FROM dist`
|
||
- `SELECT ... FROM dist GROUP BY sharding_key[, ...]`
|
||
- `SELECT ... FROM dist GROUP BY sharding_key[, ...] ORDER BY x`
|
||
- `SELECT ... FROM dist GROUP BY sharding_key[, ...] LIMIT 1`
|
||
- `SELECT ... FROM dist GROUP BY sharding_key[, ...] LIMIT 1 BY x`
|
||
|
||
The following types of queries are not supported (support for some of them may be added later):
|
||
|
||
- `SELECT ... GROUP BY sharding_key[, ...] WITH TOTALS`
|
||
- `SELECT ... GROUP BY sharding_key[, ...] WITH ROLLUP`
|
||
- `SELECT ... GROUP BY sharding_key[, ...] WITH CUBE`
|
||
- `SELECT ... GROUP BY sharding_key[, ...] SETTINGS extremes=1`
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0
|
||
|
||
See also:
|
||
|
||
- [distributed_group_by_no_merge](#distributed-group-by-no-merge)
|
||
- [distributed_push_down_limit](#distributed-push-down-limit)
|
||
- [optimize_skip_unused_shards](#optimize-skip-unused-shards)
|
||
|
||
:::note
|
||
Right now it requires `optimize_skip_unused_shards` (the reason behind this is that one day it may be enabled by default, and it will work correctly only if data was inserted via Distributed table, i.e. data is distributed according to sharding_key).
|
||
:::
|
||
|
||
## optimize_throw_if_noop {#setting-optimize_throw_if_noop}
|
||
|
||
Enables or disables throwing an exception if an [OPTIMIZE](../../sql-reference/statements/optimize.md) query didn’t perform a merge.
|
||
|
||
By default, `OPTIMIZE` returns successfully even if it didn’t do anything. This setting lets you differentiate these situations and get the reason in an exception message.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Throwing an exception is enabled.
|
||
- 0 — Throwing an exception is disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## optimize_skip_merged_partitions {#optimize-skip-merged-partitions}
|
||
|
||
Enables or disables optimization for [OPTIMIZE TABLE ... FINAL](../../sql-reference/statements/optimize.md) query if there is only one part with level > 0 and it doesn't have expired TTL.
|
||
|
||
- `OPTIMIZE TABLE ... FINAL SETTINGS optimize_skip_merged_partitions=1`
|
||
|
||
By default, `OPTIMIZE TABLE ... FINAL` query rewrites the one part even if there is only a single part.
|
||
|
||
Possible values:
|
||
|
||
- 1 - Enable optimization.
|
||
- 0 - Disable optimization.
|
||
|
||
Default value: 0.
|
||
|
||
## optimize_functions_to_subcolumns {#optimize-functions-to-subcolumns}
|
||
|
||
Enables or disables optimization by transforming some functions to reading subcolumns. This reduces the amount of data to read.
|
||
|
||
These functions can be transformed:
|
||
|
||
- [length](../../sql-reference/functions/array-functions.md/#array_functions-length) to read the [size0](../../sql-reference/data-types/array.md/#array-size) subcolumn.
|
||
- [empty](../../sql-reference/functions/array-functions.md/#function-empty) to read the [size0](../../sql-reference/data-types/array.md/#array-size) subcolumn.
|
||
- [notEmpty](../../sql-reference/functions/array-functions.md/#function-notempty) to read the [size0](../../sql-reference/data-types/array.md/#array-size) subcolumn.
|
||
- [isNull](../../sql-reference/operators/index.md#operator-is-null) to read the [null](../../sql-reference/data-types/nullable.md/#finding-null) subcolumn.
|
||
- [isNotNull](../../sql-reference/operators/index.md#is-not-null) to read the [null](../../sql-reference/data-types/nullable.md/#finding-null) subcolumn.
|
||
- [count](../../sql-reference/aggregate-functions/reference/count.md) to read the [null](../../sql-reference/data-types/nullable.md/#finding-null) subcolumn.
|
||
- [mapKeys](../../sql-reference/functions/tuple-map-functions.md/#mapkeys) to read the [keys](../../sql-reference/data-types/map.md/#map-subcolumns) subcolumn.
|
||
- [mapValues](../../sql-reference/functions/tuple-map-functions.md/#mapvalues) to read the [values](../../sql-reference/data-types/map.md/#map-subcolumns) subcolumn.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Optimization disabled.
|
||
- 1 — Optimization enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
## optimize_trivial_count_query {#optimize-trivial-count-query}
|
||
|
||
Enables or disables the optimization to trivial query `SELECT count() FROM table` using metadata from MergeTree. If you need to use row-level security, disable this setting.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Optimization disabled.
|
||
- 1 — Optimization enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
See also:
|
||
|
||
- [optimize_functions_to_subcolumns](#optimize-functions-to-subcolumns)
|
||
|
||
## distributed_replica_error_half_life {#settings-distributed_replica_error_half_life}
|
||
|
||
- Type: seconds
|
||
- Default value: 60 seconds
|
||
|
||
Controls how fast errors in distributed tables are zeroed. If a replica is unavailable for some time, accumulates 5 errors, and distributed_replica_error_half_life is set to 1 second, then the replica is considered normal 3 seconds after the last error.
|
||
|
||
See also:
|
||
|
||
- [load_balancing](#load_balancing-round_robin)
|
||
- [Table engine Distributed](../../engines/table-engines/special/distributed.md)
|
||
- [distributed_replica_error_cap](#settings-distributed_replica_error_cap)
|
||
- [distributed_replica_max_ignored_errors](#settings-distributed_replica_max_ignored_errors)
|
||
|
||
## distributed_replica_error_cap {#settings-distributed_replica_error_cap}
|
||
|
||
- Type: unsigned int
|
||
- Default value: 1000
|
||
|
||
The error count of each replica is capped at this value, preventing a single replica from accumulating too many errors.
|
||
|
||
See also:
|
||
|
||
- [load_balancing](#load_balancing-round_robin)
|
||
- [Table engine Distributed](../../engines/table-engines/special/distributed.md)
|
||
- [distributed_replica_error_half_life](#settings-distributed_replica_error_half_life)
|
||
- [distributed_replica_max_ignored_errors](#settings-distributed_replica_max_ignored_errors)
|
||
|
||
## distributed_replica_max_ignored_errors {#settings-distributed_replica_max_ignored_errors}
|
||
|
||
- Type: unsigned int
|
||
- Default value: 0
|
||
|
||
The number of errors that will be ignored while choosing replicas (according to `load_balancing` algorithm).
|
||
|
||
See also:
|
||
|
||
- [load_balancing](#load_balancing-round_robin)
|
||
- [Table engine Distributed](../../engines/table-engines/special/distributed.md)
|
||
- [distributed_replica_error_cap](#settings-distributed_replica_error_cap)
|
||
- [distributed_replica_error_half_life](#settings-distributed_replica_error_half_life)
|
||
|
||
## distributed_directory_monitor_sleep_time_ms {#distributed_directory_monitor_sleep_time_ms}
|
||
|
||
Base interval for the [Distributed](../../engines/table-engines/special/distributed.md) table engine to send data. The actual interval grows exponentially in the event of errors.
|
||
|
||
Possible values:
|
||
|
||
- A positive integer number of milliseconds.
|
||
|
||
Default value: 100 milliseconds.
|
||
|
||
## distributed_directory_monitor_max_sleep_time_ms {#distributed_directory_monitor_max_sleep_time_ms}
|
||
|
||
Maximum interval for the [Distributed](../../engines/table-engines/special/distributed.md) table engine to send data. Limits exponential growth of the interval set in the [distributed_directory_monitor_sleep_time_ms](#distributed_directory_monitor_sleep_time_ms) setting.
|
||
|
||
Possible values:
|
||
|
||
- A positive integer number of milliseconds.
|
||
|
||
Default value: 30000 milliseconds (30 seconds).
|
||
|
||
## distributed_directory_monitor_batch_inserts {#distributed_directory_monitor_batch_inserts}
|
||
|
||
Enables/disables inserted data sending in batches.
|
||
|
||
When batch sending is enabled, the [Distributed](../../engines/table-engines/special/distributed.md) table engine tries to send multiple files of inserted data in one operation instead of sending them separately. Batch sending improves cluster performance by better-utilizing server and network resources.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Enabled.
|
||
- 0 — Disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## distributed_directory_monitor_split_batch_on_failure {#distributed_directory_monitor_split_batch_on_failure}
|
||
|
||
Enables/disables splitting batches on failures.
|
||
|
||
Sometimes sending particular batch to the remote shard may fail, because of some complex pipeline after (i.e. `MATERIALIZED VIEW` with `GROUP BY`) due to `Memory limit exceeded` or similar errors. In this case, retrying will not help (and this will stuck distributed sends for the table) but sending files from that batch one by one may succeed INSERT.
|
||
|
||
So installing this setting to `1` will disable batching for such batches (i.e. temporary disables `distributed_directory_monitor_batch_inserts` for failed batches).
|
||
|
||
Possible values:
|
||
|
||
- 1 — Enabled.
|
||
- 0 — Disabled.
|
||
|
||
Default value: 0.
|
||
|
||
:::note
|
||
This setting also affects broken batches (that may appears because of abnormal server (machine) termination and no `fsync_after_insert`/`fsync_directories` for [Distributed](../../engines/table-engines/special/distributed.md) table engine).
|
||
:::
|
||
|
||
:::warning
|
||
You should not rely on automatic batch splitting, since this may hurt performance.
|
||
:::
|
||
|
||
## os_thread_priority {#setting-os-thread-priority}
|
||
|
||
Sets the priority ([nice](https://en.wikipedia.org/wiki/Nice_(Unix))) for threads that execute queries. The OS scheduler considers this priority when choosing the next thread to run on each available CPU core.
|
||
|
||
:::warning
|
||
To use this setting, you need to set the `CAP_SYS_NICE` capability. The `clickhouse-server` package sets it up during installation. Some virtual environments do not allow you to set the `CAP_SYS_NICE` capability. In this case, `clickhouse-server` shows a message about it at the start.
|
||
:::
|
||
|
||
Possible values:
|
||
|
||
- You can set values in the range `[-20, 19]`.
|
||
|
||
Lower values mean higher priority. Threads with low `nice` priority values are executed more frequently than threads with high values. High values are preferable for long-running non-interactive queries because it allows them to quickly give up resources in favour of short interactive queries when they arrive.
|
||
|
||
Default value: 0.
|
||
|
||
## query_profiler_real_time_period_ns {#query_profiler_real_time_period_ns}
|
||
|
||
Sets the period for a real clock timer of the [query profiler](../../operations/optimizing-performance/sampling-query-profiler.md). Real clock timer counts wall-clock time.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer number, in nanoseconds.
|
||
|
||
Recommended values:
|
||
|
||
- 10000000 (100 times a second) nanoseconds and less for single queries.
|
||
- 1000000000 (once a second) for cluster-wide profiling.
|
||
|
||
- 0 for turning off the timer.
|
||
|
||
Type: [UInt64](../../sql-reference/data-types/int-uint.md).
|
||
|
||
Default value: 1000000000 nanoseconds (once a second).
|
||
|
||
See also:
|
||
|
||
- System table [trace_log](../../operations/system-tables/trace_log.md/#system_tables-trace_log)
|
||
|
||
## query_profiler_cpu_time_period_ns {#query_profiler_cpu_time_period_ns}
|
||
|
||
Sets the period for a CPU clock timer of the [query profiler](../../operations/optimizing-performance/sampling-query-profiler.md). This timer counts only CPU time.
|
||
|
||
Possible values:
|
||
|
||
- A positive integer number of nanoseconds.
|
||
|
||
Recommended values:
|
||
|
||
- 10000000 (100 times a second) nanoseconds and more for single queries.
|
||
- 1000000000 (once a second) for cluster-wide profiling.
|
||
|
||
- 0 for turning off the timer.
|
||
|
||
Type: [UInt64](../../sql-reference/data-types/int-uint.md).
|
||
|
||
Default value: 1000000000 nanoseconds.
|
||
|
||
See also:
|
||
|
||
- System table [trace_log](../../operations/system-tables/trace_log.md/#system_tables-trace_log)
|
||
|
||
## memory_profiler_step {#memory_profiler_step}
|
||
|
||
Sets the step of memory profiler. Whenever query memory usage becomes larger than every next step in number of bytes the memory profiler will collect the allocating stacktrace and will write it into [trace_log](../../operations/system-tables/trace_log.md#system_tables-trace_log).
|
||
|
||
Possible values:
|
||
|
||
- A positive integer number of bytes.
|
||
|
||
- 0 for turning off the memory profiler.
|
||
|
||
Default value: 4,194,304 bytes (4 MiB).
|
||
|
||
## memory_profiler_sample_probability {#memory_profiler_sample_probability}
|
||
|
||
Sets the probability of collecting stacktraces at random allocations and deallocations and writing them into [trace_log](../../operations/system-tables/trace_log.md#system_tables-trace_log).
|
||
|
||
Possible values:
|
||
|
||
- A positive floating-point number in the range [0..1].
|
||
|
||
- 0.0 for turning off the memory sampling.
|
||
|
||
Default value: 0.0.
|
||
|
||
## trace_profile_events {#trace_profile_events}
|
||
|
||
Enables or disables collecting stacktraces on each update of profile events along with the name of profile event and the value of increment and sending them into [trace_log](../../operations/system-tables/trace_log.md#system_tables-trace_log).
|
||
|
||
Possible values:
|
||
|
||
- 1 — Tracing of profile events enabled.
|
||
- 0 — Tracing of profile events disabled.
|
||
|
||
Default value: 0.
|
||
|
||
## allow_introspection_functions {#settings-allow_introspection_functions}
|
||
|
||
Enables or disables [introspections functions](../../sql-reference/functions/introspection.md) for query profiling.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Introspection functions enabled.
|
||
- 0 — Introspection functions disabled.
|
||
|
||
Default value: 0.
|
||
|
||
**See Also**
|
||
|
||
- [Sampling Query Profiler](../../operations/optimizing-performance/sampling-query-profiler.md)
|
||
- System table [trace_log](../../operations/system-tables/trace_log.md/#system_tables-trace_log)
|
||
|
||
## input_format_parallel_parsing {#input-format-parallel-parsing}
|
||
|
||
Enables or disables order-preserving parallel parsing of data formats. Supported only for [TSV](../../interfaces/formats.md/#tabseparated), [TKSV](../../interfaces/formats.md/#tskv), [CSV](../../interfaces/formats.md/#csv) and [JSONEachRow](../../interfaces/formats.md/#jsoneachrow) formats.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Enabled.
|
||
- 0 — Disabled.
|
||
|
||
Default value: `1`.
|
||
|
||
## output_format_parallel_formatting {#output-format-parallel-formatting}
|
||
|
||
Enables or disables parallel formatting of data formats. Supported only for [TSV](../../interfaces/formats.md/#tabseparated), [TKSV](../../interfaces/formats.md/#tskv), [CSV](../../interfaces/formats.md/#csv) and [JSONEachRow](../../interfaces/formats.md/#jsoneachrow) formats.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Enabled.
|
||
- 0 — Disabled.
|
||
|
||
Default value: `1`.
|
||
|
||
## min_chunk_bytes_for_parallel_parsing {#min-chunk-bytes-for-parallel-parsing}
|
||
|
||
- Type: unsigned int
|
||
- Default value: 1 MiB
|
||
|
||
The minimum chunk size in bytes, which each thread will parse in parallel.
|
||
|
||
## merge_selecting_sleep_ms {#merge_selecting_sleep_ms}
|
||
|
||
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in `background_schedule_pool` frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: `5000`.
|
||
|
||
## parallel_distributed_insert_select {#parallel_distributed_insert_select}
|
||
|
||
Enables parallel distributed `INSERT ... SELECT` query.
|
||
|
||
If we execute `INSERT INTO distributed_table_a SELECT ... FROM distributed_table_b` queries and both tables use the same cluster, and both tables are either [replicated](../../engines/table-engines/mergetree-family/replication.md) or non-replicated, then this query is processed locally on every shard.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — `SELECT` will be executed on each shard from the underlying table of the distributed engine.
|
||
- 2 — `SELECT` and `INSERT` will be executed on each shard from/to the underlying table of the distributed engine.
|
||
|
||
Default value: 0.
|
||
|
||
## insert_distributed_sync {#insert_distributed_sync}
|
||
|
||
Enables or disables synchronous data insertion into a [Distributed](../../engines/table-engines/special/distributed.md/#distributed) table.
|
||
|
||
By default, when inserting data into a `Distributed` table, the ClickHouse server sends data to cluster nodes in asynchronous mode. When `insert_distributed_sync=1`, the data is processed synchronously, and the `INSERT` operation succeeds only after all the data is saved on all shards (at least one replica for each shard if `internal_replication` is true).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Data is inserted in asynchronous mode.
|
||
- 1 — Data is inserted in synchronous mode.
|
||
|
||
Default value: `0`.
|
||
|
||
**See Also**
|
||
|
||
- [Distributed Table Engine](../../engines/table-engines/special/distributed.md/#distributed)
|
||
- [Managing Distributed Tables](../../sql-reference/statements/system.md/#query-language-system-distributed)
|
||
|
||
## insert_shard_id {#insert_shard_id}
|
||
|
||
If not `0`, specifies the shard of [Distributed](../../engines/table-engines/special/distributed.md/#distributed) table into which the data will be inserted synchronously.
|
||
|
||
If `insert_shard_id` value is incorrect, the server will throw an exception.
|
||
|
||
To get the number of shards on `requested_cluster`, you can check server config or use this query:
|
||
|
||
``` sql
|
||
SELECT uniq(shard_num) FROM system.clusters WHERE cluster = 'requested_cluster';
|
||
```
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- Any number from `1` to `shards_num` of corresponding [Distributed](../../engines/table-engines/special/distributed.md/#distributed) table.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
```sql
|
||
CREATE TABLE x AS system.numbers ENGINE = MergeTree ORDER BY number;
|
||
CREATE TABLE x_dist AS x ENGINE = Distributed('test_cluster_two_shards_localhost', currentDatabase(), x);
|
||
INSERT INTO x_dist SELECT * FROM numbers(5) SETTINGS insert_shard_id = 1;
|
||
SELECT * FROM x_dist ORDER BY number ASC;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌─number─┐
|
||
│ 0 │
|
||
│ 0 │
|
||
│ 1 │
|
||
│ 1 │
|
||
│ 2 │
|
||
│ 2 │
|
||
│ 3 │
|
||
│ 3 │
|
||
│ 4 │
|
||
│ 4 │
|
||
└────────┘
|
||
```
|
||
|
||
## use_compact_format_in_distributed_parts_names {#use_compact_format_in_distributed_parts_names}
|
||
|
||
Uses compact format for storing blocks for async (`insert_distributed_sync`) INSERT into tables with `Distributed` engine.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Uses `user[:password]@host:port#default_database` directory format.
|
||
- 1 — Uses `[shard{shard_index}[_replica{replica_index}]]` directory format.
|
||
|
||
Default value: `1`.
|
||
|
||
:::note
|
||
- with `use_compact_format_in_distributed_parts_names=0` changes from cluster definition will not be applied for async INSERT.
|
||
- with `use_compact_format_in_distributed_parts_names=1` changing the order of the nodes in the cluster definition, will change the `shard_index`/`replica_index` so be aware.
|
||
:::
|
||
|
||
## background_buffer_flush_schedule_pool_size {#background_buffer_flush_schedule_pool_size}
|
||
|
||
Sets the number of threads performing background flush in [Buffer](../../engines/table-engines/special/buffer.md)-engine tables. This setting is applied at the ClickHouse server start and can’t be changed in a user session.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 16.
|
||
|
||
## background_move_pool_size {#background_move_pool_size}
|
||
|
||
Sets the number of threads performing background moves of data parts for [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md/#table_engine-mergetree-multiple-volumes)-engine tables. This setting is applied at the ClickHouse server start and can’t be changed in a user session.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 8.
|
||
|
||
## background_schedule_pool_size {#background_schedule_pool_size}
|
||
|
||
Sets the number of threads performing background tasks for [replicated](../../engines/table-engines/mergetree-family/replication.md) tables, [Kafka](../../engines/table-engines/integrations/kafka.md) streaming, [DNS cache updates](../../operations/server-configuration-parameters/settings.md/#server-settings-dns-cache-update-period). This setting is applied at ClickHouse server start and can’t be changed in a user session.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 128.
|
||
|
||
## background_fetches_pool_size {#background_fetches_pool_size}
|
||
|
||
Sets the number of threads performing background fetches for [replicated](../../engines/table-engines/mergetree-family/replication.md) tables. This setting is applied at the ClickHouse server start and can’t be changed in a user session. For production usage with frequent small insertions or slow ZooKeeper cluster it is recommended to use default value.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 8.
|
||
|
||
## always_fetch_merged_part {#always_fetch_merged_part}
|
||
|
||
Prohibits data parts merging in [Replicated\*MergeTree](../../engines/table-engines/mergetree-family/replication.md)-engine tables.
|
||
|
||
When merging is prohibited, the replica never merges parts and always downloads merged parts from other replicas. If there is no required data yet, the replica waits for it. CPU and disk load on the replica server decreases, but the network load on the cluster increases. This setting can be useful on servers with relatively weak CPUs or slow disks, such as servers for backups storage.
|
||
|
||
Possible values:
|
||
|
||
- 0 — `Replicated*MergeTree`-engine tables merge data parts at the replica.
|
||
- 1 — `Replicated*MergeTree`-engine tables do not merge data parts at the replica. The tables download merged data parts from other replicas.
|
||
|
||
Default value: 0.
|
||
|
||
**See Also**
|
||
|
||
- [Data Replication](../../engines/table-engines/mergetree-family/replication.md)
|
||
|
||
## background_distributed_schedule_pool_size {#background_distributed_schedule_pool_size}
|
||
|
||
Sets the number of threads performing background tasks for [distributed](../../engines/table-engines/special/distributed.md) sends. This setting is applied at the ClickHouse server start and can’t be changed in a user session.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 16.
|
||
|
||
## background_message_broker_schedule_pool_size {#background_message_broker_schedule_pool_size}
|
||
|
||
Sets the number of threads performing background tasks for message streaming. This setting is applied at the ClickHouse server start and can’t be changed in a user session.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 16.
|
||
|
||
**See Also**
|
||
|
||
- [Kafka](../../engines/table-engines/integrations/kafka.md/#kafka) engine.
|
||
- [RabbitMQ](../../engines/table-engines/integrations/rabbitmq.md/#rabbitmq-engine) engine.
|
||
|
||
## validate_polygons {#validate_polygons}
|
||
|
||
Enables or disables throwing an exception in the [pointInPolygon](../../sql-reference/functions/geo/index.md#pointinpolygon) function, if the polygon is self-intersecting or self-tangent.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Throwing an exception is disabled. `pointInPolygon` accepts invalid polygons and returns possibly incorrect results for them.
|
||
- 1 — Throwing an exception is enabled.
|
||
|
||
Default value: 1.
|
||
|
||
## transform_null_in {#transform_null_in}
|
||
|
||
Enables equality of [NULL](../../sql-reference/syntax.md/#null-literal) values for [IN](../../sql-reference/operators/in.md) operator.
|
||
|
||
By default, `NULL` values can’t be compared because `NULL` means undefined value. Thus, comparison `expr = NULL` must always return `false`. With this setting `NULL = NULL` returns `true` for `IN` operator.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Comparison of `NULL` values in `IN` operator returns `false`.
|
||
- 1 — Comparison of `NULL` values in `IN` operator returns `true`.
|
||
|
||
Default value: 0.
|
||
|
||
**Example**
|
||
|
||
Consider the `null_in` table:
|
||
|
||
``` text
|
||
┌──idx─┬─────i─┐
|
||
│ 1 │ 1 │
|
||
│ 2 │ NULL │
|
||
│ 3 │ 3 │
|
||
└──────┴───────┘
|
||
```
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
SELECT idx, i FROM null_in WHERE i IN (1, NULL) SETTINGS transform_null_in = 0;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌──idx─┬────i─┐
|
||
│ 1 │ 1 │
|
||
└──────┴──────┘
|
||
```
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
SELECT idx, i FROM null_in WHERE i IN (1, NULL) SETTINGS transform_null_in = 1;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌──idx─┬─────i─┐
|
||
│ 1 │ 1 │
|
||
│ 2 │ NULL │
|
||
└──────┴───────┘
|
||
```
|
||
|
||
**See Also**
|
||
|
||
- [NULL Processing in IN Operators](../../sql-reference/operators/in.md/#in-null-processing)
|
||
|
||
## low_cardinality_max_dictionary_size {#low_cardinality_max_dictionary_size}
|
||
|
||
Sets a maximum size in rows of a shared global dictionary for the [LowCardinality](../../sql-reference/data-types/lowcardinality.md) data type that can be written to a storage file system. This setting prevents issues with RAM in case of unlimited dictionary growth. All the data that can’t be encoded due to maximum dictionary size limitation ClickHouse writes in an ordinary method.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
|
||
Default value: 8192.
|
||
|
||
## low_cardinality_use_single_dictionary_for_part {#low_cardinality_use_single_dictionary_for_part}
|
||
|
||
Turns on or turns off using of single dictionary for the data part.
|
||
|
||
By default, the ClickHouse server monitors the size of dictionaries and if a dictionary overflows then the server starts to write the next one. To prohibit creating several dictionaries set `low_cardinality_use_single_dictionary_for_part = 1`.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Creating several dictionaries for the data part is prohibited.
|
||
- 0 — Creating several dictionaries for the data part is not prohibited.
|
||
|
||
Default value: 0.
|
||
|
||
## low_cardinality_allow_in_native_format {#low_cardinality_allow_in_native_format}
|
||
|
||
Allows or restricts using the [LowCardinality](../../sql-reference/data-types/lowcardinality.md) data type with the [Native](../../interfaces/formats.md/#native) format.
|
||
|
||
If usage of `LowCardinality` is restricted, ClickHouse server converts `LowCardinality`-columns to ordinary ones for `SELECT` queries, and convert ordinary columns to `LowCardinality`-columns for `INSERT` queries.
|
||
|
||
This setting is required mainly for third-party clients which do not support `LowCardinality` data type.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Usage of `LowCardinality` is not restricted.
|
||
- 0 — Usage of `LowCardinality` is restricted.
|
||
|
||
Default value: 1.
|
||
|
||
## allow_suspicious_low_cardinality_types {#allow_suspicious_low_cardinality_types}
|
||
|
||
Allows or restricts using [LowCardinality](../../sql-reference/data-types/lowcardinality.md) with data types with fixed size of 8 bytes or less: numeric data types and `FixedString(8_bytes_or_less)`.
|
||
|
||
For small fixed values using of `LowCardinality` is usually inefficient, because ClickHouse stores a numeric index for each row. As a result:
|
||
|
||
- Disk space usage can rise.
|
||
- RAM consumption can be higher, depending on a dictionary size.
|
||
- Some functions can work slower due to extra coding/encoding operations.
|
||
|
||
Merge times in [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md)-engine tables can grow due to all the reasons described above.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Usage of `LowCardinality` is not restricted.
|
||
- 0 — Usage of `LowCardinality` is restricted.
|
||
|
||
Default value: 0.
|
||
|
||
## min_insert_block_size_rows_for_materialized_views {#min-insert-block-size-rows-for-materialized-views}
|
||
|
||
Sets the minimum number of rows in the block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones. This setting is applied only for blocks inserted into [materialized view](../../sql-reference/statements/create/view.md). By adjusting this setting, you control blocks squashing while pushing to materialized view and avoid excessive memory usage.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
- 0 — Squashing disabled.
|
||
|
||
Default value: 1048576.
|
||
|
||
**See Also**
|
||
|
||
- [min_insert_block_size_rows](#min-insert-block-size-rows)
|
||
|
||
## min_insert_block_size_bytes_for_materialized_views {#min-insert-block-size-bytes-for-materialized-views}
|
||
|
||
Sets the minimum number of bytes in the block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones. This setting is applied only for blocks inserted into [materialized view](../../sql-reference/statements/create/view.md). By adjusting this setting, you control blocks squashing while pushing to materialized view and avoid excessive memory usage.
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
- 0 — Squashing disabled.
|
||
|
||
Default value: 268435456.
|
||
|
||
**See also**
|
||
|
||
- [min_insert_block_size_bytes](#min-insert-block-size-bytes)
|
||
|
||
## optimize_read_in_order {#optimize_read_in_order}
|
||
|
||
Enables [ORDER BY](../../sql-reference/statements/select/order-by.md/#optimize_read_in_order) optimization in [SELECT](../../sql-reference/statements/select/index.md) queries for reading data from [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) tables.
|
||
|
||
Possible values:
|
||
|
||
- 0 — `ORDER BY` optimization is disabled.
|
||
- 1 — `ORDER BY` optimization is enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
**See Also**
|
||
|
||
- [ORDER BY Clause](../../sql-reference/statements/select/order-by.md/#optimize_read_in_order)
|
||
|
||
## optimize_aggregation_in_order {#optimize_aggregation_in_order}
|
||
|
||
Enables [GROUP BY](../../sql-reference/statements/select/group-by.md) optimization in [SELECT](../../sql-reference/statements/select/index.md) queries for aggregating data in corresponding order in [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) tables.
|
||
|
||
Possible values:
|
||
|
||
- 0 — `GROUP BY` optimization is disabled.
|
||
- 1 — `GROUP BY` optimization is enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
**See Also**
|
||
|
||
- [GROUP BY optimization](../../sql-reference/statements/select/group-by.md/#aggregation-in-order)
|
||
|
||
## mutations_sync {#mutations_sync}
|
||
|
||
Allows to execute `ALTER TABLE ... UPDATE|DELETE` queries ([mutations](../../sql-reference/statements/alter/index.md#mutations)) synchronously.
|
||
|
||
Possible values:
|
||
|
||
- 0 - Mutations execute asynchronously.
|
||
- 1 - The query waits for all mutations to complete on the current server.
|
||
- 2 - The query waits for all mutations to complete on all replicas (if they exist).
|
||
|
||
Default value: `0`.
|
||
|
||
**See Also**
|
||
|
||
- [Synchronicity of ALTER Queries](../../sql-reference/statements/alter/index.md#synchronicity-of-alter-queries)
|
||
- [Mutations](../../sql-reference/statements/alter/index.md#mutations)
|
||
|
||
## ttl_only_drop_parts {#ttl_only_drop_parts}
|
||
|
||
Enables or disables complete dropping of data parts where all rows are expired in [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) tables.
|
||
|
||
When `ttl_only_drop_parts` is disabled (by default), the ClickHouse server only deletes expired rows according to their TTL.
|
||
|
||
When `ttl_only_drop_parts` is enabled, the ClickHouse server drops a whole part when all rows in it are expired.
|
||
|
||
Dropping whole parts instead of partial cleaning TTL-d rows allows having shorter `merge_with_ttl_timeout` times and lower impact on system performance.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The complete dropping of data parts is disabled.
|
||
- 1 — The complete dropping of data parts is enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
**See Also**
|
||
|
||
- [CREATE TABLE query clauses and settings](../../engines/table-engines/mergetree-family/mergetree.md/#mergetree-query-clauses) (`merge_with_ttl_timeout` setting)
|
||
- [Table TTL](../../engines/table-engines/mergetree-family/mergetree.md/#mergetree-table-ttl)
|
||
|
||
## lock_acquire_timeout {#lock_acquire_timeout}
|
||
|
||
Defines how many seconds a locking request waits before failing.
|
||
|
||
Locking timeout is used to protect from deadlocks while executing read/write operations with tables. When the timeout expires and the locking request fails, the ClickHouse server throws an exception "Locking attempt timed out! Possible deadlock avoided. Client should retry." with error code `DEADLOCK_AVOIDED`.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer (in seconds).
|
||
- 0 — No locking timeout.
|
||
|
||
Default value: `120` seconds.
|
||
|
||
## cast_keep_nullable {#cast_keep_nullable}
|
||
|
||
Enables or disables keeping of the `Nullable` data type in [CAST](../../sql-reference/functions/type-conversion-functions.md/#type_conversion_function-cast) operations.
|
||
|
||
When the setting is enabled and the argument of `CAST` function is `Nullable`, the result is also transformed to `Nullable` type. When the setting is disabled, the result always has the destination type exactly.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The `CAST` result has exactly the destination type specified.
|
||
- 1 — If the argument type is `Nullable`, the `CAST` result is transformed to `Nullable(DestinationDataType)`.
|
||
|
||
Default value: `0`.
|
||
|
||
**Examples**
|
||
|
||
The following query results in the destination data type exactly:
|
||
|
||
```sql
|
||
SET cast_keep_nullable = 0;
|
||
SELECT CAST(toNullable(toInt32(0)) AS Int32) as x, toTypeName(x);
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
┌─x─┬─toTypeName(CAST(toNullable(toInt32(0)), 'Int32'))─┐
|
||
│ 0 │ Int32 │
|
||
└───┴───────────────────────────────────────────────────┘
|
||
```
|
||
|
||
The following query results in the `Nullable` modification on the destination data type:
|
||
|
||
```sql
|
||
SET cast_keep_nullable = 1;
|
||
SELECT CAST(toNullable(toInt32(0)) AS Int32) as x, toTypeName(x);
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
┌─x─┬─toTypeName(CAST(toNullable(toInt32(0)), 'Int32'))─┐
|
||
│ 0 │ Nullable(Int32) │
|
||
└───┴───────────────────────────────────────────────────┘
|
||
```
|
||
|
||
**See Also**
|
||
|
||
- [CAST](../../sql-reference/functions/type-conversion-functions.md/#type_conversion_function-cast) function
|
||
|
||
## system_events_show_zero_values {#system_events_show_zero_values}
|
||
|
||
Allows to select zero-valued events from [`system.events`](../../operations/system-tables/events.md).
|
||
|
||
Some monitoring systems require passing all the metrics values to them for each checkpoint, even if the metric value is zero.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
**Examples**
|
||
|
||
Query
|
||
|
||
```sql
|
||
SELECT * FROM system.events WHERE event='QueryMemoryLimitExceeded';
|
||
```
|
||
|
||
Result
|
||
|
||
```text
|
||
Ok.
|
||
```
|
||
|
||
Query
|
||
```sql
|
||
SET system_events_show_zero_values = 1;
|
||
SELECT * FROM system.events WHERE event='QueryMemoryLimitExceeded';
|
||
```
|
||
|
||
Result
|
||
|
||
```text
|
||
┌─event────────────────────┬─value─┬─description───────────────────────────────────────────┐
|
||
│ QueryMemoryLimitExceeded │ 0 │ Number of times when memory limit exceeded for query. │
|
||
└──────────────────────────┴───────┴───────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
## allow_nullable_key {#allow-nullable-key}
|
||
|
||
Allows using of the [Nullable](../../sql-reference/data-types/nullable.md/#data_type-nullable)-typed values in a sorting and a primary key for [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md/#table_engines-mergetree) tables.
|
||
|
||
Possible values:
|
||
|
||
- 1 — `Nullable`-type expressions are allowed in keys.
|
||
- 0 — `Nullable`-type expressions are not allowed in keys.
|
||
|
||
Default value: `0`.
|
||
|
||
:::warning
|
||
Nullable primary key usually indicates bad design. It is forbidden in almost all main stream DBMS. The feature is mainly for [AggregatingMergeTree](../../engines/table-engines/mergetree-family/aggregatingmergetree.md) and is not heavily tested. Use with care.
|
||
:::
|
||
|
||
:::warning
|
||
Do not enable this feature in version `<= 21.8`. It's not properly implemented and may lead to server crash.
|
||
:::
|
||
|
||
## aggregate_functions_null_for_empty {#aggregate_functions_null_for_empty}
|
||
|
||
Enables or disables rewriting all aggregate functions in a query, adding [-OrNull](../../sql-reference/aggregate-functions/combinators.md/#agg-functions-combinator-ornull) suffix to them. Enable it for SQL standard compatibility.
|
||
It is implemented via query rewrite (similar to [count_distinct_implementation](#settings-count_distinct_implementation) setting) to get consistent results for distributed queries.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 0.
|
||
|
||
**Example**
|
||
|
||
Consider the following query with aggregate functions:
|
||
```sql
|
||
SELECT SUM(-1), MAX(0) FROM system.one WHERE 0;
|
||
```
|
||
|
||
With `aggregate_functions_null_for_empty = 0` it would produce:
|
||
```text
|
||
┌─SUM(-1)─┬─MAX(0)─┐
|
||
│ 0 │ 0 │
|
||
└─────────┴────────┘
|
||
```
|
||
|
||
With `aggregate_functions_null_for_empty = 1` the result would be:
|
||
```text
|
||
┌─SUMOrNull(-1)─┬─MAXOrNull(0)─┐
|
||
│ NULL │ NULL │
|
||
└───────────────┴──────────────┘
|
||
```
|
||
|
||
## union_default_mode {#union-default-mode}
|
||
|
||
Sets a mode for combining `SELECT` query results. The setting is only used when shared with [UNION](../../sql-reference/statements/select/union.md) without explicitly specifying the `UNION ALL` or `UNION DISTINCT`.
|
||
|
||
Possible values:
|
||
|
||
- `'DISTINCT'` — ClickHouse outputs rows as a result of combining queries removing duplicate rows.
|
||
- `'ALL'` — ClickHouse outputs all rows as a result of combining queries including duplicate rows.
|
||
- `''` — ClickHouse generates an exception when used with `UNION`.
|
||
|
||
Default value: `''`.
|
||
|
||
See examples in [UNION](../../sql-reference/statements/select/union.md).
|
||
|
||
## default_table_engine {#default_table_engine}
|
||
|
||
Default table engine to use when `ENGINE` is not set in a `CREATE` statement.
|
||
|
||
Possible values:
|
||
|
||
- a string representing any valid table engine name
|
||
|
||
Default value: `None`
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SET default_table_engine = 'Log';
|
||
|
||
SELECT name, value, changed FROM system.settings WHERE name = 'default_table_engine';
|
||
```
|
||
|
||
Result:
|
||
|
||
```response
|
||
┌─name─────────────────┬─value─┬─changed─┐
|
||
│ default_table_engine │ Log │ 1 │
|
||
└──────────────────────┴───────┴─────────┘
|
||
```
|
||
|
||
In this example, any new table that does not specify an `Engine` will use the `Log` table engine:
|
||
|
||
Query:
|
||
|
||
```sql
|
||
CREATE TABLE my_table (
|
||
x UInt32,
|
||
y UInt32
|
||
);
|
||
|
||
SHOW CREATE TABLE my_table;
|
||
```
|
||
|
||
Result:
|
||
|
||
```response
|
||
┌─statement────────────────────────────────────────────────────────────────┐
|
||
│ CREATE TABLE default.my_table
|
||
(
|
||
`x` UInt32,
|
||
`y` UInt32
|
||
)
|
||
ENGINE = Log
|
||
└──────────────────────────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
## data_type_default_nullable {#data_type_default_nullable}
|
||
|
||
Allows data types without explicit modifiers [NULL or NOT NULL](../../sql-reference/statements/create/table.md/#null-modifiers) in column definition will be [Nullable](../../sql-reference/data-types/nullable.md/#data_type-nullable).
|
||
|
||
Possible values:
|
||
|
||
- 1 — The data types in column definitions are set to `Nullable` by default.
|
||
- 0 — The data types in column definitions are set to not `Nullable` by default.
|
||
|
||
Default value: `0`.
|
||
|
||
## execute_merges_on_single_replica_time_threshold {#execute-merges-on-single-replica-time-threshold}
|
||
|
||
Enables special logic to perform merges on replicas.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer (in seconds).
|
||
- 0 — Special merges logic is not used. Merges happen in the usual way on all the replicas.
|
||
|
||
Default value: `0`.
|
||
|
||
**Usage**
|
||
|
||
Selects one replica to perform the merge on. Sets the time threshold from the start of the merge. Other replicas wait for the merge to finish, then download the result. If the time threshold passes and the selected replica does not perform the merge, then the merge is performed on other replicas as usual.
|
||
|
||
High values for that threshold may lead to replication delays.
|
||
|
||
It can be useful when merges are CPU bounded not IO bounded (performing heavy data compression, calculating aggregate functions or default expressions that require a large amount of calculations, or just very high number of tiny merges).
|
||
|
||
## max_final_threads {#max-final-threads}
|
||
|
||
Sets the maximum number of parallel threads for the `SELECT` query data read phase with the [FINAL](../../sql-reference/statements/select/from.md/#select-from-final) modifier.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 or 1 — Disabled. `SELECT` queries are executed in a single thread.
|
||
|
||
Default value: `16`.
|
||
|
||
## opentelemetry_start_trace_probability {#opentelemetry-start-trace-probability}
|
||
|
||
Sets the probability that the ClickHouse can start a trace for executed queries (if no parent [trace context](https://www.w3.org/TR/trace-context/) is supplied).
|
||
|
||
Possible values:
|
||
|
||
- 0 — The trace for all executed queries is disabled (if no parent trace context is supplied).
|
||
- Positive floating-point number in the range [0..1]. For example, if the setting value is `0,5`, ClickHouse can start a trace on average for half of the queries.
|
||
- 1 — The trace for all executed queries is enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
## optimize_on_insert {#optimize-on-insert}
|
||
|
||
Enables or disables data transformation before the insertion, as if merge was done on this block (according to table engine).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: 1.
|
||
|
||
**Example**
|
||
|
||
The difference between enabled and disabled:
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SET optimize_on_insert = 1;
|
||
|
||
CREATE TABLE test1 (`FirstTable` UInt32) ENGINE = ReplacingMergeTree ORDER BY FirstTable;
|
||
|
||
INSERT INTO test1 SELECT number % 2 FROM numbers(5);
|
||
|
||
SELECT * FROM test1;
|
||
|
||
SET optimize_on_insert = 0;
|
||
|
||
CREATE TABLE test2 (`SecondTable` UInt32) ENGINE = ReplacingMergeTree ORDER BY SecondTable;
|
||
|
||
INSERT INTO test2 SELECT number % 2 FROM numbers(5);
|
||
|
||
SELECT * FROM test2;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌─FirstTable─┐
|
||
│ 0 │
|
||
│ 1 │
|
||
└────────────┘
|
||
|
||
┌─SecondTable─┐
|
||
│ 0 │
|
||
│ 0 │
|
||
│ 0 │
|
||
│ 1 │
|
||
│ 1 │
|
||
└─────────────┘
|
||
```
|
||
|
||
Note that this setting influences [Materialized view](../../sql-reference/statements/create/view.md/#materialized) and [MaterializedMySQL](../../engines/database-engines/materialized-mysql.md) behaviour.
|
||
|
||
## engine_file_empty_if_not_exists {#engine-file-empty_if-not-exists}
|
||
|
||
Allows to select data from a file engine table without file.
|
||
|
||
Possible values:
|
||
- 0 — `SELECT` throws exception.
|
||
- 1 — `SELECT` returns empty result.
|
||
|
||
Default value: `0`.
|
||
|
||
## engine_file_truncate_on_insert {#engine-file-truncate-on-insert}
|
||
|
||
Enables or disables truncate before insert in [File](../../engines/table-engines/special/file.md) engine tables.
|
||
|
||
Possible values:
|
||
- 0 — `INSERT` query appends new data to the end of the file.
|
||
- 1 — `INSERT` query replaces existing content of the file with the new data.
|
||
|
||
Default value: `0`.
|
||
|
||
## s3_truncate_on_insert
|
||
|
||
Enables or disables truncate before inserts in s3 engine tables. If disabled, an exception will be thrown on insert attempts if an S3 object already exists.
|
||
|
||
Possible values:
|
||
- 0 — `INSERT` query appends new data to the end of the file.
|
||
- 1 — `INSERT` query replaces existing content of the file with the new data.
|
||
|
||
Default value: `0`.
|
||
|
||
## hdfs_truncate_on_insert
|
||
|
||
Enables or disables truncation before an insert in hdfs engine tables. If disabled, an exception will be thrown on an attempt to insert if a file in HDFS already exists.
|
||
|
||
Possible values:
|
||
- 0 — `INSERT` query appends new data to the end of the file.
|
||
- 1 — `INSERT` query replaces existing content of the file with the new data.
|
||
|
||
Default value: `0`.
|
||
|
||
## engine_file_allow_create_multiple_files
|
||
|
||
Enables or disables creating a new file on each insert in file engine tables if the format has the suffix (`JSON`, `ORC`, `Parquet`, etc.). If enabled, on each insert a new file will be created with a name following this pattern:
|
||
|
||
`data.Parquet` -> `data.1.Parquet` -> `data.2.Parquet`, etc.
|
||
|
||
Possible values:
|
||
- 0 — `INSERT` query appends new data to the end of the file.
|
||
- 1 — `INSERT` query replaces existing content of the file with the new data.
|
||
|
||
Default value: `0`.
|
||
|
||
## s3_create_new_file_on_insert
|
||
|
||
Enables or disables creating a new file on each insert in s3 engine tables. If enabled, on each insert a new S3 object will be created with the key, similar to this pattern:
|
||
|
||
initial: `data.Parquet.gz` -> `data.1.Parquet.gz` -> `data.2.Parquet.gz`, etc.
|
||
|
||
Possible values:
|
||
- 0 — `INSERT` query appends new data to the end of the file.
|
||
- 1 — `INSERT` query replaces existing content of the file with the new data.
|
||
|
||
Default value: `0`.
|
||
|
||
## hdfs_create_new_file_on_insert
|
||
|
||
Enables or disables creating a new file on each insert in HDFS engine tables. If enabled, on each insert a new HDFS file will be created with the name, similar to this pattern:
|
||
|
||
initial: `data.Parquet.gz` -> `data.1.Parquet.gz` -> `data.2.Parquet.gz`, etc.
|
||
|
||
Possible values:
|
||
- 0 — `INSERT` query appends new data to the end of the file.
|
||
- 1 — `INSERT` query replaces existing content of the file with the new data.
|
||
|
||
Default value: `0`.
|
||
|
||
## allow_experimental_geo_types {#allow-experimental-geo-types}
|
||
|
||
Allows working with experimental [geo data types](../../sql-reference/data-types/geo.md).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Working with geo data types is disabled.
|
||
- 1 — Working with geo data types is enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
## database_atomic_wait_for_drop_and_detach_synchronously {#database_atomic_wait_for_drop_and_detach_synchronously}
|
||
|
||
Adds a modifier `SYNC` to all `DROP` and `DETACH` queries.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Queries will be executed with delay.
|
||
- 1 — Queries will be executed without delay.
|
||
|
||
Default value: `0`.
|
||
|
||
## show_table_uuid_in_table_create_query_if_not_nil {#show_table_uuid_in_table_create_query_if_not_nil}
|
||
|
||
Sets the `SHOW TABLE` query display.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The query will be displayed without table UUID.
|
||
- 1 — The query will be displayed with table UUID.
|
||
|
||
Default value: `0`.
|
||
|
||
## allow_experimental_live_view {#allow-experimental-live-view}
|
||
|
||
Allows creation of experimental [live views](../../sql-reference/statements/create/view.md/#live-view).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Working with live views is disabled.
|
||
- 1 — Working with live views is enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
## live_view_heartbeat_interval {#live-view-heartbeat-interval}
|
||
|
||
Sets the heartbeat interval in seconds to indicate [live view](../../sql-reference/statements/create/view.md/#live-view) is alive .
|
||
|
||
Default value: `15`.
|
||
|
||
## max_live_view_insert_blocks_before_refresh {#max-live-view-insert-blocks-before-refresh}
|
||
|
||
Sets the maximum number of inserted blocks after which mergeable blocks are dropped and query for [live view](../../sql-reference/statements/create/view.md/#live-view) is re-executed.
|
||
|
||
Default value: `64`.
|
||
|
||
## periodic_live_view_refresh {#periodic-live-view-refresh}
|
||
|
||
Sets the interval in seconds after which periodically refreshed [live view](../../sql-reference/statements/create/view.md/#live-view) is forced to refresh.
|
||
|
||
Default value: `60`.
|
||
|
||
## http_connection_timeout {#http_connection_timeout}
|
||
|
||
HTTP connection timeout (in seconds).
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
- 0 - Disabled (infinite timeout).
|
||
|
||
Default value: 1.
|
||
|
||
## http_send_timeout {#http_send_timeout}
|
||
|
||
HTTP send timeout (in seconds).
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
- 0 - Disabled (infinite timeout).
|
||
|
||
Default value: 180.
|
||
|
||
## http_receive_timeout {#http_receive_timeout}
|
||
|
||
HTTP receive timeout (in seconds).
|
||
|
||
Possible values:
|
||
|
||
- Any positive integer.
|
||
- 0 - Disabled (infinite timeout).
|
||
|
||
Default value: 180.
|
||
|
||
## check_query_single_value_result {#check_query_single_value_result}
|
||
|
||
Defines the level of detail for the [CHECK TABLE](../../sql-reference/statements/check-table.md/#checking-mergetree-tables) query result for `MergeTree` family engines .
|
||
|
||
Possible values:
|
||
|
||
- 0 — the query shows a check status for every individual data part of a table.
|
||
- 1 — the query shows the general table check status.
|
||
|
||
Default value: `0`.
|
||
|
||
## prefer_column_name_to_alias {#prefer-column-name-to-alias}
|
||
|
||
Enables or disables using the original column names instead of aliases in query expressions and clauses. It especially matters when alias is the same as the column name, see [Expression Aliases](../../sql-reference/syntax.md/#notes-on-usage). Enable this setting to make aliases syntax rules in ClickHouse more compatible with most other database engines.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The column name is substituted with the alias.
|
||
- 1 — The column name is not substituted with the alias.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
The difference between enabled and disabled:
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SET prefer_column_name_to_alias = 0;
|
||
SELECT avg(number) AS number, max(number) FROM numbers(10);
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
Received exception from server (version 21.5.1):
|
||
Code: 184. DB::Exception: Received from localhost:9000. DB::Exception: Aggregate function avg(number) is found inside another aggregate function in query: While processing avg(number) AS number.
|
||
```
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SET prefer_column_name_to_alias = 1;
|
||
SELECT avg(number) AS number, max(number) FROM numbers(10);
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
┌─number─┬─max(number)─┐
|
||
│ 4.5 │ 9 │
|
||
└────────┴─────────────┘
|
||
```
|
||
|
||
## limit {#limit}
|
||
|
||
Sets the maximum number of rows to get from the query result. It adjusts the value set by the [LIMIT](../../sql-reference/statements/select/limit.md/#limit-clause) clause, so that the limit, specified in the query, cannot exceed the limit, set by this setting.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The number of rows is not limited.
|
||
- Positive integer.
|
||
|
||
Default value: `0`.
|
||
|
||
## offset {#offset}
|
||
|
||
Sets the number of rows to skip before starting to return rows from the query. It adjusts the offset set by the [OFFSET](../../sql-reference/statements/select/offset.md/#offset-fetch) clause, so that these two values are summarized.
|
||
|
||
Possible values:
|
||
|
||
- 0 — No rows are skipped .
|
||
- Positive integer.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
Input table:
|
||
|
||
``` sql
|
||
CREATE TABLE test (i UInt64) ENGINE = MergeTree() ORDER BY i;
|
||
INSERT INTO test SELECT number FROM numbers(500);
|
||
```
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
SET limit = 5;
|
||
SET offset = 7;
|
||
SELECT * FROM test LIMIT 10 OFFSET 100;
|
||
```
|
||
Result:
|
||
|
||
``` text
|
||
┌───i─┐
|
||
│ 107 │
|
||
│ 108 │
|
||
│ 109 │
|
||
└─────┘
|
||
```
|
||
|
||
## optimize_syntax_fuse_functions {#optimize_syntax_fuse_functions}
|
||
|
||
Enables to fuse aggregate functions with identical argument. It rewrites query contains at least two aggregate functions from [sum](../../sql-reference/aggregate-functions/reference/sum.md/#agg_function-sum), [count](../../sql-reference/aggregate-functions/reference/count.md/#agg_function-count) or [avg](../../sql-reference/aggregate-functions/reference/avg.md/#agg_function-avg) with identical argument to [sumCount](../../sql-reference/aggregate-functions/reference/sumcount.md/#agg_function-sumCount).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Functions with identical argument are not fused.
|
||
- 1 — Functions with identical argument are fused.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
CREATE TABLE fuse_tbl(a Int8, b Int8) Engine = Log;
|
||
SET optimize_syntax_fuse_functions = 1;
|
||
EXPLAIN SYNTAX SELECT sum(a), sum(b), count(b), avg(b) from fuse_tbl FORMAT TSV;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
SELECT
|
||
sum(a),
|
||
sumCount(b).1,
|
||
sumCount(b).2,
|
||
(sumCount(b).1) / (sumCount(b).2)
|
||
FROM fuse_tbl
|
||
```
|
||
|
||
## optimize_rewrite_aggregate_function_with_if
|
||
|
||
Rewrite aggregate functions with if expression as argument when logically equivalent.
|
||
For example, `avg(if(cond, col, null))` can be rewritten to `avgOrNullIf(cond, col)`. It may improve performance.
|
||
|
||
:::note
|
||
Supported only with experimental analyzer (`allow_experimental_analyzer = 1`).
|
||
:::
|
||
|
||
## allow_experimental_database_replicated {#allow_experimental_database_replicated}
|
||
|
||
Enables to create databases with [Replicated](../../engines/database-engines/replicated.md) engine.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
## database_replicated_initial_query_timeout_sec {#database_replicated_initial_query_timeout_sec}
|
||
|
||
Sets how long initial DDL query should wait for Replicated database to precess previous DDL queue entries in seconds.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Unlimited.
|
||
|
||
Default value: `300`.
|
||
|
||
## distributed_ddl_task_timeout {#distributed_ddl_task_timeout}
|
||
|
||
Sets timeout for DDL query responses from all hosts in cluster. If a DDL request has not been performed on all hosts, a response will contain a timeout error and a request will be executed in an async mode. Negative value means infinite.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Async mode.
|
||
- Negative integer — infinite timeout.
|
||
|
||
Default value: `180`.
|
||
|
||
## distributed_ddl_output_mode {#distributed_ddl_output_mode}
|
||
|
||
Sets format of distributed DDL query result.
|
||
|
||
Possible values:
|
||
|
||
- `throw` — Returns result set with query execution status for all hosts where query is finished. If query has failed on some hosts, then it will rethrow the first exception. If query is not finished yet on some hosts and [distributed_ddl_task_timeout](#distributed_ddl_task_timeout) exceeded, then it throws `TIMEOUT_EXCEEDED` exception.
|
||
- `none` — Is similar to throw, but distributed DDL query returns no result set.
|
||
- `null_status_on_timeout` — Returns `NULL` as execution status in some rows of result set instead of throwing `TIMEOUT_EXCEEDED` if query is not finished on the corresponding hosts.
|
||
- `never_throw` — Do not throw `TIMEOUT_EXCEEDED` and do not rethrow exceptions if query has failed on some hosts.
|
||
|
||
Default value: `throw`.
|
||
|
||
## flatten_nested {#flatten-nested}
|
||
|
||
Sets the data format of a [nested](../../sql-reference/data-types/nested-data-structures/nested.md) columns.
|
||
|
||
Possible values:
|
||
|
||
- 1 — Nested column is flattened to separate arrays.
|
||
- 0 — Nested column stays a single array of tuples.
|
||
|
||
Default value: `1`.
|
||
|
||
**Usage**
|
||
|
||
If the setting is set to `0`, it is possible to use an arbitrary level of nesting.
|
||
|
||
**Examples**
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
SET flatten_nested = 1;
|
||
CREATE TABLE t_nest (`n` Nested(a UInt32, b UInt32)) ENGINE = MergeTree ORDER BY tuple();
|
||
|
||
SHOW CREATE TABLE t_nest;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌─statement───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐
|
||
│ CREATE TABLE default.t_nest
|
||
(
|
||
`n.a` Array(UInt32),
|
||
`n.b` Array(UInt32)
|
||
)
|
||
ENGINE = MergeTree
|
||
ORDER BY tuple()
|
||
SETTINGS index_granularity = 8192 │
|
||
└─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
Query:
|
||
|
||
``` sql
|
||
SET flatten_nested = 0;
|
||
|
||
CREATE TABLE t_nest (`n` Nested(a UInt32, b UInt32)) ENGINE = MergeTree ORDER BY tuple();
|
||
|
||
SHOW CREATE TABLE t_nest;
|
||
```
|
||
|
||
Result:
|
||
|
||
``` text
|
||
┌─statement──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐
|
||
│ CREATE TABLE default.t_nest
|
||
(
|
||
`n` Nested(a UInt32, b UInt32)
|
||
)
|
||
ENGINE = MergeTree
|
||
ORDER BY tuple()
|
||
SETTINGS index_granularity = 8192 │
|
||
└────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
## external_table_functions_use_nulls {#external-table-functions-use-nulls}
|
||
|
||
Defines how [mysql](../../sql-reference/table-functions/mysql.md), [postgresql](../../sql-reference/table-functions/postgresql.md) and [odbc](../../sql-reference/table-functions/odbc.md)] table functions use Nullable columns.
|
||
|
||
Possible values:
|
||
|
||
- 0 — The table function explicitly uses Nullable columns.
|
||
- 1 — The table function implicitly uses Nullable columns.
|
||
|
||
Default value: `1`.
|
||
|
||
**Usage**
|
||
|
||
If the setting is set to `0`, the table function does not make Nullable columns and inserts default values instead of NULL. This is also applicable for NULL values inside arrays.
|
||
|
||
## allow_experimental_projection_optimization {#allow-experimental-projection-optimization}
|
||
|
||
Enables or disables [projection](../../engines/table-engines/mergetree-family/mergetree.md/#projections) optimization when processing `SELECT` queries.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Projection optimization disabled.
|
||
- 1 — Projection optimization enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
## force_optimize_projection {#force-optimize-projection}
|
||
|
||
Enables or disables the obligatory use of [projections](../../engines/table-engines/mergetree-family/mergetree.md/#projections) in `SELECT` queries, when projection optimization is enabled (see [allow_experimental_projection_optimization](#allow-experimental-projection-optimization) setting).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Projection optimization is not obligatory.
|
||
- 1 — Projection optimization is obligatory.
|
||
|
||
Default value: `0`.
|
||
|
||
## alter_sync {#alter-sync}
|
||
|
||
Allows to set up waiting for actions to be executed on replicas by [ALTER](../../sql-reference/statements/alter/index.md), [OPTIMIZE](../../sql-reference/statements/optimize.md) or [TRUNCATE](../../sql-reference/statements/truncate.md) queries.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Do not wait.
|
||
- 1 — Wait for own execution.
|
||
- 2 — Wait for everyone.
|
||
|
||
Default value: `1`.
|
||
|
||
## replication_wait_for_inactive_replica_timeout {#replication-wait-for-inactive-replica-timeout}
|
||
|
||
Specifies how long (in seconds) to wait for inactive replicas to execute [ALTER](../../sql-reference/statements/alter/index.md), [OPTIMIZE](../../sql-reference/statements/optimize.md) or [TRUNCATE](../../sql-reference/statements/truncate.md) queries.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Do not wait.
|
||
- Negative integer — Wait for unlimited time.
|
||
- Positive integer — The number of seconds to wait.
|
||
|
||
Default value: `120` seconds.
|
||
|
||
## regexp_max_matches_per_row {#regexp-max-matches-per-row}
|
||
|
||
Sets the maximum number of matches for a single regular expression per row. Use it to protect against memory overload when using greedy regular expression in the [extractAllGroupsHorizontal](../../sql-reference/functions/string-search-functions.md/#extractallgroups-horizontal) function.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `1000`.
|
||
|
||
## http_max_single_read_retries {#http-max-single-read-retries}
|
||
|
||
Sets the maximum number of retries during a single HTTP read.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
|
||
Default value: `1024`.
|
||
|
||
## log_queries_probability {#log-queries-probability}
|
||
|
||
Allows a user to write to [query_log](../../operations/system-tables/query_log.md), [query_thread_log](../../operations/system-tables/query_thread_log.md), and [query_views_log](../../operations/system-tables/query_views_log.md) system tables only a sample of queries selected randomly with the specified probability. It helps to reduce the load with a large volume of queries in a second.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Queries are not logged in the system tables.
|
||
- Positive floating-point number in the range [0..1]. For example, if the setting value is `0.5`, about half of the queries are logged in the system tables.
|
||
- 1 — All queries are logged in the system tables.
|
||
|
||
Default value: `1`.
|
||
|
||
## short_circuit_function_evaluation {#short-circuit-function-evaluation}
|
||
|
||
Allows calculating the [if](../../sql-reference/functions/conditional-functions.md/#if), [multiIf](../../sql-reference/functions/conditional-functions.md/#multiif), [and](../../sql-reference/functions/logical-functions.md/#logical-and-function), and [or](../../sql-reference/functions/logical-functions.md/#logical-or-function) functions according to a [short scheme](https://en.wikipedia.org/wiki/Short-circuit_evaluation). This helps optimize the execution of complex expressions in these functions and prevent possible exceptions (such as division by zero when it is not expected).
|
||
|
||
Possible values:
|
||
|
||
- `enable` — Enables short-circuit function evaluation for functions that are suitable for it (can throw an exception or computationally heavy).
|
||
- `force_enable` — Enables short-circuit function evaluation for all functions.
|
||
- `disable` — Disables short-circuit function evaluation.
|
||
|
||
Default value: `enable`.
|
||
|
||
## max_hyperscan_regexp_length {#max-hyperscan-regexp-length}
|
||
|
||
Defines the maximum length for each regular expression in the [hyperscan multi-match functions](../../sql-reference/functions/string-search-functions.md/#multimatchanyhaystack-pattern1-pattern2-patternn).
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 - The length is not limited.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SELECT multiMatchAny('abcd', ['ab','bcd','c','d']) SETTINGS max_hyperscan_regexp_length = 3;
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
┌─multiMatchAny('abcd', ['ab', 'bcd', 'c', 'd'])─┐
|
||
│ 1 │
|
||
└────────────────────────────────────────────────┘
|
||
```
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SELECT multiMatchAny('abcd', ['ab','bcd','c','d']) SETTINGS max_hyperscan_regexp_length = 2;
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
Exception: Regexp length too large.
|
||
```
|
||
|
||
**See Also**
|
||
|
||
- [max_hyperscan_regexp_total_length](#max-hyperscan-regexp-total-length)
|
||
|
||
## max_hyperscan_regexp_total_length {#max-hyperscan-regexp-total-length}
|
||
|
||
Sets the maximum length total of all regular expressions in each [hyperscan multi-match function](../../sql-reference/functions/string-search-functions.md/#multimatchanyhaystack-pattern1-pattern2-patternn).
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 - The length is not limited.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SELECT multiMatchAny('abcd', ['a','b','c','d']) SETTINGS max_hyperscan_regexp_total_length = 5;
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
┌─multiMatchAny('abcd', ['a', 'b', 'c', 'd'])─┐
|
||
│ 1 │
|
||
└─────────────────────────────────────────────┘
|
||
```
|
||
|
||
Query:
|
||
|
||
```sql
|
||
SELECT multiMatchAny('abcd', ['ab','bc','c','d']) SETTINGS max_hyperscan_regexp_total_length = 5;
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
Exception: Total regexp lengths too large.
|
||
```
|
||
|
||
**See Also**
|
||
|
||
- [max_hyperscan_regexp_length](#max-hyperscan-regexp-length)
|
||
|
||
## enable_positional_arguments {#enable-positional-arguments}
|
||
|
||
Enables or disables supporting positional arguments for [GROUP BY](../../sql-reference/statements/select/group-by.md), [LIMIT BY](../../sql-reference/statements/select/limit-by.md), [ORDER BY](../../sql-reference/statements/select/order-by.md) statements.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Positional arguments aren't supported.
|
||
- 1 — Positional arguments are supported: column numbers can use instead of column names.
|
||
|
||
Default value: `1`.
|
||
|
||
**Example**
|
||
|
||
Query:
|
||
|
||
```sql
|
||
CREATE TABLE positional_arguments(one Int, two Int, three Int) ENGINE=Memory();
|
||
|
||
INSERT INTO positional_arguments VALUES (10, 20, 30), (20, 20, 10), (30, 10, 20);
|
||
|
||
SELECT * FROM positional_arguments ORDER BY 2,3;
|
||
```
|
||
|
||
Result:
|
||
|
||
```text
|
||
┌─one─┬─two─┬─three─┐
|
||
│ 30 │ 10 │ 20 │
|
||
│ 20 │ 20 │ 10 │
|
||
│ 10 │ 20 │ 30 │
|
||
└─────┴─────┴───────┘
|
||
```
|
||
|
||
## enable_extended_results_for_datetime_functions {#enable-extended-results-for-datetime-functions}
|
||
|
||
Enables or disables returning results of type:
|
||
- `Date32` with extended range (compared to type `Date`) for functions [toStartOfYear](../../sql-reference/functions/date-time-functions.md/#tostartofyear), [toStartOfISOYear](../../sql-reference/functions/date-time-functions.md/#tostartofisoyear), [toStartOfQuarter](../../sql-reference/functions/date-time-functions.md/#tostartofquarter), [toStartOfMonth](../../sql-reference/functions/date-time-functions.md/#tostartofmonth), [toStartOfWeek](../../sql-reference/functions/date-time-functions.md/#tostartofweek), [toMonday](../../sql-reference/functions/date-time-functions.md/#tomonday) and [toLastDayOfMonth](../../sql-reference/functions/date-time-functions.md/#tolastdayofmonth).
|
||
- `DateTime64` with extended range (compared to type `DateTime`) for functions [toStartOfDay](../../sql-reference/functions/date-time-functions.md/#tostartofday), [toStartOfHour](../../sql-reference/functions/date-time-functions.md/#tostartofhour), [toStartOfMinute](../../sql-reference/functions/date-time-functions.md/#tostartofminute), [toStartOfFiveMinutes](../../sql-reference/functions/date-time-functions.md/#tostartoffiveminutes), [toStartOfTenMinutes](../../sql-reference/functions/date-time-functions.md/#tostartoftenminutes), [toStartOfFifteenMinutes](../../sql-reference/functions/date-time-functions.md/#tostartoffifteenminutes) and [timeSlot](../../sql-reference/functions/date-time-functions.md/#timeslot).
|
||
|
||
Possible values:
|
||
|
||
- 0 — Functions return `Date` or `DateTime` for all types of arguments.
|
||
- 1 — Functions return `Date32` or `DateTime64` for `Date32` or `DateTime64` arguments and `Date` or `DateTime` otherwise.
|
||
|
||
Default value: `0`.
|
||
|
||
## optimize_move_to_prewhere {#optimize_move_to_prewhere}
|
||
|
||
Enables or disables automatic [PREWHERE](../../sql-reference/statements/select/prewhere.md) optimization in [SELECT](../../sql-reference/statements/select/index.md) queries.
|
||
|
||
Works only for [*MergeTree](../../engines/table-engines/mergetree-family/index.md) tables.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Automatic `PREWHERE` optimization is disabled.
|
||
- 1 — Automatic `PREWHERE` optimization is enabled.
|
||
|
||
Default value: `1`.
|
||
|
||
## optimize_move_to_prewhere_if_final {#optimize_move_to_prewhere_if_final}
|
||
|
||
Enables or disables automatic [PREWHERE](../../sql-reference/statements/select/prewhere.md) optimization in [SELECT](../../sql-reference/statements/select/index.md) queries with [FINAL](../../sql-reference/statements/select/from.md/#select-from-final) modifier.
|
||
|
||
Works only for [*MergeTree](../../engines/table-engines/mergetree-family/index.md) tables.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Automatic `PREWHERE` optimization in `SELECT` queries with `FINAL` modifier is disabled.
|
||
- 1 — Automatic `PREWHERE` optimization in `SELECT` queries with `FINAL` modifier is enabled.
|
||
|
||
Default value: `0`.
|
||
|
||
**See Also**
|
||
|
||
- [optimize_move_to_prewhere](#optimize_move_to_prewhere) setting
|
||
|
||
## optimize_using_constraints
|
||
|
||
Use [constraints](../../sql-reference/statements/create/table#constraints) for query optimization. The default is `false`.
|
||
|
||
Possible values:
|
||
|
||
- true, false
|
||
|
||
## optimize_append_index
|
||
|
||
Use [constraints](../../sql-reference/statements/create/table#constraints) in order to append index condition. The default is `false`.
|
||
|
||
Possible values:
|
||
|
||
- true, false
|
||
|
||
## optimize_substitute_columns
|
||
|
||
Use [constraints](../../sql-reference/statements/create/table#constraints) for column substitution. The default is `false`.
|
||
|
||
Possible values:
|
||
|
||
- true, false
|
||
|
||
## describe_include_subcolumns {#describe_include_subcolumns}
|
||
|
||
Enables describing subcolumns for a [DESCRIBE](../../sql-reference/statements/describe-table.md) query. For example, members of a [Tuple](../../sql-reference/data-types/tuple.md) or subcolumns of a [Map](../../sql-reference/data-types/map.md/#map-subcolumns), [Nullable](../../sql-reference/data-types/nullable.md/#finding-null) or an [Array](../../sql-reference/data-types/array.md/#array-size) data type.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Subcolumns are not included in `DESCRIBE` queries.
|
||
- 1 — Subcolumns are included in `DESCRIBE` queries.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
See an example for the [DESCRIBE](../../sql-reference/statements/describe-table.md) statement.
|
||
|
||
|
||
## alter_partition_verbose_result {#alter-partition-verbose-result}
|
||
|
||
Enables or disables the display of information about the parts to which the manipulation operations with partitions and parts have been successfully applied.
|
||
Applicable to [ATTACH PARTITION|PART](../../sql-reference/statements/alter/partition.md/#alter_attach-partition) and to [FREEZE PARTITION](../../sql-reference/statements/alter/partition.md/#alter_freeze-partition).
|
||
|
||
Possible values:
|
||
|
||
- 0 — disable verbosity.
|
||
- 1 — enable verbosity.
|
||
|
||
Default value: `0`.
|
||
|
||
**Example**
|
||
|
||
```sql
|
||
CREATE TABLE test(a Int64, d Date, s String) ENGINE = MergeTree PARTITION BY toYYYYMM(d) ORDER BY a;
|
||
INSERT INTO test VALUES(1, '2021-01-01', '');
|
||
INSERT INTO test VALUES(1, '2021-01-01', '');
|
||
ALTER TABLE test DETACH PARTITION ID '202101';
|
||
|
||
ALTER TABLE test ATTACH PARTITION ID '202101' SETTINGS alter_partition_verbose_result = 1;
|
||
|
||
┌─command_type─────┬─partition_id─┬─part_name────┬─old_part_name─┐
|
||
│ ATTACH PARTITION │ 202101 │ 202101_7_7_0 │ 202101_5_5_0 │
|
||
│ ATTACH PARTITION │ 202101 │ 202101_8_8_0 │ 202101_6_6_0 │
|
||
└──────────────────┴──────────────┴──────────────┴───────────────┘
|
||
|
||
ALTER TABLE test FREEZE SETTINGS alter_partition_verbose_result = 1;
|
||
|
||
┌─command_type─┬─partition_id─┬─part_name────┬─backup_name─┬─backup_path───────────────────┬─part_backup_path────────────────────────────────────────────┐
|
||
│ FREEZE ALL │ 202101 │ 202101_7_7_0 │ 8 │ /var/lib/clickhouse/shadow/8/ │ /var/lib/clickhouse/shadow/8/data/default/test/202101_7_7_0 │
|
||
│ FREEZE ALL │ 202101 │ 202101_8_8_0 │ 8 │ /var/lib/clickhouse/shadow/8/ │ /var/lib/clickhouse/shadow/8/data/default/test/202101_8_8_0 │
|
||
└──────────────┴──────────────┴──────────────┴─────────────┴───────────────────────────────┴─────────────────────────────────────────────────────────────┘
|
||
```
|
||
|
||
## min_bytes_to_use_mmap_io {#min-bytes-to-use-mmap-io}
|
||
|
||
This is an experimental setting. Sets the minimum amount of memory for reading large files without copying data from the kernel to userspace. Recommended threshold is about 64 MB, because [mmap/munmap](https://en.wikipedia.org/wiki/Mmap) is slow. It makes sense only for large files and helps only if data reside in the page cache.
|
||
|
||
Possible values:
|
||
|
||
- Positive integer.
|
||
- 0 — Big files read with only copying data from kernel to userspace.
|
||
|
||
Default value: `0`.
|
||
|
||
## shutdown_wait_unfinished_queries {#shutdown_wait_unfinished_queries}
|
||
|
||
Enables or disables waiting unfinished queries when shutdown server.
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disabled.
|
||
- 1 — Enabled. The wait time equal shutdown_wait_unfinished config.
|
||
|
||
Default value: 0.
|
||
|
||
## shutdown_wait_unfinished {#shutdown_wait_unfinished}
|
||
|
||
The waiting time in seconds for currently handled connections when shutdown server.
|
||
|
||
Default Value: 5.
|
||
|
||
## memory_overcommit_ratio_denominator {#memory_overcommit_ratio_denominator}
|
||
|
||
It represents soft memory limit in case when hard limit is reached on user level.
|
||
This value is used to compute overcommit ratio for the query.
|
||
Zero means skip the query.
|
||
Read more about [memory overcommit](memory-overcommit.md).
|
||
|
||
Default value: `1GiB`.
|
||
|
||
## memory_usage_overcommit_max_wait_microseconds {#memory_usage_overcommit_max_wait_microseconds}
|
||
|
||
Maximum time thread will wait for memory to be freed in the case of memory overcommit on a user level.
|
||
If the timeout is reached and memory is not freed, an exception is thrown.
|
||
Read more about [memory overcommit](memory-overcommit.md).
|
||
|
||
Default value: `5000000`.
|
||
|
||
## memory_overcommit_ratio_denominator_for_user {#memory_overcommit_ratio_denominator_for_user}
|
||
|
||
It represents soft memory limit in case when hard limit is reached on global level.
|
||
This value is used to compute overcommit ratio for the query.
|
||
Zero means skip the query.
|
||
Read more about [memory overcommit](memory-overcommit.md).
|
||
|
||
Default value: `1GiB`.
|
||
|
||
## Schema Inference settings
|
||
|
||
### schema_inference_use_cache_for_file {schema_inference_use_cache_for_file}
|
||
|
||
Enable schemas cache for schema inference in `file` table function.
|
||
|
||
Default value: `true`.
|
||
|
||
### schema_inference_use_cache_for_s3 {schema_inference_use_cache_for_s3}
|
||
|
||
Enable schemas cache for schema inference in `s3` table function.
|
||
|
||
Default value: `true`.
|
||
|
||
### schema_inference_use_cache_for_url {schema_inference_use_cache_for_url}
|
||
|
||
Enable schemas cache for schema inference in `url` table function.
|
||
|
||
Default value: `true`.
|
||
|
||
### schema_inference_use_cache_for_hdfs {schema_inference_use_cache_for_hdfs}
|
||
|
||
Enable schemas cache for schema inference in `hdfs` table function.
|
||
|
||
Default value: `true`.
|
||
|
||
### schema_inference_cache_require_modification_time_for_url {#schema_inference_cache_require_modification_time_for_url}
|
||
|
||
Use schema from cache for URL with last modification time validation (for urls with Last-Modified header). If this setting is enabled and URL doesn't have Last-Modified header, schema from cache won't be used.
|
||
|
||
Default value: `true`.
|
||
|
||
### use_structure_from_insertion_table_in_table_functions {use_structure_from_insertion_table_in_table_functions}
|
||
|
||
Use structure from insertion table instead of schema inference from data.
|
||
|
||
Possible values:
|
||
- 0 - disabled
|
||
- 1 - enabled
|
||
- 2 - auto
|
||
|
||
Default value: 2.
|
||
|
||
## compatibility {#compatibility}
|
||
|
||
The `compatibility` setting causes ClickHouse to use the default settings of a previous version of ClickHouse, where the previous version is provided as the setting.
|
||
|
||
If settings are set to non-default values, then those settings are honored (only settings that have not been modified are affected by the `compatibility` setting).
|
||
|
||
This setting takes a ClickHouse version number as a string, like `22.3`, `22.8`. An empty value means that this setting is disabled.
|
||
|
||
Disabled by default.
|
||
|
||
:::note
|
||
In ClickHouse Cloud the compatibility setting must be set by ClickHouse Cloud support. Please [open a case](https://clickhouse.cloud/support) to have it set.
|
||
:::
|
||
|
||
## allow_settings_after_format_in_insert {#allow_settings_after_format_in_insert}
|
||
|
||
Control whether `SETTINGS` after `FORMAT` in `INSERT` queries is allowed or not. It is not recommended to use this, since this may interpret part of `SETTINGS` as values.
|
||
|
||
Example:
|
||
|
||
```sql
|
||
INSERT INTO FUNCTION null('foo String') SETTINGS max_threads=1 VALUES ('bar');
|
||
```
|
||
|
||
But the following query will work only with `allow_settings_after_format_in_insert`:
|
||
|
||
```sql
|
||
SET allow_settings_after_format_in_insert=1;
|
||
INSERT INTO FUNCTION null('foo String') VALUES ('bar') SETTINGS max_threads=1;
|
||
```
|
||
|
||
Possible values:
|
||
|
||
- 0 — Disallow.
|
||
- 1 — Allow.
|
||
|
||
Default value: `0`.
|
||
|
||
:::note
|
||
Use this setting only for backward compatibility if your use cases depend on old syntax.
|
||
:::
|
||
|
||
## final {#final}
|
||
|
||
Automatically applies [FINAL](../../sql-reference/statements/select/from/#final-modifier) modifier to all tables in a query, to tables where [FINAL](../../sql-reference/statements/select/from/#final-modifier) is applicable, including joined tables and tables in sub-queries, and
|
||
distributed tables.
|
||
|
||
Possible values:
|
||
|
||
- 0 - disabled
|
||
- 1 - enabled
|
||
|
||
Default value: `0`.
|
||
|
||
Example:
|
||
|
||
```sql
|
||
CREATE TABLE test
|
||
(
|
||
key Int64,
|
||
some String
|
||
)
|
||
ENGINE = ReplacingMergeTree
|
||
ORDER BY key;
|
||
|
||
INSERT INTO test FORMAT Values (1, 'first');
|
||
INSERT INTO test FORMAT Values (1, 'second');
|
||
|
||
SELECT * FROM test;
|
||
┌─key─┬─some───┐
|
||
│ 1 │ second │
|
||
└─────┴────────┘
|
||
┌─key─┬─some──┐
|
||
│ 1 │ first │
|
||
└─────┴───────┘
|
||
|
||
SELECT * FROM test SETTINGS final = 1;
|
||
┌─key─┬─some───┐
|
||
│ 1 │ second │
|
||
└─────┴────────┘
|
||
|
||
SET final = 1;
|
||
SELECT * FROM test;
|
||
┌─key─┬─some───┐
|
||
│ 1 │ second │
|
||
└─────┴────────┘
|
||
```
|
||
|
||
## asterisk_include_materialized_columns {#asterisk_include_materialized_columns}
|
||
|
||
Include [MATERIALIZED](../../sql-reference/statements/create/table/#materialized) columns for wildcard query (`SELECT *`).
|
||
|
||
Possible values:
|
||
|
||
- 0 - disabled
|
||
- 1 - enabled
|
||
|
||
Default value: `0`.
|
||
|
||
## asterisk_include_alias_columns {#asterisk_include_alias_columns}
|
||
|
||
Include [ALIAS](../../sql-reference/statements/create/table/#alias) columns for wildcard query (`SELECT *`).
|
||
|
||
Possible values:
|
||
|
||
- 0 - disabled
|
||
- 1 - enabled
|
||
|
||
Default value: `0`.
|