mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-29 11:02:08 +00:00
67c2e50331
* update presentations * CLICKHOUSE-2936: redirect from clickhouse.yandex.ru and clickhouse.yandex.com * update submodule * lost files * CLICKHOUSE-2981: prefer sphinx docs over original reference * CLICKHOUSE-2981: docs styles more similar to main website + add flags to switch language links * update presentations * Less confusing directory structure (docs -> doc/reference/) * Minify sphinx docs too * Website release script: fail fast + pass docker hash on deploy * Do not underline links in docs * shorter * cleanup docker images * tune nginx config * CLICKHOUSE-3043: get rid of habrastorage links * Lost translation * CLICKHOUSE-2936: temporary client-side redirect * behaves weird in test * put redirect back * CLICKHOUSE-3047: copy docs txts to public too * move to proper file * remove old pages to avoid confusion * Remove reference redirect warning for now * Refresh README.md * Yellow buttons in docs * Use svg flags instead of unicode ones in docs * fix test website instance * Put flags to separate files * wrong flag * Copy Yandex.Metrica introduction from main page to docs * Yet another home page structure change, couple new blocks (CLICKHOUSE-3045) * Update Contacts section * CLICKHOUSE-2849: more detailed legal information * CLICKHOUSE-2978 preparation - split by files * More changes in Contacts block * Tune texts on index page * update presentations * One more benchmark * Add usage sections to index page, adapted from slides * Get the roadmap started, based on slides from last ClickHouse Meetup * CLICKHOUSE-2977: some rendering tuning * Get rid of excessive section in the end of getting started * Make headers linkable * CLICKHOUSE-2981: links to editing reference - https://github.com/yandex/ClickHouse/issues/849 * CLICKHOUSE-2981: fix mobile styles in docs * Ban crawling of duplicating docs * Open some external links in new tab * Ban old docs too * Lots of trivial fixes in english docs * Lots of trivial fixes in russian docs * Remove getting started copies in markdown * Add Yandex.Webmaster * Fix some sphinx warnings * More warnings fixed in english docs * More sphinx warnings fixed * Add code-block:: text * More code-block:: text * These headers look not that well * Better switch between documentation languages * merge use_case.rst into ya_metrika_task.rst * Edit the agg_functions.rst texts * Add lost empty lines * Lost blank lines * Add new logo sizes * update presentations * Next step in migrating to new documentation * Fix all warnings in en reference * Fix all warnings in ru reference * Re-arrange existing reference * Move operation tips to main reference * Fix typos noticed by milovidov@ * Get rid of zookeeper.md * Looks like duplicate of tutorial.html * Fix some mess with html tags in tutorial * No idea why nobody noticed this before, but it was completely not clear whet to get the data * Match code block styling between main and tutorial pages (in favor of the latter) * Get rid of some copypaste in tutorial * Normalize header styles * Move example_datasets to sphinx * Move presentations submodule to website * Move and update README.md * No point in duplicating articles from habrahabr here * Move development-related docs as is for now * doc/reference/ -> docs/ (to match the URL on website) * Adapt links to match the previous commit * Adapt development docs to rst (still lacks translation and strikethrough support) * clean on release * blacklist presentations in gulp * strikethrough support in sphinx * just copy development folder for now * fix weird introduction in style article * Style guide translation (WIP) * Finish style guide translation to English * gulp clean separately * Update year in LICENSE * Initial CONTRIBUTING.md * Fix remaining links to old docs in tutorial * Some tutorial fixes * Typo * Another typo * Update list of authors from yandex-team accoding to git log
174 lines
8.0 KiB
ReStructuredText
174 lines
8.0 KiB
ReStructuredText
Restrictions on query complexity
|
|
================================
|
|
Restrictions on query complexity are part of the settings.
|
|
They are used in order to provide safer execution from the user interface.
|
|
Almost all the restrictions only apply to SELECTs.
|
|
For distributed query processing, restrictions are applied on each server separately.
|
|
|
|
Restrictions on the "maximum amount of something" can take the value 0, which means "unrestricted".
|
|
Most restrictions also have an 'overflow_mode' setting, meaning what to do when the limit is exceeded.
|
|
It can take one of two values: 'throw' or 'break'. Restrictions on aggregation (``group_by_overflow_mode``) also have the value ``any``.
|
|
|
|
``throw`` - Throw an exception (default).
|
|
|
|
``break`` - Stop executing the query and return the partial result, as if the source data ran out.
|
|
|
|
``any`` (only for group_by_overflow_mode) - Continuing aggregation for the keys that got into the set, but don't add new keys to the set.
|
|
|
|
readonly
|
|
--------
|
|
If set to 0, allows to run any queries.
|
|
If set to 1, allows to run only queries that don't change data or settings (e.g. SELECT or SHOW). INSERT and SET are forbidden.
|
|
If set to 2, allows to run queries that don't change data (SELECT, SHOW) and allows to change settings (SET).
|
|
|
|
After you set the read-only mode, you won't be able to disable it in the current session.
|
|
|
|
When using the GET method in the HTTP interface, 'readonly = 1' is set automatically. In other words, for queries that modify data, you can only use the POST method. You can send the query itself either in the POST body, or in the URL parameter.
|
|
|
|
max_memory_usage
|
|
----------------
|
|
The maximum amount of memory consumption when running a query on a single server. By default, 10 GB.
|
|
|
|
The setting doesn't consider the volume of available memory or the total volume of memory on the machine.
|
|
The restriction applies to a single query within a single server.
|
|
You can use SHOW PROCESSLIST to see the current memory consumption for each query.
|
|
In addition, the peak memory consumption is tracked for each query and written to the log.
|
|
|
|
Certain cases of memory consumption are not tracked:
|
|
* Large constants (for example, a very long string constant).
|
|
* The states of 'groupArray' aggregate functions, and also 'quantile' (it is tracked for 'quantileTiming').
|
|
|
|
Memory consumption is not fully considered for aggregate function states ``min``, ``max``, ``any``, ``anyLast``, ``argMin``, and ``argMax`` from String and Array arguments.
|
|
|
|
max_rows_to_read
|
|
----------------
|
|
The following restrictions can be checked on each block (instead of on each row). That is, the restrictions can be broken a little.
|
|
When running a query in multiple threads, the following restrictions apply to each thread separately.
|
|
|
|
Maximum number of rows that can be read from a table when running a query.
|
|
|
|
max_bytes_to_read
|
|
-----------------
|
|
Maximum number of bytes (uncompressed data) that can be read from a table when running a query.
|
|
|
|
read_overflow_mode
|
|
------------------
|
|
What to do when the volume of data read exceeds one of the limits: ``throw`` or ``break``. ``By default, throw``.
|
|
|
|
max_rows_to_group_by
|
|
--------------------
|
|
Maximum number of unique keys received from aggregation. This setting lets you limit memory consumption when aggregating.
|
|
|
|
group_by_overflow_mode
|
|
----------------------
|
|
What to do when the number of unique keys for aggregation exceeds the limit: ``throw``, ``break``, or ``any``. ``By default, throw``.
|
|
Using the 'any' value lets you run an approximation of GROUP BY. The quality of this approximation depends on the statistical nature of the data.
|
|
|
|
max_rows_to_sort
|
|
----------------
|
|
Maximum number of rows before sorting. This allows you to limit memory consumption when sorting.
|
|
|
|
max_bytes_to_sort
|
|
-----------------
|
|
Maximum number of bytes before sorting.
|
|
|
|
sort_overflow_mode
|
|
------------------
|
|
What to do if the number of rows received before sorting exceeds one of the limits: ``throw`` or ``break``. ``By default, throw``.
|
|
|
|
max_result_rows
|
|
---------------
|
|
Limit on the number of rows in the result. Also checked for subqueries, and on remote servers when running parts of a distributed query.
|
|
|
|
max_result_bytes
|
|
----------------
|
|
Limit on the number of bytes in the result. The same as the previous setting.
|
|
|
|
result_overflow_mode
|
|
--------------------
|
|
What to do if the volume of the result exceeds one of the limits: ``throw`` or ``break``. By default, throw.
|
|
Using ``break`` is similar to using ``LIMIT``.
|
|
|
|
max_execution_time
|
|
------------------
|
|
Maximum query execution time in seconds.
|
|
At this time, it is not checked for one of the sorting stages, or when merging and finalizing aggregate functions.
|
|
|
|
timeout_overflow_mode
|
|
---------------------
|
|
What to do if the query is run longer than ``max_execution_time``: ``throw`` or ``break``. ``By default, throw``.
|
|
|
|
min_execution_speed
|
|
-------------------
|
|
Minimal execution speed in rows per second. Checked on every data block when ``timeout_before_checking_execution_speed`` expires. If the execution speed is lower, an exception is thrown.
|
|
|
|
timeout_before_checking_execution_speed
|
|
---------------------------------------
|
|
Checks that execution speed is not too slow (no less than ``min_execution_speed``), after the specified time in seconds has expired.
|
|
|
|
max_columns_to_read
|
|
-------------------
|
|
Maximum number of columns that can be read from a table in a single query. If a query requires reading a greater number of columns, it throws an exception.
|
|
|
|
max_temporary_columns
|
|
---------------------
|
|
Maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns. If there are more temporary columns than this, it throws an exception.
|
|
|
|
max_temporary_non_const_columns
|
|
-------------------------------
|
|
The same thing as 'max_temporary_columns', but without counting constant columns.
|
|
Note that constant columns are formed fairly often when running a query, but they require approximately zero computing resources.
|
|
|
|
max_subquery_depth
|
|
------------------
|
|
Maximum nesting depth of subqueries. If subqueries are deeper, an exception is thrown. ``By default, 100``.
|
|
|
|
max_pipeline_depth
|
|
------------------
|
|
Maximum pipeline depth. Corresponds to the number of transformations that each data block goes through during query processing. Counted within the limits of a single server. If the pipeline depth is greater, an exception is thrown. By default, 1000.
|
|
|
|
max_ast_depth
|
|
-------------
|
|
Maximum nesting depth of a query syntactic tree. If exceeded, an exception is thrown. At this time, it isn't checked during parsing, but only after parsing the query. That is, a syntactic tree that is too deep can be created during parsing, but the query will fail. By default, 1000.
|
|
|
|
max_ast_elements
|
|
----------------
|
|
Maximum number of elements in a query syntactic tree. If exceeded, an exception is thrown.
|
|
In the same way as the previous setting, it is checked only after parsing the query. ``By default, 10,000``.
|
|
|
|
max_rows_in_set
|
|
---------------
|
|
Maximum number of rows for a data set in the IN clause created from a subquery.
|
|
|
|
max_bytes_in_set
|
|
----------------
|
|
Maximum number of bytes (uncompressed data) used by a set in the IN clause created from a subquery.
|
|
|
|
set_overflow_mode
|
|
-----------------
|
|
What to do when the amount of data exceeds one of the limits: ``throw`` or ``break``. ``By default, throw``.
|
|
|
|
max_rows_in_distinct
|
|
--------------------
|
|
Maximum number of different rows when using DISTINCT.
|
|
|
|
max_bytes_in_distinct
|
|
---------------------
|
|
Maximum number of bytes used by a hash table when using DISTINCT.
|
|
|
|
distinct_overflow_mode
|
|
----------------------
|
|
What to do when the amount of data exceeds one of the limits: ``throw`` or ``break``. ``By default, throw``.
|
|
|
|
max_rows_to_transfer
|
|
--------------------
|
|
Maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
|
|
|
|
max_bytes_to_transfer
|
|
---------------------
|
|
Maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary table when using GLOBAL IN.
|
|
|
|
transfer_overflow_mode
|
|
----------------------
|
|
What to do when the amount of data exceeds one of the limits: ``throw`` or ``break``. ``By default, throw``.
|