mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-09-20 08:40:50 +00:00
Merge branch 'master' into stress_s3
This commit is contained in:
commit
cdc7f63f4f
@ -164,7 +164,6 @@ if (HAS_RESERVED_IDENTIFIER)
|
||||
add_compile_definitions (HAS_RESERVED_IDENTIFIER)
|
||||
endif ()
|
||||
|
||||
# If turned `ON`, assumes the user has either the system GTest library or the bundled one.
|
||||
option(ENABLE_TESTS "Provide unit_test_dbms target with Google.Test unit tests" ON)
|
||||
option(ENABLE_EXAMPLES "Build all example programs in 'examples' subdirectories" OFF)
|
||||
|
||||
@ -246,7 +245,8 @@ else ()
|
||||
endif ()
|
||||
|
||||
# Create BuildID when using lld. For other linkers it is created by default.
|
||||
if (LINKER_NAME MATCHES "lld$")
|
||||
# (NOTE: LINKER_NAME can be either path or name, and in different variants)
|
||||
if (LINKER_NAME MATCHES "lld")
|
||||
# SHA1 is not cryptographically secure but it is the best what lld is offering.
|
||||
set (CMAKE_EXE_LINKER_FLAGS "${CMAKE_EXE_LINKER_FLAGS} -Wl,--build-id=sha1")
|
||||
endif ()
|
||||
@ -601,6 +601,7 @@ if (NATIVE_BUILD_TARGETS
|
||||
"-DCMAKE_C_COMPILER=${CMAKE_C_COMPILER}"
|
||||
"-DCMAKE_CXX_COMPILER=${CMAKE_CXX_COMPILER}"
|
||||
"-DENABLE_CCACHE=${ENABLE_CCACHE}"
|
||||
"-DENABLE_CLICKHOUSE_SELF_EXTRACTING=${ENABLE_CLICKHOUSE_SELF_EXTRACTING}"
|
||||
${CMAKE_SOURCE_DIR}
|
||||
WORKING_DIRECTORY "${NATIVE_BUILD_DIR}"
|
||||
COMMAND_ECHO STDOUT)
|
||||
|
@ -26,7 +26,7 @@ ARG PACKAGES="clickhouse-client clickhouse-server clickhouse-common-static"
|
||||
|
||||
# set non-empty deb_location_url url to create a docker image
|
||||
# from debs created by CI build, for example:
|
||||
# docker build . --network host --build-arg version="21.4.1.6282" --build-arg deb_location_url="https://clickhouse-builds.s3.yandex.net/21852/069cfbff388b3d478d1a16dc7060b48073f5d522/clickhouse_build_check/clang-11_relwithdebuginfo_none_bundled_unsplitted_disable_False_deb/" -t filimonovq/clickhouse-server:pr21852
|
||||
# docker build . --network host --build-arg version="21.4.1.6282" --build-arg deb_location_url="https://..." -t ...
|
||||
ARG deb_location_url=""
|
||||
|
||||
# set non-empty single_binary_location_url to create docker image
|
||||
|
@ -12,7 +12,7 @@ stage=${stage:-}
|
||||
script_dir="$( cd "$( dirname "${BASH_SOURCE[0]}" )" >/dev/null 2>&1 && pwd )"
|
||||
echo "$script_dir"
|
||||
repo_dir=ch
|
||||
BINARY_TO_DOWNLOAD=${BINARY_TO_DOWNLOAD:="clang-14_debug_none_bundled_unsplitted_disable_False_binary"}
|
||||
BINARY_TO_DOWNLOAD=${BINARY_TO_DOWNLOAD:="clang-14_debug_none_unsplitted_disable_False_binary"}
|
||||
BINARY_URL_TO_DOWNLOAD=${BINARY_URL_TO_DOWNLOAD:="https://clickhouse-builds.s3.amazonaws.com/$PR_TO_TEST/$SHA_TO_TEST/clickhouse_build_check/$BINARY_TO_DOWNLOAD/clickhouse"}
|
||||
|
||||
function clone
|
||||
|
@ -2,7 +2,7 @@
|
||||
set -euo pipefail
|
||||
|
||||
|
||||
CLICKHOUSE_PACKAGE=${CLICKHOUSE_PACKAGE:="https://clickhouse-builds.s3.amazonaws.com/$PR_TO_TEST/$SHA_TO_TEST/clickhouse_build_check/clang-14_relwithdebuginfo_none_bundled_unsplitted_disable_False_binary/clickhouse"}
|
||||
CLICKHOUSE_PACKAGE=${CLICKHOUSE_PACKAGE:="https://clickhouse-builds.s3.amazonaws.com/$PR_TO_TEST/$SHA_TO_TEST/clickhouse_build_check/clang-14_relwithdebuginfo_none_unsplitted_disable_False_binary/clickhouse"}
|
||||
CLICKHOUSE_REPO_PATH=${CLICKHOUSE_REPO_PATH:=""}
|
||||
|
||||
|
||||
|
@ -88,13 +88,15 @@ sleep 5
|
||||
function run_tests()
|
||||
{
|
||||
set -x
|
||||
# We can have several additional options so we path them as array because it's
|
||||
# more idiologically correct.
|
||||
# We can have several additional options so we pass them as array because it is more ideologically correct.
|
||||
read -ra ADDITIONAL_OPTIONS <<< "${ADDITIONAL_OPTIONS:-}"
|
||||
|
||||
HIGH_LEVEL_COVERAGE=YES
|
||||
|
||||
# Use random order in flaky check
|
||||
if [ "$NUM_TRIES" -gt "1" ]; then
|
||||
ADDITIONAL_OPTIONS+=('--order=random')
|
||||
HIGH_LEVEL_COVERAGE=NO
|
||||
fi
|
||||
|
||||
if [[ -n "$USE_S3_STORAGE_FOR_MERGE_TREE" ]] && [[ "$USE_S3_STORAGE_FOR_MERGE_TREE" -eq 1 ]]; then
|
||||
@ -117,12 +119,17 @@ function run_tests()
|
||||
ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_NUM")
|
||||
ADDITIONAL_OPTIONS+=('--run-by-hash-total')
|
||||
ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_TOTAL")
|
||||
HIGH_LEVEL_COVERAGE=NO
|
||||
fi
|
||||
|
||||
if [[ -n "$USE_DATABASE_ORDINARY" ]] && [[ "$USE_DATABASE_ORDINARY" -eq 1 ]]; then
|
||||
ADDITIONAL_OPTIONS+=('--db-engine=Ordinary')
|
||||
fi
|
||||
|
||||
if [[ "${HIGH_LEVEL_COVERAGE}" = "YES" ]]; then
|
||||
ADDITIONAL_OPTIONS+=('--report-coverage')
|
||||
fi
|
||||
|
||||
set +e
|
||||
clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time \
|
||||
--test-runs "$NUM_TRIES" "${ADDITIONAL_OPTIONS[@]}" 2>&1 \
|
||||
|
@ -105,12 +105,16 @@ EOL
|
||||
|
||||
function stop()
|
||||
{
|
||||
local pid
|
||||
# Preserve the pid, since the server can hung after the PID will be deleted.
|
||||
pid="$(cat /var/run/clickhouse-server/clickhouse-server.pid)"
|
||||
|
||||
clickhouse stop --do-not-kill && return
|
||||
# We failed to stop the server with SIGTERM. Maybe it hang, let's collect stacktraces.
|
||||
kill -TERM "$(pidof gdb)" ||:
|
||||
sleep 5
|
||||
echo "thread apply all backtrace (on stop)" >> /test_output/gdb.log
|
||||
gdb -batch -ex 'thread apply all backtrace' -p "$(cat /var/run/clickhouse-server/clickhouse-server.pid)" | ts '%Y-%m-%d %H:%M:%S' >> /test_output/gdb.log
|
||||
gdb -batch -ex 'thread apply all backtrace' -p "$pid" | ts '%Y-%m-%d %H:%M:%S' >> /test_output/gdb.log
|
||||
clickhouse stop --force
|
||||
}
|
||||
|
||||
@ -474,6 +478,13 @@ else
|
||||
[ -s /test_output/bc_check_fatal_messages.txt ] || rm /test_output/bc_check_fatal_messages.txt
|
||||
fi
|
||||
|
||||
dmesg -T > /test_output/dmesg.log
|
||||
|
||||
# OOM in dmesg -- those are real
|
||||
grep -q -F -e 'Out of memory: Killed process' -e 'oom_reaper: reaped process' -e 'oom-kill:constraint=CONSTRAINT_NONE' /test_output/dmesg.log \
|
||||
&& echo -e 'OOM in dmesg\tFAIL' >> /test_output/test_results.tsv \
|
||||
|| echo -e 'No OOM in dmesg\tOK' >> /test_output/test_results.tsv
|
||||
|
||||
tar -chf /test_output/coordination.tar /var/lib/clickhouse/coordination ||:
|
||||
mv /var/log/clickhouse-server/stderr.log /test_output/
|
||||
|
||||
@ -495,5 +506,3 @@ for core in core.*; do
|
||||
pigz $core
|
||||
mv $core.gz /test_output/
|
||||
done
|
||||
|
||||
dmesg -T > /test_output/dmesg.log
|
||||
|
@ -14,8 +14,6 @@ do
|
||||
|| curl -fO "https://packages.clickhouse.com/tgz/stable/$PKG-$LATEST_VERSION.tgz"
|
||||
done
|
||||
|
||||
exit 0
|
||||
|
||||
tar -xzvf "clickhouse-common-static-$LATEST_VERSION-${ARCH}.tgz" \
|
||||
|| tar -xzvf "clickhouse-common-static-$LATEST_VERSION.tgz"
|
||||
sudo "clickhouse-common-static-$LATEST_VERSION/install/doinst.sh"
|
||||
@ -26,7 +24,7 @@ sudo "clickhouse-common-static-dbg-$LATEST_VERSION/install/doinst.sh"
|
||||
|
||||
tar -xzvf "clickhouse-server-$LATEST_VERSION-${ARCH}.tgz" \
|
||||
|| tar -xzvf "clickhouse-server-$LATEST_VERSION.tgz"
|
||||
sudo "clickhouse-server-$LATEST_VERSION/install/doinst.sh"
|
||||
sudo "clickhouse-server-$LATEST_VERSION/install/doinst.sh" configure
|
||||
sudo /etc/init.d/clickhouse-server start
|
||||
|
||||
tar -xzvf "clickhouse-client-$LATEST_VERSION-${ARCH}.tgz" \
|
||||
|
@ -54,7 +54,7 @@ the documentation is wrong. Go to the check report and look for `ERROR` and `WAR
|
||||
Check that the description of your pull request conforms to the template
|
||||
[PULL_REQUEST_TEMPLATE.md](https://github.com/ClickHouse/ClickHouse/blob/master/.github/PULL_REQUEST_TEMPLATE.md).
|
||||
You have to specify a changelog category for your change (e.g., Bug Fix), and
|
||||
write a user-readable message describing the change for [CHANGELOG.md](../whats-new/changelog/)
|
||||
write a user-readable message describing the change for [CHANGELOG.md](../whats-new/changelog/index.md)
|
||||
|
||||
|
||||
## Push To DockerHub
|
||||
|
@ -285,9 +285,4 @@ If you are not interested in functionality provided by third-party libraries, yo
|
||||
-DENABLE_LIBRARIES=0 -DENABLE_EMBEDDED_COMPILER=0
|
||||
```
|
||||
|
||||
Compressing the binary at the end of the build may take a while, disable the self-extraction feature via
|
||||
```
|
||||
-DENABLE_CLICKHOUSE_SELF_EXTRACTING=0
|
||||
```
|
||||
|
||||
In case of problems with any of the development options, you are on your own!
|
||||
|
@ -13,6 +13,8 @@ Syntax: `URL(URL [,Format] [,CompressionMethod])`
|
||||
|
||||
- The `Format` must be one that ClickHouse can use in `SELECT` queries and, if necessary, in `INSERTs`. For the full list of supported formats, see [Formats](../../../interfaces/formats.md#formats).
|
||||
|
||||
If this argument is not specified, ClickHouse detectes the format automatically from the suffix of the `URL` parameter. If the suffix of `URL` parameter does not match any supported formats, it fails to create table. For example, for engine expression `URL('http://localhost/test.json')`, `JSON` format is applied.
|
||||
|
||||
- `CompressionMethod` indicates that whether the HTTP body should be compressed. If the compression is enabled, the HTTP packets sent by the URL engine contain 'Content-Encoding' header to indicate which compression method is used.
|
||||
|
||||
To enable compression, please first make sure the remote HTTP endpoint indicated by the `URL` parameter supports corresponding compression algorithm.
|
||||
@ -27,6 +29,11 @@ The supported `CompressionMethod` should be one of following:
|
||||
- bz2
|
||||
- snappy
|
||||
- none
|
||||
- auto
|
||||
|
||||
If `CompressionMethod` is not specified, it defaults to `auto`. This means ClickHouse detects compression method from the suffix of `URL` parameter automatically. If the suffix matches any of compression method listed above, corresponding compression is applied or there won't be any compression enabled.
|
||||
|
||||
For example, for engine expression `URL('http://localhost/test.gzip')`, `gzip` compression method is applied, but for `URL('http://localhost/test.fr')`, no compression is enabled because the suffix `fr` does not match any compression methods above.
|
||||
|
||||
## Usage {#using-the-engine-in-the-clickhouse-server}
|
||||
|
||||
|
@ -140,8 +140,6 @@ do
|
||||
|| curl -fO "https://packages.clickhouse.com/tgz/stable/$PKG-$LATEST_VERSION.tgz"
|
||||
done
|
||||
|
||||
exit 0
|
||||
|
||||
tar -xzvf "clickhouse-common-static-$LATEST_VERSION-${ARCH}.tgz" \
|
||||
|| tar -xzvf "clickhouse-common-static-$LATEST_VERSION.tgz"
|
||||
sudo "clickhouse-common-static-$LATEST_VERSION/install/doinst.sh"
|
||||
@ -152,7 +150,7 @@ sudo "clickhouse-common-static-dbg-$LATEST_VERSION/install/doinst.sh"
|
||||
|
||||
tar -xzvf "clickhouse-server-$LATEST_VERSION-${ARCH}.tgz" \
|
||||
|| tar -xzvf "clickhouse-server-$LATEST_VERSION.tgz"
|
||||
sudo "clickhouse-server-$LATEST_VERSION/install/doinst.sh"
|
||||
sudo "clickhouse-server-$LATEST_VERSION/install/doinst.sh" configure
|
||||
sudo /etc/init.d/clickhouse-server start
|
||||
|
||||
tar -xzvf "clickhouse-client-$LATEST_VERSION-${ARCH}.tgz" \
|
||||
|
@ -31,6 +31,7 @@ ClickHouse Inc does **not** maintain the libraries listed below and hasn’t don
|
||||
- [chconn](https://github.com/vahid-sohrabloo/chconn)
|
||||
- [mailrugo-clickhouse](https://github.com/mailru/go-clickhouse)
|
||||
- [golang-clickhouse](https://github.com/leprosus/golang-clickhouse)
|
||||
- [uptrace/go-clickhouse](https://clickhouse.uptrace.dev/)
|
||||
- Swift
|
||||
- [ClickHouseNIO](https://github.com/patrick-zippenfenig/ClickHouseNIO)
|
||||
- [ClickHouseVapor ORM](https://github.com/patrick-zippenfenig/ClickHouseVapor)
|
||||
|
12
docs/en/interfaces/third-party/gui.md
vendored
12
docs/en/interfaces/third-party/gui.md
vendored
@ -196,6 +196,18 @@ Features:
|
||||
|
||||
The client is available for instant usage through github pages: https://metrico.github.io/clickhouse-mate/
|
||||
|
||||
### Uptrace {#uptrace}
|
||||
|
||||
[Uptrace](https://github.com/uptrace/uptrace) is an APM tool that provides distributed tracing and metrics powered by OpenTelemetry and ClickHouse.
|
||||
|
||||
Features:
|
||||
|
||||
- [OpenTelemetry tracing](https://uptrace.dev/opentelemetry/distributed-tracing.html), metrics, and logs.
|
||||
- Email/Slack/PagerDuty notifications using AlertManager.
|
||||
- SQL-like query language to aggregate spans.
|
||||
- Promql-like language to query metrics.
|
||||
- Pre-built metrics dashboards.
|
||||
- Multiple users/projects via YAML config.
|
||||
|
||||
## Commercial {#commercial}
|
||||
|
||||
|
@ -266,8 +266,14 @@ Result:
|
||||
└────────────────┘
|
||||
```
|
||||
|
||||
:::note
|
||||
The return type `toStartOf*` functions described below is `Date` or `DateTime`. Though these functions can take `DateTime64` as an argument, passing them a `DateTime64` that is out of the normal range (years 1900 - 2299) will give an incorrect result.
|
||||
:::Attention
|
||||
The return type of `toStartOf*`, `toLastDayOfMonth`, `toMonday` functions described below is `Date` or `DateTime`.
|
||||
Though these functions can take values of the extended types `Date32` and `DateTime64` as an argument, passing them a time outside the normal range (year 1970 to 2149 for `Date` / 2106 for `DateTime`) will produce wrong results.
|
||||
In case argument is out of normal range:
|
||||
* If the argument is smaller than 1970, the result will be calculated from the argument `1970-01-01 (00:00:00)` instead.
|
||||
* If the return type is `DateTime` and the argument is larger than `2106-02-07 08:28:15`, the result will be calculated from the argument `2106-02-07 08:28:15` instead.
|
||||
* If the return type is `Date` and the argument is larger than `2149-06-06`, the result will be calculated from the argument `2149-06-06` instead.
|
||||
* If `toLastDayOfMonth` is called with an argument greater then `2149-05-31`, the result will be calculated from the argument `2149-05-31` instead.
|
||||
:::
|
||||
|
||||
## toStartOfYear
|
||||
@ -291,20 +297,23 @@ Returns the date.
|
||||
Rounds down a date or date with time to the first day of the month.
|
||||
Returns the date.
|
||||
|
||||
:::note
|
||||
The behavior of parsing incorrect dates is implementation specific. ClickHouse may return zero date, throw an exception or do “natural” overflow.
|
||||
:::
|
||||
## toLastDayOfMonth
|
||||
|
||||
Rounds up a date or date with time to the last day of the month.
|
||||
Returns the date.
|
||||
|
||||
## toMonday
|
||||
|
||||
Rounds down a date or date with time to the nearest Monday.
|
||||
As a special case, date arguments `1970-01-01`, `1970-01-02`, `1970-01-03` and `1970-01-04` return date `1970-01-01`.
|
||||
Returns the date.
|
||||
|
||||
## toStartOfWeek(t\[,mode\])
|
||||
|
||||
Rounds down a date or date with time to the nearest Sunday or Monday by mode.
|
||||
Returns the date.
|
||||
The mode argument works exactly like the mode argument to toWeek(). For the single-argument syntax, a mode value of 0 is used.
|
||||
As a special case, date arguments `1970-01-01`, `1970-01-02`, `1970-01-03` and `1970-01-04` (and `1970-01-05` if `mode` is `1`) return date `1970-01-01`.
|
||||
The `mode` argument works exactly like the mode argument to toWeek(). For the single-argument syntax, a mode value of 0 is used.
|
||||
|
||||
## toStartOfDay
|
||||
|
||||
@ -1039,7 +1048,7 @@ Example:
|
||||
SELECT timeSlots(toDateTime('2012-01-01 12:20:00'), toUInt32(600));
|
||||
SELECT timeSlots(toDateTime('1980-12-12 21:01:02', 'UTC'), toUInt32(600), 299);
|
||||
SELECT timeSlots(toDateTime64('1980-12-12 21:01:02.1234', 4, 'UTC'), toDecimal64(600.1, 1), toDecimal64(299, 0));
|
||||
```
|
||||
```
|
||||
``` text
|
||||
┌─timeSlots(toDateTime('2012-01-01 12:20:00'), toUInt32(600))─┐
|
||||
│ ['2012-01-01 12:00:00','2012-01-01 12:30:00'] │
|
||||
|
8
docs/ru/about-us/_category_.yml
Normal file
8
docs/ru/about-us/_category_.yml
Normal file
@ -0,0 +1,8 @@
|
||||
position: 80
|
||||
label: 'About Us'
|
||||
collapsible: true
|
||||
collapsed: true
|
||||
link:
|
||||
type: generated-index
|
||||
title: About Us
|
||||
slug: /ru/about-us/
|
8
docs/ru/about-us/adopters.mdx
Normal file
8
docs/ru/about-us/adopters.mdx
Normal file
@ -0,0 +1,8 @@
|
||||
---
|
||||
sidebar_label: Adopters
|
||||
title: ClickHouse Adopters
|
||||
---
|
||||
|
||||
import Adopters from '@site/docs/en/about-us/adopters.md';
|
||||
|
||||
<Adopters />
|
8
docs/ru/about-us/support.mdx
Normal file
8
docs/ru/about-us/support.mdx
Normal file
@ -0,0 +1,8 @@
|
||||
---
|
||||
sidebar_label: Commercial Support
|
||||
title: ClickHouse Commercial Support Service
|
||||
---
|
||||
|
||||
import Support from '@site/docs/en/about-us/support.md';
|
||||
|
||||
<Support />
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/commercial/cloud
|
||||
sidebar_position: 1
|
||||
sidebar_label: "Поставщики облачных услуг ClickHouse"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/commercial/
|
||||
sidebar_position: 70
|
||||
sidebar_label: "Коммерческие услуги"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/development/architecture
|
||||
sidebar_position: 62
|
||||
sidebar_label: "Обзор архитектуры ClickHouse"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/development/browse-code
|
||||
sidebar_position: 72
|
||||
sidebar_label: "Навигация по коду ClickHouse"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/development/build-osx
|
||||
sidebar_position: 65
|
||||
sidebar_label: Сборка на Mac OS X
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/development/contrib
|
||||
sidebar_position: 71
|
||||
sidebar_label: "Используемые сторонние библиотеки"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/development/developer-instruction
|
||||
sidebar_position: 61
|
||||
sidebar_label: "Инструкция для разработчиков"
|
||||
---
|
||||
|
@ -1,3 +1,6 @@
|
||||
---
|
||||
slug: /ru/development/integrating_rust_libraries
|
||||
---
|
||||
# Интеграция библиотек на языке Rust в ClickHouse.
|
||||
|
||||
Интеграция библиотек будет описываться на основе работы проведенной для библиотеки BLAKE3.
|
||||
@ -59,4 +62,4 @@ pub unsafe extern "C" fn blake3_apply_shim(
|
||||
|
||||
В заключение, стоит отметить пару пробелм, возникших при интеграции BLAKE3:
|
||||
1) Некоторые архитектуры могут потребовать настройки компиляции в build.rs и в build_rust_lib.cmake в связи со своими особенностями.
|
||||
2) MemorySanitizer плохо понимает инициализацию памяти в Rust, поэтому для избежания ложноположительных срабатываний для BLAKE3 был создан альтернативный метод, который более явно, но при этом медленнее, инициализировал память. Он компилируется только для сборки с MemorySanitizer и в релиз не попадает. Вероятно, возможны и более красивые способы решения этой проблемы, но при интеграции BLAKE3 они не были обнаружены.
|
||||
2) MemorySanitizer плохо понимает инициализацию памяти в Rust, поэтому для избежания ложноположительных срабатываний для BLAKE3 был создан альтернативный метод, который более явно, но при этом медленнее, инициализировал память. Он компилируется только для сборки с MemorySanitizer и в релиз не попадает. Вероятно, возможны и более красивые способы решения этой проблемы, но при интеграции BLAKE3 они не были обнаружены.
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/development/style
|
||||
sidebar_position: 69
|
||||
sidebar_label: "Как писать код на C++"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/atomic
|
||||
sidebar_position: 32
|
||||
sidebar_label: Atomic
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/
|
||||
sidebar_position: 27
|
||||
sidebar_label: "Движки баз данных"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/lazy
|
||||
sidebar_position: 31
|
||||
sidebar_label: Lazy
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/materialized-mysql
|
||||
sidebar_position: 29
|
||||
sidebar_label: "[experimental] MaterializedMySQL"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/materialized-postgresql
|
||||
sidebar_position: 30
|
||||
sidebar_label: MaterializedPostgreSQL
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/mysql
|
||||
sidebar_position: 30
|
||||
sidebar_label: MySQL
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/postgresql
|
||||
sidebar_position: 35
|
||||
sidebar_label: PostgreSQL
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/replicated
|
||||
sidebar_position: 36
|
||||
sidebar_label: Replicated
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/database-engines/sqlite
|
||||
sidebar_position: 32
|
||||
sidebar_label: SQLite
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/
|
||||
sidebar_label: "Движки таблиц"
|
||||
sidebar_position: 26
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/ExternalDistributed
|
||||
sidebar_position: 12
|
||||
sidebar_label: ExternalDistributed
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/embedded-rocksdb
|
||||
sidebar_position: 9
|
||||
sidebar_label: EmbeddedRocksDB
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/hdfs
|
||||
sidebar_position: 6
|
||||
sidebar_label: HDFS
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/
|
||||
sidebar_label: "Движки таблиц для интеграции"
|
||||
sidebar_position: 30
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/jdbc
|
||||
sidebar_position: 3
|
||||
sidebar_label: JDBC
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/kafka
|
||||
sidebar_position: 8
|
||||
sidebar_label: Kafka
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/materialized-postgresql
|
||||
sidebar_position: 12
|
||||
sidebar_label: MaterializedPostgreSQL
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/mongodb
|
||||
sidebar_position: 5
|
||||
sidebar_label: MongoDB
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/mysql
|
||||
sidebar_position: 4
|
||||
sidebar_label: MySQL
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/odbc
|
||||
sidebar_position: 2
|
||||
sidebar_label: ODBC
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/postgresql
|
||||
sidebar_position: 11
|
||||
sidebar_label: PostgreSQL
|
||||
---
|
||||
|
@ -1,3 +1,6 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/rabbitmq
|
||||
---
|
||||
# RabbitMQ {#rabbitmq-engine}
|
||||
|
||||
Движок работает с [RabbitMQ](https://www.rabbitmq.com).
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/s3
|
||||
sidebar_position: 4
|
||||
sidebar_label: S3
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/integrations/sqlite
|
||||
sidebar_position: 7
|
||||
sidebar_label: SQLite
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/log-family/
|
||||
sidebar_label: "Семейство Log"
|
||||
sidebar_position: 29
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/log-family/log
|
||||
sidebar_position: 33
|
||||
sidebar_label: Log
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/log-family/stripelog
|
||||
sidebar_position: 32
|
||||
sidebar_label: StripeLog
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/log-family/tinylog
|
||||
sidebar_position: 34
|
||||
sidebar_label: TinyLog
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/aggregatingmergetree
|
||||
sidebar_position: 35
|
||||
sidebar_label: AggregatingMergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/collapsingmergetree
|
||||
sidebar_position: 36
|
||||
sidebar_label: CollapsingMergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/custom-partitioning-key
|
||||
sidebar_position: 32
|
||||
sidebar_label: "Произвольный ключ партиционирования"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/graphitemergetree
|
||||
sidebar_position: 38
|
||||
sidebar_label: GraphiteMergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/
|
||||
sidebar_label: MergeTree Family
|
||||
sidebar_position: 28
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/mergetree
|
||||
sidebar_position: 30
|
||||
sidebar_label: MergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/replacingmergetree
|
||||
sidebar_position: 33
|
||||
sidebar_label: ReplacingMergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/replication
|
||||
sidebar_position: 31
|
||||
sidebar_label: "Репликация данных"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/summingmergetree
|
||||
sidebar_position: 34
|
||||
sidebar_label: SummingMergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/mergetree-family/versionedcollapsingmergetree
|
||||
sidebar_position: 37
|
||||
sidebar_label: VersionedCollapsingMergeTree
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/buffer
|
||||
sidebar_position: 45
|
||||
sidebar_label: Buffer
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/dictionary
|
||||
sidebar_position: 35
|
||||
sidebar_label: Dictionary
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/distributed
|
||||
sidebar_position: 33
|
||||
sidebar_label: Distributed
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/external-data
|
||||
sidebar_position: 45
|
||||
sidebar_label: "Внешние данные для обработки запроса"
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/file
|
||||
sidebar_position: 37
|
||||
sidebar_label: File
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/
|
||||
sidebar_label: "Специальные движки таблиц"
|
||||
sidebar_position: 31
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/join
|
||||
sidebar_position: 40
|
||||
sidebar_label: Join
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/materializedview
|
||||
sidebar_position: 43
|
||||
sidebar_label: MaterializedView
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/memory
|
||||
sidebar_position: 44
|
||||
sidebar_label: Memory
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/merge
|
||||
sidebar_position: 36
|
||||
sidebar_label: Merge
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/null
|
||||
sidebar_position: 38
|
||||
sidebar_label: 'Null'
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/set
|
||||
sidebar_position: 39
|
||||
sidebar_label: Set
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/url
|
||||
sidebar_position: 41
|
||||
sidebar_label: URL
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/engines/table-engines/special/view
|
||||
sidebar_position: 42
|
||||
sidebar_label: View
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/columnar-database
|
||||
title: Что такое столбцовая база данных?
|
||||
sidebar_position: 101
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/dbms-naming
|
||||
title: "Что означает название ClickHouse?"
|
||||
sidebar_position: 10
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/how-do-i-contribute-code-to-clickhouse
|
||||
title: How do I contribute code to ClickHouse?
|
||||
sidebar_position: 120
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/
|
||||
title: Общие вопросы о ClickHouse
|
||||
toc_hidden_folder: true
|
||||
sidebar_position: 1
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/mapreduce
|
||||
title: Why not use something like MapReduce?
|
||||
sidebar_position: 110
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/ne-tormozit
|
||||
title: "What does “не тормозит” mean?"
|
||||
sidebar_position: 11
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/olap
|
||||
title: Что такое OLAP?
|
||||
sidebar_position: 100
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/who-is-using-clickhouse
|
||||
title: Кто пользуется ClickHouse?
|
||||
sidebar_position: 9
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/general/why-clickhouse-is-so-fast
|
||||
title: Почему ClickHouse так быстро работает?
|
||||
sidebar_position: 8
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/
|
||||
sidebar_label: F.A.Q.
|
||||
sidebar_position: 76
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/integration/file-export
|
||||
title: Как экспортировать данные из ClickHouse в файл?
|
||||
sidebar_position: 10
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/integration/
|
||||
title: Интеграция ClickHouse с другими системами
|
||||
toc_hidden_folder: true
|
||||
sidebar_position: 4
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/integration/json-import
|
||||
title: Как импортировать JSON в ClickHouse?
|
||||
sidebar_position: 11
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/integration/oracle-odbc
|
||||
title: Что делать, если у меня проблема с кодировками при использовании Oracle через ODBC?
|
||||
sidebar_position: 20
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/operations/delete-old-data
|
||||
title: Возможно ли удалить старые записи из таблицы ClickHouse?
|
||||
sidebar_position: 20
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/operations/
|
||||
title: Вопросы о производительности серверов и кластеров ClickHouse
|
||||
toc_hidden_folder: true
|
||||
sidebar_position: 3
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/operations/multi-region-replication
|
||||
title: Does ClickHouse support multi-region replication?
|
||||
sidebar_position: 30
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/operations/production
|
||||
title: Какую версию ClickHouse использовать?
|
||||
sidebar_position: 10
|
||||
---
|
||||
@ -66,4 +67,4 @@ sidebar_position: 10
|
||||
Часто компании, которые изначально ориентировались на релизы `lts`, позднее переходят на `stable`, поскольку хотят быстрее получать доступ к новым возможностям.
|
||||
|
||||
:::danger "Важно"
|
||||
Мы всегда стремимся поддерживать совместимость релизов, но иногда это правило нарушается, и какие-то отдельные возможности в новых релизах становятся недоступны. Перед обновлением ClickHouse обязательно изучите [журнал изменений](../../whats-new/changelog/index.md), чтобы убедиться, что в нем нет объявлений о нарушении обратной совместимости.
|
||||
Мы всегда стремимся поддерживать совместимость релизов, но иногда это правило нарушается, и какие-то отдельные возможности в новых релизах становятся недоступны. Перед обновлением ClickHouse обязательно изучите [журнал изменений](../../whats-new/changelog/index.mdx), чтобы убедиться, что в нем нет объявлений о нарушении обратной совместимости.
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/use-cases/
|
||||
title: Вопросы о применении ClickHouse
|
||||
toc_hidden_folder: true
|
||||
sidebar_position: 2
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/use-cases/key-value
|
||||
title: Можно ли использовать ClickHouse для хранения данных вида "ключ-значение"?
|
||||
sidebar_position: 101
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/faq/use-cases/time-series
|
||||
title: Можно ли использовать ClickHouse как базу данных временных рядов?
|
||||
sidebar_position: 101
|
||||
---
|
||||
|
@ -1,4 +1,5 @@
|
||||
---
|
||||
slug: /ru/getting-started/example-datasets/amplab-benchmark
|
||||
sidebar_position: 19
|
||||
sidebar_label: AMPLab Big Data Benchmark
|
||||
---
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user