* save all merge params to zookeeper
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
* calculate hash for graphite merge params
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
* add graphite params hash to zookeeper + fix tests
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
* install new graphite for testing
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
* fix backward incompatibility
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
* minor fix test
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
* Update src/Storages/MergeTree/ReplicatedMergeTreeTableMetadata.cpp
Co-authored-by: Alexander Tokmakov <tavplubix@gmail.com>
* remove peekString and add more comments
- peekString doesn't always work even for ReadBufferFromString
- more comment re. backward compatibility
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
---------
Signed-off-by: Duc Canh Le <duccanh.le@ahrefs.com>
Co-authored-by: Alexander Tokmakov <tavplubix@gmail.com>
Previously tests assumed there is `clickhouse-client` available in test machine
`$PATH`, which is not always true and will cause a test failure when ClickHouse
is not installed system-wide. This commit fixes that by utilizing
`CLICKHOUSE_CLIENT_BINARY` more which is substitued from `CLICKHOUSE_BINARY`
which is set as `-b` argument in `clickhouse-test`.
Previously `CLICKHOUSE_BENCHMARK` was set simply as
`${CLICKHOUSE_BINARY}-benchmark` which is not applicable when testing a single
binary, e.g. downloaded via `curl https://clickhouse.com | sh`. This commit
adds support for `CLICKHOUSE_BENCHMARK_BINARY` (not exported from
`shell_config.sh`) that is set to `"$CLICKHOUSE_BINARY benchmark"` if
`CLICKHOUSE_BINARY` is set. This is the same behaviour that is already used for
`CLICKHOUSE_CLIENT_BINARY`.
Randomly modify structural characters of a valid JSON ('{', '}', '[', ']',
':', '"', ',') to generate output that cannot be parsed as JSON.
Follow-up to https://github.com/ClickHouse/ClickHouse/pull/56490
It is not safe to use statistics because of how KafkaEngine works - it
pre-creates consumers, and this leads to the situation when this
statistics entries generated (RD_KAFKA_OP_STATS), but never consumed.
Which creates a live memory leak for a server with Kafka tables, but
without materialized view attached to it (and no SELECT).
Another problem is that this makes shutdown very slow, because of how
pending queue entries are handled in librdkafka, it uses
TAILQ_INSERT_SORTED, which is sorted insert into linked list, which
works incredibly slow (likely you will never wait till it ends and kill
the server)
For instance in my production setup the server was running for ~67 days
with such table, and it got 1'942'233 `TAILQ_INSERT_SORTED` entries
(which perfectly matches by the way - `67*86400/3` = 1'929'600), and it
moved only 289'806 entries for a few hours, though I'm not sure how much
time the process was in the running state, since most of the time it was
with debugger attached.
So for now let's disable it, to make this patch easy for backporting,
and I will think about long term fix - do not pre-create consumers in
Kafka engine.
Signed-off-by: Azat Khuzhin <a.khuzhin@semrush.com>