mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-10 09:32:06 +00:00
Merge pull request #26362 from ClickHouse/fix-integration-test-3
Fix flaky integration test about "replicated max parallel fetches".
This commit is contained in:
commit
c03dd8d923
@ -224,7 +224,7 @@
|
||||
M(PerfLocalMemoryReferences, "Local NUMA node memory reads") \
|
||||
M(PerfLocalMemoryMisses, "Local NUMA node memory read misses") \
|
||||
\
|
||||
M(CreatedHTTPConnections, "Total amount of created HTTP connections (closed or opened).") \
|
||||
M(CreatedHTTPConnections, "Total amount of created HTTP connections (counter increase every time connection is created).") \
|
||||
\
|
||||
M(CannotWriteToWriteBufferDiscard, "Number of stack traces dropped by query profiler or signal handler because pipe is full or cannot write to pipe.") \
|
||||
M(QueryProfilerSignalOverruns, "Number of times we drop processing of a signal due to overrun plus the number of signals that OS has not delivered due to overrun.") \
|
||||
|
@ -676,14 +676,16 @@ void StorageMerge::convertingSourceStream(
|
||||
auto convert_actions_dag = ActionsDAG::makeConvertingActions(pipe.getHeader().getColumnsWithTypeAndName(),
|
||||
header.getColumnsWithTypeAndName(),
|
||||
ActionsDAG::MatchColumnsMode::Name);
|
||||
auto actions = std::make_shared<ExpressionActions>(convert_actions_dag, ExpressionActionsSettings::fromContext(local_context, CompileExpressions::yes));
|
||||
auto actions = std::make_shared<ExpressionActions>(
|
||||
convert_actions_dag,
|
||||
ExpressionActionsSettings::fromContext(local_context, CompileExpressions::yes));
|
||||
|
||||
pipe.addSimpleTransform([&](const Block & stream_header)
|
||||
{
|
||||
return std::make_shared<ExpressionTransform>(stream_header, actions);
|
||||
});
|
||||
}
|
||||
|
||||
|
||||
auto where_expression = query->as<ASTSelectQuery>()->where();
|
||||
|
||||
if (!where_expression)
|
||||
|
@ -11,7 +11,7 @@ def _fill_nodes(nodes, shard, connections_count):
|
||||
node.query(
|
||||
'''
|
||||
CREATE DATABASE test;
|
||||
|
||||
|
||||
CREATE TABLE test_table(date Date, id UInt32, dummy UInt32)
|
||||
ENGINE = ReplicatedMergeTree('/clickhouse/tables/test{shard}/replicated', '{replica}')
|
||||
PARTITION BY date
|
||||
@ -114,5 +114,5 @@ def test_multiple_endpoint_connections_count(start_big_cluster):
|
||||
assert_eq_with_retry(node4, "select count() from test_table", "100")
|
||||
assert_eq_with_retry(node5, "select count() from test_table", "100")
|
||||
|
||||
# two per each host
|
||||
assert node5.query("SELECT value FROM system.events where event='CreatedHTTPConnections'") == '4\n'
|
||||
# Two per each host or sometimes less, if fetches are not performed in parallel. But not more.
|
||||
assert node5.query("SELECT value FROM system.events where event='CreatedHTTPConnections'") <= '4\n'
|
||||
|
Loading…
Reference in New Issue
Block a user