mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-26 17:41:59 +00:00
Fix StorageBuffer/
This commit is contained in:
parent
9eb97e6db2
commit
00546498b6
@ -155,15 +155,8 @@ QueryProcessingStage::Enum StorageBuffer::getQueryProcessingStage(const Context
|
||||
return QueryProcessingStage::FetchColumns;
|
||||
}
|
||||
|
||||
Pipes StorageBuffer::readWithProcessors(
|
||||
const Names & column_names,
|
||||
const SelectQueryInfo & query_info,
|
||||
const Context & context,
|
||||
QueryProcessingStage::Enum processed_stage,
|
||||
size_t max_block_size,
|
||||
unsigned num_streams)
|
||||
{
|
||||
auto read_as_pipes = [](const StoragePtr & storage,
|
||||
static Pipes readAsPipes(
|
||||
const StoragePtr & storage,
|
||||
const Names & column_names,
|
||||
const SelectQueryInfo & query_info,
|
||||
const Context & context,
|
||||
@ -183,6 +176,14 @@ Pipes StorageBuffer::readWithProcessors(
|
||||
return pipes;
|
||||
};
|
||||
|
||||
Pipes StorageBuffer::readWithProcessors(
|
||||
const Names & column_names,
|
||||
const SelectQueryInfo & query_info,
|
||||
const Context & context,
|
||||
QueryProcessingStage::Enum processed_stage,
|
||||
size_t max_block_size,
|
||||
unsigned num_streams)
|
||||
{
|
||||
Pipes pipes_from_dst;
|
||||
|
||||
if (!no_destination)
|
||||
@ -206,7 +207,7 @@ Pipes StorageBuffer::readWithProcessors(
|
||||
query_info.input_sorting_info = query_info.order_by_optimizer->getInputOrder(destination);
|
||||
|
||||
/// The destination table has the same structure of the requested columns and we can simply read blocks from there.
|
||||
pipes_from_dst = read_as_pipes(destination, column_names, query_info, context, processed_stage, max_block_size, num_streams));
|
||||
pipes_from_dst = readAsPipes(destination, column_names, query_info, context, processed_stage, max_block_size, num_streams));
|
||||
}
|
||||
else
|
||||
{
|
||||
@ -241,7 +242,7 @@ Pipes StorageBuffer::readWithProcessors(
|
||||
}
|
||||
else
|
||||
{
|
||||
pipes_from_dst = read_as_pipes(destination, columns_intersection, query_info, context, processed_stage, max_block_size, num_streams);
|
||||
pipes_from_dst = readAsPipes(destination, columns_intersection, query_info, context, processed_stage, max_block_size, num_streams);
|
||||
for (auto & pipe : pipes_from_dst)
|
||||
{
|
||||
pipe.addSimpleTransform(std::make_shared<AddingMissedTransform>(
|
||||
|
Loading…
Reference in New Issue
Block a user