ClickHouse/dbms/src/Interpreters/InterpreterInsertQuery.cpp

214 lines
8.4 KiB
C++
Raw Normal View History

#include <Interpreters/InterpreterInsertQuery.h>
2017-07-13 20:58:19 +00:00
#include <DataStreams/AddingDefaultBlockOutputStream.h>
#include <DataStreams/AddingDefaultsBlockInputStream.h>
2019-05-19 05:27:00 +00:00
#include <DataStreams/CheckConstraintsBlockOutputStream.h>
#include <DataStreams/ConvertingBlockInputStream.h>
#include <DataStreams/CountingBlockOutputStream.h>
#include <DataStreams/InputStreamFromASTInsertQuery.h>
#include <DataStreams/NullAndDoCopyBlockInputStream.h>
#include <DataStreams/OwningBlockInputStream.h>
#include <DataStreams/PushingToViewsBlockOutputStream.h>
#include <DataStreams/SquashingBlockOutputStream.h>
#include <DataStreams/copyData.h>
#include <IO/ConcatReadBuffer.h>
#include <IO/ReadBufferFromMemory.h>
#include <Interpreters/InterpreterSelectWithUnionQuery.h>
#include <Access/AccessFlags.h>
#include <Parsers/ASTFunction.h>
#include <Parsers/ASTInsertQuery.h>
2018-02-25 06:34:20 +00:00
#include <Parsers/ASTSelectWithUnionQuery.h>
#include <Storages/Kafka/StorageKafka.h>
#include <TableFunctions/TableFunctionFactory.h>
#include <Common/checkStackSize.h>
2011-10-30 11:30:52 +00:00
namespace DB
{
namespace ErrorCodes
{
extern const int NO_SUCH_COLUMN_IN_TABLE;
extern const int ILLEGAL_COLUMN;
extern const int DUPLICATE_COLUMN;
}
2011-10-30 11:30:52 +00:00
2018-01-12 13:03:19 +00:00
InterpreterInsertQuery::InterpreterInsertQuery(
const ASTPtr & query_ptr_, const Context & context_, bool allow_materialized_, bool no_squash_, bool no_destination_)
: query_ptr(query_ptr_)
, context(context_)
, allow_materialized(allow_materialized_)
, no_squash(no_squash_)
, no_destination(no_destination_)
2011-10-30 11:30:52 +00:00
{
2019-08-10 17:51:47 +00:00
checkStackSize();
2011-10-30 11:30:52 +00:00
}
2020-01-24 16:20:36 +00:00
StoragePtr InterpreterInsertQuery::getTable(ASTInsertQuery & query)
2011-10-30 11:30:52 +00:00
{
if (query.table_function)
{
2019-03-11 13:22:51 +00:00
const auto * table_function = query.table_function->as<ASTFunction>();
const auto & factory = TableFunctionFactory::instance();
2019-07-18 18:29:49 +00:00
TableFunctionPtr table_function_ptr = factory.get(table_function->name, context);
return table_function_ptr->execute(query.table_function, context, table_function_ptr->getName());
}
2020-03-02 20:23:58 +00:00
query.table_id = context.resolveStorageID(query.table_id);
return DatabaseCatalog::instance().getTable(query.table_id);
2011-10-30 11:30:52 +00:00
}
Block InterpreterInsertQuery::getSampleBlock(const ASTInsertQuery & query, const StoragePtr & table)
{
2018-09-20 12:59:33 +00:00
Block table_sample_non_materialized = table->getSampleBlockNonMaterialized();
2017-04-02 17:37:49 +00:00
/// If the query does not include information about columns
if (!query.columns)
2018-09-20 12:59:33 +00:00
{
if (no_destination)
return table->getSampleBlockWithVirtuals();
2018-09-20 12:59:33 +00:00
else
return table_sample_non_materialized;
}
2013-10-25 14:56:47 +00:00
2018-09-20 12:59:33 +00:00
Block table_sample = table->getSampleBlock();
2017-04-02 17:37:49 +00:00
/// Form the block based on the column names from the query
Block res;
for (const auto & identifier : query.columns->children)
{
std::string current_name = identifier->getColumnName();
2013-10-25 14:56:47 +00:00
2017-04-02 17:37:49 +00:00
/// The table does not have a column with that name
if (!table_sample.has(current_name))
2020-03-02 20:23:58 +00:00
throw Exception("No such column " + current_name + " in table " + query.table_id.getNameForLogs(), ErrorCodes::NO_SUCH_COLUMN_IN_TABLE);
2012-03-19 12:57:56 +00:00
if (!allow_materialized && !table_sample_non_materialized.has(current_name))
throw Exception("Cannot insert column " + current_name + ", because it is MATERIALIZED column.", ErrorCodes::ILLEGAL_COLUMN);
if (res.has(current_name))
throw Exception("Column " + current_name + " specified more than once", ErrorCodes::DUPLICATE_COLUMN);
res.insert(ColumnWithTypeAndName(table_sample.getByName(current_name).type, current_name));
}
return res;
2013-10-25 14:56:47 +00:00
}
2012-03-19 12:57:56 +00:00
2011-10-30 11:30:52 +00:00
BlockIO InterpreterInsertQuery::execute()
2012-03-11 08:52:56 +00:00
{
const Settings & settings = context.getSettingsRef();
2020-01-24 16:20:36 +00:00
auto & query = query_ptr->as<ASTInsertQuery &>();
2019-12-17 08:18:17 +00:00
BlockIO res;
2012-03-11 08:52:56 +00:00
2020-01-24 16:20:36 +00:00
StoragePtr table = getTable(query);
auto table_lock = table->lockStructureForShare(true, context.getInitialQueryId());
2020-01-24 16:20:36 +00:00
auto query_sample_block = getSampleBlock(query, table);
2020-03-02 20:23:58 +00:00
if (!query.table_function)
2020-03-13 15:41:36 +00:00
context.checkAccess(AccessType::INSERT, query.table_id, query_sample_block.getNames());
2020-01-24 16:20:36 +00:00
2019-12-12 10:49:15 +00:00
BlockInputStreams in_streams;
size_t out_streams_size = 1;
2019-12-19 16:09:05 +00:00
if (query.select)
{
2019-12-12 10:49:15 +00:00
/// Passing 1 as subquery_depth will disable limiting size of intermediate result.
InterpreterSelectWithUnionQuery interpreter_select{query.select, context, SelectQueryOptions(QueryProcessingStage::Complete, 1)};
if (table->supportsParallelInsert() && settings.max_insert_threads > 1)
2019-12-12 10:49:15 +00:00
{
in_streams = interpreter_select.executeWithMultipleStreams(res.pipeline);
out_streams_size = std::min(size_t(settings.max_insert_threads), in_streams.size());
}
else
2020-01-24 09:09:47 +00:00
{
res = interpreter_select.execute();
in_streams.emplace_back(res.in);
res.in = nullptr;
res.out = nullptr;
}
}
2019-12-12 10:49:15 +00:00
BlockOutputStreams out_streams;
2012-03-11 08:52:56 +00:00
2019-12-13 03:13:11 +00:00
for (size_t i = 0; i < out_streams_size; i++)
2019-12-12 10:49:15 +00:00
{
/// We create a pipeline of several streams, into which we will write data.
BlockOutputStreamPtr out;
/// NOTE: we explicitly ignore bound materialized views when inserting into Kafka Storage.
/// Otherwise we'll get duplicates when MV reads same rows again from Kafka.
if (table->noPushingToViews() && !no_destination)
out = table->write(query_ptr, context);
else
2020-01-23 15:44:26 +00:00
out = std::make_shared<PushingToViewsBlockOutputStream>(table, context, query_ptr, no_destination);
2018-09-20 11:40:04 +00:00
2019-12-12 10:49:15 +00:00
/// Do not squash blocks if it is a sync INSERT into Distributed, since it lead to double bufferization on client and server side.
/// Client-side bufferization might cause excessive timeouts (especially in case of big blocks).
if (!(context.getSettingsRef().insert_distributed_sync && table->isRemote()) && !no_squash)
{
out = std::make_shared<SquashingBlockOutputStream>(
out, out->getHeader(), context.getSettingsRef().min_insert_block_size_rows, context.getSettingsRef().min_insert_block_size_bytes);
}
2019-05-19 05:27:00 +00:00
2019-12-12 10:49:15 +00:00
/// Actually we don't know structure of input blocks from query/table,
/// because some clients break insertion protocol (columns != header)
out = std::make_shared<AddingDefaultBlockOutputStream>(
out, query_sample_block, out->getHeader(), table->getColumns().getDefaults(), context);
2019-12-12 10:49:15 +00:00
if (const auto & constraints = table->getConstraints(); !constraints.empty())
2020-03-02 20:23:58 +00:00
out = std::make_shared<CheckConstraintsBlockOutputStream>(query.table_id,
2019-12-12 10:49:15 +00:00
out, query_sample_block, table->getConstraints(), context);
auto out_wrapper = std::make_shared<CountingBlockOutputStream>(out);
out_wrapper->setProcessListElement(context.getProcessListElement());
out = std::move(out_wrapper);
out_streams.emplace_back(std::move(out));
}
2017-04-02 17:37:49 +00:00
/// What type of query: INSERT or INSERT SELECT?
if (query.select)
{
2019-12-12 10:49:15 +00:00
for (auto & in_stream : in_streams)
{
in_stream = std::make_shared<ConvertingBlockInputStream>(
context, in_stream, out_streams.at(0)->getHeader(), ConvertingBlockInputStream::MatchColumnsMode::Position);
}
Block in_header = in_streams.at(0)->getHeader();
2019-12-13 03:13:11 +00:00
if (in_streams.size() > 1)
2019-12-12 10:49:15 +00:00
{
for (size_t i = 1; i < in_streams.size(); ++i)
assertBlocksHaveEqualStructure(in_streams[i]->getHeader(), in_header, "INSERT SELECT");
}
2019-12-12 10:49:15 +00:00
res.in = std::make_shared<NullAndDoCopyBlockInputStream>(in_streams, out_streams);
if (!allow_materialized)
{
for (const auto & column : table->getColumns())
if (column.default_desc.kind == ColumnDefaultKind::Materialized && in_header.has(column.name))
throw Exception("Cannot insert column " + column.name + ", because it is MATERIALIZED column.", ErrorCodes::ILLEGAL_COLUMN);
}
}
2019-02-08 13:24:24 +00:00
else if (query.data && !query.has_tail) /// can execute without additional data
{
2019-12-19 16:09:05 +00:00
// res.out = std::move(out_streams.at(0));
res.in = std::make_shared<InputStreamFromASTInsertQuery>(query_ptr, nullptr, query_sample_block, context, nullptr);
2019-12-19 16:09:05 +00:00
res.in = std::make_shared<NullAndDoCopyBlockInputStream>(res.in, out_streams.at(0));
}
2019-12-16 13:52:32 +00:00
else
2019-12-12 10:49:15 +00:00
res.out = std::move(out_streams.at(0));
2019-12-16 13:52:32 +00:00
res.pipeline.addStorageHolder(table);
return res;
2012-03-11 08:52:56 +00:00
}
2020-03-02 20:23:58 +00:00
StorageID InterpreterInsertQuery::getDatabaseTable() const
2018-07-16 14:52:02 +00:00
{
2020-03-02 20:23:58 +00:00
return query_ptr->as<ASTInsertQuery &>().table_id;
2018-07-16 14:52:02 +00:00
}
2011-10-30 11:30:52 +00:00
}