#include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include namespace ProfileEvents { extern const Event InsertQueriesWithSubqueries; extern const Event QueriesWithSubqueries; } namespace DB { namespace ErrorCodes { extern const int NOT_IMPLEMENTED; extern const int NO_SUCH_COLUMN_IN_TABLE; extern const int ILLEGAL_COLUMN; extern const int DUPLICATE_COLUMN; } InterpreterInsertQuery::InterpreterInsertQuery( const ASTPtr & query_ptr_, ContextPtr context_, bool allow_materialized_, bool no_squash_, bool no_destination_, bool async_insert_) : WithContext(context_) , query_ptr(query_ptr_) , allow_materialized(allow_materialized_) , no_squash(no_squash_) , no_destination(no_destination_) , async_insert(async_insert_) { checkStackSize(); if (auto quota = getContext()->getQuota()) quota->checkExceeded(QuotaType::WRITTEN_BYTES); } StoragePtr InterpreterInsertQuery::getTable(ASTInsertQuery & query) { auto current_context = getContext(); if (query.table_function) { const auto & factory = TableFunctionFactory::instance(); TableFunctionPtr table_function_ptr = factory.get(query.table_function, current_context); /// If table function needs structure hint from select query /// we can create a temporary pipeline and get the header. if (query.select && table_function_ptr->needStructureHint()) { Block header_block; auto select_query_options = SelectQueryOptions(QueryProcessingStage::Complete, 1); if (current_context->getSettingsRef().allow_experimental_analyzer) { InterpreterSelectQueryAnalyzer interpreter_select(query.select, current_context, select_query_options); header_block = interpreter_select.getSampleBlock(); } else { InterpreterSelectWithUnionQuery interpreter_select{ query.select, current_context, select_query_options}; auto tmp_pipeline = interpreter_select.buildQueryPipeline(); header_block = tmp_pipeline.getHeader(); } ColumnsDescription structure_hint{header_block.getNamesAndTypesList()}; table_function_ptr->setStructureHint(structure_hint); } return table_function_ptr->execute(query.table_function, current_context, table_function_ptr->getName(), /* cached_columns */ {}, /* use_global_context */ false, /* is_insert_query */true); } if (query.table_id) { query.table_id = current_context->resolveStorageID(query.table_id); } else { /// Insert query parser does not fill table_id because table and /// database can be parameters and be filled after parsing. StorageID local_table_id(query.getDatabase(), query.getTable()); query.table_id = current_context->resolveStorageID(local_table_id); } return DatabaseCatalog::instance().getTable(query.table_id, current_context); } Block InterpreterInsertQuery::getSampleBlock( const ASTInsertQuery & query, const StoragePtr & table, const StorageMetadataPtr & metadata_snapshot) const { /// If the query does not include information about columns if (!query.columns) { if (auto * window_view = dynamic_cast(table.get())) return window_view->getInputHeader(); else if (no_destination) return metadata_snapshot->getSampleBlockWithVirtuals(table->getVirtuals()); else return metadata_snapshot->getSampleBlockNonMaterialized(); } /// Form the block based on the column names from the query const auto columns_ast = processColumnTransformers(getContext()->getCurrentDatabase(), table, metadata_snapshot, query.columns); Names names; names.reserve(columns_ast->children.size()); for (const auto & identifier : columns_ast->children) { std::string current_name = identifier->getColumnName(); names.emplace_back(std::move(current_name)); } return getSampleBlock(names, table, metadata_snapshot); } std::optional InterpreterInsertQuery::getInsertColumnNames() const { auto const * insert_query = query_ptr->as(); if (!insert_query || !insert_query->columns) return std::nullopt; auto table = DatabaseCatalog::instance().getTable(getDatabaseTable(), getContext()); const auto columns_ast = processColumnTransformers(getContext()->getCurrentDatabase(), table, table->getInMemoryMetadataPtr(), insert_query->columns); Names names; names.reserve(columns_ast->children.size()); for (const auto & identifier : columns_ast->children) { std::string current_name = identifier->getColumnName(); names.emplace_back(std::move(current_name)); } return names; } Block InterpreterInsertQuery::getSampleBlock( const Names & names, const StoragePtr & table, const StorageMetadataPtr & metadata_snapshot) const { Block table_sample_physical = metadata_snapshot->getSampleBlock(); Block table_sample_insertable = metadata_snapshot->getSampleBlockInsertable(); Block res; for (const auto & current_name : names) { if (res.has(current_name)) throw Exception(ErrorCodes::DUPLICATE_COLUMN, "Column {} specified more than once", current_name); /// Column is not ordinary or ephemeral if (!table_sample_insertable.has(current_name)) { /// Column is materialized if (table_sample_physical.has(current_name)) { if (!allow_materialized) throw Exception(ErrorCodes::ILLEGAL_COLUMN, "Cannot insert column {}, because it is MATERIALIZED column.", current_name); res.insert(ColumnWithTypeAndName(table_sample_physical.getByName(current_name).type, current_name)); } else /// The table does not have a column with that name throw Exception(ErrorCodes::NO_SUCH_COLUMN_IN_TABLE, "No such column {} in table {}", current_name, table->getStorageID().getNameForLogs()); } else res.insert(ColumnWithTypeAndName(table_sample_insertable.getByName(current_name).type, current_name)); } return res; } static bool hasAggregateFunctions(const IAST * ast) { if (const auto * func = typeid_cast(ast)) if (AggregateUtils::isAggregateFunction(*func)) return true; for (const auto & child : ast->children) if (hasAggregateFunctions(child.get())) return true; return false; } /** A query that just reads all data without any complex computations or filetering. * If we just pipe the result to INSERT, we don't have to use too many threads for read. */ static bool isTrivialSelect(const ASTPtr & select) { if (auto * select_query = select->as()) { const auto & tables = select_query->tables(); if (!tables) return false; const auto & tables_in_select_query = tables->as(); if (tables_in_select_query.children.size() != 1) return false; const auto & child = tables_in_select_query.children.front(); const auto & table_element = child->as(); const auto & table_expr = table_element.table_expression->as(); if (table_expr.subquery) return false; /// Note: how to write it in more generic way? return (!select_query->distinct && !select_query->limit_with_ties && !select_query->prewhere() && !select_query->where() && !select_query->groupBy() && !select_query->having() && !select_query->orderBy() && !select_query->limitBy() && !hasAggregateFunctions(select_query)); } /// This query is ASTSelectWithUnionQuery subquery return false; } Chain InterpreterInsertQuery::buildChain( const StoragePtr & table, const StorageMetadataPtr & metadata_snapshot, const Names & columns, ThreadStatusesHolderPtr thread_status_holder, std::atomic_uint64_t * elapsed_counter_ms) { ProfileEvents::increment(ProfileEvents::InsertQueriesWithSubqueries); ProfileEvents::increment(ProfileEvents::QueriesWithSubqueries); ThreadGroupPtr running_group; if (current_thread) running_group = current_thread->getThreadGroup(); if (!running_group) running_group = std::make_shared(getContext()); auto sample = getSampleBlock(columns, table, metadata_snapshot); Chain sink = buildSink(table, metadata_snapshot, thread_status_holder, running_group, elapsed_counter_ms); Chain chain = buildPreSinkChain(sink.getInputHeader(), table, metadata_snapshot, sample, thread_status_holder); chain.appendChain(std::move(sink)); return chain; } Chain InterpreterInsertQuery::buildSink( const StoragePtr & table, const StorageMetadataPtr & metadata_snapshot, ThreadStatusesHolderPtr thread_status_holder, ThreadGroupPtr running_group, std::atomic_uint64_t * elapsed_counter_ms) { ThreadStatus * thread_status = current_thread; if (!thread_status_holder) thread_status = nullptr; auto context_ptr = getContext(); Chain out; /// Keep a reference to the context to make sure it stays alive until the chain is executed and destroyed out.addInterpreterContext(context_ptr); /// NOTE: we explicitly ignore bound materialized views when inserting into Kafka Storage. /// Otherwise we'll get duplicates when MV reads same rows again from Kafka. if (table->noPushingToViews() && !no_destination) { auto sink = table->write(query_ptr, metadata_snapshot, context_ptr, async_insert); sink->setRuntimeData(thread_status, elapsed_counter_ms); out.addSource(std::move(sink)); } else { out = buildPushingToViewsChain(table, metadata_snapshot, context_ptr, query_ptr, no_destination, thread_status_holder, running_group, elapsed_counter_ms, async_insert); } return out; } Chain InterpreterInsertQuery::buildPreSinkChain( const Block & subsequent_header, const StoragePtr & table, const StorageMetadataPtr & metadata_snapshot, const Block & query_sample_block, ThreadStatusesHolderPtr thread_status_holder) { ThreadStatus * thread_status = current_thread; if (!thread_status_holder) thread_status = nullptr; auto context_ptr = getContext(); const ASTInsertQuery * query = nullptr; if (query_ptr) query = query_ptr->as(); const Settings & settings = context_ptr->getSettingsRef(); bool null_as_default = query && query->select && context_ptr->getSettingsRef().insert_null_as_default; /// We create a pipeline of several streams, into which we will write data. Chain out; auto input_header = [&]() -> const Block & { return out.empty() ? subsequent_header : out.getInputHeader(); }; /// Note that we wrap transforms one on top of another, so we write them in reverse of data processing order. /// Checking constraints. It must be done after calculation of all defaults, so we can check them on calculated columns. if (const auto & constraints = metadata_snapshot->getConstraints(); !constraints.empty()) out.addSource(std::make_shared( table->getStorageID(), input_header(), metadata_snapshot->getConstraints(), context_ptr)); auto adding_missing_defaults_dag = addMissingDefaults( query_sample_block, input_header().getNamesAndTypesList(), metadata_snapshot->getColumns(), context_ptr, null_as_default); auto adding_missing_defaults_actions = std::make_shared(adding_missing_defaults_dag); /// Actually we don't know structure of input blocks from query/table, /// because some clients break insertion protocol (columns != header) out.addSource(std::make_shared(query_sample_block, adding_missing_defaults_actions)); /// It's important to squash blocks as early as possible (before other transforms), /// because other transforms may work inefficient if block size is small. /// Do not squash blocks if it is a sync INSERT into Distributed, since it lead to double bufferization on client and server side. /// Client-side bufferization might cause excessive timeouts (especially in case of big blocks). if (!(settings.distributed_foreground_insert && table->isRemote()) && !async_insert && !no_squash && !(query && query->watch)) { bool table_prefers_large_blocks = table->prefersLargeBlocks(); out.addSource(std::make_shared( input_header(), table_prefers_large_blocks ? settings.min_insert_block_size_rows : settings.max_block_size, table_prefers_large_blocks ? settings.min_insert_block_size_bytes : 0ULL)); } auto counting = std::make_shared(input_header(), thread_status, getContext()->getQuota()); counting->setProcessListElement(context_ptr->getProcessListElement()); counting->setProgressCallback(context_ptr->getProgressCallback()); out.addSource(std::move(counting)); return out; } BlockIO InterpreterInsertQuery::execute() { const Settings & settings = getContext()->getSettingsRef(); auto & query = query_ptr->as(); QueryPipelineBuilder pipeline; std::optional distributed_pipeline; QueryPlanResourceHolder resources; StoragePtr table = getTable(query); checkStorageSupportsTransactionsIfNeeded(table, getContext()); StoragePtr inner_table; if (const auto * mv = dynamic_cast(table.get())) inner_table = mv->getTargetTable(); if (query.partition_by && !table->supportsPartitionBy()) throw Exception(ErrorCodes::NOT_IMPLEMENTED, "PARTITION BY clause is not supported by storage"); auto table_lock = table->lockForShare(getContext()->getInitialQueryId(), settings.lock_acquire_timeout); auto metadata_snapshot = table->getInMemoryMetadataPtr(); auto query_sample_block = getSampleBlock(query, table, metadata_snapshot); /// For table functions we check access while executing /// getTable() -> ITableFunction::execute(). if (!query.table_function) getContext()->checkAccess(AccessType::INSERT, query.table_id, query_sample_block.getNames()); if (query.select && settings.parallel_distributed_insert_select) // Distributed INSERT SELECT distributed_pipeline = table->distributedWrite(query, getContext()); std::vector presink_chains; std::vector sink_chains; if (!distributed_pipeline || query.watch) { /// Number of streams works like this: /// * For the SELECT, use `max_threads`, or `max_insert_threads`, or whatever /// InterpreterSelectQuery ends up with. /// * Use `max_insert_threads` streams for various insert-preparation steps, e.g. /// materializing and squashing (too slow to do in one thread). That's `presink_chains`. /// * If the table supports parallel inserts, use the same streams for writing to IStorage. /// Otherwise ResizeProcessor them down to 1 stream. /// * If it's not an INSERT SELECT, forget all that and use one stream. size_t pre_streams_size = 1; size_t sink_streams_size = 1; if (query.select) { bool is_trivial_insert_select = false; if (settings.optimize_trivial_insert_select) { const auto & select_query = query.select->as(); const auto & selects = select_query.list_of_selects->children; const auto & union_modes = select_query.list_of_modes; /// ASTSelectWithUnionQuery is not normalized now, so it may pass some queries which can be Trivial select queries const auto mode_is_all = [](const auto & mode) { return mode == SelectUnionMode::UNION_ALL; }; is_trivial_insert_select = std::all_of(union_modes.begin(), union_modes.end(), std::move(mode_is_all)) && std::all_of(selects.begin(), selects.end(), isTrivialSelect); } if (is_trivial_insert_select) { /** When doing trivial INSERT INTO ... SELECT ... FROM table, * don't need to process SELECT with more than max_insert_threads * and it's reasonable to set block size for SELECT to the desired block size for INSERT * to avoid unnecessary squashing. */ Settings new_settings = getContext()->getSettings(); new_settings.max_threads = std::max(1, settings.max_insert_threads); if (table->prefersLargeBlocks()) { if (settings.min_insert_block_size_rows) new_settings.max_block_size = settings.min_insert_block_size_rows; if (settings.min_insert_block_size_bytes) new_settings.preferred_block_size_bytes = settings.min_insert_block_size_bytes; } auto new_context = Context::createCopy(context); new_context->setSettings(new_settings); new_context->setInsertionTable(getContext()->getInsertionTable(), getContext()->getInsertionTableColumnNames()); auto select_query_options = SelectQueryOptions(QueryProcessingStage::Complete, 1); if (settings.allow_experimental_analyzer) { InterpreterSelectQueryAnalyzer interpreter_select_analyzer(query.select, new_context, select_query_options); pipeline = interpreter_select_analyzer.buildQueryPipeline(); } else { InterpreterSelectWithUnionQuery interpreter_select(query.select, new_context, select_query_options); pipeline = interpreter_select.buildQueryPipeline(); } } else { /// Passing 1 as subquery_depth will disable limiting size of intermediate result. auto select_query_options = SelectQueryOptions(QueryProcessingStage::Complete, 1); if (settings.allow_experimental_analyzer) { InterpreterSelectQueryAnalyzer interpreter_select_analyzer(query.select, getContext(), select_query_options); pipeline = interpreter_select_analyzer.buildQueryPipeline(); } else { InterpreterSelectWithUnionQuery interpreter_select(query.select, getContext(), select_query_options); pipeline = interpreter_select.buildQueryPipeline(); } } pipeline.dropTotalsAndExtremes(); if (settings.max_insert_threads > 1) { pre_streams_size = std::min(static_cast(settings.max_insert_threads), pipeline.getNumStreams()); if (table->supportsParallelInsert()) sink_streams_size = pre_streams_size; } pipeline.resize(pre_streams_size); /// Allow to insert Nullable into non-Nullable columns, NULL values will be added as defaults values. if (getContext()->getSettingsRef().insert_null_as_default) { const auto & input_columns = pipeline.getHeader().getColumnsWithTypeAndName(); const auto & query_columns = query_sample_block.getColumnsWithTypeAndName(); const auto & output_columns = metadata_snapshot->getColumns(); if (input_columns.size() == query_columns.size()) { for (size_t col_idx = 0; col_idx < query_columns.size(); ++col_idx) { /// Change query sample block columns to Nullable to allow inserting nullable columns, where NULL values will be substituted with /// default column values (in AddingDefaultsTransform), so all values will be cast correctly. if (isNullableOrLowCardinalityNullable(input_columns[col_idx].type) && !isNullableOrLowCardinalityNullable(query_columns[col_idx].type) && output_columns.has(query_columns[col_idx].name)) query_sample_block.setColumn(col_idx, ColumnWithTypeAndName(makeNullableOrLowCardinalityNullable(query_columns[col_idx].column), makeNullableOrLowCardinalityNullable(query_columns[col_idx].type), query_columns[col_idx].name)); } } } } else if (query.watch) { InterpreterWatchQuery interpreter_watch{ query.watch, getContext() }; pipeline = interpreter_watch.buildQueryPipeline(); } ThreadGroupPtr running_group; if (current_thread) running_group = current_thread->getThreadGroup(); if (!running_group) running_group = std::make_shared(getContext()); for (size_t i = 0; i < sink_streams_size; ++i) { auto out = buildSink(table, metadata_snapshot, /* thread_status_holder= */ nullptr, running_group, /* elapsed_counter_ms= */ nullptr); sink_chains.emplace_back(std::move(out)); } for (size_t i = 0; i < pre_streams_size; ++i) { auto out = buildPreSinkChain(sink_chains[0].getInputHeader(), table, metadata_snapshot, query_sample_block, /* thread_status_holder= */ nullptr); presink_chains.emplace_back(std::move(out)); } } BlockIO res; /// What type of query: INSERT or INSERT SELECT or INSERT WATCH? if (distributed_pipeline) { res.pipeline = std::move(*distributed_pipeline); } else if (query.select || query.watch) { const auto & header = presink_chains.at(0).getInputHeader(); auto actions_dag = ActionsDAG::makeConvertingActions( pipeline.getHeader().getColumnsWithTypeAndName(), header.getColumnsWithTypeAndName(), ActionsDAG::MatchColumnsMode::Position); auto actions = std::make_shared(actions_dag, ExpressionActionsSettings::fromContext(getContext(), CompileExpressions::yes)); pipeline.addSimpleTransform([&](const Block & in_header) -> ProcessorPtr { return std::make_shared(in_header, actions); }); /// We need to convert Sparse columns to full, because it's destination storage /// may not support it or may have different settings for applying Sparse serialization. pipeline.addSimpleTransform([&](const Block & in_header) -> ProcessorPtr { return std::make_shared(in_header); }); size_t num_select_threads = pipeline.getNumThreads(); for (auto & chain : presink_chains) resources = chain.detachResources(); for (auto & chain : sink_chains) resources = chain.detachResources(); pipeline.addChains(std::move(presink_chains)); pipeline.resize(sink_chains.size()); pipeline.addChains(std::move(sink_chains)); if (!settings.parallel_view_processing) { /// Don't use more threads for INSERT than for SELECT to reduce memory consumption. if (pipeline.getNumThreads() > num_select_threads) pipeline.setMaxThreads(num_select_threads); } else if (pipeline.getNumThreads() < settings.max_threads) { /// It is possible for query to have max_threads=1, due to optimize_trivial_insert_select, /// however in case of parallel_view_processing and multiple views, views can still be processed in parallel. /// /// Note, number of threads will be limited by buildPushingToViewsChain() to max_threads. pipeline.setMaxThreads(settings.max_threads); } pipeline.setSinks([&](const Block & cur_header, QueryPipelineBuilder::StreamType) -> ProcessorPtr { return std::make_shared(cur_header); }); if (!allow_materialized) { for (const auto & column : metadata_snapshot->getColumns()) if (column.default_desc.kind == ColumnDefaultKind::Materialized && header.has(column.name)) throw Exception(ErrorCodes::ILLEGAL_COLUMN, "Cannot insert column {}, because it is MATERIALIZED column.", column.name); } res.pipeline = QueryPipelineBuilder::getPipeline(std::move(pipeline)); } else { presink_chains.at(0).appendChain(std::move(sink_chains.at(0))); res.pipeline = QueryPipeline(std::move(presink_chains[0])); res.pipeline.setNumThreads(std::min(res.pipeline.getNumThreads(), settings.max_threads)); res.pipeline.setConcurrencyControl(settings.use_concurrency_control); if (query.hasInlinedData() && !async_insert) { /// can execute without additional data auto format = getInputFormatFromASTInsertQuery(query_ptr, true, query_sample_block, getContext(), nullptr); for (auto && buffer : owned_buffers) format->addBuffer(std::move(buffer)); auto pipe = getSourceFromInputFormat(query_ptr, std::move(format), getContext(), nullptr); res.pipeline.complete(std::move(pipe)); } } res.pipeline.addResources(std::move(resources)); res.pipeline.addStorageHolder(table); if (inner_table) res.pipeline.addStorageHolder(inner_table); return res; } StorageID InterpreterInsertQuery::getDatabaseTable() const { return query_ptr->as().table_id; } void InterpreterInsertQuery::extendQueryLogElemImpl(QueryLogElement & elem, ContextPtr context_) { const auto & insert_table = context_->getInsertionTable(); if (!insert_table.empty()) { elem.query_databases.insert(insert_table.getDatabaseName()); elem.query_tables.insert(insert_table.getFullNameNotQuoted()); } } void InterpreterInsertQuery::extendQueryLogElemImpl(QueryLogElement & elem, const ASTPtr &, ContextPtr context_) const { extendQueryLogElemImpl(elem, context_); } }