mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-13 18:02:24 +00:00
981 lines
35 KiB
C++
981 lines
35 KiB
C++
#include <Storages/StorageDistributed.h>
|
|
|
|
#include <Databases/IDatabase.h>
|
|
#include <Disks/StoragePolicy.h>
|
|
#include <Disks/DiskLocal.h>
|
|
|
|
#include <DataTypes/DataTypeFactory.h>
|
|
#include <DataTypes/DataTypesNumber.h>
|
|
|
|
#include <Storages/Distributed/DistributedBlockOutputStream.h>
|
|
#include <Storages/StorageFactory.h>
|
|
#include <Storages/AlterCommands.h>
|
|
|
|
#include <Columns/ColumnConst.h>
|
|
|
|
#include <Common/Macros.h>
|
|
#include <Common/escapeForFileName.h>
|
|
#include <Common/typeid_cast.h>
|
|
#include <Common/quoteString.h>
|
|
|
|
#include <Parsers/ASTDropQuery.h>
|
|
#include <Parsers/ASTExpressionList.h>
|
|
#include <Parsers/ASTIdentifier.h>
|
|
#include <Parsers/ASTInsertQuery.h>
|
|
#include <Parsers/ASTLiteral.h>
|
|
#include <Parsers/ASTSelectQuery.h>
|
|
#include <Parsers/ASTTablesInSelectQuery.h>
|
|
#include <Parsers/ParserAlterQuery.h>
|
|
#include <Parsers/TablePropertiesQueriesASTs.h>
|
|
#include <Parsers/parseQuery.h>
|
|
|
|
#include <Interpreters/ClusterProxy/SelectStreamFactory.h>
|
|
#include <Interpreters/ClusterProxy/executeQuery.h>
|
|
#include <Interpreters/Cluster.h>
|
|
#include <Interpreters/ExpressionAnalyzer.h>
|
|
#include <Interpreters/InterpreterAlterQuery.h>
|
|
#include <Interpreters/InterpreterDescribeQuery.h>
|
|
#include <Interpreters/InterpreterSelectQuery.h>
|
|
#include <Interpreters/TranslateQualifiedNamesVisitor.h>
|
|
#include <Interpreters/TreeRewriter.h>
|
|
#include <Interpreters/Context.h>
|
|
#include <Interpreters/createBlockSelector.h>
|
|
#include <Interpreters/evaluateConstantExpression.h>
|
|
#include <Interpreters/getClusterName.h>
|
|
#include <Interpreters/getTableExpressions.h>
|
|
#include <Functions/IFunction.h>
|
|
|
|
#include <Core/Field.h>
|
|
#include <Core/Settings.h>
|
|
|
|
#include <IO/ReadHelpers.h>
|
|
#include <IO/WriteBufferFromString.h>
|
|
#include <IO/Operators.h>
|
|
|
|
#include <Poco/DirectoryIterator.h>
|
|
|
|
#include <memory>
|
|
#include <filesystem>
|
|
#include <optional>
|
|
#include <cassert>
|
|
|
|
|
|
namespace
|
|
{
|
|
const UInt64 FORCE_OPTIMIZE_SKIP_UNUSED_SHARDS_HAS_SHARDING_KEY = 1;
|
|
const UInt64 FORCE_OPTIMIZE_SKIP_UNUSED_SHARDS_ALWAYS = 2;
|
|
|
|
const UInt64 DISTRIBUTED_GROUP_BY_NO_MERGE_AFTER_AGGREGATION = 2;
|
|
}
|
|
|
|
namespace DB
|
|
{
|
|
|
|
namespace ErrorCodes
|
|
{
|
|
extern const int NOT_IMPLEMENTED;
|
|
extern const int STORAGE_REQUIRES_PARAMETER;
|
|
extern const int BAD_ARGUMENTS;
|
|
extern const int NUMBER_OF_ARGUMENTS_DOESNT_MATCH;
|
|
extern const int INCORRECT_NUMBER_OF_COLUMNS;
|
|
extern const int INFINITE_LOOP;
|
|
extern const int TYPE_MISMATCH;
|
|
extern const int TOO_MANY_ROWS;
|
|
extern const int UNABLE_TO_SKIP_UNUSED_SHARDS;
|
|
}
|
|
|
|
namespace ActionLocks
|
|
{
|
|
extern const StorageActionBlockType DistributedSend;
|
|
}
|
|
|
|
namespace
|
|
{
|
|
|
|
/// select query has database, table and table function names as AST pointers
|
|
/// Creates a copy of query, changes database, table and table function names.
|
|
ASTPtr rewriteSelectQuery(const ASTPtr & query, const std::string & database, const std::string & table, ASTPtr table_function_ptr = nullptr)
|
|
{
|
|
auto modified_query_ast = query->clone();
|
|
|
|
ASTSelectQuery & select_query = modified_query_ast->as<ASTSelectQuery &>();
|
|
|
|
// Get rid of the settings clause so we don't send them to remote. Thus newly non-important
|
|
// settings won't break any remote parser. It's also more reasonable since the query settings
|
|
// are written into the query context and will be sent by the query pipeline.
|
|
select_query.setExpression(ASTSelectQuery::Expression::SETTINGS, {});
|
|
|
|
if (table_function_ptr)
|
|
select_query.addTableFunction(table_function_ptr);
|
|
else
|
|
select_query.replaceDatabaseAndTable(database, table);
|
|
|
|
/// Restore long column names (cause our short names are ambiguous).
|
|
/// TODO: aliased table functions & CREATE TABLE AS table function cases
|
|
if (!table_function_ptr)
|
|
{
|
|
RestoreQualifiedNamesVisitor::Data data;
|
|
data.distributed_table = DatabaseAndTableWithAlias(*getTableExpression(query->as<ASTSelectQuery &>(), 0));
|
|
data.remote_table.database = database;
|
|
data.remote_table.table = table;
|
|
data.rename = true;
|
|
RestoreQualifiedNamesVisitor(data).visit(modified_query_ast);
|
|
}
|
|
|
|
return modified_query_ast;
|
|
}
|
|
|
|
/// The columns list in the original INSERT query is incorrect because inserted blocks are transformed
|
|
/// to the form of the sample block of the Distributed table. So we rewrite it and add all columns from
|
|
/// the sample block instead.
|
|
ASTPtr createInsertToRemoteTableQuery(const std::string & database, const std::string & table, const Block & sample_block_non_materialized)
|
|
{
|
|
auto query = std::make_shared<ASTInsertQuery>();
|
|
query->table_id = StorageID(database, table);
|
|
|
|
auto columns = std::make_shared<ASTExpressionList>();
|
|
query->columns = columns;
|
|
query->children.push_back(columns);
|
|
for (const auto & col : sample_block_non_materialized)
|
|
columns->children.push_back(std::make_shared<ASTIdentifier>(col.name));
|
|
|
|
return query;
|
|
}
|
|
|
|
/// Calculate maximum number in file names in directory and all subdirectories.
|
|
/// To ensure global order of data blocks yet to be sent across server restarts.
|
|
UInt64 getMaximumFileNumber(const std::string & dir_path)
|
|
{
|
|
UInt64 res = 0;
|
|
|
|
std::filesystem::recursive_directory_iterator begin(dir_path);
|
|
std::filesystem::recursive_directory_iterator end;
|
|
for (auto it = begin; it != end; ++it)
|
|
{
|
|
const auto & file_path = it->path();
|
|
|
|
if (!std::filesystem::is_regular_file(*it) || !endsWith(file_path.filename().string(), ".bin"))
|
|
continue;
|
|
|
|
UInt64 num = 0;
|
|
try
|
|
{
|
|
num = parse<UInt64>(file_path.filename().stem().string());
|
|
}
|
|
catch (Exception & e)
|
|
{
|
|
e.addMessage("Unexpected file name " + file_path.filename().string() + " found at " + file_path.parent_path().string() + ", should have numeric base name.");
|
|
throw;
|
|
}
|
|
|
|
if (num > res)
|
|
res = num;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
std::string makeFormattedListOfShards(const ClusterPtr & cluster)
|
|
{
|
|
WriteBufferFromOwnString buf;
|
|
|
|
bool head = true;
|
|
buf << "[";
|
|
for (const auto & shard_info : cluster->getShardsInfo())
|
|
{
|
|
(head ? buf : buf << ", ") << shard_info.shard_num;
|
|
head = false;
|
|
}
|
|
buf << "]";
|
|
|
|
return buf.str();
|
|
}
|
|
|
|
ExpressionActionsPtr buildShardingKeyExpression(const ASTPtr & sharding_key, const Context & context, const NamesAndTypesList & columns, bool project)
|
|
{
|
|
ASTPtr query = sharding_key;
|
|
auto syntax_result = TreeRewriter(context).analyze(query, columns);
|
|
return ExpressionAnalyzer(query, syntax_result, context).getActions(project);
|
|
}
|
|
|
|
bool isExpressionActionsDeterministics(const ExpressionActionsPtr & actions)
|
|
{
|
|
for (const auto & action : actions->getActions())
|
|
{
|
|
if (action.node->type != ActionsDAG::ActionType::FUNCTION)
|
|
continue;
|
|
if (!action.node->function_base->isDeterministic())
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
class ReplacingConstantExpressionsMatcher
|
|
{
|
|
public:
|
|
using Data = Block;
|
|
|
|
static bool needChildVisit(ASTPtr &, const ASTPtr &)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
static void visit(ASTPtr & node, Block & block_with_constants)
|
|
{
|
|
if (!node->as<ASTFunction>())
|
|
return;
|
|
|
|
std::string name = node->getColumnName();
|
|
if (block_with_constants.has(name))
|
|
{
|
|
auto result = block_with_constants.getByName(name);
|
|
if (!isColumnConst(*result.column))
|
|
return;
|
|
|
|
node = std::make_shared<ASTLiteral>(assert_cast<const ColumnConst &>(*result.column).getField());
|
|
}
|
|
}
|
|
};
|
|
|
|
void replaceConstantExpressions(
|
|
ASTPtr & node,
|
|
const Context & context,
|
|
const NamesAndTypesList & columns,
|
|
ConstStoragePtr storage,
|
|
const StorageMetadataPtr & metadata_snapshot)
|
|
{
|
|
auto syntax_result = TreeRewriter(context).analyze(node, columns, storage, metadata_snapshot);
|
|
Block block_with_constants = KeyCondition::getBlockWithConstants(node, syntax_result, context);
|
|
|
|
InDepthNodeVisitor<ReplacingConstantExpressionsMatcher, true> visitor(block_with_constants);
|
|
visitor.visit(node);
|
|
}
|
|
|
|
/// Returns one of the following:
|
|
/// - QueryProcessingStage::Complete
|
|
/// - QueryProcessingStage::WithMergeableStateAfterAggregation
|
|
/// - none (in this case regular WithMergeableState should be used)
|
|
std::optional<QueryProcessingStage::Enum> getOptimizedQueryProcessingStage(const ASTPtr & query_ptr, bool extremes, const Block & sharding_key_block)
|
|
{
|
|
const auto & select = query_ptr->as<ASTSelectQuery &>();
|
|
|
|
auto sharding_block_has = [&](const auto & exprs, size_t limit = SIZE_MAX) -> bool
|
|
{
|
|
size_t i = 0;
|
|
for (auto & expr : exprs)
|
|
{
|
|
++i;
|
|
if (i > limit)
|
|
break;
|
|
|
|
auto id = expr->template as<ASTIdentifier>();
|
|
if (!id)
|
|
return false;
|
|
/// TODO: if GROUP BY contains multiIf()/if() it should contain only columns from sharding_key
|
|
if (!sharding_key_block.has(id->name()))
|
|
return false;
|
|
}
|
|
return true;
|
|
};
|
|
|
|
// GROUP BY qualifiers
|
|
// - TODO: WITH TOTALS can be implemented
|
|
// - TODO: WITH ROLLUP can be implemented (I guess)
|
|
if (select.group_by_with_totals || select.group_by_with_rollup || select.group_by_with_cube)
|
|
return {};
|
|
|
|
// TODO: extremes support can be implemented
|
|
if (extremes)
|
|
return {};
|
|
|
|
// DISTINCT
|
|
if (select.distinct)
|
|
{
|
|
if (!sharding_block_has(select.select()->children))
|
|
return {};
|
|
}
|
|
|
|
// GROUP BY
|
|
const ASTPtr group_by = select.groupBy();
|
|
if (!group_by)
|
|
{
|
|
if (!select.distinct)
|
|
return {};
|
|
}
|
|
else
|
|
{
|
|
if (!sharding_block_has(group_by->children, 1))
|
|
return {};
|
|
}
|
|
|
|
// ORDER BY
|
|
const ASTPtr order_by = select.orderBy();
|
|
if (order_by)
|
|
return QueryProcessingStage::WithMergeableStateAfterAggregation;
|
|
|
|
// LIMIT BY
|
|
// LIMIT
|
|
if (select.limitBy() || select.limitLength())
|
|
return QueryProcessingStage::WithMergeableStateAfterAggregation;
|
|
|
|
// Only simple SELECT FROM GROUP BY sharding_key can use Complete state.
|
|
return QueryProcessingStage::Complete;
|
|
}
|
|
|
|
size_t getClusterQueriedNodes(const Settings & settings, const ClusterPtr & cluster)
|
|
{
|
|
size_t num_local_shards = cluster->getLocalShardCount();
|
|
size_t num_remote_shards = cluster->getRemoteShardCount();
|
|
return (num_remote_shards * settings.max_parallel_replicas) + num_local_shards;
|
|
}
|
|
|
|
}
|
|
|
|
|
|
/// For destruction of std::unique_ptr of type that is incomplete in class definition.
|
|
StorageDistributed::~StorageDistributed() = default;
|
|
|
|
|
|
NamesAndTypesList StorageDistributed::getVirtuals() const
|
|
{
|
|
/// NOTE This is weird. Most of these virtual columns are part of MergeTree
|
|
/// tables info. But Distributed is general-purpose engine.
|
|
return NamesAndTypesList{
|
|
NameAndTypePair("_table", std::make_shared<DataTypeString>()),
|
|
NameAndTypePair("_part", std::make_shared<DataTypeString>()),
|
|
NameAndTypePair("_part_index", std::make_shared<DataTypeUInt64>()),
|
|
NameAndTypePair("_partition_id", std::make_shared<DataTypeString>()),
|
|
NameAndTypePair("_sample_factor", std::make_shared<DataTypeFloat64>()),
|
|
NameAndTypePair("_shard_num", std::make_shared<DataTypeUInt32>()),
|
|
};
|
|
}
|
|
|
|
StorageDistributed::StorageDistributed(
|
|
const StorageID & id_,
|
|
const ColumnsDescription & columns_,
|
|
const ConstraintsDescription & constraints_,
|
|
const String & remote_database_,
|
|
const String & remote_table_,
|
|
const String & cluster_name_,
|
|
const Context & context_,
|
|
const ASTPtr & sharding_key_,
|
|
const String & storage_policy_name_,
|
|
const String & relative_data_path_,
|
|
bool attach_,
|
|
ClusterPtr owned_cluster_)
|
|
: IStorage(id_)
|
|
, remote_database(remote_database_)
|
|
, remote_table(remote_table_)
|
|
, global_context(std::make_unique<Context>(context_))
|
|
, log(&Poco::Logger::get("StorageDistributed (" + id_.table_name + ")"))
|
|
, owned_cluster(std::move(owned_cluster_))
|
|
, cluster_name(global_context->getMacros()->expand(cluster_name_))
|
|
, has_sharding_key(sharding_key_)
|
|
, relative_data_path(relative_data_path_)
|
|
{
|
|
StorageInMemoryMetadata storage_metadata;
|
|
storage_metadata.setColumns(columns_);
|
|
storage_metadata.setConstraints(constraints_);
|
|
setInMemoryMetadata(storage_metadata);
|
|
|
|
if (sharding_key_)
|
|
{
|
|
sharding_key_expr = buildShardingKeyExpression(sharding_key_, *global_context, storage_metadata.getColumns().getAllPhysical(), false);
|
|
sharding_key_column_name = sharding_key_->getColumnName();
|
|
sharding_key_is_deterministic = isExpressionActionsDeterministics(sharding_key_expr);
|
|
}
|
|
|
|
if (!relative_data_path.empty())
|
|
{
|
|
storage_policy = global_context->getStoragePolicy(storage_policy_name_);
|
|
data_volume = storage_policy->getVolume(0);
|
|
if (storage_policy->getVolumes().size() > 1)
|
|
LOG_WARNING(log, "Storage policy for Distributed table has multiple volumes. "
|
|
"Only {} volume will be used to store data. Other will be ignored.", data_volume->getName());
|
|
}
|
|
|
|
/// Sanity check. Skip check if the table is already created to allow the server to start.
|
|
if (!attach_ && !cluster_name.empty())
|
|
{
|
|
size_t num_local_shards = global_context->getCluster(cluster_name)->getLocalShardCount();
|
|
if (num_local_shards && remote_database == id_.database_name && remote_table == id_.table_name)
|
|
throw Exception("Distributed table " + id_.table_name + " looks at itself", ErrorCodes::INFINITE_LOOP);
|
|
}
|
|
}
|
|
|
|
|
|
StorageDistributed::StorageDistributed(
|
|
const StorageID & id_,
|
|
const ColumnsDescription & columns_,
|
|
const ConstraintsDescription & constraints_,
|
|
ASTPtr remote_table_function_ptr_,
|
|
const String & cluster_name_,
|
|
const Context & context_,
|
|
const ASTPtr & sharding_key_,
|
|
const String & storage_policy_name_,
|
|
const String & relative_data_path_,
|
|
bool attach,
|
|
ClusterPtr owned_cluster_)
|
|
: StorageDistributed(id_, columns_, constraints_, String{}, String{}, cluster_name_, context_, sharding_key_, storage_policy_name_, relative_data_path_, attach, std::move(owned_cluster_))
|
|
{
|
|
remote_table_function_ptr = std::move(remote_table_function_ptr_);
|
|
}
|
|
|
|
QueryProcessingStage::Enum StorageDistributed::getQueryProcessingStage(
|
|
const Context & context, QueryProcessingStage::Enum to_stage, SelectQueryInfo & query_info) const
|
|
{
|
|
const auto & settings = context.getSettingsRef();
|
|
auto metadata_snapshot = getInMemoryMetadataPtr();
|
|
|
|
ClusterPtr cluster = getCluster();
|
|
query_info.cluster = cluster;
|
|
|
|
/// Always calculate optimized cluster here, to avoid conditions during read()
|
|
/// (Anyway it will be calculated in the read())
|
|
if (settings.optimize_skip_unused_shards)
|
|
{
|
|
ClusterPtr optimized_cluster = getOptimizedCluster(context, metadata_snapshot, query_info.query);
|
|
if (optimized_cluster)
|
|
{
|
|
LOG_DEBUG(log, "Skipping irrelevant shards - the query will be sent to the following shards of the cluster (shard numbers): {}", makeFormattedListOfShards(optimized_cluster));
|
|
cluster = optimized_cluster;
|
|
query_info.cluster = cluster;
|
|
}
|
|
else
|
|
{
|
|
LOG_DEBUG(log, "Unable to figure out irrelevant shards from WHERE/PREWHERE clauses - the query will be sent to all shards of the cluster{}", has_sharding_key ? "" : " (no sharding key)");
|
|
}
|
|
}
|
|
|
|
if (settings.distributed_group_by_no_merge)
|
|
{
|
|
if (settings.distributed_group_by_no_merge == DISTRIBUTED_GROUP_BY_NO_MERGE_AFTER_AGGREGATION)
|
|
return QueryProcessingStage::WithMergeableStateAfterAggregation;
|
|
else
|
|
return QueryProcessingStage::Complete;
|
|
}
|
|
|
|
/// Nested distributed query cannot return Complete stage,
|
|
/// since the parent query need to aggregate the results after.
|
|
if (to_stage == QueryProcessingStage::WithMergeableState)
|
|
return QueryProcessingStage::WithMergeableState;
|
|
|
|
/// If there is only one node, the query can be fully processed by the
|
|
/// shard, initiator will work as a proxy only.
|
|
if (getClusterQueriedNodes(settings, cluster) == 1)
|
|
return QueryProcessingStage::Complete;
|
|
|
|
if (settings.optimize_skip_unused_shards &&
|
|
settings.optimize_distributed_group_by_sharding_key &&
|
|
has_sharding_key &&
|
|
(settings.allow_nondeterministic_optimize_skip_unused_shards || sharding_key_is_deterministic))
|
|
{
|
|
Block sharding_key_block = sharding_key_expr->getSampleBlock();
|
|
auto stage = getOptimizedQueryProcessingStage(query_info.query, settings.extremes, sharding_key_block);
|
|
if (stage)
|
|
{
|
|
LOG_DEBUG(log, "Force processing stage to {}", QueryProcessingStage::toString(*stage));
|
|
return *stage;
|
|
}
|
|
}
|
|
|
|
return QueryProcessingStage::WithMergeableState;
|
|
}
|
|
|
|
Pipe StorageDistributed::read(
|
|
const Names & column_names,
|
|
const StorageMetadataPtr & metadata_snapshot,
|
|
SelectQueryInfo & query_info,
|
|
const Context & context,
|
|
QueryProcessingStage::Enum processed_stage,
|
|
const size_t max_block_size,
|
|
const unsigned num_streams)
|
|
{
|
|
QueryPlan plan;
|
|
read(plan, column_names, metadata_snapshot, query_info, context, processed_stage, max_block_size, num_streams);
|
|
return plan.convertToPipe();
|
|
}
|
|
|
|
void StorageDistributed::read(
|
|
QueryPlan & query_plan,
|
|
const Names & column_names,
|
|
const StorageMetadataPtr & metadata_snapshot,
|
|
SelectQueryInfo & query_info,
|
|
const Context & context,
|
|
QueryProcessingStage::Enum processed_stage,
|
|
const size_t /*max_block_size*/,
|
|
const unsigned /*num_streams*/)
|
|
{
|
|
const auto & modified_query_ast = rewriteSelectQuery(
|
|
query_info.query, remote_database, remote_table, remote_table_function_ptr);
|
|
|
|
Block header =
|
|
InterpreterSelectQuery(query_info.query, context, SelectQueryOptions(processed_stage)).getSampleBlock();
|
|
|
|
const Scalars & scalars = context.hasQueryContext() ? context.getQueryContext().getScalars() : Scalars{};
|
|
|
|
bool has_virtual_shard_num_column = std::find(column_names.begin(), column_names.end(), "_shard_num") != column_names.end();
|
|
if (has_virtual_shard_num_column && !isVirtualColumn("_shard_num", metadata_snapshot))
|
|
has_virtual_shard_num_column = false;
|
|
|
|
ClusterProxy::SelectStreamFactory select_stream_factory = remote_table_function_ptr
|
|
? ClusterProxy::SelectStreamFactory(
|
|
header, processed_stage, remote_table_function_ptr, scalars, has_virtual_shard_num_column, context.getExternalTables())
|
|
: ClusterProxy::SelectStreamFactory(
|
|
header, processed_stage, StorageID{remote_database, remote_table}, scalars, has_virtual_shard_num_column, context.getExternalTables());
|
|
|
|
ClusterProxy::executeQuery(query_plan, select_stream_factory, log,
|
|
modified_query_ast, context, query_info);
|
|
}
|
|
|
|
|
|
BlockOutputStreamPtr StorageDistributed::write(const ASTPtr &, const StorageMetadataPtr & metadata_snapshot, const Context & context)
|
|
{
|
|
auto cluster = getCluster();
|
|
const auto & settings = context.getSettingsRef();
|
|
|
|
/// Ban an attempt to make async insert into the table belonging to DatabaseMemory
|
|
if (!storage_policy && !owned_cluster && !settings.insert_distributed_sync)
|
|
{
|
|
throw Exception("Storage " + getName() + " must has own data directory to enable asynchronous inserts",
|
|
ErrorCodes::BAD_ARGUMENTS);
|
|
}
|
|
|
|
/// If sharding key is not specified, then you can only write to a shard containing only one shard
|
|
if (!has_sharding_key && ((cluster->getLocalShardCount() + cluster->getRemoteShardCount()) >= 2))
|
|
{
|
|
throw Exception("Method write is not supported by storage " + getName() + " with more than one shard and no sharding key provided",
|
|
ErrorCodes::STORAGE_REQUIRES_PARAMETER);
|
|
}
|
|
|
|
/// Force sync insertion if it is remote() table function
|
|
bool insert_sync = settings.insert_distributed_sync || owned_cluster;
|
|
auto timeout = settings.insert_distributed_timeout;
|
|
|
|
/// DistributedBlockOutputStream will not own cluster, but will own ConnectionPools of the cluster
|
|
return std::make_shared<DistributedBlockOutputStream>(
|
|
context, *this, metadata_snapshot, createInsertToRemoteTableQuery(remote_database, remote_table, metadata_snapshot->getSampleBlockNonMaterialized()), cluster,
|
|
insert_sync, timeout);
|
|
}
|
|
|
|
|
|
void StorageDistributed::checkAlterIsPossible(const AlterCommands & commands, const Settings & /* settings */) const
|
|
{
|
|
for (const auto & command : commands)
|
|
{
|
|
if (command.type != AlterCommand::Type::ADD_COLUMN
|
|
&& command.type != AlterCommand::Type::MODIFY_COLUMN
|
|
&& command.type != AlterCommand::Type::DROP_COLUMN
|
|
&& command.type != AlterCommand::Type::COMMENT_COLUMN
|
|
&& command.type != AlterCommand::Type::RENAME_COLUMN)
|
|
|
|
throw Exception("Alter of type '" + alterTypeToString(command.type) + "' is not supported by storage " + getName(),
|
|
ErrorCodes::NOT_IMPLEMENTED);
|
|
}
|
|
}
|
|
|
|
void StorageDistributed::alter(const AlterCommands & params, const Context & context, TableLockHolder &)
|
|
{
|
|
auto table_id = getStorageID();
|
|
|
|
checkAlterIsPossible(params, context.getSettingsRef());
|
|
StorageInMemoryMetadata new_metadata = getInMemoryMetadata();
|
|
params.apply(new_metadata, context);
|
|
DatabaseCatalog::instance().getDatabase(table_id.database_name)->alterTable(context, table_id, new_metadata);
|
|
setInMemoryMetadata(new_metadata);
|
|
}
|
|
|
|
|
|
void StorageDistributed::startup()
|
|
{
|
|
if (remote_database.empty() && !remote_table_function_ptr)
|
|
LOG_WARNING(log, "Name of remote database is empty. Default database will be used implicitly.");
|
|
|
|
if (!storage_policy)
|
|
return;
|
|
|
|
for (const DiskPtr & disk : data_volume->getDisks())
|
|
createDirectoryMonitors(disk->getPath());
|
|
|
|
for (const String & path : getDataPaths())
|
|
{
|
|
UInt64 inc = getMaximumFileNumber(path);
|
|
if (inc > file_names_increment.value)
|
|
file_names_increment.value.store(inc);
|
|
}
|
|
LOG_DEBUG(log, "Auto-increment is {}", file_names_increment.value);
|
|
}
|
|
|
|
|
|
void StorageDistributed::shutdown()
|
|
{
|
|
monitors_blocker.cancelForever();
|
|
|
|
std::lock_guard lock(cluster_nodes_mutex);
|
|
|
|
LOG_DEBUG(log, "Joining background threads for async INSERT");
|
|
cluster_nodes_data.clear();
|
|
LOG_DEBUG(log, "Background threads for async INSERT joined");
|
|
}
|
|
void StorageDistributed::drop()
|
|
{
|
|
// Some INSERT in-between shutdown() and drop() can call
|
|
// requireDirectoryMonitor() again, so call shutdown() to clear them, but
|
|
// when the drop() (this function) executed none of INSERT is allowed in
|
|
// parallel.
|
|
//
|
|
// And second time shutdown() should be fast, since none of
|
|
// DirectoryMonitor should do anything, because ActionBlocker is canceled
|
|
// (in shutdown()).
|
|
shutdown();
|
|
|
|
// Distributed table w/o sharding_key does not allows INSERTs
|
|
if (relative_data_path.empty())
|
|
return;
|
|
|
|
LOG_DEBUG(log, "Removing pending blocks for async INSERT from filesystem on DROP TABLE");
|
|
|
|
auto disks = data_volume->getDisks();
|
|
for (const auto & disk : disks)
|
|
disk->removeRecursive(relative_data_path);
|
|
|
|
LOG_DEBUG(log, "Removed");
|
|
}
|
|
|
|
Strings StorageDistributed::getDataPaths() const
|
|
{
|
|
Strings paths;
|
|
|
|
if (relative_data_path.empty())
|
|
return paths;
|
|
|
|
for (const DiskPtr & disk : data_volume->getDisks())
|
|
paths.push_back(disk->getPath() + relative_data_path);
|
|
|
|
return paths;
|
|
}
|
|
|
|
void StorageDistributed::truncate(const ASTPtr &, const StorageMetadataPtr &, const Context &, TableExclusiveLockHolder &)
|
|
{
|
|
std::lock_guard lock(cluster_nodes_mutex);
|
|
|
|
LOG_DEBUG(log, "Removing pending blocks for async INSERT from filesystem on TRUNCATE TABLE");
|
|
|
|
for (auto it = cluster_nodes_data.begin(); it != cluster_nodes_data.end();)
|
|
{
|
|
it->second.shutdownAndDropAllData();
|
|
it = cluster_nodes_data.erase(it);
|
|
}
|
|
|
|
LOG_DEBUG(log, "Removed");
|
|
}
|
|
|
|
StoragePolicyPtr StorageDistributed::getStoragePolicy() const
|
|
{
|
|
return storage_policy;
|
|
}
|
|
|
|
void StorageDistributed::createDirectoryMonitors(const std::string & disk)
|
|
{
|
|
const std::string path(disk + relative_data_path);
|
|
Poco::File{path}.createDirectories();
|
|
|
|
std::filesystem::directory_iterator begin(path);
|
|
std::filesystem::directory_iterator end;
|
|
for (auto it = begin; it != end; ++it)
|
|
{
|
|
const auto & dir_path = it->path();
|
|
if (std::filesystem::is_directory(dir_path))
|
|
{
|
|
const auto & tmp_path = dir_path / "tmp";
|
|
|
|
/// "tmp" created by DistributedBlockOutputStream
|
|
if (std::filesystem::is_directory(tmp_path) && std::filesystem::is_empty(tmp_path))
|
|
std::filesystem::remove(tmp_path);
|
|
|
|
if (std::filesystem::is_empty(dir_path))
|
|
{
|
|
LOG_DEBUG(log, "Removing {} (used for async INSERT into Distributed)", dir_path.string());
|
|
/// Will be created by DistributedBlockOutputStream on demand.
|
|
std::filesystem::remove(dir_path);
|
|
}
|
|
else
|
|
{
|
|
requireDirectoryMonitor(disk, dir_path.filename().string());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
StorageDistributedDirectoryMonitor& StorageDistributed::requireDirectoryMonitor(const std::string & disk, const std::string & name)
|
|
{
|
|
const std::string path(disk + relative_data_path + name);
|
|
const std::string key(disk + name);
|
|
|
|
std::lock_guard lock(cluster_nodes_mutex);
|
|
auto & node_data = cluster_nodes_data[key];
|
|
if (!node_data.directory_monitor)
|
|
{
|
|
node_data.connection_pool = StorageDistributedDirectoryMonitor::createPool(name, *this);
|
|
node_data.directory_monitor = std::make_unique<StorageDistributedDirectoryMonitor>(
|
|
*this, path, node_data.connection_pool, monitors_blocker, global_context->getDistributedSchedulePool());
|
|
}
|
|
return *node_data.directory_monitor;
|
|
}
|
|
|
|
std::vector<StorageDistributedDirectoryMonitor::Status> StorageDistributed::getDirectoryMonitorsStatuses() const
|
|
{
|
|
std::vector<StorageDistributedDirectoryMonitor::Status> statuses;
|
|
std::lock_guard lock(cluster_nodes_mutex);
|
|
statuses.reserve(cluster_nodes_data.size());
|
|
for (const auto & node : cluster_nodes_data)
|
|
statuses.push_back(node.second.directory_monitor->getStatus());
|
|
return statuses;
|
|
}
|
|
|
|
size_t StorageDistributed::getShardCount() const
|
|
{
|
|
return getCluster()->getShardCount();
|
|
}
|
|
|
|
ClusterPtr StorageDistributed::getCluster() const
|
|
{
|
|
return owned_cluster ? owned_cluster : global_context->getCluster(cluster_name);
|
|
}
|
|
|
|
ClusterPtr StorageDistributed::getOptimizedCluster(const Context & context, const StorageMetadataPtr & metadata_snapshot, const ASTPtr & query_ptr) const
|
|
{
|
|
ClusterPtr cluster = getCluster();
|
|
const Settings & settings = context.getSettingsRef();
|
|
|
|
bool sharding_key_is_usable = settings.allow_nondeterministic_optimize_skip_unused_shards || sharding_key_is_deterministic;
|
|
|
|
if (has_sharding_key && sharding_key_is_usable)
|
|
{
|
|
ClusterPtr optimized = skipUnusedShards(cluster, query_ptr, metadata_snapshot, context);
|
|
if (optimized)
|
|
return optimized;
|
|
}
|
|
|
|
UInt64 force = settings.force_optimize_skip_unused_shards;
|
|
if (force)
|
|
{
|
|
WriteBufferFromOwnString exception_message;
|
|
if (!has_sharding_key)
|
|
exception_message << "No sharding key";
|
|
else if (!sharding_key_is_usable)
|
|
exception_message << "Sharding key is not deterministic";
|
|
else
|
|
exception_message << "Sharding key " << sharding_key_column_name << " is not used";
|
|
|
|
if (force == FORCE_OPTIMIZE_SKIP_UNUSED_SHARDS_ALWAYS)
|
|
throw Exception(exception_message.str(), ErrorCodes::UNABLE_TO_SKIP_UNUSED_SHARDS);
|
|
if (force == FORCE_OPTIMIZE_SKIP_UNUSED_SHARDS_HAS_SHARDING_KEY && has_sharding_key)
|
|
throw Exception(exception_message.str(), ErrorCodes::UNABLE_TO_SKIP_UNUSED_SHARDS);
|
|
}
|
|
|
|
return cluster;
|
|
}
|
|
|
|
void StorageDistributed::ClusterNodeData::flushAllData() const
|
|
{
|
|
directory_monitor->flushAllData();
|
|
}
|
|
|
|
void StorageDistributed::ClusterNodeData::shutdownAndDropAllData() const
|
|
{
|
|
directory_monitor->shutdownAndDropAllData();
|
|
}
|
|
|
|
IColumn::Selector StorageDistributed::createSelector(const ClusterPtr cluster, const ColumnWithTypeAndName & result)
|
|
{
|
|
const auto & slot_to_shard = cluster->getSlotToShard();
|
|
|
|
// If result.type is DataTypeLowCardinality, do shard according to its dictionaryType
|
|
#define CREATE_FOR_TYPE(TYPE) \
|
|
if (typeid_cast<const DataType##TYPE *>(result.type.get())) \
|
|
return createBlockSelector<TYPE>(*result.column, slot_to_shard); \
|
|
else if (auto * type_low_cardinality = typeid_cast<const DataTypeLowCardinality *>(result.type.get())) \
|
|
if (typeid_cast<const DataType ## TYPE *>(type_low_cardinality->getDictionaryType().get())) \
|
|
return createBlockSelector<TYPE>(*result.column->convertToFullColumnIfLowCardinality(), slot_to_shard);
|
|
|
|
CREATE_FOR_TYPE(UInt8)
|
|
CREATE_FOR_TYPE(UInt16)
|
|
CREATE_FOR_TYPE(UInt32)
|
|
CREATE_FOR_TYPE(UInt64)
|
|
CREATE_FOR_TYPE(Int8)
|
|
CREATE_FOR_TYPE(Int16)
|
|
CREATE_FOR_TYPE(Int32)
|
|
CREATE_FOR_TYPE(Int64)
|
|
|
|
#undef CREATE_FOR_TYPE
|
|
|
|
throw Exception{"Sharding key expression does not evaluate to an integer type", ErrorCodes::TYPE_MISMATCH};
|
|
}
|
|
|
|
/// Returns a new cluster with fewer shards if constant folding for `sharding_key_expr` is possible
|
|
/// using constraints from "PREWHERE" and "WHERE" conditions, otherwise returns `nullptr`
|
|
ClusterPtr StorageDistributed::skipUnusedShards(
|
|
ClusterPtr cluster,
|
|
const ASTPtr & query_ptr,
|
|
const StorageMetadataPtr & metadata_snapshot,
|
|
const Context & context) const
|
|
{
|
|
const auto & select = query_ptr->as<ASTSelectQuery &>();
|
|
|
|
if (!select.prewhere() && !select.where())
|
|
{
|
|
return nullptr;
|
|
}
|
|
|
|
ASTPtr condition_ast;
|
|
if (select.prewhere() && select.where())
|
|
{
|
|
condition_ast = makeASTFunction("and", select.prewhere()->clone(), select.where()->clone());
|
|
}
|
|
else
|
|
{
|
|
condition_ast = select.prewhere() ? select.prewhere()->clone() : select.where()->clone();
|
|
}
|
|
|
|
replaceConstantExpressions(condition_ast, context, metadata_snapshot->getColumns().getAll(), shared_from_this(), metadata_snapshot);
|
|
const auto blocks = evaluateExpressionOverConstantCondition(condition_ast, sharding_key_expr);
|
|
|
|
// Can't get definite answer if we can skip any shards
|
|
if (!blocks)
|
|
{
|
|
return nullptr;
|
|
}
|
|
|
|
std::set<int> shards;
|
|
|
|
for (const auto & block : *blocks)
|
|
{
|
|
if (!block.has(sharding_key_column_name))
|
|
throw Exception("sharding_key_expr should evaluate as a single row", ErrorCodes::TOO_MANY_ROWS);
|
|
|
|
const ColumnWithTypeAndName & result = block.getByName(sharding_key_column_name);
|
|
const auto selector = createSelector(cluster, result);
|
|
|
|
shards.insert(selector.begin(), selector.end());
|
|
}
|
|
|
|
return cluster->getClusterWithMultipleShards({shards.begin(), shards.end()});
|
|
}
|
|
|
|
ActionLock StorageDistributed::getActionLock(StorageActionBlockType type)
|
|
{
|
|
if (type == ActionLocks::DistributedSend)
|
|
return monitors_blocker.cancel();
|
|
return {};
|
|
}
|
|
|
|
void StorageDistributed::flushClusterNodesAllData()
|
|
{
|
|
std::lock_guard lock(cluster_nodes_mutex);
|
|
|
|
/// TODO: Maybe it should be executed in parallel
|
|
for (auto & node : cluster_nodes_data)
|
|
node.second.flushAllData();
|
|
}
|
|
|
|
void StorageDistributed::rename(const String & new_path_to_table_data, const StorageID & new_table_id)
|
|
{
|
|
assert(relative_data_path != new_path_to_table_data);
|
|
if (!relative_data_path.empty())
|
|
renameOnDisk(new_path_to_table_data);
|
|
renameInMemory(new_table_id);
|
|
}
|
|
|
|
|
|
void StorageDistributed::renameOnDisk(const String & new_path_to_table_data)
|
|
{
|
|
for (const DiskPtr & disk : data_volume->getDisks())
|
|
{
|
|
disk->moveDirectory(relative_data_path, new_path_to_table_data);
|
|
|
|
auto new_path = disk->getPath() + new_path_to_table_data;
|
|
LOG_DEBUG(log, "Updating path to {}", new_path);
|
|
|
|
std::lock_guard lock(cluster_nodes_mutex);
|
|
for (auto & node : cluster_nodes_data)
|
|
node.second.directory_monitor->updatePath(new_path);
|
|
}
|
|
|
|
relative_data_path = new_path_to_table_data;
|
|
}
|
|
|
|
|
|
void registerStorageDistributed(StorageFactory & factory)
|
|
{
|
|
factory.registerStorage("Distributed", [](const StorageFactory::Arguments & args)
|
|
{
|
|
/** Arguments of engine is following:
|
|
* - name of cluster in configuration;
|
|
* - name of remote database;
|
|
* - name of remote table;
|
|
* - policy to store data in;
|
|
*
|
|
* Remote database may be specified in following form:
|
|
* - identifier;
|
|
* - constant expression with string result, like currentDatabase();
|
|
* -- string literal as specific case;
|
|
* - empty string means 'use default database from cluster'.
|
|
*/
|
|
|
|
ASTs & engine_args = args.engine_args;
|
|
|
|
if (engine_args.size() < 3 || engine_args.size() > 5)
|
|
throw Exception(
|
|
"Storage Distributed requires from 3 to 5 parameters - "
|
|
"name of configuration section with list of remote servers, "
|
|
"name of remote database, "
|
|
"name of remote table, "
|
|
"sharding key expression (optional), "
|
|
"policy to store data in (optional).",
|
|
ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
|
|
|
|
String cluster_name = getClusterNameAndMakeLiteral(engine_args[0]);
|
|
|
|
engine_args[1] = evaluateConstantExpressionOrIdentifierAsLiteral(engine_args[1], args.local_context);
|
|
engine_args[2] = evaluateConstantExpressionOrIdentifierAsLiteral(engine_args[2], args.local_context);
|
|
|
|
String remote_database = engine_args[1]->as<ASTLiteral &>().value.safeGet<String>();
|
|
String remote_table = engine_args[2]->as<ASTLiteral &>().value.safeGet<String>();
|
|
|
|
const auto & sharding_key = engine_args.size() >= 4 ? engine_args[3] : nullptr;
|
|
const auto & storage_policy = engine_args.size() >= 5 ? engine_args[4]->as<ASTLiteral &>().value.safeGet<String>() : "default";
|
|
|
|
/// Check that sharding_key exists in the table and has numeric type.
|
|
if (sharding_key)
|
|
{
|
|
auto sharding_expr = buildShardingKeyExpression(sharding_key, args.context, args.columns.getAllPhysical(), true);
|
|
const Block & block = sharding_expr->getSampleBlock();
|
|
|
|
if (block.columns() != 1)
|
|
throw Exception("Sharding expression must return exactly one column", ErrorCodes::INCORRECT_NUMBER_OF_COLUMNS);
|
|
|
|
auto type = block.getByPosition(0).type;
|
|
|
|
if (!type->isValueRepresentedByInteger())
|
|
throw Exception("Sharding expression has type " + type->getName() +
|
|
", but should be one of integer type", ErrorCodes::TYPE_MISMATCH);
|
|
}
|
|
|
|
return StorageDistributed::create(
|
|
args.table_id, args.columns, args.constraints,
|
|
remote_database, remote_table, cluster_name,
|
|
args.context,
|
|
sharding_key,
|
|
storage_policy,
|
|
args.relative_data_path,
|
|
args.attach);
|
|
},
|
|
{
|
|
.source_access_type = AccessType::REMOTE,
|
|
});
|
|
}
|
|
|
|
}
|