2017-04-01 09:19:00 +00:00
|
|
|
#include <DataStreams/RemoteBlockInputStream.h>
|
|
|
|
#include <DataStreams/BlockExtraInfoInputStream.h>
|
|
|
|
#include <DataStreams/UnionBlockInputStream.h>
|
|
|
|
|
|
|
|
#include <Databases/IDatabase.h>
|
|
|
|
|
|
|
|
#include <Storages/StorageDistributed.h>
|
|
|
|
#include <Storages/VirtualColumnFactory.h>
|
|
|
|
#include <Storages/Distributed/DistributedBlockOutputStream.h>
|
|
|
|
#include <Storages/Distributed/DirectoryMonitor.h>
|
|
|
|
#include <Storages/MergeTree/ReshardingWorker.h>
|
|
|
|
|
|
|
|
#include <Common/escapeForFileName.h>
|
2017-07-13 20:58:19 +00:00
|
|
|
#include <Common/typeid_cast.h>
|
2017-04-01 09:19:00 +00:00
|
|
|
|
|
|
|
#include <Parsers/ASTInsertQuery.h>
|
|
|
|
#include <Parsers/ASTSelectQuery.h>
|
|
|
|
#include <Parsers/ASTIdentifier.h>
|
|
|
|
#include <Parsers/TablePropertiesQueriesASTs.h>
|
|
|
|
#include <Parsers/ParserAlterQuery.h>
|
|
|
|
#include <Parsers/parseQuery.h>
|
|
|
|
#include <Parsers/ASTWeightedZooKeeperPath.h>
|
|
|
|
#include <Parsers/ASTLiteral.h>
|
|
|
|
#include <Parsers/ASTExpressionList.h>
|
|
|
|
#include <Parsers/queryToString.h>
|
|
|
|
|
|
|
|
#include <Interpreters/InterpreterSelectQuery.h>
|
|
|
|
#include <Interpreters/InterpreterAlterQuery.h>
|
|
|
|
#include <Interpreters/InterpreterDescribeQuery.h>
|
|
|
|
#include <Interpreters/ExpressionAnalyzer.h>
|
|
|
|
#include <Interpreters/ClusterProxy/Query.h>
|
|
|
|
#include <Interpreters/ClusterProxy/SelectQueryConstructor.h>
|
|
|
|
#include <Interpreters/ClusterProxy/DescribeQueryConstructor.h>
|
|
|
|
#include <Interpreters/ClusterProxy/AlterQueryConstructor.h>
|
|
|
|
|
|
|
|
#include <Core/Field.h>
|
2012-05-21 20:38:34 +00:00
|
|
|
|
2016-12-12 03:33:34 +00:00
|
|
|
#include <Poco/DirectoryIterator.h>
|
|
|
|
|
2015-02-10 21:10:58 +00:00
|
|
|
#include <memory>
|
2014-08-19 08:04:13 +00:00
|
|
|
|
2017-05-10 06:39:37 +00:00
|
|
|
#include <boost/filesystem.hpp>
|
|
|
|
|
2016-12-12 03:33:34 +00:00
|
|
|
|
2012-05-21 20:38:34 +00:00
|
|
|
namespace DB
|
|
|
|
{
|
|
|
|
|
2016-01-11 21:46:36 +00:00
|
|
|
namespace ErrorCodes
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
extern const int STORAGE_REQUIRES_PARAMETER;
|
|
|
|
extern const int RESHARDING_NO_WORKER;
|
|
|
|
extern const int RESHARDING_INVALID_PARAMETERS;
|
|
|
|
extern const int RESHARDING_INITIATOR_CHECK_FAILED;
|
2016-01-11 21:46:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-08-21 12:07:29 +00:00
|
|
|
namespace
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-05-10 06:39:37 +00:00
|
|
|
/// select query has database and table names as AST pointers
|
|
|
|
/// Creates a copy of query, changes database and table names.
|
|
|
|
ASTPtr rewriteSelectQuery(const ASTPtr & query, const std::string & database, const std::string & table)
|
|
|
|
{
|
2017-06-15 14:07:31 +00:00
|
|
|
auto modified_query_ast = typeid_cast<const ASTSelectQuery &>(*query).cloneFirstSelect();
|
|
|
|
modified_query_ast->replaceDatabaseAndTable(database, table);
|
2017-05-10 06:39:37 +00:00
|
|
|
return modified_query_ast;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// insert query has database and table names as bare strings
|
|
|
|
/// Creates a copy of query, changes the database and table names.
|
|
|
|
ASTPtr rewriteInsertQuery(const ASTPtr & query, const std::string & database, const std::string & table)
|
|
|
|
{
|
|
|
|
auto modified_query_ast = query->clone();
|
|
|
|
|
|
|
|
auto & actual_query = typeid_cast<ASTInsertQuery &>(*modified_query_ast);
|
|
|
|
actual_query.database = database;
|
|
|
|
actual_query.table = table;
|
|
|
|
/// make sure query is not INSERT SELECT
|
|
|
|
actual_query.select = nullptr;
|
|
|
|
|
|
|
|
return modified_query_ast;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Calculate maximum number in file names in directory and all subdirectories.
|
|
|
|
/// To ensure global order of data blocks yet to be sent across server restarts.
|
|
|
|
UInt64 getMaximumFileNumber(const std::string & path)
|
|
|
|
{
|
|
|
|
UInt64 res = 0;
|
|
|
|
|
|
|
|
boost::filesystem::recursive_directory_iterator begin(path);
|
|
|
|
boost::filesystem::recursive_directory_iterator end;
|
|
|
|
for (auto it = begin; it != end; ++it)
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2017-05-10 06:39:37 +00:00
|
|
|
const auto & path = it->path();
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-05-10 06:39:37 +00:00
|
|
|
if (it->status().type() != boost::filesystem::regular_file || !endsWith(path.filename().string(), ".bin"))
|
|
|
|
continue;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-05-10 06:39:37 +00:00
|
|
|
UInt64 num = 0;
|
|
|
|
try
|
|
|
|
{
|
|
|
|
num = parse<UInt64>(path.filename().stem().string());
|
|
|
|
}
|
|
|
|
catch (Exception & e)
|
|
|
|
{
|
|
|
|
e.addMessage("Unexpected file name " + path.filename().string() + " found at " + path.parent_path().string() + ", should have numeric base name.");
|
|
|
|
throw;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (num > res)
|
|
|
|
res = num;
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
2017-05-10 06:39:37 +00:00
|
|
|
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
|
|
|
void initializeFileNamesIncrement(const std::string & path, SimpleIncrement & increment)
|
|
|
|
{
|
|
|
|
if (!path.empty())
|
|
|
|
increment.set(getMaximumFileNumber(path));
|
|
|
|
}
|
|
|
|
|
2014-08-13 12:52:30 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2017-06-06 18:48:38 +00:00
|
|
|
/// For destruction of std::unique_ptr of type that is incomplete in class definition.
|
|
|
|
StorageDistributed::~StorageDistributed() = default;
|
|
|
|
|
|
|
|
|
2012-11-06 17:04:38 +00:00
|
|
|
StorageDistributed::StorageDistributed(
|
2017-04-01 07:20:54 +00:00
|
|
|
const std::string & name_,
|
|
|
|
NamesAndTypesListPtr columns_,
|
|
|
|
const String & remote_database_,
|
|
|
|
const String & remote_table_,
|
|
|
|
const String & cluster_name_,
|
2017-05-23 18:37:14 +00:00
|
|
|
const Context & context_,
|
2017-04-01 07:20:54 +00:00
|
|
|
const ASTPtr & sharding_key_,
|
|
|
|
const String & data_path_)
|
|
|
|
: name(name_), columns(columns_),
|
|
|
|
remote_database(remote_database_), remote_table(remote_table_),
|
|
|
|
context(context_), cluster_name(cluster_name_), has_sharding_key(sharding_key_),
|
|
|
|
sharding_key_expr(sharding_key_ ? ExpressionAnalyzer(sharding_key_, context, nullptr, *columns).getActions(false) : nullptr),
|
|
|
|
sharding_key_column_name(sharding_key_ ? sharding_key_->getColumnName() : String{}),
|
|
|
|
path(data_path_.empty() ? "" : (data_path_ + escapeForFileName(name) + '/'))
|
2013-11-28 10:31:17 +00:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2014-09-30 03:08:47 +00:00
|
|
|
StorageDistributed::StorageDistributed(
|
2017-04-01 07:20:54 +00:00
|
|
|
const std::string & name_,
|
|
|
|
NamesAndTypesListPtr columns_,
|
|
|
|
const NamesAndTypesList & materialized_columns_,
|
|
|
|
const NamesAndTypesList & alias_columns_,
|
|
|
|
const ColumnDefaults & column_defaults_,
|
|
|
|
const String & remote_database_,
|
|
|
|
const String & remote_table_,
|
|
|
|
const String & cluster_name_,
|
2017-05-23 18:37:14 +00:00
|
|
|
const Context & context_,
|
2017-04-01 07:20:54 +00:00
|
|
|
const ASTPtr & sharding_key_,
|
|
|
|
const String & data_path_)
|
|
|
|
: IStorage{materialized_columns_, alias_columns_, column_defaults_},
|
|
|
|
name(name_), columns(columns_),
|
|
|
|
remote_database(remote_database_), remote_table(remote_table_),
|
|
|
|
context(context_), cluster_name(cluster_name_), has_sharding_key(sharding_key_),
|
|
|
|
sharding_key_expr(sharding_key_ ? ExpressionAnalyzer(sharding_key_, context, nullptr, *columns).getActions(false) : nullptr),
|
|
|
|
sharding_key_column_name(sharding_key_ ? sharding_key_->getColumnName() : String{}),
|
|
|
|
path(data_path_.empty() ? "" : (data_path_ + escapeForFileName(name) + '/'))
|
2014-09-30 03:08:47 +00:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2017-06-06 18:48:38 +00:00
|
|
|
StoragePtr StorageDistributed::createWithOwnCluster(
|
2017-04-01 07:20:54 +00:00
|
|
|
const std::string & name_,
|
|
|
|
NamesAndTypesListPtr columns_,
|
|
|
|
const String & remote_database_,
|
|
|
|
const String & remote_table_,
|
|
|
|
ClusterPtr & owned_cluster_,
|
2017-05-23 18:37:14 +00:00
|
|
|
const Context & context_)
|
2014-02-07 15:11:57 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
auto res = make_shared(
|
|
|
|
name_, columns_, remote_database_,
|
2017-06-06 17:06:14 +00:00
|
|
|
remote_table_, String{}, context_);
|
2014-02-23 02:27:09 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
res->owned_cluster = owned_cluster_;
|
2014-02-23 02:27:09 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
return res;
|
2014-02-07 15:11:57 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2012-05-21 20:38:34 +00:00
|
|
|
BlockInputStreams StorageDistributed::read(
|
2017-04-01 07:20:54 +00:00
|
|
|
const Names & column_names,
|
2017-07-15 03:48:36 +00:00
|
|
|
const SelectQueryInfo & query_info,
|
2017-04-01 07:20:54 +00:00
|
|
|
const Context & context,
|
|
|
|
QueryProcessingStage::Enum & processed_stage,
|
|
|
|
const size_t max_block_size,
|
2017-06-02 15:54:39 +00:00
|
|
|
const unsigned num_streams)
|
2012-05-21 20:38:34 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
auto cluster = getCluster();
|
2016-10-10 08:44:52 +00:00
|
|
|
|
2017-05-24 21:06:29 +00:00
|
|
|
const Settings & settings = context.getSettingsRef();
|
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
size_t result_size = (cluster->getRemoteShardCount() * settings.max_parallel_replicas) + cluster->getLocalShardCount();
|
2014-02-11 17:10:48 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
processed_stage = result_size == 1 || settings.distributed_group_by_no_merge
|
|
|
|
? QueryProcessingStage::Complete
|
|
|
|
: QueryProcessingStage::WithMergeableState;
|
2014-02-11 17:10:48 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
const auto & modified_query_ast = rewriteSelectQuery(
|
2017-07-15 03:48:36 +00:00
|
|
|
query_info.query, remote_database, remote_table);
|
2015-02-10 20:48:17 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
Tables external_tables;
|
2015-09-05 01:22:09 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
if (settings.global_subqueries_method == GlobalSubqueriesMethod::PUSH)
|
|
|
|
external_tables = context.getExternalTables();
|
2015-03-15 08:05:28 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/// Disable multiplexing of shards if there is an ORDER BY without GROUP BY.
|
|
|
|
//const ASTSelectQuery & ast = *(static_cast<const ASTSelectQuery *>(modified_query_ast.get()));
|
2015-11-12 01:17:44 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/** The functionality of shard_multiplexing is not completed - turn it off.
|
|
|
|
* (Because connecting to different shards within a single thread is not done in parallel.)
|
|
|
|
*/
|
|
|
|
//bool enable_shard_multiplexing = !(ast.order_expression_list && !ast.group_expression_list);
|
|
|
|
bool enable_shard_multiplexing = false;
|
2015-11-06 17:44:01 +00:00
|
|
|
|
2017-04-17 16:16:04 +00:00
|
|
|
ClusterProxy::SelectQueryConstructor select_query_constructor(
|
2017-07-03 21:04:10 +00:00
|
|
|
processed_stage, QualifiedTableName{remote_database, remote_table}, external_tables);
|
2015-11-06 17:44:01 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
return ClusterProxy::Query{select_query_constructor, cluster, modified_query_ast,
|
|
|
|
context, settings, enable_shard_multiplexing}.execute();
|
2012-05-21 20:38:34 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2017-05-21 22:25:25 +00:00
|
|
|
BlockOutputStreamPtr StorageDistributed::write(const ASTPtr & query, const Settings & settings)
|
2014-08-12 13:46:46 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
auto cluster = context.getCluster(cluster_name);
|
2016-10-10 08:44:52 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/// TODO: !path.empty() can be replaced by !owned_cluster or !cluster_name.empty() ?
|
|
|
|
bool write_enabled = !path.empty() && (((cluster->getLocalShardCount() + cluster->getRemoteShardCount()) < 2) || has_sharding_key);
|
2016-10-10 08:44:52 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
if (!write_enabled)
|
|
|
|
throw Exception{
|
|
|
|
"Method write is not supported by storage " + getName() +
|
|
|
|
" with more than one shard and no sharding key provided",
|
|
|
|
ErrorCodes::STORAGE_REQUIRES_PARAMETER};
|
2014-08-21 12:07:29 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/// DistributedBlockOutputStream will not own cluster, but will own ConnectionPools of the cluster
|
|
|
|
return std::make_shared<DistributedBlockOutputStream>(
|
2017-07-25 19:42:36 +00:00
|
|
|
*this, rewriteInsertQuery(query, remote_database, remote_table), cluster, settings.insert_distributed_sync, settings.insert_distributed_timeout);
|
2014-08-12 13:46:46 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2016-01-28 01:00:27 +00:00
|
|
|
void StorageDistributed::alter(const AlterCommands & params, const String & database_name, const String & table_name, const Context & context)
|
2013-09-23 12:01:19 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
for (const auto & param : params)
|
|
|
|
if (param.type == AlterCommand::MODIFY_PRIMARY_KEY)
|
|
|
|
throw Exception("Storage engine " + getName() + " doesn't support primary key.", ErrorCodes::NOT_IMPLEMENTED);
|
2016-05-05 18:28:46 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
auto lock = lockStructureForAlter();
|
|
|
|
params.apply(*columns, materialized_columns, alias_columns, column_defaults);
|
2016-05-13 21:08:19 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
context.getDatabase(database_name)->alterTable(
|
|
|
|
context, table_name,
|
|
|
|
*columns, materialized_columns, alias_columns, column_defaults, {});
|
2013-09-23 12:01:19 +00:00
|
|
|
}
|
2014-02-04 15:44:15 +00:00
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2017-06-06 17:06:14 +00:00
|
|
|
void StorageDistributed::startup()
|
|
|
|
{
|
|
|
|
createDirectoryMonitors();
|
|
|
|
initializeFileNamesIncrement(path, file_names_increment);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-08-13 11:26:13 +00:00
|
|
|
void StorageDistributed::shutdown()
|
|
|
|
{
|
2017-07-27 15:24:39 +00:00
|
|
|
cluster_nodes_data.clear();
|
2014-08-13 11:26:13 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2017-05-21 22:25:25 +00:00
|
|
|
void StorageDistributed::reshardPartitions(
|
|
|
|
const ASTPtr & query, const String & database_name,
|
2017-07-10 03:28:12 +00:00
|
|
|
const Field & partition,
|
2017-04-01 07:20:54 +00:00
|
|
|
const WeightedZooKeeperPaths & weighted_zookeeper_paths,
|
|
|
|
const ASTPtr & sharding_key_expr, bool do_copy, const Field & coordinator,
|
2017-05-21 22:25:25 +00:00
|
|
|
Context & context)
|
2015-10-12 14:53:16 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
auto & resharding_worker = context.getReshardingWorker();
|
|
|
|
if (!resharding_worker.isStarted())
|
|
|
|
throw Exception{"Resharding background thread is not running", ErrorCodes::RESHARDING_NO_WORKER};
|
|
|
|
|
|
|
|
if (!coordinator.isNull())
|
|
|
|
throw Exception{"Use of COORDINATE WITH is forbidden in ALTER TABLE ... RESHARD"
|
|
|
|
" queries for distributed tables",
|
|
|
|
ErrorCodes::RESHARDING_INVALID_PARAMETERS};
|
|
|
|
|
|
|
|
auto cluster = getCluster();
|
|
|
|
|
|
|
|
/// resharding_worker doesn't need to own cluster, here only meta-information of cluster is used
|
|
|
|
std::string coordinator_id = resharding_worker.createCoordinator(*cluster);
|
|
|
|
|
|
|
|
std::atomic<bool> has_notified_error{false};
|
|
|
|
|
|
|
|
std::string dumped_coordinator_state;
|
|
|
|
|
|
|
|
auto handle_exception = [&](const std::string & msg = "")
|
|
|
|
{
|
|
|
|
try
|
|
|
|
{
|
|
|
|
if (!has_notified_error)
|
|
|
|
resharding_worker.setStatus(coordinator_id, ReshardingWorker::STATUS_ERROR, msg);
|
|
|
|
dumped_coordinator_state = resharding_worker.dumpCoordinatorState(coordinator_id);
|
|
|
|
resharding_worker.deleteCoordinator(coordinator_id);
|
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
try
|
|
|
|
{
|
|
|
|
/// Create query ALTER TABLE ... RESHARD [COPY] PARTITION ... COORDINATE WITH ...
|
|
|
|
|
|
|
|
ASTPtr alter_query_ptr = std::make_shared<ASTAlterQuery>();
|
|
|
|
auto & alter_query = static_cast<ASTAlterQuery &>(*alter_query_ptr);
|
|
|
|
|
|
|
|
alter_query.database = remote_database;
|
|
|
|
alter_query.table = remote_table;
|
|
|
|
|
|
|
|
alter_query.parameters.emplace_back();
|
|
|
|
ASTAlterQuery::Parameters & parameters = alter_query.parameters.back();
|
|
|
|
|
|
|
|
parameters.type = ASTAlterQuery::RESHARD_PARTITION;
|
2017-07-10 03:28:12 +00:00
|
|
|
if (!partition.isNull())
|
|
|
|
parameters.partition = std::make_shared<ASTLiteral>(StringRange(), partition);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
ASTPtr expr_list = std::make_shared<ASTExpressionList>();
|
|
|
|
for (const auto & entry : weighted_zookeeper_paths)
|
|
|
|
{
|
|
|
|
ASTPtr weighted_path_ptr = std::make_shared<ASTWeightedZooKeeperPath>();
|
|
|
|
auto & weighted_path = static_cast<ASTWeightedZooKeeperPath &>(*weighted_path_ptr);
|
|
|
|
weighted_path.path = entry.first;
|
|
|
|
weighted_path.weight = entry.second;
|
|
|
|
expr_list->children.push_back(weighted_path_ptr);
|
|
|
|
}
|
|
|
|
|
|
|
|
parameters.weighted_zookeeper_paths = expr_list;
|
|
|
|
parameters.sharding_key_expr = sharding_key_expr;
|
|
|
|
parameters.do_copy = do_copy;
|
|
|
|
parameters.coordinator = std::make_shared<ASTLiteral>(StringRange(), Field(coordinator_id));
|
|
|
|
|
|
|
|
resharding_worker.registerQuery(coordinator_id, queryToString(alter_query_ptr));
|
|
|
|
|
|
|
|
/** The functionality of shard_multiplexing is not completed - turn it off.
|
|
|
|
* (Because connecting to different shards within a single thread is not done in parallel.)
|
|
|
|
*/
|
|
|
|
bool enable_shard_multiplexing = false;
|
|
|
|
|
|
|
|
ClusterProxy::AlterQueryConstructor alter_query_constructor;
|
|
|
|
|
|
|
|
BlockInputStreams streams = ClusterProxy::Query{alter_query_constructor, cluster, alter_query_ptr,
|
2017-05-21 22:25:25 +00:00
|
|
|
context, context.getSettingsRef(), enable_shard_multiplexing}.execute();
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
/// This callback is called if an exception has occurred while attempting to read
|
|
|
|
/// a block from a shard. This is to avoid a potential deadlock if other shards are
|
|
|
|
/// waiting inside a barrier. Actually, even without this solution, we would avoid
|
|
|
|
/// such a deadlock because we would eventually time out while trying to get remote
|
|
|
|
/// blocks. Nevertheless this is not the ideal way of sorting out this issue since
|
|
|
|
/// we would then not get to know the actual cause of the failure.
|
|
|
|
auto exception_callback = [&resharding_worker, coordinator_id, &has_notified_error]()
|
|
|
|
{
|
|
|
|
try
|
|
|
|
{
|
|
|
|
resharding_worker.setStatus(coordinator_id, ReshardingWorker::STATUS_ERROR);
|
|
|
|
has_notified_error = true;
|
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
streams[0] = std::make_shared<UnionBlockInputStream<>>(
|
2017-05-21 22:25:25 +00:00
|
|
|
streams, nullptr, context.getSettingsRef().max_distributed_connections, exception_callback);
|
2017-04-01 07:20:54 +00:00
|
|
|
streams.resize(1);
|
|
|
|
|
|
|
|
auto stream_ptr = dynamic_cast<IProfilingBlockInputStream *>(&*streams[0]);
|
|
|
|
if (stream_ptr == nullptr)
|
|
|
|
throw Exception{"StorageDistributed: Internal error", ErrorCodes::LOGICAL_ERROR};
|
|
|
|
auto & stream = *stream_ptr;
|
|
|
|
|
|
|
|
stream.readPrefix();
|
|
|
|
|
|
|
|
while (!stream.isCancelled() && stream.read())
|
|
|
|
;
|
|
|
|
|
|
|
|
if (!stream.isCancelled())
|
|
|
|
stream.readSuffix();
|
|
|
|
}
|
|
|
|
catch (const Exception & ex)
|
|
|
|
{
|
|
|
|
handle_exception(ex.message());
|
|
|
|
LOG_ERROR(log, dumped_coordinator_state);
|
|
|
|
throw;
|
|
|
|
}
|
|
|
|
catch (const std::exception & ex)
|
|
|
|
{
|
|
|
|
handle_exception(ex.what());
|
|
|
|
LOG_ERROR(log, dumped_coordinator_state);
|
|
|
|
throw;
|
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
handle_exception();
|
|
|
|
LOG_ERROR(log, dumped_coordinator_state);
|
|
|
|
throw;
|
|
|
|
}
|
2016-01-28 01:00:27 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2016-01-28 01:00:27 +00:00
|
|
|
BlockInputStreams StorageDistributed::describe(const Context & context, const Settings & settings)
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
/// Create DESCRIBE TABLE query.
|
|
|
|
auto cluster = getCluster();
|
2016-01-28 01:00:27 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
ASTPtr describe_query_ptr = std::make_shared<ASTDescribeQuery>();
|
|
|
|
auto & describe_query = static_cast<ASTDescribeQuery &>(*describe_query_ptr);
|
2016-01-28 01:00:27 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
describe_query.database = remote_database;
|
|
|
|
describe_query.table = remote_table;
|
2016-01-28 01:00:27 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/** The functionality of shard_multiplexing is not completed - turn it off.
|
|
|
|
* (Because connecting connections to different shards within a single thread is not done in parallel.)
|
|
|
|
*/
|
|
|
|
bool enable_shard_multiplexing = false;
|
2016-01-28 01:00:27 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
ClusterProxy::DescribeQueryConstructor describe_query_constructor;
|
2015-10-12 14:53:16 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
return ClusterProxy::Query{describe_query_constructor, cluster, describe_query_ptr,
|
|
|
|
context, settings, enable_shard_multiplexing}.execute();
|
2015-10-12 14:53:16 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2014-07-31 09:11:49 +00:00
|
|
|
NameAndTypePair StorageDistributed::getColumn(const String & column_name) const
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
if (const auto & type = VirtualColumnFactory::tryGetType(column_name))
|
|
|
|
return { column_name, type };
|
2014-07-31 09:11:49 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
return getRealColumn(column_name);
|
2014-07-31 09:11:49 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2014-07-31 09:11:49 +00:00
|
|
|
bool StorageDistributed::hasColumn(const String & column_name) const
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
return VirtualColumnFactory::hasColumn(column_name) || IStorage::hasColumn(column_name);
|
2014-07-31 09:11:49 +00:00
|
|
|
}
|
|
|
|
|
2014-08-15 09:50:05 +00:00
|
|
|
void StorageDistributed::createDirectoryMonitors()
|
2014-08-13 11:26:13 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
if (path.empty())
|
|
|
|
return;
|
2014-12-30 03:53:41 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
Poco::File{path}.createDirectory();
|
2014-08-14 11:50:36 +00:00
|
|
|
|
2017-05-10 06:39:37 +00:00
|
|
|
boost::filesystem::directory_iterator begin(path);
|
|
|
|
boost::filesystem::directory_iterator end;
|
|
|
|
for (auto it = begin; it != end; ++it)
|
|
|
|
if (it->status().type() == boost::filesystem::directory_file)
|
2017-07-27 15:24:39 +00:00
|
|
|
requireDirectoryMonitor(it->path().filename().string());
|
2014-08-13 11:26:13 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2017-07-27 15:24:39 +00:00
|
|
|
void StorageDistributed::requireDirectoryMonitor(const std::string & name)
|
|
|
|
{
|
|
|
|
cluster_nodes_data[name].requireDirectoryMonitor(name, *this);
|
|
|
|
}
|
|
|
|
|
|
|
|
ConnectionPoolPtr StorageDistributed::requireConnectionPool(const std::string & name)
|
2014-08-19 08:04:13 +00:00
|
|
|
{
|
2017-07-27 15:24:39 +00:00
|
|
|
auto & node_data = cluster_nodes_data[name];
|
|
|
|
node_data.requireConnectionPool(name, *this);
|
|
|
|
return node_data.conneciton_pool;
|
2014-08-19 08:04:13 +00:00
|
|
|
}
|
|
|
|
|
2015-09-18 13:36:10 +00:00
|
|
|
size_t StorageDistributed::getShardCount() const
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
return getCluster()->getRemoteShardCount();
|
2016-10-10 08:44:52 +00:00
|
|
|
}
|
|
|
|
|
2017-03-09 03:34:09 +00:00
|
|
|
|
2016-10-10 08:44:52 +00:00
|
|
|
ClusterPtr StorageDistributed::getCluster() const
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
return (owned_cluster) ? owned_cluster : context.getCluster(cluster_name);
|
2015-09-18 13:36:10 +00:00
|
|
|
}
|
|
|
|
|
2017-07-27 15:24:39 +00:00
|
|
|
void StorageDistributed::ClusterNodeData::requireConnectionPool(const std::string & name, const StorageDistributed & storage)
|
|
|
|
{
|
|
|
|
if (!conneciton_pool)
|
|
|
|
conneciton_pool = StorageDistributedDirectoryMonitor::createPool(name, storage);
|
|
|
|
}
|
|
|
|
|
|
|
|
void StorageDistributed::ClusterNodeData::requireDirectoryMonitor(const std::string & name, StorageDistributed & storage)
|
|
|
|
{
|
|
|
|
requireConnectionPool(name, storage);
|
|
|
|
if (!directory_monitor)
|
|
|
|
directory_monitor = std::make_unique<StorageDistributedDirectoryMonitor>(storage, name, conneciton_pool);
|
|
|
|
}
|
|
|
|
|
2012-05-21 20:38:34 +00:00
|
|
|
}
|