mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-15 19:02:04 +00:00
1187 lines
43 KiB
C++
1187 lines
43 KiB
C++
#include <Interpreters/DDLWorker.h>
|
|
#include <Interpreters/DDLTask.h>
|
|
#include <Parsers/ASTAlterQuery.h>
|
|
#include <Parsers/ASTDropQuery.h>
|
|
#include <Parsers/ASTOptimizeQuery.h>
|
|
#include <Parsers/ASTQueryWithOnCluster.h>
|
|
#include <Parsers/ASTQueryWithTableAndOutput.h>
|
|
#include <Parsers/ParserQuery.h>
|
|
#include <Parsers/parseQuery.h>
|
|
#include <Parsers/queryToString.h>
|
|
#include <IO/WriteHelpers.h>
|
|
#include <IO/ReadHelpers.h>
|
|
#include <IO/ReadBufferFromString.h>
|
|
#include <Storages/IStorage.h>
|
|
#include <Interpreters/executeQuery.h>
|
|
#include <Interpreters/Cluster.h>
|
|
#include <Interpreters/Context.h>
|
|
#include <Common/setThreadName.h>
|
|
#include <Common/randomSeed.h>
|
|
#include <Common/ZooKeeper/ZooKeeper.h>
|
|
#include <Common/ZooKeeper/KeeperException.h>
|
|
#include <Common/isLocalAddress.h>
|
|
#include <Storages/StorageReplicatedMergeTree.h>
|
|
#include <Poco/Timestamp.h>
|
|
#include <common/sleep.h>
|
|
#include <common/getFQDNOrHostName.h>
|
|
#include <common/logger_useful.h>
|
|
#include <random>
|
|
#include <pcg_random.hpp>
|
|
|
|
|
|
namespace DB
|
|
{
|
|
|
|
namespace ErrorCodes
|
|
{
|
|
extern const int NOT_IMPLEMENTED;
|
|
extern const int LOGICAL_ERROR;
|
|
extern const int INCONSISTENT_CLUSTER_DEFINITION;
|
|
extern const int TIMEOUT_EXCEEDED;
|
|
extern const int UNKNOWN_TYPE_OF_QUERY;
|
|
extern const int UNFINISHED;
|
|
extern const int QUERY_IS_PROHIBITED;
|
|
}
|
|
|
|
|
|
namespace
|
|
{
|
|
|
|
/** Caveats: usage of locks in ZooKeeper is incorrect in 99% of cases,
|
|
* and highlights your poor understanding of distributed systems.
|
|
*
|
|
* It's only correct if all the operations that are performed under lock
|
|
* are atomically checking that the lock still holds
|
|
* or if we ensure that these operations will be undone if lock is lost
|
|
* (due to ZooKeeper session loss) that's very difficult to achieve.
|
|
*
|
|
* It's Ok if every operation that we perform under lock is actually operation in ZooKeeper.
|
|
*
|
|
* In 1% of cases when you can correctly use Lock, the logic is complex enough, so you don't need this class.
|
|
*
|
|
* TLDR: Don't use this code.
|
|
* We only have a few cases of it's usage and it will be removed.
|
|
*/
|
|
class ZooKeeperLock
|
|
{
|
|
public:
|
|
/// lock_prefix - path where the ephemeral lock node will be created
|
|
/// lock_name - the name of the ephemeral lock node
|
|
ZooKeeperLock(
|
|
const zkutil::ZooKeeperPtr & zookeeper_,
|
|
const std::string & lock_prefix_,
|
|
const std::string & lock_name_,
|
|
const std::string & lock_message_ = "")
|
|
:
|
|
zookeeper(zookeeper_),
|
|
lock_path(lock_prefix_ + "/" + lock_name_),
|
|
lock_message(lock_message_),
|
|
log(&Poco::Logger::get("zkutil::Lock"))
|
|
{
|
|
zookeeper->createIfNotExists(lock_prefix_, "");
|
|
}
|
|
|
|
~ZooKeeperLock()
|
|
{
|
|
try
|
|
{
|
|
unlock();
|
|
}
|
|
catch (...)
|
|
{
|
|
DB::tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
}
|
|
}
|
|
|
|
void unlock()
|
|
{
|
|
Coordination::Stat stat;
|
|
std::string dummy;
|
|
bool result = zookeeper->tryGet(lock_path, dummy, &stat);
|
|
|
|
if (result && stat.ephemeralOwner == zookeeper->getClientID())
|
|
zookeeper->remove(lock_path, -1);
|
|
else
|
|
LOG_WARNING(log, "Lock is lost. It is normal if session was expired. Path: {}/{}", lock_path, lock_message);
|
|
}
|
|
|
|
bool tryLock()
|
|
{
|
|
std::string dummy;
|
|
Coordination::Error code = zookeeper->tryCreate(lock_path, lock_message, zkutil::CreateMode::Ephemeral, dummy);
|
|
|
|
if (code == Coordination::Error::ZNODEEXISTS)
|
|
{
|
|
return false;
|
|
}
|
|
else if (code == Coordination::Error::ZOK)
|
|
{
|
|
return true;
|
|
}
|
|
else
|
|
{
|
|
throw Coordination::Exception(code);
|
|
}
|
|
}
|
|
|
|
private:
|
|
zkutil::ZooKeeperPtr zookeeper;
|
|
|
|
std::string lock_path;
|
|
std::string lock_message;
|
|
Poco::Logger * log;
|
|
|
|
};
|
|
|
|
std::unique_ptr<ZooKeeperLock> createSimpleZooKeeperLock(
|
|
const zkutil::ZooKeeperPtr & zookeeper, const String & lock_prefix, const String & lock_name, const String & lock_message)
|
|
{
|
|
return std::make_unique<ZooKeeperLock>(zookeeper, lock_prefix, lock_name, lock_message);
|
|
}
|
|
|
|
}
|
|
|
|
|
|
String DatabaseReplicatedExtensions::getLogEntryName(UInt32 log_entry_number)
|
|
{
|
|
constexpr size_t seq_node_digits = 10;
|
|
String number = toString(log_entry_number);
|
|
String name = "query-" + String(seq_node_digits - number.size(), '0') + number;
|
|
return name;
|
|
}
|
|
|
|
UInt32 DatabaseReplicatedExtensions::getLogEntryNumber(const String & log_entry_name)
|
|
{
|
|
constexpr const char * name = "query-";
|
|
assert(startsWith(log_entry_name, name));
|
|
return parse<UInt32>(log_entry_name.substr(strlen(name)));
|
|
}
|
|
|
|
|
|
DDLWorker::DDLWorker(int pool_size_, const std::string & zk_root_dir, const Context & context_, const Poco::Util::AbstractConfiguration * config, const String & prefix,
|
|
std::optional<DatabaseReplicatedExtensions> database_replicated_ext_)
|
|
: context(context_)
|
|
, log(&Poco::Logger::get(database_replicated_ext_ ? fmt::format("DDLWorker ({})", database_replicated_ext_->database_name) : "DDLWorker"))
|
|
, database_replicated_ext(std::move(database_replicated_ext_))
|
|
, pool_size(pool_size_)
|
|
, worker_pool(pool_size_)
|
|
{
|
|
assert(!database_replicated_ext || pool_size == 1);
|
|
last_tasks.reserve(pool_size);
|
|
|
|
queue_dir = zk_root_dir;
|
|
if (queue_dir.back() == '/')
|
|
queue_dir.resize(queue_dir.size() - 1);
|
|
|
|
if (config)
|
|
{
|
|
task_max_lifetime = config->getUInt64(prefix + ".task_max_lifetime", static_cast<UInt64>(task_max_lifetime));
|
|
cleanup_delay_period = config->getUInt64(prefix + ".cleanup_delay_period", static_cast<UInt64>(cleanup_delay_period));
|
|
max_tasks_in_queue = std::max<UInt64>(1, config->getUInt64(prefix + ".max_tasks_in_queue", max_tasks_in_queue));
|
|
|
|
if (config->has(prefix + ".profile"))
|
|
context.setSetting("profile", config->getString(prefix + ".profile"));
|
|
}
|
|
|
|
if (context.getSettingsRef().readonly)
|
|
{
|
|
LOG_WARNING(log, "Distributed DDL worker is run with readonly settings, it will not be able to execute DDL queries Set appropriate system_profile or distributed_ddl.profile to fix this.");
|
|
}
|
|
|
|
host_fqdn = getFQDNOrHostName();
|
|
host_fqdn_id = Cluster::Address::toString(host_fqdn, context.getTCPPort());
|
|
|
|
main_thread = ThreadFromGlobalPool(&DDLWorker::runMainThread, this);
|
|
cleanup_thread = ThreadFromGlobalPool(&DDLWorker::runCleanupThread, this);
|
|
}
|
|
|
|
void DDLWorker::shutdown()
|
|
{
|
|
stop_flag = true;
|
|
queue_updated_event->set();
|
|
cleanup_event->set();
|
|
}
|
|
|
|
DDLWorker::~DDLWorker()
|
|
{
|
|
shutdown();
|
|
worker_pool.wait();
|
|
main_thread.join();
|
|
cleanup_thread.join();
|
|
}
|
|
|
|
|
|
ZooKeeperPtr DDLWorker::tryGetZooKeeper() const
|
|
{
|
|
std::lock_guard lock(zookeeper_mutex);
|
|
return current_zookeeper;
|
|
}
|
|
|
|
ZooKeeperPtr DDLWorker::getAndSetZooKeeper()
|
|
{
|
|
std::lock_guard lock(zookeeper_mutex);
|
|
|
|
if (!current_zookeeper || current_zookeeper->expired())
|
|
current_zookeeper = context.getZooKeeper();
|
|
|
|
return current_zookeeper;
|
|
}
|
|
|
|
void DDLWorker::recoverZooKeeper()
|
|
{
|
|
LOG_DEBUG(log, "Recovering ZooKeeper session after: {}", getCurrentExceptionMessage(false));
|
|
|
|
while (!stop_flag)
|
|
{
|
|
try
|
|
{
|
|
getAndSetZooKeeper();
|
|
break;
|
|
}
|
|
catch (...)
|
|
{
|
|
tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
sleepForSeconds(5);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
DDLTaskPtr DDLWorker::initAndCheckTask(const String & entry_name, String & out_reason, const ZooKeeperPtr & zookeeper)
|
|
{
|
|
String node_data;
|
|
String entry_path = queue_dir + "/" + entry_name;
|
|
|
|
auto task = std::make_unique<DDLTask>();
|
|
task->entry_name = entry_name;
|
|
task->entry_path = entry_path;
|
|
|
|
if (database_replicated_ext)
|
|
{
|
|
String initiator_name;
|
|
zkutil::EventPtr wait_committed_or_failed = std::make_shared<Poco::Event>();
|
|
|
|
if (zookeeper->tryGet(entry_path + "/try", initiator_name, nullptr, wait_committed_or_failed))
|
|
{
|
|
task->we_are_initiator = initiator_name == database_replicated_ext->getFullReplicaName();
|
|
/// Query is not committed yet. We cannot just skip it and execute next one, because reordering may break replication.
|
|
//FIXME add some timeouts
|
|
if (!task->we_are_initiator)
|
|
{
|
|
LOG_TRACE(log, "Waiting for initiator {} to commit or rollback entry {}", initiator_name, entry_path);
|
|
wait_committed_or_failed->wait();
|
|
}
|
|
}
|
|
|
|
if (!task->we_are_initiator && !zookeeper->exists(entry_path + "/committed"))
|
|
{
|
|
out_reason = "Entry " + entry_name + " hasn't been committed";
|
|
return {};
|
|
}
|
|
}
|
|
|
|
if (!zookeeper->tryGet(entry_path, node_data))
|
|
{
|
|
if (database_replicated_ext)
|
|
database_replicated_ext->lost_callback(entry_name, zookeeper);
|
|
/// It is Ok that node could be deleted just now. It means that there are no current host in node's host list.
|
|
out_reason = "The task was deleted";
|
|
return {};
|
|
}
|
|
|
|
try
|
|
{
|
|
task->entry.parse(node_data);
|
|
}
|
|
catch (...)
|
|
{
|
|
/// What should we do if we even cannot parse host name and therefore cannot properly submit execution status?
|
|
/// We can try to create fail node using FQDN if it equal to host name in cluster config attempt will be successful.
|
|
/// Otherwise, that node will be ignored by DDLQueryStatusInputStream.
|
|
|
|
tryLogCurrentException(log, "Cannot parse DDL task " + entry_name + ", will try to send error status");
|
|
|
|
String status = ExecutionStatus::fromCurrentException().serializeText();
|
|
try
|
|
{
|
|
createStatusDirs(entry_path, zookeeper);
|
|
zookeeper->tryCreate(entry_path + "/finished/" + host_fqdn_id, status, zkutil::CreateMode::Persistent);
|
|
}
|
|
catch (...)
|
|
{
|
|
tryLogCurrentException(log, "Can't report the task has invalid format");
|
|
}
|
|
|
|
out_reason = "Incorrect task format";
|
|
return {};
|
|
}
|
|
|
|
if (database_replicated_ext)
|
|
{
|
|
task->host_id.host_name = host_fqdn;
|
|
task->host_id.port = context.getTCPPort();
|
|
task->host_id_str = database_replicated_ext->shard_name + '|' + database_replicated_ext->replica_name;
|
|
return task;
|
|
}
|
|
|
|
bool host_in_hostlist = false;
|
|
for (const HostID & host : task->entry.hosts)
|
|
{
|
|
auto maybe_secure_port = context.getTCPPortSecure();
|
|
|
|
/// The port is considered local if it matches TCP or TCP secure port that the server is listening.
|
|
bool is_local_port = (maybe_secure_port && host.isLocalAddress(*maybe_secure_port))
|
|
|| host.isLocalAddress(context.getTCPPort());
|
|
|
|
if (!is_local_port)
|
|
continue;
|
|
|
|
if (host_in_hostlist)
|
|
{
|
|
/// This check could be slow a little bit
|
|
LOG_WARNING(log, "There are two the same ClickHouse instances in task {}: {} and {}. Will use the first one only.", entry_name, task->host_id.readableString(), host.readableString());
|
|
}
|
|
else
|
|
{
|
|
host_in_hostlist = true;
|
|
task->host_id = host;
|
|
task->host_id_str = host.toString();
|
|
}
|
|
}
|
|
|
|
if (!host_in_hostlist)
|
|
{
|
|
out_reason = "There is no a local address in host list";
|
|
return {};
|
|
}
|
|
|
|
return task;
|
|
}
|
|
|
|
|
|
static void filterAndSortQueueNodes(Strings & all_nodes)
|
|
{
|
|
all_nodes.erase(std::remove_if(all_nodes.begin(), all_nodes.end(), [] (const String & s) { return !startsWith(s, "query-"); }), all_nodes.end());
|
|
std::sort(all_nodes.begin(), all_nodes.end());
|
|
}
|
|
|
|
void DDLWorker::scheduleTasks()
|
|
{
|
|
LOG_DEBUG(log, "Scheduling tasks");
|
|
auto zookeeper = tryGetZooKeeper();
|
|
|
|
Strings queue_nodes = zookeeper->getChildren(queue_dir, nullptr, queue_updated_event);
|
|
filterAndSortQueueNodes(queue_nodes);
|
|
if (queue_nodes.empty())
|
|
{
|
|
LOG_TRACE(log, "No tasks to schedule");
|
|
return;
|
|
}
|
|
|
|
bool server_startup = last_tasks.empty();
|
|
|
|
auto begin_node = server_startup
|
|
? queue_nodes.begin()
|
|
: std::upper_bound(queue_nodes.begin(), queue_nodes.end(), last_tasks.back());
|
|
|
|
for (auto it = begin_node; it != queue_nodes.end() && !stop_flag; ++it)
|
|
{
|
|
String entry_name = *it;
|
|
LOG_TRACE(log, "Checking task {}", entry_name);
|
|
|
|
String reason;
|
|
auto task = initAndCheckTask(entry_name, reason, zookeeper);
|
|
if (!task)
|
|
{
|
|
LOG_DEBUG(log, "Will not execute task {}: {}", entry_name, reason);
|
|
saveTask(entry_name);
|
|
continue;
|
|
}
|
|
|
|
bool already_processed = zookeeper->exists(task->entry_path + "/finished/" + task->host_id_str);
|
|
if (!server_startup && !task->was_executed && already_processed)
|
|
{
|
|
throw Exception(ErrorCodes::LOGICAL_ERROR,
|
|
"Server expects that DDL task {} should be processed, but it was already processed according to ZK",
|
|
entry_name);
|
|
}
|
|
|
|
if (!already_processed)
|
|
{
|
|
if (database_replicated_ext)
|
|
{
|
|
enqueueTask(DDLTaskPtr(task.release()));
|
|
}
|
|
else
|
|
{
|
|
worker_pool.scheduleOrThrowOnError([this, task_ptr = task.release()]() {
|
|
setThreadName("DDLWorkerExec");
|
|
enqueueTask(DDLTaskPtr(task_ptr));
|
|
});
|
|
}
|
|
}
|
|
else
|
|
{
|
|
LOG_DEBUG(log, "Task {} ({}) has been already processed", entry_name, task->entry.query);
|
|
}
|
|
|
|
saveTask(entry_name);
|
|
}
|
|
}
|
|
|
|
void DDLWorker::saveTask(const String & entry_name)
|
|
{
|
|
if (last_tasks.size() == pool_size)
|
|
{
|
|
last_tasks.erase(last_tasks.begin());
|
|
}
|
|
last_tasks.emplace_back(entry_name);
|
|
}
|
|
|
|
/// Parses query and resolves cluster and host in cluster
|
|
void DDLWorker::parseQueryAndResolveHost(DDLTask & task)
|
|
{
|
|
{
|
|
const char * begin = task.entry.query.data();
|
|
const char * end = begin + task.entry.query.size();
|
|
|
|
ParserQuery parser_query(end);
|
|
String description;
|
|
task.query = parseQuery(parser_query, begin, end, description, 0, context.getSettingsRef().max_parser_depth);
|
|
}
|
|
|
|
// XXX: serious design flaw since `ASTQueryWithOnCluster` is not inherited from `IAST`!
|
|
if (!task.query || !(task.query_on_cluster = dynamic_cast<ASTQueryWithOnCluster *>(task.query.get())))
|
|
throw Exception("Received unknown DDL query", ErrorCodes::UNKNOWN_TYPE_OF_QUERY);
|
|
|
|
if (database_replicated_ext)
|
|
return;
|
|
|
|
task.cluster_name = task.query_on_cluster->cluster;
|
|
task.cluster = context.tryGetCluster(task.cluster_name);
|
|
if (!task.cluster)
|
|
throw Exception(ErrorCodes::INCONSISTENT_CLUSTER_DEFINITION,
|
|
"DDL task {} contains current host {} in cluster {}, but there are no such cluster here.",
|
|
task.entry_name, task.host_id.readableString(), task.cluster_name);
|
|
|
|
/// Try to find host from task host list in cluster
|
|
/// At the first, try find exact match (host name and ports should be literally equal)
|
|
/// If the attempt fails, try find it resolving host name of each instance
|
|
const auto & shards = task.cluster->getShardsAddresses();
|
|
|
|
bool found_exact_match = false;
|
|
String default_database;
|
|
for (size_t shard_num = 0; shard_num < shards.size(); ++shard_num)
|
|
{
|
|
for (size_t replica_num = 0; replica_num < shards[shard_num].size(); ++replica_num)
|
|
{
|
|
const Cluster::Address & address = shards[shard_num][replica_num];
|
|
|
|
if (address.host_name == task.host_id.host_name && address.port == task.host_id.port)
|
|
{
|
|
if (found_exact_match)
|
|
{
|
|
if (default_database == address.default_database)
|
|
{
|
|
throw Exception(ErrorCodes::INCONSISTENT_CLUSTER_DEFINITION,
|
|
"There are two exactly the same ClickHouse instances {} in cluster {}",
|
|
address.readableString(), task.cluster_name);
|
|
}
|
|
else
|
|
{
|
|
/* Circular replication is used.
|
|
* It is when every physical node contains
|
|
* replicas of different shards of the same table.
|
|
* To distinguish one replica from another on the same node,
|
|
* every shard is placed into separate database.
|
|
* */
|
|
is_circular_replicated = true;
|
|
auto * query_with_table = dynamic_cast<ASTQueryWithTableAndOutput *>(task.query.get());
|
|
if (!query_with_table || query_with_table->database.empty())
|
|
{
|
|
throw Exception(ErrorCodes::INCONSISTENT_CLUSTER_DEFINITION,
|
|
"For a distributed DDL on circular replicated cluster its table name must be qualified by database name.");
|
|
}
|
|
if (default_database == query_with_table->database)
|
|
return;
|
|
}
|
|
}
|
|
found_exact_match = true;
|
|
task.host_shard_num = shard_num;
|
|
task.host_replica_num = replica_num;
|
|
task.address_in_cluster = address;
|
|
default_database = address.default_database;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (found_exact_match)
|
|
return;
|
|
|
|
LOG_WARNING(log, "Not found the exact match of host {} from task {} in cluster {} definition. Will try to find it using host name resolving.", task.host_id.readableString(), task.entry_name, task.cluster_name);
|
|
|
|
bool found_via_resolving = false;
|
|
for (size_t shard_num = 0; shard_num < shards.size(); ++shard_num)
|
|
{
|
|
for (size_t replica_num = 0; replica_num < shards[shard_num].size(); ++replica_num)
|
|
{
|
|
const Cluster::Address & address = shards[shard_num][replica_num];
|
|
|
|
if (auto resolved = address.getResolvedAddress();
|
|
resolved && (isLocalAddress(*resolved, context.getTCPPort())
|
|
|| (context.getTCPPortSecure() && isLocalAddress(*resolved, *context.getTCPPortSecure()))))
|
|
{
|
|
if (found_via_resolving)
|
|
{
|
|
throw Exception(ErrorCodes::INCONSISTENT_CLUSTER_DEFINITION,
|
|
"There are two the same ClickHouse instances in cluster {} : {} and {}",
|
|
task.cluster_name, task.address_in_cluster.readableString(), address.readableString());
|
|
}
|
|
else
|
|
{
|
|
found_via_resolving = true;
|
|
task.host_shard_num = shard_num;
|
|
task.host_replica_num = replica_num;
|
|
task.address_in_cluster = address;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!found_via_resolving)
|
|
{
|
|
throw Exception(ErrorCodes::INCONSISTENT_CLUSTER_DEFINITION,
|
|
"Not found host {} in definition of cluster {}",
|
|
task.host_id.readableString(), task.cluster_name);
|
|
}
|
|
else
|
|
{
|
|
LOG_INFO(log, "Resolved host {} from task {} as host {} in definition of cluster {}", task.host_id.readableString(), task.entry_name, task.address_in_cluster.readableString(), task.cluster_name);
|
|
}
|
|
}
|
|
|
|
|
|
bool DDLWorker::tryExecuteQuery(const String & query, const DDLTask & task, ExecutionStatus & status)
|
|
{
|
|
/// Add special comment at the start of query to easily identify DDL-produced queries in query_log
|
|
String query_prefix = "/* ddl_entry=" + task.entry_name + " */ ";
|
|
String query_to_execute = query_prefix + query;
|
|
|
|
ReadBufferFromString istr(query_to_execute);
|
|
String dummy_string;
|
|
WriteBufferFromString ostr(dummy_string);
|
|
|
|
try
|
|
{
|
|
auto current_context = std::make_unique<Context>(context);
|
|
current_context->makeQueryContext();
|
|
current_context->setCurrentQueryId(""); // generate random query_id
|
|
|
|
if (database_replicated_ext)
|
|
{
|
|
current_context->getClientInfo().query_kind
|
|
= ClientInfo::QueryKind::REPLICATED_LOG_QUERY; //FIXME why do we need separate query kind?
|
|
current_context->setCurrentDatabase(database_replicated_ext->database_name);
|
|
|
|
if (task.we_are_initiator)
|
|
{
|
|
auto txn = std::make_shared<MetadataTransaction>();
|
|
current_context->initMetadataTransaction(txn);
|
|
txn->current_zookeeper = current_zookeeper;
|
|
txn->zookeeper_path = database_replicated_ext->zookeeper_path;
|
|
txn->ops.emplace_back(zkutil::makeRemoveRequest(task.entry_path + "/try", -1));
|
|
txn->ops.emplace_back(zkutil::makeCreateRequest(task.entry_path + "/committed",
|
|
database_replicated_ext->getFullReplicaName(), zkutil::CreateMode::Persistent));
|
|
txn->ops.emplace_back(zkutil::makeRemoveRequest(task.active_path, -1));
|
|
if (!task.shard_path.empty())
|
|
txn->ops.emplace_back(zkutil::makeCreateRequest(task.shard_path, task.host_id_str, zkutil::CreateMode::Persistent));
|
|
txn->ops.emplace_back(zkutil::makeCreateRequest(task.finished_path, task.execution_status.serializeText(), zkutil::CreateMode::Persistent));
|
|
//txn->ops.emplace_back(zkutil::makeSetRequest(database_replicated_ext->getReplicaPath() + "/log_ptr", toString(database_replicated_ext->first_not_executed), -1));
|
|
}
|
|
}
|
|
else
|
|
current_context->getClientInfo().query_kind = ClientInfo::QueryKind::SECONDARY_QUERY;
|
|
|
|
executeQuery(istr, ostr, false, *current_context, {});
|
|
}
|
|
catch (...)
|
|
{
|
|
status = ExecutionStatus::fromCurrentException();
|
|
tryLogCurrentException(log, "Query " + query + " wasn't finished successfully");
|
|
|
|
return false;
|
|
}
|
|
|
|
status = ExecutionStatus(0);
|
|
LOG_DEBUG(log, "Executed query: {}", query);
|
|
|
|
return true;
|
|
}
|
|
|
|
void DDLWorker::attachToThreadGroup()
|
|
{
|
|
if (thread_group)
|
|
{
|
|
/// Put all threads to one thread pool
|
|
CurrentThread::attachToIfDetached(thread_group);
|
|
}
|
|
else
|
|
{
|
|
CurrentThread::initializeQuery();
|
|
thread_group = CurrentThread::getGroup();
|
|
}
|
|
}
|
|
|
|
|
|
void DDLWorker::enqueueTask(DDLTaskPtr task_ptr)
|
|
{
|
|
auto & task = *task_ptr;
|
|
|
|
while (!stop_flag)
|
|
{
|
|
try
|
|
{
|
|
processTask(task);
|
|
return;
|
|
}
|
|
catch (const Coordination::Exception & e)
|
|
{
|
|
if (Coordination::isHardwareError(e.code))
|
|
{
|
|
recoverZooKeeper();
|
|
}
|
|
else if (e.code == Coordination::Error::ZNONODE)
|
|
{
|
|
LOG_ERROR(log, "ZooKeeper error: {}", getCurrentExceptionMessage(true));
|
|
// TODO: retry?
|
|
}
|
|
else
|
|
{
|
|
LOG_ERROR(log, "Unexpected ZooKeeper error: {}.", getCurrentExceptionMessage(true));
|
|
return;
|
|
}
|
|
}
|
|
catch (...)
|
|
{
|
|
LOG_WARNING(log, "An error occurred while processing task {} ({}) : {}", task.entry_name, task.entry.query, getCurrentExceptionMessage(true));
|
|
}
|
|
}
|
|
}
|
|
|
|
void DDLWorker::processTask(DDLTask & task)
|
|
{
|
|
auto zookeeper = tryGetZooKeeper();
|
|
|
|
LOG_DEBUG(log, "Processing task {} ({})", task.entry_name, task.entry.query);
|
|
|
|
String dummy;
|
|
//FIXME duplicate
|
|
String active_node_path = task.active_path = task.entry_path + "/active/" + task.host_id_str;
|
|
String finished_node_path = task.finished_path = task.entry_path + "/finished/" + task.host_id_str;
|
|
|
|
auto code = zookeeper->tryCreate(active_node_path, "", zkutil::CreateMode::Ephemeral, dummy);
|
|
|
|
if (code == Coordination::Error::ZOK || code == Coordination::Error::ZNODEEXISTS)
|
|
{
|
|
// Ok
|
|
}
|
|
else if (code == Coordination::Error::ZNONODE)
|
|
{
|
|
/// There is no parent
|
|
//TODO why not to create parent before active_node?
|
|
createStatusDirs(task.entry_path, zookeeper);
|
|
if (Coordination::Error::ZOK != zookeeper->tryCreate(active_node_path, "", zkutil::CreateMode::Ephemeral, dummy))
|
|
throw Coordination::Exception(code, active_node_path);
|
|
}
|
|
else
|
|
throw Coordination::Exception(code, active_node_path);
|
|
|
|
//FIXME
|
|
bool is_dummy_query = database_replicated_ext && task.entry.query.empty();
|
|
if (!task.was_executed && !is_dummy_query)
|
|
{
|
|
try
|
|
{
|
|
is_circular_replicated = false;
|
|
parseQueryAndResolveHost(task);
|
|
|
|
ASTPtr rewritten_ast = task.query_on_cluster->getRewrittenASTWithoutOnCluster(task.address_in_cluster.default_database);
|
|
String rewritten_query = queryToString(rewritten_ast);
|
|
LOG_DEBUG(log, "Executing query: {}", rewritten_query);
|
|
|
|
if (auto * query_with_table = dynamic_cast<ASTQueryWithTableAndOutput *>(rewritten_ast.get()); query_with_table)
|
|
{
|
|
StoragePtr storage;
|
|
if (!query_with_table->table.empty())
|
|
{
|
|
/// It's not CREATE DATABASE
|
|
auto table_id = context.tryResolveStorageID(*query_with_table, Context::ResolveOrdinary);
|
|
storage = DatabaseCatalog::instance().tryGetTable(table_id, context);
|
|
}
|
|
|
|
if (storage && taskShouldBeExecutedOnLeader(rewritten_ast, storage) && !is_circular_replicated)
|
|
tryExecuteQueryOnLeaderReplica(task, storage, rewritten_query, task.entry_path, zookeeper);
|
|
else
|
|
tryExecuteQuery(rewritten_query, task, task.execution_status);
|
|
}
|
|
else
|
|
tryExecuteQuery(rewritten_query, task, task.execution_status);
|
|
}
|
|
catch (const Coordination::Exception &)
|
|
{
|
|
throw;
|
|
}
|
|
catch (...)
|
|
{
|
|
tryLogCurrentException(log, "An error occurred before execution of DDL task: ");
|
|
task.execution_status = ExecutionStatus::fromCurrentException("An error occurred before execution");
|
|
}
|
|
|
|
/// We need to distinguish ZK errors occurred before and after query executing
|
|
task.was_executed = true;
|
|
}
|
|
|
|
/// FIXME: if server fails right here, the task will be executed twice. We need WAL here.
|
|
|
|
/// Delete active flag and create finish flag
|
|
Coordination::Requests ops;
|
|
ops.emplace_back(zkutil::makeRemoveRequest(active_node_path, -1));
|
|
ops.emplace_back(zkutil::makeCreateRequest(finished_node_path, task.execution_status.serializeText(), zkutil::CreateMode::Persistent));
|
|
|
|
//FIXME replace with multi(...) or use MetadataTransaction
|
|
Coordination::Responses responses;
|
|
auto res = zookeeper->tryMulti(ops, responses);
|
|
if (res != Coordination::Error::ZNODEEXISTS && res != Coordination::Error::ZNONODE)
|
|
zkutil::KeeperMultiException::check(res, ops, responses);
|
|
|
|
if (database_replicated_ext)
|
|
{
|
|
database_replicated_ext->executed_callback(task.entry_name, zookeeper);
|
|
++(database_replicated_ext->first_not_executed);
|
|
}
|
|
}
|
|
|
|
|
|
bool DDLWorker::taskShouldBeExecutedOnLeader(const ASTPtr ast_ddl, const StoragePtr storage)
|
|
{
|
|
/// Pure DROP queries have to be executed on each node separately
|
|
if (auto * query = ast_ddl->as<ASTDropQuery>(); query && query->kind != ASTDropQuery::Kind::Truncate)
|
|
return false;
|
|
|
|
if (!ast_ddl->as<ASTAlterQuery>() && !ast_ddl->as<ASTOptimizeQuery>() && !ast_ddl->as<ASTDropQuery>())
|
|
return false;
|
|
|
|
return storage->supportsReplication();
|
|
}
|
|
|
|
bool DDLWorker::tryExecuteQueryOnLeaderReplica(
|
|
DDLTask & task,
|
|
StoragePtr storage,
|
|
const String & rewritten_query,
|
|
const String & node_path,
|
|
const ZooKeeperPtr & zookeeper)
|
|
{
|
|
StorageReplicatedMergeTree * replicated_storage = dynamic_cast<StorageReplicatedMergeTree *>(storage.get());
|
|
|
|
/// If we will develop new replicated storage
|
|
if (!replicated_storage)
|
|
throw Exception(ErrorCodes::NOT_IMPLEMENTED, "Storage type '{}' is not supported by distributed DDL", storage->getName());
|
|
|
|
/// Generate unique name for shard node, it will be used to execute the query by only single host
|
|
/// Shard node name has format 'replica_name1,replica_name2,...,replica_nameN'
|
|
/// Where replica_name is 'replica_config_host_name:replica_port'
|
|
auto get_shard_name = [] (const Cluster::Addresses & shard_addresses)
|
|
{
|
|
Strings replica_names;
|
|
for (const Cluster::Address & address : shard_addresses)
|
|
replica_names.emplace_back(address.readableString());
|
|
std::sort(replica_names.begin(), replica_names.end());
|
|
|
|
String res;
|
|
for (auto it = replica_names.begin(); it != replica_names.end(); ++it)
|
|
res += *it + (std::next(it) != replica_names.end() ? "," : "");
|
|
|
|
return res;
|
|
};
|
|
|
|
String shard_node_name;
|
|
if (database_replicated_ext)
|
|
shard_node_name = database_replicated_ext->shard_name;
|
|
else
|
|
shard_node_name = get_shard_name(task.cluster->getShardsAddresses().at(task.host_shard_num));
|
|
String shard_path = node_path + "/shards/" + shard_node_name;
|
|
String is_executed_path = shard_path + "/executed";
|
|
task.shard_path = is_executed_path; //FIXME duplicate
|
|
String tries_to_execute_path = shard_path + "/tries_to_execute";
|
|
zookeeper->createAncestors(shard_path + "/");
|
|
|
|
/// Node exists, or we will create or we will get an exception
|
|
zookeeper->tryCreate(tries_to_execute_path, "0", zkutil::CreateMode::Persistent);
|
|
|
|
static constexpr int MAX_TRIES_TO_EXECUTE = 3;
|
|
static constexpr int MAX_EXECUTION_TIMEOUT_SEC = 3600;
|
|
|
|
String executed_by;
|
|
|
|
zkutil::EventPtr event = std::make_shared<Poco::Event>();
|
|
/// We must use exists request instead of get, because zookeeper will not setup event
|
|
/// for non existing node after get request
|
|
if (zookeeper->exists(is_executed_path, nullptr, event))
|
|
{
|
|
LOG_DEBUG(log, "Task {} has already been executed by replica ({}) of the same shard.", task.entry_name, zookeeper->get(is_executed_path));
|
|
return true;
|
|
}
|
|
|
|
pcg64 rng(randomSeed());
|
|
|
|
auto lock = createSimpleZooKeeperLock(zookeeper, shard_path, "lock", task.host_id_str);
|
|
|
|
Stopwatch stopwatch;
|
|
|
|
bool executed_by_leader = false;
|
|
/// Defensive programming. One hour is more than enough to execute almost all DDL queries.
|
|
/// If it will be very long query like ALTER DELETE for a huge table it's still will be executed,
|
|
/// but DDL worker can continue processing other queries.
|
|
while (stopwatch.elapsedSeconds() <= MAX_EXECUTION_TIMEOUT_SEC)
|
|
{
|
|
StorageReplicatedMergeTree::Status status;
|
|
replicated_storage->getStatus(status);
|
|
|
|
/// Any replica which is leader tries to take lock
|
|
if (status.is_leader && lock->tryLock())
|
|
{
|
|
/// In replicated merge tree we can have multiple leaders. So we can
|
|
/// be "leader" and took lock, but another "leader" replica may have
|
|
/// already executed this task.
|
|
if (zookeeper->tryGet(is_executed_path, executed_by))
|
|
{
|
|
LOG_DEBUG(log, "Task {} has already been executed by replica ({}) of the same shard.", task.entry_name, executed_by);
|
|
executed_by_leader = true;
|
|
break;
|
|
}
|
|
|
|
/// Checking and incrementing counter exclusively.
|
|
size_t counter = parse<int>(zookeeper->get(tries_to_execute_path));
|
|
if (counter > MAX_TRIES_TO_EXECUTE)
|
|
break;
|
|
|
|
zookeeper->set(tries_to_execute_path, toString(counter + 1));
|
|
|
|
/// If the leader will unexpectedly changed this method will return false
|
|
/// and on the next iteration new leader will take lock
|
|
if (tryExecuteQuery(rewritten_query, task, task.execution_status))
|
|
{
|
|
//FIXME replace with create(...) or remove and use MetadataTransaction
|
|
zookeeper->createIfNotExists(is_executed_path, task.host_id_str);
|
|
executed_by_leader = true;
|
|
break;
|
|
}
|
|
|
|
lock->unlock();
|
|
}
|
|
|
|
/// Waiting for someone who will execute query and change is_executed_path node
|
|
if (event->tryWait(std::uniform_int_distribution<int>(0, 1000)(rng)))
|
|
{
|
|
LOG_DEBUG(log, "Task {} has already been executed by replica ({}) of the same shard.", task.entry_name, zookeeper->get(is_executed_path));
|
|
executed_by_leader = true;
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
String tries_count;
|
|
zookeeper->tryGet(tries_to_execute_path, tries_count);
|
|
if (parse<int>(tries_count) > MAX_TRIES_TO_EXECUTE)
|
|
{
|
|
/// Nobody will try to execute query again
|
|
LOG_WARNING(log, "Maximum retries count for task {} exceeded, cannot execute replicated DDL query", task.entry_name);
|
|
break;
|
|
}
|
|
else
|
|
{
|
|
/// Will try to wait or execute
|
|
LOG_TRACE(log, "Task {} still not executed, will try to wait for it or execute ourselves, tries count {}", task.entry_name, tries_count);
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Not executed by leader so was not executed at all
|
|
if (!executed_by_leader)
|
|
{
|
|
/// If we failed with timeout
|
|
if (stopwatch.elapsedSeconds() >= MAX_EXECUTION_TIMEOUT_SEC)
|
|
{
|
|
LOG_WARNING(log, "Task {} was not executed by anyone, maximum timeout {} seconds exceeded", task.entry_name, MAX_EXECUTION_TIMEOUT_SEC);
|
|
task.execution_status = ExecutionStatus(ErrorCodes::TIMEOUT_EXCEEDED, "Cannot execute replicated DDL query, timeout exceeded");
|
|
}
|
|
else /// If we exceeded amount of tries
|
|
{
|
|
LOG_WARNING(log, "Task {} was not executed by anyone, maximum number of retries exceeded", task.entry_name);
|
|
task.execution_status = ExecutionStatus(ErrorCodes::UNFINISHED, "Cannot execute replicated DDL query, maximum retires exceeded");
|
|
}
|
|
return false;
|
|
}
|
|
|
|
LOG_DEBUG(log, "Task {} has already been executed by replica ({}) of the same shard.", task.entry_name, zookeeper->get(is_executed_path));
|
|
return true;
|
|
}
|
|
|
|
|
|
void DDLWorker::cleanupQueue(Int64 current_time_seconds, const ZooKeeperPtr & zookeeper)
|
|
{
|
|
LOG_DEBUG(log, "Cleaning queue");
|
|
|
|
Strings queue_nodes = zookeeper->getChildren(queue_dir);
|
|
filterAndSortQueueNodes(queue_nodes);
|
|
|
|
size_t num_outdated_nodes = (queue_nodes.size() > max_tasks_in_queue) ? queue_nodes.size() - max_tasks_in_queue : 0;
|
|
auto first_non_outdated_node = queue_nodes.begin() + num_outdated_nodes;
|
|
|
|
for (auto it = queue_nodes.cbegin(); it < queue_nodes.cend(); ++it)
|
|
{
|
|
if (stop_flag)
|
|
return;
|
|
|
|
String node_name = *it;
|
|
String node_path = queue_dir + "/" + node_name;
|
|
String lock_path = node_path + "/lock";
|
|
|
|
Coordination::Stat stat;
|
|
String dummy;
|
|
|
|
try
|
|
{
|
|
/// Already deleted
|
|
if (!zookeeper->exists(node_path, &stat))
|
|
continue;
|
|
|
|
/// Delete node if its lifetime is expired (according to task_max_lifetime parameter)
|
|
constexpr UInt64 zookeeper_time_resolution = 1000;
|
|
Int64 zookeeper_time_seconds = stat.ctime / zookeeper_time_resolution;
|
|
bool node_lifetime_is_expired = zookeeper_time_seconds + task_max_lifetime < current_time_seconds;
|
|
|
|
/// If too many nodes in task queue (> max_tasks_in_queue), delete oldest one
|
|
bool node_is_outside_max_window = it < first_non_outdated_node;
|
|
|
|
if (!node_lifetime_is_expired && !node_is_outside_max_window)
|
|
continue;
|
|
|
|
/// Skip if there are active nodes (it is weak guard)
|
|
if (zookeeper->exists(node_path + "/active", &stat) && stat.numChildren > 0)
|
|
{
|
|
LOG_INFO(log, "Task {} should be deleted, but there are active workers. Skipping it.", node_name);
|
|
continue;
|
|
}
|
|
|
|
/// Usage of the lock is not necessary now (tryRemoveRecursive correctly removes node in a presence of concurrent cleaners)
|
|
/// But the lock will be required to implement system.distributed_ddl_queue table
|
|
auto lock = createSimpleZooKeeperLock(zookeeper, node_path, "lock", host_fqdn_id);
|
|
if (!lock->tryLock())
|
|
{
|
|
LOG_INFO(log, "Task {} should be deleted, but it is locked. Skipping it.", node_name);
|
|
continue;
|
|
}
|
|
|
|
if (node_lifetime_is_expired)
|
|
LOG_INFO(log, "Lifetime of task {} is expired, deleting it", node_name);
|
|
else if (node_is_outside_max_window)
|
|
LOG_INFO(log, "Task {} is outdated, deleting it", node_name);
|
|
|
|
/// Deleting
|
|
{
|
|
Strings children = zookeeper->getChildren(node_path);
|
|
for (const String & child : children)
|
|
{
|
|
if (child != "lock")
|
|
zookeeper->tryRemoveRecursive(node_path + "/" + child);
|
|
}
|
|
|
|
/// Remove the lock node and its parent atomically
|
|
Coordination::Requests ops;
|
|
ops.emplace_back(zkutil::makeRemoveRequest(lock_path, -1));
|
|
ops.emplace_back(zkutil::makeRemoveRequest(node_path, -1));
|
|
zookeeper->multi(ops);
|
|
}
|
|
}
|
|
catch (...)
|
|
{
|
|
LOG_INFO(log, "An error occurred while checking and cleaning task {} from queue: {}", node_name, getCurrentExceptionMessage(false));
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/// Try to create nonexisting "status" dirs for a node
|
|
void DDLWorker::createStatusDirs(const std::string & node_path, const ZooKeeperPtr & zookeeper)
|
|
{
|
|
Coordination::Requests ops;
|
|
{
|
|
Coordination::CreateRequest request;
|
|
request.path = node_path + "/active";
|
|
ops.emplace_back(std::make_shared<Coordination::CreateRequest>(std::move(request)));
|
|
}
|
|
{
|
|
Coordination::CreateRequest request;
|
|
request.path = node_path + "/finished";
|
|
ops.emplace_back(std::make_shared<Coordination::CreateRequest>(std::move(request)));
|
|
}
|
|
Coordination::Responses responses;
|
|
Coordination::Error code = zookeeper->tryMulti(ops, responses);
|
|
if (code != Coordination::Error::ZOK
|
|
&& code != Coordination::Error::ZNODEEXISTS)
|
|
throw Coordination::Exception(code);
|
|
}
|
|
|
|
|
|
String DDLWorker::enqueueQuery(DDLLogEntry & entry)
|
|
{
|
|
if (entry.hosts.empty() && !database_replicated_ext)
|
|
throw Exception("Empty host list in a distributed DDL task", ErrorCodes::LOGICAL_ERROR);
|
|
|
|
auto zookeeper = getAndSetZooKeeper();
|
|
|
|
String query_path_prefix = queue_dir + "/query-";
|
|
zookeeper->createAncestors(query_path_prefix);
|
|
|
|
String node_path;
|
|
if (database_replicated_ext)
|
|
{
|
|
/// We cannot create sequential node and it's ephemeral child in a single transaction, so allocate sequential number another way
|
|
String counter_prefix = database_replicated_ext->zookeeper_path + "/counter/cnt-";
|
|
String counter_path = zookeeper->create(counter_prefix, "", zkutil::CreateMode::EphemeralSequential);
|
|
node_path = query_path_prefix + counter_path.substr(counter_prefix.size());
|
|
|
|
Coordination::Requests ops;
|
|
/// Query is not committed yet, but we have to write it into log to avoid reordering
|
|
ops.emplace_back(zkutil::makeCreateRequest(node_path, entry.toString(), zkutil::CreateMode::Persistent));
|
|
/// '/try' will be replaced with '/committed' or will be removed due to expired session or other error
|
|
ops.emplace_back(zkutil::makeCreateRequest(node_path + "/try", database_replicated_ext->getFullReplicaName(), zkutil::CreateMode::Ephemeral));
|
|
/// We don't need it anymore
|
|
ops.emplace_back(zkutil::makeRemoveRequest(counter_path, -1));
|
|
zookeeper->multi(ops);
|
|
}
|
|
else
|
|
{
|
|
node_path = zookeeper->create(query_path_prefix, entry.toString(), zkutil::CreateMode::PersistentSequential);
|
|
}
|
|
|
|
/// Optional step
|
|
try
|
|
{
|
|
createStatusDirs(node_path, zookeeper);
|
|
}
|
|
catch (...)
|
|
{
|
|
LOG_INFO(log, "An error occurred while creating auxiliary ZooKeeper directories in {} . They will be created later. Error : {}", node_path, getCurrentExceptionMessage(true));
|
|
}
|
|
|
|
return node_path;
|
|
}
|
|
|
|
|
|
void DDLWorker::runMainThread()
|
|
{
|
|
setThreadName("DDLWorker");
|
|
LOG_DEBUG(log, "Started DDLWorker thread");
|
|
|
|
bool initialized = false;
|
|
do
|
|
{
|
|
try
|
|
{
|
|
auto zookeeper = getAndSetZooKeeper();
|
|
zookeeper->createAncestors(queue_dir + "/");
|
|
initialized = true;
|
|
}
|
|
catch (const Coordination::Exception & e)
|
|
{
|
|
if (!Coordination::isHardwareError(e.code))
|
|
throw; /// A logical error.
|
|
|
|
tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
|
|
/// Avoid busy loop when ZooKeeper is not available.
|
|
sleepForSeconds(1);
|
|
}
|
|
catch (...)
|
|
{
|
|
tryLogCurrentException(log, "Terminating. Cannot initialize DDL queue.");
|
|
return;
|
|
}
|
|
}
|
|
while (!initialized && !stop_flag);
|
|
|
|
while (!stop_flag)
|
|
{
|
|
try
|
|
{
|
|
attachToThreadGroup();
|
|
|
|
cleanup_event->set();
|
|
scheduleTasks();
|
|
|
|
LOG_DEBUG(log, "Waiting a watch");
|
|
queue_updated_event->wait();
|
|
}
|
|
catch (const Coordination::Exception & e)
|
|
{
|
|
if (Coordination::isHardwareError(e.code))
|
|
{
|
|
recoverZooKeeper();
|
|
}
|
|
else if (e.code == Coordination::Error::ZNONODE)
|
|
{
|
|
LOG_ERROR(log, "ZooKeeper error: {}", getCurrentExceptionMessage(true));
|
|
}
|
|
else
|
|
{
|
|
LOG_ERROR(log, "Unexpected ZooKeeper error: {}. Terminating.", getCurrentExceptionMessage(true));
|
|
return;
|
|
}
|
|
}
|
|
catch (...)
|
|
{
|
|
tryLogCurrentException(log, "Unexpected error, will terminate:");
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
void DDLWorker::runCleanupThread()
|
|
{
|
|
setThreadName("DDLWorkerClnr");
|
|
LOG_DEBUG(log, "Started DDLWorker cleanup thread");
|
|
|
|
Int64 last_cleanup_time_seconds = 0;
|
|
while (!stop_flag)
|
|
{
|
|
try
|
|
{
|
|
cleanup_event->wait();
|
|
if (stop_flag)
|
|
break;
|
|
|
|
Int64 current_time_seconds = Poco::Timestamp().epochTime();
|
|
if (last_cleanup_time_seconds && current_time_seconds < last_cleanup_time_seconds + cleanup_delay_period)
|
|
{
|
|
LOG_TRACE(log, "Too early to clean queue, will do it later.");
|
|
continue;
|
|
}
|
|
|
|
auto zookeeper = tryGetZooKeeper();
|
|
if (zookeeper->expired())
|
|
continue;
|
|
|
|
cleanupQueue(current_time_seconds, zookeeper);
|
|
last_cleanup_time_seconds = current_time_seconds;
|
|
}
|
|
catch (...)
|
|
{
|
|
tryLogCurrentException(log, __PRETTY_FUNCTION__);
|
|
}
|
|
}
|
|
}
|
|
|
|
}
|