#pragma once #include #include #include #include #include #include #include #include #include #include #include #include #include #include namespace DB { namespace ErrorCodes { extern const int LOGICAL_ERROR; } struct TaskRuntimeData; using TaskRuntimeDataPtr = std::shared_ptr; /** * Has RAII class to determine how many tasks are waiting for the execution and executing at the moment. * Also has some flags and primitives to wait for current task to be executed. */ struct TaskRuntimeData { TaskRuntimeData(ExecutableTaskPtr && task_, CurrentMetrics::Metric metric_) : task(std::move(task_)) , increment(std::move(metric_)) {} ExecutableTaskPtr task; CurrentMetrics::Increment increment; std::atomic_bool is_currently_deleting{false}; /// Actually autoreset=false is needed only for unit test /// where multiple threads could remove tasks corresponding to the same storage /// This scenario in not possible in reality. Poco::Event is_done{/*autoreset=*/false}; /// This is equal to task->getPriority() not to do useless virtual calls in comparator UInt64 priority{0}; /// By default priority queue will have max element at top static bool comparePtrByPriority(const TaskRuntimeDataPtr & lhs, const TaskRuntimeDataPtr & rhs) { return lhs->priority > rhs->priority; } }; class OrdinaryRuntimeQueue { public: TaskRuntimeDataPtr pop() { auto result = std::move(queue.front()); queue.pop_front(); return result; } void push(TaskRuntimeDataPtr item) { queue.push_back(std::move(item));} void remove(StorageID id) { auto it = std::remove_if(queue.begin(), queue.end(), [&] (auto item) -> bool { return item->task->getStorageID() == id; }); queue.erase(it, queue.end()); } void setCapacity(size_t count) { queue.set_capacity(count); } bool empty() { return queue.empty(); } private: boost::circular_buffer queue{0}; }; /// Uses a heap to pop a task with minimal priority class MergeMutateRuntimeQueue { public: TaskRuntimeDataPtr pop() { std::pop_heap(buffer.begin(), buffer.end(), TaskRuntimeData::comparePtrByPriority); auto result = std::move(buffer.back()); buffer.pop_back(); return result; } void push(TaskRuntimeDataPtr item) { item->priority = item->task->getPriority(); buffer.push_back(std::move(item)); std::push_heap(buffer.begin(), buffer.end(), TaskRuntimeData::comparePtrByPriority); } void remove(StorageID id) { auto it = std::remove_if(buffer.begin(), buffer.end(), [&] (auto item) -> bool { return item->task->getStorageID() == id; }); buffer.erase(it, buffer.end()); std::make_heap(buffer.begin(), buffer.end(), TaskRuntimeData::comparePtrByPriority); } void setCapacity(size_t count) { buffer.reserve(count); } bool empty() { return buffer.empty(); } private: std::vector buffer{}; }; /** * Executor for a background MergeTree related operations such as merges, mutations, fetches an so on. * It can execute only successors of ExecutableTask interface. * Which is a self-written coroutine. It suspends, when returns true from executeStep() method. * * There are two queues of a tasks: pending (main queue for all the tasks) and active (currently executing). * Pending queue is needed since the number of tasks will be more than thread to execute. * Pending tasks are tasks that successfully scheduled to an executor or tasks that have some extra steps to execute. * There is an invariant, that task may occur only in one of these queue. It can occur in both queues only in critical sections. * * Pending: Active: * * |s| |s| |s| |s| |s| |s| |s| |s| |s| |s| |s| * |s| |s| |s| |s| |s| |s| |s| |s| |s| |s| * |s| |s| |s| |s| |s| |s| |s| * |s| |s| |s| |s| * |s| |s| * |s| * * Each task is simply a sequence of steps. Heavier tasks have longer sequences. * When a step of a task is executed, we move tasks to pending queue. And take another from the queue's head. * With these architecture all small merges / mutations will be executed faster, than bigger ones. * * We use boost::circular_buffer as a container for queues not to do any allocations. * * Another nuisance that we faces with is than background operations always interact with an associated Storage. * So, when a Storage want to shutdown, it must wait until all its background operaions are finished. */ template class MergeTreeBackgroundExecutor final : public shared_ptr_helper> { public: MergeTreeBackgroundExecutor( String name_, size_t threads_count_, size_t max_tasks_count_, CurrentMetrics::Metric metric_) : name(name_) , threads_count(threads_count_) , max_tasks_count(max_tasks_count_) , metric(metric_) { if (max_tasks_count == 0) throw Exception(ErrorCodes::LOGICAL_ERROR, "Task count for MergeTreeBackgroundExecutor must not be zero"); pending.setCapacity(max_tasks_count); active.set_capacity(max_tasks_count); pool.setMaxThreads(std::max(1UL, threads_count)); pool.setMaxFreeThreads(std::max(1UL, threads_count)); pool.setQueueSize(std::max(1UL, threads_count)); for (size_t number = 0; number < threads_count; ++number) pool.scheduleOrThrowOnError([this] { threadFunction(); }); } ~MergeTreeBackgroundExecutor() { wait(); } bool trySchedule(ExecutableTaskPtr task); void removeTasksCorrespondingToStorage(StorageID id); void wait(); private: String name; size_t threads_count{0}; size_t max_tasks_count{0}; CurrentMetrics::Metric metric; void routine(TaskRuntimeDataPtr item); void threadFunction(); /// Initially it will be empty Queue pending{}; boost::circular_buffer active{0}; std::mutex mutex; std::condition_variable has_tasks; std::atomic_bool shutdown{false}; ThreadPool pool; }; extern template class MergeTreeBackgroundExecutor; extern template class MergeTreeBackgroundExecutor; using MergeMutateBackgroundExecutor = MergeTreeBackgroundExecutor; using OrdinaryBackgroundExecutor = MergeTreeBackgroundExecutor; }