#pragma once #include #include #include #include #include #include #include #include #include #include #include #include #include #include namespace DB { using ChannelPtr = std::shared_ptr; class StorageRabbitMQ final: public shared_ptr_helper, public IStorage, WithContext { friend struct shared_ptr_helper; public: std::string getName() const override { return "RabbitMQ"; } bool noPushingToViews() const override { return true; } void startup() override; void shutdown() override; /// This is a bad way to let storage know in shutdown() that table is going to be dropped. There are some actions which need /// to be done only when table is dropped (not when detached). Also connection must be closed only in shutdown, but those /// actions require an open connection. Therefore there needs to be a way inside shutdown() method to know whether it is called /// because of drop query. And drop() method is not suitable at all, because it will not only require to reopen connection, but also /// it can be called considerable time after table is dropped (for example, in case of Atomic database), which is not appropriate for the case. void checkTableCanBeDropped() const override { drop_table = true; } /// Always return virtual columns in addition to required columns Pipe read( const Names & column_names, const StorageMetadataPtr & metadata_snapshot, SelectQueryInfo & query_info, ContextPtr context, QueryProcessingStage::Enum processed_stage, size_t max_block_size, unsigned num_streams) override; BlockOutputStreamPtr write( const ASTPtr & query, const StorageMetadataPtr & metadata_snapshot, ContextPtr context) override; void pushReadBuffer(ConsumerBufferPtr buf); ConsumerBufferPtr popReadBuffer(); ConsumerBufferPtr popReadBuffer(std::chrono::milliseconds timeout); ProducerBufferPtr createWriteBuffer(); const String & getFormatName() const { return format_name; } NamesAndTypesList getVirtuals() const override; String getExchange() const { return exchange_name; } void unbindExchange(); bool exchangeRemoved() { return exchange_removed.load(); } bool updateChannel(ChannelPtr & channel); void updateQueues(std::vector & queues_) { queues_ = queues; } protected: StorageRabbitMQ( const StorageID & table_id_, ContextPtr context_, const ColumnsDescription & columns_, std::unique_ptr rabbitmq_settings_); private: ContextMutablePtr rabbitmq_context; std::unique_ptr rabbitmq_settings; const String exchange_name; const String format_name; AMQP::ExchangeType exchange_type; Names routing_keys; char row_delimiter; const String schema_name; size_t num_consumers; size_t num_queues; String queue_base; Names queue_settings_list; /// For insert query. Mark messages as durable. const bool persistent; /// A table setting. It is possible not to perform any RabbitMQ setup, which is supposed to be consumer-side setup: /// declaring exchanges, queues, bindings. Instead everything needed from RabbitMQ table is to connect to a specific queue. /// This solution disables all optimizations and is not really optimal, but allows user to fully control all RabbitMQ setup. bool use_user_setup; bool hash_exchange; Poco::Logger * log; String address; std::pair parsed_address; std::pair login_password; String vhost; UVLoop loop; std::shared_ptr event_handler; std::unique_ptr connection; /// Connection for all consumers size_t num_created_consumers = 0; Poco::Semaphore semaphore; std::mutex buffers_mutex; std::vector buffers; /// available buffers for RabbitMQ consumers String unique_strbase; /// to make unique consumer channel id /// maximum number of messages in RabbitMQ queue (x-max-length). Also used /// to setup size of inner buffer for received messages uint32_t queue_size; String sharding_exchange, bridge_exchange, consumer_exchange; size_t consumer_id = 0; /// counter for consumer buffer, needed for channel id std::atomic producer_id = 1; /// counter for producer buffer, needed for channel id std::atomic wait_confirm = true; /// needed to break waiting for confirmations for producer std::atomic exchange_removed = false, rabbit_is_ready = false; std::vector queues; std::once_flag flag; /// remove exchange only once std::mutex task_mutex; BackgroundSchedulePool::TaskHolder streaming_task; BackgroundSchedulePool::TaskHolder looping_task; BackgroundSchedulePool::TaskHolder connection_task; uint64_t milliseconds_to_wait; std::atomic stream_cancelled{false}; size_t read_attempts = 0; mutable bool drop_table = false; ConsumerBufferPtr createReadBuffer(); /// Functions working in the background void streamingToViewsFunc(); void loopingFunc(); void connectionFunc(); static Names parseSettings(String settings_list); static AMQP::ExchangeType defineExchangeType(String exchange_type_); static String getTableBasedName(String name, const StorageID & table_id); std::shared_ptr addSettings(ContextPtr context) const; size_t getMaxBlockSize() const; void deactivateTask(BackgroundSchedulePool::TaskHolder & task, bool wait, bool stop_loop); void initRabbitMQ(); void cleanupRabbitMQ() const; void initExchange(AMQP::TcpChannel & rabbit_channel); void bindExchange(AMQP::TcpChannel & rabbit_channel); void bindQueue(size_t queue_id, AMQP::TcpChannel & rabbit_channel); bool restoreConnection(bool reconnecting); bool streamToViews(); bool checkDependencies(const StorageID & table_id); String getRandomName() const { std::uniform_int_distribution distribution('a', 'z'); String random_str(32, ' '); for (auto & c : random_str) c = distribution(thread_local_rng); return random_str; } }; }