2018-12-01 12:42:56 +00:00
|
|
|
#include <boost/range/algorithm_ext/erase.hpp>
|
2017-04-01 09:19:00 +00:00
|
|
|
#include <Interpreters/InterpreterSelectQuery.h>
|
|
|
|
#include <Interpreters/InterpreterInsertQuery.h>
|
|
|
|
#include <Interpreters/InterpreterAlterQuery.h>
|
2018-11-19 06:14:36 +00:00
|
|
|
#include <Interpreters/castColumn.h>
|
2017-12-30 00:36:06 +00:00
|
|
|
#include <Interpreters/evaluateConstantExpression.h>
|
2020-01-29 18:14:40 +00:00
|
|
|
#include <Processors/Transforms/AddingMissedTransform.h>
|
2019-01-23 14:48:50 +00:00
|
|
|
#include <DataStreams/IBlockInputStream.h>
|
2017-04-01 09:19:00 +00:00
|
|
|
#include <Storages/StorageBuffer.h>
|
2017-12-28 21:36:27 +00:00
|
|
|
#include <Storages/StorageFactory.h>
|
2018-12-25 23:14:39 +00:00
|
|
|
#include <Storages/AlterCommands.h>
|
2017-04-01 09:19:00 +00:00
|
|
|
#include <Parsers/ASTInsertQuery.h>
|
|
|
|
#include <Parsers/ASTIdentifier.h>
|
2017-12-30 00:36:06 +00:00
|
|
|
#include <Parsers/ASTLiteral.h>
|
2017-04-01 09:19:00 +00:00
|
|
|
#include <Parsers/ASTExpressionList.h>
|
|
|
|
#include <Common/setThreadName.h>
|
|
|
|
#include <Common/CurrentMetrics.h>
|
2017-04-08 01:32:05 +00:00
|
|
|
#include <Common/MemoryTracker.h>
|
2017-12-30 00:36:06 +00:00
|
|
|
#include <Common/FieldVisitors.h>
|
2019-10-08 18:42:22 +00:00
|
|
|
#include <Common/quoteString.h>
|
2017-12-30 00:36:06 +00:00
|
|
|
#include <Common/typeid_cast.h>
|
2018-06-05 19:46:49 +00:00
|
|
|
#include <Common/ProfileEvents.h>
|
2017-01-21 04:24:28 +00:00
|
|
|
#include <common/logger_useful.h>
|
2020-02-02 02:35:47 +00:00
|
|
|
#include <common/getThreadId.h>
|
2017-06-06 17:18:32 +00:00
|
|
|
#include <ext/range.h>
|
2020-01-29 18:14:40 +00:00
|
|
|
#include <Processors/Transforms/ConvertingTransform.h>
|
|
|
|
#include <Processors/Transforms/FilterTransform.h>
|
|
|
|
#include <Processors/Transforms/ExpressionTransform.h>
|
2020-01-30 10:22:59 +00:00
|
|
|
#include <Processors/Sources/SourceFromInputStream.h>
|
2014-10-27 04:18:13 +00:00
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2016-12-19 23:55:13 +00:00
|
|
|
namespace ProfileEvents
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
extern const Event StorageBufferFlush;
|
|
|
|
extern const Event StorageBufferErrorOnFlush;
|
|
|
|
extern const Event StorageBufferPassedAllMinThresholds;
|
|
|
|
extern const Event StorageBufferPassedTimeMaxThreshold;
|
|
|
|
extern const Event StorageBufferPassedRowsMaxThreshold;
|
|
|
|
extern const Event StorageBufferPassedBytesMaxThreshold;
|
2016-12-19 23:55:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
namespace CurrentMetrics
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
extern const Metric StorageBufferRows;
|
|
|
|
extern const Metric StorageBufferBytes;
|
2016-12-19 23:55:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
namespace DB
|
|
|
|
{
|
|
|
|
|
2016-01-11 21:46:36 +00:00
|
|
|
namespace ErrorCodes
|
|
|
|
{
|
2020-02-25 18:02:41 +00:00
|
|
|
extern const int NOT_IMPLEMENTED;
|
|
|
|
extern const int LOGICAL_ERROR;
|
2017-04-01 07:20:54 +00:00
|
|
|
extern const int INFINITE_LOOP;
|
2017-12-30 00:36:06 +00:00
|
|
|
extern const int NUMBER_OF_ARGUMENTS_DOESNT_MATCH;
|
2016-01-11 21:46:36 +00:00
|
|
|
}
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2019-12-04 16:06:55 +00:00
|
|
|
StorageBuffer::StorageBuffer(
|
|
|
|
const StorageID & table_id_,
|
|
|
|
const ColumnsDescription & columns_,
|
|
|
|
const ConstraintsDescription & constraints_,
|
2017-04-01 07:20:54 +00:00
|
|
|
Context & context_,
|
2019-12-04 16:06:55 +00:00
|
|
|
size_t num_shards_,
|
|
|
|
const Thresholds & min_thresholds_,
|
|
|
|
const Thresholds & max_thresholds_,
|
2020-02-17 19:28:25 +00:00
|
|
|
const StorageID & destination_id_,
|
2019-12-04 16:06:55 +00:00
|
|
|
bool allow_materialized_)
|
|
|
|
: IStorage(table_id_)
|
|
|
|
, global_context(context_)
|
|
|
|
, num_shards(num_shards_), buffers(num_shards_)
|
|
|
|
, min_thresholds(min_thresholds_)
|
|
|
|
, max_thresholds(max_thresholds_)
|
2020-02-17 19:28:25 +00:00
|
|
|
, destination_id(destination_id_)
|
2019-12-04 16:06:55 +00:00
|
|
|
, allow_materialized(allow_materialized_)
|
|
|
|
, log(&Logger::get("StorageBuffer (" + table_id_.getFullTableName() + ")"))
|
2014-10-26 00:01:36 +00:00
|
|
|
{
|
2019-08-24 21:20:20 +00:00
|
|
|
setColumns(columns_);
|
|
|
|
setConstraints(constraints_);
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
2018-09-06 15:06:54 +00:00
|
|
|
StorageBuffer::~StorageBuffer()
|
|
|
|
{
|
|
|
|
// Should not happen if shutdown was called
|
|
|
|
if (flush_thread.joinable())
|
|
|
|
{
|
|
|
|
shutdown_event.set();
|
|
|
|
flush_thread.join();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2017-03-12 19:18:07 +00:00
|
|
|
/// Reads from one buffer (from one block) under its mutex.
|
2020-01-29 18:14:40 +00:00
|
|
|
class BufferSource : public SourceWithProgress
|
2014-10-26 00:01:36 +00:00
|
|
|
{
|
|
|
|
public:
|
2020-01-29 18:14:40 +00:00
|
|
|
BufferSource(const Names & column_names_, StorageBuffer::Buffer & buffer_, const StorageBuffer & storage)
|
2020-01-30 07:28:49 +00:00
|
|
|
: SourceWithProgress(storage.getSampleBlockForColumns(column_names_))
|
2020-01-29 18:14:40 +00:00
|
|
|
, column_names(column_names_.begin(), column_names_.end()), buffer(buffer_) {}
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2018-01-06 18:10:44 +00:00
|
|
|
String getName() const override { return "Buffer"; }
|
2014-10-26 00:01:36 +00:00
|
|
|
|
|
|
|
protected:
|
2020-01-29 18:14:40 +00:00
|
|
|
Chunk generate() override
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2020-01-29 18:14:40 +00:00
|
|
|
Chunk res;
|
2014-10-26 00:12:39 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
if (has_been_read)
|
|
|
|
return res;
|
|
|
|
has_been_read = true;
|
2014-10-26 00:12:39 +00:00
|
|
|
|
2019-01-02 06:44:36 +00:00
|
|
|
std::lock_guard lock(buffer.mutex);
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
if (!buffer.data.rows())
|
|
|
|
return res;
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2020-01-29 18:14:40 +00:00
|
|
|
Columns columns;
|
|
|
|
columns.reserve(column_names.size());
|
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
for (const auto & name : column_names)
|
2020-01-29 18:14:40 +00:00
|
|
|
columns.push_back(buffer.data.getByName(name).column);
|
|
|
|
|
|
|
|
UInt64 size = columns.at(0)->size();
|
|
|
|
res.setColumns(std::move(columns), size);
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
return res;
|
|
|
|
}
|
2014-10-26 00:01:36 +00:00
|
|
|
|
|
|
|
private:
|
2017-04-11 18:40:47 +00:00
|
|
|
Names column_names;
|
2017-04-01 07:20:54 +00:00
|
|
|
StorageBuffer::Buffer & buffer;
|
|
|
|
bool has_been_read = false;
|
2014-10-26 00:01:36 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
|
2020-03-22 17:43:01 +00:00
|
|
|
QueryProcessingStage::Enum StorageBuffer::getQueryProcessingStage(const Context & context, const ASTPtr & query_ptr) const
|
2018-04-19 14:47:09 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
if (destination_id)
|
2018-04-19 14:47:09 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
auto destination = DatabaseCatalog::instance().getTable(destination_id);
|
2018-04-19 14:47:09 +00:00
|
|
|
|
|
|
|
if (destination.get() == this)
|
|
|
|
throw Exception("Destination table is myself. Read will cause infinite loop.", ErrorCodes::INFINITE_LOOP);
|
|
|
|
|
2020-03-22 17:43:01 +00:00
|
|
|
return destination->getQueryProcessingStage(context, query_ptr);
|
2018-04-19 14:47:09 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return QueryProcessingStage::FetchColumns;
|
|
|
|
}
|
|
|
|
|
2020-01-30 10:22:59 +00:00
|
|
|
|
2020-02-19 16:07:28 +00:00
|
|
|
Pipes StorageBuffer::read(
|
2020-01-30 10:26:25 +00:00
|
|
|
const Names & column_names,
|
|
|
|
const SelectQueryInfo & query_info,
|
|
|
|
const Context & context,
|
|
|
|
QueryProcessingStage::Enum processed_stage,
|
|
|
|
size_t max_block_size,
|
|
|
|
unsigned num_streams)
|
|
|
|
{
|
2020-01-29 18:14:40 +00:00
|
|
|
Pipes pipes_from_dst;
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2020-02-17 19:28:25 +00:00
|
|
|
if (destination_id)
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
auto destination = DatabaseCatalog::instance().getTable(destination_id);
|
2015-02-27 20:39:34 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
if (destination.get() == this)
|
|
|
|
throw Exception("Destination table is myself. Read will cause infinite loop.", ErrorCodes::INFINITE_LOOP);
|
2015-02-27 20:39:34 +00:00
|
|
|
|
2020-04-01 12:43:09 +00:00
|
|
|
auto destination_lock = destination->lockStructureForShare(false, context.getCurrentQueryId());
|
2018-11-30 15:34:24 +00:00
|
|
|
|
2018-12-01 12:42:56 +00:00
|
|
|
const bool dst_has_same_structure = std::all_of(column_names.begin(), column_names.end(), [this, destination](const String& column_name)
|
2018-11-19 06:14:36 +00:00
|
|
|
{
|
2018-12-01 12:42:56 +00:00
|
|
|
return destination->hasColumn(column_name) &&
|
|
|
|
destination->getColumn(column_name).type->equals(*getColumn(column_name).type);
|
|
|
|
});
|
|
|
|
|
|
|
|
if (dst_has_same_structure)
|
|
|
|
{
|
2019-12-11 13:09:46 +00:00
|
|
|
if (query_info.order_by_optimizer)
|
2019-12-20 13:15:17 +00:00
|
|
|
query_info.input_sorting_info = query_info.order_by_optimizer->getInputOrder(destination);
|
2019-12-11 13:09:46 +00:00
|
|
|
|
2018-12-01 12:42:56 +00:00
|
|
|
/// The destination table has the same structure of the requested columns and we can simply read blocks from there.
|
2020-02-19 16:02:37 +00:00
|
|
|
pipes_from_dst = destination->read(column_names, query_info, context, processed_stage, max_block_size, num_streams);
|
2018-12-01 12:42:56 +00:00
|
|
|
}
|
|
|
|
else
|
2018-11-19 06:14:36 +00:00
|
|
|
{
|
2018-12-01 12:42:56 +00:00
|
|
|
/// There is a struct mismatch and we need to convert read blocks from the destination table.
|
|
|
|
const Block header = getSampleBlock();
|
|
|
|
Names columns_intersection = column_names;
|
|
|
|
Block header_after_adding_defaults = header;
|
|
|
|
for (const String & column_name : column_names)
|
2018-11-27 00:43:58 +00:00
|
|
|
{
|
2018-12-01 12:42:56 +00:00
|
|
|
if (!destination->hasColumn(column_name))
|
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
LOG_WARNING(log, "Destination table " << destination_id.getNameForLogs()
|
2018-12-01 12:42:56 +00:00
|
|
|
<< " doesn't have column " << backQuoteIfNeed(column_name) << ". The default values are used.");
|
|
|
|
boost::range::remove_erase(columns_intersection, column_name);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const auto & dst_col = destination->getColumn(column_name);
|
|
|
|
const auto & col = getColumn(column_name);
|
|
|
|
if (!dst_col.type->equals(*col.type))
|
2018-11-27 00:43:58 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
LOG_WARNING(log, "Destination table " << destination_id.getNameForLogs()
|
2018-11-27 00:43:58 +00:00
|
|
|
<< " has different type of column " << backQuoteIfNeed(column_name) << " ("
|
2018-12-01 12:42:56 +00:00
|
|
|
<< dst_col.type->getName() << " != " << col.type->getName() << "). Data from destination table are converted.");
|
|
|
|
header_after_adding_defaults.getByName(column_name) = ColumnWithTypeAndName(dst_col.type, column_name);
|
2018-11-27 00:43:58 +00:00
|
|
|
}
|
|
|
|
}
|
2018-12-01 12:42:56 +00:00
|
|
|
|
|
|
|
if (columns_intersection.empty())
|
2018-11-27 00:43:58 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
LOG_WARNING(log, "Destination table " << destination_id.getNameForLogs()
|
2018-12-01 12:42:56 +00:00
|
|
|
<< " has no common columns with block in buffer. Block of data is skipped.");
|
2018-11-27 00:43:58 +00:00
|
|
|
}
|
2018-12-01 12:42:56 +00:00
|
|
|
else
|
2018-11-27 00:43:58 +00:00
|
|
|
{
|
2020-02-19 16:02:37 +00:00
|
|
|
pipes_from_dst = destination->read(columns_intersection, query_info, context, processed_stage, max_block_size, num_streams);
|
2020-01-29 18:14:40 +00:00
|
|
|
for (auto & pipe : pipes_from_dst)
|
2018-12-01 12:42:56 +00:00
|
|
|
{
|
2020-01-29 18:14:40 +00:00
|
|
|
pipe.addSimpleTransform(std::make_shared<AddingMissedTransform>(
|
|
|
|
pipe.getHeader(), header_after_adding_defaults, getColumns().getDefaults(), context));
|
|
|
|
|
|
|
|
pipe.addSimpleTransform(std::make_shared<ConvertingTransform>(
|
|
|
|
pipe.getHeader(), header, ConvertingTransform::MatchColumnsMode::Name, context));
|
2018-12-01 12:42:56 +00:00
|
|
|
}
|
2018-11-27 00:43:58 +00:00
|
|
|
}
|
|
|
|
}
|
2018-12-01 12:42:56 +00:00
|
|
|
|
2020-01-29 18:14:40 +00:00
|
|
|
for (auto & pipe : pipes_from_dst)
|
|
|
|
pipe.addTableLock(destination_lock);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2020-01-29 18:14:40 +00:00
|
|
|
Pipes pipes_from_buffers;
|
|
|
|
pipes_from_buffers.reserve(num_shards);
|
2017-04-01 07:20:54 +00:00
|
|
|
for (auto & buf : buffers)
|
2020-01-29 18:14:40 +00:00
|
|
|
pipes_from_buffers.emplace_back(std::make_shared<BufferSource>(column_names, buf, *this));
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/** If the sources from the table were processed before some non-initial stage of query execution,
|
|
|
|
* then sources from the buffers must also be wrapped in the processing pipeline before the same stage.
|
|
|
|
*/
|
|
|
|
if (processed_stage > QueryProcessingStage::FetchColumns)
|
2020-01-29 18:14:40 +00:00
|
|
|
for (auto & pipe : pipes_from_buffers)
|
|
|
|
pipe = InterpreterSelectQuery(query_info.query, context, std::move(pipe), SelectQueryOptions(processed_stage)).executeWithProcessors().getPipe();
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2019-03-14 08:05:18 +00:00
|
|
|
if (query_info.prewhere_info)
|
|
|
|
{
|
2020-01-29 18:14:40 +00:00
|
|
|
for (auto & pipe : pipes_from_buffers)
|
|
|
|
pipe.addSimpleTransform(std::make_shared<FilterTransform>(pipe.getHeader(), query_info.prewhere_info->prewhere_actions,
|
|
|
|
query_info.prewhere_info->prewhere_column_name, query_info.prewhere_info->remove_prewhere_column));
|
2019-03-14 08:05:18 +00:00
|
|
|
|
|
|
|
if (query_info.prewhere_info->alias_actions)
|
|
|
|
{
|
2020-01-29 18:14:40 +00:00
|
|
|
for (auto & pipe : pipes_from_buffers)
|
|
|
|
pipe.addSimpleTransform(std::make_shared<ExpressionTransform>(pipe.getHeader(), query_info.prewhere_info->alias_actions));
|
2019-03-14 08:05:18 +00:00
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-29 18:14:40 +00:00
|
|
|
for (auto & pipe : pipes_from_buffers)
|
|
|
|
pipes_from_dst.emplace_back(std::move(pipe));
|
|
|
|
|
|
|
|
return pipes_from_dst;
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-10-27 04:18:13 +00:00
|
|
|
static void appendBlock(const Block & from, Block & to)
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
if (!to)
|
|
|
|
throw Exception("Cannot append to empty block", ErrorCodes::LOGICAL_ERROR);
|
|
|
|
|
2018-02-20 01:14:38 +00:00
|
|
|
assertBlocksHaveEqualStructure(from, to, "Buffer");
|
2017-12-15 20:48:46 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
from.checkNumberOfRows();
|
|
|
|
to.checkNumberOfRows();
|
|
|
|
|
|
|
|
size_t rows = from.rows();
|
|
|
|
size_t bytes = from.bytes();
|
|
|
|
|
|
|
|
CurrentMetrics::add(CurrentMetrics::StorageBufferRows, rows);
|
|
|
|
CurrentMetrics::add(CurrentMetrics::StorageBufferBytes, bytes);
|
|
|
|
|
|
|
|
size_t old_rows = to.rows();
|
|
|
|
|
2019-10-30 19:58:19 +00:00
|
|
|
auto temporarily_disable_memory_tracker = getCurrentMemoryTrackerActionLock();
|
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
try
|
|
|
|
{
|
|
|
|
for (size_t column_no = 0, columns = to.columns(); column_no < columns; ++column_no)
|
|
|
|
{
|
2017-12-15 20:48:46 +00:00
|
|
|
const IColumn & col_from = *from.getByPosition(column_no).column.get();
|
2018-03-20 10:58:16 +00:00
|
|
|
MutableColumnPtr col_to = (*std::move(to.getByPosition(column_no).column)).mutate();
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-12-15 20:48:46 +00:00
|
|
|
col_to->insertRangeFrom(col_from, 0, rows);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-12-15 20:48:46 +00:00
|
|
|
to.getByPosition(column_no).column = std::move(col_to);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
/// Rollback changes.
|
|
|
|
try
|
|
|
|
{
|
|
|
|
for (size_t column_no = 0, columns = to.columns(); column_no < columns; ++column_no)
|
|
|
|
{
|
2017-12-15 20:48:46 +00:00
|
|
|
ColumnPtr & col_to = to.getByPosition(column_no).column;
|
2017-04-01 07:20:54 +00:00
|
|
|
if (col_to->size() != old_rows)
|
2018-03-20 10:58:16 +00:00
|
|
|
col_to = (*std::move(col_to)).mutate()->cut(0, old_rows);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
/// In case when we cannot rollback, do not leave incorrect state in memory.
|
|
|
|
std::terminate();
|
|
|
|
}
|
|
|
|
|
|
|
|
throw;
|
|
|
|
}
|
2014-10-27 04:18:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
class BufferBlockOutputStream : public IBlockOutputStream
|
|
|
|
{
|
|
|
|
public:
|
2017-09-07 21:04:48 +00:00
|
|
|
explicit BufferBlockOutputStream(StorageBuffer & storage_) : storage(storage_) {}
|
2018-02-19 00:45:32 +00:00
|
|
|
|
|
|
|
Block getHeader() const override { return storage.getSampleBlock(); }
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
void write(const Block & block) override
|
|
|
|
{
|
|
|
|
if (!block)
|
|
|
|
return;
|
|
|
|
|
2018-11-19 06:14:36 +00:00
|
|
|
// Check table structure.
|
|
|
|
storage.check(block, true);
|
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
size_t rows = block.rows();
|
|
|
|
if (!rows)
|
|
|
|
return;
|
|
|
|
|
|
|
|
StoragePtr destination;
|
2020-02-17 19:28:25 +00:00
|
|
|
if (storage.destination_id)
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
destination = DatabaseCatalog::instance().tryGetTable(storage.destination_id);
|
2018-11-19 06:14:36 +00:00
|
|
|
if (destination.get() == &storage)
|
|
|
|
throw Exception("Destination table is myself. Write will cause infinite loop.", ErrorCodes::INFINITE_LOOP);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
size_t bytes = block.bytes();
|
|
|
|
|
|
|
|
/// If the block already exceeds the maximum limit, then we skip the buffer.
|
|
|
|
if (rows > storage.max_thresholds.rows || bytes > storage.max_thresholds.bytes)
|
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
if (storage.destination_id)
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
|
|
|
LOG_TRACE(storage.log, "Writing block with " << rows << " rows, " << bytes << " bytes directly.");
|
|
|
|
storage.writeBlockToDestination(block, destination);
|
2019-10-30 19:58:19 +00:00
|
|
|
}
|
2017-04-01 07:20:54 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// We distribute the load on the shards by the stream number.
|
2020-02-02 02:35:47 +00:00
|
|
|
const auto start_shard_num = getThreadId() % storage.num_shards;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
/// We loop through the buffers, trying to lock mutex. No more than one lap.
|
|
|
|
auto shard_num = start_shard_num;
|
|
|
|
|
|
|
|
StorageBuffer::Buffer * least_busy_buffer = nullptr;
|
|
|
|
std::unique_lock<std::mutex> least_busy_lock;
|
|
|
|
size_t least_busy_shard_rows = 0;
|
|
|
|
|
|
|
|
for (size_t try_no = 0; try_no < storage.num_shards; ++try_no)
|
|
|
|
{
|
2019-01-02 06:44:36 +00:00
|
|
|
std::unique_lock lock(storage.buffers[shard_num].mutex, std::try_to_lock);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
if (lock.owns_lock())
|
|
|
|
{
|
|
|
|
size_t num_rows = storage.buffers[shard_num].data.rows();
|
|
|
|
if (!least_busy_buffer || num_rows < least_busy_shard_rows)
|
|
|
|
{
|
|
|
|
least_busy_buffer = &storage.buffers[shard_num];
|
|
|
|
least_busy_lock = std::move(lock);
|
|
|
|
least_busy_shard_rows = num_rows;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
shard_num = (shard_num + 1) % storage.num_shards;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// If you still can not lock anything at once, then we'll wait on mutex.
|
|
|
|
if (!least_busy_buffer)
|
2018-08-24 14:51:34 +00:00
|
|
|
{
|
|
|
|
least_busy_buffer = &storage.buffers[start_shard_num];
|
2019-01-02 06:44:36 +00:00
|
|
|
least_busy_lock = std::unique_lock(least_busy_buffer->mutex);
|
2018-08-24 14:51:34 +00:00
|
|
|
}
|
|
|
|
insertIntoBuffer(block, *least_busy_buffer);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
2014-10-26 00:01:36 +00:00
|
|
|
private:
|
2017-04-01 07:20:54 +00:00
|
|
|
StorageBuffer & storage;
|
|
|
|
|
2018-08-24 14:51:34 +00:00
|
|
|
void insertIntoBuffer(const Block & block, StorageBuffer::Buffer & buffer)
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2017-08-04 14:00:26 +00:00
|
|
|
time_t current_time = time(nullptr);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
/// Sort the columns in the block. This is necessary to make it easier to concatenate the blocks later.
|
|
|
|
Block sorted_block = block.sortColumns();
|
|
|
|
|
|
|
|
if (!buffer.data)
|
|
|
|
{
|
|
|
|
buffer.data = sorted_block.cloneEmpty();
|
|
|
|
}
|
|
|
|
else if (storage.checkThresholds(buffer, current_time, sorted_block.rows(), sorted_block.bytes()))
|
|
|
|
{
|
|
|
|
/** If, after inserting the buffer, the constraints are exceeded, then we will reset the buffer.
|
|
|
|
* This also protects against unlimited consumption of RAM, since if it is impossible to write to the table,
|
|
|
|
* an exception will be thrown, and new data will not be added to the buffer.
|
|
|
|
*/
|
|
|
|
|
2019-12-22 22:35:53 +00:00
|
|
|
storage.flushBuffer(buffer, false /* check_thresholds */, true /* locked */);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!buffer.first_write_time)
|
|
|
|
buffer.first_write_time = current_time;
|
|
|
|
|
|
|
|
appendBlock(sorted_block, buffer.data);
|
|
|
|
}
|
2014-10-26 00:01:36 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
|
2019-02-27 18:26:24 +00:00
|
|
|
BlockOutputStreamPtr StorageBuffer::write(const ASTPtr & /*query*/, const Context & /*context*/)
|
2014-10-26 00:01:36 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
return std::make_shared<BufferBlockOutputStream>(*this);
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2019-02-27 18:26:24 +00:00
|
|
|
bool StorageBuffer::mayBenefitFromIndexForIn(const ASTPtr & left_in_operand, const Context & query_context) const
|
2018-03-16 09:00:04 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
if (!destination_id)
|
2018-03-16 09:00:04 +00:00
|
|
|
return false;
|
|
|
|
|
2020-02-17 19:28:25 +00:00
|
|
|
auto destination = DatabaseCatalog::instance().getTable(destination_id);
|
2018-03-16 09:00:04 +00:00
|
|
|
|
|
|
|
if (destination.get() == this)
|
|
|
|
throw Exception("Destination table is myself. Read will cause infinite loop.", ErrorCodes::INFINITE_LOOP);
|
|
|
|
|
2019-02-27 18:26:24 +00:00
|
|
|
return destination->mayBenefitFromIndexForIn(left_in_operand, query_context);
|
2018-03-16 09:00:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2017-06-06 17:06:14 +00:00
|
|
|
void StorageBuffer::startup()
|
|
|
|
{
|
2019-01-04 12:10:00 +00:00
|
|
|
if (global_context.getSettingsRef().readonly)
|
2018-02-01 13:52:29 +00:00
|
|
|
{
|
|
|
|
LOG_WARNING(log, "Storage " << getName() << " is run with readonly settings, it will not be able to insert data."
|
2020-01-11 09:50:41 +00:00
|
|
|
<< " Set appropriate system_profile to fix this.");
|
2018-02-01 13:52:29 +00:00
|
|
|
}
|
|
|
|
|
2019-01-14 19:22:09 +00:00
|
|
|
flush_thread = ThreadFromGlobalPool(&StorageBuffer::flushThread, this);
|
2017-06-06 17:06:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
void StorageBuffer::shutdown()
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
shutdown_event.set();
|
|
|
|
|
|
|
|
if (flush_thread.joinable())
|
|
|
|
flush_thread.join();
|
|
|
|
|
|
|
|
try
|
|
|
|
{
|
2019-01-04 12:10:00 +00:00
|
|
|
optimize(nullptr /*query*/, {} /*partition*/, false /*final*/, false /*deduplicate*/, global_context);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
|
|
}
|
2014-10-26 00:21:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2016-11-18 09:02:49 +00:00
|
|
|
/** NOTE If you do OPTIMIZE after insertion,
|
|
|
|
* it does not guarantee, that all data will be in destination table at the time of next SELECT just after OPTIMIZE.
|
|
|
|
*
|
|
|
|
* Because in case if there was already running flushBuffer method,
|
|
|
|
* then call to flushBuffer inside OPTIMIZE will see empty buffer and return quickly,
|
|
|
|
* but at the same time, the already running flushBuffer method possibly is not finished,
|
|
|
|
* so next SELECT will observe missing data.
|
|
|
|
*
|
|
|
|
* This kind of race condition make very hard to implement proper tests.
|
|
|
|
*/
|
2017-12-01 21:13:25 +00:00
|
|
|
bool StorageBuffer::optimize(const ASTPtr & /*query*/, const ASTPtr & partition, bool final, bool deduplicate, const Context & /*context*/)
|
2014-10-26 00:21:06 +00:00
|
|
|
{
|
2017-09-06 20:34:26 +00:00
|
|
|
if (partition)
|
2017-04-01 07:20:54 +00:00
|
|
|
throw Exception("Partition cannot be specified when optimizing table of type Buffer", ErrorCodes::NOT_IMPLEMENTED);
|
2016-05-16 18:43:38 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
if (final)
|
|
|
|
throw Exception("FINAL cannot be specified when optimizing table of type Buffer", ErrorCodes::NOT_IMPLEMENTED);
|
2014-10-26 00:21:06 +00:00
|
|
|
|
2017-04-13 13:13:08 +00:00
|
|
|
if (deduplicate)
|
|
|
|
throw Exception("DEDUPLICATE cannot be specified when optimizing table of type Buffer", ErrorCodes::NOT_IMPLEMENTED);
|
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
flushAllBuffers(false);
|
|
|
|
return true;
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2015-12-09 06:55:49 +00:00
|
|
|
bool StorageBuffer::checkThresholds(const Buffer & buffer, time_t current_time, size_t additional_rows, size_t additional_bytes) const
|
2014-10-26 00:01:36 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
time_t time_passed = 0;
|
|
|
|
if (buffer.first_write_time)
|
|
|
|
time_passed = current_time - buffer.first_write_time;
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
size_t rows = buffer.data.rows() + additional_rows;
|
|
|
|
size_t bytes = buffer.data.bytes() + additional_bytes;
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
return checkThresholdsImpl(rows, bytes, time_passed);
|
2015-12-09 06:55:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
bool StorageBuffer::checkThresholdsImpl(size_t rows, size_t bytes, time_t time_passed) const
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
if (time_passed > min_thresholds.time && rows > min_thresholds.rows && bytes > min_thresholds.bytes)
|
|
|
|
{
|
|
|
|
ProfileEvents::increment(ProfileEvents::StorageBufferPassedAllMinThresholds);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (time_passed > max_thresholds.time)
|
|
|
|
{
|
|
|
|
ProfileEvents::increment(ProfileEvents::StorageBufferPassedTimeMaxThreshold);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rows > max_thresholds.rows)
|
|
|
|
{
|
|
|
|
ProfileEvents::increment(ProfileEvents::StorageBufferPassedRowsMaxThreshold);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (bytes > max_thresholds.bytes)
|
|
|
|
{
|
|
|
|
ProfileEvents::increment(ProfileEvents::StorageBufferPassedBytesMaxThreshold);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-12-03 13:28:17 +00:00
|
|
|
void StorageBuffer::flushAllBuffers(const bool check_thresholds)
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
for (auto & buf : buffers)
|
|
|
|
flushBuffer(buf, check_thresholds);
|
2014-12-03 13:28:17 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2018-08-24 14:51:34 +00:00
|
|
|
void StorageBuffer::flushBuffer(Buffer & buffer, bool check_thresholds, bool locked)
|
2014-10-26 00:01:36 +00:00
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
Block block_to_write;
|
2017-08-04 14:00:26 +00:00
|
|
|
time_t current_time = time(nullptr);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
size_t rows = 0;
|
|
|
|
size_t bytes = 0;
|
|
|
|
time_t time_passed = 0;
|
|
|
|
|
2019-01-02 06:44:36 +00:00
|
|
|
std::unique_lock lock(buffer.mutex, std::defer_lock);
|
2018-08-24 14:51:34 +00:00
|
|
|
if (!locked)
|
|
|
|
lock.lock();
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
block_to_write = buffer.data.cloneEmpty();
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
rows = buffer.data.rows();
|
|
|
|
bytes = buffer.data.bytes();
|
|
|
|
if (buffer.first_write_time)
|
|
|
|
time_passed = current_time - buffer.first_write_time;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
if (check_thresholds)
|
|
|
|
{
|
|
|
|
if (!checkThresholdsImpl(rows, bytes, time_passed))
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
if (rows == 0)
|
|
|
|
return;
|
|
|
|
}
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
buffer.data.swap(block_to_write);
|
|
|
|
buffer.first_write_time = 0;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
CurrentMetrics::sub(CurrentMetrics::StorageBufferRows, block_to_write.rows());
|
|
|
|
CurrentMetrics::sub(CurrentMetrics::StorageBufferBytes, block_to_write.bytes());
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
ProfileEvents::increment(ProfileEvents::StorageBufferFlush);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
LOG_TRACE(log, "Flushing buffer with " << rows << " rows, " << bytes << " bytes, age " << time_passed << " seconds.");
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2020-02-17 19:28:25 +00:00
|
|
|
if (!destination_id)
|
2017-09-07 21:04:48 +00:00
|
|
|
return;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
/** For simplicity, buffer is locked during write.
|
2017-11-15 19:47:49 +00:00
|
|
|
* We could unlock buffer temporary, but it would lead to too many difficulties:
|
2017-09-07 21:04:48 +00:00
|
|
|
* - data, that is written, will not be visible for SELECTs;
|
|
|
|
* - new data could be appended to buffer, and in case of exception, we must merge it with old data, that has not been written;
|
|
|
|
* - this could lead to infinite memory growth.
|
|
|
|
*/
|
|
|
|
try
|
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
writeBlockToDestination(block_to_write, DatabaseCatalog::instance().tryGetTable(destination_id));
|
2017-09-07 21:04:48 +00:00
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
ProfileEvents::increment(ProfileEvents::StorageBufferErrorOnFlush);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
/// Return the block to its place in the buffer.
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
CurrentMetrics::add(CurrentMetrics::StorageBufferRows, block_to_write.rows());
|
|
|
|
CurrentMetrics::add(CurrentMetrics::StorageBufferBytes, block_to_write.bytes());
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
buffer.data.swap(block_to_write);
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
if (!buffer.first_write_time)
|
|
|
|
buffer.first_write_time = current_time;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2017-09-07 21:04:48 +00:00
|
|
|
/// After a while, the next write attempt will happen.
|
|
|
|
throw;
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
2014-10-27 04:18:13 +00:00
|
|
|
}
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2014-10-27 04:18:13 +00:00
|
|
|
void StorageBuffer::writeBlockToDestination(const Block & block, StoragePtr table)
|
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
if (!destination_id || !block)
|
2017-04-01 07:20:54 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
if (!table)
|
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
LOG_ERROR(log, "Destination table " << destination_id.getNameForLogs() << " doesn't exist. Block of data is discarded.");
|
2017-04-01 07:20:54 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-10-30 19:58:19 +00:00
|
|
|
auto temporarily_disable_memory_tracker = getCurrentMemoryTrackerActionLock();
|
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
auto insert = std::make_shared<ASTInsertQuery>();
|
2020-03-02 20:23:58 +00:00
|
|
|
insert->table_id = destination_id;
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
/** We will insert columns that are the intersection set of columns of the buffer table and the subordinate table.
|
|
|
|
* This will support some of the cases (but not all) when the table structure does not match.
|
|
|
|
*/
|
2018-01-01 11:54:14 +00:00
|
|
|
Block structure_of_destination_table = allow_materialized ? table->getSampleBlock() : table->getSampleBlockNonMaterialized();
|
2018-11-19 06:14:36 +00:00
|
|
|
Block block_to_write;
|
2017-04-01 07:20:54 +00:00
|
|
|
for (size_t i : ext::range(0, structure_of_destination_table.columns()))
|
|
|
|
{
|
|
|
|
auto dst_col = structure_of_destination_table.getByPosition(i);
|
|
|
|
if (block.has(dst_col.name))
|
|
|
|
{
|
2018-11-19 06:14:36 +00:00
|
|
|
auto column = block.getByName(dst_col.name);
|
|
|
|
if (!column.type->equals(*dst_col.type))
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
LOG_WARNING(log, "Destination table " << destination_id.getNameForLogs()
|
2018-11-19 15:20:05 +00:00
|
|
|
<< " have different type of column " << backQuoteIfNeed(column.name) << " ("
|
2018-11-27 00:43:58 +00:00
|
|
|
<< dst_col.type->getName() << " != " << column.type->getName()
|
2018-11-19 06:14:36 +00:00
|
|
|
<< "). Block of data is converted.");
|
2019-01-04 12:10:00 +00:00
|
|
|
column.column = castColumn(column, dst_col.type, global_context);
|
2018-11-19 06:14:36 +00:00
|
|
|
column.type = dst_col.type;
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
|
2018-11-19 06:14:36 +00:00
|
|
|
block_to_write.insert(column);
|
2017-04-01 07:20:54 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-11-19 06:14:36 +00:00
|
|
|
if (block_to_write.columns() == 0)
|
2017-04-01 07:20:54 +00:00
|
|
|
{
|
2020-02-17 19:28:25 +00:00
|
|
|
LOG_ERROR(log, "Destination table " << destination_id.getNameForLogs()
|
2018-11-19 15:20:05 +00:00
|
|
|
<< " have no common columns with block in buffer. Block of data is discarded.");
|
2017-04-01 07:20:54 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2018-11-19 06:14:36 +00:00
|
|
|
if (block_to_write.columns() != block.columns())
|
2017-04-01 07:20:54 +00:00
|
|
|
LOG_WARNING(log, "Not all columns from block in buffer exist in destination table "
|
2020-02-17 19:28:25 +00:00
|
|
|
<< destination_id.getNameForLogs() << ". Some columns are discarded.");
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
auto list_of_columns = std::make_shared<ASTExpressionList>();
|
|
|
|
insert->columns = list_of_columns;
|
2018-11-19 06:14:36 +00:00
|
|
|
list_of_columns->children.reserve(block_to_write.columns());
|
2018-11-19 15:20:34 +00:00
|
|
|
for (const auto & column : block_to_write)
|
2018-11-19 06:14:36 +00:00
|
|
|
list_of_columns->children.push_back(std::make_shared<ASTIdentifier>(column.name));
|
2017-04-01 07:20:54 +00:00
|
|
|
|
2019-01-04 12:10:00 +00:00
|
|
|
InterpreterInsertQuery interpreter{insert, global_context, allow_materialized};
|
2017-04-01 07:20:54 +00:00
|
|
|
|
|
|
|
auto block_io = interpreter.execute();
|
|
|
|
block_io.out->writePrefix();
|
2018-09-09 03:28:45 +00:00
|
|
|
block_io.out->write(block_to_write);
|
2017-04-01 07:20:54 +00:00
|
|
|
block_io.out->writeSuffix();
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void StorageBuffer::flushThread()
|
|
|
|
{
|
2017-04-01 07:20:54 +00:00
|
|
|
setThreadName("BufferFlush");
|
|
|
|
|
|
|
|
do
|
|
|
|
{
|
|
|
|
try
|
|
|
|
{
|
|
|
|
flushAllBuffers(true);
|
|
|
|
}
|
|
|
|
catch (...)
|
|
|
|
{
|
|
|
|
tryLogCurrentException(__PRETTY_FUNCTION__);
|
|
|
|
}
|
|
|
|
} while (!shutdown_event.tryWait(1000));
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
2019-12-26 18:17:05 +00:00
|
|
|
void StorageBuffer::checkAlterIsPossible(const AlterCommands & commands, const Settings & /* settings */)
|
|
|
|
{
|
|
|
|
for (const auto & command : commands)
|
|
|
|
{
|
|
|
|
if (command.type != AlterCommand::Type::ADD_COLUMN && command.type != AlterCommand::Type::MODIFY_COLUMN
|
|
|
|
&& command.type != AlterCommand::Type::DROP_COLUMN && command.type != AlterCommand::Type::COMMENT_COLUMN)
|
|
|
|
throw Exception(
|
|
|
|
"Alter of type '" + alterTypeToString(command.type) + "' is not supported by storage " + getName(),
|
|
|
|
ErrorCodes::NOT_IMPLEMENTED);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-03-29 07:50:47 +00:00
|
|
|
std::optional<UInt64> StorageBuffer::totalRows() const
|
|
|
|
{
|
|
|
|
std::optional<UInt64> underlying_rows;
|
|
|
|
auto underlying = DatabaseCatalog::instance().tryGetTable(destination_id);
|
|
|
|
|
|
|
|
if (underlying)
|
|
|
|
underlying_rows = underlying->totalRows();
|
|
|
|
if (!underlying_rows)
|
|
|
|
return underlying_rows;
|
|
|
|
|
|
|
|
UInt64 rows = 0;
|
|
|
|
for (auto & buffer : buffers)
|
|
|
|
{
|
|
|
|
std::lock_guard lock(buffer.mutex);
|
|
|
|
rows += buffer.data.rows();
|
|
|
|
}
|
|
|
|
return rows + *underlying_rows;
|
|
|
|
}
|
2014-10-26 00:01:36 +00:00
|
|
|
|
2020-03-29 08:54:00 +00:00
|
|
|
std::optional<UInt64> StorageBuffer::totalBytes() const
|
|
|
|
{
|
|
|
|
UInt64 bytes = 0;
|
|
|
|
for (auto & buffer : buffers)
|
|
|
|
{
|
|
|
|
std::lock_guard lock(buffer.mutex);
|
|
|
|
bytes += buffer.data.bytes();
|
|
|
|
}
|
|
|
|
return bytes;
|
|
|
|
}
|
|
|
|
|
2019-08-26 14:50:34 +00:00
|
|
|
void StorageBuffer::alter(const AlterCommands & params, const Context & context, TableStructureWriteLockHolder & table_lock_holder)
|
2014-10-26 00:01:36 +00:00
|
|
|
{
|
2019-03-07 20:52:25 +00:00
|
|
|
lockStructureExclusively(table_lock_holder, context.getCurrentQueryId());
|
2014-10-27 04:18:13 +00:00
|
|
|
|
2019-12-10 20:47:05 +00:00
|
|
|
auto table_id = getStorageID();
|
2019-12-26 18:17:05 +00:00
|
|
|
checkAlterIsPossible(params, context.getSettingsRef());
|
2019-08-26 14:50:34 +00:00
|
|
|
|
2017-04-01 07:20:54 +00:00
|
|
|
/// So that no blocks of the old structure remain.
|
2017-09-06 20:34:26 +00:00
|
|
|
optimize({} /*query*/, {} /*partition_id*/, false /*final*/, false /*deduplicate*/, context);
|
2014-10-27 04:18:13 +00:00
|
|
|
|
2019-12-26 18:17:05 +00:00
|
|
|
StorageInMemoryMetadata metadata = getInMemoryMetadata();
|
|
|
|
params.apply(metadata);
|
2020-02-10 13:10:17 +00:00
|
|
|
DatabaseCatalog::instance().getDatabase(table_id.database_name)->alterTable(context, table_id.table_name, metadata);
|
2019-12-26 18:17:05 +00:00
|
|
|
setColumns(std::move(metadata.columns));
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|
|
|
|
|
2017-12-28 21:36:27 +00:00
|
|
|
|
|
|
|
void registerStorageBuffer(StorageFactory & factory)
|
|
|
|
{
|
|
|
|
/** Buffer(db, table, num_buckets, min_time, max_time, min_rows, max_rows, min_bytes, max_bytes)
|
|
|
|
*
|
|
|
|
* db, table - in which table to put data from buffer.
|
|
|
|
* num_buckets - level of parallelism.
|
|
|
|
* min_time, max_time, min_rows, max_rows, min_bytes, max_bytes - conditions for flushing the buffer.
|
|
|
|
*/
|
|
|
|
|
2017-12-30 00:36:06 +00:00
|
|
|
factory.registerStorage("Buffer", [](const StorageFactory::Arguments & args)
|
2017-12-28 21:36:27 +00:00
|
|
|
{
|
2017-12-30 00:36:06 +00:00
|
|
|
ASTs & engine_args = args.engine_args;
|
|
|
|
|
|
|
|
if (engine_args.size() != 9)
|
2017-12-28 21:36:27 +00:00
|
|
|
throw Exception("Storage Buffer requires 9 parameters: "
|
|
|
|
" destination_database, destination_table, num_buckets, min_time, max_time, min_rows, max_rows, min_bytes, max_bytes.",
|
|
|
|
ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
|
|
|
|
|
2020-02-19 18:58:29 +00:00
|
|
|
engine_args[0] = evaluateConstantExpressionForDatabaseName(engine_args[0], args.local_context);
|
2017-12-30 00:36:06 +00:00
|
|
|
engine_args[1] = evaluateConstantExpressionOrIdentifierAsLiteral(engine_args[1], args.local_context);
|
2017-12-28 21:36:27 +00:00
|
|
|
|
2019-03-15 17:09:14 +00:00
|
|
|
String destination_database = engine_args[0]->as<ASTLiteral &>().value.safeGet<String>();
|
|
|
|
String destination_table = engine_args[1]->as<ASTLiteral &>().value.safeGet<String>();
|
2017-12-28 21:36:27 +00:00
|
|
|
|
2019-03-15 17:09:14 +00:00
|
|
|
UInt64 num_buckets = applyVisitor(FieldVisitorConvertToNumber<UInt64>(), engine_args[2]->as<ASTLiteral &>().value);
|
2017-12-28 21:36:27 +00:00
|
|
|
|
2019-03-15 17:09:14 +00:00
|
|
|
Int64 min_time = applyVisitor(FieldVisitorConvertToNumber<Int64>(), engine_args[3]->as<ASTLiteral &>().value);
|
|
|
|
Int64 max_time = applyVisitor(FieldVisitorConvertToNumber<Int64>(), engine_args[4]->as<ASTLiteral &>().value);
|
|
|
|
UInt64 min_rows = applyVisitor(FieldVisitorConvertToNumber<UInt64>(), engine_args[5]->as<ASTLiteral &>().value);
|
|
|
|
UInt64 max_rows = applyVisitor(FieldVisitorConvertToNumber<UInt64>(), engine_args[6]->as<ASTLiteral &>().value);
|
|
|
|
UInt64 min_bytes = applyVisitor(FieldVisitorConvertToNumber<UInt64>(), engine_args[7]->as<ASTLiteral &>().value);
|
|
|
|
UInt64 max_bytes = applyVisitor(FieldVisitorConvertToNumber<UInt64>(), engine_args[8]->as<ASTLiteral &>().value);
|
2017-12-28 21:36:27 +00:00
|
|
|
|
2020-02-17 19:28:25 +00:00
|
|
|
/// If destination_id is not set, do not write data from the buffer, but simply empty the buffer.
|
|
|
|
StorageID destination_id = StorageID::createEmpty();
|
|
|
|
if (!destination_table.empty())
|
|
|
|
{
|
|
|
|
destination_id.database_name = args.context.resolveDatabase(destination_database);
|
|
|
|
destination_id.table_name = destination_table;
|
|
|
|
}
|
|
|
|
|
2017-12-28 21:36:27 +00:00
|
|
|
return StorageBuffer::create(
|
2019-12-04 16:06:55 +00:00
|
|
|
args.table_id,
|
|
|
|
args.columns,
|
|
|
|
args.constraints,
|
2017-12-30 00:36:06 +00:00
|
|
|
args.context,
|
2017-12-28 21:36:27 +00:00
|
|
|
num_buckets,
|
|
|
|
StorageBuffer::Thresholds{min_time, min_rows, min_bytes},
|
|
|
|
StorageBuffer::Thresholds{max_time, max_rows, max_bytes},
|
2020-02-17 19:28:25 +00:00
|
|
|
destination_id,
|
2018-01-12 13:03:19 +00:00
|
|
|
static_cast<bool>(args.local_context.getSettingsRef().insert_allow_materialized_columns));
|
2017-12-28 21:36:27 +00:00
|
|
|
});
|
|
|
|
}
|
|
|
|
|
2014-10-26 00:01:36 +00:00
|
|
|
}
|