mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-16 03:12:43 +00:00
248 lines
7.6 KiB
C++
248 lines
7.6 KiB
C++
#include <Common/Exception.h>
|
|
|
|
#include <DataStreams/IBlockInputStream.h>
|
|
|
|
#include <Storages/StorageMemory.h>
|
|
#include <Storages/StorageFactory.h>
|
|
|
|
#include <IO/WriteHelpers.h>
|
|
#include <Processors/Sources/SourceWithProgress.h>
|
|
#include <Processors/Pipe.h>
|
|
|
|
|
|
namespace DB
|
|
{
|
|
|
|
namespace ErrorCodes
|
|
{
|
|
extern const int NUMBER_OF_ARGUMENTS_DOESNT_MATCH;
|
|
}
|
|
|
|
|
|
class MemorySource : public SourceWithProgress
|
|
{
|
|
using InitializerFunc = std::function<void(BlocksList::const_iterator &, size_t &)>;
|
|
public:
|
|
/// Blocks are stored in std::list which may be appended in another thread.
|
|
/// We use pointer to the beginning of the list and its current size.
|
|
/// We don't need synchronisation in this reader, because while we hold SharedLock on storage,
|
|
/// only new elements can be added to the back of the list, so our iterators remain valid
|
|
|
|
MemorySource(
|
|
Names column_names_,
|
|
BlocksList::const_iterator first_,
|
|
size_t num_blocks_,
|
|
const StorageMemory & storage,
|
|
const StorageMetadataPtr & metadata_snapshot,
|
|
InitializerFunc initializer_func_ = [](BlocksList::const_iterator &, size_t &) {})
|
|
: SourceWithProgress(metadata_snapshot->getSampleBlockForColumns(column_names_, storage.getVirtuals(), storage.getStorageID()))
|
|
, column_names(std::move(column_names_))
|
|
, current_it(first_)
|
|
, num_blocks(num_blocks_)
|
|
, initializer_func(std::move(initializer_func_))
|
|
{
|
|
}
|
|
|
|
String getName() const override { return "Memory"; }
|
|
|
|
protected:
|
|
Chunk generate() override
|
|
{
|
|
if (!postponed_init_done)
|
|
{
|
|
initializer_func(current_it, num_blocks);
|
|
postponed_init_done = true;
|
|
}
|
|
|
|
if (current_block_idx == num_blocks)
|
|
return {};
|
|
|
|
const Block & src = *current_it;
|
|
Columns columns;
|
|
columns.reserve(column_names.size());
|
|
|
|
/// Add only required columns to `res`.
|
|
for (const auto & name : column_names)
|
|
columns.push_back(src.getByName(name).column);
|
|
|
|
if (++current_block_idx < num_blocks)
|
|
++current_it;
|
|
|
|
return Chunk(std::move(columns), src.rows());
|
|
}
|
|
|
|
private:
|
|
const Names column_names;
|
|
BlocksList::const_iterator current_it;
|
|
size_t num_blocks;
|
|
size_t current_block_idx = 0;
|
|
|
|
bool postponed_init_done = false;
|
|
InitializerFunc initializer_func;
|
|
};
|
|
|
|
|
|
class MemoryBlockOutputStream : public IBlockOutputStream
|
|
{
|
|
public:
|
|
explicit MemoryBlockOutputStream(
|
|
StorageMemory & storage_,
|
|
const StorageMetadataPtr & metadata_snapshot_)
|
|
: storage(storage_)
|
|
, metadata_snapshot(metadata_snapshot_)
|
|
{}
|
|
|
|
Block getHeader() const override { return metadata_snapshot->getSampleBlock(); }
|
|
|
|
void write(const Block & block) override
|
|
{
|
|
const auto size_bytes_diff = block.allocatedBytes();
|
|
const auto size_rows_diff = block.rows();
|
|
|
|
metadata_snapshot->check(block, true);
|
|
{
|
|
std::lock_guard lock(storage.mutex);
|
|
storage.data.push_back(block);
|
|
|
|
storage.total_size_bytes.fetch_add(size_bytes_diff, std::memory_order_relaxed);
|
|
storage.total_size_rows.fetch_add(size_rows_diff, std::memory_order_relaxed);
|
|
}
|
|
|
|
}
|
|
private:
|
|
StorageMemory & storage;
|
|
StorageMetadataPtr metadata_snapshot;
|
|
};
|
|
|
|
|
|
StorageMemory::StorageMemory(const StorageID & table_id_, ColumnsDescription columns_description_, ConstraintsDescription constraints_)
|
|
: IStorage(table_id_)
|
|
{
|
|
StorageInMemoryMetadata storage_metadata;
|
|
storage_metadata.setColumns(std::move(columns_description_));
|
|
storage_metadata.setConstraints(std::move(constraints_));
|
|
setInMemoryMetadata(storage_metadata);
|
|
}
|
|
|
|
|
|
Pipe StorageMemory::read(
|
|
const Names & column_names,
|
|
const StorageMetadataPtr & metadata_snapshot,
|
|
SelectQueryInfo & /*query_info*/,
|
|
const Context & /*context*/,
|
|
QueryProcessingStage::Enum /*processed_stage*/,
|
|
size_t /*max_block_size*/,
|
|
unsigned num_streams)
|
|
{
|
|
metadata_snapshot->check(column_names, getVirtuals(), getStorageID());
|
|
|
|
if (delay_read_for_global_subqueries)
|
|
{
|
|
/// Note: for global subquery we use single source.
|
|
/// Mainly, the reason is that at this point table is empty,
|
|
/// and we don't know the number of blocks are going to be inserted into it.
|
|
///
|
|
/// It may seem to be not optimal, but actually data from such table is used to fill
|
|
/// set for IN or hash table for JOIN, which can't be done concurrently.
|
|
/// Since no other manipulation with data is done, multiple sources shouldn't give any profit.
|
|
|
|
return Pipe(
|
|
std::make_shared<MemorySource>(
|
|
column_names, data.end(), 0, *this, metadata_snapshot,
|
|
/// This hack is needed for global subqueries.
|
|
/// It allows to set up this Source for read AFTER Storage::read() has been called and just before actual reading
|
|
[this](BlocksList::const_iterator & current_it, size_t & num_blocks)
|
|
{
|
|
std::lock_guard guard(mutex);
|
|
current_it = data.begin();
|
|
num_blocks = data.size();
|
|
}
|
|
));
|
|
}
|
|
|
|
std::lock_guard lock(mutex);
|
|
|
|
size_t size = data.size();
|
|
|
|
if (num_streams > size)
|
|
num_streams = size;
|
|
|
|
Pipes pipes;
|
|
|
|
BlocksList::const_iterator it = data.begin();
|
|
|
|
size_t offset = 0;
|
|
for (size_t stream = 0; stream < num_streams; ++stream)
|
|
{
|
|
size_t next_offset = (stream + 1) * size / num_streams;
|
|
size_t num_blocks = next_offset - offset;
|
|
|
|
assert(num_blocks > 0);
|
|
|
|
pipes.emplace_back(std::make_shared<MemorySource>(column_names, it, num_blocks, *this, metadata_snapshot));
|
|
|
|
while (offset < next_offset)
|
|
{
|
|
++it;
|
|
++offset;
|
|
}
|
|
}
|
|
|
|
return Pipe::unitePipes(std::move(pipes));
|
|
}
|
|
|
|
|
|
BlockOutputStreamPtr StorageMemory::write(const ASTPtr & /*query*/, const StorageMetadataPtr & metadata_snapshot, const Context & /*context*/)
|
|
{
|
|
return std::make_shared<MemoryBlockOutputStream>(*this, metadata_snapshot);
|
|
}
|
|
|
|
|
|
void StorageMemory::drop()
|
|
{
|
|
std::lock_guard lock(mutex);
|
|
data.clear();
|
|
total_size_bytes.store(0, std::memory_order_relaxed);
|
|
total_size_rows.store(0, std::memory_order_relaxed);
|
|
}
|
|
|
|
|
|
void StorageMemory::truncate(
|
|
const ASTPtr &, const StorageMetadataPtr &, const Context &, TableExclusiveLockHolder &)
|
|
{
|
|
std::lock_guard lock(mutex);
|
|
data.clear();
|
|
total_size_bytes.store(0, std::memory_order_relaxed);
|
|
total_size_rows.store(0, std::memory_order_relaxed);
|
|
}
|
|
|
|
|
|
std::optional<UInt64> StorageMemory::totalRows() const
|
|
{
|
|
/// All modifications of these counters are done under mutex which automatically guarantees synchronization/consistency
|
|
/// When run concurrently we are fine with any value: "before" or "after"
|
|
return total_size_rows.load(std::memory_order_relaxed);
|
|
}
|
|
|
|
|
|
std::optional<UInt64> StorageMemory::totalBytes() const
|
|
{
|
|
return total_size_bytes.load(std::memory_order_relaxed);
|
|
}
|
|
|
|
|
|
void registerStorageMemory(StorageFactory & factory)
|
|
{
|
|
factory.registerStorage("Memory", [](const StorageFactory::Arguments & args)
|
|
{
|
|
if (!args.engine_args.empty())
|
|
throw Exception(
|
|
"Engine " + args.engine_name + " doesn't support any arguments (" + toString(args.engine_args.size()) + " given)",
|
|
ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
|
|
|
|
return StorageMemory::create(args.table_id, args.columns, args.constraints);
|
|
});
|
|
}
|
|
|
|
}
|