mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-21 23:21:59 +00:00
97f2a2213e
* Move some code outside dbms/src folder * Fix paths
341 lines
13 KiB
C++
341 lines
13 KiB
C++
#include <optional>
|
|
#include <Storages/System/StorageSystemColumns.h>
|
|
#include <Storages/MergeTree/MergeTreeData.h>
|
|
#include <Columns/ColumnsNumber.h>
|
|
#include <Columns/ColumnString.h>
|
|
#include <DataTypes/DataTypeString.h>
|
|
#include <DataTypes/DataTypesNumber.h>
|
|
#include <DataStreams/NullBlockInputStream.h>
|
|
#include <Storages/VirtualColumnUtils.h>
|
|
#include <Parsers/queryToString.h>
|
|
#include <Parsers/ASTSelectQuery.h>
|
|
#include <Access/ContextAccess.h>
|
|
#include <Databases/IDatabase.h>
|
|
#include <Processors/Sources/NullSource.h>
|
|
|
|
|
|
namespace DB
|
|
{
|
|
|
|
namespace ErrorCodes
|
|
{
|
|
extern const int TABLE_IS_DROPPED;
|
|
}
|
|
|
|
StorageSystemColumns::StorageSystemColumns(const std::string & name_)
|
|
: IStorage({"system", name_})
|
|
{
|
|
setColumns(ColumnsDescription(
|
|
{
|
|
{ "database", std::make_shared<DataTypeString>() },
|
|
{ "table", std::make_shared<DataTypeString>() },
|
|
{ "name", std::make_shared<DataTypeString>() },
|
|
{ "type", std::make_shared<DataTypeString>() },
|
|
{ "default_kind", std::make_shared<DataTypeString>() },
|
|
{ "default_expression", std::make_shared<DataTypeString>() },
|
|
{ "data_compressed_bytes", std::make_shared<DataTypeUInt64>() },
|
|
{ "data_uncompressed_bytes", std::make_shared<DataTypeUInt64>() },
|
|
{ "marks_bytes", std::make_shared<DataTypeUInt64>() },
|
|
{ "comment", std::make_shared<DataTypeString>() },
|
|
{ "is_in_partition_key", std::make_shared<DataTypeUInt8>() },
|
|
{ "is_in_sorting_key", std::make_shared<DataTypeUInt8>() },
|
|
{ "is_in_primary_key", std::make_shared<DataTypeUInt8>() },
|
|
{ "is_in_sampling_key", std::make_shared<DataTypeUInt8>() },
|
|
{ "compression_codec", std::make_shared<DataTypeString>() },
|
|
}));
|
|
}
|
|
|
|
|
|
namespace
|
|
{
|
|
using Storages = std::map<std::pair<std::string, std::string>, StoragePtr>;
|
|
}
|
|
|
|
|
|
class ColumnsSource : public SourceWithProgress
|
|
{
|
|
public:
|
|
ColumnsSource(
|
|
std::vector<UInt8> columns_mask_,
|
|
Block header_,
|
|
UInt64 max_block_size_,
|
|
ColumnPtr databases_,
|
|
ColumnPtr tables_,
|
|
Storages storages_,
|
|
const std::shared_ptr<const ContextAccess> & access_,
|
|
String query_id_)
|
|
: SourceWithProgress(header_)
|
|
, columns_mask(std::move(columns_mask_)), max_block_size(max_block_size_)
|
|
, databases(std::move(databases_)), tables(std::move(tables_)), storages(std::move(storages_))
|
|
, query_id(std::move(query_id_)), total_tables(tables->size()), access(access_)
|
|
{
|
|
}
|
|
|
|
String getName() const override { return "Columns"; }
|
|
|
|
protected:
|
|
Chunk generate() override
|
|
{
|
|
if (db_table_num >= total_tables)
|
|
return {};
|
|
|
|
MutableColumns res_columns = getPort().getHeader().cloneEmptyColumns();
|
|
size_t rows_count = 0;
|
|
|
|
const bool check_access_for_tables = !access->isGranted(AccessType::SHOW_COLUMNS);
|
|
|
|
while (rows_count < max_block_size && db_table_num < total_tables)
|
|
{
|
|
const std::string database_name = (*databases)[db_table_num].get<std::string>();
|
|
const std::string table_name = (*tables)[db_table_num].get<std::string>();
|
|
++db_table_num;
|
|
|
|
ColumnsDescription columns;
|
|
Names cols_required_for_partition_key;
|
|
Names cols_required_for_sorting_key;
|
|
Names cols_required_for_primary_key;
|
|
Names cols_required_for_sampling;
|
|
MergeTreeData::ColumnSizeByName column_sizes;
|
|
|
|
{
|
|
StoragePtr storage = storages.at(std::make_pair(database_name, table_name));
|
|
TableStructureReadLockHolder table_lock;
|
|
|
|
try
|
|
{
|
|
table_lock = storage->lockStructureForShare(false, query_id);
|
|
}
|
|
catch (const Exception & e)
|
|
{
|
|
/** There are case when IStorage::drop was called,
|
|
* but we still own the object.
|
|
* Then table will throw exception at attempt to lock it.
|
|
* Just skip the table.
|
|
*/
|
|
if (e.code() == ErrorCodes::TABLE_IS_DROPPED)
|
|
continue;
|
|
else
|
|
throw;
|
|
}
|
|
|
|
columns = storage->getColumns();
|
|
|
|
cols_required_for_partition_key = storage->getColumnsRequiredForPartitionKey();
|
|
cols_required_for_sorting_key = storage->getColumnsRequiredForSortingKey();
|
|
cols_required_for_primary_key = storage->getColumnsRequiredForPrimaryKey();
|
|
cols_required_for_sampling = storage->getColumnsRequiredForSampling();
|
|
|
|
column_sizes = storage->getColumnSizes();
|
|
}
|
|
|
|
bool check_access_for_columns = check_access_for_tables && !access->isGranted(AccessType::SHOW_COLUMNS, database_name, table_name);
|
|
|
|
for (const auto & column : columns)
|
|
{
|
|
if (column.is_virtual)
|
|
continue;
|
|
|
|
if (check_access_for_columns && !access->isGranted(AccessType::SHOW_COLUMNS, database_name, table_name, column.name))
|
|
continue;
|
|
|
|
size_t src_index = 0;
|
|
size_t res_index = 0;
|
|
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(database_name);
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(table_name);
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(column.name);
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(column.type->getName());
|
|
|
|
if (column.default_desc.expression)
|
|
{
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(toString(column.default_desc.kind));
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(queryToString(column.default_desc.expression));
|
|
}
|
|
else
|
|
{
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insertDefault();
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insertDefault();
|
|
}
|
|
|
|
{
|
|
const auto it = column_sizes.find(column.name);
|
|
if (it == std::end(column_sizes))
|
|
{
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insertDefault();
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insertDefault();
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insertDefault();
|
|
}
|
|
else
|
|
{
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(it->second.data_compressed);
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(it->second.data_uncompressed);
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(it->second.marks);
|
|
}
|
|
}
|
|
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(column.comment);
|
|
|
|
{
|
|
auto find_in_vector = [&key = column.name](const Names& names)
|
|
{
|
|
return std::find(names.cbegin(), names.cend(), key) != names.end();
|
|
};
|
|
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(find_in_vector(cols_required_for_partition_key));
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(find_in_vector(cols_required_for_sorting_key));
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(find_in_vector(cols_required_for_primary_key));
|
|
if (columns_mask[src_index++])
|
|
res_columns[res_index++]->insert(find_in_vector(cols_required_for_sampling));
|
|
}
|
|
|
|
if (columns_mask[src_index++])
|
|
{
|
|
if (column.codec)
|
|
res_columns[res_index++]->insert("CODEC(" + column.codec->getCodecDesc() + ")");
|
|
else
|
|
res_columns[res_index++]->insertDefault();
|
|
}
|
|
|
|
++rows_count;
|
|
}
|
|
}
|
|
|
|
return Chunk(std::move(res_columns), rows_count);
|
|
}
|
|
|
|
private:
|
|
std::vector<UInt8> columns_mask;
|
|
UInt64 max_block_size;
|
|
ColumnPtr databases;
|
|
ColumnPtr tables;
|
|
Storages storages;
|
|
String query_id;
|
|
size_t db_table_num = 0;
|
|
size_t total_tables;
|
|
std::shared_ptr<const ContextAccess> access;
|
|
};
|
|
|
|
|
|
Pipes StorageSystemColumns::read(
|
|
const Names & column_names,
|
|
const SelectQueryInfo & query_info,
|
|
const Context & context,
|
|
QueryProcessingStage::Enum /*processed_stage*/,
|
|
const size_t max_block_size,
|
|
const unsigned /*num_streams*/)
|
|
{
|
|
check(column_names);
|
|
|
|
/// Create a mask of what columns are needed in the result.
|
|
|
|
NameSet names_set(column_names.begin(), column_names.end());
|
|
|
|
Block sample_block = getSampleBlock();
|
|
Block header;
|
|
|
|
std::vector<UInt8> columns_mask(sample_block.columns());
|
|
for (size_t i = 0, size = columns_mask.size(); i < size; ++i)
|
|
{
|
|
if (names_set.count(sample_block.getByPosition(i).name))
|
|
{
|
|
columns_mask[i] = 1;
|
|
header.insert(sample_block.getByPosition(i));
|
|
}
|
|
}
|
|
|
|
Block block_to_filter;
|
|
Storages storages;
|
|
Pipes pipes;
|
|
|
|
{
|
|
Databases databases = DatabaseCatalog::instance().getDatabases();
|
|
|
|
/// Add `database` column.
|
|
MutableColumnPtr database_column_mut = ColumnString::create();
|
|
for (const auto & database : databases)
|
|
{
|
|
/// We are skipping "Lazy" database because we cannot afford initialization of all its tables.
|
|
/// This should be documented.
|
|
|
|
if (database.second->getEngineName() != "Lazy")
|
|
database_column_mut->insert(database.first);
|
|
}
|
|
|
|
block_to_filter.insert(ColumnWithTypeAndName(std::move(database_column_mut), std::make_shared<DataTypeString>(), "database"));
|
|
|
|
/// Filter block with `database` column.
|
|
VirtualColumnUtils::filterBlockWithQuery(query_info.query, block_to_filter, context);
|
|
|
|
if (!block_to_filter.rows())
|
|
{
|
|
pipes.emplace_back(std::make_shared<NullSource>(header));
|
|
return pipes;
|
|
}
|
|
|
|
ColumnPtr & database_column = block_to_filter.getByName("database").column;
|
|
size_t rows = database_column->size();
|
|
|
|
/// Add `table` column.
|
|
MutableColumnPtr table_column_mut = ColumnString::create();
|
|
IColumn::Offsets offsets(rows);
|
|
for (size_t i = 0; i < rows; ++i)
|
|
{
|
|
const std::string database_name = (*database_column)[i].get<std::string>();
|
|
const DatabasePtr database = databases.at(database_name);
|
|
offsets[i] = i ? offsets[i - 1] : 0;
|
|
|
|
for (auto iterator = database->getTablesWithDictionaryTablesIterator(context); iterator->isValid(); iterator->next())
|
|
{
|
|
const String & table_name = iterator->name();
|
|
storages.emplace(std::piecewise_construct,
|
|
std::forward_as_tuple(database_name, table_name),
|
|
std::forward_as_tuple(iterator->table()));
|
|
table_column_mut->insert(table_name);
|
|
++offsets[i];
|
|
}
|
|
}
|
|
|
|
database_column = database_column->replicate(offsets);
|
|
block_to_filter.insert(ColumnWithTypeAndName(std::move(table_column_mut), std::make_shared<DataTypeString>(), "table"));
|
|
}
|
|
|
|
/// Filter block with `database` and `table` columns.
|
|
VirtualColumnUtils::filterBlockWithQuery(query_info.query, block_to_filter, context);
|
|
|
|
if (!block_to_filter.rows())
|
|
{
|
|
pipes.emplace_back(std::make_shared<NullSource>(header));
|
|
return pipes;
|
|
}
|
|
|
|
ColumnPtr filtered_database_column = block_to_filter.getByName("database").column;
|
|
ColumnPtr filtered_table_column = block_to_filter.getByName("table").column;
|
|
|
|
pipes.emplace_back(std::make_shared<ColumnsSource>(
|
|
std::move(columns_mask), std::move(header), max_block_size,
|
|
std::move(filtered_database_column), std::move(filtered_table_column), std::move(storages),
|
|
context.getAccess(), context.getCurrentQueryId()));
|
|
|
|
return pipes;
|
|
}
|
|
|
|
}
|