2017-03-24 13:52:50 +00:00
|
|
|
#include <Storages/MergeTree/MergeTreeBlockReadUtils.h>
|
|
|
|
#include <Storages/MergeTree/MergeTreeData.h>
|
2017-07-19 19:19:27 +00:00
|
|
|
#include <Common/typeid_cast.h>
|
|
|
|
#include <Columns/ColumnConst.h>
|
2017-07-20 20:07:59 +00:00
|
|
|
#include <unordered_set>
|
2017-03-24 13:52:50 +00:00
|
|
|
|
|
|
|
|
2018-01-10 00:04:08 +00:00
|
|
|
namespace DB
|
|
|
|
{
|
2020-02-25 18:10:48 +00:00
|
|
|
namespace ErrorCodes
|
|
|
|
{
|
|
|
|
extern const int LOGICAL_ERROR;
|
|
|
|
}
|
2017-03-24 13:52:50 +00:00
|
|
|
|
2019-11-21 16:10:22 +00:00
|
|
|
|
2017-03-24 13:52:50 +00:00
|
|
|
NameSet injectRequiredColumns(const MergeTreeData & storage, const MergeTreeData::DataPartPtr & part, Names & columns)
|
|
|
|
{
|
|
|
|
NameSet required_columns{std::begin(columns), std::end(columns)};
|
|
|
|
NameSet injected_columns;
|
|
|
|
|
|
|
|
auto all_column_files_missing = true;
|
|
|
|
|
2020-04-24 10:20:03 +00:00
|
|
|
const auto & storage_columns = storage.getColumns();
|
2020-05-15 10:26:44 +00:00
|
|
|
auto alter_conversions = storage.getAlterConversionsForPart(part);
|
2017-03-24 13:52:50 +00:00
|
|
|
for (size_t i = 0; i < columns.size(); ++i)
|
|
|
|
{
|
2020-05-15 10:26:44 +00:00
|
|
|
/// possibly renamed
|
|
|
|
auto column_name_in_part = columns[i];
|
|
|
|
|
|
|
|
if (alter_conversions.isColumnRenamed(column_name_in_part))
|
|
|
|
column_name_in_part = alter_conversions.getColumnOldName(column_name_in_part);
|
2017-03-24 13:52:50 +00:00
|
|
|
|
|
|
|
/// column has files and hence does not require evaluation
|
2020-05-15 10:26:44 +00:00
|
|
|
if (part->hasColumnFiles(column_name_in_part, *storage_columns.getPhysical(columns[i]).type))
|
2017-03-24 13:52:50 +00:00
|
|
|
{
|
|
|
|
all_column_files_missing = false;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2020-05-15 10:26:44 +00:00
|
|
|
const auto column_default = storage_columns.getDefault(columns[i]);
|
2019-03-14 15:20:51 +00:00
|
|
|
if (!column_default)
|
2017-03-24 13:52:50 +00:00
|
|
|
continue;
|
|
|
|
|
|
|
|
/// collect identifiers required for evaluation
|
|
|
|
IdentifierNameSet identifiers;
|
2019-03-14 15:20:51 +00:00
|
|
|
column_default->expression->collectIdentifierNames(identifiers);
|
2017-03-24 13:52:50 +00:00
|
|
|
|
|
|
|
for (const auto & identifier : identifiers)
|
|
|
|
{
|
2020-04-27 15:10:50 +00:00
|
|
|
if (storage_columns.hasPhysical(identifier))
|
2017-03-24 13:52:50 +00:00
|
|
|
{
|
|
|
|
/// ensure each column is added only once
|
|
|
|
if (required_columns.count(identifier) == 0)
|
|
|
|
{
|
|
|
|
columns.emplace_back(identifier);
|
|
|
|
required_columns.emplace(identifier);
|
|
|
|
injected_columns.emplace(identifier);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/** Add a column of the minimum size.
|
|
|
|
* Used in case when no column is needed or files are missing, but at least you need to know number of rows.
|
|
|
|
* Adds to the columns.
|
|
|
|
*/
|
|
|
|
if (all_column_files_missing)
|
|
|
|
{
|
|
|
|
const auto minimum_size_column_name = part->getColumnNameWithMinumumCompressedSize();
|
|
|
|
columns.push_back(minimum_size_column_name);
|
|
|
|
/// correctly report added column
|
|
|
|
injected_columns.insert(columns.back());
|
|
|
|
}
|
|
|
|
|
|
|
|
return injected_columns;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2017-04-06 17:21:45 +00:00
|
|
|
MergeTreeReadTask::MergeTreeReadTask(
|
2019-08-03 11:02:40 +00:00
|
|
|
const MergeTreeData::DataPartPtr & data_part_, const MarkRanges & mark_ranges_, const size_t part_index_in_query_,
|
|
|
|
const Names & ordered_names_, const NameSet & column_name_set_, const NamesAndTypesList & columns_,
|
|
|
|
const NamesAndTypesList & pre_columns_, const bool remove_prewhere_column_, const bool should_reorder_,
|
|
|
|
MergeTreeBlockSizePredictorPtr && size_predictor_)
|
|
|
|
: data_part{data_part_}, mark_ranges{mark_ranges_}, part_index_in_query{part_index_in_query_},
|
|
|
|
ordered_names{ordered_names_}, column_name_set{column_name_set_}, columns{columns_}, pre_columns{pre_columns_},
|
|
|
|
remove_prewhere_column{remove_prewhere_column_}, should_reorder{should_reorder_}, size_predictor{std::move(size_predictor_)}
|
2020-02-07 21:07:18 +00:00
|
|
|
{
|
|
|
|
}
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2017-04-05 20:34:19 +00:00
|
|
|
MergeTreeReadTask::~MergeTreeReadTask() = default;
|
|
|
|
|
2017-04-06 17:21:45 +00:00
|
|
|
|
|
|
|
MergeTreeBlockSizePredictor::MergeTreeBlockSizePredictor(
|
2017-05-18 16:03:08 +00:00
|
|
|
const MergeTreeData::DataPartPtr & data_part_, const Names & columns, const Block & sample_block)
|
2017-07-19 19:19:27 +00:00
|
|
|
: data_part(data_part_)
|
2017-04-06 17:21:45 +00:00
|
|
|
{
|
2017-10-24 14:11:53 +00:00
|
|
|
number_of_rows_in_part = data_part->rows_count;
|
2019-01-22 19:56:53 +00:00
|
|
|
/// Initialize with sample block until update won't called.
|
2019-09-26 17:29:41 +00:00
|
|
|
initialize(sample_block, {}, columns);
|
2017-07-19 19:19:27 +00:00
|
|
|
}
|
|
|
|
|
2019-09-26 17:29:41 +00:00
|
|
|
void MergeTreeBlockSizePredictor::initialize(const Block & sample_block, const Columns & columns, const Names & names, bool from_update)
|
2017-07-19 19:19:27 +00:00
|
|
|
{
|
|
|
|
fixed_columns_bytes_per_row = 0;
|
|
|
|
dynamic_columns_infos.clear();
|
|
|
|
|
2017-07-20 20:07:59 +00:00
|
|
|
std::unordered_set<String> names_set;
|
|
|
|
if (!from_update)
|
2019-09-26 17:29:41 +00:00
|
|
|
names_set.insert(names.begin(), names.end());
|
2017-07-20 20:07:59 +00:00
|
|
|
|
2019-09-26 17:29:41 +00:00
|
|
|
size_t num_columns = sample_block.columns();
|
|
|
|
for (size_t pos = 0; pos < num_columns; ++pos)
|
2017-04-06 17:21:45 +00:00
|
|
|
{
|
2019-09-26 17:29:41 +00:00
|
|
|
const auto & column_with_type_and_name = sample_block.getByPosition(pos);
|
2017-07-19 19:19:27 +00:00
|
|
|
const String & column_name = column_with_type_and_name.name;
|
2019-09-26 17:29:41 +00:00
|
|
|
const ColumnPtr & column_data = from_update ? columns[pos]
|
|
|
|
: column_with_type_and_name.column;
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2017-07-20 20:07:59 +00:00
|
|
|
if (!from_update && !names_set.count(column_name))
|
|
|
|
continue;
|
|
|
|
|
2017-07-19 19:19:27 +00:00
|
|
|
/// At least PREWHERE filter column might be const.
|
2017-07-21 06:40:05 +00:00
|
|
|
if (typeid_cast<const ColumnConst *>(column_data.get()))
|
2017-07-19 19:19:27 +00:00
|
|
|
continue;
|
2017-05-18 16:03:08 +00:00
|
|
|
|
2017-12-09 10:14:45 +00:00
|
|
|
if (column_data->valuesHaveFixedSize())
|
2017-04-06 17:21:45 +00:00
|
|
|
{
|
2017-12-09 10:14:45 +00:00
|
|
|
size_t size_of_value = column_data->sizeOfValueIfFixed();
|
|
|
|
fixed_columns_bytes_per_row += column_data->sizeOfValueIfFixed();
|
|
|
|
max_size_per_row_fixed = std::max<size_t>(max_size_per_row_fixed, size_of_value);
|
2017-04-06 17:21:45 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ColumnInfo info;
|
2017-05-18 16:03:08 +00:00
|
|
|
info.name = column_name;
|
2017-07-19 19:19:27 +00:00
|
|
|
/// If column isn't fixed and doesn't have checksum, than take first
|
2019-07-16 17:13:12 +00:00
|
|
|
ColumnSize column_size = data_part->getColumnSize(
|
2018-03-26 14:18:04 +00:00
|
|
|
column_name, *column_with_type_and_name.type);
|
|
|
|
|
|
|
|
info.bytes_per_row_global = column_size.data_uncompressed
|
|
|
|
? column_size.data_uncompressed / number_of_rows_in_part
|
2017-07-19 19:19:27 +00:00
|
|
|
: column_data->byteSize() / std::max<size_t>(1, column_data->size());
|
2017-04-06 17:21:45 +00:00
|
|
|
|
|
|
|
dynamic_columns_infos.emplace_back(info);
|
|
|
|
}
|
2017-07-19 19:19:27 +00:00
|
|
|
}
|
2017-04-06 17:21:45 +00:00
|
|
|
|
|
|
|
bytes_per_row_global = fixed_columns_bytes_per_row;
|
|
|
|
for (auto & info : dynamic_columns_infos)
|
|
|
|
{
|
|
|
|
info.bytes_per_row = info.bytes_per_row_global;
|
|
|
|
bytes_per_row_global += info.bytes_per_row_global;
|
2017-06-28 09:44:02 +00:00
|
|
|
|
|
|
|
max_size_per_row_dynamic = std::max<double>(max_size_per_row_dynamic, info.bytes_per_row);
|
2017-04-06 17:21:45 +00:00
|
|
|
}
|
|
|
|
bytes_per_row_current = bytes_per_row_global;
|
|
|
|
}
|
|
|
|
|
|
|
|
void MergeTreeBlockSizePredictor::startBlock()
|
|
|
|
{
|
|
|
|
block_size_bytes = 0;
|
|
|
|
block_size_rows = 0;
|
|
|
|
for (auto & info : dynamic_columns_infos)
|
|
|
|
info.size_bytes = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2017-05-18 16:03:08 +00:00
|
|
|
/// TODO: add last_read_row_in_part parameter to take into account gaps between adjacent ranges
|
2019-09-26 17:29:41 +00:00
|
|
|
void MergeTreeBlockSizePredictor::update(const Block & sample_block, const Columns & columns, size_t num_rows, double decay)
|
2017-04-07 11:43:24 +00:00
|
|
|
{
|
2019-09-26 17:29:41 +00:00
|
|
|
if (columns.size() != sample_block.columns())
|
|
|
|
throw Exception("Inconsistent number of columns passed to MergeTreeBlockSizePredictor. "
|
|
|
|
"Have " + toString(sample_block.columns()) + " in sample block "
|
|
|
|
"and " + toString(columns.size()) + " columns in list", ErrorCodes::LOGICAL_ERROR);
|
|
|
|
|
2017-07-19 19:19:27 +00:00
|
|
|
if (!is_initialized_in_update)
|
|
|
|
{
|
|
|
|
/// Reinitialize with read block to update estimation for DEFAULT and MATERIALIZED columns without data.
|
2019-09-26 17:29:41 +00:00
|
|
|
initialize(sample_block, columns, {}, true);
|
2017-07-19 19:19:27 +00:00
|
|
|
is_initialized_in_update = true;
|
|
|
|
}
|
2019-09-26 17:29:41 +00:00
|
|
|
|
|
|
|
if (num_rows < block_size_rows)
|
2017-04-10 14:06:44 +00:00
|
|
|
{
|
2019-09-26 17:29:41 +00:00
|
|
|
throw Exception("Updated block has less rows (" + toString(num_rows) + ") than previous one (" + toString(block_size_rows) + ")",
|
2017-04-10 14:06:44 +00:00
|
|
|
ErrorCodes::LOGICAL_ERROR);
|
|
|
|
}
|
|
|
|
|
2019-09-26 17:29:41 +00:00
|
|
|
size_t diff_rows = num_rows - block_size_rows;
|
|
|
|
block_size_bytes = num_rows * fixed_columns_bytes_per_row;
|
2017-04-07 11:43:24 +00:00
|
|
|
bytes_per_row_current = fixed_columns_bytes_per_row;
|
2019-09-26 17:29:41 +00:00
|
|
|
block_size_rows = num_rows;
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2017-04-10 14:06:44 +00:00
|
|
|
/// Make recursive updates for each read row: v_{i+1} = (1 - decay) v_{i} + decay v_{target}
|
2019-01-22 19:56:53 +00:00
|
|
|
/// Use sum of geometric sequence formula to update multiple rows: v{n} = (1 - decay)^n v_{0} + (1 - (1 - decay)^n) v_{target}
|
2017-07-19 19:19:27 +00:00
|
|
|
/// NOTE: DEFAULT and MATERIALIZED columns without data has inaccurate estimation of v_{target}
|
2017-04-15 03:32:33 +00:00
|
|
|
double alpha = std::pow(1. - decay, diff_rows);
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2017-06-28 09:44:02 +00:00
|
|
|
max_size_per_row_dynamic = 0;
|
2017-04-07 11:43:24 +00:00
|
|
|
for (auto & info : dynamic_columns_infos)
|
|
|
|
{
|
2019-09-26 17:29:41 +00:00
|
|
|
size_t new_size = columns[sample_block.getPositionByName(info.name)]->byteSize();
|
2017-04-15 03:32:33 +00:00
|
|
|
size_t diff_size = new_size - info.size_bytes;
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2017-04-15 03:32:33 +00:00
|
|
|
double local_bytes_per_row = static_cast<double>(diff_size) / diff_rows;
|
2017-04-07 11:43:24 +00:00
|
|
|
info.bytes_per_row = alpha * info.bytes_per_row + (1. - alpha) * local_bytes_per_row;
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2017-04-07 11:43:24 +00:00
|
|
|
info.size_bytes = new_size;
|
|
|
|
block_size_bytes += new_size;
|
|
|
|
bytes_per_row_current += info.bytes_per_row;
|
2017-06-28 09:44:02 +00:00
|
|
|
|
|
|
|
max_size_per_row_dynamic = std::max<double>(max_size_per_row_dynamic, info.bytes_per_row);
|
2017-04-06 17:21:45 +00:00
|
|
|
}
|
2017-04-07 11:43:24 +00:00
|
|
|
}
|
2017-04-06 17:21:45 +00:00
|
|
|
|
2019-07-19 14:56:00 +00:00
|
|
|
|
|
|
|
MergeTreeReadTaskColumns getReadTaskColumns(const MergeTreeData & storage, const MergeTreeData::DataPartPtr & data_part,
|
|
|
|
const Names & required_columns, const PrewhereInfoPtr & prewhere_info, bool check_columns)
|
|
|
|
{
|
|
|
|
Names column_names = required_columns;
|
|
|
|
Names pre_column_names;
|
|
|
|
|
|
|
|
/// inject columns required for defaults evaluation
|
|
|
|
bool should_reorder = !injectRequiredColumns(storage, data_part, column_names).empty();
|
|
|
|
|
|
|
|
if (prewhere_info)
|
|
|
|
{
|
|
|
|
if (prewhere_info->alias_actions)
|
|
|
|
pre_column_names = prewhere_info->alias_actions->getRequiredColumns();
|
|
|
|
else
|
|
|
|
pre_column_names = prewhere_info->prewhere_actions->getRequiredColumns();
|
|
|
|
|
|
|
|
if (pre_column_names.empty())
|
|
|
|
pre_column_names.push_back(column_names[0]);
|
|
|
|
|
|
|
|
const auto injected_pre_columns = injectRequiredColumns(storage, data_part, pre_column_names);
|
|
|
|
if (!injected_pre_columns.empty())
|
|
|
|
should_reorder = true;
|
|
|
|
|
|
|
|
const NameSet pre_name_set(pre_column_names.begin(), pre_column_names.end());
|
|
|
|
|
|
|
|
Names post_column_names;
|
|
|
|
for (const auto & name : column_names)
|
|
|
|
if (!pre_name_set.count(name))
|
|
|
|
post_column_names.push_back(name);
|
|
|
|
|
|
|
|
column_names = post_column_names;
|
|
|
|
}
|
|
|
|
|
|
|
|
MergeTreeReadTaskColumns result;
|
|
|
|
|
|
|
|
if (check_columns)
|
|
|
|
{
|
2020-04-27 14:17:12 +00:00
|
|
|
const NamesAndTypesList & physical_columns = storage.getColumns().getAllPhysical();
|
|
|
|
result.pre_columns = physical_columns.addTypes(pre_column_names);
|
|
|
|
result.columns = physical_columns.addTypes(column_names);
|
2019-07-19 14:56:00 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2020-01-16 16:15:01 +00:00
|
|
|
result.pre_columns = data_part->getColumns().addTypes(pre_column_names);
|
|
|
|
result.columns = data_part->getColumns().addTypes(column_names);
|
2019-07-19 14:56:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
result.should_reorder = should_reorder;
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2017-03-24 13:52:50 +00:00
|
|
|
}
|