#pragma once #include #include #include #include #include #include namespace DB { class IDataType; class CachedCompressedReadBuffer; class CompressedReadBufferFromFile; /// Reads the data between pairs of marks in the same part. When reading consecutive ranges, avoids unnecessary seeks. /// When ranges are almost consecutive, seeks are fast because they are performed inside the buffer. /// Avoids loading the marks file if it is not needed (e.g. when reading the whole part). class MergeTreeReader : private boost::noncopyable { public: using ValueSizeMap = std::map; using DeserializeBinaryBulkStateMap = std::map; MergeTreeReader(const String & path, /// Path to the directory containing the part const MergeTreeData::DataPartPtr & data_part, const NamesAndTypesList & columns, UncompressedCache * uncompressed_cache, MarkCache * mark_cache, bool save_marks_in_cache, MergeTreeData & storage, const MarkRanges & all_mark_ranges, size_t aio_threshold, size_t max_read_buffer_size, const ValueSizeMap & avg_value_size_hints = ValueSizeMap{}, const ReadBufferFromFileBase::ProfileCallback & profile_callback = ReadBufferFromFileBase::ProfileCallback{}, clockid_t clock_type = CLOCK_MONOTONIC_COARSE); ~MergeTreeReader(); const ValueSizeMap & getAvgValueSizeHints() const; /// Add columns from ordered_names that are not present in the block. /// Missing columns are added in the order specified by ordered_names. /// If at least one column was added, reorders all columns in the block according to ordered_names. void fillMissingColumns(Block & res, bool & should_reorder, bool & should_evaluate_missing_defaults); /// Sort columns to ensure consistent order among all blocks. /// If filter_name is not nullptr and block has filter column, move it to the end of block. void reorderColumns(Block & res, const Names & ordered_names, const String * filter_name); /// Evaluate defaulted columns if necessary. void evaluateMissingDefaults(Block & res); const NamesAndTypesList & getColumns() const { return columns; } private: class Stream { public: Stream( const String & path_prefix_, const String & extension_, size_t marks_count_, const MarkRanges & all_mark_ranges, MarkCache * mark_cache, bool save_marks_in_cache, UncompressedCache * uncompressed_cache, size_t aio_threshold, size_t max_read_buffer_size, const ReadBufferFromFileBase::ProfileCallback & profile_callback, clockid_t clock_type); void seekToMark(size_t index); void seekToStart(); ReadBuffer * data_buffer; private: Stream() = default; /// NOTE: lazily loads marks from the marks cache. const MarkInCompressedFile & getMark(size_t index); void loadMarks(); std::string path_prefix; std::string extension; size_t marks_count; MarkCache * mark_cache; bool save_marks_in_cache; MarkCache::MappedPtr marks; std::unique_ptr cached_buffer; std::unique_ptr non_cached_buffer; }; using FileStreams = std::map>; /// avg_value_size_hints are used to reduce the number of reallocations when creating columns of variable size. ValueSizeMap avg_value_size_hints; /// Stores states for IDataType::deserializeBinaryBulk DeserializeBinaryBulkStateMap deserialize_binary_bulk_state_map; String path; MergeTreeData::DataPartPtr data_part; FileStreams streams; /// Columns that are read. NamesAndTypesList columns; UncompressedCache * uncompressed_cache; MarkCache * mark_cache; /// If save_marks_in_cache is false, then, if marks are not in cache, we will load them but won't save in the cache, to avoid evicting other data. bool save_marks_in_cache; MergeTreeData & storage; MarkRanges all_mark_ranges; size_t aio_threshold; size_t max_read_buffer_size; size_t index_granularity; void addStreams(const String & name, const IDataType & type, const MarkRanges & all_mark_ranges, const ReadBufferFromFileBase::ProfileCallback & profile_callback, clockid_t clock_type); void readData( const String & name, const IDataType & type, IColumn & column, size_t from_mark, bool continue_reading, size_t max_rows_to_read, bool read_offsets = true); /// Return the number of rows has been read or zero if there is no columns to read. /// If continue_reading is true, continue reading from last state, otherwise seek to from_mark size_t readRows(size_t from_mark, bool continue_reading, size_t max_rows_to_read, Block & res); friend class MergeTreeRangeReader::DelayedStream; }; }