mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-16 03:12:43 +00:00
fix building and stateless test
This commit is contained in:
parent
e5ff05c50d
commit
05dfc4eb25
@ -907,7 +907,7 @@ void IMergeTreeDataPart::loadDefaultCompressionCodec()
|
||||
}
|
||||
}
|
||||
|
||||
void IMergeTreeDataPart::appendFilesOfDefaultCompressionCodec(Strings & files) const
|
||||
void IMergeTreeDataPart::appendFilesOfDefaultCompressionCodec(Strings & files)
|
||||
{
|
||||
files.push_back(DEFAULT_COMPRESSION_CODEC_FILE_NAME);
|
||||
}
|
||||
@ -1075,7 +1075,7 @@ void IMergeTreeDataPart::loadChecksums(bool require)
|
||||
}
|
||||
}
|
||||
|
||||
void IMergeTreeDataPart::appendFilesOfChecksums(Strings & files) const
|
||||
void IMergeTreeDataPart::appendFilesOfChecksums(Strings & files)
|
||||
{
|
||||
files.push_back("checksums.txt");
|
||||
}
|
||||
@ -1215,7 +1215,7 @@ void IMergeTreeDataPart::loadRowsCount()
|
||||
}
|
||||
}
|
||||
|
||||
void IMergeTreeDataPart::appendFilesOfRowsCount(Strings & files) const
|
||||
void IMergeTreeDataPart::appendFilesOfRowsCount(Strings & files)
|
||||
{
|
||||
files.push_back("count.txt");
|
||||
}
|
||||
@ -1269,7 +1269,7 @@ void IMergeTreeDataPart::loadTTLInfos()
|
||||
}
|
||||
|
||||
|
||||
void IMergeTreeDataPart::appendFilesOfTTLInfos(Strings & files) const
|
||||
void IMergeTreeDataPart::appendFilesOfTTLInfos(Strings & files)
|
||||
{
|
||||
files.push_back("ttl.txt");
|
||||
}
|
||||
@ -1307,7 +1307,7 @@ void IMergeTreeDataPart::loadUUID()
|
||||
}
|
||||
}
|
||||
|
||||
void IMergeTreeDataPart::appendFilesOfUUID(Strings & files) const
|
||||
void IMergeTreeDataPart::appendFilesOfUUID(Strings & files)
|
||||
{
|
||||
files.push_back(UUID_FILE_NAME);
|
||||
}
|
||||
@ -1388,7 +1388,7 @@ void IMergeTreeDataPart::loadColumns(bool require)
|
||||
setColumns(loaded_columns, infos);
|
||||
}
|
||||
|
||||
void IMergeTreeDataPart::appendFilesOfColumns(Strings & files) const
|
||||
void IMergeTreeDataPart::appendFilesOfColumns(Strings & files)
|
||||
{
|
||||
files.push_back("columns.txt");
|
||||
}
|
||||
@ -1475,7 +1475,6 @@ void IMergeTreeDataPart::modifyAllMetadataCaches(ModifyCacheType type, bool incl
|
||||
assert(use_metadata_cache);
|
||||
|
||||
Strings files;
|
||||
files.reserve(16);
|
||||
appendFilesOfColumnsChecksumsIndexes(files, include_projection);
|
||||
LOG_TRACE(
|
||||
storage.log,
|
||||
|
@ -503,17 +503,17 @@ private:
|
||||
/// Reads part unique identifier (if exists) from uuid.txt
|
||||
void loadUUID();
|
||||
|
||||
void appendFilesOfUUID(Strings & files) const;
|
||||
static void appendFilesOfUUID(Strings & files);
|
||||
|
||||
/// Reads columns names and types from columns.txt
|
||||
void loadColumns(bool require);
|
||||
|
||||
void appendFilesOfColumns(Strings & files) const;
|
||||
static void appendFilesOfColumns(Strings & files);
|
||||
|
||||
/// If checksums.txt exists, reads file's checksums (and sizes) from it
|
||||
void loadChecksums(bool require);
|
||||
|
||||
void appendFilesOfChecksums(Strings & files) const;
|
||||
static void appendFilesOfChecksums(Strings & files);
|
||||
|
||||
/// Loads marks index granularity into memory
|
||||
virtual void loadIndexGranularity();
|
||||
@ -529,12 +529,12 @@ private:
|
||||
/// For the older format version calculates rows count from the size of a column with a fixed size.
|
||||
void loadRowsCount();
|
||||
|
||||
void appendFilesOfRowsCount(Strings & files) const;
|
||||
static void appendFilesOfRowsCount(Strings & files);
|
||||
|
||||
/// Loads ttl infos in json format from file ttl.txt. If file doesn't exists assigns ttl infos with all zeros
|
||||
void loadTTLInfos();
|
||||
|
||||
void appendFilesOfTTLInfos(Strings & files) const;
|
||||
static void appendFilesOfTTLInfos(Strings & files);
|
||||
|
||||
void loadPartitionAndMinMaxIndex();
|
||||
|
||||
@ -549,7 +549,7 @@ private:
|
||||
/// any specifial compression.
|
||||
void loadDefaultCompressionCodec();
|
||||
|
||||
void appendFilesOfDefaultCompressionCodec(Strings & files) const;
|
||||
static void appendFilesOfDefaultCompressionCodec(Strings & files);
|
||||
|
||||
/// Found column without specific compression and return codec
|
||||
/// for this column with default parameters.
|
||||
|
@ -2,8 +2,8 @@
|
||||
|
||||
-- Create table under database with engine ordinary.
|
||||
set mutations_sync = 1;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
DROP TABLE IF EXISTS test_metadata_cache.check_part_metadata_cache;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
CREATE DATABASE test_metadata_cache ENGINE = Ordinary;
|
||||
CREATE TABLE test_metadata_cache.check_part_metadata_cache( p Date, k UInt64, v1 UInt64, v2 Int64) ENGINE MergeTree() PARTITION BY toYYYYMM(p) ORDER BY k settings use_metadata_cache = 1;
|
||||
with arrayJoin(checkPartMetadataCache('test_metadata_cache', 'check_part_metadata_cache')) as info select countIf(info.5 = 0);
|
||||
|
@ -2,8 +2,8 @@
|
||||
|
||||
-- Create table under database with engine atomic.
|
||||
set mutations_sync = 1;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
DROP TABLE IF EXISTS test_metadata_cache.check_part_metadata_cache SYNC;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
CREATE DATABASE test_metadata_cache ENGINE = Atomic;
|
||||
CREATE TABLE test_metadata_cache.check_part_metadata_cache( p Date, k UInt64, v1 UInt64, v2 Int64) ENGINE MergeTree() PARTITION BY toYYYYMM(p) ORDER BY k settings use_metadata_cache = 1;
|
||||
with arrayJoin(checkPartMetadataCache('test_metadata_cache', 'check_part_metadata_cache')) as info select countIf(info.5 = 0);
|
||||
|
@ -4,8 +4,8 @@
|
||||
-- Create table under database with engine ordinary.
|
||||
set mutations_sync = 1;
|
||||
set replication_alter_partitions_sync = 2;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
DROP TABLE IF EXISTS test_metadata_cache.check_part_metadata_cache;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
CREATE DATABASE test_metadata_cache ENGINE = Ordinary;
|
||||
CREATE TABLE test_metadata_cache.check_part_metadata_cache ( p Date, k UInt64, v1 UInt64, v2 Int64) ENGINE ReplicatedMergeTree('/clickhouse/tables/{database}/check_part_metadata_cache', '{replica}') PARTITION BY toYYYYMM(p) ORDER BY k settings use_metadata_cache = 1;
|
||||
with arrayJoin(checkPartMetadataCache('test_metadata_cache', 'check_part_metadata_cache')) as info select countIf(info.5 = 0);
|
||||
|
@ -4,8 +4,8 @@
|
||||
-- Create table under database with engine ordinary.
|
||||
set mutations_sync = 1;
|
||||
set replication_alter_partitions_sync = 2;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache ;
|
||||
DROP TABLE IF EXISTS test_metadata_cache.check_part_metadata_cache SYNC;
|
||||
DROP DATABASE IF EXISTS test_metadata_cache;
|
||||
CREATE DATABASE test_metadata_cache ENGINE = Atomic;
|
||||
CREATE TABLE test_metadata_cache.check_part_metadata_cache ( p Date, k UInt64, v1 UInt64, v2 Int64) ENGINE ReplicatedMergeTree('/clickhouse/tables/{database}/check_part_metadata_cache', '{replica}') PARTITION BY toYYYYMM(p) ORDER BY k settings use_metadata_cache = 1;
|
||||
with arrayJoin(checkPartMetadataCache('test_metadata_cache', 'check_part_metadata_cache')) as info select countIf(info.5 = 0);
|
||||
|
Loading…
Reference in New Issue
Block a user