mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-19 06:01:57 +00:00
479 lines
12 KiB
C++
479 lines
12 KiB
C++
#include <iostream>
|
|
#include <iomanip>
|
|
#include <vector>
|
|
|
|
#include <unordered_map>
|
|
|
|
#include <sparsehash/dense_hash_map>
|
|
#include <sparsehash/sparse_hash_map>
|
|
|
|
#include <statdaemons/Stopwatch.h>
|
|
|
|
//#define DBMS_HASH_MAP_COUNT_COLLISIONS
|
|
#define DBMS_HASH_MAP_DEBUG_RESIZES
|
|
|
|
#include <DB/Core/Types.h>
|
|
#include <DB/IO/ReadBufferFromFile.h>
|
|
#include <DB/IO/ReadHelpers.h>
|
|
#include <DB/IO/CompressedReadBuffer.h>
|
|
#include <DB/Core/StringRef.h>
|
|
#include <DB/Common/HashTable/HashMap.h>
|
|
#include <DB/Interpreters/AggregationCommon.h>
|
|
|
|
|
|
struct CompactStringRef
|
|
{
|
|
union
|
|
{
|
|
const char * data_mixed = nullptr;
|
|
struct
|
|
{
|
|
char dummy[6];
|
|
UInt16 size;
|
|
};
|
|
};
|
|
|
|
CompactStringRef(const char * data_, size_t size_)
|
|
{
|
|
data_mixed = data_;
|
|
size = size_;
|
|
}
|
|
|
|
CompactStringRef(const unsigned char * data_, size_t size_) : CompactStringRef(reinterpret_cast<const char *>(data_), size_) {}
|
|
CompactStringRef(const std::string & s) : CompactStringRef(s.data(), s.size()) {}
|
|
CompactStringRef() {}
|
|
|
|
const char * data() const { return reinterpret_cast<const char *>(reinterpret_cast<intptr_t>(data_mixed) & 0x0000FFFFFFFFFFFFULL); }
|
|
|
|
std::string toString() const { return std::string(data(), size); }
|
|
};
|
|
|
|
inline bool operator==(CompactStringRef lhs, CompactStringRef rhs)
|
|
{
|
|
if (lhs.size != rhs.size)
|
|
return false;
|
|
|
|
const char * lhs_data = lhs.data();
|
|
const char * rhs_data = rhs.data();
|
|
for (size_t pos = lhs.size - 1; pos < lhs.size; --pos)
|
|
if (lhs_data[pos] != rhs_data[pos])
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
namespace ZeroTraits
|
|
{
|
|
template <>
|
|
inline bool check<CompactStringRef>(CompactStringRef x) { return nullptr == x.data_mixed; }
|
|
|
|
template <>
|
|
inline void set<CompactStringRef>(CompactStringRef & x) { x.data_mixed = nullptr; }
|
|
};
|
|
|
|
template <>
|
|
struct DefaultHash<CompactStringRef>
|
|
{
|
|
size_t operator() (CompactStringRef x) const
|
|
{
|
|
return CityHash64(x.data(), x.size);
|
|
}
|
|
};
|
|
|
|
|
|
#define mix(h) ({ \
|
|
(h) ^= (h) >> 23; \
|
|
(h) *= 0x2127599bf4325c37ULL; \
|
|
(h) ^= (h) >> 47; })
|
|
|
|
struct FastHash64
|
|
{
|
|
size_t operator() (CompactStringRef x) const
|
|
{
|
|
const char * buf = x.data();
|
|
size_t len = x.size;
|
|
|
|
const uint64_t m = 0x880355f21e6d1965ULL;
|
|
const uint64_t *pos = (const uint64_t *)buf;
|
|
const uint64_t *end = pos + (len / 8);
|
|
const unsigned char *pos2;
|
|
uint64_t h = len * m;
|
|
uint64_t v;
|
|
|
|
while (pos != end) {
|
|
v = *pos++;
|
|
h ^= mix(v);
|
|
h *= m;
|
|
}
|
|
|
|
pos2 = (const unsigned char*)pos;
|
|
v = 0;
|
|
|
|
switch (len & 7) {
|
|
case 7: v ^= (uint64_t)pos2[6] << 48;
|
|
case 6: v ^= (uint64_t)pos2[5] << 40;
|
|
case 5: v ^= (uint64_t)pos2[4] << 32;
|
|
case 4: v ^= (uint64_t)pos2[3] << 24;
|
|
case 3: v ^= (uint64_t)pos2[2] << 16;
|
|
case 2: v ^= (uint64_t)pos2[1] << 8;
|
|
case 1: v ^= (uint64_t)pos2[0];
|
|
h ^= mix(v);
|
|
h *= m;
|
|
}
|
|
|
|
return mix(h);
|
|
}
|
|
};
|
|
|
|
|
|
struct CrapWow
|
|
{
|
|
size_t operator() (CompactStringRef x) const
|
|
{
|
|
const char * key = x.data();
|
|
size_t len = x.size;
|
|
size_t seed = 0;
|
|
|
|
const UInt64 m = 0x95b47aa3355ba1a1, n = 0x8a970be7488fda55;
|
|
UInt64 hash;
|
|
// 3 = m, 4 = n
|
|
// r12 = h, r13 = k, ecx = seed, r12 = key
|
|
asm(
|
|
"leaq (%%rcx,%4), %%r13\n"
|
|
"movq %%rdx, %%r14\n"
|
|
"movq %%rcx, %%r15\n"
|
|
"movq %%rcx, %%r12\n"
|
|
"addq %%rax, %%r13\n"
|
|
"andq $0xfffffffffffffff0, %%rcx\n"
|
|
"jz QW%=\n"
|
|
"addq %%rcx, %%r14\n\n"
|
|
"negq %%rcx\n"
|
|
"XW%=:\n"
|
|
"movq %4, %%rax\n"
|
|
"mulq (%%r14,%%rcx)\n"
|
|
"xorq %%rax, %%r12\n"
|
|
"xorq %%rdx, %%r13\n"
|
|
"movq %3, %%rax\n"
|
|
"mulq 8(%%r14,%%rcx)\n"
|
|
"xorq %%rdx, %%r12\n"
|
|
"xorq %%rax, %%r13\n"
|
|
"addq $16, %%rcx\n"
|
|
"jnz XW%=\n"
|
|
"QW%=:\n"
|
|
"movq %%r15, %%rcx\n"
|
|
"andq $8, %%r15\n"
|
|
"jz B%=\n"
|
|
"movq %4, %%rax\n"
|
|
"mulq (%%r14)\n"
|
|
"addq $8, %%r14\n"
|
|
"xorq %%rax, %%r12\n"
|
|
"xorq %%rdx, %%r13\n"
|
|
"B%=:\n"
|
|
"andq $7, %%rcx\n"
|
|
"jz F%=\n"
|
|
"movq $1, %%rdx\n"
|
|
"shlq $3, %%rcx\n"
|
|
"movq %3, %%rax\n"
|
|
"shlq %%cl, %%rdx\n"
|
|
"addq $-1, %%rdx\n"
|
|
"andq (%%r14), %%rdx\n"
|
|
"mulq %%rdx\n"
|
|
"xorq %%rdx, %%r12\n"
|
|
"xorq %%rax, %%r13\n"
|
|
"F%=:\n"
|
|
"leaq (%%r13,%4), %%rax\n"
|
|
"xorq %%r12, %%rax\n"
|
|
"mulq %4\n"
|
|
"xorq %%rdx, %%rax\n"
|
|
"xorq %%r12, %%rax\n"
|
|
"xorq %%r13, %%rax\n"
|
|
: "=a"(hash), "=c"(key), "=d"(key)
|
|
: "r"(m), "r"(n), "a"(seed), "c"(len), "d"(key)
|
|
: "%r12", "%r13", "%r14", "%r15", "cc"
|
|
);
|
|
return hash;
|
|
}
|
|
};
|
|
|
|
|
|
struct SimpleHash
|
|
{
|
|
size_t operator() (CompactStringRef x) const
|
|
{
|
|
const char * pos = x.data();
|
|
size_t size = x.size;
|
|
|
|
const char * end = pos + size;
|
|
|
|
size_t res = 0;
|
|
|
|
if (size == 0)
|
|
return 0;
|
|
|
|
if (size < 8)
|
|
{
|
|
memcpy(reinterpret_cast<char *>(&res), pos, size);
|
|
return intHash64(res);
|
|
}
|
|
|
|
while (pos + 8 < end)
|
|
{
|
|
UInt64 word = *reinterpret_cast<const UInt64 *>(pos);
|
|
res = intHash64(word ^ res);
|
|
|
|
pos += 8;
|
|
}
|
|
|
|
UInt64 word = *reinterpret_cast<const UInt64 *>(end - 8);
|
|
res = intHash64(word ^ res);
|
|
|
|
return res;
|
|
}
|
|
};
|
|
|
|
|
|
typedef CompactStringRef Key;
|
|
typedef UInt64 Value;
|
|
|
|
|
|
struct Grower : public HashTableGrower<>
|
|
{
|
|
/// Состояние этой структуры достаточно, чтобы получить размер буфера хэш-таблицы.
|
|
|
|
/// Определяет начальный размер хэш-таблицы.
|
|
static const size_t initial_size_degree = 16;
|
|
Grower() { size_degree = initial_size_degree; }
|
|
|
|
size_t max_fill = (1 << initial_size_degree) * 0.9;
|
|
|
|
/// Размер хэш-таблицы в ячейках.
|
|
size_t bufSize() const { return 1 << size_degree; }
|
|
|
|
size_t maxFill() const { return max_fill /*1 << (size_degree - 1)*/; }
|
|
size_t mask() const { return bufSize() - 1; }
|
|
|
|
/// Из значения хэш-функции получить номер ячейки в хэш-таблице.
|
|
size_t place(size_t x) const { return x & mask(); }
|
|
|
|
/// Следующая ячейка в цепочке разрешения коллизий.
|
|
size_t next(size_t pos) const { ++pos; return pos & mask(); }
|
|
|
|
/// Является ли хэш-таблица достаточно заполненной. Нужно увеличить размер хэш-таблицы, или удалить из неё что-нибудь ненужное.
|
|
bool overflow(size_t elems) const { return elems > maxFill(); }
|
|
|
|
/// Увеличить размер хэш-таблицы.
|
|
void increaseSize()
|
|
{
|
|
size_degree += size_degree >= 23 ? 1 : 2;
|
|
max_fill = (1 << size_degree) * 0.9;
|
|
}
|
|
|
|
/// Установить размер буфера по количеству элементов хэш-таблицы. Используется при десериализации хэш-таблицы.
|
|
void set(size_t num_elems)
|
|
{
|
|
throw Poco::Exception(__PRETTY_FUNCTION__);
|
|
}
|
|
};
|
|
|
|
|
|
int main(int argc, char ** argv)
|
|
{
|
|
size_t n = atoi(argv[1]);
|
|
size_t m = atoi(argv[2]);
|
|
|
|
DB::Arena pool;
|
|
std::vector<Key> data(n);
|
|
|
|
std::cerr << "sizeof(Key) = " << sizeof(Key) << ", sizeof(Value) = " << sizeof(Value) << std::endl;
|
|
|
|
{
|
|
Stopwatch watch;
|
|
DB::ReadBufferFromFileDescriptor in1(STDIN_FILENO);
|
|
DB::CompressedReadBuffer in2(in1);
|
|
|
|
std::string tmp;
|
|
for (size_t i = 0; i < n && !in2.eof(); ++i)
|
|
{
|
|
DB::readStringBinary(tmp, in2);
|
|
data[i] = Key(pool.insert(tmp.data(), tmp.size()), tmp.size());
|
|
}
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "Vector. Size: " << n
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 1)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
//typedef HashMap<Key, Value> Map;
|
|
|
|
/// Сохранение хэша ускоряет ресайзы примерно в 2 раза, и общую производительность - на 6-8%.
|
|
typedef HashMapWithSavedHash<Key, Value, DefaultHash<Key>, Grower> Map;
|
|
|
|
Map map;
|
|
Map::iterator it;
|
|
bool inserted;
|
|
|
|
for (size_t i = 0; i < n; ++i)
|
|
{
|
|
map.emplace(data[i], it, inserted);
|
|
if (inserted)
|
|
it->second = 0;
|
|
++it->second;
|
|
}
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "HashMap (CityHash64). Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
#ifdef DBMS_HASH_MAP_COUNT_COLLISIONS
|
|
<< ", collisions: " << map.getCollisions()
|
|
#endif
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 2)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
typedef HashMapWithSavedHash<Key, Value, FastHash64, Grower> Map;
|
|
|
|
Map map;
|
|
Map::iterator it;
|
|
bool inserted;
|
|
|
|
for (size_t i = 0; i < n; ++i)
|
|
{
|
|
map.emplace(data[i], it, inserted);
|
|
if (inserted)
|
|
it->second = 0;
|
|
++it->second;
|
|
}
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "HashMap (FastHash64). Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
#ifdef DBMS_HASH_MAP_COUNT_COLLISIONS
|
|
<< ", collisions: " << map.getCollisions()
|
|
#endif
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 3)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
typedef HashMapWithSavedHash<Key, Value, CrapWow, Grower> Map;
|
|
|
|
Map map;
|
|
Map::iterator it;
|
|
bool inserted;
|
|
|
|
for (size_t i = 0; i < n; ++i)
|
|
{
|
|
map.emplace(data[i], it, inserted);
|
|
if (inserted)
|
|
it->second = 0;
|
|
++it->second;
|
|
}
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "HashMap (CrapWow). Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
#ifdef DBMS_HASH_MAP_COUNT_COLLISIONS
|
|
<< ", collisions: " << map.getCollisions()
|
|
#endif
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 4)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
typedef HashMapWithSavedHash<Key, Value, SimpleHash, Grower> Map;
|
|
|
|
Map map;
|
|
Map::iterator it;
|
|
bool inserted;
|
|
|
|
for (size_t i = 0; i < n; ++i)
|
|
{
|
|
map.emplace(data[i], it, inserted);
|
|
if (inserted)
|
|
it->second = 0;
|
|
++it->second;
|
|
}
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "HashMap (SimpleHash). Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
#ifdef DBMS_HASH_MAP_COUNT_COLLISIONS
|
|
<< ", collisions: " << map.getCollisions()
|
|
#endif
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 5)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
std::unordered_map<Key, Value, DefaultHash<Key> > map;
|
|
for (size_t i = 0; i < n; ++i)
|
|
++map[data[i]];
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "std::unordered_map. Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 6)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
google::dense_hash_map<Key, Value, DefaultHash<Key> > map;
|
|
map.set_empty_key(Key("\0", 1));
|
|
for (size_t i = 0; i < n; ++i)
|
|
++map[data[i]];
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "google::dense_hash_map. Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
<< std::endl;
|
|
}
|
|
|
|
if (!m || m == 7)
|
|
{
|
|
Stopwatch watch;
|
|
|
|
google::sparse_hash_map<Key, Value, DefaultHash<Key> > map;
|
|
for (size_t i = 0; i < n; ++i)
|
|
++map[data[i]];
|
|
|
|
watch.stop();
|
|
std::cerr << std::fixed << std::setprecision(2)
|
|
<< "google::sparse_hash_map. Size: " << map.size()
|
|
<< ", elapsed: " << watch.elapsedSeconds()
|
|
<< " (" << n / watch.elapsedSeconds() << " elem/sec.)"
|
|
<< std::endl;
|
|
}
|
|
|
|
return 0;
|
|
}
|