ClickHouse/src/IO/HadoopSnappyReadBuffer.h
Nikolay Degterinsky 1be9371fb5 Better
2023-05-01 13:01:23 +00:00

118 lines
3.2 KiB
C++

#pragma once
#include "config.h"
#if USE_SNAPPY
#include <memory>
#include <IO/ReadBuffer.h>
#include <IO/CompressedReadBufferWrapper.h>
namespace DB
{
/*
* Hadoop-snappy format is one of the compression formats base on Snappy used in Hadoop. It uses its own framing format as follows:
* 1. A compressed file consists of one or more blocks.
* 2. A block consists of uncompressed length (big endian 4 byte integer) and one or more subblocks.
* 3. A subblock consists of compressed length (big endian 4 byte integer) and raw compressed data.
*
* HadoopSnappyDecoder implements the decompression of data compressed with hadoop-snappy format.
*/
class HadoopSnappyDecoder
{
public:
enum class Status : int
{
OK = 0,
INVALID_INPUT = 1,
BUFFER_TOO_SMALL = 2,
NEEDS_MORE_INPUT = 3,
TOO_LARGE_COMPRESSED_BLOCK = 4,
};
HadoopSnappyDecoder() = default;
~HadoopSnappyDecoder() = default;
Status readBlock(size_t * avail_in, const char ** next_in, size_t * avail_out, char ** next_out);
inline void reset()
{
buffer_length = 0;
block_length = -1;
compressed_length = -1;
total_uncompressed_length = 0;
}
Status result = Status::OK;
private:
inline bool checkBufferLength(int max) const;
inline static bool checkAvailIn(size_t avail_in, int min);
inline void copyToBuffer(size_t * avail_in, const char ** next_in);
inline static uint32_t readLength(const char * in);
inline Status readLength(size_t * avail_in, const char ** next_in, int * length);
inline Status readBlockLength(size_t * avail_in, const char ** next_in);
inline Status readCompressedLength(size_t * avail_in, const char ** next_in);
inline Status readCompressedData(size_t * avail_in, const char ** next_in, size_t * avail_out, char ** next_out);
char buffer[DBMS_DEFAULT_BUFFER_SIZE] = {0};
int buffer_length = 0;
int block_length = -1;
int compressed_length = -1;
int total_uncompressed_length = 0;
};
/// HadoopSnappyReadBuffer implements read buffer for data compressed with hadoop-snappy format.
class HadoopSnappyReadBuffer : public CompressedReadBufferWrapper
{
public:
using Status = HadoopSnappyDecoder::Status;
inline static String statusToString(Status status)
{
switch (status)
{
case Status::OK:
return "OK";
case Status::INVALID_INPUT:
return "INVALID_INPUT";
case Status::BUFFER_TOO_SMALL:
return "BUFFER_TOO_SMALL";
case Status::NEEDS_MORE_INPUT:
return "NEEDS_MORE_INPUT";
case Status::TOO_LARGE_COMPRESSED_BLOCK:
return "TOO_LARGE_COMPRESSED_BLOCK";
}
UNREACHABLE();
}
explicit HadoopSnappyReadBuffer(
std::unique_ptr<ReadBuffer> in_,
size_t buf_size = DBMS_DEFAULT_BUFFER_SIZE,
char * existing_memory = nullptr,
size_t alignment = 0);
~HadoopSnappyReadBuffer() override;
private:
bool nextImpl() override;
std::unique_ptr<HadoopSnappyDecoder> decoder;
size_t in_available;
const char * in_data;
size_t out_capacity;
char * out_data;
bool eof;
};
}
#endif