Merge remote-tracking branch 'rschu1ze/master' into byteSlice

This commit is contained in:
Robert Schulze 2024-03-01 10:10:30 +00:00
commit 0986d80a1c
No known key found for this signature in database
GPG Key ID: 26703B55FB13728A
42 changed files with 1490 additions and 258 deletions

View File

@ -157,7 +157,7 @@ if (TARGET ch_contrib::zlib)
endif() endif()
if (TARGET ch_contrib::zstd) if (TARGET ch_contrib::zstd)
target_compile_definitions(_libarchive PUBLIC HAVE_ZSTD_H=1 HAVE_LIBZSTD=1) target_compile_definitions(_libarchive PUBLIC HAVE_ZSTD_H=1 HAVE_LIBZSTD=1 HAVE_LIBZSTD_COMPRESSOR=1)
target_link_libraries(_libarchive PRIVATE ch_contrib::zstd) target_link_libraries(_libarchive PRIVATE ch_contrib::zstd)
endif() endif()

View File

@ -19,6 +19,8 @@ CREATE TABLE azure_blob_storage_table (name String, value UInt32)
### Engine parameters ### Engine parameters
- `endpoint` — AzureBlobStorage endpoint URL with container & prefix. Optionally can contain account_name if the authentication method used needs it. (http://azurite1:{port}/[account_name]{container_name}/{data_prefix}) or these parameters can be provided separately using storage_account_url, account_name & container. For specifying prefix, endpoint should be used.
- `endpoint_contains_account_name` - This flag is used to specify if endpoint contains account_name as it is only needed for certain authentication methods. (Default : true)
- `connection_string|storage_account_url` — connection_string includes account name & key ([Create connection string](https://learn.microsoft.com/en-us/azure/storage/common/storage-configure-connection-string?toc=%2Fazure%2Fstorage%2Fblobs%2Ftoc.json&bc=%2Fazure%2Fstorage%2Fblobs%2Fbreadcrumb%2Ftoc.json#configure-a-connection-string-for-an-azure-storage-account)) or you could also provide the storage account url here and account name & account key as separate parameters (see parameters account_name & account_key) - `connection_string|storage_account_url` — connection_string includes account name & key ([Create connection string](https://learn.microsoft.com/en-us/azure/storage/common/storage-configure-connection-string?toc=%2Fazure%2Fstorage%2Fblobs%2Ftoc.json&bc=%2Fazure%2Fstorage%2Fblobs%2Fbreadcrumb%2Ftoc.json#configure-a-connection-string-for-an-azure-storage-account)) or you could also provide the storage account url here and account name & account key as separate parameters (see parameters account_name & account_key)
- `container_name` - Container name - `container_name` - Container name
- `blobpath` - file path. Supports following wildcards in readonly mode: `*`, `**`, `?`, `{abc,def}` and `{N..M}` where `N`, `M` — numbers, `'abc'`, `'def'` — strings. - `blobpath` - file path. Supports following wildcards in readonly mode: `*`, `**`, `?`, `{abc,def}` and `{N..M}` where `N`, `M` — numbers, `'abc'`, `'def'` — strings.

View File

@ -1242,7 +1242,9 @@ Configuration markup:
``` ```
Connection parameters: Connection parameters:
* `storage_account_url` - **Required**, Azure Blob Storage account URL, like `http://account.blob.core.windows.net` or `http://azurite1:10000/devstoreaccount1`. * `endpoint` — AzureBlobStorage endpoint URL with container & prefix. Optionally can contain account_name if the authentication method used needs it. (`http://account.blob.core.windows.net:{port}/[account_name]{container_name}/{data_prefix}`) or these parameters can be provided separately using storage_account_url, account_name & container. For specifying prefix, endpoint should be used.
* `endpoint_contains_account_name` - This flag is used to specify if endpoint contains account_name as it is only needed for certain authentication methods. (Default : true)
* `storage_account_url` - Required if endpoint is not specified, Azure Blob Storage account URL, like `http://account.blob.core.windows.net` or `http://azurite1:10000/devstoreaccount1`.
* `container_name` - Target container name, defaults to `default-container`. * `container_name` - Target container name, defaults to `default-container`.
* `container_already_exists` - If set to `false`, a new container `container_name` is created in the storage account, if set to `true`, disk connects to the container directly, and if left unset, disk connects to the account, checks if the container `container_name` exists, and creates it if it doesn't exist yet. * `container_already_exists` - If set to `false`, a new container `container_name` is created in the storage account, if set to `true`, disk connects to the container directly, and if left unset, disk connects to the account, checks if the container `container_name` exists, and creates it if it doesn't exist yet.

View File

@ -168,6 +168,28 @@ RESTORE TABLE test.table PARTITIONS '2', '3'
FROM Disk('backups', 'filename.zip') FROM Disk('backups', 'filename.zip')
``` ```
### Backups as tar archives
Backups can also be stored as tar archives. The functionality is the same as for zip, except that a password is not supported.
Write a backup as a tar:
```
BACKUP TABLE test.table TO Disk('backups', '1.tar')
```
Corresponding restore:
```
RESTORE TABLE test.table FROM Disk('backups', '1.tar')
```
To change the compression method, the correct file suffix should be appended to the backup name. I.E to compress the tar archive using gzip:
```
BACKUP TABLE test.table TO Disk('backups', '1.tar.gz')
```
The supported compression file suffixes are `tar.gz`, `.tgz` `tar.bz2`, `tar.lzma`, `.tar.zst`, `.tzst` and `.tar.xz`.
### Check the status of backups ### Check the status of backups
The backup command returns an `id` and `status`, and that `id` can be used to get the status of the backup. This is very useful to check the progress of long ASYNC backups. The example below shows a failure that happened when trying to overwrite an existing backup file: The backup command returns an `id` and `status`, and that `id` can be used to get the status of the backup. This is very useful to check the progress of long ASYNC backups. The example below shows a failure that happened when trying to overwrite an existing backup file:

View File

@ -927,7 +927,7 @@ void BackupImpl::writeFile(const BackupFileInfo & info, BackupEntryPtr entry)
const auto write_info_to_archive = [&](const auto & file_name) const auto write_info_to_archive = [&](const auto & file_name)
{ {
auto out = archive_writer->writeFile(file_name); auto out = archive_writer->writeFile(file_name, info.size);
auto read_buffer = entry->getReadBuffer(writer->getReadSettings()); auto read_buffer = entry->getReadBuffer(writer->getReadSettings());
if (info.base_size != 0) if (info.base_size != 0)
read_buffer->seek(info.base_size, SEEK_SET); read_buffer->seek(info.base_size, SEEK_SET);

View File

@ -20,13 +20,6 @@ namespace ErrorCodes
extern const int BAD_ARGUMENTS; extern const int BAD_ARGUMENTS;
} }
struct AzureBlobStorageEndpoint
{
const String storage_account_url;
const String container_name;
const std::optional<bool> container_already_exists;
};
void validateStorageAccountUrl(const String & storage_account_url) void validateStorageAccountUrl(const String & storage_account_url)
{ {
@ -58,28 +51,89 @@ void validateContainerName(const String & container_name)
AzureBlobStorageEndpoint processAzureBlobStorageEndpoint(const Poco::Util::AbstractConfiguration & config, const String & config_prefix) AzureBlobStorageEndpoint processAzureBlobStorageEndpoint(const Poco::Util::AbstractConfiguration & config, const String & config_prefix)
{ {
std::string storage_url; String storage_url;
if (config.has(config_prefix + ".storage_account_url")) String account_name;
String container_name;
String prefix;
if (config.has(config_prefix + ".endpoint"))
{
String endpoint = config.getString(config_prefix + ".endpoint");
/// For some authentication methods account name is not present in the endpoint
/// 'endpoint_contains_account_name' bool is used to understand how to split the endpoint (default : true)
bool endpoint_contains_account_name = config.getBool(config_prefix + ".endpoint_contains_account_name", true);
size_t pos = endpoint.find("//");
if (pos == std::string::npos)
throw Exception(ErrorCodes::BAD_ARGUMENTS, "Expected '//' in endpoint");
if (endpoint_contains_account_name)
{
size_t acc_pos_begin = endpoint.find('/', pos+2);
if (acc_pos_begin == std::string::npos)
throw Exception(ErrorCodes::BAD_ARGUMENTS, "Expected account_name in endpoint");
storage_url = endpoint.substr(0,acc_pos_begin);
size_t acc_pos_end = endpoint.find('/',acc_pos_begin+1);
if (acc_pos_end == std::string::npos)
throw Exception(ErrorCodes::BAD_ARGUMENTS, "Expected container_name in endpoint");
account_name = endpoint.substr(acc_pos_begin+1,(acc_pos_end-acc_pos_begin)-1);
size_t cont_pos_end = endpoint.find('/', acc_pos_end+1);
if (cont_pos_end != std::string::npos)
{
container_name = endpoint.substr(acc_pos_end+1,(cont_pos_end-acc_pos_end)-1);
prefix = endpoint.substr(cont_pos_end+1);
}
else
{
container_name = endpoint.substr(acc_pos_end+1);
}
}
else
{
size_t cont_pos_begin = endpoint.find('/', pos+2);
if (cont_pos_begin == std::string::npos)
throw Exception(ErrorCodes::BAD_ARGUMENTS, "Expected container_name in endpoint");
storage_url = endpoint.substr(0,cont_pos_begin);
size_t cont_pos_end = endpoint.find('/',cont_pos_begin+1);
if (cont_pos_end != std::string::npos)
{
container_name = endpoint.substr(cont_pos_begin+1,(cont_pos_end-cont_pos_begin)-1);
prefix = endpoint.substr(cont_pos_end+1);
}
else
{
container_name = endpoint.substr(cont_pos_begin+1);
}
}
}
else if (config.has(config_prefix + ".connection_string"))
{
storage_url = config.getString(config_prefix + ".connection_string");
container_name = config.getString(config_prefix + ".container_name");
}
else if (config.has(config_prefix + ".storage_account_url"))
{ {
storage_url = config.getString(config_prefix + ".storage_account_url"); storage_url = config.getString(config_prefix + ".storage_account_url");
validateStorageAccountUrl(storage_url); validateStorageAccountUrl(storage_url);
container_name = config.getString(config_prefix + ".container_name");
} }
else else
{ throw Exception(ErrorCodes::BAD_ARGUMENTS, "Expected either `storage_account_url` or `connection_string` or `endpoint` in config");
if (config.has(config_prefix + ".connection_string"))
storage_url = config.getString(config_prefix + ".connection_string");
else if (config.has(config_prefix + ".endpoint"))
storage_url = config.getString(config_prefix + ".endpoint");
else
throw Exception(ErrorCodes::BAD_ARGUMENTS, "Expected either `connection_string` or `endpoint` in config");
}
String container_name = config.getString(config_prefix + ".container_name", "default-container"); if (!container_name.empty())
validateContainerName(container_name); validateContainerName(container_name);
std::optional<bool> container_already_exists {}; std::optional<bool> container_already_exists {};
if (config.has(config_prefix + ".container_already_exists")) if (config.has(config_prefix + ".container_already_exists"))
container_already_exists = {config.getBool(config_prefix + ".container_already_exists")}; container_already_exists = {config.getBool(config_prefix + ".container_already_exists")};
return {storage_url, container_name, container_already_exists}; return {storage_url, account_name, container_name, prefix, container_already_exists};
} }
@ -133,15 +187,13 @@ std::unique_ptr<BlobContainerClient> getAzureBlobContainerClient(
{ {
auto endpoint = processAzureBlobStorageEndpoint(config, config_prefix); auto endpoint = processAzureBlobStorageEndpoint(config, config_prefix);
auto container_name = endpoint.container_name; auto container_name = endpoint.container_name;
auto final_url = container_name.empty() auto final_url = endpoint.getEndpoint();
? endpoint.storage_account_url
: (std::filesystem::path(endpoint.storage_account_url) / container_name).string();
if (endpoint.container_already_exists.value_or(false)) if (endpoint.container_already_exists.value_or(false))
return getAzureBlobStorageClientWithAuth<BlobContainerClient>(final_url, container_name, config, config_prefix); return getAzureBlobStorageClientWithAuth<BlobContainerClient>(final_url, container_name, config, config_prefix);
auto blob_service_client = getAzureBlobStorageClientWithAuth<BlobServiceClient>( auto blob_service_client = getAzureBlobStorageClientWithAuth<BlobServiceClient>(
endpoint.storage_account_url, container_name, config, config_prefix); endpoint.getEndpointWithoutContainer(), container_name, config, config_prefix);
try try
{ {

View File

@ -10,9 +10,46 @@
namespace DB namespace DB
{ {
struct AzureBlobStorageEndpoint
{
const String storage_account_url;
const String account_name;
const String container_name;
const String prefix;
const std::optional<bool> container_already_exists;
String getEndpoint()
{
String url = storage_account_url;
if (!account_name.empty())
url += "/" + account_name;
if (!container_name.empty())
url += "/" + container_name;
if (!prefix.empty())
url += "/" + prefix;
return url;
}
String getEndpointWithoutContainer()
{
String url = storage_account_url;
if (!account_name.empty())
url += "/" + account_name;
return url;
}
};
std::unique_ptr<Azure::Storage::Blobs::BlobContainerClient> getAzureBlobContainerClient( std::unique_ptr<Azure::Storage::Blobs::BlobContainerClient> getAzureBlobContainerClient(
const Poco::Util::AbstractConfiguration & config, const String & config_prefix); const Poco::Util::AbstractConfiguration & config, const String & config_prefix);
AzureBlobStorageEndpoint processAzureBlobStorageEndpoint(const Poco::Util::AbstractConfiguration & config, const String & config_prefix);
std::unique_ptr<AzureObjectStorageSettings> getAzureBlobStorageSettings(const Poco::Util::AbstractConfiguration & config, const String & config_prefix, ContextPtr /*context*/); std::unique_ptr<AzureObjectStorageSettings> getAzureBlobStorageSettings(const Poco::Util::AbstractConfiguration & config, const String & config_prefix, ContextPtr /*context*/);
} }

View File

@ -93,11 +93,11 @@ AzureObjectStorage::AzureObjectStorage(
const String & name_, const String & name_,
AzureClientPtr && client_, AzureClientPtr && client_,
SettingsPtr && settings_, SettingsPtr && settings_,
const String & container_) const String & object_namespace_)
: name(name_) : name(name_)
, client(std::move(client_)) , client(std::move(client_))
, settings(std::move(settings_)) , settings(std::move(settings_))
, container(container_) , object_namespace(object_namespace_)
, log(getLogger("AzureObjectStorage")) , log(getLogger("AzureObjectStorage"))
{ {
} }
@ -379,7 +379,7 @@ std::unique_ptr<IObjectStorage> AzureObjectStorage::cloneObjectStorage(const std
name, name,
getAzureBlobContainerClient(config, config_prefix), getAzureBlobContainerClient(config, config_prefix),
getAzureBlobStorageSettings(config, config_prefix, context), getAzureBlobStorageSettings(config, config_prefix, context),
container object_namespace
); );
} }

View File

@ -67,7 +67,7 @@ public:
const String & name_, const String & name_,
AzureClientPtr && client_, AzureClientPtr && client_,
SettingsPtr && settings_, SettingsPtr && settings_,
const String & container_); const String & object_namespace_);
void listObjects(const std::string & path, RelativePathsWithMetadata & children, int max_keys) const override; void listObjects(const std::string & path, RelativePathsWithMetadata & children, int max_keys) const override;
@ -130,7 +130,7 @@ public:
const std::string & config_prefix, const std::string & config_prefix,
ContextPtr context) override; ContextPtr context) override;
String getObjectsNamespace() const override { return container ; } String getObjectsNamespace() const override { return object_namespace ; }
std::unique_ptr<IObjectStorage> cloneObjectStorage( std::unique_ptr<IObjectStorage> cloneObjectStorage(
const std::string & new_namespace, const std::string & new_namespace,
@ -154,7 +154,7 @@ private:
/// client used to access the files in the Blob Storage cloud /// client used to access the files in the Blob Storage cloud
MultiVersion<Azure::Storage::Blobs::BlobContainerClient> client; MultiVersion<Azure::Storage::Blobs::BlobContainerClient> client;
MultiVersion<AzureObjectStorageSettings> settings; MultiVersion<AzureObjectStorageSettings> settings;
const String container; const String object_namespace; /// container + prefix
LoggerPtr log; LoggerPtr log;
}; };

View File

@ -213,12 +213,12 @@ void registerAzureObjectStorage(ObjectStorageFactory & factory)
const ContextPtr & context, const ContextPtr & context,
bool /* skip_access_check */) -> ObjectStoragePtr bool /* skip_access_check */) -> ObjectStoragePtr
{ {
String container_name = config.getString(config_prefix + ".container_name", "default-container"); AzureBlobStorageEndpoint endpoint = processAzureBlobStorageEndpoint(config, config_prefix);
return std::make_unique<AzureObjectStorage>( return std::make_unique<AzureObjectStorage>(
name, name,
getAzureBlobContainerClient(config, config_prefix), getAzureBlobContainerClient(config, config_prefix),
getAzureBlobStorageSettings(config, config_prefix, context), getAzureBlobStorageSettings(config, config_prefix, context),
container_name); endpoint.prefix.empty() ? endpoint.container_name : endpoint.container_name + "/" + endpoint.prefix);
}); });
} }

View File

@ -22,6 +22,8 @@ public:
/// of the function `writeFile()` should be destroyed before next call of `writeFile()`. /// of the function `writeFile()` should be destroyed before next call of `writeFile()`.
virtual std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename) = 0; virtual std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename) = 0;
virtual std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename, size_t size) = 0;
/// Returns true if there is an active instance of WriteBuffer returned by writeFile(). /// Returns true if there is an active instance of WriteBuffer returned by writeFile().
/// This function should be used mostly for debugging purposes. /// This function should be used mostly for debugging purposes.
virtual bool isWritingFile() const = 0; virtual bool isWritingFile() const = 0;

View File

@ -1,11 +1,9 @@
#include <IO/Archives/ArchiveUtils.h>
#include <IO/Archives/LibArchiveReader.h> #include <IO/Archives/LibArchiveReader.h>
#include <IO/ReadBufferFromFileBase.h> #include <IO/ReadBufferFromFileBase.h>
#include <Common/quoteString.h> #include <Common/quoteString.h>
#include <Common/scope_guard_safe.h> #include <Common/scope_guard_safe.h>
#include <IO/Archives/ArchiveUtils.h>
#include <mutex>
namespace DB namespace DB
{ {
@ -20,29 +18,52 @@ namespace ErrorCodes
extern const int UNSUPPORTED_METHOD; extern const int UNSUPPORTED_METHOD;
} }
class LibArchiveReader::StreamInfo
{
public:
explicit StreamInfo(std::unique_ptr<SeekableReadBuffer> read_buffer_) : read_buffer(std::move(read_buffer_)) { }
static ssize_t read(struct archive *, void * client_data, const void ** buff)
{
auto * read_stream = reinterpret_cast<StreamInfo *>(client_data);
*buff = reinterpret_cast<void *>(read_stream->buf);
return read_stream->read_buffer->read(read_stream->buf, DBMS_DEFAULT_BUFFER_SIZE);
}
std::unique_ptr<SeekableReadBuffer> read_buffer;
char buf[DBMS_DEFAULT_BUFFER_SIZE];
};
class LibArchiveReader::Handle class LibArchiveReader::Handle
{ {
public: public:
explicit Handle(std::string path_to_archive_, bool lock_on_reading_) explicit Handle(std::string path_to_archive_, bool lock_on_reading_)
: path_to_archive(path_to_archive_), lock_on_reading(lock_on_reading_) : path_to_archive(std::move(path_to_archive_)), lock_on_reading(lock_on_reading_)
{ {
current_archive = open(path_to_archive); current_archive = openWithPath(path_to_archive);
}
explicit Handle(std::string path_to_archive_, bool lock_on_reading_, const ReadArchiveFunction & archive_read_function_)
: path_to_archive(std::move(path_to_archive_)), archive_read_function(archive_read_function_), lock_on_reading(lock_on_reading_)
{
read_stream = std::make_unique<StreamInfo>(archive_read_function());
current_archive = openWithReader(read_stream.get());
} }
Handle(const Handle &) = delete; Handle(const Handle &) = delete;
Handle(Handle && other) noexcept Handle(Handle && other) noexcept
: current_archive(other.current_archive) : read_stream(std::move(other.read_stream))
, current_archive(other.current_archive)
, current_entry(other.current_entry) , current_entry(other.current_entry)
, archive_read_function(std::move(other.archive_read_function))
, lock_on_reading(other.lock_on_reading) , lock_on_reading(other.lock_on_reading)
{ {
other.current_archive = nullptr; other.current_archive = nullptr;
other.current_entry = nullptr; other.current_entry = nullptr;
} }
~Handle() ~Handle() { close(current_archive); }
{
close(current_archive);
}
bool locateFile(const std::string & filename) bool locateFile(const std::string & filename)
{ {
@ -64,10 +85,14 @@ public:
break; break;
if (filter(archive_entry_pathname(current_entry))) if (filter(archive_entry_pathname(current_entry)))
{
valid = true;
return true; return true;
} }
}
checkError(err); checkError(err);
valid = false;
return false; return false;
} }
@ -81,17 +106,19 @@ public:
} while (err == ARCHIVE_RETRY); } while (err == ARCHIVE_RETRY);
checkError(err); checkError(err);
return err == ARCHIVE_OK; valid = err == ARCHIVE_OK;
return valid;
} }
std::vector<std::string> getAllFiles(NameFilter filter) std::vector<std::string> getAllFiles(NameFilter filter)
{ {
auto * archive = open(path_to_archive); std::unique_ptr<LibArchiveReader::StreamInfo> rs
SCOPE_EXIT( = archive_read_function ? std::make_unique<StreamInfo>(archive_read_function()) : nullptr;
close(archive); auto * archive = rs ? openWithReader(rs.get()) : openWithPath(path_to_archive);
);
struct archive_entry * entry = nullptr; SCOPE_EXIT(close(archive););
Entry entry = nullptr;
std::vector<std::string> files; std::vector<std::string> files;
int error = readNextHeader(archive, &entry); int error = readNextHeader(archive, &entry);
@ -112,6 +139,8 @@ public:
const String & getFileName() const const String & getFileName() const
{ {
chassert(current_entry); chassert(current_entry);
if (!valid)
throw Exception(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "No current file");
if (!file_name) if (!file_name)
file_name.emplace(archive_entry_pathname(current_entry)); file_name.emplace(archive_entry_pathname(current_entry));
@ -121,6 +150,8 @@ public:
const FileInfo & getFileInfo() const const FileInfo & getFileInfo() const
{ {
chassert(current_entry); chassert(current_entry);
if (!valid)
throw Exception(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "No current file");
if (!file_info) if (!file_info)
{ {
file_info.emplace(); file_info.emplace();
@ -132,13 +163,21 @@ public:
return *file_info; return *file_info;
} }
struct archive * current_archive; la_ssize_t readData(void * buf, size_t len) { return archive_read_data(current_archive, buf, len); }
struct archive_entry * current_entry = nullptr;
const char * getArchiveError() { return archive_error_string(current_archive); }
private: private:
using Archive = struct archive *;
using Entry = struct archive_entry *;
void checkError(int error) const void checkError(int error) const
{ {
if (error == ARCHIVE_FATAL) if (error == ARCHIVE_FATAL)
throw Exception(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Failed to read archive while fetching all files: {}", archive_error_string(current_archive)); throw Exception(
ErrorCodes::CANNOT_UNPACK_ARCHIVE,
"Failed to read archive while fetching all files: {}",
archive_error_string(current_archive));
} }
void resetFileInfo() void resetFileInfo()
@ -147,7 +186,7 @@ private:
file_info.reset(); file_info.reset();
} }
static struct archive * open(const String & path_to_archive) Archive openWithReader(StreamInfo * read_stream_)
{ {
auto * archive = archive_read_new(); auto * archive = archive_read_new();
try try
@ -158,13 +197,18 @@ private:
archive_read_support_filter_xz(archive); archive_read_support_filter_xz(archive);
archive_read_support_filter_lz4(archive); archive_read_support_filter_lz4(archive);
archive_read_support_filter_zstd(archive); archive_read_support_filter_zstd(archive);
archive_read_support_filter_lzma(archive);
// Support tar, 7zip and zip // Support tar, 7zip and zip
archive_read_support_format_tar(archive); archive_read_support_format_tar(archive);
archive_read_support_format_7zip(archive); archive_read_support_format_7zip(archive);
archive_read_support_format_zip(archive); archive_read_support_format_zip(archive);
if (archive_read_open_filename(archive, path_to_archive.c_str(), 10240) != ARCHIVE_OK) if (archive_read_open(archive, read_stream_, nullptr, StreamInfo::read, nullptr) != ARCHIVE_OK)
throw Exception(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Couldn't open archive {}: {}", quoteString(path_to_archive), archive_error_string(archive)); throw Exception(
ErrorCodes::CANNOT_UNPACK_ARCHIVE,
"Couldn't open archive {}: {}",
quoteString(path_to_archive),
archive_error_string(archive));
} }
catch (...) catch (...)
{ {
@ -175,7 +219,39 @@ private:
return archive; return archive;
} }
static void close(struct archive * archive) Archive openWithPath(const String & path_to_archive_)
{
auto * archive = archive_read_new();
try
{
// Support for bzip2, gzip, lzip, xz, zstd and lz4
archive_read_support_filter_bzip2(archive);
archive_read_support_filter_gzip(archive);
archive_read_support_filter_xz(archive);
archive_read_support_filter_lz4(archive);
archive_read_support_filter_zstd(archive);
archive_read_support_filter_lzma(archive);
// Support tar, 7zip and zip
archive_read_support_format_tar(archive);
archive_read_support_format_7zip(archive);
archive_read_support_format_zip(archive);
if (archive_read_open_filename(archive, path_to_archive_.c_str(), 10240) != ARCHIVE_OK)
throw Exception(
ErrorCodes::CANNOT_UNPACK_ARCHIVE,
"Couldn't open archive {}: {}",
quoteString(path_to_archive),
archive_error_string(archive));
}
catch (...)
{
close(archive);
throw;
}
return archive;
}
static void close(Archive archive)
{ {
if (archive) if (archive)
{ {
@ -193,7 +269,12 @@ private:
return archive_read_next_header(archive, entry); return archive_read_next_header(archive, entry);
} }
const String path_to_archive; String path_to_archive;
std::unique_ptr<StreamInfo> read_stream;
Archive current_archive;
Entry current_entry = nullptr;
bool valid = true;
IArchiveReader::ReadArchiveFunction archive_read_function;
/// for some archive types when we are reading headers static variables are used /// for some archive types when we are reading headers static variables are used
/// which are not thread-safe /// which are not thread-safe
@ -215,6 +296,7 @@ public:
/// Releases owned handle to pass it to a read buffer. /// Releases owned handle to pass it to a read buffer.
Handle releaseHandle() && { return std::move(handle); } Handle releaseHandle() && { return std::move(handle); }
private: private:
Handle handle; Handle handle;
}; };
@ -226,36 +308,33 @@ public:
: ReadBufferFromFileBase(DBMS_DEFAULT_BUFFER_SIZE, nullptr, 0) : ReadBufferFromFileBase(DBMS_DEFAULT_BUFFER_SIZE, nullptr, 0)
, handle(std::move(handle_)) , handle(std::move(handle_))
, path_to_archive(std::move(path_to_archive_)) , path_to_archive(std::move(path_to_archive_))
{} {
}
off_t seek(off_t /* off */, int /* whence */) override off_t seek(off_t /* off */, int /* whence */) override
{ {
throw Exception(ErrorCodes::UNSUPPORTED_METHOD, "Seek is not supported when reading from archive"); throw Exception(ErrorCodes::UNSUPPORTED_METHOD, "Seek is not supported when reading from archive");
} }
bool checkIfActuallySeekable() override { return false; } bool checkIfActuallySeekable() override { return false; }
off_t getPosition() override off_t getPosition() override { throw Exception(ErrorCodes::UNSUPPORTED_METHOD, "getPosition not supported when reading from archive"); }
{
throw Exception(ErrorCodes::UNSUPPORTED_METHOD, "getPosition not supported when reading from archive");
}
String getFileName() const override { return handle.getFileName(); } String getFileName() const override { return handle.getFileName(); }
size_t getFileSize() override { return handle.getFileInfo().uncompressed_size; } size_t getFileSize() override { return handle.getFileInfo().uncompressed_size; }
Handle releaseHandle() && Handle releaseHandle() && { return std::move(handle); }
{
return std::move(handle);
}
private: private:
bool nextImpl() override bool nextImpl() override
{ {
auto bytes_read = archive_read_data(handle.current_archive, internal_buffer.begin(), static_cast<int>(internal_buffer.size())); auto bytes_read = handle.readData(internal_buffer.begin(), internal_buffer.size());
if (bytes_read < 0) if (bytes_read < 0)
throw Exception(ErrorCodes::CANNOT_READ_ALL_DATA, "Failed to read file {} from {}: {}", handle.getFileName(), path_to_archive, archive_error_string(handle.current_archive)); throw Exception(
ErrorCodes::CANNOT_READ_ALL_DATA,
"Failed to read file {} from {}: {}",
handle.getFileName(),
path_to_archive,
handle.getArchiveError());
if (!bytes_read) if (!bytes_read)
return false; return false;
@ -274,7 +353,17 @@ private:
LibArchiveReader::LibArchiveReader(std::string archive_name_, bool lock_on_reading_, std::string path_to_archive_) LibArchiveReader::LibArchiveReader(std::string archive_name_, bool lock_on_reading_, std::string path_to_archive_)
: archive_name(std::move(archive_name_)), lock_on_reading(lock_on_reading_), path_to_archive(std::move(path_to_archive_)) : archive_name(std::move(archive_name_)), lock_on_reading(lock_on_reading_), path_to_archive(std::move(path_to_archive_))
{} {
}
LibArchiveReader::LibArchiveReader(
std::string archive_name_, bool lock_on_reading_, std::string path_to_archive_, const ReadArchiveFunction & archive_read_function_)
: archive_name(std::move(archive_name_))
, lock_on_reading(lock_on_reading_)
, path_to_archive(std::move(path_to_archive_))
, archive_read_function(archive_read_function_)
{
}
LibArchiveReader::~LibArchiveReader() = default; LibArchiveReader::~LibArchiveReader() = default;
@ -285,21 +374,25 @@ const std::string & LibArchiveReader::getPath() const
bool LibArchiveReader::fileExists(const String & filename) bool LibArchiveReader::fileExists(const String & filename)
{ {
Handle handle(path_to_archive, lock_on_reading); Handle handle = acquireHandle();
return handle.locateFile(filename); return handle.locateFile(filename);
} }
LibArchiveReader::FileInfo LibArchiveReader::getFileInfo(const String & filename) LibArchiveReader::FileInfo LibArchiveReader::getFileInfo(const String & filename)
{ {
Handle handle(path_to_archive, lock_on_reading); Handle handle = acquireHandle();
if (!handle.locateFile(filename)) if (!handle.locateFile(filename))
throw Exception(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Couldn't unpack archive {}: file not found", path_to_archive); throw Exception(
ErrorCodes::CANNOT_UNPACK_ARCHIVE,
"Couldn't unpack archive {}: File {} was not found in archive",
path_to_archive,
quoteString(filename));
return handle.getFileInfo(); return handle.getFileInfo();
} }
std::unique_ptr<LibArchiveReader::FileEnumerator> LibArchiveReader::firstFile() std::unique_ptr<LibArchiveReader::FileEnumerator> LibArchiveReader::firstFile()
{ {
Handle handle(path_to_archive, lock_on_reading); Handle handle = acquireHandle();
if (!handle.nextFile()) if (!handle.nextFile())
return nullptr; return nullptr;
@ -308,17 +401,28 @@ std::unique_ptr<LibArchiveReader::FileEnumerator> LibArchiveReader::firstFile()
std::unique_ptr<ReadBufferFromFileBase> LibArchiveReader::readFile(const String & filename, bool throw_on_not_found) std::unique_ptr<ReadBufferFromFileBase> LibArchiveReader::readFile(const String & filename, bool throw_on_not_found)
{ {
return readFile([&](const std::string & file) { return file == filename; }, throw_on_not_found); Handle handle = acquireHandle();
if (!handle.locateFile(filename))
{
if (throw_on_not_found)
throw Exception(
ErrorCodes::CANNOT_UNPACK_ARCHIVE,
"Couldn't unpack archive {}: File {} was not found in archive",
path_to_archive,
quoteString(filename));
return nullptr;
}
return std::make_unique<ReadBufferFromLibArchive>(std::move(handle), path_to_archive);
} }
std::unique_ptr<ReadBufferFromFileBase> LibArchiveReader::readFile(NameFilter filter, bool throw_on_not_found) std::unique_ptr<ReadBufferFromFileBase> LibArchiveReader::readFile(NameFilter filter, bool throw_on_not_found)
{ {
Handle handle(path_to_archive, lock_on_reading); Handle handle = acquireHandle();
if (!handle.locateFile(filter)) if (!handle.locateFile(filter))
{ {
if (throw_on_not_found) if (throw_on_not_found)
throw Exception( throw Exception(
ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Couldn't unpack archive {}: no file found satisfying the filter", path_to_archive); ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Couldn't unpack archive {}: No file satisfying filter in archive", path_to_archive);
return nullptr; return nullptr;
} }
return std::make_unique<ReadBufferFromLibArchive>(std::move(handle), path_to_archive); return std::make_unique<ReadBufferFromLibArchive>(std::move(handle), path_to_archive);
@ -337,7 +441,8 @@ std::unique_ptr<LibArchiveReader::FileEnumerator> LibArchiveReader::nextFile(std
{ {
if (!dynamic_cast<ReadBufferFromLibArchive *>(read_buffer.get())) if (!dynamic_cast<ReadBufferFromLibArchive *>(read_buffer.get()))
throw Exception(ErrorCodes::LOGICAL_ERROR, "Wrong ReadBuffer passed to nextFile()"); throw Exception(ErrorCodes::LOGICAL_ERROR, "Wrong ReadBuffer passed to nextFile()");
auto read_buffer_from_libarchive = std::unique_ptr<ReadBufferFromLibArchive>(static_cast<ReadBufferFromLibArchive *>(read_buffer.release())); auto read_buffer_from_libarchive
= std::unique_ptr<ReadBufferFromLibArchive>(static_cast<ReadBufferFromLibArchive *>(read_buffer.release()));
auto handle = std::move(*read_buffer_from_libarchive).releaseHandle(); auto handle = std::move(*read_buffer_from_libarchive).releaseHandle();
if (!handle.nextFile()) if (!handle.nextFile())
return nullptr; return nullptr;
@ -348,7 +453,8 @@ std::unique_ptr<LibArchiveReader::FileEnumerator> LibArchiveReader::currentFile(
{ {
if (!dynamic_cast<ReadBufferFromLibArchive *>(read_buffer.get())) if (!dynamic_cast<ReadBufferFromLibArchive *>(read_buffer.get()))
throw Exception(ErrorCodes::LOGICAL_ERROR, "Wrong ReadBuffer passed to nextFile()"); throw Exception(ErrorCodes::LOGICAL_ERROR, "Wrong ReadBuffer passed to nextFile()");
auto read_buffer_from_libarchive = std::unique_ptr<ReadBufferFromLibArchive>(static_cast<ReadBufferFromLibArchive *>(read_buffer.release())); auto read_buffer_from_libarchive
= std::unique_ptr<ReadBufferFromLibArchive>(static_cast<ReadBufferFromLibArchive *>(read_buffer.release()));
auto handle = std::move(*read_buffer_from_libarchive).releaseHandle(); auto handle = std::move(*read_buffer_from_libarchive).releaseHandle();
return std::make_unique<FileEnumeratorImpl>(std::move(handle)); return std::make_unique<FileEnumeratorImpl>(std::move(handle));
} }
@ -360,15 +466,24 @@ std::vector<std::string> LibArchiveReader::getAllFiles()
std::vector<std::string> LibArchiveReader::getAllFiles(NameFilter filter) std::vector<std::string> LibArchiveReader::getAllFiles(NameFilter filter)
{ {
Handle handle(path_to_archive, lock_on_reading); Handle handle = acquireHandle();
return handle.getAllFiles(filter); return handle.getAllFiles(filter);
} }
void LibArchiveReader::setPassword(const String & /*password_*/) void LibArchiveReader::setPassword(const String & password_)
{ {
if (password_.empty())
return;
throw Exception(ErrorCodes::LOGICAL_ERROR, "Cannot set password to {} archive", archive_name); throw Exception(ErrorCodes::LOGICAL_ERROR, "Cannot set password to {} archive", archive_name);
} }
LibArchiveReader::Handle LibArchiveReader::acquireHandle()
{
std::lock_guard lock{mutex};
return archive_read_function ? Handle{path_to_archive, lock_on_reading, archive_read_function}
: Handle{path_to_archive, lock_on_reading};
}
#endif #endif
} }

View File

@ -1,8 +1,9 @@
#pragma once #pragma once
#include "config.h" #include <mutex>
#include <IO/Archives/IArchiveReader.h> #include <IO/Archives/IArchiveReader.h>
#include <IO/Archives/LibArchiveReader.h>
#include "config.h"
namespace DB namespace DB
@ -52,26 +53,44 @@ protected:
/// Constructs an archive's reader that will read from a file in the local filesystem. /// Constructs an archive's reader that will read from a file in the local filesystem.
LibArchiveReader(std::string archive_name_, bool lock_on_reading_, std::string path_to_archive_); LibArchiveReader(std::string archive_name_, bool lock_on_reading_, std::string path_to_archive_);
LibArchiveReader(
std::string archive_name_, bool lock_on_reading_, std::string path_to_archive_, const ReadArchiveFunction & archive_read_function_);
private: private:
class ReadBufferFromLibArchive; class ReadBufferFromLibArchive;
class Handle; class Handle;
class FileEnumeratorImpl; class FileEnumeratorImpl;
class StreamInfo;
Handle acquireHandle();
const std::string archive_name; const std::string archive_name;
const bool lock_on_reading; const bool lock_on_reading;
const String path_to_archive; const String path_to_archive;
const ReadArchiveFunction archive_read_function;
mutable std::mutex mutex;
}; };
class TarArchiveReader : public LibArchiveReader class TarArchiveReader : public LibArchiveReader
{ {
public: public:
explicit TarArchiveReader(std::string path_to_archive) : LibArchiveReader("tar", /*lock_on_reading_=*/ true, std::move(path_to_archive)) { } explicit TarArchiveReader(std::string path_to_archive) : LibArchiveReader("tar", /*lock_on_reading_=*/true, std::move(path_to_archive))
{
}
explicit TarArchiveReader(std::string path_to_archive, const ReadArchiveFunction & archive_read_function)
: LibArchiveReader("tar", /*lock_on_reading_=*/true, std::move(path_to_archive), archive_read_function)
{
}
}; };
class SevenZipArchiveReader : public LibArchiveReader class SevenZipArchiveReader : public LibArchiveReader
{ {
public: public:
explicit SevenZipArchiveReader(std::string path_to_archive) : LibArchiveReader("7z", /*lock_on_reading_=*/ false, std::move(path_to_archive)) { } explicit SevenZipArchiveReader(std::string path_to_archive)
: LibArchiveReader("7z", /*lock_on_reading_=*/false, std::move(path_to_archive))
{
}
}; };
#endif #endif

View File

@ -0,0 +1,248 @@
#include <IO/Archives/LibArchiveWriter.h>
#include <filesystem>
#include <IO/WriteBufferFromFileBase.h>
#include <Common/quoteString.h>
#include <Common/scope_guard_safe.h>
#include <mutex>
#if USE_LIBARCHIVE
// this implemation follows the ZipArchiveWriter implemation as closely as possible.
namespace DB
{
namespace ErrorCodes
{
extern const int CANNOT_PACK_ARCHIVE;
extern const int NOT_IMPLEMENTED;
}
namespace
{
void checkResultCodeImpl(int code, const String & filename)
{
if (code == ARCHIVE_OK)
return;
throw Exception(
ErrorCodes::CANNOT_PACK_ARCHIVE, "Couldn't pack archive: LibArchive Code = {}, filename={}", code, quoteString(filename));
}
}
// this is a thin wrapper for libarchive to be able to write the archive to a WriteBuffer
class LibArchiveWriter::StreamInfo
{
public:
explicit StreamInfo(std::unique_ptr<WriteBuffer> archive_write_buffer_) : archive_write_buffer(std::move(archive_write_buffer_)) { }
static ssize_t memory_write(struct archive *, void * client_data, const void * buff, size_t length)
{
auto * stream_info = reinterpret_cast<StreamInfo *>(client_data);
stream_info->archive_write_buffer->write(reinterpret_cast<const char *>(buff), length);
return length;
}
std::unique_ptr<WriteBuffer> archive_write_buffer;
};
class LibArchiveWriter::WriteBufferFromLibArchive : public WriteBufferFromFileBase
{
public:
WriteBufferFromLibArchive(std::shared_ptr<LibArchiveWriter> archive_writer_, const String & filename_, const size_t & size_)
: WriteBufferFromFileBase(DBMS_DEFAULT_BUFFER_SIZE, nullptr, 0), archive_writer(archive_writer_), filename(filename_), size(size_)
{
startWritingFile();
archive = archive_writer_->getArchive();
entry = nullptr;
}
~WriteBufferFromLibArchive() override
{
try
{
closeFile(/* throw_if_error= */ false);
endWritingFile();
}
catch (...)
{
tryLogCurrentException("WriteBufferFromTarArchive");
}
}
void finalizeImpl() override
{
next();
closeFile(/* throw_if_error=*/true);
endWritingFile();
}
void sync() override { next(); }
std::string getFileName() const override { return filename; }
private:
void nextImpl() override
{
if (!offset())
return;
if (entry == nullptr)
writeEntry();
ssize_t to_write = offset();
ssize_t written = archive_write_data(archive, working_buffer.begin(), offset());
if (written != to_write)
{
throw Exception(
ErrorCodes::CANNOT_PACK_ARCHIVE,
"Couldn't pack tar archive: Failed to write all bytes, {} of {}, filename={}",
written,
to_write,
quoteString(filename));
}
}
void writeEntry()
{
expected_size = getSize();
entry = archive_entry_new();
archive_entry_set_pathname(entry, filename.c_str());
archive_entry_set_size(entry, expected_size);
archive_entry_set_filetype(entry, static_cast<__LA_MODE_T>(0100000));
archive_entry_set_perm(entry, 0644);
checkResult(archive_write_header(archive, entry));
}
size_t getSize() const
{
if (size)
return size;
else
return offset();
}
void closeFile(bool throw_if_error)
{
if (entry)
{
archive_entry_free(entry);
entry = nullptr;
}
if (throw_if_error and bytes != expected_size)
{
throw Exception(
ErrorCodes::CANNOT_PACK_ARCHIVE,
"Couldn't pack tar archive: Wrote {} of expected {} , filename={}",
bytes,
expected_size,
quoteString(filename));
}
}
void endWritingFile()
{
if (auto archive_writer_ptr = archive_writer.lock())
archive_writer_ptr->endWritingFile();
}
void startWritingFile()
{
if (auto archive_writer_ptr = archive_writer.lock())
archive_writer_ptr->startWritingFile();
}
void checkResult(int code) { checkResultCodeImpl(code, filename); }
std::weak_ptr<LibArchiveWriter> archive_writer;
const String filename;
Entry entry;
Archive archive;
size_t size;
size_t expected_size;
};
LibArchiveWriter::LibArchiveWriter(const String & path_to_archive_, std::unique_ptr<WriteBuffer> archive_write_buffer_)
: path_to_archive(path_to_archive_)
{
if (archive_write_buffer_)
stream_info = std::make_unique<StreamInfo>(std::move(archive_write_buffer_));
}
void LibArchiveWriter::createArchive()
{
std::lock_guard lock{mutex};
archive = archive_write_new();
setFormatAndSettings();
if (stream_info)
{
//This allows use to write directly to a writebuffer rather than an intermediate buffer in libarchive.
//This has to be set otherwise zstd breaks due to extra bytes being written at the end of the archive.
archive_write_set_bytes_per_block(archive, 0);
archive_write_open2(archive, stream_info.get(), nullptr, &StreamInfo::memory_write, nullptr, nullptr);
}
else
archive_write_open_filename(archive, path_to_archive.c_str());
}
LibArchiveWriter::~LibArchiveWriter()
{
chassert((finalized || std::uncaught_exceptions() || std::current_exception()) && "LibArchiveWriter is not finalized in destructor.");
if (archive)
archive_write_free(archive);
}
std::unique_ptr<WriteBufferFromFileBase> LibArchiveWriter::writeFile(const String & filename, size_t size)
{
return std::make_unique<WriteBufferFromLibArchive>(std::static_pointer_cast<LibArchiveWriter>(shared_from_this()), filename, size);
}
std::unique_ptr<WriteBufferFromFileBase> LibArchiveWriter::writeFile(const String & filename)
{
return std::make_unique<WriteBufferFromLibArchive>(std::static_pointer_cast<LibArchiveWriter>(shared_from_this()), filename, 0);
}
bool LibArchiveWriter::isWritingFile() const
{
std::lock_guard lock{mutex};
return is_writing_file;
}
void LibArchiveWriter::endWritingFile()
{
std::lock_guard lock{mutex};
is_writing_file = false;
}
void LibArchiveWriter::startWritingFile()
{
std::lock_guard lock{mutex};
if (std::exchange(is_writing_file, true))
throw Exception(ErrorCodes::NOT_IMPLEMENTED, "Cannot write two files to a tar archive in parallel");
}
void LibArchiveWriter::finalize()
{
std::lock_guard lock{mutex};
if (finalized)
return;
if (archive)
archive_write_close(archive);
if (stream_info)
{
stream_info->archive_write_buffer->finalize();
stream_info.reset();
}
finalized = true;
}
void LibArchiveWriter::setPassword(const String & password_)
{
if (password_.empty())
return;
throw Exception(ErrorCodes::NOT_IMPLEMENTED, "Setting a password is not currently supported for libarchive");
}
LibArchiveWriter::Archive LibArchiveWriter::getArchive()
{
std::lock_guard lock{mutex};
return archive;
}
}
#endif

View File

@ -0,0 +1,77 @@
#pragma once
#include "config.h"
#if USE_LIBARCHIVE
# include <IO/Archives/ArchiveUtils.h>
# include <IO/Archives/IArchiveWriter.h>
# include <IO/WriteBufferFromFileBase.h>
# include <base/defines.h>
namespace DB
{
class WriteBufferFromFileBase;
/// Interface for writing an archive.
class LibArchiveWriter : public IArchiveWriter
{
public:
/// Constructs an archive that will be written as a file in the local filesystem.
explicit LibArchiveWriter(const String & path_to_archive_, std::unique_ptr<WriteBuffer> archive_write_buffer_);
/// Call finalize() before destructing IArchiveWriter.
~LibArchiveWriter() override;
/// Starts writing a file to the archive. The function returns a write buffer,
/// any data written to that buffer will be compressed and then put to the archive.
/// You can keep only one such buffer at a time, a buffer returned by previous call
/// of the function `writeFile()` should be destroyed before next call of `writeFile()`.
std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename) override;
/// LibArchive needs to know the size of the file being written. If the file size is not
/// passed in the the archive writer tries to infer the size by looking at the available
/// data in the buffer, if next is called before all data is written to the buffer
/// an exception is thrown.
std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename, size_t size) override;
/// Returns true if there is an active instance of WriteBuffer returned by writeFile().
/// This function should be used mostly for debugging purposes.
bool isWritingFile() const override;
/// Finalizes writing of the archive. This function must be always called at the end of writing.
/// (Unless an error appeared and the archive is in fact no longer needed.)
void finalize() override;
/// Sets compression method and level.
/// Changing them will affect next file in the archive.
//void setCompression(const String & compression_method_, int compression_level_) override;
/// Sets password. If the password is not empty it will enable encryption in the archive.
void setPassword(const String & password) override;
protected:
using Archive = struct archive *;
using Entry = struct archive_entry *;
/// derived classes must call createArchive. CreateArchive calls setFormatAndSettings.
void createArchive();
virtual void setFormatAndSettings() = 0;
Archive archive = nullptr;
String path_to_archive;
private:
class WriteBufferFromLibArchive;
class StreamInfo;
Archive getArchive();
void startWritingFile();
void endWritingFile();
std::unique_ptr<StreamInfo> stream_info TSA_GUARDED_BY(mutex) = nullptr;
bool is_writing_file TSA_GUARDED_BY(mutex) = false;
bool finalized TSA_GUARDED_BY(mutex) = false;
mutable std::mutex mutex;
};
}
#endif

View File

@ -0,0 +1,42 @@
#include <IO/Archives/TarArchiveWriter.h>
#if USE_LIBARCHIVE
namespace DB
{
namespace ErrorCodes
{
extern const int NOT_IMPLEMENTED;
extern const int CANNOT_PACK_ARCHIVE;
}
void TarArchiveWriter::setCompression(const String & compression_method_, int compression_level_)
{
// throw an error unless setCompression is passed the default value
if (compression_method_.empty() && compression_level_ == -1)
return;
throw Exception(
ErrorCodes::NOT_IMPLEMENTED, "Using compression_method and compression_level options are not supported for tar archives");
}
void TarArchiveWriter::setFormatAndSettings()
{
archive_write_set_format_pax_restricted(archive);
inferCompressionFromPath();
}
void TarArchiveWriter::inferCompressionFromPath()
{
if (path_to_archive.ends_with(".tar.gz") || path_to_archive.ends_with(".tgz"))
archive_write_add_filter_gzip(archive);
else if (path_to_archive.ends_with(".tar.bz2"))
archive_write_add_filter_bzip2(archive);
else if (path_to_archive.ends_with(".tar.lzma"))
archive_write_add_filter_lzma(archive);
else if (path_to_archive.ends_with(".tar.zst") || path_to_archive.ends_with(".tzst"))
archive_write_add_filter_zstd(archive);
else if (path_to_archive.ends_with(".tar.xz"))
archive_write_add_filter_xz(archive);
else if (!path_to_archive.ends_with(".tar"))
throw Exception(ErrorCodes::CANNOT_PACK_ARCHIVE, "Unknown compression format");
}
}
#endif

View File

@ -0,0 +1,26 @@
#pragma once
#include "config.h"
#if USE_LIBARCHIVE
# include <IO/Archives/LibArchiveWriter.h>
namespace DB
{
using namespace std::literals;
class TarArchiveWriter : public LibArchiveWriter
{
public:
explicit TarArchiveWriter(const String & path_to_archive_, std::unique_ptr<WriteBuffer> archive_write_buffer_)
: LibArchiveWriter(path_to_archive_, std::move(archive_write_buffer_))
{
createArchive();
}
void setCompression(const String & compression_method_, int compression_level_) override;
void setFormatAndSettings() override;
void inferCompressionFromPath();
};
}
#endif

View File

@ -274,6 +274,11 @@ std::unique_ptr<WriteBufferFromFileBase> ZipArchiveWriter::writeFile(const Strin
return std::make_unique<WriteBufferFromZipArchive>(std::static_pointer_cast<ZipArchiveWriter>(shared_from_this()), filename); return std::make_unique<WriteBufferFromZipArchive>(std::static_pointer_cast<ZipArchiveWriter>(shared_from_this()), filename);
} }
std::unique_ptr<WriteBufferFromFileBase> ZipArchiveWriter::writeFile(const String & filename, [[maybe_unused]] size_t size)
{
return ZipArchiveWriter::writeFile(filename);
}
bool ZipArchiveWriter::isWritingFile() const bool ZipArchiveWriter::isWritingFile() const
{ {
std::lock_guard lock{mutex}; std::lock_guard lock{mutex};

View File

@ -32,6 +32,9 @@ public:
/// of the function `writeFile()` should be destroyed before next call of `writeFile()`. /// of the function `writeFile()` should be destroyed before next call of `writeFile()`.
std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename) override; std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename) override;
std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & filename, size_t size) override;
/// Returns true if there is an active instance of WriteBuffer returned by writeFile(). /// Returns true if there is an active instance of WriteBuffer returned by writeFile().
/// This function should be used mostly for debugging purposes. /// This function should be used mostly for debugging purposes.
bool isWritingFile() const override; bool isWritingFile() const override;

View File

@ -1,6 +1,6 @@
#include <IO/Archives/createArchiveReader.h>
#include <IO/Archives/ZipArchiveReader.h>
#include <IO/Archives/LibArchiveReader.h> #include <IO/Archives/LibArchiveReader.h>
#include <IO/Archives/ZipArchiveReader.h>
#include <IO/Archives/createArchiveReader.h>
#include <Common/Exception.h> #include <Common/Exception.h>
@ -25,16 +25,8 @@ std::shared_ptr<IArchiveReader> createArchiveReader(
[[maybe_unused]] size_t archive_size) [[maybe_unused]] size_t archive_size)
{ {
using namespace std::literals; using namespace std::literals;
static constexpr std::array tar_extensions static constexpr std::array tar_extensions{
{ ".tar"sv, ".tar.gz"sv, ".tgz"sv, ".tar.zst"sv, ".tzst"sv, ".tar.xz"sv, ".tar.bz2"sv, ".tar.lzma"sv};
".tar"sv,
".tar.gz"sv,
".tgz"sv,
".tar.zst"sv,
".tzst"sv,
".tar.xz"sv,
".tar.bz2"sv
};
if (path_to_archive.ends_with(".zip") || path_to_archive.ends_with(".zipx")) if (path_to_archive.ends_with(".zip") || path_to_archive.ends_with(".zipx"))
{ {
@ -48,7 +40,7 @@ std::shared_ptr<IArchiveReader> createArchiveReader(
tar_extensions.begin(), tar_extensions.end(), [&](const auto extension) { return path_to_archive.ends_with(extension); })) tar_extensions.begin(), tar_extensions.end(), [&](const auto extension) { return path_to_archive.ends_with(extension); }))
{ {
#if USE_LIBARCHIVE #if USE_LIBARCHIVE
return std::make_shared<TarArchiveReader>(path_to_archive); return std::make_shared<TarArchiveReader>(path_to_archive, archive_read_function);
#else #else
throw Exception(ErrorCodes::SUPPORT_IS_DISABLED, "libarchive library is disabled"); throw Exception(ErrorCodes::SUPPORT_IS_DISABLED, "libarchive library is disabled");
#endif #endif

View File

@ -1,5 +1,7 @@
#include <IO/Archives/createArchiveWriter.h> #include <IO/Archives/LibArchiveWriter.h>
#include <IO/Archives/TarArchiveWriter.h>
#include <IO/Archives/ZipArchiveWriter.h> #include <IO/Archives/ZipArchiveWriter.h>
#include <IO/Archives/createArchiveWriter.h>
#include <IO/WriteBuffer.h> #include <IO/WriteBuffer.h>
#include <Common/Exception.h> #include <Common/Exception.h>
@ -19,20 +21,30 @@ std::shared_ptr<IArchiveWriter> createArchiveWriter(const String & path_to_archi
} }
std::shared_ptr<IArchiveWriter> createArchiveWriter( std::shared_ptr<IArchiveWriter>
const String & path_to_archive, createArchiveWriter(const String & path_to_archive, [[maybe_unused]] std::unique_ptr<WriteBuffer> archive_write_buffer)
[[maybe_unused]] std::unique_ptr<WriteBuffer> archive_write_buffer)
{ {
using namespace std::literals;
static constexpr std::array tar_extensions{
".tar"sv, ".tar.gz"sv, ".tgz"sv, ".tar.bz2"sv, ".tar.lzma"sv, ".tar.zst"sv, ".tzst"sv, ".tar.xz"sv};
if (path_to_archive.ends_with(".zip") || path_to_archive.ends_with(".zipx")) if (path_to_archive.ends_with(".zip") || path_to_archive.ends_with(".zipx"))
{ {
#if USE_MINIZIP #if USE_MINIZIP
return std::make_shared<ZipArchiveWriter>(path_to_archive, std::move(archive_write_buffer)); return std::make_shared<ZipArchiveWriter>(path_to_archive, std::move(archive_write_buffer));
#else #else
throw Exception(ErrorCodes::SUPPORT_IS_DISABLED, "minizip library is disabled"); throw Exception(ErrorCodes::SUPPORT_IS_DISABLED, "minizip library is disabled");
#endif
}
else if (std::any_of(
tar_extensions.begin(), tar_extensions.end(), [&](const auto extension) { return path_to_archive.ends_with(extension); }))
{
#if USE_LIBARCHIVE
return std::make_shared<TarArchiveWriter>(path_to_archive, std::move(archive_write_buffer));
#else
throw Exception(ErrorCodes::SUPPORT_IS_DISABLED, "libarchive library is disabled");
#endif #endif
} }
else else
throw Exception(ErrorCodes::CANNOT_PACK_ARCHIVE, "Cannot determine the type of archive {}", path_to_archive); throw Exception(ErrorCodes::CANNOT_PACK_ARCHIVE, "Cannot determine the type of archive {}", path_to_archive);
} }
} }

View File

@ -6,7 +6,10 @@ namespace DB
bool hasRegisteredArchiveFileExtension(const String & path) bool hasRegisteredArchiveFileExtension(const String & path)
{ {
return path.ends_with(".zip") || path.ends_with(".zipx"); using namespace std::literals;
static constexpr std::array archive_extensions{
".zip"sv, ".zipx"sv, ".tar"sv, ".tar.gz"sv, ".tgz"sv, ".tar.bz2"sv, ".tar.lzma"sv, ".tar.zst"sv, ".tzst"sv, ".tar.xz"sv};
return std::any_of(
archive_extensions.begin(), archive_extensions.end(), [&](const auto extension) { return path.ends_with(extension); });
} }
} }

View File

@ -1,20 +1,23 @@
#include <gtest/gtest.h> #include <gtest/gtest.h>
#include "config.h" #include "config.h"
#include <filesystem>
#include <format>
#include <IO/Archives/ArchiveUtils.h> #include <IO/Archives/ArchiveUtils.h>
#include <IO/Archives/IArchiveReader.h> #include <IO/Archives/IArchiveReader.h>
#include <IO/Archives/IArchiveWriter.h> #include <IO/Archives/IArchiveWriter.h>
#include <IO/Archives/createArchiveReader.h> #include <IO/Archives/createArchiveReader.h>
#include <IO/Archives/createArchiveWriter.h> #include <IO/Archives/createArchiveWriter.h>
#include <IO/ReadBufferFromFile.h>
#include <IO/ReadBufferFromFileBase.h> #include <IO/ReadBufferFromFileBase.h>
#include <IO/ReadBufferFromString.h> #include <IO/ReadBufferFromString.h>
#include <IO/ReadHelpers.h> #include <IO/ReadHelpers.h>
#include <IO/WriteBufferFromFile.h>
#include <IO/WriteBufferFromFileBase.h> #include <IO/WriteBufferFromFileBase.h>
#include <IO/WriteBufferFromString.h> #include <IO/WriteBufferFromString.h>
#include <IO/WriteHelpers.h> #include <IO/WriteHelpers.h>
#include <Common/Exception.h>
#include <Poco/TemporaryFile.h> #include <Poco/TemporaryFile.h>
#include <filesystem> #include <Common/Exception.h>
namespace DB::ErrorCodes namespace DB::ErrorCodes
@ -49,7 +52,8 @@ bool createArchiveWithFiles(const std::string & archivename, const std::map<std:
archive_write_open_filename(a, archivename.c_str()); archive_write_open_filename(a, archivename.c_str());
for (const auto & [filename, content] : files) { for (const auto & [filename, content] : files)
{
entry = archive_entry_new(); entry = archive_entry_new();
archive_entry_set_pathname(entry, filename.c_str()); archive_entry_set_pathname(entry, filename.c_str());
archive_entry_set_size(entry, content.size()); archive_entry_set_size(entry, content.size());
@ -64,7 +68,6 @@ bool createArchiveWithFiles(const std::string & archivename, const std::map<std:
archive_write_free(a); archive_write_free(a);
return true; return true;
} }
class ArchiveReaderAndWriterTest : public ::testing::TestWithParam<const char *> class ArchiveReaderAndWriterTest : public ::testing::TestWithParam<const char *>
@ -114,10 +117,12 @@ TEST_P(ArchiveReaderAndWriterTest, EmptyArchive)
EXPECT_FALSE(reader->fileExists("nofile.txt")); EXPECT_FALSE(reader->fileExists("nofile.txt"));
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "File 'nofile.txt' was not found in archive", expectException(
[&]{ reader->getFileInfo("nofile.txt"); }); ErrorCodes::CANNOT_UNPACK_ARCHIVE, "File 'nofile.txt' was not found in archive", [&] { reader->getFileInfo("nofile.txt"); });
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "File 'nofile.txt' was not found in archive", expectException(
ErrorCodes::CANNOT_UNPACK_ARCHIVE,
"File 'nofile.txt' was not found in archive",
[&] { reader->readFile("nofile.txt", /*throw_on_not_found=*/true); }); [&] { reader->readFile("nofile.txt", /*throw_on_not_found=*/true); });
EXPECT_EQ(reader->firstFile(), nullptr); EXPECT_EQ(reader->firstFile(), nullptr);
@ -182,11 +187,9 @@ TEST_P(ArchiveReaderAndWriterTest, SingleFileInArchive)
auto enumerator = reader->firstFile(); auto enumerator = reader->firstFile();
ASSERT_NE(enumerator, nullptr); ASSERT_NE(enumerator, nullptr);
EXPECT_FALSE(enumerator->nextFile()); EXPECT_FALSE(enumerator->nextFile());
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "No current file", expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "No current file", [&] { enumerator->getFileName(); });
[&]{ enumerator->getFileName(); });
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "No current file", expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "No current file", [&] { reader->readFile(std::move(enumerator)); });
[&] { reader->readFile(std::move(enumerator)); });
} }
} }
@ -217,6 +220,10 @@ TEST_P(ArchiveReaderAndWriterTest, TwoFilesInArchive)
ASSERT_TRUE(reader->fileExists("a.txt")); ASSERT_TRUE(reader->fileExists("a.txt"));
ASSERT_TRUE(reader->fileExists("b/c.txt")); ASSERT_TRUE(reader->fileExists("b/c.txt"));
// Get all files
auto files = reader->getAllFiles();
EXPECT_EQ(files.size(), 2);
EXPECT_EQ(reader->getFileInfo("a.txt").uncompressed_size, a_contents.size()); EXPECT_EQ(reader->getFileInfo("a.txt").uncompressed_size, a_contents.size());
EXPECT_EQ(reader->getFileInfo("b/c.txt").uncompressed_size, c_contents.size()); EXPECT_EQ(reader->getFileInfo("b/c.txt").uncompressed_size, c_contents.size());
@ -272,6 +279,10 @@ TEST_P(ArchiveReaderAndWriterTest, TwoFilesInArchive)
enumerator = reader->nextFile(std::move(in)); enumerator = reader->nextFile(std::move(in));
EXPECT_EQ(enumerator, nullptr); EXPECT_EQ(enumerator, nullptr);
} }
// Get all files one last time
files = reader->getAllFiles();
EXPECT_EQ(files.size(), 2);
} }
@ -301,7 +312,8 @@ TEST_P(ArchiveReaderAndWriterTest, InMemory)
ASSERT_FALSE(fs::exists(getPathToArchive())); ASSERT_FALSE(fs::exists(getPathToArchive()));
/// Read the archive. /// Read the archive.
auto read_archive_func = [&]() -> std::unique_ptr<SeekableReadBuffer> { return std::make_unique<ReadBufferFromString>(archive_in_memory); }; auto read_archive_func
= [&]() -> std::unique_ptr<SeekableReadBuffer> { return std::make_unique<ReadBufferFromString>(archive_in_memory); };
auto reader = createArchiveReader(getPathToArchive(), read_archive_func, archive_in_memory.size()); auto reader = createArchiveReader(getPathToArchive(), read_archive_func, archive_in_memory.size());
ASSERT_TRUE(reader->fileExists("a.txt")); ASSERT_TRUE(reader->fileExists("a.txt"));
@ -334,12 +346,69 @@ TEST_P(ArchiveReaderAndWriterTest, InMemory)
} }
TEST_P(ArchiveReaderAndWriterTest, ManyFilesInMemory)
{
String archive_in_memory;
int files = 1000;
size_t times = 1;
/// Make an archive.
{
auto writer = createArchiveWriter(getPathToArchive(), std::make_unique<WriteBufferFromString>(archive_in_memory));
{
for (int i = 0; i < files; i++)
{
auto filename = std::format("{}.txt", i);
auto contents = std::format("The contents of {}.txt", i);
auto out = writer->writeFile(filename, times * contents.size());
for (int j = 0; j < times; j++)
writeString(contents, *out);
out->finalize();
}
}
writer->finalize();
}
/// The created archive is really in memory.
ASSERT_FALSE(fs::exists(getPathToArchive()));
/// Read the archive.
auto read_archive_func
= [&]() -> std::unique_ptr<SeekableReadBuffer> { return std::make_unique<ReadBufferFromString>(archive_in_memory); };
auto reader = createArchiveReader(getPathToArchive(), read_archive_func, archive_in_memory.size());
for (int i = 0; i < files; i++)
{
auto filename = std::format("{}.txt", i);
auto contents = std::format("The contents of {}.txt", i);
ASSERT_TRUE(reader->fileExists(filename));
EXPECT_EQ(reader->getFileInfo(filename).uncompressed_size, times * contents.size());
{
auto in = reader->readFile(filename, /*throw_on_not_found=*/true);
for (int j = 0; j < times; j++)
ASSERT_TRUE(checkString(String(contents), *in));
}
}
}
TEST_P(ArchiveReaderAndWriterTest, Password) TEST_P(ArchiveReaderAndWriterTest, Password)
{
auto writer = createArchiveWriter(getPathToArchive());
//don't support passwords for tar archives
if (getPathToArchive().ends_with(".tar") || getPathToArchive().ends_with(".tar.gz") || getPathToArchive().ends_with(".tar.bz2")
|| getPathToArchive().ends_with(".tar.lzma") || getPathToArchive().ends_with(".tar.zst") || getPathToArchive().ends_with(".tar.xz"))
{
expectException(
ErrorCodes::NOT_IMPLEMENTED,
"Setting a password is not currently supported for libarchive",
[&] { writer->setPassword("a.txt"); });
writer->finalize();
}
else
{ {
/// Make an archive. /// Make an archive.
std::string_view contents = "The contents of a.txt"; std::string_view contents = "The contents of a.txt";
{ {
auto writer = createArchiveWriter(getPathToArchive());
writer->setPassword("Qwe123"); writer->setPassword("Qwe123");
{ {
auto out = writer->writeFile("a.txt"); auto out = writer->writeFile("a.txt");
@ -353,14 +422,14 @@ TEST_P(ArchiveReaderAndWriterTest, Password)
auto reader = createArchiveReader(getPathToArchive()); auto reader = createArchiveReader(getPathToArchive());
/// Try to read without a password. /// Try to read without a password.
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Password is required", expectException(
[&]{ reader->readFile("a.txt", /*throw_on_not_found=*/true); }); ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Password is required", [&] { reader->readFile("a.txt", /*throw_on_not_found=*/true); });
{ {
/// Try to read with a wrong password. /// Try to read with a wrong password.
reader->setPassword("123Qwe"); reader->setPassword("123Qwe");
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Wrong password", expectException(
[&]{ reader->readFile("a.txt", /*throw_on_not_found=*/true); }); ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Wrong password", [&] { reader->readFile("a.txt", /*throw_on_not_found=*/true); });
} }
{ {
@ -372,15 +441,101 @@ TEST_P(ArchiveReaderAndWriterTest, Password)
EXPECT_EQ(str, contents); EXPECT_EQ(str, contents);
} }
} }
}
TEST_P(ArchiveReaderAndWriterTest, ArchiveNotExist) TEST_P(ArchiveReaderAndWriterTest, ArchiveNotExist)
{ {
expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Couldn't open", expectException(ErrorCodes::CANNOT_UNPACK_ARCHIVE, "Couldn't open", [&] { createArchiveReader(getPathToArchive()); });
[&]{ createArchiveReader(getPathToArchive()); });
} }
TEST(TarArchiveReaderTest, FileExists) {
TEST_P(ArchiveReaderAndWriterTest, ManyFilesOnDisk)
{
int files = 1000;
size_t times = 1;
/// Make an archive.
{
auto writer = createArchiveWriter(getPathToArchive());
{
for (int i = 0; i < files; i++)
{
auto filename = std::format("{}.txt", i);
auto contents = std::format("The contents of {}.txt", i);
auto out = writer->writeFile(filename, times * contents.size());
for (int j = 0; j < times; j++)
writeString(contents, *out);
out->finalize();
}
}
writer->finalize();
}
/// The created archive is really in memory.
ASSERT_TRUE(fs::exists(getPathToArchive()));
/// Read the archive.
auto reader = createArchiveReader(getPathToArchive());
for (int i = 0; i < files; i++)
{
auto filename = std::format("{}.txt", i);
auto contents = std::format("The contents of {}.txt", i);
ASSERT_TRUE(reader->fileExists(filename));
EXPECT_EQ(reader->getFileInfo(filename).uncompressed_size, times * contents.size());
{
auto in = reader->readFile(filename, /*throw_on_not_found=*/true);
for (int j = 0; j < times; j++)
ASSERT_TRUE(checkString(String(contents), *in));
}
}
}
TEST_P(ArchiveReaderAndWriterTest, LargeFile)
{
/// Make an archive.
std::string_view contents = "The contents of a.txt\n";
int times = 10000000;
{
auto writer = createArchiveWriter(getPathToArchive());
{
auto out = writer->writeFile("a.txt", times * contents.size());
for (int i = 0; i < times; i++)
writeString(contents, *out);
out->finalize();
}
writer->finalize();
}
/// Read the archive.
auto reader = createArchiveReader(getPathToArchive());
ASSERT_TRUE(reader->fileExists("a.txt"));
auto file_info = reader->getFileInfo("a.txt");
EXPECT_EQ(file_info.uncompressed_size, contents.size() * times);
EXPECT_GT(file_info.compressed_size, 0);
{
auto in = reader->readFile("a.txt", /*throw_on_not_found=*/true);
for (int i = 0; i < times; i++)
ASSERT_TRUE(checkString(String(contents), *in));
}
{
/// Use an enumerator.
auto enumerator = reader->firstFile();
ASSERT_NE(enumerator, nullptr);
EXPECT_EQ(enumerator->getFileName(), "a.txt");
EXPECT_EQ(enumerator->getFileInfo().uncompressed_size, contents.size() * times);
EXPECT_GT(enumerator->getFileInfo().compressed_size, 0);
EXPECT_FALSE(enumerator->nextFile());
}
}
TEST(TarArchiveReaderTest, FileExists)
{
String archive_path = "archive.tar"; String archive_path = "archive.tar";
String filename = "file.txt"; String filename = "file.txt";
String contents = "test"; String contents = "test";
@ -391,7 +546,8 @@ TEST(TarArchiveReaderTest, FileExists) {
fs::remove(archive_path); fs::remove(archive_path);
} }
TEST(TarArchiveReaderTest, ReadFile) { TEST(TarArchiveReaderTest, ReadFile)
{
String archive_path = "archive.tar"; String archive_path = "archive.tar";
String filename = "file.txt"; String filename = "file.txt";
String contents = "test"; String contents = "test";
@ -405,7 +561,8 @@ TEST(TarArchiveReaderTest, ReadFile) {
fs::remove(archive_path); fs::remove(archive_path);
} }
TEST(TarArchiveReaderTest, ReadTwoFiles) { TEST(TarArchiveReaderTest, ReadTwoFiles)
{
String archive_path = "archive.tar"; String archive_path = "archive.tar";
String file1 = "file1.txt"; String file1 = "file1.txt";
String contents1 = "test1"; String contents1 = "test1";
@ -428,7 +585,8 @@ TEST(TarArchiveReaderTest, ReadTwoFiles) {
} }
TEST(TarArchiveReaderTest, CheckFileInfo) { TEST(TarArchiveReaderTest, CheckFileInfo)
{
String archive_path = "archive.tar"; String archive_path = "archive.tar";
String filename = "file.txt"; String filename = "file.txt";
String contents = "test"; String contents = "test";
@ -441,7 +599,8 @@ TEST(TarArchiveReaderTest, CheckFileInfo) {
fs::remove(archive_path); fs::remove(archive_path);
} }
TEST(SevenZipArchiveReaderTest, FileExists) { TEST(SevenZipArchiveReaderTest, FileExists)
{
String archive_path = "archive.7z"; String archive_path = "archive.7z";
String filename = "file.txt"; String filename = "file.txt";
String contents = "test"; String contents = "test";
@ -452,7 +611,8 @@ TEST(SevenZipArchiveReaderTest, FileExists) {
fs::remove(archive_path); fs::remove(archive_path);
} }
TEST(SevenZipArchiveReaderTest, ReadFile) { TEST(SevenZipArchiveReaderTest, ReadFile)
{
String archive_path = "archive.7z"; String archive_path = "archive.7z";
String filename = "file.txt"; String filename = "file.txt";
String contents = "test"; String contents = "test";
@ -466,7 +626,8 @@ TEST(SevenZipArchiveReaderTest, ReadFile) {
fs::remove(archive_path); fs::remove(archive_path);
} }
TEST(SevenZipArchiveReaderTest, CheckFileInfo) { TEST(SevenZipArchiveReaderTest, CheckFileInfo)
{
String archive_path = "archive.7z"; String archive_path = "archive.7z";
String filename = "file.txt"; String filename = "file.txt";
String contents = "test"; String contents = "test";
@ -479,7 +640,8 @@ TEST(SevenZipArchiveReaderTest, CheckFileInfo) {
fs::remove(archive_path); fs::remove(archive_path);
} }
TEST(SevenZipArchiveReaderTest, ReadTwoFiles) { TEST(SevenZipArchiveReaderTest, ReadTwoFiles)
{
String archive_path = "archive.7z"; String archive_path = "archive.7z";
String file1 = "file1.txt"; String file1 = "file1.txt";
String contents1 = "test1"; String contents1 = "test1";
@ -502,16 +664,21 @@ TEST(SevenZipArchiveReaderTest, ReadTwoFiles) {
} }
#if USE_MINIZIP
namespace namespace
{ {
const char * supported_archive_file_exts[] = const char * supported_archive_file_exts[] = {
{ #if USE_MINIZIP
".zip" ".zip",
#endif
#if USE_LIBARCHIVE
".tar",
".tar.gz",
".tar.bz2",
".tar.lzma",
".tar.zst",
".tar.xz",
#endif
}; };
} }
INSTANTIATE_TEST_SUITE_P(All, ArchiveReaderAndWriterTest, ::testing::ValuesIn(supported_archive_file_exts)); INSTANTIATE_TEST_SUITE_P(All, ArchiveReaderAndWriterTest, ::testing::ValuesIn(supported_archive_file_exts));
#endif

View File

@ -566,7 +566,7 @@ namespace
}; };
} }
static void executeAction(const ExpressionActions::Action & action, ExecutionContext & execution_context, bool dry_run) static void executeAction(const ExpressionActions::Action & action, ExecutionContext & execution_context, bool dry_run, bool allow_duplicates_in_input)
{ {
auto & inputs = execution_context.inputs; auto & inputs = execution_context.inputs;
auto & columns = execution_context.columns; auto & columns = execution_context.columns;
@ -696,15 +696,20 @@ static void executeAction(const ExpressionActions::Action & action, ExecutionCon
"Not found column {} in block", "Not found column {} in block",
action.node->result_name); action.node->result_name);
} }
else
{
if (allow_duplicates_in_input)
columns[action.result_position] = inputs[pos];
else else
columns[action.result_position] = std::move(inputs[pos]); columns[action.result_position] = std::move(inputs[pos]);
}
break; break;
} }
} }
} }
void ExpressionActions::execute(Block & block, size_t & num_rows, bool dry_run) const void ExpressionActions::execute(Block & block, size_t & num_rows, bool dry_run, bool allow_duplicates_in_input) const
{ {
ExecutionContext execution_context ExecutionContext execution_context
{ {
@ -725,6 +730,7 @@ void ExpressionActions::execute(Block & block, size_t & num_rows, bool dry_run)
if (execution_context.inputs_pos[input_pos] < 0) if (execution_context.inputs_pos[input_pos] < 0)
{ {
execution_context.inputs_pos[input_pos] = pos; execution_context.inputs_pos[input_pos] = pos;
if (!allow_duplicates_in_input)
break; break;
} }
} }
@ -737,12 +743,8 @@ void ExpressionActions::execute(Block & block, size_t & num_rows, bool dry_run)
{ {
try try
{ {
executeAction(action, execution_context, dry_run); executeAction(action, execution_context, dry_run, allow_duplicates_in_input);
checkLimits(execution_context.columns); checkLimits(execution_context.columns);
//std::cerr << "Action: " << action.toString() << std::endl;
//for (const auto & col : execution_context.columns)
// std::cerr << col.dumpStructure() << std::endl;
} }
catch (Exception & e) catch (Exception & e)
{ {
@ -755,6 +757,12 @@ void ExpressionActions::execute(Block & block, size_t & num_rows, bool dry_run)
{ {
block.clear(); block.clear();
} }
else if (allow_duplicates_in_input)
{
/// This case is the same as when the input is projected
/// since we do not need any input columns.
block.clear();
}
else else
{ {
::sort(execution_context.inputs_pos.rbegin(), execution_context.inputs_pos.rend()); ::sort(execution_context.inputs_pos.rbegin(), execution_context.inputs_pos.rend());
@ -777,11 +785,11 @@ void ExpressionActions::execute(Block & block, size_t & num_rows, bool dry_run)
num_rows = execution_context.num_rows; num_rows = execution_context.num_rows;
} }
void ExpressionActions::execute(Block & block, bool dry_run) const void ExpressionActions::execute(Block & block, bool dry_run, bool allow_duplicates_in_input) const
{ {
size_t num_rows = block.rows(); size_t num_rows = block.rows();
execute(block, num_rows, dry_run); execute(block, num_rows, dry_run, allow_duplicates_in_input);
if (!block) if (!block)
block.insert({DataTypeUInt8().createColumnConst(num_rows, 0), std::make_shared<DataTypeUInt8>(), "_dummy"}); block.insert({DataTypeUInt8().createColumnConst(num_rows, 0), std::make_shared<DataTypeUInt8>(), "_dummy"});

View File

@ -98,9 +98,15 @@ public:
const NamesAndTypesList & getRequiredColumnsWithTypes() const { return required_columns; } const NamesAndTypesList & getRequiredColumnsWithTypes() const { return required_columns; }
/// Execute the expression on the block. The block must contain all the columns returned by getRequiredColumns. /// Execute the expression on the block. The block must contain all the columns returned by getRequiredColumns.
void execute(Block & block, size_t & num_rows, bool dry_run = false) const; ///
/// @param allow_duplicates_in_input - actions are allowed to have
/// duplicated input (that will refer into the block). This is needed for
/// preliminary query filtering (filterBlockWithDAG()), because they just
/// pass available virtual columns, which cannot be moved in case they are
/// used multiple times.
void execute(Block & block, size_t & num_rows, bool dry_run = false, bool allow_duplicates_in_input = false) const;
/// The same, but without `num_rows`. If result block is empty, adds `_dummy` column to keep block size. /// The same, but without `num_rows`. If result block is empty, adds `_dummy` column to keep block size.
void execute(Block & block, bool dry_run = false) const; void execute(Block & block, bool dry_run = false, bool allow_duplicates_in_input = false) const;
bool hasArrayJoin() const; bool hasArrayJoin() const;
void assertDeterministic() const; void assertDeterministic() const;

View File

@ -448,21 +448,16 @@ static ColumnPtr andFilters(ColumnPtr c1, ColumnPtr c2)
throw Exception(ErrorCodes::LOGICAL_ERROR, "Size of filters don't match: {} and {}", throw Exception(ErrorCodes::LOGICAL_ERROR, "Size of filters don't match: {} and {}",
c1->size(), c2->size()); c1->size(), c2->size());
// TODO: use proper vectorized implementation of AND?
auto res = ColumnUInt8::create(c1->size()); auto res = ColumnUInt8::create(c1->size());
auto & res_data = res->getData(); auto & res_data = res->getData();
const auto & c1_data = typeid_cast<const ColumnUInt8&>(*c1).getData(); const auto & c1_data = typeid_cast<const ColumnUInt8&>(*c1).getData();
const auto & c2_data = typeid_cast<const ColumnUInt8&>(*c2).getData(); const auto & c2_data = typeid_cast<const ColumnUInt8&>(*c2).getData();
const size_t size = c1->size(); const size_t size = c1->size();
const size_t step = 16; /// The double NOT operators (!!) convert the non-zeros to the bool value of true (0x01) and zeros to false (0x00).
size_t i = 0; /// After casting them to UInt8, '&' could replace '&&' for the 'AND' operation implementation and at the same
/// NOTE: '&&' must be used instead of '&' for 'AND' operation because UInt8 columns might contain any non-zero /// time enable the auto vectorization.
/// value for true and we cannot bitwise AND them to get the correct result. for (size_t i = 0; i < size; ++i)
for (; i + step < size; i += step) res_data[i] = (static_cast<UInt8>(!!c1_data[i]) & static_cast<UInt8>(!!c2_data[i]));
for (size_t j = 0; j < step; ++j)
res_data[i+j] = (c1_data[i+j] && c2_data[i+j]);
for (; i < size; ++i)
res_data[i] = (c1_data[i] && c2_data[i]);
return res; return res;
} }

View File

@ -138,6 +138,57 @@ bool testCombineColumns(size_t size)
return true; return true;
} }
/* To ensure the vectorized DB::andFilters works as its scalar implementation, this test validates the AND (&&)
* of any combinations of the UInt8 values.
*/
bool testAndFilters(size_t size)
{
auto generateFastIncrementColumn = [](size_t len)->ColumnPtr
{
auto filter = ColumnUInt8::create(len);
auto & filter_data = filter->getData();
for (size_t i = 0; i < len; ++i)
filter_data[i] = static_cast<UInt8>(i & 0xFF);
return filter;
};
auto generateSlowIncrementColumn = [](size_t len)->ColumnPtr
{
auto filter = ColumnUInt8::create(len);
auto & filter_data = filter->getData();
for (size_t i = 0; i < len; ++i)
filter_data[i] = static_cast<UInt8>((i >> 8) & 0xFF);
return filter;
};
auto first_filter = generateFastIncrementColumn(size);
auto second_filter = generateSlowIncrementColumn(size);
auto result = andFilters(first_filter, second_filter);
const auto & first_filter_data = typeid_cast<const ColumnUInt8 *>(first_filter.get())->getData();
const auto & second_filter_data = typeid_cast<const ColumnUInt8 *>(second_filter.get())->getData();
const auto & result_data = typeid_cast<const ColumnUInt8 *>(result.get())->getData();
if (result->size() != size)
{
return false;
}
for (size_t i = 0; i < size; i++)
{
UInt8 expected = first_filter_data[i] && second_filter_data[i];
if (result_data[i] != expected)
return false;
}
return true;
}
TEST(MergeTree, CombineFilters) TEST(MergeTree, CombineFilters)
{ {
/// Tests with only 0/1 and fixed intervals. /// Tests with only 0/1 and fixed intervals.
@ -159,3 +210,18 @@ TEST(MergeTree, CombineFilters)
EXPECT_TRUE(testCombineColumns(2000)); EXPECT_TRUE(testCombineColumns(2000));
EXPECT_TRUE(testCombineColumns(200000)); EXPECT_TRUE(testCombineColumns(200000));
} }
TEST(MergeTree, AndFilters)
{
EXPECT_TRUE(testAndFilters(1));
EXPECT_TRUE(testAndFilters(2));
EXPECT_TRUE(testAndFilters(15));
EXPECT_TRUE(testAndFilters(16));
EXPECT_TRUE(testAndFilters(200));
EXPECT_TRUE(testAndFilters(201));
EXPECT_TRUE(testAndFilters(2000));
EXPECT_TRUE(testAndFilters(65535));
EXPECT_TRUE(testAndFilters(65536));
EXPECT_TRUE(testAndFilters(65537));
EXPECT_TRUE(testAndFilters(200000));
}

View File

@ -223,7 +223,7 @@ bool prepareFilterBlockWithQuery(const ASTPtr & query, ContextPtr context, Block
auto expression_actions = std::make_shared<ExpressionActions>(actions); auto expression_actions = std::make_shared<ExpressionActions>(actions);
auto block_with_constants = block; auto block_with_constants = block;
expression_actions->execute(block_with_constants); expression_actions->execute(block_with_constants, /*dry_run=*/ false, /*allow_duplicates_in_input=*/ true);
return block_with_constants.has(expr_column_name) && isColumnConst(*block_with_constants.getByName(expr_column_name).column); return block_with_constants.has(expr_column_name) && isColumnConst(*block_with_constants.getByName(expr_column_name).column);
}; };
@ -266,7 +266,7 @@ void filterBlockWithDAG(ActionsDAGPtr dag, Block & block, ContextPtr context)
auto actions = std::make_shared<ExpressionActions>(dag); auto actions = std::make_shared<ExpressionActions>(dag);
makeSets(actions, context); makeSets(actions, context);
Block block_with_filter = block; Block block_with_filter = block;
actions->execute(block_with_filter); actions->execute(block_with_filter, /*dry_run=*/ false, /*allow_duplicates_in_input=*/ true);
/// Filter the block. /// Filter the block.
String filter_column_name = dag->getOutputs().at(0)->result_name; String filter_column_name = dag->getOutputs().at(0)->result_name;
@ -313,7 +313,7 @@ void filterBlockWithQuery(const ASTPtr & query, Block & block, ContextPtr contex
makeSets(actions, context); makeSets(actions, context);
Block block_with_filter = block; Block block_with_filter = block;
actions->execute(block_with_filter); actions->execute(block_with_filter, /*dry_run=*/ false, /*allow_duplicates_in_input=*/ true);
/// Filter the block. /// Filter the block.
String filter_column_name = expression_ast->getColumnName(); String filter_column_name = expression_ast->getColumnName();

View File

@ -953,10 +953,18 @@ def _mark_success_action(
# FIXME: find generic design for propagating and handling job status (e.g. stop using statuses in GH api) # FIXME: find generic design for propagating and handling job status (e.g. stop using statuses in GH api)
# now job ca be build job w/o status data, any other job that exit with 0 with or w/o status data # now job ca be build job w/o status data, any other job that exit with 0 with or w/o status data
if CI_CONFIG.is_build_job(job): if CI_CONFIG.is_build_job(job):
# there is no status for build jobs # there is no CommitStatus for build jobs
# create dummy success to mark it as done # create dummy status relying on JobReport
# FIXME: consider creating commit status for build jobs too, to treat everything the same way # FIXME: consider creating commit status for build jobs too, to treat everything the same way
CommitStatusData(SUCCESS, "dummy description", "dummy_url").dump_status() job_report = JobReport.load() if JobReport.exist() else None
if job_report and job_report.status == SUCCESS:
CommitStatusData(
SUCCESS,
"dummy description",
"dummy_url",
pr_num=pr_info.number,
sha=pr_info.sha,
).dump_status()
job_status = None job_status = None
if CommitStatusData.exist(): if CommitStatusData.exist():

View File

@ -1,19 +1,7 @@
#!/usr/bin/env bash #!/usr/bin/env bash
set -xeuo pipefail set -xeuo pipefail
# Add cloudflare DNS as a fallback bash /usr/local/share/scripts/init-network.sh
# Get default gateway interface
IFACE=$(ip --json route list | jq '.[]|select(.dst == "default").dev' --raw-output)
# `Link 2 (eth0): 172.31.0.2`
ETH_DNS=$(resolvectl dns "$IFACE") || :
CLOUDFLARE_NS=1.1.1.1
if [[ "$ETH_DNS" ]] && [[ "${ETH_DNS#*: }" != *"$CLOUDFLARE_NS"* ]]; then
# Cut the leading legend
ETH_DNS=${ETH_DNS#*: }
# shellcheck disable=SC2206
new_dns=(${ETH_DNS} "$CLOUDFLARE_NS")
resolvectl dns "$IFACE" "${new_dns[@]}"
fi
# tune sysctl for network performance # tune sysctl for network performance
cat > /etc/sysctl.d/10-network-memory.conf << EOF cat > /etc/sysctl.d/10-network-memory.conf << EOF

View File

@ -60,19 +60,7 @@ export RUNNER_URL="https://github.com/${RUNNER_ORG}"
INSTANCE_ID=$(ec2metadata --instance-id) INSTANCE_ID=$(ec2metadata --instance-id)
export INSTANCE_ID export INSTANCE_ID
# Add cloudflare DNS as a fallback bash /usr/local/share/scripts/init-network.sh
# Get default gateway interface
IFACE=$(ip --json route list | jq '.[]|select(.dst == "default").dev' --raw-output)
# `Link 2 (eth0): 172.31.0.2`
ETH_DNS=$(resolvectl dns "$IFACE") || :
CLOUDFLARE_NS=1.1.1.1
if [[ "$ETH_DNS" ]] && [[ "${ETH_DNS#*: }" != *"$CLOUDFLARE_NS"* ]]; then
# Cut the leading legend
ETH_DNS=${ETH_DNS#*: }
# shellcheck disable=SC2206
new_dns=(${ETH_DNS} "$CLOUDFLARE_NS")
resolvectl dns "$IFACE" "${new_dns[@]}"
fi
# combine labels # combine labels
RUNNER_TYPE=$(/usr/local/bin/aws ec2 describe-tags --filters "Name=resource-id,Values=$INSTANCE_ID" --query "Tags[?Key=='github:runner-type'].Value" --output text) RUNNER_TYPE=$(/usr/local/bin/aws ec2 describe-tags --filters "Name=resource-id,Values=$INSTANCE_ID" --query "Tags[?Key=='github:runner-type'].Value" --output text)

View File

@ -9,7 +9,7 @@ set -xeuo pipefail
echo "Running prepare script" echo "Running prepare script"
export DEBIAN_FRONTEND=noninteractive export DEBIAN_FRONTEND=noninteractive
export RUNNER_VERSION=2.311.0 export RUNNER_VERSION=2.313.0
export RUNNER_HOME=/home/ubuntu/actions-runner export RUNNER_HOME=/home/ubuntu/actions-runner
deb_arch() { deb_arch() {
@ -138,6 +138,49 @@ dpkg -i /tmp/amazon-cloudwatch-agent.deb
aws ssm get-parameter --region us-east-1 --name AmazonCloudWatch-github-runners --query 'Parameter.Value' --output text > /opt/aws/amazon-cloudwatch-agent/etc/amazon-cloudwatch-agent.json aws ssm get-parameter --region us-east-1 --name AmazonCloudWatch-github-runners --query 'Parameter.Value' --output text > /opt/aws/amazon-cloudwatch-agent/etc/amazon-cloudwatch-agent.json
systemctl enable amazon-cloudwatch-agent.service systemctl enable amazon-cloudwatch-agent.service
echo "Install tailscale"
# Build get-authkey for tailscale
docker run --rm -v /usr/local/bin/:/host-local-bin -i golang:alpine sh -ex <<'EOF'
CGO_ENABLED=0 go install -tags tag:svc-core-ci-github tailscale.com/cmd/get-authkey@main
mv /go/bin/get-authkey /host-local-bin
EOF
# install tailscale
curl -fsSL "https://pkgs.tailscale.com/stable/ubuntu/$(lsb_release -cs).noarmor.gpg" > /usr/share/keyrings/tailscale-archive-keyring.gpg
curl -fsSL "https://pkgs.tailscale.com/stable/ubuntu/$(lsb_release -cs).tailscale-keyring.list" > /etc/apt/sources.list.d/tailscale.list
apt-get update
apt-get install tailscale --yes --no-install-recommends
# Create a common script for the instances
mkdir /usr/local/share/scripts -p
cat > /usr/local/share/scripts/init-network.sh << 'EOF'
#!/usr/bin/env bash
# Add cloudflare DNS as a fallback
# Get default gateway interface
IFACE=$(ip --json route list | jq '.[]|select(.dst == "default").dev' --raw-output)
# `Link 2 (eth0): 172.31.0.2`
ETH_DNS=$(resolvectl dns "$IFACE") || :
CLOUDFLARE_NS=1.1.1.1
if [[ "$ETH_DNS" ]] && [[ "${ETH_DNS#*: }" != *"$CLOUDFLARE_NS"* ]]; then
# Cut the leading legend
ETH_DNS=${ETH_DNS#*: }
# shellcheck disable=SC2206
new_dns=(${ETH_DNS} "$CLOUDFLARE_NS")
resolvectl dns "$IFACE" "${new_dns[@]}"
fi
# Setup tailscale, the very first action
TS_API_CLIENT_ID=$(aws ssm get-parameter --region us-east-1 --name /tailscale/api-client-id --query 'Parameter.Value' --output text --with-decryption)
TS_API_CLIENT_SECRET=$(aws ssm get-parameter --region us-east-1 --name /tailscale/api-client-secret --query 'Parameter.Value' --output text --with-decryption)
export TS_API_CLIENT_ID TS_API_CLIENT_SECRET
TS_AUTHKEY=$(get-authkey -tags tag:svc-core-ci-github -reusable -ephemeral)
tailscale up --ssh --auth-key="$TS_AUTHKEY" --hostname="ci-runner-$INSTANCE_ID"
EOF
# The following line is used in aws TOE check. # The following line is used in aws TOE check.
touch /var/tmp/clickhouse-ci-ami.success touch /var/tmp/clickhouse-ci-ami.success
# END OF THE SCRIPT # END OF THE SCRIPT

View File

@ -591,6 +591,138 @@ def test_zip_archive_with_bad_compression_method():
) )
def test_tar_archive():
backup_name = f"Disk('backups', 'archive.tar')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
instance.query(f"BACKUP TABLE test.table TO {backup_name}")
assert os.path.isfile(get_path_to_backup(backup_name))
instance.query("DROP TABLE test.table")
assert instance.query("EXISTS test.table") == "0\n"
instance.query(f"RESTORE TABLE test.table FROM {backup_name}")
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
def test_tar_bz2_archive():
backup_name = f"Disk('backups', 'archive.tar.bz2')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
instance.query(f"BACKUP TABLE test.table TO {backup_name}")
assert os.path.isfile(get_path_to_backup(backup_name))
instance.query("DROP TABLE test.table")
assert instance.query("EXISTS test.table") == "0\n"
instance.query(f"RESTORE TABLE test.table FROM {backup_name}")
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
def test_tar_gz_archive():
backup_name = f"Disk('backups', 'archive.tar.gz')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
instance.query(f"BACKUP TABLE test.table TO {backup_name}")
assert os.path.isfile(get_path_to_backup(backup_name))
instance.query("DROP TABLE test.table")
assert instance.query("EXISTS test.table") == "0\n"
instance.query(f"RESTORE TABLE test.table FROM {backup_name}")
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
def test_tar_lzma_archive():
backup_name = f"Disk('backups', 'archive.tar.lzma')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
instance.query(f"BACKUP TABLE test.table TO {backup_name}")
assert os.path.isfile(get_path_to_backup(backup_name))
instance.query("DROP TABLE test.table")
assert instance.query("EXISTS test.table") == "0\n"
instance.query(f"RESTORE TABLE test.table FROM {backup_name}")
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
def test_tar_zst_archive():
backup_name = f"Disk('backups', 'archive.tar.zst')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
instance.query(f"BACKUP TABLE test.table TO {backup_name}")
assert os.path.isfile(get_path_to_backup(backup_name))
instance.query("DROP TABLE test.table")
assert instance.query("EXISTS test.table") == "0\n"
instance.query(f"RESTORE TABLE test.table FROM {backup_name}")
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
def test_tar_xz_archive():
backup_name = f"Disk('backups', 'archive.tar.xz')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
instance.query(f"BACKUP TABLE test.table TO {backup_name}")
assert os.path.isfile(get_path_to_backup(backup_name))
instance.query("DROP TABLE test.table")
assert instance.query("EXISTS test.table") == "0\n"
instance.query(f"RESTORE TABLE test.table FROM {backup_name}")
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
def test_tar_archive_with_password():
backup_name = f"Disk('backups', 'archive_with_password.tar')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
expected_error = "Setting a password is not currently supported for libarchive"
assert expected_error in instance.query_and_get_error(
f"BACKUP TABLE test.table TO {backup_name} SETTINGS id='tar_archive_with_password', password='password123'"
)
assert (
instance.query(
"SELECT status FROM system.backups WHERE id='tar_archive_with_password'"
)
== "BACKUP_FAILED\n"
)
def test_tar_archive_with_bad_compression_method():
backup_name = f"Disk('backups', 'archive_with_bad_compression_method.tar')"
create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"
expected_error = "Using compression_method and compression_level options are not supported for tar archives"
assert expected_error in instance.query_and_get_error(
f"BACKUP TABLE test.table TO {backup_name} SETTINGS id='tar_archive_with_bad_compression_method', compression_method='foobar'"
)
assert (
instance.query(
"SELECT status FROM system.backups WHERE id='tar_archive_with_bad_compression_method'"
)
== "BACKUP_FAILED\n"
)
def test_async(): def test_async():
create_and_fill_table() create_and_fill_table()
assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n" assert instance.query("SELECT count(), sum(x) FROM test.table") == "100\t4950\n"

View File

@ -454,6 +454,48 @@ def test_backup_to_zip():
check_backup_and_restore(storage_policy, backup_destination) check_backup_and_restore(storage_policy, backup_destination)
def test_backup_to_tar():
storage_policy = "default"
backup_name = new_backup_name()
backup_destination = f"S3('http://minio1:9001/root/data/backups/{backup_name}.tar', 'minio', 'minio123')"
check_backup_and_restore(storage_policy, backup_destination)
def test_backup_to_tar_gz():
storage_policy = "default"
backup_name = new_backup_name()
backup_destination = f"S3('http://minio1:9001/root/data/backups/{backup_name}.tar.gz', 'minio', 'minio123')"
check_backup_and_restore(storage_policy, backup_destination)
def test_backup_to_tar_bz2():
storage_policy = "default"
backup_name = new_backup_name()
backup_destination = f"S3('http://minio1:9001/root/data/backups/{backup_name}.tar.bz2', 'minio', 'minio123')"
check_backup_and_restore(storage_policy, backup_destination)
def test_backup_to_tar_lzma():
storage_policy = "default"
backup_name = new_backup_name()
backup_destination = f"S3('http://minio1:9001/root/data/backups/{backup_name}.tar.lzma', 'minio', 'minio123')"
check_backup_and_restore(storage_policy, backup_destination)
def test_backup_to_tar_zst():
storage_policy = "default"
backup_name = new_backup_name()
backup_destination = f"S3('http://minio1:9001/root/data/backups/{backup_name}.tar.zst', 'minio', 'minio123')"
check_backup_and_restore(storage_policy, backup_destination)
def test_backup_to_tar_xz():
storage_policy = "default"
backup_name = new_backup_name()
backup_destination = f"S3('http://minio1:9001/root/data/backups/{backup_name}.tar.xz', 'minio', 'minio123')"
check_backup_and_restore(storage_policy, backup_destination)
def test_user_specific_auth(start_cluster): def test_user_specific_auth(start_cluster):
def create_user(user): def create_user(user):
node.query(f"CREATE USER {user}") node.query(f"CREATE USER {user}")

View File

@ -613,7 +613,8 @@ def test_endpoint(cluster):
container_client = cluster.blob_service_client.get_container_client(container_name) container_client = cluster.blob_service_client.get_container_client(container_name)
container_client.create_container() container_client.create_container()
node.query( azure_query(
node,
f""" f"""
DROP TABLE IF EXISTS test SYNC; DROP TABLE IF EXISTS test SYNC;
@ -622,13 +623,128 @@ def test_endpoint(cluster):
SETTINGS disk = disk( SETTINGS disk = disk(
type = azure_blob_storage, type = azure_blob_storage,
endpoint = 'http://azurite1:{port}/{account_name}/{container_name}/{data_prefix}', endpoint = 'http://azurite1:{port}/{account_name}/{container_name}/{data_prefix}',
endpoint_contains_account_name = 'true',
account_name = 'devstoreaccount1', account_name = 'devstoreaccount1',
account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==',
container_already_exists = 1, container_already_exists = 1,
skip_access_check = 0); skip_access_check = 0);
INSERT INTO test SELECT number FROM numbers(10); INSERT INTO test SELECT number FROM numbers(10);
""" """,
) )
assert 10 == int(node.query("SELECT count() FROM test")) assert 10 == int(node.query("SELECT count() FROM test"))
def test_endpoint_new_container(cluster):
node = cluster.instances[NODE_NAME]
account_name = "devstoreaccount1"
container_name = "cont3"
data_prefix = "data_prefix"
port = cluster.azurite_port
azure_query(
node,
f"""
DROP TABLE IF EXISTS test SYNC;
CREATE TABLE test (a Int32)
ENGINE = MergeTree() ORDER BY tuple()
SETTINGS disk = disk(
type = azure_blob_storage,
endpoint = 'http://azurite1:{port}/{account_name}/{container_name}/{data_prefix}',
endpoint_contains_account_name = 'true',
account_name = 'devstoreaccount1',
account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==',
skip_access_check = 0);
INSERT INTO test SELECT number FROM numbers(10);
""",
)
assert 10 == int(node.query("SELECT count() FROM test"))
def test_endpoint_without_prefix(cluster):
node = cluster.instances[NODE_NAME]
account_name = "devstoreaccount1"
container_name = "cont4"
port = cluster.azurite_port
azure_query(
node,
f"""
DROP TABLE IF EXISTS test SYNC;
CREATE TABLE test (a Int32)
ENGINE = MergeTree() ORDER BY tuple()
SETTINGS disk = disk(
type = azure_blob_storage,
endpoint = 'http://azurite1:{port}/{account_name}/{container_name}',
endpoint_contains_account_name = 'true',
account_name = 'devstoreaccount1',
account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==',
skip_access_check = 0);
INSERT INTO test SELECT number FROM numbers(10);
""",
)
assert 10 == int(node.query("SELECT count() FROM test"))
def test_endpoint_error_check(cluster):
node = cluster.instances[NODE_NAME]
account_name = "devstoreaccount1"
port = cluster.azurite_port
query = f"""
DROP TABLE IF EXISTS test SYNC;
CREATE TABLE test (a Int32)
ENGINE = MergeTree() ORDER BY tuple()
SETTINGS disk = disk(
type = azure_blob_storage,
endpoint = 'http://azurite1:{port}/{account_name}',
endpoint_contains_account_name = 'true',
account_name = 'devstoreaccount1',
account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==',
skip_access_check = 0);
"""
expected_err_msg = "Expected container_name in endpoint"
assert expected_err_msg in azure_query(node, query, expect_error="true")
query = f"""
DROP TABLE IF EXISTS test SYNC;
CREATE TABLE test (a Int32)
ENGINE = MergeTree() ORDER BY tuple()
SETTINGS disk = disk(
type = azure_blob_storage,
endpoint = 'http://azurite1:{port}',
endpoint_contains_account_name = 'true',
account_name = 'devstoreaccount1',
account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==',
skip_access_check = 0);
"""
expected_err_msg = "Expected account_name in endpoint"
assert expected_err_msg in azure_query(node, query, expect_error="true")
query = f"""
DROP TABLE IF EXISTS test SYNC;
CREATE TABLE test (a Int32)
ENGINE = MergeTree() ORDER BY tuple()
SETTINGS disk = disk(
type = azure_blob_storage,
endpoint = 'http://azurite1:{port}',
endpoint_contains_account_name = 'false',
account_name = 'devstoreaccount1',
account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==',
skip_access_check = 0);
"""
expected_err_msg = "Expected container_name in endpoint"
assert expected_err_msg in azure_query(node, query, expect_error="true")

View File

@ -0,0 +1,6 @@
-- It is special because actions cannot be reused for SimpleAggregateFunction (see https://github.com/ClickHouse/ClickHouse/pull/54436)
drop table if exists data;
create table data (key Int) engine=AggregatingMergeTree() order by tuple();
insert into data values (0);
select * from data final prewhere indexHint(_partition_id = 'all') or indexHint(_partition_id = 'all');
select * from data final prewhere indexHint(_partition_id = 'all') or indexHint(_partition_id = 'all') or indexHint(_partition_id = 'all');

View File

@ -0,0 +1,5 @@
-- It is special because actions cannot be reused for SimpleAggregateFunction (see https://github.com/ClickHouse/ClickHouse/pull/54436)
drop table if exists data;
create table data (key SimpleAggregateFunction(max, Int)) engine=AggregatingMergeTree() order by tuple();
insert into data values (0);
select * from data final prewhere indexHint(_partition_id = 'all') and key >= -1 where key >= 0;

View File

@ -50,11 +50,6 @@ find $ROOT_PATH/{src,base,programs,utils} -name '*.h' -or -name '*.cpp' 2>/dev/n
# Broken symlinks # Broken symlinks
find -L $ROOT_PATH -type l 2>/dev/null | grep -v contrib && echo "^ Broken symlinks found" find -L $ROOT_PATH -type l 2>/dev/null | grep -v contrib && echo "^ Broken symlinks found"
# Double whitespaces
find $ROOT_PATH/{src,base,programs,utils} -name '*.h' -or -name '*.cpp' 2>/dev/null |
grep -vP $EXCLUDE_DIRS |
while read i; do $ROOT_PATH/utils/check-style/double-whitespaces.pl < $i || echo -e "^ File $i contains double whitespaces\n"; done
# Unused/Undefined/Duplicates ErrorCodes/ProfileEvents/CurrentMetrics # Unused/Undefined/Duplicates ErrorCodes/ProfileEvents/CurrentMetrics
declare -A EXTERN_TYPES declare -A EXTERN_TYPES
EXTERN_TYPES[ErrorCodes]=int EXTERN_TYPES[ErrorCodes]=int

View File

@ -2,8 +2,9 @@
ROOT_PATH=$(git rev-parse --show-toplevel) ROOT_PATH=$(git rev-parse --show-toplevel)
EXCLUDE_DIRS='build/|integration/|widechar_width/|glibc-compatibility/|memcpy/|consistent-hashing/|Parsers/New' EXCLUDE_DIRS='build/|integration/|widechar_width/|glibc-compatibility/|memcpy/|consistent-hashing/|Parsers/New'
NPROC=$(($(nproc) + 3))
# Double whitespaces # Double whitespaces
find $ROOT_PATH/{src,base,programs,utils} -name '*.h' -or -name '*.cpp' 2>/dev/null | find $ROOT_PATH/{src,base,programs,utils} -name '*.h' -or -name '*.cpp' 2>/dev/null |
grep -vP $EXCLUDE_DIRS | grep -vP $EXCLUDE_DIRS |
while read i; do $ROOT_PATH/utils/check-style/double-whitespaces.pl < $i || echo -e "^ File $i contains double whitespaces\n"; done xargs -P "$NPROC" -n 20 "${ROOT_PATH}/utils/check-style/double-whitespaces.pl"

View File

@ -5,15 +5,18 @@ use strict;
# Find double whitespace such as "a, b, c" that looks very ugly and annoying. # Find double whitespace such as "a, b, c" that looks very ugly and annoying.
# But skip double whitespaces if they are used as an alignment - by comparing to surrounding lines. # But skip double whitespaces if they are used as an alignment - by comparing to surrounding lines.
my $ret = 0;
foreach my $file (@ARGV)
{
my @array; my @array;
while (<>) open (FH,'<',$file);
while (<FH>)
{ {
push @array, $_; push @array, $_;
} }
my $ret = 0;
for (my $i = 1; $i < $#array; ++$i) for (my $i = 1; $i < $#array; ++$i)
{ {
if ($array[$i] =~ ',( {2,3})[^ /]') if ($array[$i] =~ ',( {2,3})[^ /]')
@ -24,10 +27,11 @@ for (my $i = 1; $i < $#array; ++$i)
&& (substr($array[$i + 1], $+[1] - 1, 2) !~ /^[ -][^ ]$/) && (substr($array[$i + 1], $+[1] - 1, 2) !~ /^[ -][^ ]$/)
&& $array[$i] !~ /(-?\d+\w*,\s+){3,}/) # this is not a number table like { 10, -1, 2 } && $array[$i] !~ /(-?\d+\w*,\s+){3,}/) # this is not a number table like { 10, -1, 2 }
{ {
print(($i + 1) . ":" . $array[$i]); print($file . ":" . ($i + 1) . $array[$i]);
$ret = 1; $ret = 1;
} }
} }
} }
}
exit $ret; exit $ret;