Fix checks

This commit is contained in:
kssenii 2021-05-21 08:35:13 +00:00
parent 074428264f
commit 6931573f69
10 changed files with 14 additions and 17 deletions

View File

@ -427,7 +427,6 @@ endif()
if (USE_HDFS)
dbms_target_link_libraries(PRIVATE ${HDFS3_LIBRARY})
dbms_target_include_directories (SYSTEM BEFORE PUBLIC ${HDFS3_INCLUDE_DIR})
target_include_directories (clickhouse_common_io SYSTEM BEFORE PUBLIC ${HDFS3_INCLUDE_DIR})
endif()
if (USE_AWS_S3)

View File

@ -3,8 +3,8 @@
#include <Storages/HDFS/ReadBufferFromHDFS.h>
#include <Storages/HDFS/WriteBufferFromHDFS.h>
#include <IO/SeekAvoidingReadBuffer.h>
#include <IO/ReadIndirectBufferFromRemoteFS.h>
#include <IO/WriteIndirectBufferFromRemoteFS.h>
#include <Disks/ReadIndirectBufferFromRemoteFS.h>
#include <Disks/WriteIndirectBufferFromRemoteFS.h>
#include <Common/checkStackSize.h>
#include <Common/quoteString.h>
#include <common/logger_useful.h>
@ -17,9 +17,6 @@ namespace ErrorCodes
{
extern const int BAD_ARGUMENTS;
extern const int LOGICAL_ERROR;
extern const int PATH_ACCESS_DENIED;
extern const int CANNOT_DELETE_DIRECTORY;
extern const int UNKNOWN_FORMAT;
}
/// Reads data from HDFS using stored paths in metadata.
@ -109,7 +106,7 @@ void DiskHDFS::removeFromRemoteFS(const RemoteFSPathKeeper & fs_paths_keeper)
{
for (const auto & hdfs_object_path : chunk)
{
const String hdfs_path = hdfs_object_path.GetKey();
const String & hdfs_path = hdfs_object_path.GetKey();
const size_t begin_of_path = hdfs_path.find('/', hdfs_path.find("//") + 2);
/// Add path from root to file name

View File

@ -47,7 +47,7 @@ public:
std::unique_ptr<WriteBufferFromFileBase> writeFile(const String & path, size_t buf_size, WriteMode mode) override;
void removeFromRemoteFS(const RemoteFSPathKeeper & keys) override;
void removeFromRemoteFS(const RemoteFSPathKeeper & remote_paths_keeper) override;
private:
String getRandomName() { return toString(UUIDHelpers::generateV4()); }

View File

@ -357,20 +357,20 @@ void IDiskRemote::removeFileIfExists(const String & path)
}
void IDiskRemote::removeSharedFile(const String & path, bool keep_remote_fs)
void IDiskRemote::removeSharedFile(const String & path, bool keep_in_remote_fs)
{
RemoteFSPathKeeper fs_paths_keeper;
removeMeta(path, fs_paths_keeper);
if (!keep_remote_fs)
if (!keep_in_remote_fs)
removeFromRemoteFS(fs_paths_keeper);
}
void IDiskRemote::removeSharedRecursive(const String & path, bool keep_remote_fs)
void IDiskRemote::removeSharedRecursive(const String & path, bool keep_in_remote_fs)
{
RemoteFSPathKeeper fs_paths_keeper;
removeMetaRecursive(path, fs_paths_keeper);
if (!keep_remote_fs)
if (!keep_in_remote_fs)
removeFromRemoteFS(fs_paths_keeper);
}

View File

@ -170,13 +170,13 @@ struct IDiskRemote::Metadata
String metadata_file_path;
/// Total size of all remote FS (S3, HDFS) objects.
size_t total_size;
size_t total_size = 0;
/// Remote FS (S3, HDFS) objects paths and their sizes.
std::vector<PathAndSize> remote_fs_objects;
/// Number of references (hardlinks) to this metadata file.
UInt32 ref_count;
UInt32 ref_count = 0;
/// Flag indicates that file is read only.
bool read_only = false;

View File

@ -1,3 +1,4 @@
#pragma once
#include <Common/config.h>
#if USE_AWS_S3 || USE_HDFS

View File

@ -8,8 +8,8 @@
#include <IO/ReadBufferFromString.h>
#include <Interpreters/Context.h>
#include <IO/ReadBufferFromS3.h>
#include <IO/ReadIndirectBufferFromRemoteFS.h>
#include <IO/WriteIndirectBufferFromRemoteFS.h>
#include <Disks/ReadIndirectBufferFromRemoteFS.h>
#include <Disks/WriteIndirectBufferFromRemoteFS.h>
#include <IO/ReadHelpers.h>
#include <IO/SeekAvoidingReadBuffer.h>
#include <IO/WriteBufferFromFile.h>
@ -31,7 +31,6 @@ namespace ErrorCodes
{
extern const int S3_ERROR;
extern const int FILE_ALREADY_EXISTS;
extern const int CANNOT_SEEK_THROUGH_FILE;
extern const int UNKNOWN_FORMAT;
extern const int BAD_ARGUMENTS;
extern const int LOGICAL_ERROR;

View File

@ -1,3 +1,4 @@
#pragma once
#include <Common/config.h>
#if USE_AWS_S3 || USE_HDFS