mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-25 17:12:03 +00:00
Handle cluster macro in s3Cluster table function
This commit is contained in:
parent
dd2e085ae5
commit
233f0b4b52
@ -286,8 +286,9 @@ Cluster::Address Cluster::Address::fromFullString(const String & full_string)
|
||||
|
||||
/// Implementation of Clusters class
|
||||
|
||||
Clusters::Clusters(const Poco::Util::AbstractConfiguration & config, const Settings & settings, const String & config_prefix)
|
||||
Clusters::Clusters(const Poco::Util::AbstractConfiguration & config, const Settings & settings, MultiVersion<Macros>::Version macros, const String & config_prefix)
|
||||
{
|
||||
this->macros_ = macros;
|
||||
updateClusters(config, settings, config_prefix);
|
||||
}
|
||||
|
||||
@ -296,7 +297,8 @@ ClusterPtr Clusters::getCluster(const std::string & cluster_name) const
|
||||
{
|
||||
std::lock_guard lock(mutex);
|
||||
|
||||
auto it = impl.find(cluster_name);
|
||||
auto expanded_cluster_name = macros_->expand(cluster_name);
|
||||
auto it = impl.find(expanded_cluster_name);
|
||||
return (it != impl.end()) ? it->second : nullptr;
|
||||
}
|
||||
|
||||
|
@ -2,6 +2,8 @@
|
||||
|
||||
#include <Client/ConnectionPool.h>
|
||||
#include <Client/ConnectionPoolWithFailover.h>
|
||||
#include <Common/Macros.h>
|
||||
#include <Common/MultiVersion.h>
|
||||
|
||||
#include <Poco/Net/SocketAddress.h>
|
||||
|
||||
@ -287,7 +289,7 @@ using ClusterPtr = std::shared_ptr<Cluster>;
|
||||
class Clusters
|
||||
{
|
||||
public:
|
||||
Clusters(const Poco::Util::AbstractConfiguration & config, const Settings & settings, const String & config_prefix = "remote_servers");
|
||||
Clusters(const Poco::Util::AbstractConfiguration & config, const Settings & settings, MultiVersion<Macros>::Version macros, const String & config_prefix = "remote_servers");
|
||||
|
||||
Clusters(const Clusters &) = delete;
|
||||
Clusters & operator=(const Clusters &) = delete;
|
||||
@ -306,6 +308,8 @@ protected:
|
||||
/// setup outside of this class, stored to prevent deleting from impl on config update
|
||||
std::unordered_set<std::string> automatic_clusters;
|
||||
|
||||
MultiVersion<Macros>::Version macros_;
|
||||
|
||||
Impl impl;
|
||||
mutable std::mutex mutex;
|
||||
};
|
||||
|
@ -2265,7 +2265,7 @@ void Context::reloadClusterConfig() const
|
||||
}
|
||||
|
||||
const auto & config = cluster_config ? *cluster_config : getConfigRef();
|
||||
auto new_clusters = std::make_shared<Clusters>(config, settings);
|
||||
auto new_clusters = std::make_shared<Clusters>(config, settings, getMacros());
|
||||
|
||||
{
|
||||
std::lock_guard lock(shared->clusters_mutex);
|
||||
@ -2287,7 +2287,7 @@ std::shared_ptr<Clusters> Context::getClusters() const
|
||||
if (!shared->clusters)
|
||||
{
|
||||
const auto & config = shared->clusters_config ? *shared->clusters_config : getConfigRef();
|
||||
shared->clusters = std::make_shared<Clusters>(config, settings);
|
||||
shared->clusters = std::make_shared<Clusters>(config, settings, getMacros());
|
||||
}
|
||||
|
||||
return shared->clusters;
|
||||
@ -2318,7 +2318,7 @@ void Context::setClustersConfig(const ConfigurationPtr & config, bool enable_dis
|
||||
shared->clusters_config = config;
|
||||
|
||||
if (!shared->clusters)
|
||||
shared->clusters = std::make_shared<Clusters>(*shared->clusters_config, settings, config_name);
|
||||
shared->clusters = std::make_shared<Clusters>(*shared->clusters_config, settings, getMacros(), config_name);
|
||||
else
|
||||
shared->clusters->updateClusters(*shared->clusters_config, settings, config_name, old_clusters_config);
|
||||
}
|
||||
|
@ -201,11 +201,10 @@ void TableFunctionRemote::parseArguments(const ASTPtr & ast_function, ContextPtr
|
||||
if (!cluster_name.empty())
|
||||
{
|
||||
/// Use an existing cluster from the main config
|
||||
String cluster_name_expanded = context->getMacros()->expand(cluster_name);
|
||||
if (name != "clusterAllReplicas")
|
||||
cluster = context->getCluster(cluster_name_expanded);
|
||||
cluster = context->getCluster(cluster_name);
|
||||
else
|
||||
cluster = context->getCluster(cluster_name_expanded)->getClusterWithReplicasAsShards(context->getSettingsRef());
|
||||
cluster = context->getCluster(cluster_name)->getClusterWithReplicasAsShards(context->getSettingsRef());
|
||||
}
|
||||
else
|
||||
{
|
||||
|
@ -21,4 +21,7 @@
|
||||
</cluster_simple>
|
||||
|
||||
</remote_servers>
|
||||
<macros>
|
||||
<default_cluster_macro>cluster_simple</default_cluster_macro>
|
||||
</macros>
|
||||
</clickhouse>
|
@ -95,6 +95,29 @@ def test_count(started_cluster):
|
||||
assert TSV(pure_s3) == TSV(s3_distibuted)
|
||||
|
||||
|
||||
def test_count_macro(started_cluster):
|
||||
node = started_cluster.instances["s0_0_0"]
|
||||
|
||||
s3_macro = node.query(
|
||||
"""
|
||||
SELECT count(*) from s3Cluster(
|
||||
'{default_cluster_macro}', 'http://minio1:9001/root/data/{clickhouse,database}/*',
|
||||
'minio', 'minio123', 'CSV',
|
||||
'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))')"""
|
||||
)
|
||||
# print(s3_distibuted)
|
||||
s3_distibuted = node.query(
|
||||
"""
|
||||
SELECT count(*) from s3Cluster(
|
||||
'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*',
|
||||
'minio', 'minio123', 'CSV',
|
||||
'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))')"""
|
||||
)
|
||||
# print(s3_distibuted)
|
||||
|
||||
assert TSV(s3_macro) == TSV(s3_distibuted)
|
||||
|
||||
|
||||
def test_union_all(started_cluster):
|
||||
node = started_cluster.instances["s0_0_0"]
|
||||
pure_s3 = node.query(
|
||||
|
5
tests/integration/test_storage_hdfs/configs/macro.xml
Normal file
5
tests/integration/test_storage_hdfs/configs/macro.xml
Normal file
@ -0,0 +1,5 @@
|
||||
<clickhouse>
|
||||
<macros>
|
||||
<default_cluster_macro>test_cluster_two_shards</default_cluster_macro>
|
||||
</macros>
|
||||
</clickhouse>
|
@ -2,10 +2,13 @@ import os
|
||||
|
||||
import pytest
|
||||
from helpers.cluster import ClickHouseCluster
|
||||
from helpers.test_tools import TSV
|
||||
from pyhdfs import HdfsClient
|
||||
|
||||
cluster = ClickHouseCluster(__file__)
|
||||
node1 = cluster.add_instance("node1", with_hdfs=True)
|
||||
node1 = cluster.add_instance(
|
||||
"node1", main_configs=["configs/macro.xml"], with_hdfs=True
|
||||
)
|
||||
|
||||
|
||||
@pytest.fixture(scope="module")
|
||||
@ -565,6 +568,22 @@ def test_cluster_join(started_cluster):
|
||||
assert "AMBIGUOUS_COLUMN_NAME" not in result
|
||||
|
||||
|
||||
def test_cluster_macro(started_cluster):
|
||||
with_macro = node1.query(
|
||||
"""
|
||||
SELECT id FROM hdfsCluster('{default_cluster_macro}', 'hdfs://hdfs1:9000/test_hdfsCluster/file*', 'TSV', 'id UInt32')
|
||||
"""
|
||||
)
|
||||
|
||||
no_macro = node1.query(
|
||||
"""
|
||||
SELECT id FROM hdfsCluster('test_cluster_two_shards', 'hdfs://hdfs1:9000/test_hdfsCluster/file*', 'TSV', 'id UInt32')
|
||||
"""
|
||||
)
|
||||
|
||||
assert TSV(with_macro) == TSV(no_macro)
|
||||
|
||||
|
||||
def test_virtual_columns_2(started_cluster):
|
||||
hdfs_api = started_cluster.hdfs_api
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user