2019-11-20 11:56:38 +00:00
import base64
import errno
2024-03-20 09:46:07 +00:00
from functools import cache
2020-10-02 16:54:07 +00:00
import http . client
2020-08-12 08:55:04 +00:00
import logging
2017-05-19 18:54:05 +00:00
import os
2024-03-20 09:46:07 +00:00
import platform
2021-03-05 13:39:51 +00:00
import stat
2017-05-19 18:54:05 +00:00
import os . path as p
2020-08-12 08:55:04 +00:00
import pprint
2017-05-23 14:24:04 +00:00
import pwd
2017-05-19 18:54:05 +00:00
import re
import shutil
import socket
2019-11-20 11:56:38 +00:00
import subprocess
2017-05-19 18:54:05 +00:00
import time
2020-09-09 11:11:59 +00:00
import traceback
2020-10-02 16:54:07 +00:00
import urllib . parse
2021-02-23 16:53:14 +00:00
import shlex
2021-04-29 11:57:48 +00:00
import urllib3
2020-09-16 04:26:10 +00:00
import requests
2022-03-21 13:22:15 +00:00
try :
# Please, add modules that required for specific tests only here.
# So contributors will be able to run most tests locally
# without installing tons of unneeded packages that may be not so easy to install.
2022-05-17 13:58:09 +00:00
import asyncio
2022-03-21 13:22:15 +00:00
from cassandra . policies import RoundRobinPolicy
import cassandra . cluster
import psycopg2
from psycopg2 . extensions import ISOLATION_LEVEL_AUTOCOMMIT
import pymongo
import pymysql
2022-05-17 13:58:09 +00:00
import nats
2022-07-24 14:13:17 +00:00
import ssl
2022-03-23 10:58:57 +00:00
from confluent_kafka . avro . cached_schema_registry_client import (
CachedSchemaRegistryClient ,
)
2023-06-21 18:45:28 +00:00
from . hdfs_api import HDFSApi # imports requests_kerberos
2022-03-21 13:22:15 +00:00
except Exception as e :
2022-03-22 14:34:20 +00:00
logging . warning ( f " Cannot import some modules, some tests may not work: { e } " )
2022-03-21 13:22:15 +00:00
2021-01-27 09:50:11 +00:00
from dict2xml import dict2xml
2019-11-20 11:56:38 +00:00
from kazoo . client import KazooClient
from kazoo . exceptions import KazooException
from minio import Minio
2021-07-12 08:32:20 +00:00
2021-10-19 10:19:43 +00:00
from helpers . test_tools import assert_eq_with_retry , exec_query_with_retry
2021-07-01 14:41:59 +00:00
from helpers import pytest_xdist_logging_to_separate_files
2021-10-13 11:55:58 +00:00
from helpers . client import QueryRuntimeException
2017-05-19 18:54:05 +00:00
2021-02-24 11:46:58 +00:00
import docker
2019-11-20 11:56:38 +00:00
from . client import Client
2024-07-24 10:44:06 +00:00
from . retry_decorator import retry
2017-05-19 18:54:05 +00:00
2022-04-04 18:49:30 +00:00
from . config_cluster import *
2017-05-19 18:54:05 +00:00
HELPERS_DIR = p . dirname ( __file__ )
2020-04-17 16:01:16 +00:00
CLICKHOUSE_ROOT_DIR = p . join ( p . dirname ( __file__ ) , " ../../.. " )
2024-03-19 13:08:38 +00:00
LOCAL_DOCKER_COMPOSE_DIR = p . join ( CLICKHOUSE_ROOT_DIR , " tests/integration/compose/ " )
2021-02-12 15:51:21 +00:00
DEFAULT_ENV_NAME = " .env "
2017-05-19 18:54:05 +00:00
2019-11-14 16:00:02 +00:00
SANITIZER_SIGN = " ================== "
2023-07-06 06:16:18 +00:00
CLICKHOUSE_START_COMMAND = (
" clickhouse server --config-file=/etc/clickhouse-server/ {main_config_file} "
)
CLICKHOUSE_LOG_FILE = " /var/log/clickhouse-server/clickhouse-server.log "
CLICKHOUSE_ERROR_LOG_FILE = " /var/log/clickhouse-server/clickhouse-server.err.log "
2023-03-24 10:54:38 +00:00
2024-02-26 12:45:20 +00:00
# Minimum version we use in integration tests to check compatibility with old releases
# Keep in mind that we only support upgrading between releases that are at most 1 year different.
# This means that this minimum need to be, at least, 1 year older than the current release
2024-07-03 21:51:40 +00:00
CLICKHOUSE_CI_MIN_TESTED_VERSION = " 23.3 "
2024-02-26 12:45:20 +00:00
2023-03-24 10:54:38 +00:00
2021-02-19 12:58:11 +00:00
# to create docker-compose env file
2021-02-12 15:51:21 +00:00
def _create_env_file ( path , variables ) :
2021-02-24 11:46:58 +00:00
logging . debug ( f " Env { variables } stored in { path } " )
2021-02-12 15:51:21 +00:00
with open ( path , " w " ) as f :
2020-10-02 16:54:07 +00:00
for var , value in list ( variables . items ( ) ) :
2018-07-28 14:38:08 +00:00
f . write ( " = " . join ( [ var , value ] ) + " \n " )
2021-02-12 15:51:21 +00:00
return path
2022-03-22 16:39:58 +00:00
2021-06-04 10:14:32 +00:00
def run_and_check (
args ,
env = None ,
shell = False ,
stdout = subprocess . PIPE ,
stderr = subprocess . PIPE ,
timeout = 300 ,
nothrow = False ,
detach = False ,
) :
2021-06-01 09:53:36 +00:00
if detach :
2021-06-02 15:08:16 +00:00
subprocess . Popen (
args ,
stdout = subprocess . DEVNULL ,
stderr = subprocess . DEVNULL ,
env = env ,
shell = shell ,
)
2021-06-01 09:53:36 +00:00
return
2021-07-01 14:41:59 +00:00
logging . debug ( f " Command: { args } " )
2021-05-19 05:54:48 +00:00
res = subprocess . run (
args , stdout = stdout , stderr = stderr , env = env , shell = shell , timeout = timeout
)
2024-02-21 16:16:34 +00:00
out = res . stdout . decode ( " utf-8 " , " ignore " )
err = res . stderr . decode ( " utf-8 " , " ignore " )
2021-07-01 14:41:59 +00:00
# check_call(...) from subprocess does not print stderr, so we do it manually
2021-07-26 16:32:13 +00:00
for outline in out . splitlines ( ) :
logging . debug ( f " Stdout: { outline } " )
for errline in err . splitlines ( ) :
logging . debug ( f " Stderr: { errline } " )
2021-01-22 14:27:23 +00:00
if res . returncode != 0 :
2021-07-01 14:41:59 +00:00
logging . debug ( f " Exitcode: { res . returncode } " )
if env :
logging . debug ( f " Env: { env } " )
2021-06-01 09:53:36 +00:00
if not nothrow :
raise Exception (
f " Command { args } return non-zero code { res . returncode } : { res . stderr . decode ( ' utf-8 ' ) } "
)
2021-07-01 14:41:59 +00:00
return out
2021-01-22 14:27:23 +00:00
2022-03-22 16:39:58 +00:00
2023-04-04 14:56:42 +00:00
# Based on https://stackoverflow.com/a/1365284/3706827
2021-05-11 14:27:38 +00:00
def get_free_port ( ) :
2023-04-04 14:56:42 +00:00
with socket . socket ( ) as s :
s . bind ( ( " " , 0 ) )
return s . getsockname ( ) [ 1 ]
2021-01-22 14:27:23 +00:00
2022-03-22 16:39:58 +00:00
2024-08-02 11:32:59 +00:00
def is_port_free ( port : int ) - > bool :
try :
with socket . socket ( socket . AF_INET , socket . SOCK_STREAM ) as s :
s . bind ( ( " " , port ) )
return True
except socket . error :
return False
class PortPoolManager :
"""
This class is used for distribution of ports allocated to single pytest - xdist worker
It can be used by multiple ClickHouseCluster instances
"""
# Shared between instances
all_ports = None
free_ports = None
def __init__ ( self ) :
self . used_ports = [ ]
if self . all_ports is None :
worker_ports = os . getenv ( " WORKER_FREE_PORTS " )
ports = [ int ( p ) for p in worker_ports . split ( " " ) ]
# Static vars
PortPoolManager . all_ports = ports
PortPoolManager . free_ports = ports
def get_port ( self ) :
for port in self . free_ports :
if is_port_free ( port ) :
self . free_ports . remove ( port )
self . used_ports . append ( port )
return port
raise Exception (
f " No free ports: { self . all_ports } " ,
)
def return_used_ports ( self ) :
self . free_ports . extend ( self . used_ports )
self . used_ports . clear ( )
2021-04-05 04:48:45 +00:00
def retry_exception ( num , delay , func , exception = Exception , * args , * * kwargs ) :
"""
Retry if ` func ( ) ` throws , ` num ` times .
: param func : func to run
: param num : number of retries
: throws StopIteration
"""
i = 0
while i < = num :
try :
func ( * args , * * kwargs )
time . sleep ( delay )
except exception : # pylint: disable=broad-except
i + = 1
continue
return
raise StopIteration ( " Function did not finished successfully " )
2022-03-22 16:39:58 +00:00
2021-04-05 04:48:45 +00:00
2021-06-01 09:53:36 +00:00
def subprocess_check_call ( args , detach = False , nothrow = False ) :
2018-09-28 14:53:20 +00:00
# Uncomment for debugging
2021-06-01 09:53:36 +00:00
# logging.info('run:' + ' '.join(args))
2021-06-01 14:18:35 +00:00
return run_and_check ( args , detach = detach , nothrow = nothrow )
2018-09-28 14:53:20 +00:00
2021-07-27 15:54:13 +00:00
2019-01-29 17:17:31 +00:00
def get_odbc_bridge_path ( ) :
path = os . environ . get ( " CLICKHOUSE_TESTS_ODBC_BRIDGE_BIN_PATH " )
if path is None :
server_path = os . environ . get ( " CLICKHOUSE_TESTS_SERVER_BIN_PATH " )
if server_path is not None :
return os . path . join ( os . path . dirname ( server_path ) , " clickhouse-odbc-bridge " )
else :
return " /usr/bin/clickhouse-odbc-bridge "
return path
2022-03-22 16:39:58 +00:00
2021-03-11 17:48:47 +00:00
def get_library_bridge_path ( ) :
path = os . environ . get ( " CLICKHOUSE_TESTS_LIBRARY_BRIDGE_BIN_PATH " )
if path is None :
server_path = os . environ . get ( " CLICKHOUSE_TESTS_SERVER_BIN_PATH " )
if server_path is not None :
return os . path . join (
os . path . dirname ( server_path ) , " clickhouse-library-bridge "
)
else :
return " /usr/bin/clickhouse-library-bridge "
return path
2020-09-16 04:26:10 +00:00
2022-03-22 16:39:58 +00:00
2020-07-06 13:45:54 +00:00
def get_docker_compose_path ( ) :
2024-03-19 13:08:38 +00:00
return LOCAL_DOCKER_COMPOSE_DIR
2020-07-06 13:45:54 +00:00
2022-03-22 16:39:58 +00:00
2021-02-24 11:46:58 +00:00
def check_kafka_is_available ( kafka_id , kafka_port ) :
p = subprocess . Popen (
2022-03-22 16:39:58 +00:00
(
2021-02-24 11:46:58 +00:00
" docker " ,
" exec " ,
" -i " ,
kafka_id ,
" /usr/bin/kafka-broker-api-versions " ,
" --bootstrap-server " ,
f " INSIDE://localhost: { kafka_port } " ,
) ,
stdout = subprocess . PIPE ,
stderr = subprocess . PIPE ,
)
p . communicate ( )
return p . returncode == 0
2022-12-19 11:28:49 +00:00
2022-12-15 11:50:23 +00:00
def check_kerberos_kdc_is_available ( kerberos_kdc_id ) :
p = subprocess . Popen (
(
" docker " ,
" exec " ,
" -i " ,
kerberos_kdc_id ,
" /etc/rc.d/init.d/krb5kdc " ,
" status " ,
) ,
stdout = subprocess . PIPE ,
stderr = subprocess . PIPE ,
)
p . communicate ( )
return p . returncode == 0
2022-03-22 16:39:58 +00:00
2023-01-03 19:30:14 +00:00
def check_postgresql_java_client_is_available ( postgresql_java_client_id ) :
p = subprocess . Popen (
( " docker " , " exec " , " -i " , postgresql_java_client_id , " java " , " -version " ) ,
stdout = subprocess . PIPE ,
)
p . communicate ( )
return p . returncode == 0
2022-03-22 16:39:58 +00:00
2023-01-03 21:28:54 +00:00
2023-10-09 18:33:03 +00:00
def check_rabbitmq_is_available ( rabbitmq_id , cookie ) :
2021-04-30 09:18:12 +00:00
p = subprocess . Popen (
2023-10-09 19:10:41 +00:00
(
" docker " ,
" exec " ,
" -e " ,
f " RABBITMQ_ERLANG_COOKIE= { cookie } " ,
" -i " ,
rabbitmq_id ,
" rabbitmqctl " ,
" await_startup " ,
) ,
2021-04-30 09:18:12 +00:00
stdout = subprocess . PIPE ,
)
2024-08-01 11:38:41 +00:00
p . wait ( timeout = 60 )
2021-04-30 09:18:12 +00:00
return p . returncode == 0
2022-03-22 16:39:58 +00:00
2023-10-09 18:33:03 +00:00
def rabbitmq_debuginfo ( rabbitmq_id , cookie ) :
p = subprocess . Popen (
2023-10-09 19:10:41 +00:00
(
" docker " ,
" exec " ,
" -e " ,
f " RABBITMQ_ERLANG_COOKIE= { cookie } " ,
" -i " ,
rabbitmq_id ,
" rabbitmq-diagnostics " ,
" status " ,
) ,
2023-10-09 18:33:03 +00:00
stdout = subprocess . PIPE ,
)
p . communicate ( )
p = subprocess . Popen (
2023-10-09 19:10:41 +00:00
(
" docker " ,
" exec " ,
" -e " ,
f " RABBITMQ_ERLANG_COOKIE= { cookie } " ,
" -i " ,
rabbitmq_id ,
" rabbitmq-diagnostics " ,
" listeners " ,
) ,
2023-10-09 18:33:03 +00:00
stdout = subprocess . PIPE ,
)
p . communicate ( )
p = subprocess . Popen (
2023-10-09 19:10:41 +00:00
(
" docker " ,
" exec " ,
" -e " ,
f " RABBITMQ_ERLANG_COOKIE= { cookie } " ,
" -i " ,
rabbitmq_id ,
" rabbitmq-diagnostics " ,
" environment " ,
) ,
2023-10-09 18:33:03 +00:00
stdout = subprocess . PIPE ,
)
p . communicate ( )
2022-03-22 16:39:58 +00:00
2023-10-09 19:10:41 +00:00
2022-07-24 14:13:17 +00:00
async def check_nats_is_available ( nats_port , ssl_ctx = None ) :
nc = await nats_connect_ssl (
nats_port , user = " click " , password = " house " , ssl_ctx = ssl_ctx
)
available = nc . is_connected
await nc . close ( )
return available
async def nats_connect_ssl ( nats_port , user , password , ssl_ctx = None ) :
if not ssl_ctx :
ssl_ctx = ssl . create_default_context ( )
ssl_ctx . check_hostname = False
ssl_ctx . verify_mode = ssl . CERT_NONE
nc = await nats . connect (
" tls://localhost: {} " . format ( nats_port ) ,
user = user ,
password = password ,
tls = ssl_ctx ,
)
return nc
2022-05-17 13:58:09 +00:00
2023-10-09 18:33:03 +00:00
def enable_consistent_hash_plugin ( rabbitmq_id , cookie ) :
2021-04-30 09:18:12 +00:00
p = subprocess . Popen (
2022-03-22 16:39:58 +00:00
(
2021-04-30 09:18:12 +00:00
" docker " ,
" exec " ,
2023-10-09 18:33:03 +00:00
" -e " ,
f " RABBITMQ_ERLANG_COOKIE= { cookie } " ,
2021-04-30 09:18:12 +00:00
" -i " ,
rabbitmq_id ,
" rabbitmq-plugins " ,
" enable " ,
" rabbitmq_consistent_hash_exchange " ,
) ,
stdout = subprocess . PIPE ,
)
p . communicate ( )
return p . returncode == 0
2022-03-22 16:39:58 +00:00
2022-07-07 20:19:15 +00:00
def get_instances_dir ( name ) :
instances_dir_name = " _instances "
run_id = os . environ . get ( " INTEGRATION_TESTS_RUN_ID " , " " )
if name :
instances_dir_name + = " _ " + name
if run_id :
instances_dir_name + = " _ " + shlex . quote ( run_id )
return instances_dir_name
2021-06-21 08:02:27 +00:00
2022-07-18 12:25:14 +00:00
def extract_test_name ( base_path ) :
""" Extracts the name of the test based to a path to its test*.py file
Must be unique in each test directory ( because it ' s used to make instances dir and to stop docker containers from previous run)
"""
name = p . basename ( base_path )
2022-07-18 15:19:02 +00:00
if name == " test.py " :
name = " "
elif name . startswith ( " test_ " ) and name . endswith ( " .py " ) :
2022-07-18 18:49:30 +00:00
name = name [ len ( " test_ " ) : ( len ( name ) - len ( " .py " ) ) ]
2022-07-18 12:25:14 +00:00
return name
2017-05-19 18:54:05 +00:00
class ClickHouseCluster :
""" ClickHouse cluster with several instances and (possibly) ZooKeeper.
Add instances with several calls to add_instance ( ) , then start them with the start ( ) call .
Directories for instances are created in the directory of base_path . After cluster is started ,
these directories will contain logs , database files , docker - compose config , ClickHouse configs etc .
"""
2020-08-14 15:51:28 +00:00
def __init__ (
self ,
base_path ,
name = None ,
base_config_dir = None ,
server_bin_path = None ,
client_bin_path = None ,
2021-05-21 13:29:43 +00:00
odbc_bridge_bin_path = None ,
library_bridge_bin_path = None ,
zookeeper_config_path = None ,
2023-12-01 17:07:45 +00:00
keeper_config_dir = None ,
2021-05-21 13:29:43 +00:00
custom_dockerd_host = None ,
zookeeper_keyfile = None ,
zookeeper_certfile = None ,
2023-04-13 13:10:49 +00:00
with_spark = False ,
2021-05-21 13:29:43 +00:00
) :
2020-10-02 16:54:07 +00:00
for param in list ( os . environ . keys ( ) ) :
2021-02-18 21:21:50 +00:00
logging . debug ( " ENV %40s %s " % ( param , os . environ [ param ] ) )
2021-07-01 14:41:59 +00:00
self . base_path = base_path
2017-05-19 18:54:05 +00:00
self . base_dir = p . dirname ( base_path )
2022-07-18 12:25:14 +00:00
self . name = name if name is not None else extract_test_name ( base_path )
2017-05-19 18:54:05 +00:00
2020-08-12 08:55:04 +00:00
self . base_config_dir = base_config_dir or os . environ . get (
" CLICKHOUSE_TESTS_BASE_CONFIG_DIR " , " /etc/clickhouse-server/ "
2020-09-16 04:26:10 +00:00
)
2019-11-20 11:56:38 +00:00
self . server_bin_path = p . realpath (
server_bin_path
or os . environ . get ( " CLICKHOUSE_TESTS_SERVER_BIN_PATH " , " /usr/bin/clickhouse " )
2022-03-22 16:39:58 +00:00
)
2019-01-29 17:17:31 +00:00
self . odbc_bridge_bin_path = p . realpath (
odbc_bridge_bin_path or get_odbc_bridge_path ( )
2022-03-22 16:39:58 +00:00
)
2021-03-12 14:07:20 +00:00
self . library_bridge_bin_path = p . realpath (
library_bridge_bin_path or get_library_bridge_path ( )
2022-03-22 16:39:58 +00:00
)
2019-11-20 11:56:38 +00:00
self . client_bin_path = p . realpath (
client_bin_path
or os . environ . get (
" CLICKHOUSE_TESTS_CLIENT_BIN_PATH " , " /usr/bin/clickhouse-client "
)
2022-03-22 16:39:58 +00:00
)
2019-11-20 11:56:38 +00:00
self . zookeeper_config_path = (
p . join ( self . base_dir , zookeeper_config_path )
if zookeeper_config_path
else p . join ( HELPERS_DIR , " zookeeper_config.xml " )
2022-03-22 16:39:58 +00:00
)
2023-12-01 19:39:38 +00:00
self . keeper_config_dir = (
p . join ( self . base_dir , keeper_config_dir )
if keeper_config_dir
else HELPERS_DIR
)
2023-12-01 17:07:45 +00:00
2021-01-27 16:21:54 +00:00
project_name = (
pwd . getpwuid ( os . getuid ( ) ) . pw_name + p . basename ( self . base_dir ) + self . name
2022-03-22 16:39:58 +00:00
)
2017-05-19 18:54:05 +00:00
# docker-compose removes everything non-alphanumeric from project names so we do it too.
2021-01-27 16:21:54 +00:00
self . project_name = re . sub ( r " [^a-z0-9] " , " " , project_name . lower ( ) )
2022-07-07 20:19:15 +00:00
self . instances_dir_name = get_instances_dir ( self . name )
2024-07-26 08:42:52 +00:00
xdist_worker = os . getenv ( " PYTEST_XDIST_WORKER " )
if xdist_worker :
self . project_name + = f " _ { xdist_worker } "
self . instances_dir_name + = f " _ { xdist_worker } "
2022-07-07 20:19:15 +00:00
self . instances_dir = p . join ( self . base_dir , self . instances_dir_name )
2019-11-14 16:00:02 +00:00
self . docker_logs_path = p . join ( self . instances_dir , " docker.log " )
2021-02-16 07:10:01 +00:00
self . env_file = p . join ( self . instances_dir , DEFAULT_ENV_NAME )
self . env_variables = { }
2023-04-10 13:06:40 +00:00
# Problems with glibc 2.36+ [1]
#
# [1]: https://github.com/ClickHouse/ClickHouse/issues/43426#issuecomment-1368512678
2023-02-13 12:42:21 +00:00
self . env_variables [ " ASAN_OPTIONS " ] = " use_sigaltstack=0 "
2023-04-19 15:35:57 +00:00
self . env_variables [ " TSAN_OPTIONS " ] = " use_sigaltstack=0 "
2021-10-26 14:17:51 +00:00
self . env_variables [ " CLICKHOUSE_WATCHDOG_ENABLE " ] = " 0 "
2022-07-24 14:13:17 +00:00
self . env_variables [ " CLICKHOUSE_NATS_TLS_SECURE " ] = " 0 "
2021-02-17 15:40:01 +00:00
self . up_called = False
2017-05-19 18:54:05 +00:00
2018-08-23 15:38:25 +00:00
custom_dockerd_host = custom_dockerd_host or os . environ . get (
" CLICKHOUSE_TESTS_DOCKERD_HOST "
)
2018-08-24 11:19:06 +00:00
self . docker_api_version = os . environ . get ( " DOCKER_API_VERSION " )
2020-09-03 13:03:26 +00:00
self . docker_base_tag = os . environ . get ( " DOCKER_BASE_TAG " , " latest " )
2018-08-23 15:31:20 +00:00
self . base_cmd = [ " docker-compose " ]
if custom_dockerd_host :
self . base_cmd + = [ " --host " , custom_dockerd_host ]
2021-02-16 07:10:01 +00:00
self . base_cmd + = [ " --env-file " , self . env_file ]
2021-01-27 16:21:54 +00:00
self . base_cmd + = [ " --project-name " , self . project_name ]
2018-08-23 15:31:20 +00:00
2017-08-30 16:25:34 +00:00
self . base_zookeeper_cmd = None
2024-02-13 12:15:46 +00:00
self . base_mysql57_cmd = [ ]
2018-07-18 05:22:01 +00:00
self . base_kafka_cmd = [ ]
2020-09-29 08:56:37 +00:00
self . base_kerberized_kafka_cmd = [ ]
2022-12-19 11:28:49 +00:00
self . base_kerberos_kdc_cmd = [ ]
2020-05-20 06:22:12 +00:00
self . base_rabbitmq_cmd = [ ]
2022-05-17 13:58:09 +00:00
self . base_nats_cmd = [ ]
2020-05-19 02:21:27 +00:00
self . base_cassandra_cmd = [ ]
2021-06-07 12:56:29 +00:00
self . base_jdbc_bridge_cmd = [ ]
2021-02-18 21:21:50 +00:00
self . base_redis_cmd = [ ]
2018-01-09 19:12:43 +00:00
self . pre_zookeeper_commands = [ ]
2017-05-19 18:54:05 +00:00
self . instances = { }
self . with_zookeeper = False
2021-05-21 13:29:43 +00:00
self . with_zookeeper_secure = False
2021-04-13 14:55:31 +00:00
self . with_mysql_client = False
2024-02-13 12:15:46 +00:00
self . with_mysql57 = False
2021-02-16 07:10:01 +00:00
self . with_mysql8 = False
2021-03-31 17:04:57 +00:00
self . with_mysql_cluster = False
2018-10-15 14:49:23 +00:00
self . with_postgres = False
2021-03-31 17:04:57 +00:00
self . with_postgres_cluster = False
2023-01-03 19:30:14 +00:00
self . with_postgresql_java_client = False
2018-07-18 05:22:01 +00:00
self . with_kafka = False
2020-09-29 08:56:37 +00:00
self . with_kerberized_kafka = False
2022-12-14 13:39:23 +00:00
self . with_kerberos_kdc = False
2020-05-20 06:22:12 +00:00
self . with_rabbitmq = False
2022-05-17 13:58:09 +00:00
self . with_nats = False
2018-08-22 15:42:27 +00:00
self . with_odbc_drivers = False
2018-12-05 13:24:45 +00:00
self . with_hdfs = False
2020-09-10 10:02:46 +00:00
self . with_kerberized_hdfs = False
2019-02-25 10:45:22 +00:00
self . with_mongo = False
2021-07-27 15:54:13 +00:00
self . with_mongo_secure = False
2019-06-05 09:23:41 +00:00
self . with_net_trics = False
2019-03-21 18:10:55 +00:00
self . with_redis = False
2020-05-19 02:21:27 +00:00
self . with_cassandra = False
2023-09-29 07:16:14 +00:00
self . with_ldap = False
2021-06-07 12:56:29 +00:00
self . with_jdbc_bridge = False
2021-06-13 12:56:22 +00:00
self . with_nginx = False
2021-11-19 01:58:34 +00:00
self . with_hive = False
2022-07-20 17:09:38 +00:00
self . with_coredns = False
2018-07-28 14:38:08 +00:00
2023-01-03 19:30:14 +00:00
# available when with_minio == True
2019-11-20 11:56:38 +00:00
self . with_minio = False
2021-02-19 12:58:11 +00:00
self . minio_dir = os . path . join ( self . instances_dir , " minio " )
2021-06-13 12:56:22 +00:00
self . minio_certs_dir = None # source for certificates
2023-01-03 19:30:14 +00:00
self . minio_data_dir = p . join ( self . minio_dir , " data " )
2019-11-20 11:56:38 +00:00
self . minio_host = " minio1 "
2021-04-14 11:21:40 +00:00
self . minio_ip = None
2019-11-20 11:56:38 +00:00
self . minio_bucket = " root "
2021-01-12 17:18:40 +00:00
self . minio_bucket_2 = " root2 "
2019-11-20 11:56:38 +00:00
self . minio_port = 9001
self . minio_client = None # type: Minio
2020-07-10 19:42:18 +00:00
self . minio_redirect_host = " proxy1 "
2021-04-29 11:57:48 +00:00
self . minio_redirect_ip = None
2020-07-10 19:42:18 +00:00
self . minio_redirect_port = 8080
2024-05-21 11:21:53 +00:00
self . minio_docker_id = self . get_instance_docker_id ( self . minio_host )
2019-11-20 11:56:38 +00:00
2023-04-13 09:12:24 +00:00
self . spark_session = None
2023-04-11 15:23:05 +00:00
2021-11-26 14:04:53 +00:00
self . with_azurite = False
2024-08-15 09:41:03 +00:00
self . azurite_container = " azurite-container "
2024-08-14 14:15:05 +00:00
self . blob_service_client = None
2023-10-18 20:20:39 +00:00
self . _azurite_port = 0
2021-11-26 14:04:53 +00:00
2021-02-18 21:21:50 +00:00
# available when with_hdfs == True
self . hdfs_host = " hdfs1 "
2021-05-27 04:24:16 +00:00
self . hdfs_ip = None
2021-06-09 09:23:02 +00:00
self . hdfs_name_port = 50070
self . hdfs_data_port = 50075
2021-02-18 21:21:50 +00:00
self . hdfs_dir = p . abspath ( p . join ( self . instances_dir , " hdfs " ) )
self . hdfs_logs_dir = os . path . join ( self . hdfs_dir , " logs " )
2021-06-09 09:23:02 +00:00
self . hdfs_api = None # also for kerberized hdfs
2021-02-18 21:21:50 +00:00
2021-02-19 14:42:43 +00:00
# available when with_kerberized_hdfs == True
self . hdfs_kerberized_host = " kerberizedhdfs1 "
2021-06-09 09:23:02 +00:00
self . hdfs_kerberized_ip = None
self . hdfs_kerberized_name_port = 50070
self . hdfs_kerberized_data_port = 1006
2021-02-19 14:42:43 +00:00
self . hdfs_kerberized_dir = p . abspath (
p . join ( self . instances_dir , " kerberized_hdfs " )
2022-03-22 16:39:58 +00:00
)
2021-02-19 14:42:43 +00:00
self . hdfs_kerberized_logs_dir = os . path . join ( self . hdfs_kerberized_dir , " logs " )
2020-02-03 00:02:19 +00:00
# available when with_kafka == True
2021-02-18 12:57:45 +00:00
self . kafka_host = " kafka1 "
2023-01-17 16:25:05 +00:00
self . kafka_dir = os . path . join ( self . instances_dir , " kafka " )
2023-04-04 15:18:24 +00:00
self . _kafka_port = 0
2021-02-18 12:57:45 +00:00
self . kafka_docker_id = None
2020-02-03 00:02:19 +00:00
self . schema_registry_host = " schema-registry "
2023-04-04 15:18:24 +00:00
self . _schema_registry_port = 0
2023-05-02 12:31:00 +00:00
self . schema_registry_auth_host = " schema-registry-auth "
self . _schema_registry_auth_port = 0
2021-02-18 12:57:45 +00:00
self . kafka_docker_id = self . get_instance_docker_id ( self . kafka_host )
2022-07-20 17:09:38 +00:00
self . coredns_host = " coredns "
2021-02-18 12:57:45 +00:00
# available when with_kerberozed_kafka == True
2023-01-17 16:25:05 +00:00
# reuses kafka_dir
2021-02-18 12:57:45 +00:00
self . kerberized_kafka_host = " kerberized_kafka1 "
2023-04-04 15:18:24 +00:00
self . _kerberized_kafka_port = 0
2021-02-18 21:21:50 +00:00
self . kerberized_kafka_docker_id = self . get_instance_docker_id (
self . kerberized_kafka_host
)
2020-02-03 00:02:19 +00:00
2022-12-14 13:39:23 +00:00
# available when with_kerberos_kdc == True
self . kerberos_kdc_host = " kerberoskdc "
2022-12-19 11:28:49 +00:00
self . keberos_kdc_docker_id = self . get_instance_docker_id ( self . kerberos_kdc_host )
2022-12-14 13:39:23 +00:00
2021-02-12 15:51:21 +00:00
# available when with_mongo == True
self . mongo_host = " mongo1 "
2023-04-04 15:18:24 +00:00
self . _mongo_port = 0
2022-01-14 13:00:38 +00:00
self . mongo_no_cred_host = " mongo2 "
2023-04-04 15:18:24 +00:00
self . _mongo_no_cred_port = 0
2021-02-12 15:51:21 +00:00
2021-02-15 09:35:45 +00:00
# available when with_cassandra == True
self . cassandra_host = " cassandra1 "
2021-04-07 12:22:53 +00:00
self . cassandra_port = 9042
self . cassandra_ip = None
self . cassandra_id = self . get_instance_docker_id ( self . cassandra_host )
2021-02-15 09:35:45 +00:00
2023-09-29 07:16:14 +00:00
# available when with_ldap == True
self . ldap_host = " openldap "
self . ldap_container = None
self . ldap_port = 1389
self . ldap_id = self . get_instance_docker_id ( self . ldap_host )
2021-02-16 14:16:15 +00:00
# available when with_rabbitmq == True
self . rabbitmq_host = " rabbitmq1 "
2021-04-30 09:18:12 +00:00
self . rabbitmq_ip = None
self . rabbitmq_port = 5672
2023-11-16 12:09:13 +00:00
self . rabbitmq_secure_port = 5671
2021-05-17 11:16:16 +00:00
self . rabbitmq_dir = p . abspath ( p . join ( self . instances_dir , " rabbitmq " ) )
2023-10-09 18:33:03 +00:00
self . rabbitmq_cookie_file = os . path . join ( self . rabbitmq_dir , " erlang.cookie " )
2021-05-17 11:16:16 +00:00
self . rabbitmq_logs_dir = os . path . join ( self . rabbitmq_dir , " logs " )
2023-10-09 18:33:03 +00:00
self . rabbitmq_cookie = self . get_instance_docker_id ( self . rabbitmq_host )
2021-05-17 11:16:16 +00:00
2022-05-17 13:58:09 +00:00
self . nats_host = " nats1 "
self . nats_port = 4444
self . nats_docker_id = None
2022-07-24 14:13:17 +00:00
self . nats_dir = p . abspath ( p . join ( self . instances_dir , " nats " ) )
self . nats_cert_dir = os . path . join ( self . nats_dir , " cert " )
self . nats_ssl_context = None
2022-05-17 13:58:09 +00:00
2021-06-13 14:02:08 +00:00
# available when with_nginx == True
2021-06-13 12:56:22 +00:00
self . nginx_host = " nginx "
self . nginx_ip = None
self . nginx_port = 80
self . nginx_id = self . get_instance_docker_id ( self . nginx_host )
2021-02-16 14:16:15 +00:00
# available when with_redis == True
self . redis_host = " redis1 "
2023-04-04 15:18:24 +00:00
self . _redis_port = 0
2021-02-16 14:16:15 +00:00
2021-02-17 15:40:01 +00:00
# available when with_postgres == True
self . postgres_host = " postgres1 "
2021-03-19 12:33:14 +00:00
self . postgres_ip = None
2021-07-12 08:32:20 +00:00
self . postgres_conn = None
2021-03-19 12:33:14 +00:00
self . postgres2_host = " postgres2 "
self . postgres2_ip = None
2021-07-12 08:32:20 +00:00
self . postgres2_conn = None
2021-04-08 09:30:24 +00:00
self . postgres3_host = " postgres3 "
self . postgres3_ip = None
2021-07-12 08:32:20 +00:00
self . postgres3_conn = None
2021-04-08 09:30:24 +00:00
self . postgres4_host = " postgres4 "
self . postgres4_ip = None
2021-07-12 08:32:20 +00:00
self . postgres4_conn = None
2021-03-19 12:33:14 +00:00
self . postgres_port = 5432
self . postgres_dir = p . abspath ( p . join ( self . instances_dir , " postgres " ) )
self . postgres_logs_dir = os . path . join ( self . postgres_dir , " postgres1 " )
self . postgres2_logs_dir = os . path . join ( self . postgres_dir , " postgres2 " )
2021-04-08 09:30:24 +00:00
self . postgres3_logs_dir = os . path . join ( self . postgres_dir , " postgres3 " )
self . postgres4_logs_dir = os . path . join ( self . postgres_dir , " postgres4 " )
2023-01-03 19:30:14 +00:00
self . postgres_id = self . get_instance_docker_id ( self . postgres_host )
# available when with_postgresql_java_client = True
self . postgresql_java_client_host = " java "
2023-01-03 21:28:54 +00:00
self . postgresql_java_client_docker_id = self . get_instance_docker_id (
self . postgresql_java_client_host
)
2021-02-17 15:40:01 +00:00
2021-04-13 14:55:31 +00:00
# available when with_mysql_client == True
self . mysql_client_host = " mysql_client "
self . mysql_client_container = None
2021-06-13 12:56:22 +00:00
2024-02-13 12:15:46 +00:00
# available when with_mysql57 == True
self . mysql57_host = " mysql57 "
self . mysql57_port = 3306
self . mysql57_ip = None
self . mysql57_dir = p . abspath ( p . join ( self . instances_dir , " mysql " ) )
self . mysql57_logs_dir = os . path . join ( self . mysql57_dir , " logs " )
# available when with_mysql8 == True
self . mysql8_host = " mysql80 "
self . mysql8_port = 3306
self . mysql8_ip = None
self . mysql8_dir = p . abspath ( p . join ( self . instances_dir , " mysql8 " ) )
self . mysql8_logs_dir = os . path . join ( self . mysql8_dir , " logs " )
2021-02-15 09:35:45 +00:00
2021-04-08 09:30:24 +00:00
# available when with_mysql_cluster == True
self . mysql2_host = " mysql2 "
self . mysql3_host = " mysql3 "
self . mysql4_host = " mysql4 "
self . mysql2_ip = None
self . mysql3_ip = None
self . mysql4_ip = None
self . mysql_cluster_dir = p . abspath ( p . join ( self . instances_dir , " mysql " ) )
2024-02-13 12:15:46 +00:00
self . mysql_cluster_logs_dir = os . path . join ( self . mysql8_dir , " logs " )
2021-02-16 07:10:01 +00:00
2021-05-21 13:29:43 +00:00
# available when with_zookeper_secure == True
self . zookeeper_secure_port = 2281
self . zookeeper_keyfile = zookeeper_keyfile
self . zookeeper_certfile = zookeeper_certfile
2021-05-20 16:12:04 +00:00
# available when with_zookeper == True
2021-05-16 14:18:21 +00:00
self . use_keeper = True
2021-05-21 13:29:43 +00:00
self . zookeeper_port = 2181
2021-05-20 16:12:04 +00:00
self . keeper_instance_dir_prefix = p . join (
p . abspath ( self . instances_dir ) , " keeper "
) # if use_keeper = True
self . zookeeper_instance_dir_prefix = p . join ( self . instances_dir , " zk " )
self . zookeeper_dirs_to_create = [ ]
2020-05-19 15:27:10 +00:00
2021-07-27 07:33:49 +00:00
# available when with_jdbc_bridge == True
self . jdbc_bridge_host = " bridge1 "
self . jdbc_bridge_ip = None
self . jdbc_bridge_port = 9019
self . jdbc_driver_dir = p . abspath ( p . join ( self . instances_dir , " jdbc_driver " ) )
self . jdbc_driver_logs_dir = os . path . join ( self . jdbc_driver_dir , " logs " )
2024-03-11 22:58:05 +00:00
# available when with_prometheus == True
self . with_prometheus = False
self . prometheus_writer_host = " prometheus_writer "
self . prometheus_writer_port = 9090
self . prometheus_writer_logs_dir = p . abspath (
p . join ( self . instances_dir , " prometheus_writer/logs " )
)
self . prometheus_reader_host = " prometheus_reader "
self . prometheus_reader_port = 9091
self . prometheus_reader_logs_dir = p . abspath (
p . join ( self . instances_dir , " prometheus_reader/logs " )
)
self . prometheus_remote_write_handler_host = None
self . prometheus_remote_write_handler_port = 9092
self . prometheus_remote_write_handler_path = " /write "
self . prometheus_remote_read_handler_host = None
self . prometheus_remote_read_handler_port = 9092
self . prometheus_remote_read_handler_path = " /read "
2017-05-23 17:13:36 +00:00
self . docker_client = None
2017-05-19 18:54:05 +00:00
self . is_up = False
2021-05-14 12:30:49 +00:00
self . env = os . environ . copy ( )
2021-02-24 11:46:58 +00:00
logging . debug ( f " CLUSTER INIT base_config_dir: { self . base_config_dir } " )
2022-07-07 20:19:15 +00:00
if p . exists ( self . instances_dir ) :
shutil . rmtree ( self . instances_dir , ignore_errors = True )
logging . debug ( f " Removed : { self . instances_dir } " )
2023-04-13 13:10:49 +00:00
if with_spark :
2023-10-15 16:02:34 +00:00
import pyspark
2023-10-16 11:03:12 +00:00
2023-04-13 13:10:49 +00:00
# if you change packages, don't forget to update them in docker/test/integration/runner/dockerd-entrypoint.sh
(
pyspark . sql . SparkSession . builder . appName ( " spark_test " )
2023-06-29 10:24:19 +00:00
# The jars are now linked to "$SPARK_HOME/jars" and we don't
# need packages to be downloaded once and once again
# .config(
# "spark.jars.packages",
# "org.apache.hudi:hudi-spark3.3-bundle_2.12:0.13.0,io.delta:delta-core_2.12:2.2.0,org.apache.iceberg:iceberg-spark-runtime-3.3_2.12:1.1.0",
# )
2023-04-13 13:10:49 +00:00
. master ( " local " )
. getOrCreate ( )
. stop ( )
)
2024-08-02 11:32:59 +00:00
self . port_pool = PortPoolManager ( )
2023-04-04 15:18:24 +00:00
@property
def kafka_port ( self ) :
if self . _kafka_port :
return self . _kafka_port
2024-08-02 11:32:59 +00:00
self . _kafka_port = self . port_pool . get_port ( )
2023-04-04 15:18:24 +00:00
return self . _kafka_port
@property
def schema_registry_port ( self ) :
if self . _schema_registry_port :
return self . _schema_registry_port
2024-08-02 11:32:59 +00:00
self . _schema_registry_port = self . port_pool . get_port ( )
2023-04-04 15:18:24 +00:00
return self . _schema_registry_port
2023-05-02 12:31:00 +00:00
@property
def schema_registry_auth_port ( self ) :
if self . _schema_registry_auth_port :
return self . _schema_registry_auth_port
2024-08-02 11:32:59 +00:00
self . _schema_registry_auth_port = self . port_pool . get_port ( )
2023-05-02 12:31:00 +00:00
return self . _schema_registry_auth_port
2023-04-04 15:18:24 +00:00
@property
def kerberized_kafka_port ( self ) :
if self . _kerberized_kafka_port :
return self . _kerberized_kafka_port
2024-08-02 11:32:59 +00:00
self . _kerberized_kafka_port = self . port_pool . get_port ( )
2023-04-04 15:18:24 +00:00
return self . _kerberized_kafka_port
2023-10-18 20:20:39 +00:00
@property
def azurite_port ( self ) :
if self . _azurite_port :
return self . _azurite_port
2024-08-02 11:32:59 +00:00
self . _azurite_port = self . port_pool . get_port ( )
2023-10-18 20:20:39 +00:00
return self . _azurite_port
2023-04-04 15:18:24 +00:00
@property
def mongo_port ( self ) :
if self . _mongo_port :
return self . _mongo_port
2024-08-02 11:32:59 +00:00
self . _mongo_port = self . port_pool . get_port ( )
2023-04-04 15:18:24 +00:00
return self . _mongo_port
@property
def mongo_no_cred_port ( self ) :
if self . _mongo_no_cred_port :
return self . _mongo_no_cred_port
2024-08-02 11:32:59 +00:00
self . _mongo_no_cred_port = self . port_pool . get_port ( )
2023-04-04 15:18:24 +00:00
return self . _mongo_no_cred_port
@property
def redis_port ( self ) :
if self . _redis_port :
return self . _redis_port
2024-08-02 11:32:59 +00:00
self . _redis_port = self . port_pool . get_port ( )
2023-04-04 15:18:24 +00:00
return self . _redis_port
2024-08-02 11:32:59 +00:00
def __exit__ ( self , exc_type , exc_val , exc_tb ) :
self . port_pool . return_used_ports ( )
2022-07-07 20:19:15 +00:00
def print_all_docker_pieces ( self ) :
res_networks = subprocess . check_output (
f " docker network ls --filter name= ' { self . project_name } * ' " ,
shell = True ,
universal_newlines = True ,
)
2022-07-07 20:42:41 +00:00
logging . debug (
f " Docker networks for project { self . project_name } are { res_networks } "
)
2022-07-07 20:19:15 +00:00
res_containers = subprocess . check_output (
f " docker container ls -a --filter name= ' { self . project_name } * ' " ,
shell = True ,
universal_newlines = True ,
)
2022-07-07 20:42:41 +00:00
logging . debug (
f " Docker containers for project { self . project_name } are { res_containers } "
)
2022-07-07 20:19:15 +00:00
res_volumes = subprocess . check_output (
f " docker volume ls --filter name= ' { self . project_name } * ' " ,
shell = True ,
universal_newlines = True ,
)
2022-07-07 20:42:41 +00:00
logging . debug (
f " Docker volumes for project { self . project_name } are { res_volumes } "
)
2017-05-19 18:54:05 +00:00
2021-04-27 09:07:21 +00:00
def cleanup ( self ) :
2022-07-07 20:42:41 +00:00
logging . debug ( " Cleanup called " )
2022-07-07 20:19:15 +00:00
self . print_all_docker_pieces ( )
2021-10-26 14:17:51 +00:00
if (
os . environ
and " DISABLE_CLEANUP " in os . environ
and os . environ [ " DISABLE_CLEANUP " ] == " 1 "
) :
logging . warning ( " Cleanup is disabled " )
return
2021-04-27 09:07:21 +00:00
# Just in case kill unstopped containers from previous launch
try :
2021-11-21 15:01:28 +00:00
unstopped_containers = self . get_running_containers ( )
2022-07-07 20:19:15 +00:00
logging . debug ( f " Unstopped containers: { unstopped_containers } " )
if len ( unstopped_containers ) :
2021-11-21 15:01:28 +00:00
logging . debug (
f " Trying to kill unstopped containers: { unstopped_containers } "
)
2021-11-11 07:41:48 +00:00
for id in unstopped_containers :
2021-07-26 16:32:13 +00:00
run_and_check ( f " docker kill { id } " , shell = True , nothrow = True )
run_and_check ( f " docker rm { id } " , shell = True , nothrow = True )
2021-11-21 15:01:28 +00:00
unstopped_containers = self . get_running_containers ( )
if unstopped_containers :
logging . debug ( f " Left unstopped containers: { unstopped_containers } " )
else :
logging . debug ( f " Unstopped containers killed. " )
2021-04-27 15:34:33 +00:00
else :
logging . debug ( f " No running containers for project: { self . project_name } " )
2022-07-07 20:19:15 +00:00
except Exception as ex :
logging . debug ( f " Got exception removing containers { str ( ex ) } " )
2021-04-27 09:07:21 +00:00
# # Just in case remove unused networks
2022-07-07 20:19:15 +00:00
try :
logging . debug ( " Trying to prune unused networks... " )
list_networks = subprocess . check_output (
f " docker network ls -q --filter name= ' { self . project_name } ' " ,
shell = True ,
universal_newlines = True ,
) . splitlines ( )
if list_networks :
logging . debug ( f " Trying to remove networks: { list_networks } " )
2022-07-07 20:42:41 +00:00
run_and_check ( f " docker network rm { ' ' . join ( list_networks ) } " )
2022-07-07 20:19:15 +00:00
logging . debug ( f " Networks removed: { list_networks } " )
except :
pass
2021-04-27 09:07:21 +00:00
2021-04-27 15:34:33 +00:00
# Remove unused images
2022-07-07 20:19:15 +00:00
try :
logging . debug ( " Trying to prune unused images... " )
2021-04-27 09:07:21 +00:00
2022-07-07 20:42:41 +00:00
run_and_check ( [ " docker " , " image " , " prune " , " -f " ] )
2022-07-07 20:19:15 +00:00
logging . debug ( " Images pruned " )
except :
pass
2021-04-27 09:07:21 +00:00
# Remove unused volumes
try :
logging . debug ( " Trying to prune unused volumes... " )
2021-06-09 13:53:16 +00:00
result = run_and_check ( [ " docker volume ls | wc -l " ] , shell = True )
if int ( result > 0 ) :
run_and_check ( [ " docker " , " volume " , " prune " , " -f " ] )
logging . debug ( f " Volumes pruned: { result } " )
2021-04-27 09:07:21 +00:00
except :
pass
2021-04-13 14:55:31 +00:00
def get_docker_handle ( self , docker_id ) :
2021-06-21 08:14:26 +00:00
exception = None
2024-03-25 10:15:43 +00:00
for i in range ( 20 ) :
2021-06-21 08:14:26 +00:00
try :
return self . docker_client . containers . get ( docker_id )
except Exception as ex :
print ( " Got exception getting docker handle " , str ( ex ) )
2024-03-25 10:15:43 +00:00
time . sleep ( 0.5 )
2021-06-21 08:14:26 +00:00
exception = ex
raise exception
2021-04-13 14:55:31 +00:00
2018-09-07 11:51:51 +00:00
def get_client_cmd ( self ) :
cmd = self . client_bin_path
if p . basename ( cmd ) == " clickhouse " :
cmd + = " client "
return cmd
2021-11-21 15:01:28 +00:00
# Returns the list of currently running docker containers corresponding to this ClickHouseCluster.
def get_running_containers ( self ) :
# docker-compose names containers using the following formula:
# container_name = project_name + '_' + instance_name + '_1'
# We need to have "^/" and "$" in the "--filter name" option below to filter by exact name of the container, see
# https://stackoverflow.com/questions/48767760/how-to-make-docker-container-ls-f-name-filter-by-exact-name
filter_name = f " ^/ { self . project_name } _.*_1$ "
# We want the command "docker container list" to show only containers' ID and their names, separated by colon.
format = " {{ .ID}}: {{ .Names}} "
containers = run_and_check (
f " docker container list --all --filter name= ' { filter_name } ' --format ' { format } ' " ,
shell = True ,
)
2022-07-07 20:42:41 +00:00
containers = dict ( line . split ( " : " , 1 ) for line in containers . splitlines ( ) )
2021-11-21 15:01:28 +00:00
return containers
2021-08-18 13:34:57 +00:00
def copy_file_from_container_to_container (
self , src_node , src_path , dst_node , dst_path
) :
fname = os . path . basename ( src_path )
run_and_check (
[ f " docker cp { src_node . docker_id } : { src_path } { self . instances_dir } " ] ,
shell = True ,
)
run_and_check (
[ f " docker cp { self . instances_dir } / { fname } { dst_node . docker_id } : { dst_path } " ] ,
2021-10-08 09:16:23 +00:00
shell = True ,
2022-03-22 16:39:58 +00:00
)
2021-05-21 13:29:43 +00:00
def setup_zookeeper_secure_cmd (
self , instance , env_variables , docker_compose_yml_dir
) :
logging . debug ( " Setup ZooKeeper Secure " )
zookeeper_docker_compose_path = p . join (
docker_compose_yml_dir , " docker_compose_zookeeper_secure.yml "
)
env_variables [ " ZOO_SECURE_CLIENT_PORT " ] = str ( self . zookeeper_secure_port )
env_variables [ " ZK_FS " ] = " bind "
for i in range ( 1 , 4 ) :
zk_data_path = os . path . join (
self . zookeeper_instance_dir_prefix + str ( i ) , " data "
)
zk_log_path = os . path . join (
self . zookeeper_instance_dir_prefix + str ( i ) , " log "
2022-03-22 16:39:58 +00:00
)
2021-05-21 13:29:43 +00:00
env_variables [ " ZK_DATA " + str ( i ) ] = zk_data_path
env_variables [ " ZK_DATA_LOG " + str ( i ) ] = zk_log_path
self . zookeeper_dirs_to_create + = [ zk_data_path , zk_log_path ]
logging . debug ( f " DEBUG ZK: { self . zookeeper_dirs_to_create } " )
self . with_zookeeper_secure = True
self . base_cmd . extend ( [ " --file " , zookeeper_docker_compose_path ] )
self . base_zookeeper_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
zookeeper_docker_compose_path ,
]
return self . base_zookeeper_cmd
2021-05-20 16:12:04 +00:00
def setup_zookeeper_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
logging . debug ( " Setup ZooKeeper " )
zookeeper_docker_compose_path = p . join (
docker_compose_yml_dir , " docker_compose_zookeeper.yml "
)
env_variables [ " ZK_FS " ] = " bind "
for i in range ( 1 , 4 ) :
zk_data_path = os . path . join (
self . zookeeper_instance_dir_prefix + str ( i ) , " data "
)
zk_log_path = os . path . join (
self . zookeeper_instance_dir_prefix + str ( i ) , " log "
2022-03-22 16:39:58 +00:00
)
2021-05-20 16:12:04 +00:00
env_variables [ " ZK_DATA " + str ( i ) ] = zk_data_path
env_variables [ " ZK_DATA_LOG " + str ( i ) ] = zk_log_path
self . zookeeper_dirs_to_create + = [ zk_data_path , zk_log_path ]
logging . debug ( f " DEBUG ZK: { self . zookeeper_dirs_to_create } " )
self . with_zookeeper = True
self . base_cmd . extend ( [ " --file " , zookeeper_docker_compose_path ] )
self . base_zookeeper_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
zookeeper_docker_compose_path ,
]
return self . base_zookeeper_cmd
def setup_keeper_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
logging . debug ( " Setup Keeper " )
keeper_docker_compose_path = p . join (
docker_compose_yml_dir , " docker_compose_keeper.yml "
)
binary_path = self . server_bin_path
2022-04-20 10:11:55 +00:00
binary_dir = os . path . dirname ( self . server_bin_path )
# always prefer clickhouse-keeper standalone binary
2022-09-09 20:09:54 +00:00
if os . path . exists (
os . path . join ( binary_dir , " clickhouse-keeper " )
) and not os . path . islink ( os . path . join ( binary_dir , " clickhouse-keeper " ) ) :
2022-04-20 10:11:55 +00:00
binary_path = os . path . join ( binary_dir , " clickhouse-keeper " )
keeper_cmd_prefix = " clickhouse-keeper "
else :
if binary_path . endswith ( " -server " ) :
binary_path = binary_path [ : - len ( " -server " ) ]
keeper_cmd_prefix = " clickhouse keeper "
2021-05-20 16:12:04 +00:00
env_variables [ " keeper_binary " ] = binary_path
2022-04-20 10:11:55 +00:00
env_variables [ " keeper_cmd_prefix " ] = keeper_cmd_prefix
2021-09-08 10:03:54 +00:00
env_variables [ " image " ] = " clickhouse/integration-test: " + self . docker_base_tag
2021-05-20 16:12:04 +00:00
env_variables [ " user " ] = str ( os . getuid ( ) )
env_variables [ " keeper_fs " ] = " bind "
for i in range ( 1 , 4 ) :
keeper_instance_dir = self . keeper_instance_dir_prefix + f " { i } "
logs_dir = os . path . join ( keeper_instance_dir , " log " )
configs_dir = os . path . join ( keeper_instance_dir , " config " )
coordination_dir = os . path . join ( keeper_instance_dir , " coordination " )
env_variables [ f " keeper_logs_dir { i } " ] = logs_dir
env_variables [ f " keeper_config_dir { i } " ] = configs_dir
env_variables [ f " keeper_db_dir { i } " ] = coordination_dir
self . zookeeper_dirs_to_create + = [ logs_dir , configs_dir , coordination_dir ]
self . with_zookeeper = True
self . base_cmd . extend ( [ " --file " , keeper_docker_compose_path ] )
self . base_zookeeper_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
keeper_docker_compose_path ,
]
return self . base_zookeeper_cmd
2021-04-13 14:55:31 +00:00
def setup_mysql_client_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_mysql_client = True
self . base_cmd . extend (
2022-03-22 16:39:58 +00:00
[
2021-04-13 14:55:31 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_mysql_client.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-04-13 14:55:31 +00:00
)
self . base_mysql_client_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_mysql_client.yml " ) ,
]
return self . base_mysql_client_cmd
2024-02-13 12:15:46 +00:00
def setup_mysql57_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_mysql57 = True
env_variables [ " MYSQL_HOST " ] = self . mysql57_host
env_variables [ " MYSQL_PORT " ] = str ( self . mysql57_port )
2021-03-05 13:39:51 +00:00
env_variables [ " MYSQL_ROOT_HOST " ] = " % "
2024-02-13 12:15:46 +00:00
env_variables [ " MYSQL_LOGS " ] = self . mysql57_logs_dir
2021-03-05 13:39:51 +00:00
env_variables [ " MYSQL_LOGS_FS " ] = " bind "
2024-03-23 13:53:13 +00:00
env_variables [ " MYSQL_DOCKER_USER " ] = str ( os . getuid ( ) )
2021-02-15 09:35:45 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_mysql.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2024-02-13 12:15:46 +00:00
self . base_mysql57_cmd = [
2021-02-19 12:58:11 +00:00
" docker-compose " ,
" --env-file " ,
2021-11-26 14:04:53 +00:00
instance . env_file ,
" --project-name " ,
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-15 09:35:45 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_mysql.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-15 09:35:45 +00:00
2024-02-13 12:15:46 +00:00
return self . base_mysql57_cmd
2021-02-15 09:35:45 +00:00
2021-02-16 07:10:01 +00:00
def setup_mysql8_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_mysql8 = True
env_variables [ " MYSQL8_HOST " ] = self . mysql8_host
2021-03-09 07:32:10 +00:00
env_variables [ " MYSQL8_PORT " ] = str ( self . mysql8_port )
env_variables [ " MYSQL8_ROOT_HOST " ] = " % "
2021-03-05 13:39:51 +00:00
env_variables [ " MYSQL8_LOGS " ] = self . mysql8_logs_dir
env_variables [ " MYSQL8_LOGS_FS " ] = " bind "
2024-03-23 13:53:13 +00:00
env_variables [ " MYSQL8_DOCKER_USER " ] = str ( os . getuid ( ) )
2021-02-16 07:10:01 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_mysql_8_0.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-16 07:10:01 +00:00
self . base_mysql8_cmd = [
2021-06-07 12:56:29 +00:00
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-06-13 12:56:22 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-16 07:10:01 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_mysql_8_0.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-16 07:10:01 +00:00
return self . base_mysql8_cmd
2021-06-13 12:56:22 +00:00
def setup_mysql_cluster_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
2021-04-08 09:30:24 +00:00
self . with_mysql_cluster = True
2024-02-13 12:15:46 +00:00
env_variables [ " MYSQL_CLUSTER_PORT " ] = str ( self . mysql8_port )
2021-04-08 09:30:24 +00:00
env_variables [ " MYSQL_CLUSTER_ROOT_HOST " ] = " % "
env_variables [ " MYSQL_CLUSTER_LOGS " ] = self . mysql_cluster_logs_dir
env_variables [ " MYSQL_CLUSTER_LOGS_FS " ] = " bind "
2024-03-23 13:53:13 +00:00
env_variables [ " MYSQL_CLUSTER_DOCKER_USER " ] = str ( os . getuid ( ) )
2022-03-22 16:39:58 +00:00
2021-04-08 09:30:24 +00:00
self . base_cmd . extend (
2022-03-22 16:39:58 +00:00
[
2021-04-08 09:30:24 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_mysql_cluster.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-04-08 09:30:24 +00:00
)
self . base_mysql_cluster_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_mysql_cluster.yml " ) ,
]
return self . base_mysql_cluster_cmd
2021-03-19 12:33:14 +00:00
def setup_postgres_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_postgres.yml " ) ]
)
env_variables [ " POSTGRES_PORT " ] = str ( self . postgres_port )
2021-03-19 16:44:08 +00:00
env_variables [ " POSTGRES_DIR " ] = self . postgres_logs_dir
2021-03-19 12:33:14 +00:00
env_variables [ " POSTGRES_LOGS_FS " ] = " bind "
self . with_postgres = True
self . base_postgres_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_postgres.yml " ) ,
]
return self . base_postgres_cmd
2021-06-13 12:56:22 +00:00
def setup_postgres_cluster_cmd (
self , instance , env_variables , docker_compose_yml_dir
) :
2021-04-08 14:43:57 +00:00
self . with_postgres_cluster = True
env_variables [ " POSTGRES_PORT " ] = str ( self . postgres_port )
env_variables [ " POSTGRES2_DIR " ] = self . postgres2_logs_dir
env_variables [ " POSTGRES3_DIR " ] = self . postgres3_logs_dir
env_variables [ " POSTGRES4_DIR " ] = self . postgres4_logs_dir
env_variables [ " POSTGRES_LOGS_FS " ] = " bind "
2021-04-13 16:08:29 +00:00
self . base_cmd . extend (
2022-03-22 16:39:58 +00:00
[
2021-04-13 16:08:29 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_postgres_cluster.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-04-13 16:08:29 +00:00
)
2021-04-08 14:43:57 +00:00
self . base_postgres_cluster_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_postgres_cluster.yml " ) ,
]
2023-01-03 19:30:14 +00:00
def setup_postgresql_java_client_cmd (
self , instance , env_variables , docker_compose_yml_dir
) :
self . with_postgresql_java_client = True
self . base_cmd . extend (
[
" --file " ,
2023-01-03 21:28:54 +00:00
p . join (
docker_compose_yml_dir , " docker_compose_postgresql_java_client.yml "
) ,
2023-01-03 19:30:14 +00:00
]
)
self . base_postgresql_java_client_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_postgresql_java_client.yml " ) ,
]
2021-02-18 21:21:50 +00:00
def setup_hdfs_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_hdfs = True
env_variables [ " HDFS_HOST " ] = self . hdfs_host
2021-06-09 09:23:02 +00:00
env_variables [ " HDFS_NAME_PORT " ] = str ( self . hdfs_name_port )
env_variables [ " HDFS_DATA_PORT " ] = str ( self . hdfs_data_port )
2021-02-18 21:21:50 +00:00
env_variables [ " HDFS_LOGS " ] = self . hdfs_logs_dir
env_variables [ " HDFS_FS " ] = " bind "
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_hdfs.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-19 14:42:43 +00:00
self . base_hdfs_cmd = [
2021-11-18 08:17:49 +00:00
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-19 14:42:43 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_hdfs.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-06-11 12:00:40 +00:00
logging . debug ( " HDFS BASE CMD: { self.base_hdfs_cmd)} " )
2021-02-18 21:21:50 +00:00
return self . base_hdfs_cmd
2021-02-19 14:42:43 +00:00
def setup_kerberized_hdfs_cmd (
self , instance , env_variables , docker_compose_yml_dir
) :
self . with_kerberized_hdfs = True
env_variables [ " KERBERIZED_HDFS_HOST " ] = self . hdfs_kerberized_host
2021-06-09 09:23:02 +00:00
env_variables [ " KERBERIZED_HDFS_NAME_PORT " ] = str ( self . hdfs_kerberized_name_port )
env_variables [ " KERBERIZED_HDFS_DATA_PORT " ] = str ( self . hdfs_kerberized_data_port )
2021-02-19 14:42:43 +00:00
env_variables [ " KERBERIZED_HDFS_LOGS " ] = self . hdfs_kerberized_logs_dir
env_variables [ " KERBERIZED_HDFS_FS " ] = " bind "
env_variables [ " KERBERIZED_HDFS_DIR " ] = instance . path + " / "
self . base_cmd . extend (
2022-03-22 16:39:58 +00:00
[
2021-02-19 14:42:43 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_kerberized_hdfs.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-19 14:42:43 +00:00
)
self . base_kerberized_hdfs_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_kerberized_hdfs.yml " ) ,
]
return self . base_kerberized_hdfs_cmd
2023-05-08 20:49:43 +00:00
def setup_kafka_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
2021-02-24 11:46:58 +00:00
self . with_kafka = True
2021-02-18 12:57:45 +00:00
env_variables [ " KAFKA_HOST " ] = self . kafka_host
env_variables [ " KAFKA_EXTERNAL_PORT " ] = str ( self . kafka_port )
2023-04-29 21:23:55 +00:00
env_variables [ " SCHEMA_REGISTRY_DIR " ] = instance . path + " / "
2021-02-18 12:57:45 +00:00
env_variables [ " SCHEMA_REGISTRY_EXTERNAL_PORT " ] = str ( self . schema_registry_port )
2023-05-08 20:49:43 +00:00
env_variables [ " SCHEMA_REGISTRY_AUTH_EXTERNAL_PORT " ] = str (
self . schema_registry_auth_port
)
2021-02-18 12:57:45 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_kafka.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-18 21:21:50 +00:00
self . base_kafka_cmd = [
2021-11-18 08:17:49 +00:00
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-18 12:57:45 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_kafka.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-18 21:21:50 +00:00
return self . base_kafka_cmd
2021-02-18 12:57:45 +00:00
def setup_kerberized_kafka_cmd (
self , instance , env_variables , docker_compose_yml_dir
) :
2021-02-24 11:46:58 +00:00
self . with_kerberized_kafka = True
2021-02-18 12:57:45 +00:00
env_variables [ " KERBERIZED_KAFKA_DIR " ] = instance . path + " / "
env_variables [ " KERBERIZED_KAFKA_HOST " ] = self . kerberized_kafka_host
env_variables [ " KERBERIZED_KAFKA_EXTERNAL_PORT " ] = str (
self . kerberized_kafka_port
)
self . base_cmd . extend (
2022-03-22 16:39:58 +00:00
[
2021-02-18 12:57:45 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_kerberized_kafka.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-18 12:57:45 +00:00
)
2021-02-18 21:21:50 +00:00
self . base_kerberized_kafka_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
2021-02-18 12:57:45 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_kerberized_kafka.yml " ) ,
]
2021-02-18 21:21:50 +00:00
return self . base_kerberized_kafka_cmd
2021-02-18 12:57:45 +00:00
2022-12-19 11:28:49 +00:00
def setup_kerberos_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
2022-12-14 13:39:23 +00:00
self . with_kerberos_kdc = True
2022-12-19 11:28:49 +00:00
env_variables [ " KERBEROS_KDC_DIR " ] = self . instances_dir + " / "
2022-12-14 13:39:23 +00:00
env_variables [ " KERBEROS_KDC_HOST " ] = self . kerberos_kdc_host
self . base_cmd . extend (
[
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_kerberos_kdc.yml " ) ,
]
)
2022-12-19 11:28:49 +00:00
self . base_kerberos_kdc_cmd = [
2022-12-14 13:39:23 +00:00
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_kerberos_kdc.yml " ) ,
]
2022-12-19 11:28:49 +00:00
return self . base_kerberos_kdc_cmd
2022-12-14 13:39:23 +00:00
2021-02-16 14:16:15 +00:00
def setup_redis_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_redis = True
env_variables [ " REDIS_HOST " ] = self . redis_host
env_variables [ " REDIS_EXTERNAL_PORT " ] = str ( self . redis_port )
env_variables [ " REDIS_INTERNAL_PORT " ] = " 6379 "
2022-03-22 16:39:58 +00:00
2021-02-16 14:16:15 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_redis.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-16 14:16:15 +00:00
self . base_redis_cmd = [
2021-11-18 08:17:49 +00:00
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-16 14:16:15 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_redis.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-16 14:16:15 +00:00
return self . base_redis_cmd
def setup_rabbitmq_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_rabbitmq = True
env_variables [ " RABBITMQ_HOST " ] = self . rabbitmq_host
2021-04-30 09:18:12 +00:00
env_variables [ " RABBITMQ_PORT " ] = str ( self . rabbitmq_port )
2023-11-16 12:09:13 +00:00
env_variables [ " RABBITMQ_SECURE_PORT " ] = str ( self . rabbitmq_secure_port )
2021-05-17 11:16:16 +00:00
env_variables [ " RABBITMQ_LOGS " ] = self . rabbitmq_logs_dir
env_variables [ " RABBITMQ_LOGS_FS " ] = " bind "
2023-10-09 18:33:03 +00:00
env_variables [ " RABBITMQ_COOKIE_FILE " ] = self . rabbitmq_cookie_file
env_variables [ " RABBITMQ_COOKIE_FILE_FS " ] = " bind "
2022-03-22 16:39:58 +00:00
2021-02-16 14:16:15 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_rabbitmq.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-16 14:16:15 +00:00
self . base_rabbitmq_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-16 14:16:15 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_rabbitmq.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-16 14:16:15 +00:00
return self . base_rabbitmq_cmd
2022-05-17 13:58:09 +00:00
def setup_nats_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_nats = True
env_variables [ " NATS_HOST " ] = self . nats_host
env_variables [ " NATS_INTERNAL_PORT " ] = " 4444 "
env_variables [ " NATS_EXTERNAL_PORT " ] = str ( self . nats_port )
2022-07-24 14:13:17 +00:00
env_variables [ " NATS_CERT_DIR " ] = self . nats_cert_dir
2022-05-17 13:58:09 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_nats.yml " ) ]
)
self . base_nats_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_nats.yml " ) ,
]
return self . base_nats_cmd
2021-07-27 15:54:13 +00:00
def setup_mongo_secure_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_mongo = self . with_mongo_secure = True
env_variables [ " MONGO_HOST " ] = self . mongo_host
env_variables [ " MONGO_EXTERNAL_PORT " ] = str ( self . mongo_port )
env_variables [ " MONGO_INTERNAL_PORT " ] = " 27017 "
env_variables [ " MONGO_CONFIG_PATH " ] = HELPERS_DIR
self . base_cmd . extend (
2022-03-22 16:39:58 +00:00
[
2021-07-27 15:54:13 +00:00
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_mongo_secure.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-07-27 15:54:13 +00:00
)
self . base_mongo_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_mongo_secure.yml " ) ,
]
return self . base_mongo_cmd
2021-02-18 21:21:50 +00:00
def setup_mongo_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_mongo = True
env_variables [ " MONGO_HOST " ] = self . mongo_host
env_variables [ " MONGO_EXTERNAL_PORT " ] = str ( self . mongo_port )
env_variables [ " MONGO_INTERNAL_PORT " ] = " 27017 "
2022-01-14 13:00:38 +00:00
env_variables [ " MONGO_NO_CRED_EXTERNAL_PORT " ] = str ( self . mongo_no_cred_port )
env_variables [ " MONGO_NO_CRED_INTERNAL_PORT " ] = " 27017 "
2021-02-18 21:21:50 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_mongo.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-18 21:21:50 +00:00
self . base_mongo_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-18 21:21:50 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_mongo.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-18 21:21:50 +00:00
return self . base_mongo_cmd
2022-07-20 17:09:38 +00:00
def setup_coredns_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_coredns = True
env_variables [ " COREDNS_CONFIG_DIR " ] = instance . path + " / " + " coredns_config "
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_coredns.yml " ) ]
)
self . base_coredns_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_coredns.yml " ) ,
]
return self . base_coredns_cmd
2021-02-19 12:58:11 +00:00
def setup_minio_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
2021-06-13 12:56:22 +00:00
self . with_minio = True
2021-02-19 12:58:11 +00:00
cert_d = p . join ( self . minio_dir , " certs " )
env_variables [ " MINIO_CERTS_DIR " ] = cert_d
2023-01-03 19:30:14 +00:00
env_variables [ " MINIO_DATA_DIR " ] = self . minio_data_dir
2021-04-14 11:21:40 +00:00
env_variables [ " MINIO_PORT " ] = str ( self . minio_port )
2021-02-19 12:58:11 +00:00
env_variables [ " SSL_CERT_FILE " ] = p . join ( self . base_dir , cert_d , " public.crt " )
2022-03-22 16:39:58 +00:00
2021-02-19 12:58:11 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_minio.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-02-19 12:58:11 +00:00
self . base_minio_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-02-19 12:58:11 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_minio.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-02-19 12:58:11 +00:00
return self . base_minio_cmd
2021-11-26 14:04:53 +00:00
def setup_azurite_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_azurite = True
2023-10-18 20:20:39 +00:00
env_variables [ " AZURITE_PORT " ] = str ( self . azurite_port )
2024-07-10 14:19:06 +00:00
env_variables [ " AZURITE_STORAGE_ACCOUNT_URL " ] = (
f " http://azurite1: { env_variables [ ' AZURITE_PORT ' ] } /devstoreaccount1 "
)
2023-10-23 12:28:02 +00:00
env_variables [ " AZURITE_CONNECTION_STRING " ] = (
f " DefaultEndpointsProtocol=http;AccountName=devstoreaccount1; "
f " AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==; "
f " BlobEndpoint= { env_variables [ ' AZURITE_STORAGE_ACCOUNT_URL ' ] } ; "
)
2023-10-23 12:13:14 +00:00
2021-11-26 14:04:53 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_azurite.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-11-26 14:04:53 +00:00
self . base_azurite_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-11-26 14:04:53 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_azurite.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-11-26 14:04:53 +00:00
return self . base_azurite_cmd
2021-05-20 16:12:04 +00:00
def setup_cassandra_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_cassandra = True
env_variables [ " CASSANDRA_PORT " ] = str ( self . cassandra_port )
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_cassandra.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-05-20 16:12:04 +00:00
self . base_cassandra_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-05-20 16:12:04 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_cassandra.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-05-20 16:12:04 +00:00
return self . base_cassandra_cmd
2023-09-29 07:16:14 +00:00
def setup_ldap_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_ldap = True
env_variables [ " LDAP_EXTERNAL_PORT " ] = str ( self . ldap_port )
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_ldap.yml " ) ]
)
self . base_ldap_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_ldap.yml " ) ,
]
return self . base_ldap_cmd
2021-06-07 12:56:29 +00:00
def setup_jdbc_bridge_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_jdbc_bridge = True
2021-07-27 07:33:49 +00:00
env_variables [ " JDBC_DRIVER_LOGS " ] = self . jdbc_driver_logs_dir
env_variables [ " JDBC_DRIVER_FS " ] = " bind "
2021-06-07 12:56:29 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_jdbc_bridge.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-06-07 12:56:29 +00:00
self . base_jdbc_bridge_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-06-07 12:56:29 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_jdbc_bridge.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-06-07 12:56:29 +00:00
return self . base_jdbc_bridge_cmd
2021-06-13 12:56:22 +00:00
def setup_nginx_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_nginx = True
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_nginx.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-06-13 12:56:22 +00:00
self . base_nginx_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-06-13 12:56:22 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_nginx.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-06-13 12:56:22 +00:00
return self . base_nginx_cmd
2021-11-18 08:17:49 +00:00
def setup_hive ( self , instance , env_variables , docker_compose_yml_dir ) :
self . with_hive = True
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_hive.yml " ) ]
2022-03-22 16:39:58 +00:00
)
2021-11-18 08:17:49 +00:00
self . base_hive_cmd = [
2021-06-04 10:14:32 +00:00
" docker-compose " ,
2021-11-18 08:17:49 +00:00
" --env-file " ,
instance . env_file ,
" --project-name " ,
2021-01-27 16:21:54 +00:00
self . project_name ,
2022-03-22 16:39:58 +00:00
" --file " ,
2021-11-18 08:17:49 +00:00
p . join ( docker_compose_yml_dir , " docker_compose_hive.yml " ) ,
2022-03-22 16:39:58 +00:00
]
2021-11-18 08:17:49 +00:00
return self . base_hive_cmd
2024-03-11 22:58:05 +00:00
def setup_prometheus_cmd ( self , instance , env_variables , docker_compose_yml_dir ) :
env_variables [ " PROMETHEUS_WRITER_HOST " ] = self . prometheus_writer_host
env_variables [ " PROMETHEUS_WRITER_PORT " ] = str ( self . prometheus_writer_port )
env_variables [ " PROMETHEUS_WRITER_LOGS " ] = self . prometheus_writer_logs_dir
env_variables [ " PROMETHEUS_WRITER_LOGS_FS " ] = " bind "
env_variables [ " PROMETHEUS_READER_HOST " ] = self . prometheus_reader_host
env_variables [ " PROMETHEUS_READER_PORT " ] = str ( self . prometheus_reader_port )
env_variables [ " PROMETHEUS_READER_LOGS " ] = self . prometheus_reader_logs_dir
env_variables [ " PROMETHEUS_READER_LOGS_FS " ] = " bind "
if self . prometheus_remote_write_handler_host :
env_variables [ " PROMETHEUS_REMOTE_WRITE_HANDLER " ] = (
f " http:// { self . prometheus_remote_write_handler_host } : { self . prometheus_remote_write_handler_port } / { self . prometheus_remote_write_handler_path . strip ( ' / ' ) } "
)
if self . prometheus_remote_read_handler_host :
env_variables [ " PROMETHEUS_REMOTE_READ_HANDLER " ] = (
f " http:// { self . prometheus_remote_read_handler_host } : { self . prometheus_remote_read_handler_port } / { self . prometheus_remote_read_handler_path . strip ( ' / ' ) } "
)
if not self . with_prometheus :
self . with_prometheus = True
self . base_cmd . extend (
[
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_prometheus.yml " ) ,
]
)
self . base_prometheus_cmd = [
" docker-compose " ,
" --env-file " ,
instance . env_file ,
" --project-name " ,
self . project_name ,
" --file " ,
p . join ( docker_compose_yml_dir , " docker_compose_prometheus.yml " ) ,
]
return self . base_prometheus_cmd
2020-10-30 19:40:16 +00:00
def add_instance (
self ,
name ,
base_config_dir = None ,
main_configs = None ,
user_configs = None ,
dictionaries = None ,
2021-05-21 13:29:43 +00:00
macros = None ,
with_zookeeper = False ,
with_zookeeper_secure = False ,
2021-06-13 12:56:22 +00:00
with_mysql_client = False ,
2024-02-13 12:15:46 +00:00
with_mysql57 = False ,
2021-06-13 12:56:22 +00:00
with_mysql8 = False ,
with_mysql_cluster = False ,
2021-04-08 09:30:24 +00:00
with_kafka = False ,
with_kerberized_kafka = False ,
2022-12-14 13:39:23 +00:00
with_kerberos_kdc = False ,
2023-04-29 21:23:55 +00:00
with_secrets = False ,
2021-04-08 09:30:24 +00:00
with_rabbitmq = False ,
2022-05-17 13:58:09 +00:00
with_nats = False ,
2021-04-08 09:30:24 +00:00
clickhouse_path_dir = None ,
2021-07-27 15:54:13 +00:00
with_odbc_drivers = False ,
with_postgres = False ,
with_postgres_cluster = False ,
2023-01-03 21:28:54 +00:00
with_postgresql_java_client = False ,
2023-07-06 06:16:18 +00:00
clickhouse_log_file = CLICKHOUSE_LOG_FILE ,
clickhouse_error_log_file = CLICKHOUSE_ERROR_LOG_FILE ,
2021-07-27 15:54:13 +00:00
with_hdfs = False ,
2021-08-06 19:41:03 +00:00
with_kerberized_hdfs = False ,
with_mongo = False ,
with_mongo_secure = False ,
with_nginx = False ,
2021-12-14 02:39:21 +00:00
with_redis = False ,
with_minio = False ,
with_azurite = False ,
with_cassandra = False ,
2023-09-29 07:16:14 +00:00
with_ldap = False ,
2021-12-14 02:39:21 +00:00
with_jdbc_bridge = False ,
with_hive = False ,
2022-07-20 17:09:38 +00:00
with_coredns = False ,
2024-03-11 22:58:05 +00:00
with_prometheus = False ,
handle_prometheus_remote_write = False ,
handle_prometheus_remote_read = False ,
2024-04-27 08:32:27 +00:00
use_old_analyzer = None ,
2021-09-08 10:03:54 +00:00
hostname = None ,
env_variables = None ,
2023-10-03 16:53:38 +00:00
instance_env_variables = False ,
2021-09-08 10:03:54 +00:00
image = " clickhouse/integration-test " ,
tag = None ,
2021-11-09 08:34:37 +00:00
stay_alive = False ,
ipv4_address = None ,
ipv6_address = None ,
with_installed_binary = False ,
external_dirs = None ,
tmpfs = None ,
2024-03-12 22:31:33 +00:00
mem_limit = None ,
2021-05-24 06:59:16 +00:00
zookeeper_docker_compose_path = None ,
minio_certs_dir = None ,
2023-01-03 19:30:14 +00:00
minio_data_dir = None ,
2021-05-24 06:59:16 +00:00
use_keeper = True ,
2021-11-23 02:19:02 +00:00
main_config_name = " config.xml " ,
users_config_name = " users.xml " ,
copy_common_configs = True ,
config_root_name = " clickhouse " ,
extra_configs = [ ] ,
) - > " ClickHouseInstance " :
2017-05-19 18:54:05 +00:00
""" Add an instance to the cluster.
name - the name of the instance directory and the value of the ' instance ' macro in ClickHouse .
2020-08-12 08:55:04 +00:00
base_config_dir - a directory with config . xml and users . xml files which will be copied to / etc / clickhouse - server / directory
2017-05-30 11:49:17 +00:00
main_configs - a list of config files that will be added to config . d / directory
user_configs - a list of config files that will be added to users . d / directory
2017-05-19 18:54:05 +00:00
with_zookeeper - if True , add ZooKeeper configuration to configs and ZooKeeper instances to the cluster .
2021-05-21 13:29:43 +00:00
with_zookeeper_secure - if True , add ZooKeeper Secure configuration to configs and ZooKeeper instances to the cluster .
2021-11-19 01:58:34 +00:00
extra_configs - config files cannot put into config . d and users . d
2017-05-19 18:54:05 +00:00
"""
if self . is_up :
2017-05-23 17:13:36 +00:00
raise Exception ( " Can ' t add instance %s : cluster is already up! " % name )
2017-05-19 18:54:05 +00:00
if name in self . instances :
2017-05-23 17:13:36 +00:00
raise Exception (
" Can ' t add instance ` %s ' : there is already an instance with the same name! "
% name
)
2017-05-19 18:54:05 +00:00
2020-09-01 06:38:23 +00:00
if tag is None :
tag = self . docker_base_tag
2020-12-24 15:51:12 +00:00
if not env_variables :
env_variables = { }
2021-05-16 14:18:21 +00:00
self . use_keeper = use_keeper
2020-12-24 15:51:12 +00:00
# Code coverage files will be placed in database directory
# (affect only WITH_COVERAGE=1 build)
2024-07-10 14:19:06 +00:00
env_variables [ " LLVM_PROFILE_FILE " ] = (
" /var/lib/clickhouse/server_ % h_ % p_ % m.profraw "
)
2020-09-01 06:38:23 +00:00
2023-07-06 06:16:18 +00:00
clickhouse_start_command = CLICKHOUSE_START_COMMAND
if clickhouse_log_file :
clickhouse_start_command + = " --log-file= " + clickhouse_log_file
if clickhouse_error_log_file :
clickhouse_start_command + = " --errorlog-file= " + clickhouse_error_log_file
logging . debug ( f " clickhouse_start_command: { clickhouse_start_command } " )
2017-08-02 14:42:35 +00:00
instance = ClickHouseInstance (
2020-09-02 12:28:47 +00:00
cluster = self ,
base_path = self . base_dir ,
name = name ,
2024-07-10 14:19:06 +00:00
base_config_dir = (
base_config_dir if base_config_dir else self . base_config_dir
) ,
2020-09-02 12:28:47 +00:00
custom_main_configs = main_configs or [ ] ,
custom_user_configs = user_configs or [ ] ,
custom_dictionaries = dictionaries or [ ] ,
macros = macros or { } ,
with_zookeeper = with_zookeeper ,
zookeeper_config_path = self . zookeeper_config_path ,
2021-04-13 14:55:31 +00:00
with_mysql_client = with_mysql_client ,
2024-02-13 12:15:46 +00:00
with_mysql57 = with_mysql57 ,
2021-02-16 07:10:01 +00:00
with_mysql8 = with_mysql8 ,
2021-03-31 17:04:57 +00:00
with_mysql_cluster = with_mysql_cluster ,
2020-09-02 12:28:47 +00:00
with_kafka = with_kafka ,
2020-09-29 08:56:37 +00:00
with_kerberized_kafka = with_kerberized_kafka ,
2022-12-14 13:39:23 +00:00
with_kerberos_kdc = with_kerberos_kdc ,
2020-09-02 12:28:47 +00:00
with_rabbitmq = with_rabbitmq ,
2022-05-17 13:58:09 +00:00
with_nats = with_nats ,
2021-06-13 12:56:22 +00:00
with_nginx = with_nginx ,
2020-09-10 10:02:46 +00:00
with_kerberized_hdfs = with_kerberized_hdfs ,
2023-05-08 20:49:43 +00:00
with_secrets = with_secrets
or with_kerberized_hdfs
or with_kerberos_kdc
or with_kerberized_kafka ,
2021-07-27 15:54:13 +00:00
with_mongo = with_mongo or with_mongo_secure ,
2020-09-02 12:28:47 +00:00
with_redis = with_redis ,
with_minio = with_minio ,
2021-11-26 14:04:53 +00:00
with_azurite = with_azurite ,
2021-06-07 12:56:29 +00:00
with_jdbc_bridge = with_jdbc_bridge ,
2021-11-18 08:17:49 +00:00
with_hive = with_hive ,
2022-07-20 17:09:38 +00:00
with_coredns = with_coredns ,
2023-07-06 06:16:18 +00:00
with_cassandra = with_cassandra ,
2023-09-29 07:16:14 +00:00
with_ldap = with_ldap ,
2024-03-20 12:43:18 +00:00
use_old_analyzer = use_old_analyzer ,
2020-09-02 12:28:47 +00:00
server_bin_path = self . server_bin_path ,
odbc_bridge_bin_path = self . odbc_bridge_bin_path ,
2021-03-11 17:48:47 +00:00
library_bridge_bin_path = self . library_bridge_bin_path ,
2020-09-02 12:28:47 +00:00
clickhouse_path_dir = clickhouse_path_dir ,
with_odbc_drivers = with_odbc_drivers ,
2021-04-13 14:55:31 +00:00
with_postgres = with_postgres ,
with_postgres_cluster = with_postgres_cluster ,
2023-01-03 19:30:14 +00:00
with_postgresql_java_client = with_postgresql_java_client ,
2023-07-06 06:16:18 +00:00
clickhouse_start_command = clickhouse_start_command ,
main_config_name = main_config_name ,
users_config_name = users_config_name ,
copy_common_configs = copy_common_configs ,
2020-09-02 12:28:47 +00:00
hostname = hostname ,
2020-12-24 15:51:12 +00:00
env_variables = env_variables ,
2023-10-03 16:53:38 +00:00
instance_env_variables = instance_env_variables ,
2020-09-02 12:28:47 +00:00
image = image ,
tag = tag ,
stay_alive = stay_alive ,
ipv4_address = ipv4_address ,
2019-11-20 11:56:38 +00:00
ipv6_address = ipv6_address ,
2020-09-02 12:28:47 +00:00
with_installed_binary = with_installed_binary ,
2021-11-09 08:34:37 +00:00
external_dirs = external_dirs ,
2021-09-22 15:00:08 +00:00
tmpfs = tmpfs or [ ] ,
2024-03-12 22:31:33 +00:00
mem_limit = mem_limit ,
2021-11-18 08:17:49 +00:00
config_root_name = config_root_name ,
2021-11-19 01:58:34 +00:00
extra_configs = extra_configs ,
)
2017-08-02 14:42:35 +00:00
2020-07-06 13:45:54 +00:00
docker_compose_yml_dir = get_docker_compose_path ( )
2017-05-19 18:54:05 +00:00
self . instances [ name ] = instance
2019-06-04 20:59:31 +00:00
if ipv4_address is not None or ipv6_address is not None :
2019-06-05 09:23:41 +00:00
self . with_net_trics = True
2020-07-06 13:45:54 +00:00
self . base_cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_net.yml " ) ]
)
2019-06-04 20:59:31 +00:00
2017-05-19 18:54:05 +00:00
self . base_cmd . extend ( [ " --file " , instance . docker_compose_path ] )
2019-06-04 20:59:31 +00:00
2019-06-05 09:23:41 +00:00
cmds = [ ]
2021-05-21 13:29:43 +00:00
if with_zookeeper_secure and not self . with_zookeeper_secure :
cmds . append (
self . setup_zookeeper_secure_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-05-21 13:29:43 +00:00
2017-05-19 18:54:05 +00:00
if with_zookeeper and not self . with_zookeeper :
2021-05-20 16:12:04 +00:00
if self . use_keeper :
cmds . append (
self . setup_keeper_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-05-20 16:12:04 +00:00
else :
cmds . append (
self . setup_zookeeper_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2018-07-28 14:38:08 +00:00
2021-04-13 14:55:31 +00:00
if with_mysql_client and not self . with_mysql_client :
cmds . append (
self . setup_mysql_client_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-04-13 14:55:31 +00:00
2024-02-13 12:15:46 +00:00
if with_mysql57 and not self . with_mysql57 :
2021-02-15 09:35:45 +00:00
cmds . append (
2024-02-13 12:15:46 +00:00
self . setup_mysql57_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2019-06-05 09:23:41 +00:00
2021-02-16 07:10:01 +00:00
if with_mysql8 and not self . with_mysql8 :
cmds . append (
self . setup_mysql8_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2021-02-16 07:10:01 +00:00
2021-03-31 17:04:57 +00:00
if with_mysql_cluster and not self . with_mysql_cluster :
2021-04-08 09:30:24 +00:00
cmds . append (
self . setup_mysql_cluster_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-03-31 17:04:57 +00:00
2018-10-15 14:49:23 +00:00
if with_postgres and not self . with_postgres :
2021-02-17 15:40:01 +00:00
cmds . append (
self . setup_postgres_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2018-10-15 14:49:23 +00:00
2021-03-31 17:04:57 +00:00
if with_postgres_cluster and not self . with_postgres_cluster :
2021-04-08 14:43:57 +00:00
cmds . append (
self . setup_postgres_cluster_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-03-31 17:04:57 +00:00
2023-01-03 19:30:14 +00:00
if with_postgresql_java_client and not self . with_postgresql_java_client :
cmds . append (
self . setup_postgresql_java_client_cmd (
instance , env_variables , docker_compose_yml_dir
)
)
2018-08-22 15:42:27 +00:00
if with_odbc_drivers and not self . with_odbc_drivers :
self . with_odbc_drivers = True
2024-02-13 12:15:46 +00:00
if not self . with_mysql8 :
2021-02-15 09:35:45 +00:00
cmds . append (
2024-02-13 12:15:46 +00:00
self . setup_mysql8_cmd (
2021-02-15 09:35:45 +00:00
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2019-06-05 09:23:41 +00:00
2018-10-15 14:49:23 +00:00
if not self . with_postgres :
2021-02-17 15:40:01 +00:00
cmds . append (
self . setup_postgres_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2018-10-15 14:49:23 +00:00
2018-07-18 05:22:01 +00:00
if with_kafka and not self . with_kafka :
2021-02-18 12:57:45 +00:00
cmds . append (
self . setup_kafka_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2018-07-18 05:22:01 +00:00
2020-09-29 08:56:37 +00:00
if with_kerberized_kafka and not self . with_kerberized_kafka :
2021-02-18 12:57:45 +00:00
cmds . append (
self . setup_kerberized_kafka_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2020-09-29 08:56:37 +00:00
2022-12-14 13:39:23 +00:00
if with_kerberos_kdc and not self . with_kerberos_kdc :
cmds . append (
2022-12-19 11:28:49 +00:00
self . setup_kerberos_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-12-14 13:39:23 +00:00
)
2020-05-20 06:22:12 +00:00
if with_rabbitmq and not self . with_rabbitmq :
2021-02-16 14:16:15 +00:00
cmds . append (
self . setup_rabbitmq_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2020-05-20 06:22:12 +00:00
2022-05-17 13:58:09 +00:00
if with_nats and not self . with_nats :
cmds . append (
self . setup_nats_cmd ( instance , env_variables , docker_compose_yml_dir )
)
2021-06-13 12:56:22 +00:00
if with_nginx and not self . with_nginx :
cmds . append (
self . setup_nginx_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2021-06-13 12:56:22 +00:00
2018-12-05 13:24:45 +00:00
if with_hdfs and not self . with_hdfs :
2021-02-18 21:21:50 +00:00
cmds . append (
self . setup_hdfs_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2018-12-05 13:24:45 +00:00
2020-09-10 10:02:46 +00:00
if with_kerberized_hdfs and not self . with_kerberized_hdfs :
2021-02-19 14:42:43 +00:00
cmds . append (
self . setup_kerberized_hdfs_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-07-27 15:54:13 +00:00
if ( with_mongo or with_mongo_secure ) and not (
self . with_mongo or self . with_mongo_secure
) :
if with_mongo_secure :
cmds . append (
self . setup_mongo_secure_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-07-27 17:43:41 +00:00
else :
2021-07-27 15:54:13 +00:00
cmds . append (
self . setup_mongo_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2019-06-05 09:23:41 +00:00
2022-07-20 17:09:38 +00:00
if with_coredns and not self . with_coredns :
cmds . append (
self . setup_coredns_cmd ( instance , env_variables , docker_compose_yml_dir )
)
2019-06-05 09:23:41 +00:00
if self . with_net_trics :
for cmd in cmds :
2020-07-06 13:45:54 +00:00
cmd . extend (
[ " --file " , p . join ( docker_compose_yml_dir , " docker_compose_net.yml " ) ]
)
2018-12-05 13:24:45 +00:00
2019-03-21 18:10:55 +00:00
if with_redis and not self . with_redis :
2021-02-16 14:16:15 +00:00
cmds . append (
self . setup_redis_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2019-03-21 18:10:55 +00:00
2019-11-20 11:56:38 +00:00
if with_minio and not self . with_minio :
2021-02-19 12:58:11 +00:00
cmds . append (
self . setup_minio_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2021-02-19 12:58:11 +00:00
2021-11-26 14:04:53 +00:00
if with_azurite and not self . with_azurite :
cmds . append (
self . setup_azurite_cmd ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2021-11-26 14:04:53 +00:00
2021-02-19 12:58:11 +00:00
if minio_certs_dir is not None :
if self . minio_certs_dir is None :
self . minio_certs_dir = minio_certs_dir
2021-02-12 15:51:21 +00:00
else :
2021-06-13 12:56:22 +00:00
raise Exception ( " Overwriting minio certs dir " )
2019-03-21 18:10:55 +00:00
2023-01-03 19:30:14 +00:00
if minio_data_dir is not None :
if self . minio_data_dir is None :
self . minio_data_dir = minio_data_dir
else :
raise Exception ( " Overwriting minio data dir " )
2020-05-19 02:21:27 +00:00
if with_cassandra and not self . with_cassandra :
2021-05-20 16:12:04 +00:00
cmds . append (
self . setup_cassandra_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2020-05-19 02:21:27 +00:00
2023-09-29 07:16:14 +00:00
if with_ldap and not self . with_ldap :
cmds . append (
2023-09-30 05:04:05 +00:00
self . setup_ldap_cmd ( instance , env_variables , docker_compose_yml_dir )
2023-09-29 07:16:14 +00:00
)
2021-06-07 12:56:29 +00:00
if with_jdbc_bridge and not self . with_jdbc_bridge :
cmds . append (
self . setup_jdbc_bridge_cmd (
instance , env_variables , docker_compose_yml_dir
)
2022-03-22 16:39:58 +00:00
)
2021-06-07 12:56:29 +00:00
2021-11-18 08:17:49 +00:00
if with_hive :
cmds . append (
self . setup_hive ( instance , env_variables , docker_compose_yml_dir )
2022-03-22 16:39:58 +00:00
)
2024-03-11 22:58:05 +00:00
if with_prometheus :
if handle_prometheus_remote_write :
self . prometheus_remote_write_handler_host = instance . hostname
if handle_prometheus_remote_read :
self . prometheus_remote_read_handler_host = instance . hostname
cmds . append (
self . setup_prometheus_cmd (
instance , env_variables , docker_compose_yml_dir
)
)
2021-02-18 21:21:50 +00:00
logging . debug (
" Cluster name: {} project_name: {} . Added instance name: {} tag: {} base_cmd: {} docker_compose_yml_dir: {} " . format (
2021-01-27 16:21:54 +00:00
self . name ,
self . project_name ,
name ,
tag ,
self . base_cmd ,
docker_compose_yml_dir ,
)
2022-03-22 16:39:58 +00:00
)
2017-05-19 18:54:05 +00:00
return instance
2017-05-30 11:49:17 +00:00
def get_instance_docker_id ( self , instance_name ) :
# According to how docker-compose names containers.
return self . project_name + " _ " + instance_name + " _1 "
2019-06-04 20:59:31 +00:00
def _replace ( self , path , what , to ) :
with open ( path , " r " ) as p :
data = p . read ( )
data = data . replace ( what , to )
with open ( path , " w " ) as p :
p . write ( data )
def restart_instance_with_ip_change ( self , node , new_ip ) :
if " :: " in new_ip :
if node . ipv6_address is None :
2019-07-21 11:45:01 +00:00
raise Exception ( " You should specity ipv6_address in add_node method " )
2019-06-04 20:59:31 +00:00
self . _replace ( node . docker_compose_path , node . ipv6_address , new_ip )
node . ipv6_address = new_ip
else :
if node . ipv4_address is None :
2019-07-21 11:45:01 +00:00
raise Exception ( " You should specity ipv4_address in add_node method " )
2019-06-04 20:59:31 +00:00
self . _replace ( node . docker_compose_path , node . ipv4_address , new_ip )
node . ipv4_address = new_ip
2021-01-22 14:27:23 +00:00
run_and_check ( self . base_cmd + [ " stop " , node . name ] )
run_and_check ( self . base_cmd + [ " rm " , " --force " , " --stop " , node . name ] )
run_and_check (
self . base_cmd + [ " up " , " --force-recreate " , " --no-deps " , " -d " , node . name ]
)
2019-06-04 20:59:31 +00:00
node . ip_address = self . get_instance_ip ( node . name )
node . client = Client ( node . ip_address , command = self . client_bin_path )
2021-06-01 14:23:01 +00:00
logging . info ( " Restart node with ip change " )
2021-05-28 14:37:05 +00:00
# In builds with sanitizer the server can take a long time to start
2021-05-25 15:32:24 +00:00
node . wait_for_start ( start_timeout = 180.0 , connection_timeout = 600.0 ) # seconds
2021-11-11 07:41:48 +00:00
res = node . client . query ( " SELECT 30 " )
logging . debug ( f " Read ' { res } ' " )
assert " 30 \n " == res
2021-06-01 14:23:01 +00:00
logging . info ( " Restarted " )
2019-06-04 20:59:31 +00:00
return node
2017-05-30 11:49:17 +00:00
2021-02-16 07:10:01 +00:00
def restart_service ( self , service_name ) :
run_and_check ( self . base_cmd + [ " restart " , service_name ] )
2018-01-25 18:14:37 +00:00
def get_instance_ip ( self , instance_name ) :
2021-02-18 21:21:50 +00:00
logging . debug ( " get_instance_ip instance_name= {} " . format ( instance_name ) )
2018-01-25 18:14:37 +00:00
docker_id = self . get_instance_docker_id ( instance_name )
2021-05-21 13:29:43 +00:00
# for cont in self.docker_client.containers.list():
# logging.debug("CONTAINERS LIST: ID={} NAME={} STATUS={}".format(cont.id, cont.name, cont.status))
2018-01-25 18:14:37 +00:00
handle = self . docker_client . containers . get ( docker_id )
2020-10-02 16:54:07 +00:00
return list ( handle . attrs [ " NetworkSettings " ] [ " Networks " ] . values ( ) ) [ 0 ] [
" IPAddress "
]
2018-01-25 18:14:37 +00:00
2022-07-20 17:09:38 +00:00
def get_instance_global_ipv6 ( self , instance_name ) :
logging . debug ( " get_instance_ip instance_name= {} " . format ( instance_name ) )
docker_id = self . get_instance_docker_id ( instance_name )
# for cont in self.docker_client.containers.list():
# logging.debug("CONTAINERS LIST: ID={} NAME={} STATUS={}".format(cont.id, cont.name, cont.status))
handle = self . docker_client . containers . get ( docker_id )
return list ( handle . attrs [ " NetworkSettings " ] [ " Networks " ] . values ( ) ) [ 0 ] [
" GlobalIPv6Address "
]
2020-04-29 08:39:00 +00:00
def get_container_id ( self , instance_name ) :
2021-06-01 14:18:35 +00:00
return self . get_instance_docker_id ( instance_name )
# docker_id = self.get_instance_docker_id(instance_name)
# handle = self.docker_client.containers.get(docker_id)
# return handle.attrs['Id']
2020-04-29 08:39:00 +00:00
def get_container_logs ( self , instance_name ) :
container_id = self . get_container_id ( instance_name )
2020-10-02 16:54:07 +00:00
return self . docker_client . api . logs ( container_id ) . decode ( )
2020-04-29 08:39:00 +00:00
2021-06-01 09:53:36 +00:00
def exec_in_container (
self , container_id , cmd , detach = False , nothrow = False , use_cli = True , * * kwargs
) :
if use_cli :
2021-06-01 14:18:35 +00:00
logging . debug (
f " run container_id: { container_id } detach: { detach } nothrow: { nothrow } cmd: { cmd } "
)
2021-06-10 12:15:07 +00:00
exec_cmd = [ " docker " , " exec " ]
if " user " in kwargs :
exec_cmd + = [ " -u " , kwargs [ " user " ] ]
2022-11-30 15:47:47 +00:00
if " privileged " in kwargs :
exec_cmd + = [ " --privileged " ]
2021-06-10 12:15:07 +00:00
result = subprocess_check_call (
exec_cmd + [ container_id ] + cmd , detach = detach , nothrow = nothrow
)
2021-06-01 09:53:36 +00:00
return result
else :
exec_id = self . docker_client . api . exec_create ( container_id , cmd , * * kwargs )
output = self . docker_client . api . exec_start ( exec_id , detach = detach )
exit_code = self . docker_client . api . exec_inspect ( exec_id ) [ " ExitCode " ]
if exit_code :
container_info = self . docker_client . api . inspect_container ( container_id )
image_id = container_info . get ( " Image " )
image_info = self . docker_client . api . inspect_image ( image_id )
logging . debug ( ( " Command failed in container {} : " . format ( container_id ) ) )
pprint . pprint ( container_info )
logging . debug ( " " )
logging . debug (
( " Container {} uses image {} : " . format ( container_id , image_id ) )
2022-03-22 16:39:58 +00:00
)
2021-06-01 09:53:36 +00:00
pprint . pprint ( image_info )
logging . debug ( " " )
message = ' Cmd " {} " failed in container {} . Return code {} . Output: {} ' . format (
" " . join ( cmd ) , container_id , exit_code , output
)
if nothrow :
logging . debug ( message )
else :
raise Exception ( message )
if not detach :
return output . decode ( )
return output
2020-05-08 10:53:12 +00:00
def copy_file_to_container ( self , container_id , local_path , dest_path ) :
2023-05-31 11:11:10 +00:00
with open ( local_path , " rb " ) as fdata :
2020-05-08 10:53:12 +00:00
data = fdata . read ( )
2023-05-31 11:11:10 +00:00
encodedBytes = base64 . b64encode ( data )
2020-10-02 16:54:07 +00:00
encodedStr = str ( encodedBytes , " utf-8 " )
2020-09-16 04:26:10 +00:00
self . exec_in_container (
container_id ,
2020-10-02 16:54:07 +00:00
[
" bash " ,
" -c " ,
" echo {} | base64 --decode > {} " . format ( encodedStr , dest_path ) ,
] ,
2020-05-08 10:53:12 +00:00
)
2022-03-22 16:39:58 +00:00
2021-06-07 12:56:29 +00:00
def wait_for_url (
self , url = " http://localhost:8123/ping " , conn_timeout = 2 , interval = 2 , timeout = 60
) :
if not url . startswith ( " http " ) :
url = " http:// " + url
if interval < = 0 :
interval = 2
if timeout < = 0 :
timeout = 60
attempts = 1
errors = [ ]
start = time . time ( )
while time . time ( ) - start < timeout :
try :
requests . get (
url , allow_redirects = True , timeout = conn_timeout , verify = False
) . raise_for_status ( )
logging . debug (
" {} is available after {} seconds " . format ( url , time . time ( ) - start )
2022-03-22 16:39:58 +00:00
)
2021-06-07 12:56:29 +00:00
return
except Exception as ex :
logging . debug (
" {} Attempt {} failed, retrying in {} seconds " . format (
ex , attempts , interval
)
2022-03-22 16:39:58 +00:00
)
2021-06-07 12:56:29 +00:00
attempts + = 1
errors + = [ str ( ex ) ]
time . sleep ( interval )
2021-09-02 12:43:15 +00:00
run_and_check ( [ " docker " , " ps " , " --all " ] )
2021-06-07 12:56:29 +00:00
logging . error ( " Can ' t connect to URL: {} " . format ( errors ) )
raise Exception (
" Cannot wait URL {} (interval= {} , timeout= {} , attempts= {} ) " . format (
url , interval , timeout , attempts
)
2022-03-22 16:39:58 +00:00
)
2021-06-07 12:56:29 +00:00
2021-04-13 14:55:31 +00:00
def wait_mysql_client_to_start ( self , timeout = 180 ) :
start = time . time ( )
errors = [ ]
self . mysql_client_container = self . get_docker_handle (
self . get_instance_docker_id ( self . mysql_client_host )
2022-03-22 16:39:58 +00:00
)
2021-04-13 14:55:31 +00:00
while time . time ( ) - start < timeout :
try :
info = self . mysql_client_container . client . api . inspect_container (
self . mysql_client_container . name
)
if info [ " State " ] [ " Health " ] [ " Status " ] == " healthy " :
logging . debug ( " Mysql Client Container Started " )
2021-12-14 07:01:59 +00:00
return
2021-04-13 14:55:31 +00:00
time . sleep ( 1 )
except Exception as ex :
errors + = [ str ( ex ) ]
time . sleep ( 1 )
tests/integration: fix printing all available services in wait_mysql_client_to_start()
There is no docker-compose.yml in that place, since mysql client has
different name:
2021-12-14 04:03:02 [ 862 ] DEBUG : Command:['docker-compose', 'ps', '--services', '--all'] (cluster.py:63, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Stderr: (cluster.py:71, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Stderr: Can't find a suitable configuration file in this directory or any (cluster.py:71, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Stderr: parent. Are you in the right directory? (cluster.py:71, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Stderr: (cluster.py:71, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Stderr: Supported filenames: docker-compose.yml, docker-compose.yaml (cluster.py:71, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Stderr: (cluster.py:71, run_and_check)
2021-12-14 04:03:03 [ 862 ] DEBUG : Exitcode:1 (cluster.py:73, run_and_check)
2021-12-14 06:59:54 +00:00
run_and_check ( [ " docker " , " ps " , " --all " ] )
2021-04-13 14:55:31 +00:00
logging . error ( " Can ' t connect to MySQL Client: {} " . format ( errors ) )
raise Exception ( " Cannot wait MySQL Client container " )
2024-02-13 12:15:46 +00:00
def wait_mysql57_to_start ( self , timeout = 180 ) :
self . mysql57_ip = self . get_instance_ip ( " mysql57 " )
2018-08-22 15:42:27 +00:00
start = time . time ( )
2021-02-16 07:10:01 +00:00
errors = [ ]
2018-08-22 15:42:27 +00:00
while time . time ( ) - start < timeout :
try :
2021-03-05 13:39:51 +00:00
conn = pymysql . connect (
2022-04-04 18:49:30 +00:00
user = mysql_user ,
password = mysql_pass ,
2024-02-13 12:15:46 +00:00
host = self . mysql57_ip ,
port = self . mysql57_port ,
2021-03-05 13:39:51 +00:00
)
2018-08-22 15:42:27 +00:00
conn . close ( )
2021-02-18 21:21:50 +00:00
logging . debug ( " Mysql Started " )
2018-08-22 15:42:27 +00:00
return
2018-08-27 14:45:37 +00:00
except Exception as ex :
2021-02-16 07:10:01 +00:00
errors + = [ str ( ex ) ]
2018-08-22 15:42:27 +00:00
time . sleep ( 0.5 )
2022-07-07 20:19:15 +00:00
run_and_check ( [ " docker " , " ps " , " --all " ] )
2021-02-16 07:10:01 +00:00
logging . error ( " Can ' t connect to MySQL: {} " . format ( errors ) )
2018-08-22 15:42:27 +00:00
raise Exception ( " Cannot wait MySQL container " )
2021-04-13 16:22:10 +00:00
def wait_mysql8_to_start ( self , timeout = 180 ) :
2021-03-19 11:44:03 +00:00
self . mysql8_ip = self . get_instance_ip ( " mysql80 " )
2021-02-16 07:10:01 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2021-03-19 11:44:03 +00:00
conn = pymysql . connect (
2022-04-04 18:49:30 +00:00
user = mysql8_user ,
password = mysql8_pass ,
2021-03-19 11:44:03 +00:00
host = self . mysql8_ip ,
port = self . mysql8_port ,
)
2021-02-16 07:10:01 +00:00
conn . close ( )
2021-02-18 21:21:50 +00:00
logging . debug ( " Mysql 8 Started " )
2021-02-16 07:10:01 +00:00
return
except Exception as ex :
2021-02-18 21:21:50 +00:00
logging . debug ( " Can ' t connect to MySQL 8 " + str ( ex ) )
2021-02-16 07:10:01 +00:00
time . sleep ( 0.5 )
2022-07-07 20:19:15 +00:00
run_and_check ( [ " docker " , " ps " , " --all " ] )
2021-02-16 07:10:01 +00:00
raise Exception ( " Cannot wait MySQL 8 container " )
2021-04-13 10:52:22 +00:00
def wait_mysql_cluster_to_start ( self , timeout = 180 ) :
2021-04-08 09:30:24 +00:00
self . mysql2_ip = self . get_instance_ip ( self . mysql2_host )
self . mysql3_ip = self . get_instance_ip ( self . mysql3_host )
self . mysql4_ip = self . get_instance_ip ( self . mysql4_host )
start = time . time ( )
errors = [ ]
while time . time ( ) - start < timeout :
try :
for ip in [ self . mysql2_ip , self . mysql3_ip , self . mysql4_ip ] :
conn = pymysql . connect (
2022-04-04 18:49:30 +00:00
user = mysql_user ,
password = mysql_pass ,
2021-04-08 09:30:24 +00:00
host = ip ,
2024-02-13 12:15:46 +00:00
port = self . mysql8_port ,
2021-04-08 09:30:24 +00:00
)
conn . close ( )
logging . debug ( f " Mysql Started { ip } " )
2018-10-15 14:49:23 +00:00
return
except Exception as ex :
2021-04-08 09:30:24 +00:00
errors + = [ str ( ex ) ]
2018-10-15 14:49:23 +00:00
time . sleep ( 0.5 )
2022-07-07 20:19:15 +00:00
run_and_check ( [ " docker " , " ps " , " --all " ] )
2021-04-08 09:30:24 +00:00
logging . error ( " Can ' t connect to MySQL: {} " . format ( errors ) )
raise Exception ( " Cannot wait MySQL container " )
2021-07-04 14:56:31 +00:00
def wait_postgres_to_start ( self , timeout = 260 ) :
2021-03-19 12:33:14 +00:00
self . postgres_ip = self . get_instance_ip ( self . postgres_host )
2021-04-08 14:43:57 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2021-07-16 13:26:35 +00:00
self . postgres_conn = psycopg2 . connect (
host = self . postgres_ip ,
port = self . postgres_port ,
2022-04-04 18:49:30 +00:00
database = pg_db ,
user = pg_user ,
password = pg_pass ,
2021-07-16 13:26:35 +00:00
)
2021-07-12 08:32:20 +00:00
self . postgres_conn . set_isolation_level ( ISOLATION_LEVEL_AUTOCOMMIT )
self . postgres_conn . autocommit = True
2021-04-08 14:43:57 +00:00
logging . debug ( " Postgres Started " )
return
except Exception as ex :
logging . debug ( " Can ' t connect to Postgres " + str ( ex ) )
time . sleep ( 0.5 )
raise Exception ( " Cannot wait Postgres container " )
2021-04-13 10:52:22 +00:00
def wait_postgres_cluster_to_start ( self , timeout = 180 ) :
2021-03-19 12:33:14 +00:00
self . postgres2_ip = self . get_instance_ip ( self . postgres2_host )
2021-04-08 09:30:24 +00:00
self . postgres3_ip = self . get_instance_ip ( self . postgres3_host )
self . postgres4_ip = self . get_instance_ip ( self . postgres4_host )
2018-10-15 14:49:23 +00:00
start = time . time ( )
2021-07-12 08:32:20 +00:00
while time . time ( ) - start < timeout :
try :
2021-07-16 13:26:35 +00:00
self . postgres2_conn = psycopg2 . connect (
host = self . postgres2_ip ,
port = self . postgres_port ,
2022-04-04 18:49:30 +00:00
database = pg_db ,
user = pg_user ,
password = pg_pass ,
2021-07-16 13:26:35 +00:00
)
2021-07-12 08:32:20 +00:00
self . postgres2_conn . set_isolation_level ( ISOLATION_LEVEL_AUTOCOMMIT )
self . postgres2_conn . autocommit = True
logging . debug ( " Postgres Cluster host 2 started " )
break
except Exception as ex :
logging . debug ( " Can ' t connect to Postgres host 2 " + str ( ex ) )
time . sleep ( 0.5 )
while time . time ( ) - start < timeout :
try :
2021-07-16 13:26:35 +00:00
self . postgres3_conn = psycopg2 . connect (
host = self . postgres3_ip ,
port = self . postgres_port ,
2022-04-04 18:49:30 +00:00
database = pg_db ,
user = pg_user ,
password = pg_pass ,
2021-07-16 13:26:35 +00:00
)
2021-07-12 08:32:20 +00:00
self . postgres3_conn . set_isolation_level ( ISOLATION_LEVEL_AUTOCOMMIT )
self . postgres3_conn . autocommit = True
logging . debug ( " Postgres Cluster host 3 started " )
break
except Exception as ex :
logging . debug ( " Can ' t connect to Postgres host 3 " + str ( ex ) )
time . sleep ( 0.5 )
while time . time ( ) - start < timeout :
try :
2021-07-16 13:26:35 +00:00
self . postgres4_conn = psycopg2 . connect (
host = self . postgres4_ip ,
port = self . postgres_port ,
2022-04-04 18:49:30 +00:00
database = pg_db ,
user = pg_user ,
password = pg_pass ,
2021-07-16 13:26:35 +00:00
)
2021-07-12 08:32:20 +00:00
self . postgres4_conn . set_isolation_level ( ISOLATION_LEVEL_AUTOCOMMIT )
self . postgres4_conn . autocommit = True
logging . debug ( " Postgres Cluster host 4 started " )
return
except Exception as ex :
logging . debug ( " Can ' t connect to Postgres host 4 " + str ( ex ) )
time . sleep ( 0.5 )
2018-10-15 14:49:23 +00:00
raise Exception ( " Cannot wait Postgres container " )
2023-01-03 19:30:14 +00:00
def wait_postgresql_java_client ( self , timeout = 180 ) :
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2023-01-03 21:28:54 +00:00
if check_postgresql_java_client_is_available (
self . postgresql_java_client_docker_id
) :
2023-01-03 19:30:14 +00:00
logging . debug ( " PostgreSQL Java Client is available " )
return True
time . sleep ( 0.5 )
except Exception as ex :
logging . debug ( " Can ' t find PostgreSQL Java Client " + str ( ex ) )
time . sleep ( 0.5 )
raise Exception ( " Cannot wait PostgreSQL Java Client container " )
2024-08-04 07:06:50 +00:00
def wait_rabbitmq_to_start ( self , timeout = 60 ) :
2023-10-09 18:33:03 +00:00
self . print_all_docker_pieces ( )
2021-04-30 09:18:12 +00:00
self . rabbitmq_ip = self . get_instance_ip ( self . rabbitmq_host )
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2023-10-09 19:10:41 +00:00
if check_rabbitmq_is_available (
self . rabbitmq_docker_id , self . rabbitmq_cookie
) :
2021-04-30 09:18:12 +00:00
logging . debug ( " RabbitMQ is available " )
2023-10-09 19:10:41 +00:00
if enable_consistent_hash_plugin (
self . rabbitmq_docker_id , self . rabbitmq_cookie
) :
2021-04-30 09:18:12 +00:00
logging . debug ( " RabbitMQ consistent hash plugin is available " )
2023-10-09 18:33:03 +00:00
return True
2021-04-30 09:18:12 +00:00
time . sleep ( 0.5 )
except Exception as ex :
logging . debug ( " Can ' t connect to RabbitMQ " + str ( ex ) )
time . sleep ( 0.5 )
2023-10-09 18:33:03 +00:00
try :
with open ( os . path . join ( self . rabbitmq_dir , " docker.log " ) , " w+ " ) as f :
subprocess . check_call ( # STYLE_CHECK_ALLOW_SUBPROCESS_CHECK_CALL
self . base_rabbitmq_cmd + [ " logs " ] , stdout = f
)
rabbitmq_debuginfo ( self . rabbitmq_docker_id , self . rabbitmq_cookie )
except Exception as e :
2024-08-04 07:07:05 +00:00
logging . debug ( f " Unable to get logs from docker: { e } . " )
2023-10-09 18:33:03 +00:00
raise Exception ( " Cannot wait RabbitMQ container " )
2021-04-30 09:18:12 +00:00
2022-07-24 14:13:17 +00:00
def wait_nats_is_available ( self , max_retries = 5 ) :
2022-05-17 13:58:09 +00:00
retries = 0
while True :
2022-07-24 14:13:17 +00:00
if asyncio . run (
check_nats_is_available ( self . nats_port , ssl_ctx = self . nats_ssl_context )
) :
2022-05-17 13:58:09 +00:00
break
else :
retries + = 1
if retries > max_retries :
raise Exception ( " NATS is not available " )
logging . debug ( " Waiting for NATS to start up " )
time . sleep ( 1 )
2021-05-21 13:29:43 +00:00
def wait_zookeeper_secure_to_start ( self , timeout = 20 ) :
logging . debug ( " Wait ZooKeeper Secure to start " )
start = time . time ( )
while time . time ( ) - start < timeout :
try :
for instance in [ " zoo1 " , " zoo2 " , " zoo3 " ] :
2021-05-24 08:23:04 +00:00
conn = self . get_kazoo_client ( instance )
2021-05-21 13:29:43 +00:00
conn . get_children ( " / " )
2021-06-01 14:18:35 +00:00
conn . stop ( )
2021-05-21 13:29:43 +00:00
logging . debug ( " All instances of ZooKeeper Secure started " )
return
except Exception as ex :
logging . debug ( " Can ' t connect to ZooKeeper secure " + str ( ex ) )
time . sleep ( 0.5 )
raise Exception ( " Cannot wait ZooKeeper secure container " )
2021-04-13 10:52:22 +00:00
def wait_zookeeper_to_start ( self , timeout = 180 ) :
2021-05-21 13:29:43 +00:00
logging . debug ( " Wait ZooKeeper to start " )
2018-08-27 13:42:39 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
for instance in [ " zoo1 " , " zoo2 " , " zoo3 " ] :
2021-05-24 08:23:04 +00:00
conn = self . get_kazoo_client ( instance )
2018-08-27 13:42:39 +00:00
conn . get_children ( " / " )
2021-06-01 14:18:35 +00:00
conn . stop ( )
2021-02-18 21:21:50 +00:00
logging . debug ( " All instances of ZooKeeper started " )
2018-08-27 13:42:39 +00:00
return
2018-08-27 14:45:37 +00:00
except Exception as ex :
2022-11-15 16:16:04 +00:00
logging . debug ( f " Can ' t connect to ZooKeeper { instance } : { ex } " )
2018-08-27 13:42:39 +00:00
time . sleep ( 0.5 )
2022-11-15 16:40:01 +00:00
raise Exception (
" Cannot wait ZooKeeper container (probably it ' s a `iptables-nft` issue, you may try to `sudo iptables -P FORWARD ACCEPT`) "
)
2018-01-25 18:14:37 +00:00
2021-04-13 16:22:10 +00:00
def make_hdfs_api ( self , timeout = 180 , kerberized = False ) :
2020-09-10 10:02:46 +00:00
if kerberized :
keytab = p . abspath (
p . join ( self . instances [ " node1 " ] . path , " secrets/clickhouse.keytab " )
2022-03-22 16:39:58 +00:00
)
2020-10-30 19:40:16 +00:00
krb_conf = p . abspath (
p . join ( self . instances [ " node1 " ] . path , " secrets/krb_long.conf " )
2022-03-22 16:39:58 +00:00
)
2021-06-09 09:23:02 +00:00
self . hdfs_kerberized_ip = self . get_instance_ip ( self . hdfs_kerberized_host )
2020-09-28 17:20:04 +00:00
kdc_ip = self . get_instance_ip ( " hdfskerberos " )
2022-03-22 16:39:58 +00:00
2021-06-09 09:23:02 +00:00
self . hdfs_api = HDFSApi (
user = " root " ,
timeout = timeout ,
kerberized = True ,
principal = " root@TEST.CLICKHOUSE.TECH " ,
keytab = keytab ,
krb_conf = krb_conf ,
host = self . hdfs_kerberized_host ,
protocol = " http " ,
proxy_port = self . hdfs_kerberized_name_port ,
data_port = self . hdfs_kerberized_data_port ,
hdfs_ip = self . hdfs_kerberized_ip ,
2021-06-13 12:56:22 +00:00
kdc_ip = kdc_ip ,
)
2020-09-10 10:02:46 +00:00
else :
2021-06-09 09:23:02 +00:00
self . hdfs_ip = self . get_instance_ip ( self . hdfs_host )
self . hdfs_api = HDFSApi (
user = " root " ,
host = self . hdfs_host ,
data_port = self . hdfs_data_port ,
proxy_port = self . hdfs_name_port ,
hdfs_ip = self . hdfs_ip ,
)
2020-09-10 10:02:46 +00:00
2021-02-24 11:46:58 +00:00
def wait_kafka_is_available ( self , kafka_docker_id , kafka_port , max_retries = 50 ) :
retries = 0
while True :
if check_kafka_is_available ( kafka_docker_id , kafka_port ) :
2023-01-17 16:25:05 +00:00
return
2021-02-24 11:46:58 +00:00
else :
retries + = 1
if retries > max_retries :
2023-01-17 16:25:05 +00:00
break
2021-02-24 11:46:58 +00:00
logging . debug ( " Waiting for Kafka to start up " )
time . sleep ( 1 )
2023-01-17 16:25:05 +00:00
try :
with open ( os . path . join ( self . kafka_dir , " docker.log " ) , " w+ " ) as f :
subprocess . check_call ( # STYLE_CHECK_ALLOW_SUBPROCESS_CHECK_CALL
self . base_kafka_cmd + [ " logs " ] , stdout = f
)
except Exception as e :
logging . debug ( " Unable to get logs from docker. " )
raise Exception ( " Kafka is not available " )
2022-12-15 11:50:23 +00:00
def wait_kerberos_kdc_is_available ( self , kerberos_kdc_docker_id , max_retries = 50 ) :
retries = 0
while True :
if check_kerberos_kdc_is_available ( kerberos_kdc_docker_id ) :
break
else :
retries + = 1
if retries > max_retries :
raise Exception ( " Kerberos KDC is not available " )
logging . debug ( " Waiting for Kerberos KDC to start up " )
time . sleep ( 1 )
2021-07-01 23:12:31 +00:00
def wait_hdfs_to_start ( self , timeout = 300 , check_marker = False ) :
2020-10-30 19:40:16 +00:00
start = time . time ( )
2018-12-05 13:24:45 +00:00
while time . time ( ) - start < timeout :
try :
2021-06-09 09:23:02 +00:00
self . hdfs_api . write_data ( " /somefilewithrandomname222 " , " 1 " )
2021-02-18 21:21:50 +00:00
logging . debug ( " Connected to HDFS and SafeMode disabled! " )
2021-07-01 23:12:31 +00:00
if check_marker :
self . hdfs_api . read_data ( " /preparations_done_marker " )
2018-12-05 13:24:45 +00:00
return
except Exception as ex :
2021-07-01 23:12:31 +00:00
logging . exception (
" Can ' t connect to HDFS or preparations are not done yet " + str ( ex )
2022-03-22 16:39:58 +00:00
)
2018-12-05 13:24:45 +00:00
time . sleep ( 1 )
raise Exception ( " Can ' t wait HDFS to start " )
2021-07-27 15:54:13 +00:00
def wait_mongo_to_start ( self , timeout = 30 , secure = False ) :
2019-02-25 10:45:22 +00:00
connection_str = " mongodb:// {user} : {password} @ {host} : {port} " . format (
2022-04-04 18:49:30 +00:00
host = " localhost " , port = self . mongo_port , user = mongo_user , password = mongo_pass
2021-02-12 15:51:21 +00:00
)
2021-07-27 15:54:13 +00:00
if secure :
connection_str + = " /?tls=true&tlsAllowInvalidCertificates=true "
2019-02-25 10:45:22 +00:00
connection = pymongo . MongoClient ( connection_str )
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2020-09-22 13:52:56 +00:00
connection . list_database_names ( )
2021-05-26 12:06:13 +00:00
logging . debug ( f " Connected to Mongo dbs: { connection . database_names ( ) } " )
2019-02-25 10:45:22 +00:00
return
except Exception as ex :
2021-02-18 21:21:50 +00:00
logging . debug ( " Can ' t connect to Mongo " + str ( ex ) )
2019-02-25 10:45:22 +00:00
time . sleep ( 1 )
2021-04-13 10:52:22 +00:00
def wait_minio_to_start ( self , timeout = 180 , secure = False ) :
2021-04-14 11:21:40 +00:00
self . minio_ip = self . get_instance_ip ( self . minio_host )
2021-04-29 11:57:48 +00:00
self . minio_redirect_ip = self . get_instance_ip ( self . minio_redirect_host )
2021-02-24 18:24:16 +00:00
os . environ [ " SSL_CERT_FILE " ] = p . join (
self . base_dir , self . minio_dir , " certs " , " public.crt "
)
2021-04-14 11:21:40 +00:00
minio_client = Minio (
f " { self . minio_ip } : { self . minio_port } " ,
2022-04-04 18:49:30 +00:00
access_key = minio_access_key ,
secret_key = minio_secret_key ,
2021-04-29 11:57:48 +00:00
secure = secure ,
http_client = urllib3 . PoolManager ( cert_reqs = " CERT_NONE " ) ,
) # disable SSL check as we test ClickHouse and not Python library
2019-11-20 11:56:38 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2020-05-22 11:48:08 +00:00
minio_client . list_buckets ( )
2021-02-18 21:21:50 +00:00
logging . debug ( " Connected to Minio. " )
2020-05-22 11:48:08 +00:00
2021-01-12 17:18:40 +00:00
buckets = [ self . minio_bucket , self . minio_bucket_2 ]
2020-05-22 11:48:08 +00:00
2021-01-12 17:18:40 +00:00
for bucket in buckets :
if minio_client . bucket_exists ( bucket ) :
2021-04-30 12:22:37 +00:00
delete_object_list = map (
2021-06-21 08:02:27 +00:00
lambda x : x . object_name ,
minio_client . list_objects_v2 ( bucket , recursive = True ) ,
2021-04-30 12:22:37 +00:00
)
errors = minio_client . remove_objects ( bucket , delete_object_list )
for error in errors :
logging . error ( f " Error occured when deleting object { error } " )
2021-01-12 17:18:40 +00:00
minio_client . remove_bucket ( bucket )
minio_client . make_bucket ( bucket )
2021-02-18 21:21:50 +00:00
logging . debug ( " S3 bucket ' %s ' created " , bucket )
2020-05-22 11:48:08 +00:00
2019-11-20 11:56:38 +00:00
self . minio_client = minio_client
return
except Exception as ex :
2021-02-18 21:21:50 +00:00
logging . debug ( " Can ' t connect to Minio: %s " , str ( ex ) )
2019-11-20 11:56:38 +00:00
time . sleep ( 1 )
2022-07-07 20:19:15 +00:00
try :
with open ( os . path . join ( self . minio_dir , " docker.log " ) , " w+ " ) as f :
subprocess . check_call ( # STYLE_CHECK_ALLOW_SUBPROCESS_CHECK_CALL
self . base_minio_cmd + [ " logs " ] , stdout = f
)
except Exception as e :
logging . debug ( " Unable to get logs from docker. " )
2020-05-22 11:48:08 +00:00
raise Exception ( " Can ' t wait Minio to start " )
2021-11-26 14:04:53 +00:00
def wait_azurite_to_start ( self , timeout = 180 ) :
from azure . storage . blob import BlobServiceClient
2022-03-22 16:39:58 +00:00
2023-10-23 12:28:02 +00:00
connection_string = (
f " DefaultEndpointsProtocol=http;AccountName=devstoreaccount1; "
f " AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==; "
f " BlobEndpoint=http://127.0.0.1: { self . env_variables [ ' AZURITE_PORT ' ] } /devstoreaccount1; "
2023-10-18 20:39:41 +00:00
)
2021-12-03 15:53:23 +00:00
time . sleep ( 1 )
2021-11-26 14:04:53 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
blob_service_client = BlobServiceClient . from_connection_string (
connection_string
)
logging . debug ( blob_service_client . get_account_information ( ) )
2024-08-14 16:13:17 +00:00
containers = [
c
for c in blob_service_client . list_containers (
name_starts_with = self . azurite_container
)
if c . name == self . azurite_container
]
2024-08-14 14:15:05 +00:00
if len ( containers ) > 0 :
for c in containers :
blob_service_client . delete_container ( c )
2024-08-14 16:13:17 +00:00
container_client = blob_service_client . get_container_client (
self . azurite_container
)
2024-08-14 14:15:05 +00:00
if container_client . exists ( ) :
2024-08-14 16:13:17 +00:00
logging . debug (
f " azurite container ' { self . azurite_container } ' exist, deleting all blobs "
)
2024-08-14 14:15:05 +00:00
for b in container_client . list_blobs ( ) :
container_client . delete_blob ( b . name )
else :
2024-08-14 16:13:17 +00:00
logging . debug (
f " azurite container ' { self . azurite_container } ' doesn ' t exist, creating it "
)
2024-08-14 14:15:05 +00:00
container_client . create_container ( )
2021-12-07 15:55:25 +00:00
self . blob_service_client = blob_service_client
2021-11-26 14:04:53 +00:00
return
except Exception as ex :
logging . debug ( " Can ' t connect to Azurite: %s " , str ( ex ) )
time . sleep ( 1 )
2021-12-02 12:28:28 +00:00
raise Exception ( " Can ' t wait Azurite to start " )
2021-04-13 10:52:22 +00:00
def wait_schema_registry_to_start ( self , timeout = 180 ) :
2023-05-08 07:08:04 +00:00
for port in self . schema_registry_port , self . schema_registry_auth_port :
2023-05-08 20:49:43 +00:00
reg_url = " http://localhost: {} " . format ( port )
arg = { " url " : reg_url }
2023-05-08 07:08:04 +00:00
sr_client = CachedSchemaRegistryClient ( arg )
start = time . time ( )
sr_started = False
sr_auth_started = False
while time . time ( ) - start < timeout :
try :
sr_client . _send_request ( sr_client . url )
logging . debug ( " Connected to SchemaRegistry " )
# don't care about possible auth errors
sr_started = True
break
except Exception as ex :
logging . debug ( ( " Can ' t connect to SchemaRegistry: %s " , str ( ex ) ) )
time . sleep ( 1 )
2023-05-02 12:31:00 +00:00
2023-05-08 07:08:04 +00:00
if not sr_started :
raise Exception ( " Can ' t wait Schema Registry to start " )
2021-04-07 12:22:53 +00:00
2021-04-13 10:52:22 +00:00
def wait_cassandra_to_start ( self , timeout = 180 ) :
2021-04-07 12:22:53 +00:00
self . cassandra_ip = self . get_instance_ip ( self . cassandra_host )
cass_client = cassandra . cluster . Cluster (
[ self . cassandra_ip ] ,
port = self . cassandra_port ,
load_balancing_policy = RoundRobinPolicy ( ) ,
)
2020-05-26 19:21:18 +00:00
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2021-04-07 12:22:53 +00:00
logging . info (
f " Check Cassandra Online { self . cassandra_id } { self . cassandra_ip } { self . cassandra_port } "
)
check = self . exec_in_container (
self . cassandra_id ,
[
" bash " ,
" -c " ,
f " /opt/cassandra/bin/cqlsh -u cassandra -p cassandra -e ' describe keyspaces ' { self . cassandra_ip } { self . cassandra_port } " ,
] ,
user = " root " ,
)
logging . info ( " Cassandra Online " )
2020-05-27 20:13:25 +00:00
cass_client . connect ( )
2021-04-07 12:22:53 +00:00
logging . info ( " Connected Clients to Cassandra " )
2020-05-26 19:21:18 +00:00
return
except Exception as ex :
2020-05-27 20:13:25 +00:00
logging . warning ( " Can ' t connect to Cassandra: %s " , str ( ex ) )
2020-05-26 19:21:18 +00:00
time . sleep ( 1 )
2021-04-07 12:22:53 +00:00
raise Exception ( " Can ' t wait Cassandra to start " )
2023-09-29 07:16:14 +00:00
def wait_ldap_to_start ( self , timeout = 180 ) :
self . ldap_container = self . get_docker_handle ( self . ldap_id )
start = time . time ( )
while time . time ( ) - start < timeout :
try :
2023-11-10 03:28:21 +00:00
logging . info ( f " Check LDAP Online { self . ldap_host } { self . ldap_port } " )
2023-09-29 07:16:14 +00:00
self . exec_in_container (
self . ldap_id ,
[
" bash " ,
" -c " ,
2024-06-29 17:00:33 +00:00
f " /opt/bitnami/openldap/bin/ldapsearch -x -H ldap:// { self . ldap_host } : { self . ldap_port } -D cn=admin,dc=example,dc=org -w clickhouse -b dc=example,dc=org "
2024-06-27 19:30:45 +00:00
f ' | grep -c -E " member: cn=j(ohn|ane)doe " '
2024-06-29 17:00:33 +00:00
f " | grep 2 >> /dev/null " ,
2023-09-29 07:16:14 +00:00
] ,
user = " root " ,
)
logging . info ( " LDAP Online " )
return
except Exception as ex :
logging . warning ( " Can ' t connect to LDAP: %s " , str ( ex ) )
time . sleep ( 1 )
raise Exception ( " Can ' t wait LDAP to start " )
2022-07-07 20:19:15 +00:00
def start ( self ) :
2021-07-01 14:41:59 +00:00
pytest_xdist_logging_to_separate_files . setup ( )
logging . info ( " Running tests in {} " . format ( self . base_path ) )
2023-01-03 19:30:14 +00:00
if not os . path . exists ( self . instances_dir ) :
os . mkdir ( self . instances_dir )
else :
2023-01-03 21:28:54 +00:00
logging . warning (
" Instance directory already exists. Did you call cluster.start() for second time? "
)
2022-07-07 20:19:15 +00:00
logging . debug ( f " Cluster start called. is_up= { self . is_up } " )
self . print_all_docker_pieces ( )
2017-05-19 18:54:05 +00:00
if self . is_up :
return
2017-07-27 22:29:48 +00:00
try :
2021-04-27 09:07:21 +00:00
self . cleanup ( )
except Exception as e :
logging . warning ( " Cleanup failed: {e} " )
2021-03-05 13:39:51 +00:00
2020-02-29 11:14:36 +00:00
try :
2020-10-02 16:54:07 +00:00
for instance in list ( self . instances . values ( ) ) :
2022-07-07 20:19:15 +00:00
logging . debug ( f " Setup directory for instance: { instance . name } " )
instance . create_dir ( )
2020-02-29 11:14:36 +00:00
2021-05-20 16:12:04 +00:00
_create_env_file ( os . path . join ( self . env_file ) , self . env_variables )
2021-06-01 08:43:44 +00:00
self . docker_client = docker . DockerClient (
base_url = " unix:///var/run/docker.sock " ,
version = self . docker_api_version ,
timeout = 600 ,
)
2020-02-29 11:14:36 +00:00
2021-10-26 14:17:51 +00:00
common_opts = [ " --verbose " , " up " , " -d " ]
2020-02-29 11:14:36 +00:00
2021-12-27 09:24:19 +00:00
images_pull_cmd = self . base_cmd + [ " pull " ]
# sometimes dockerhub/proxy can be flaky
2024-07-24 10:44:06 +00:00
2024-07-31 12:00:09 +00:00
def logging_pulling_images ( * * kwargs ) :
if " exception " in kwargs :
logging . info (
" Got exception pulling images: %s " , kwargs [ " exception " ]
)
retry ( log_function = logging_pulling_images ) ( run_and_check ) ( images_pull_cmd )
2021-12-27 09:24:19 +00:00
2021-05-21 13:29:43 +00:00
if self . with_zookeeper_secure and self . base_zookeeper_cmd :
logging . debug ( " Setup ZooKeeper Secure " )
logging . debug (
f " Creating internal ZooKeeper dirs: { self . zookeeper_dirs_to_create } "
)
for i in range ( 1 , 3 ) :
if os . path . exists ( self . zookeeper_instance_dir_prefix + f " { i } " ) :
shutil . rmtree ( self . zookeeper_instance_dir_prefix + f " { i } " )
for dir in self . zookeeper_dirs_to_create :
os . makedirs ( dir )
run_and_check ( self . base_zookeeper_cmd + common_opts , env = self . env )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-05-21 13:29:43 +00:00
self . wait_zookeeper_secure_to_start ( )
for command in self . pre_zookeeper_commands :
2021-05-24 08:23:04 +00:00
self . run_kazoo_commands_with_retries ( command , repeats = 5 )
2021-05-21 13:29:43 +00:00
2020-02-29 11:14:36 +00:00
if self . with_zookeeper and self . base_zookeeper_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup ZooKeeper " )
2021-05-20 16:12:04 +00:00
logging . debug (
f " Creating internal ZooKeeper dirs: { self . zookeeper_dirs_to_create } "
)
2021-05-20 18:19:45 +00:00
if self . use_keeper :
for i in range ( 1 , 4 ) :
if os . path . exists ( self . keeper_instance_dir_prefix + f " { i } " ) :
shutil . rmtree ( self . keeper_instance_dir_prefix + f " { i } " )
else :
for i in range ( 1 , 3 ) :
if os . path . exists ( self . zookeeper_instance_dir_prefix + f " { i } " ) :
shutil . rmtree ( self . zookeeper_instance_dir_prefix + f " { i } " )
2021-05-20 16:12:04 +00:00
for dir in self . zookeeper_dirs_to_create :
os . makedirs ( dir )
2021-06-13 12:56:22 +00:00
2021-05-20 16:12:04 +00:00
if self . use_keeper : # TODO: remove hardcoded paths from here
for i in range ( 1 , 4 ) :
shutil . copy (
2023-12-01 19:39:38 +00:00
os . path . join (
self . keeper_config_dir , f " keeper_config { i } .xml "
) ,
2021-05-20 16:12:04 +00:00
os . path . join (
self . keeper_instance_dir_prefix + f " { i } " , " config "
2022-03-22 16:39:58 +00:00
) ,
2021-05-20 16:12:04 +00:00
)
2021-05-14 12:30:49 +00:00
run_and_check ( self . base_zookeeper_cmd + common_opts , env = self . env )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-05-20 16:12:04 +00:00
2021-05-21 13:29:43 +00:00
self . wait_zookeeper_to_start ( )
2020-02-29 11:14:36 +00:00
for command in self . pre_zookeeper_commands :
self . run_kazoo_commands_with_retries ( command , repeats = 5 )
2021-04-13 14:55:31 +00:00
if self . with_mysql_client and self . base_mysql_client_cmd :
logging . debug ( " Setup MySQL Client " )
subprocess_check_call ( self . base_mysql_client_cmd + common_opts )
self . wait_mysql_client_to_start ( )
2024-02-13 12:15:46 +00:00
if self . with_mysql57 and self . base_mysql57_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup MySQL " )
2024-02-13 12:15:46 +00:00
if os . path . exists ( self . mysql57_dir ) :
shutil . rmtree ( self . mysql57_dir )
os . makedirs ( self . mysql57_logs_dir )
os . chmod ( self . mysql57_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
subprocess_check_call ( self . base_mysql57_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2024-02-13 12:15:46 +00:00
self . wait_mysql57_to_start ( )
2020-02-29 11:14:36 +00:00
2021-02-16 07:10:01 +00:00
if self . with_mysql8 and self . base_mysql8_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup MySQL 8 " )
2021-03-09 07:32:10 +00:00
if os . path . exists ( self . mysql8_dir ) :
shutil . rmtree ( self . mysql8_dir )
2021-03-05 13:39:51 +00:00
os . makedirs ( self . mysql8_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . mysql8_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-02-16 07:10:01 +00:00
subprocess_check_call ( self . base_mysql8_cmd + common_opts )
2021-04-13 10:52:22 +00:00
self . wait_mysql8_to_start ( )
2021-02-16 07:10:01 +00:00
2021-03-31 17:04:57 +00:00
if self . with_mysql_cluster and self . base_mysql_cluster_cmd :
print ( " Setup MySQL " )
2021-04-08 09:30:24 +00:00
if os . path . exists ( self . mysql_cluster_dir ) :
shutil . rmtree ( self . mysql_cluster_dir )
2024-02-13 12:15:46 +00:00
os . makedirs ( self . mysql_cluster_logs_dir , exist_ok = True )
2021-12-21 10:50:54 +00:00
os . chmod ( self . mysql_cluster_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-04-08 09:30:24 +00:00
2021-03-31 17:04:57 +00:00
subprocess_check_call ( self . base_mysql_cluster_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-04-13 10:52:22 +00:00
self . wait_mysql_cluster_to_start ( )
2021-03-31 17:04:57 +00:00
2020-02-29 11:14:36 +00:00
if self . with_postgres and self . base_postgres_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup Postgres " )
2021-03-19 16:44:08 +00:00
if os . path . exists ( self . postgres_dir ) :
shutil . rmtree ( self . postgres_dir )
os . makedirs ( self . postgres_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . postgres_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-03-19 16:44:08 +00:00
2020-02-29 11:14:36 +00:00
subprocess_check_call ( self . base_postgres_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-04-13 10:52:22 +00:00
self . wait_postgres_to_start ( )
2020-02-29 11:14:36 +00:00
2021-03-31 17:04:57 +00:00
if self . with_postgres_cluster and self . base_postgres_cluster_cmd :
2023-01-03 19:30:14 +00:00
logging . debug ( " Setup Postgres " )
2021-04-08 14:43:57 +00:00
os . makedirs ( self . postgres2_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . postgres2_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-04-08 14:43:57 +00:00
os . makedirs ( self . postgres3_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . postgres3_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-04-08 14:43:57 +00:00
os . makedirs ( self . postgres4_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . postgres4_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-03-31 17:04:57 +00:00
subprocess_check_call ( self . base_postgres_cluster_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-04-13 10:52:22 +00:00
self . wait_postgres_cluster_to_start ( )
2021-03-31 17:04:57 +00:00
2023-01-03 21:28:54 +00:00
if (
self . with_postgresql_java_client
and self . base_postgresql_java_client_cmd
) :
2023-01-03 19:30:14 +00:00
logging . debug ( " Setup Postgres Java Client " )
2023-01-03 21:28:54 +00:00
subprocess_check_call (
self . base_postgresql_java_client_cmd + common_opts
)
2023-01-03 19:30:14 +00:00
self . up_called = True
self . wait_postgresql_java_client ( )
2020-02-29 11:14:36 +00:00
if self . with_kafka and self . base_kafka_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup Kafka " )
2023-01-17 16:25:05 +00:00
os . mkdir ( self . kafka_dir )
2020-02-29 11:14:36 +00:00
subprocess_check_call (
self . base_kafka_cmd + common_opts + [ " --renew-anon-volumes " ]
)
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-02-24 11:46:58 +00:00
self . wait_kafka_is_available ( self . kafka_docker_id , self . kafka_port )
2021-04-13 10:52:22 +00:00
self . wait_schema_registry_to_start ( )
2020-02-29 11:14:36 +00:00
2020-09-29 08:56:37 +00:00
if self . with_kerberized_kafka and self . base_kerberized_kafka_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup kerberized kafka " )
2023-01-17 16:25:05 +00:00
os . mkdir ( self . kafka_dir )
2021-02-12 15:51:21 +00:00
run_and_check (
self . base_kerberized_kafka_cmd
+ common_opts
+ [ " --renew-anon-volumes " ]
)
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-02-24 11:46:58 +00:00
self . wait_kafka_is_available (
self . kerberized_kafka_docker_id , self . kerberized_kafka_port , 100
)
2021-02-18 21:21:50 +00:00
2022-12-19 11:28:49 +00:00
if self . with_kerberos_kdc and self . base_kerberos_kdc_cmd :
2022-12-14 13:39:23 +00:00
logging . debug ( " Setup Kerberos KDC " )
run_and_check (
2022-12-19 11:28:49 +00:00
self . base_kerberos_kdc_cmd + common_opts + [ " --renew-anon-volumes " ]
2022-12-14 13:39:23 +00:00
)
self . up_called = True
2022-12-15 11:50:23 +00:00
self . wait_kerberos_kdc_is_available ( self . keberos_kdc_docker_id )
2022-12-14 13:39:23 +00:00
2020-05-20 06:22:12 +00:00
if self . with_rabbitmq and self . base_rabbitmq_cmd :
2021-05-17 11:16:16 +00:00
logging . debug ( " Setup RabbitMQ " )
os . makedirs ( self . rabbitmq_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . rabbitmq_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-07-27 06:16:40 +00:00
2023-10-09 18:33:03 +00:00
with open ( self . rabbitmq_cookie_file , " w " ) as f :
f . write ( self . rabbitmq_cookie )
os . chmod ( self . rabbitmq_cookie_file , stat . S_IRUSR )
subprocess_check_call (
self . base_rabbitmq_cmd + common_opts + [ " --renew-anon-volumes " ]
)
self . up_called = True
self . rabbitmq_docker_id = self . get_instance_docker_id ( " rabbitmq1 " )
time . sleep ( 2 )
logging . debug ( f " RabbitMQ checking container try " )
self . wait_rabbitmq_to_start ( )
2020-05-20 06:22:12 +00:00
2022-05-17 13:58:09 +00:00
if self . with_nats and self . base_nats_cmd :
logging . debug ( " Setup NATS " )
2022-07-24 14:13:17 +00:00
os . makedirs ( self . nats_cert_dir )
env = os . environ . copy ( )
env [ " NATS_CERT_DIR " ] = self . nats_cert_dir
run_and_check (
p . join ( self . base_dir , " nats_certs.sh " ) ,
env = env ,
detach = False ,
nothrow = False ,
)
self . nats_ssl_context = ssl . create_default_context ( )
self . nats_ssl_context . load_verify_locations (
p . join ( self . nats_cert_dir , " ca " , " ca-cert.pem " )
)
2022-05-17 13:58:09 +00:00
subprocess_check_call ( self . base_nats_cmd + common_opts )
self . nats_docker_id = self . get_instance_docker_id ( " nats1 " )
self . up_called = True
2022-07-24 14:13:17 +00:00
self . wait_nats_is_available ( )
2022-05-17 13:58:09 +00:00
2020-02-29 11:14:36 +00:00
if self . with_hdfs and self . base_hdfs_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup HDFS " )
os . makedirs ( self . hdfs_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . hdfs_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2020-02-29 11:14:36 +00:00
subprocess_check_call ( self . base_hdfs_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-06-09 09:23:02 +00:00
self . make_hdfs_api ( )
self . wait_hdfs_to_start ( )
2020-02-29 11:14:36 +00:00
2020-09-10 10:02:46 +00:00
if self . with_kerberized_hdfs and self . base_kerberized_hdfs_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup kerberized HDFS " )
2021-02-19 14:42:43 +00:00
os . makedirs ( self . hdfs_kerberized_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . hdfs_kerberized_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-02-12 15:51:21 +00:00
run_and_check ( self . base_kerberized_hdfs_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-06-09 09:23:02 +00:00
self . make_hdfs_api ( kerberized = True )
2021-07-01 23:12:31 +00:00
self . wait_hdfs_to_start ( check_marker = True )
2020-09-10 10:02:46 +00:00
2021-06-13 12:56:22 +00:00
if self . with_nginx and self . base_nginx_cmd :
logging . debug ( " Setup nginx " )
subprocess_check_call (
self . base_nginx_cmd + common_opts + [ " --renew-anon-volumes " ]
)
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-06-13 12:56:22 +00:00
self . nginx_docker_id = self . get_instance_docker_id ( " nginx " )
2020-02-29 11:14:36 +00:00
if self . with_mongo and self . base_mongo_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup Mongo " )
2021-01-22 14:27:23 +00:00
run_and_check ( self . base_mongo_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-07-27 15:54:13 +00:00
self . wait_mongo_to_start ( 30 , secure = self . with_mongo_secure )
2020-02-29 11:14:36 +00:00
2022-07-20 17:09:38 +00:00
if self . with_coredns and self . base_coredns_cmd :
logging . debug ( " Setup coredns " )
run_and_check ( self . base_coredns_cmd + common_opts )
self . up_called = True
time . sleep ( 10 )
2020-02-29 11:14:36 +00:00
if self . with_redis and self . base_redis_cmd :
2021-02-18 21:21:50 +00:00
logging . debug ( " Setup Redis " )
2021-02-12 15:51:21 +00:00
subprocess_check_call ( self . base_redis_cmd + common_opts )
2021-09-24 10:46:55 +00:00
self . up_called = True
2020-02-29 11:14:36 +00:00
time . sleep ( 10 )
2021-11-18 08:17:49 +00:00
if self . with_hive and self . base_hive_cmd :
logging . debug ( " Setup hive " )
subprocess_check_call ( self . base_hive_cmd + common_opts )
self . up_called = True
2022-06-21 07:59:54 +00:00
time . sleep ( 30 )
2021-11-18 08:17:49 +00:00
2020-02-29 11:14:36 +00:00
if self . with_minio and self . base_minio_cmd :
2021-02-19 12:58:11 +00:00
# Copy minio certificates to minio/certs
os . mkdir ( self . minio_dir )
if self . minio_certs_dir is None :
os . mkdir ( os . path . join ( self . minio_dir , " certs " ) )
else :
2021-02-24 11:46:58 +00:00
shutil . copytree (
os . path . join ( self . base_dir , self . minio_certs_dir ) ,
os . path . join ( self . minio_dir , " certs " ) ,
)
2023-01-03 19:30:14 +00:00
os . mkdir ( self . minio_data_dir )
os . chmod ( self . minio_data_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-02-19 12:58:11 +00:00
2020-02-29 11:14:36 +00:00
minio_start_cmd = self . base_minio_cmd + common_opts
2020-07-10 19:42:18 +00:00
2020-02-29 11:14:36 +00:00
logging . info (
" Trying to create Minio instance by command %s " ,
" " . join ( map ( str , minio_start_cmd ) ) ,
)
2021-02-12 15:51:21 +00:00
run_and_check ( minio_start_cmd )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-02-12 15:51:21 +00:00
logging . info ( " Trying to connect to Minio... " )
self . wait_minio_to_start ( secure = self . minio_certs_dir is not None )
2020-02-29 11:14:36 +00:00
2021-11-26 14:04:53 +00:00
if self . with_azurite and self . base_azurite_cmd :
azurite_start_cmd = self . base_azurite_cmd + common_opts
logging . info (
" Trying to create Azurite instance by command %s " ,
" " . join ( map ( str , azurite_start_cmd ) ) ,
)
2024-07-31 14:16:40 +00:00
2024-07-31 12:00:09 +00:00
def logging_azurite_initialization ( exception , retry_number , sleep_time ) :
logging . info (
2024-07-24 10:44:06 +00:00
f " Azurite initialization failed with error: { exception } "
2024-07-31 12:00:09 +00:00
)
retry (
log_function = logging_azurite_initialization ,
) (
run_and_check
) ( azurite_start_cmd )
2021-11-26 14:04:53 +00:00
self . up_called = True
logging . info ( " Trying to connect to Azurite " )
self . wait_azurite_to_start ( )
2020-05-19 02:21:27 +00:00
if self . with_cassandra and self . base_cassandra_cmd :
2021-04-05 04:40:43 +00:00
subprocess_check_call ( self . base_cassandra_cmd + [ " up " , " -d " ] )
2021-09-24 10:46:55 +00:00
self . up_called = True
2020-05-26 19:21:18 +00:00
self . wait_cassandra_to_start ( )
2020-05-19 02:21:27 +00:00
2023-09-29 07:16:14 +00:00
if self . with_ldap and self . base_ldap_cmd :
2023-11-09 20:24:52 +00:00
ldap_start_cmd = self . base_ldap_cmd + common_opts
subprocess_check_call ( ldap_start_cmd )
2023-09-29 07:16:14 +00:00
self . up_called = True
self . wait_ldap_to_start ( )
2021-06-07 12:56:29 +00:00
if self . with_jdbc_bridge and self . base_jdbc_bridge_cmd :
2021-07-27 07:33:49 +00:00
os . makedirs ( self . jdbc_driver_logs_dir )
2021-12-21 10:50:54 +00:00
os . chmod ( self . jdbc_driver_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2021-07-27 07:33:49 +00:00
2021-06-07 12:56:29 +00:00
subprocess_check_call ( self . base_jdbc_bridge_cmd + [ " up " , " -d " ] )
2021-09-24 10:46:55 +00:00
self . up_called = True
2021-07-27 07:33:49 +00:00
self . jdbc_bridge_ip = self . get_instance_ip ( self . jdbc_bridge_host )
self . wait_for_url (
f " http:// { self . jdbc_bridge_ip } : { self . jdbc_bridge_port } /ping "
)
2022-03-22 16:39:58 +00:00
2024-03-11 22:58:05 +00:00
if self . with_prometheus :
os . makedirs ( self . prometheus_writer_logs_dir )
os . chmod ( self . prometheus_writer_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
os . makedirs ( self . prometheus_reader_logs_dir )
os . chmod ( self . prometheus_reader_logs_dir , stat . S_IRWXU | stat . S_IRWXO )
2020-02-29 11:14:36 +00:00
clickhouse_start_cmd = self . base_cmd + [ " up " , " -d " , " --no-recreate " ]
2021-02-18 21:21:50 +00:00
logging . debug (
2022-03-22 16:39:58 +00:00
(
2021-02-18 21:21:50 +00:00
" Trying to create ClickHouse instance by command %s " ,
" " . join ( map ( str , clickhouse_start_cmd ) ) ,
)
2022-03-22 16:39:58 +00:00
)
2021-02-17 15:40:01 +00:00
self . up_called = True
2021-05-20 16:12:04 +00:00
run_and_check ( clickhouse_start_cmd )
2021-02-18 21:21:50 +00:00
logging . debug ( " ClickHouse instance created " )
2020-08-12 08:55:04 +00:00
2021-06-04 10:14:32 +00:00
start_timeout = 300.0 # seconds
2020-10-02 16:54:07 +00:00
for instance in self . instances . values ( ) :
2020-02-29 11:14:36 +00:00
instance . docker_client = self . docker_client
instance . ip_address = self . get_instance_ip ( instance . name )
2021-06-30 11:16:37 +00:00
logging . debug (
f " Waiting for ClickHouse start in { instance . name } , ip: { instance . ip_address } ... "
)
2021-05-18 07:38:32 +00:00
instance . wait_for_start ( start_timeout )
2021-06-30 11:16:37 +00:00
logging . debug ( f " ClickHouse { instance . name } started " )
2020-02-29 11:14:36 +00:00
instance . client = Client (
instance . ip_address , command = self . client_bin_path
)
self . is_up = True
2020-03-26 14:43:22 +00:00
2020-10-02 16:54:07 +00:00
except BaseException as e :
2021-02-18 21:21:50 +00:00
logging . debug ( " Failed to start cluster: " )
logging . debug ( str ( e ) )
logging . debug ( traceback . print_exc ( ) )
2021-05-17 11:16:16 +00:00
self . shutdown ( )
2020-02-29 11:14:36 +00:00
raise
2017-05-19 18:54:05 +00:00
2021-10-08 09:16:23 +00:00
def shutdown ( self , kill = True , ignore_fatal = True ) :
2019-11-14 16:00:02 +00:00
sanitizer_assert_instance = None
2021-03-03 08:09:44 +00:00
fatal_log = None
2021-08-19 19:13:12 +00:00
2021-02-17 15:40:01 +00:00
if self . up_called :
with open ( self . docker_logs_path , " w+ " ) as f :
try :
subprocess . check_call ( # STYLE_CHECK_ALLOW_SUBPROCESS_CHECK_CALL
self . base_cmd + [ " logs " ] , stdout = f
)
except Exception as e :
2021-02-18 21:21:50 +00:00
logging . debug ( " Unable to get logs from docker. " )
2021-02-17 15:40:01 +00:00
f . seek ( 0 )
for line in f :
if SANITIZER_SIGN in line :
sanitizer_assert_instance = line . split ( " | " ) [ 0 ] . strip ( )
break
if kill :
try :
2021-05-20 16:12:04 +00:00
run_and_check ( self . base_cmd + [ " stop " , " --timeout " , " 20 " ] )
2021-02-17 15:40:01 +00:00
except Exception as e :
2021-02-18 21:21:50 +00:00
logging . debug (
" Kill command failed during shutdown. {} " . format ( repr ( e ) )
2022-03-22 16:39:58 +00:00
)
2021-02-18 21:21:50 +00:00
logging . debug ( " Trying to kill forcefully " )
2021-05-20 16:12:04 +00:00
run_and_check ( self . base_cmd + [ " kill " ] )
2021-02-17 15:40:01 +00:00
2021-08-19 19:13:12 +00:00
# Check server logs for Fatal messages and sanitizer failures.
# NOTE: we cannot do this via docker since in case of Fatal message container may already die.
for name , instance in self . instances . items ( ) :
2022-10-20 12:09:31 +00:00
if instance . contains_in_log (
SANITIZER_SIGN , from_host = True , filename = " stderr.log "
) :
2021-10-19 10:19:43 +00:00
sanitizer_assert_instance = instance . grep_in_log (
2023-02-04 10:15:40 +00:00
SANITIZER_SIGN ,
from_host = True ,
filename = " stderr.log " ,
after = 1000 ,
2021-10-19 10:19:43 +00:00
)
2021-08-19 19:13:12 +00:00
logging . error (
" Sanitizer in instance %s log %s " ,
name ,
sanitizer_assert_instance ,
)
2022-03-22 16:39:58 +00:00
2021-10-08 09:16:23 +00:00
if not ignore_fatal and instance . contains_in_log (
" Fatal " , from_host = True
) :
2021-08-19 19:13:12 +00:00
fatal_log = instance . grep_in_log ( " Fatal " , from_host = True )
2021-09-21 14:29:05 +00:00
if " Child process was terminated by signal 9 (KILL) " in fatal_log :
fatal_log = None
continue
2021-08-19 19:13:12 +00:00
logging . error ( " Crash in instance %s fatal log %s " , name , fatal_log )
2020-08-12 08:55:04 +00:00
try :
2021-02-17 15:40:01 +00:00
subprocess_check_call ( self . base_cmd + [ " down " , " --volumes " ] )
2020-08-12 08:55:04 +00:00
except Exception as e :
2021-10-26 14:17:51 +00:00
logging . debug (
" Down + remove orphans failed during shutdown. {} " . format ( repr ( e ) )
2022-03-22 16:39:58 +00:00
)
2021-09-24 10:46:55 +00:00
else :
logging . warning (
" docker-compose up was not called. Trying to export docker.log for running containers "
)
2019-11-20 11:56:38 +00:00
2021-04-27 15:34:33 +00:00
self . cleanup ( )
2019-11-20 11:56:38 +00:00
2017-05-19 18:54:05 +00:00
self . is_up = False
2017-05-23 17:13:36 +00:00
self . docker_client = None
2020-10-02 16:54:07 +00:00
for instance in list ( self . instances . values ( ) ) :
2017-05-23 17:13:36 +00:00
instance . docker_client = None
2017-05-19 18:54:05 +00:00
instance . ip_address = None
instance . client = None
2019-11-14 16:00:02 +00:00
if sanitizer_assert_instance is not None :
2020-09-16 04:26:10 +00:00
raise Exception (
2023-02-04 10:16:37 +00:00
" Sanitizer assert found for instance {} " . format (
sanitizer_assert_instance
2020-09-16 04:26:10 +00:00
)
2022-03-22 16:39:58 +00:00
)
2021-08-19 19:13:12 +00:00
if fatal_log is not None :
raise Exception ( " Fatal messages found: {} " . format ( fatal_log ) )
2019-11-14 16:00:02 +00:00
2020-03-26 14:43:22 +00:00
def pause_container ( self , instance_name ) :
subprocess_check_call ( self . base_cmd + [ " pause " , instance_name ] )
2020-09-16 04:26:10 +00:00
2020-03-26 14:43:22 +00:00
def unpause_container ( self , instance_name ) :
subprocess_check_call ( self . base_cmd + [ " unpause " , instance_name ] )
2020-09-16 04:26:10 +00:00
2019-06-18 07:20:14 +00:00
def open_bash_shell ( self , instance_name ) :
os . system ( " " . join ( self . base_cmd + [ " exec " , instance_name , " /bin/bash " ] ) )
2017-05-19 18:54:05 +00:00
2021-05-24 08:23:04 +00:00
def get_kazoo_client ( self , zoo_instance_name ) :
use_ssl = False
if self . with_zookeeper_secure :
port = self . zookeeper_secure_port
use_ssl = True
elif self . with_zookeeper :
port = self . zookeeper_port
else :
raise Exception ( " Cluster has no ZooKeeper " )
2021-05-20 16:12:04 +00:00
ip = self . get_instance_ip ( zoo_instance_name )
2021-05-21 13:29:43 +00:00
logging . debug (
f " get_kazoo_client: { zoo_instance_name } , ip: { ip } , port: { port } , use_ssl: { use_ssl } "
)
zk = KazooClient (
hosts = f " { ip } : { port } " ,
use_ssl = use_ssl ,
verify_certs = False ,
certfile = self . zookeeper_certfile ,
keyfile = self . zookeeper_keyfile ,
)
2018-01-25 18:14:37 +00:00
zk . start ( )
return zk
2021-05-24 08:23:04 +00:00
def run_kazoo_commands_with_retries (
self , kazoo_callback , zoo_instance_name = " zoo1 " , repeats = 1 , sleep_for = 1
) :
2021-06-01 14:18:35 +00:00
zk = self . get_kazoo_client ( zoo_instance_name )
2021-05-24 08:23:04 +00:00
logging . debug (
f " run_kazoo_commands_with_retries: { zoo_instance_name } , { kazoo_callback } "
)
2018-01-25 18:14:37 +00:00
for i in range ( repeats - 1 ) :
2017-08-30 16:25:34 +00:00
try :
2021-06-01 14:18:35 +00:00
kazoo_callback ( zk )
2018-01-25 18:14:37 +00:00
return
except KazooException as e :
2021-02-18 21:21:50 +00:00
logging . debug ( repr ( e ) )
2017-08-30 16:25:34 +00:00
time . sleep ( sleep_for )
2021-06-01 14:18:35 +00:00
kazoo_callback ( zk )
zk . stop ( )
2018-01-25 18:14:37 +00:00
2017-08-30 16:25:34 +00:00
def add_zookeeper_startup_command ( self , command ) :
2018-01-09 19:12:43 +00:00
self . pre_zookeeper_commands . append ( command )
2017-08-30 16:25:34 +00:00
2020-09-10 04:00:33 +00:00
def stop_zookeeper_nodes ( self , zk_nodes ) :
for n in zk_nodes :
logging . info ( " Stopping zookeeper node: %s " , n )
subprocess_check_call ( self . base_zookeeper_cmd + [ " stop " , n ] )
def start_zookeeper_nodes ( self , zk_nodes ) :
for n in zk_nodes :
logging . info ( " Starting zookeeper node: %s " , n )
subprocess_check_call ( self . base_zookeeper_cmd + [ " start " , n ] )
2017-08-30 16:25:34 +00:00
2017-05-19 18:54:05 +00:00
DOCKER_COMPOSE_TEMPLATE = """
2020-05-19 15:27:10 +00:00
version : ' 2.3 '
2017-05-19 18:54:05 +00:00
services :
{ name } :
2020-09-01 06:38:23 +00:00
image : { image } : { tag }
2017-08-02 14:42:35 +00:00
hostname : { hostname }
2017-05-19 18:54:05 +00:00
volumes :
2020-08-12 08:55:04 +00:00
- { instance_config_dir } : / etc / clickhouse - server /
2017-05-19 18:54:05 +00:00
- { db_dir } : / var / lib / clickhouse /
- { logs_dir } : / var / log / clickhouse - server /
2020-09-29 08:56:37 +00:00
- / etc / passwd : / etc / passwd : ro
2019-06-20 16:25:32 +00:00
{ binary_volume }
{ odbc_bridge_volume }
2021-03-11 17:48:47 +00:00
{ library_bridge_volume }
2021-11-09 08:34:37 +00:00
{ external_dirs_volumes }
2018-08-22 15:42:27 +00:00
{ odbc_ini_path }
2020-09-29 08:56:37 +00:00
{ keytab_path }
{ krb5_conf }
2018-11-22 15:59:00 +00:00
entrypoint : { entrypoint_cmd }
2019-05-22 16:40:30 +00:00
tmpfs : { tmpfs }
2024-03-12 22:31:33 +00:00
{ mem_limit }
2018-12-27 19:42:25 +00:00
cap_add :
- SYS_PTRACE
2020-10-27 12:24:10 +00:00
- NET_ADMIN
2021-02-23 10:18:24 +00:00
- IPC_LOCK
- SYS_NICE
2022-11-30 15:47:12 +00:00
# for umount/mount on fly
- SYS_ADMIN
2017-05-19 18:54:05 +00:00
depends_on : { depends_on }
2019-02-21 17:34:19 +00:00
user : ' {user} '
2018-07-28 14:38:08 +00:00
env_file :
- { env_file }
2019-02-21 17:34:19 +00:00
security_opt :
- label : disable
2020-08-12 08:55:04 +00:00
dns_opt :
2020-09-02 08:07:46 +00:00
- attempts : 2
2020-08-12 08:55:04 +00:00
- timeout : 1
2020-09-02 08:07:46 +00:00
- inet6
- rotate
2018-12-27 15:55:51 +00:00
{ networks }
{ app_net }
{ ipv4_address }
{ ipv6_address }
2019-12-18 13:42:39 +00:00
{ net_aliases }
{ net_alias1 }
2022-03-22 16:39:58 +00:00
"""
2017-05-19 18:54:05 +00:00
2019-06-04 20:59:31 +00:00
2017-05-19 18:54:05 +00:00
class ClickHouseInstance :
def __init__ (
2020-09-16 04:26:10 +00:00
self ,
cluster ,
base_path ,
name ,
base_config_dir ,
custom_main_configs ,
custom_user_configs ,
custom_dictionaries ,
2021-04-13 14:55:31 +00:00
macros ,
with_zookeeper ,
zookeeper_config_path ,
with_mysql_client ,
2024-02-13 12:15:46 +00:00
with_mysql57 ,
2021-04-13 14:55:31 +00:00
with_mysql8 ,
with_mysql_cluster ,
with_kafka ,
with_kerberized_kafka ,
2022-12-14 13:39:23 +00:00
with_kerberos_kdc ,
2021-12-14 08:06:30 +00:00
with_rabbitmq ,
2022-05-17 13:58:09 +00:00
with_nats ,
2021-12-14 08:06:30 +00:00
with_nginx ,
with_kerberized_hdfs ,
2023-04-29 21:23:55 +00:00
with_secrets ,
2021-12-14 08:06:30 +00:00
with_mongo ,
with_redis ,
with_minio ,
with_azurite ,
with_jdbc_bridge ,
with_hive ,
2022-07-20 17:09:38 +00:00
with_coredns ,
2021-04-13 14:55:31 +00:00
with_cassandra ,
2023-09-29 07:16:14 +00:00
with_ldap ,
2024-03-20 12:43:18 +00:00
use_old_analyzer ,
2021-04-13 14:55:31 +00:00
server_bin_path ,
odbc_bridge_bin_path ,
library_bridge_bin_path ,
clickhouse_path_dir ,
with_odbc_drivers ,
with_postgres ,
with_postgres_cluster ,
2023-01-03 19:30:14 +00:00
with_postgresql_java_client ,
2021-05-21 18:56:22 +00:00
clickhouse_start_command = CLICKHOUSE_START_COMMAND ,
main_config_name = " config.xml " ,
users_config_name = " users.xml " ,
copy_common_configs = True ,
2020-09-16 04:26:10 +00:00
hostname = None ,
env_variables = None ,
2023-10-03 16:53:38 +00:00
instance_env_variables = False ,
2021-09-08 10:03:54 +00:00
image = " clickhouse/integration-test " ,
tag = " latest " ,
2021-11-19 01:58:34 +00:00
stay_alive = False ,
ipv4_address = None ,
ipv6_address = None ,
with_installed_binary = False ,
external_dirs = None ,
tmpfs = None ,
2024-03-12 22:31:33 +00:00
mem_limit = None ,
2021-11-19 01:58:34 +00:00
config_root_name = " clickhouse " ,
extra_configs = [ ] ,
) :
2017-05-19 18:54:05 +00:00
self . name = name
2020-04-11 10:57:13 +00:00
self . base_cmd = cluster . base_cmd
2017-05-30 11:49:17 +00:00
self . docker_id = cluster . get_instance_docker_id ( self . name )
self . cluster = cluster
2017-08-02 14:42:35 +00:00
self . hostname = hostname if hostname is not None else self . name
2017-05-30 11:49:17 +00:00
2021-11-09 08:34:37 +00:00
self . external_dirs = external_dirs
2020-04-11 10:57:13 +00:00
self . tmpfs = tmpfs or [ ]
2024-03-12 22:31:33 +00:00
if mem_limit is not None :
self . mem_limit = " mem_limit : " + mem_limit
else :
self . mem_limit = " "
2020-08-12 08:55:04 +00:00
self . base_config_dir = (
p . abspath ( p . join ( base_path , base_config_dir ) ) if base_config_dir else None
2022-03-22 16:39:58 +00:00
)
2017-05-30 11:49:17 +00:00
self . custom_main_config_paths = [
p . abspath ( p . join ( base_path , c ) ) for c in custom_main_configs
]
self . custom_user_config_paths = [
p . abspath ( p . join ( base_path , c ) ) for c in custom_user_configs
]
2020-08-12 08:55:04 +00:00
self . custom_dictionaries_paths = [
p . abspath ( p . join ( base_path , c ) ) for c in custom_dictionaries
]
2021-11-19 01:58:34 +00:00
self . custom_extra_config_paths = [
p . abspath ( p . join ( base_path , c ) ) for c in extra_configs
]
2017-06-15 20:08:26 +00:00
self . clickhouse_path_dir = (
p . abspath ( p . join ( base_path , clickhouse_path_dir ) )
if clickhouse_path_dir
else None
2022-03-22 16:39:58 +00:00
)
2023-04-29 21:23:55 +00:00
self . secrets_dir = p . abspath ( p . join ( base_path , " secrets " ) )
2018-07-25 16:00:51 +00:00
self . macros = macros if macros is not None else { }
2017-05-19 18:54:05 +00:00
self . with_zookeeper = with_zookeeper
2017-08-30 16:25:34 +00:00
self . zookeeper_config_path = zookeeper_config_path
2017-05-19 18:54:05 +00:00
self . server_bin_path = server_bin_path
2019-01-29 17:17:31 +00:00
self . odbc_bridge_bin_path = odbc_bridge_bin_path
2021-03-11 17:48:47 +00:00
self . library_bridge_bin_path = library_bridge_bin_path
2017-05-19 18:54:05 +00:00
2021-04-13 14:55:31 +00:00
self . with_mysql_client = with_mysql_client
2024-02-13 12:15:46 +00:00
self . with_mysql57 = with_mysql57
2021-02-16 07:10:01 +00:00
self . with_mysql8 = with_mysql8
2021-03-31 17:04:57 +00:00
self . with_mysql_cluster = with_mysql_cluster
2021-04-13 14:55:31 +00:00
self . with_postgres = with_postgres
self . with_postgres_cluster = with_postgres_cluster
2023-01-03 19:30:14 +00:00
self . with_postgresql_java_client = with_postgresql_java_client
2018-07-18 05:22:01 +00:00
self . with_kafka = with_kafka
2020-09-29 08:56:37 +00:00
self . with_kerberized_kafka = with_kerberized_kafka
2022-12-14 13:39:23 +00:00
self . with_kerberos_kdc = with_kerberos_kdc
2020-05-20 06:22:12 +00:00
self . with_rabbitmq = with_rabbitmq
2022-05-17 13:58:09 +00:00
self . with_nats = with_nats
2021-06-13 12:56:22 +00:00
self . with_nginx = with_nginx
2020-09-10 10:02:46 +00:00
self . with_kerberized_hdfs = with_kerberized_hdfs
2023-04-29 21:23:55 +00:00
self . with_secrets = with_secrets
2019-02-25 10:45:22 +00:00
self . with_mongo = with_mongo
2019-03-21 18:10:55 +00:00
self . with_redis = with_redis
2019-11-20 11:56:38 +00:00
self . with_minio = with_minio
2021-11-26 14:04:53 +00:00
self . with_azurite = with_azurite
2020-05-19 02:21:27 +00:00
self . with_cassandra = with_cassandra
2023-09-29 07:16:14 +00:00
self . with_ldap = with_ldap
2021-06-07 12:56:29 +00:00
self . with_jdbc_bridge = with_jdbc_bridge
2021-12-14 02:41:51 +00:00
self . with_hive = with_hive
2022-07-20 17:09:38 +00:00
self . with_coredns = with_coredns
self . coredns_config_dir = p . abspath ( p . join ( base_path , " coredns_config " ) )
2024-03-20 12:43:18 +00:00
self . use_old_analyzer = use_old_analyzer
2018-05-14 11:10:07 +00:00
2021-05-21 18:56:22 +00:00
self . main_config_name = main_config_name
self . users_config_name = users_config_name
self . copy_common_configs = copy_common_configs
self . clickhouse_start_command = clickhouse_start_command . replace (
" {main_config_file} " , self . main_config_name
)
2023-07-11 02:32:33 +00:00
self . clickhouse_stay_alive_command = " bash -c \" trap ' pkill tail ' INT TERM; {} --daemon; coproc tail -f /dev/null; wait $$! \" " . format (
clickhouse_start_command
)
2021-05-21 18:56:22 +00:00
2017-07-26 12:31:55 +00:00
self . path = p . join ( self . cluster . instances_dir , name )
2021-01-27 11:26:49 +00:00
self . docker_compose_path = p . join ( self . path , " docker-compose.yml " )
2020-04-11 10:57:13 +00:00
self . env_variables = env_variables or { }
2023-10-03 16:53:38 +00:00
self . instance_env_variables = instance_env_variables
2021-05-20 16:12:04 +00:00
self . env_file = self . cluster . env_file
2018-08-22 15:42:27 +00:00
if with_odbc_drivers :
2020-08-12 08:55:04 +00:00
self . odbc_ini_path = self . path + " /odbc.ini:/etc/odbc.ini "
2024-02-13 12:15:46 +00:00
self . with_mysql8 = True
2018-08-22 15:42:27 +00:00
else :
self . odbc_ini_path = " "
2017-05-19 18:54:05 +00:00
2022-12-14 13:39:23 +00:00
if with_kerberized_kafka or with_kerberized_hdfs or with_kerberos_kdc :
2022-12-19 11:28:49 +00:00
if with_kerberos_kdc :
base_secrets_dir = self . cluster . instances_dir
else :
base_secrets_dir = os . path . dirname ( self . docker_compose_path )
self . keytab_path = " - " + base_secrets_dir + " /secrets:/tmp/keytab "
2020-10-30 19:40:16 +00:00
self . krb5_conf = (
2022-12-19 11:28:49 +00:00
" - " + base_secrets_dir + " /secrets/krb.conf:/etc/krb5.conf:ro "
2022-03-22 16:39:58 +00:00
)
2020-09-29 08:56:37 +00:00
else :
self . keytab_path = " "
self . krb5_conf = " "
2017-05-23 17:13:36 +00:00
self . docker_client = None
2017-05-19 18:54:05 +00:00
self . ip_address = None
self . client = None
2018-08-22 15:42:27 +00:00
self . image = image
2020-09-01 06:38:23 +00:00
self . tag = tag
2018-11-22 15:59:00 +00:00
self . stay_alive = stay_alive
2018-12-27 15:55:51 +00:00
self . ipv4_address = ipv4_address
self . ipv6_address = ipv6_address
2019-06-20 16:25:32 +00:00
self . with_installed_binary = with_installed_binary
2021-03-05 13:39:51 +00:00
self . is_up = False
2021-09-22 15:00:08 +00:00
self . config_root_name = config_root_name
2021-02-12 15:51:21 +00:00
2021-04-25 02:25:46 +00:00
def is_built_with_sanitizer ( self , sanitizer_name = " " ) :
2020-06-22 13:10:25 +00:00
build_opts = self . query (
" SELECT value FROM system.build_options WHERE name = ' CXX_FLAGS ' "
)
2021-04-25 02:25:46 +00:00
return " -fsanitize= {} " . format ( sanitizer_name ) in build_opts
2021-08-19 11:32:32 +00:00
def is_debug_build ( self ) :
build_opts = self . query (
" SELECT value FROM system.build_options WHERE name = ' CXX_FLAGS ' "
)
return " NDEBUG " not in build_opts
2021-04-25 02:25:46 +00:00
def is_built_with_thread_sanitizer ( self ) :
2021-04-25 03:09:42 +00:00
return self . is_built_with_sanitizer ( " thread " )
2020-06-22 13:10:25 +00:00
2020-11-23 15:18:09 +00:00
def is_built_with_address_sanitizer ( self ) :
2021-04-25 03:09:42 +00:00
return self . is_built_with_sanitizer ( " address " )
2021-04-25 02:25:46 +00:00
def is_built_with_memory_sanitizer ( self ) :
2021-04-25 03:09:42 +00:00
return self . is_built_with_sanitizer ( " memory " )
2020-11-23 15:18:09 +00:00
2017-08-14 01:29:19 +00:00
# Connects to the instance via clickhouse-client, sends a query (1st argument) and returns the answer
2021-09-28 18:27:32 +00:00
def query (
self ,
sql ,
stdin = None ,
timeout = None ,
settings = None ,
user = None ,
password = None ,
database = None ,
2022-06-07 06:28:33 +00:00
host = None ,
2021-09-28 18:27:32 +00:00
ignore_error = False ,
query_id = None ,
) :
2022-07-12 18:12:59 +00:00
sql_for_log = " "
2022-07-12 12:05:59 +00:00
if len ( sql ) > 1000 :
sql_for_log = sql [ : 1000 ]
else :
sql_for_log = sql
logging . debug ( " Executing query %s on %s " , sql_for_log , self . name )
2021-09-28 18:27:32 +00:00
return self . client . query (
sql ,
stdin = stdin ,
timeout = timeout ,
settings = settings ,
user = user ,
password = password ,
database = database ,
ignore_error = ignore_error ,
query_id = query_id ,
2022-06-07 06:28:33 +00:00
host = host ,
2021-09-28 18:27:32 +00:00
)
2022-03-22 16:39:58 +00:00
2020-09-16 04:26:10 +00:00
def query_with_retry (
self ,
sql ,
stdin = None ,
timeout = None ,
settings = None ,
user = None ,
password = None ,
database = None ,
2022-06-07 06:28:33 +00:00
host = None ,
2020-09-16 04:26:10 +00:00
ignore_error = False ,
2019-11-20 11:56:38 +00:00
retry_count = 20 ,
sleep_time = 0.5 ,
check_callback = lambda x : True ,
) :
2023-02-13 23:53:15 +00:00
# logging.debug(f"Executing query {sql} on {self.name}")
2018-09-03 14:06:00 +00:00
result = None
2024-03-23 13:56:26 +00:00
exception_msg = " "
2018-09-03 14:06:00 +00:00
for i in range ( retry_count ) :
try :
2020-09-16 04:26:10 +00:00
result = self . query (
sql ,
stdin = stdin ,
timeout = timeout ,
settings = settings ,
user = user ,
password = password ,
database = database ,
2022-06-07 06:28:33 +00:00
host = host ,
2020-09-16 04:26:10 +00:00
ignore_error = ignore_error ,
)
2018-09-03 14:06:00 +00:00
if check_callback ( result ) :
return result
time . sleep ( sleep_time )
2024-03-03 19:38:34 +00:00
except QueryRuntimeException as ex :
2024-03-23 13:56:26 +00:00
exception_msg = f " { type ( ex ) . __name__ } : { str ( ex ) } "
2024-03-03 19:38:34 +00:00
# Container is down, this is likely due to server crash.
if " No route to host " in str ( ex ) :
raise
time . sleep ( sleep_time )
2018-09-03 14:06:00 +00:00
except Exception as ex :
2023-02-13 23:53:15 +00:00
# logging.debug("Retry {} got exception {}".format(i + 1, ex))
2024-03-23 13:56:26 +00:00
exception_msg = f " { type ( ex ) . __name__ } : { str ( ex ) } "
2018-09-03 14:06:00 +00:00
time . sleep ( sleep_time )
if result is not None :
return result
2024-03-23 13:56:26 +00:00
raise Exception ( f " Can ' t execute query { sql } \n { exception_msg } " )
2017-05-30 11:49:17 +00:00
2017-07-26 12:31:55 +00:00
# As query() but doesn't wait response and returns response handler
2021-07-01 14:41:59 +00:00
def get_query_request ( self , sql , * args , * * kwargs ) :
logging . debug ( f " Executing query { sql } on { self . name } " )
return self . client . get_query_request ( sql , * args , * * kwargs )
2017-05-30 11:49:17 +00:00
2019-04-07 00:31:20 +00:00
# Connects to the instance via clickhouse-client, sends a query (1st argument), expects an error and return its code
2020-09-16 04:26:10 +00:00
def query_and_get_error (
self ,
sql ,
stdin = None ,
timeout = None ,
settings = None ,
user = None ,
password = None ,
database = None ,
2023-06-03 18:59:04 +00:00
query_id = None ,
2020-09-16 04:26:10 +00:00
) :
2021-07-01 14:41:59 +00:00
logging . debug ( f " Executing query { sql } on { self . name } " )
2020-09-16 04:26:10 +00:00
return self . client . query_and_get_error (
sql ,
stdin = stdin ,
timeout = timeout ,
settings = settings ,
user = user ,
password = password ,
database = database ,
2023-06-03 18:59:04 +00:00
query_id = query_id ,
2020-09-16 04:26:10 +00:00
)
2019-04-07 00:31:20 +00:00
2023-01-06 11:21:35 +00:00
def query_and_get_error_with_retry (
self ,
sql ,
stdin = None ,
timeout = None ,
settings = None ,
user = None ,
password = None ,
database = None ,
retry_count = 20 ,
sleep_time = 0.5 ,
) :
logging . debug ( f " Executing query { sql } on { self . name } " )
result = None
for i in range ( retry_count ) :
try :
result = self . client . query_and_get_error (
sql ,
stdin = stdin ,
timeout = timeout ,
settings = settings ,
user = user ,
password = password ,
database = database ,
)
time . sleep ( sleep_time )
2023-07-04 09:31:15 +00:00
if result is not None :
return result
2023-01-06 11:21:35 +00:00
except QueryRuntimeException as ex :
logging . debug ( " Retry {} got exception {} " . format ( i + 1 , ex ) )
time . sleep ( sleep_time )
2023-07-04 09:31:15 +00:00
raise Exception ( " Query {} did not fail " . format ( sql ) )
2023-01-06 11:21:35 +00:00
2019-07-17 11:55:18 +00:00
# The same as query_and_get_error but ignores successful query.
2020-09-16 04:26:10 +00:00
def query_and_get_answer_with_error (
self ,
sql ,
stdin = None ,
timeout = None ,
settings = None ,
user = None ,
password = None ,
database = None ,
2023-08-23 06:05:26 +00:00
query_id = None ,
2020-09-16 04:26:10 +00:00
) :
2021-07-01 14:41:59 +00:00
logging . debug ( f " Executing query { sql } on { self . name } " )
2020-09-16 04:26:10 +00:00
return self . client . query_and_get_answer_with_error (
sql ,
stdin = stdin ,
timeout = timeout ,
settings = settings ,
user = user ,
password = password ,
database = database ,
2023-08-23 06:05:26 +00:00
query_id = query_id ,
2020-09-16 04:26:10 +00:00
)
2019-07-17 11:55:18 +00:00
2019-03-30 18:40:52 +00:00
# Connects to the instance via HTTP interface, sends a query and returns the answer
2021-10-22 07:15:34 +00:00
def http_query (
self ,
sql ,
data = None ,
2022-06-19 15:17:46 +00:00
method = None ,
2021-10-22 07:15:34 +00:00
params = None ,
user = None ,
password = None ,
2022-08-03 12:04:18 +00:00
port = 8123 ,
timeout = None ,
retry_strategy = None ,
2023-02-15 14:17:28 +00:00
content = False ,
2022-08-03 12:04:18 +00:00
) :
output , error = self . http_query_and_get_answer_with_error (
sql ,
data = data ,
method = method ,
params = params ,
user = user ,
password = password ,
port = port ,
timeout = timeout ,
retry_strategy = retry_strategy ,
2023-02-15 14:17:28 +00:00
content = content ,
2022-08-03 12:04:18 +00:00
)
if error :
raise Exception ( " ClickHouse HTTP server returned " + error )
return output
# Connects to the instance via HTTP interface, sends a query, expects an error and return the error message
def http_query_and_get_error (
self ,
sql ,
data = None ,
method = None ,
params = None ,
user = None ,
password = None ,
port = 8123 ,
timeout = None ,
retry_strategy = None ,
) :
output , error = self . http_query_and_get_answer_with_error (
sql ,
data = data ,
method = method ,
params = params ,
user = user ,
password = password ,
port = port ,
timeout = timeout ,
retry_strategy = retry_strategy ,
)
if not error :
raise Exception (
" ClickHouse HTTP server is expected to fail, but succeeded: " + output
)
return error
2023-07-14 09:08:40 +00:00
def append_hosts ( self , name , ip ) :
self . exec_in_container (
( [ " bash " , " -c " , " echo ' {} ' {} >> /etc/hosts " . format ( ip , name ) ] ) ,
privileged = True ,
user = " root " ,
)
def set_hosts ( self , hosts ) :
entries = [ " 127.0.0.1 localhost " , " ::1 localhost " ]
for host in hosts :
entries . append ( f " { host [ 0 ] } { host [ 1 ] } " )
self . exec_in_container (
[ " bash " , " -c " , ' echo -e " {} " > /etc/hosts ' . format ( " \\ n " . join ( entries ) ) ] ,
privileged = True ,
user = " root " ,
)
2022-08-03 12:04:18 +00:00
# Connects to the instance via HTTP interface, sends a query and returns both the answer and the error message
# as a tuple (output, error).
def http_query_and_get_answer_with_error (
self ,
sql ,
data = None ,
method = None ,
params = None ,
user = None ,
password = None ,
2021-10-22 07:15:34 +00:00
port = 8123 ,
timeout = None ,
retry_strategy = None ,
2023-02-15 14:17:28 +00:00
content = False ,
2021-10-22 07:15:34 +00:00
) :
2021-07-01 14:41:59 +00:00
logging . debug ( f " Executing query { sql } on { self . name } via HTTP interface " )
2019-12-17 17:20:15 +00:00
if params is None :
params = { }
else :
params = params . copy ( )
2022-01-19 18:29:31 +00:00
if sql is not None :
params [ " query " ] = sql
2019-12-17 17:20:15 +00:00
2020-10-02 16:54:07 +00:00
auth = None
2020-02-29 12:57:52 +00:00
if user and password :
2020-10-02 16:54:07 +00:00
auth = requests . auth . HTTPBasicAuth ( user , password )
2020-02-29 12:57:52 +00:00
elif user :
2020-10-02 16:54:07 +00:00
auth = requests . auth . HTTPBasicAuth ( user , " " )
2021-10-22 07:15:34 +00:00
url = f " http:// { self . ip_address } : { port } /? " + urllib . parse . urlencode ( params )
2019-12-17 17:20:15 +00:00
2021-10-22 07:15:34 +00:00
if retry_strategy is None :
requester = requests
else :
adapter = requests . adapters . HTTPAdapter ( max_retries = retry_strategy )
requester = requests . Session ( )
requester . mount ( " https:// " , adapter )
requester . mount ( " http:// " , adapter )
2022-06-18 11:56:04 +00:00
if method is None :
method = " POST " if data else " GET "
r = requester . request ( method , url , data = data , auth = auth , timeout = timeout )
2023-10-30 07:43:55 +00:00
# Force encoding to UTF-8
r . encoding = " UTF-8 "
2020-03-29 13:21:26 +00:00
2022-08-03 12:04:18 +00:00
if r . ok :
2023-02-15 14:17:28 +00:00
return ( r . content if content else r . text , None )
2019-11-08 12:37:31 +00:00
2022-08-03 12:04:18 +00:00
code = r . status_code
return ( None , str ( code ) + " " + http . client . responses [ code ] + " : " + r . text )
2020-03-29 13:21:26 +00:00
2019-11-14 02:20:06 +00:00
# Connects to the instance via HTTP interface, sends a query and returns the answer
def http_request ( self , url , method = " GET " , params = None , data = None , headers = None ) :
2021-07-01 14:41:59 +00:00
logging . debug ( f " Sending HTTP request { url } to { self . name } " )
2020-09-16 04:26:10 +00:00
url = " http:// " + self . ip_address + " :8123/ " + url
2020-04-21 11:30:45 +00:00
return requests . request (
method = method , url = url , params = params , data = data , headers = headers
)
2019-11-08 12:37:31 +00:00
2021-04-15 09:44:45 +00:00
def stop_clickhouse ( self , stop_wait_sec = 30 , kill = False ) :
2019-03-14 13:39:47 +00:00
if not self . stay_alive :
2021-01-11 17:37:08 +00:00
raise Exception (
" clickhouse can be stopped only with stay_alive=True instance "
)
2021-04-07 12:22:53 +00:00
try :
2021-10-26 14:17:51 +00:00
ps_clickhouse = self . exec_in_container (
[ " bash " , " -c " , " ps -C clickhouse " ] , nothrow = True , user = " root "
)
2021-04-07 12:22:53 +00:00
if ps_clickhouse == " PID TTY STAT TIME COMMAND " :
logging . warning ( " ClickHouse process already stopped " )
return
2021-10-01 09:52:24 +00:00
self . exec_in_container (
[ " bash " , " -c " , " pkill {} clickhouse " . format ( " -9 " if kill else " " ) ] ,
user = " root " ,
)
2021-10-13 11:55:58 +00:00
start_time = time . time ( )
2021-09-24 15:25:19 +00:00
stopped = False
2021-10-13 11:55:58 +00:00
while time . time ( ) < = start_time + stop_wait_sec :
2021-10-26 14:17:51 +00:00
pid = self . get_process_pid ( " clickhouse " )
if pid is None :
2021-09-24 15:25:19 +00:00
stopped = True
break
2021-10-26 14:17:51 +00:00
else :
time . sleep ( 1 )
2021-09-24 15:25:19 +00:00
if not stopped :
2021-10-19 10:19:43 +00:00
pid = self . get_process_pid ( " clickhouse " )
if pid is not None :
2021-10-26 14:17:51 +00:00
logging . warning (
f " Force kill clickhouse in stop_clickhouse. ps: { pid } "
)
self . exec_in_container (
2022-03-22 16:39:58 +00:00
[
2021-10-26 14:17:51 +00:00
" bash " ,
" -c " ,
f " gdb -batch -ex ' thread apply all bt full ' -p { pid } > { os . path . join ( self . path , ' logs/stdout.log ' ) } " ,
] ,
user = " root " ,
)
self . stop_clickhouse ( kill = True )
else :
ps_all = self . exec_in_container (
[ " bash " , " -c " , " ps aux " ] , nothrow = True , user = " root "
)
logging . warning (
f " We want force stop clickhouse, but no clickhouse-server is running \n { ps_all } "
)
return
2021-04-07 12:22:53 +00:00
except Exception as e :
logging . warning ( f " Stop ClickHouse raised an error { e } " )
2021-01-11 17:37:08 +00:00
2024-01-24 12:15:19 +00:00
def start_clickhouse (
self , start_wait_sec = 60 , retry_start = True , expected_to_fail = False
) :
2021-01-11 17:37:08 +00:00
if not self . stay_alive :
2021-10-26 08:28:28 +00:00
raise Exception (
" ClickHouse can be started again only with stay_alive=True instance "
)
2021-11-11 07:41:48 +00:00
start_time = time . time ( )
2021-10-26 08:28:28 +00:00
time_to_sleep = 0.5
2021-11-11 07:41:48 +00:00
while start_time + start_wait_sec > = time . time ( ) :
2021-10-26 08:28:28 +00:00
# sometimes after SIGKILL (hard reset) server may refuse to start for some time
2021-10-26 08:29:22 +00:00
# for different reasons.
2021-11-11 07:41:48 +00:00
pid = self . get_process_pid ( " clickhouse " )
if pid is None :
logging . debug ( " No clickhouse process running. Start new one. " )
self . exec_in_container (
[ " bash " , " -c " , " {} --daemon " . format ( self . clickhouse_start_command ) ] ,
user = str ( os . getuid ( ) ) ,
)
2024-01-24 12:01:32 +00:00
if expected_to_fail :
self . wait_start_failed ( start_wait_sec + start_time - time . time ( ) )
return
2021-11-11 07:41:48 +00:00
time . sleep ( 1 )
continue
else :
logging . debug ( " Clickhouse process running. " )
2024-01-24 12:01:32 +00:00
if expected_to_fail :
raise Exception ( " ClickHouse was expected not to be running. " )
2021-10-26 08:28:28 +00:00
try :
2021-11-11 07:41:48 +00:00
self . wait_start ( start_wait_sec + start_time - time . time ( ) )
return
except Exception as e :
logging . warning (
f " Current start attempt failed. Will kill { pid } just in case. "
)
2021-11-11 13:28:29 +00:00
self . exec_in_container (
[ " bash " , " -c " , f " kill -9 { pid } " ] , user = " root " , nothrow = True
)
2022-11-12 12:20:35 +00:00
if not retry_start :
raise
2021-12-09 14:40:51 +00:00
time . sleep ( time_to_sleep )
2021-01-11 17:37:08 +00:00
2021-11-11 07:41:48 +00:00
raise Exception ( " Cannot start ClickHouse, see additional info in logs " )
2021-01-11 17:37:08 +00:00
2021-11-11 07:41:48 +00:00
def wait_start ( self , start_wait_sec ) :
2021-10-26 14:17:51 +00:00
start_time = time . time ( )
last_err = None
2021-12-14 06:38:17 +00:00
while True :
2021-10-26 14:17:51 +00:00
try :
pid = self . get_process_pid ( " clickhouse " )
if pid is None :
raise Exception ( " ClickHouse server is not running. Check logs. " )
2021-11-11 07:41:48 +00:00
exec_query_with_retry ( self , " select 20 " , retry_count = 10 , silent = True )
return
2021-10-26 14:17:51 +00:00
except QueryRuntimeException as err :
last_err = err
pid = self . get_process_pid ( " clickhouse " )
if pid is not None :
logging . warning ( f " ERROR { err } " )
else :
raise Exception ( " ClickHouse server is not running. Check logs. " )
2021-12-14 06:38:17 +00:00
if time . time ( ) > start_time + start_wait_sec :
break
2021-10-26 14:17:51 +00:00
logging . error (
f " No time left to start. But process is still running. Will dump threads. "
)
2021-11-11 07:41:48 +00:00
ps_clickhouse = self . exec_in_container (
[ " bash " , " -c " , " ps -C clickhouse " ] , nothrow = True , user = " root "
)
logging . info ( f " PS RESULT: \n { ps_clickhouse } " )
2021-10-26 14:17:51 +00:00
pid = self . get_process_pid ( " clickhouse " )
if pid is not None :
self . exec_in_container (
[ " bash " , " -c " , f " gdb -batch -ex ' thread apply all bt full ' -p { pid } " ] ,
user = " root " ,
)
if last_err is not None :
raise last_err
2021-01-11 17:37:08 +00:00
2024-01-24 12:01:32 +00:00
def wait_start_failed ( self , start_wait_sec ) :
start_time = time . time ( )
while time . time ( ) < = start_time + start_wait_sec :
pid = self . get_process_pid ( " clickhouse " )
if pid is None :
return
time . sleep ( 1 )
logging . error (
f " No time left to shutdown. Process is still running. Will dump threads. "
)
ps_clickhouse = self . exec_in_container (
[ " bash " , " -c " , " ps -C clickhouse " ] , nothrow = True , user = " root "
)
logging . info ( f " PS RESULT: \n { ps_clickhouse } " )
pid = self . get_process_pid ( " clickhouse " )
if pid is not None :
self . exec_in_container (
[ " bash " , " -c " , f " gdb -batch -ex ' thread apply all bt full ' -p { pid } " ] ,
user = " root " ,
)
2024-01-24 12:15:19 +00:00
raise Exception (
" ClickHouse server is still running, but was expected to shutdown. Check logs. "
)
2024-01-24 12:01:32 +00:00
2021-10-13 11:55:58 +00:00
def restart_clickhouse ( self , stop_start_wait_sec = 60 , kill = False ) :
2021-01-11 17:37:08 +00:00
self . stop_clickhouse ( stop_start_wait_sec , kill )
self . start_clickhouse ( stop_start_wait_sec )
2017-05-19 18:54:05 +00:00
2020-08-03 14:40:02 +00:00
def exec_in_container ( self , cmd , detach = False , nothrow = False , * * kwargs ) :
2021-06-11 12:00:40 +00:00
return self . cluster . exec_in_container (
self . docker_id , cmd , detach , nothrow , * * kwargs
)
2017-08-02 14:42:35 +00:00
2021-10-06 13:08:25 +00:00
def rotate_logs ( self ) :
2021-10-07 21:42:12 +00:00
self . exec_in_container (
[ " bash " , " -c " , f " kill -HUP { self . get_process_pid ( ' clickhouse server ' ) } " ] ,
user = " root " ,
)
2021-10-06 13:08:25 +00:00
2021-10-19 10:19:43 +00:00
def contains_in_log (
2024-07-31 09:06:30 +00:00
self ,
substring ,
from_host = False ,
filename = " clickhouse-server.log " ,
exclusion_substring = " " ,
2021-10-19 10:19:43 +00:00
) :
2021-08-19 19:13:12 +00:00
if from_host :
2021-10-19 10:19:43 +00:00
# We check fist file exists but want to look for all rotated logs as well
2021-08-19 19:13:12 +00:00
result = subprocess_check_call (
2022-03-22 16:39:58 +00:00
[
2021-08-19 19:13:12 +00:00
" bash " ,
" -c " ,
2024-07-31 09:06:30 +00:00
f ' [ -f { self . logs_dir } / { filename } ] && zgrep -aH " { substring } " { self . logs_dir } / { filename } * | ( [ -z " { exclusion_substring } " ] && cat || grep -v " $ { exclusion_substring } " ) || true ' ,
2021-08-19 19:13:12 +00:00
]
)
else :
result = self . exec_in_container (
2022-03-22 16:39:58 +00:00
[
2021-08-19 19:13:12 +00:00
" bash " ,
" -c " ,
2024-07-31 09:06:30 +00:00
f ' [ -f /var/log/clickhouse-server/ { filename } ] && zgrep -aH " { substring } " /var/log/clickhouse-server/ { filename } | ( [ -z " { exclusion_substring } " ] && cat || grep -v " $ { exclusion_substring } " ) || true ' ,
2021-08-19 19:13:12 +00:00
]
)
2019-03-29 18:10:03 +00:00
return len ( result ) > 0
2023-02-04 10:15:40 +00:00
def grep_in_log (
self , substring , from_host = False , filename = " clickhouse-server.log " , after = None
) :
2021-08-19 19:13:12 +00:00
logging . debug ( f " grep in log called %s " , substring )
2023-02-04 10:15:40 +00:00
if after is not None :
after_opt = " -A {} " . format ( after )
else :
after_opt = " "
2021-08-19 19:13:12 +00:00
if from_host :
2021-10-19 10:19:43 +00:00
# We check fist file exists but want to look for all rotated logs as well
2021-08-19 19:13:12 +00:00
result = subprocess_check_call (
2022-03-22 16:39:58 +00:00
[
2021-08-19 19:13:12 +00:00
" bash " ,
" -c " ,
2023-02-04 10:15:40 +00:00
f ' [ -f { self . logs_dir } / { filename } ] && zgrep { after_opt } -a " { substring } " { self . logs_dir } / { filename } * || true ' ,
2021-08-19 19:13:12 +00:00
]
)
else :
result = self . exec_in_container (
2022-03-22 16:39:58 +00:00
[
2021-08-19 19:13:12 +00:00
" bash " ,
" -c " ,
2023-02-04 10:15:40 +00:00
f ' [ -f /var/log/clickhouse-server/ { filename } ] && zgrep { after_opt } -a " { substring } " /var/log/clickhouse-server/ { filename } * || true ' ,
2021-08-19 19:13:12 +00:00
]
)
logging . debug ( " grep result %s " , result )
2021-03-03 08:09:44 +00:00
return result
2021-03-19 16:12:33 +00:00
def count_in_log ( self , substring ) :
result = self . exec_in_container (
2021-10-10 09:55:37 +00:00
[
" bash " ,
" -c " ,
' grep -a " {} " /var/log/clickhouse-server/clickhouse-server.log | wc -l ' . format (
substring
2022-03-22 16:39:58 +00:00
) ,
]
2021-10-10 09:55:37 +00:00
)
2021-03-19 20:57:00 +00:00
return result
2021-03-19 16:12:33 +00:00
2021-02-23 16:53:14 +00:00
def wait_for_log_line (
self ,
regexp ,
filename = " /var/log/clickhouse-server/clickhouse-server.log " ,
timeout = 30 ,
repetitions = 1 ,
look_behind_lines = 100 ,
) :
start_time = time . time ( )
result = self . exec_in_container (
2021-02-24 15:08:58 +00:00
[
" bash " ,
" -c " ,
' timeout {} tail -Fn {} " {} " | grep -Em {} {} ' . format (
timeout ,
look_behind_lines ,
filename ,
repetitions ,
shlex . quote ( regexp ) ,
2022-03-22 16:39:58 +00:00
) ,
]
2021-02-24 15:08:58 +00:00
)
# if repetitions>1 grep will return success even if not enough lines were collected,
if repetitions > 1 and len ( result . splitlines ( ) ) < repetitions :
2021-03-03 08:56:15 +00:00
logging . debug (
" wait_for_log_line: those lines were found during {} seconds: " . format (
timeout
)
2022-03-22 16:39:58 +00:00
)
2021-03-03 08:56:15 +00:00
logging . debug ( result )
2021-02-24 15:08:58 +00:00
raise Exception (
" wait_for_log_line: Not enough repetitions: {} found, while {} expected " . format (
len ( result . splitlines ( ) ) , repetitions
)
2022-03-22 16:39:58 +00:00
)
2021-02-24 15:08:58 +00:00
wait_duration = time . time ( ) - start_time
2021-06-03 18:38:12 +00:00
logging . debug (
' {} log line(s) matching " {} " appeared in a {:.3f} seconds ' . format (
repetitions , regexp , wait_duration
)
2022-03-22 16:39:58 +00:00
)
2021-03-04 09:07:59 +00:00
return wait_duration
2021-02-23 16:53:14 +00:00
2021-11-29 16:07:00 +00:00
def path_exists ( self , path ) :
2020-11-17 14:36:04 +00:00
return (
self . exec_in_container (
[
" bash " ,
" -c " ,
" echo $(if [ -e ' {} ' ]; then echo ' yes ' ; else echo ' no ' ; fi) " . format (
path
2022-03-22 16:39:58 +00:00
) ,
]
2020-11-17 14:36:04 +00:00
)
== " yes \n "
2022-03-22 16:39:58 +00:00
)
2020-11-17 14:36:04 +00:00
2019-02-21 17:34:19 +00:00
def copy_file_to_container ( self , local_path , dest_path ) :
2021-06-11 12:00:40 +00:00
return self . cluster . copy_file_to_container (
self . docker_id , local_path , dest_path
)
2019-02-21 17:34:19 +00:00
2019-06-21 08:03:13 +00:00
def get_process_pid ( self , process_name ) :
2021-01-12 17:18:40 +00:00
output = self . exec_in_container (
2022-03-22 16:39:58 +00:00
[
2021-01-12 17:18:40 +00:00
" bash " ,
" -c " ,
2021-11-11 07:41:48 +00:00
" ps ax | grep ' {} ' | grep -v ' grep ' | grep -v ' coproc ' | grep -v ' bash -c ' | awk ' {{ print $1}} ' " . format (
2021-01-12 17:18:40 +00:00
process_name
2022-03-22 16:39:58 +00:00
) ,
]
2021-01-12 17:18:40 +00:00
)
2019-06-21 08:03:13 +00:00
if output :
try :
pid = int ( output . split ( " \n " ) [ 0 ] . strip ( ) )
return pid
except :
return None
return None
2021-08-20 15:58:43 +00:00
def restart_with_original_version (
2023-09-28 15:47:24 +00:00
self ,
stop_start_wait_sec = 300 ,
callback_onstop = None ,
signal = 15 ,
clear_data_dir = False ,
2021-08-20 15:58:43 +00:00
) :
2021-11-11 07:41:48 +00:00
begin_time = time . time ( )
2021-08-20 15:58:43 +00:00
if not self . stay_alive :
raise Exception ( " Cannot restart not stay alive container " )
self . exec_in_container (
[ " bash " , " -c " , " pkill - {} clickhouse " . format ( signal ) ] , user = " root "
)
retries = int ( stop_start_wait_sec / 0.5 )
local_counter = 0
# wait stop
while local_counter < retries :
if not self . get_process_pid ( " clickhouse server " ) :
break
time . sleep ( 0.5 )
local_counter + = 1
# force kill if server hangs
if self . get_process_pid ( " clickhouse server " ) :
# server can die before kill, so don't throw exception, it's expected
self . exec_in_container (
[ " bash " , " -c " , " pkill - {} clickhouse " . format ( 9 ) ] ,
nothrow = True ,
user = " root " ,
)
if callback_onstop :
callback_onstop ( self )
2023-09-28 15:28:26 +00:00
if clear_data_dir :
self . exec_in_container (
[
" bash " ,
" -c " ,
" rm -rf /var/lib/clickhouse/metadata && rm -rf /var/lib/clickhouse/data " ,
] ,
user = " root " ,
)
2021-08-20 15:58:43 +00:00
self . exec_in_container (
[
" bash " ,
" -c " ,
2021-11-11 07:41:48 +00:00
" echo ' restart_with_original_version: From version ' && /usr/bin/clickhouse server --version && echo ' To version ' && /usr/share/clickhouse_original server --version " ,
2022-03-22 16:39:58 +00:00
]
)
2019-11-20 11:56:38 +00:00
self . exec_in_container (
2022-03-22 16:39:58 +00:00
[
" bash " ,
" -c " ,
2021-08-20 15:58:43 +00:00
" cp /usr/share/clickhouse_original /usr/bin/clickhouse && chmod 777 /usr/bin/clickhouse " ,
] ,
user = " root " ,
)
self . exec_in_container (
[ " bash " , " -c " , " {} --daemon " . format ( self . clickhouse_start_command ) ] ,
user = str ( os . getuid ( ) ) ,
)
# wait start
2021-11-11 07:41:48 +00:00
time_left = begin_time + stop_start_wait_sec - time . time ( )
if time_left < = 0 :
raise Exception ( f " No time left during restart " )
else :
self . wait_start ( time_left )
2021-08-20 15:58:43 +00:00
2021-06-16 11:15:08 +00:00
def restart_with_latest_version (
2022-06-23 19:40:05 +00:00
self ,
stop_start_wait_sec = 300 ,
callback_onstop = None ,
signal = 15 ,
fix_metadata = False ,
2021-06-16 11:15:08 +00:00
) :
2021-11-11 07:41:48 +00:00
begin_time = time . time ( )
2019-06-20 16:25:32 +00:00
if not self . stay_alive :
raise Exception ( " Cannot restart not stay alive container " )
self . exec_in_container (
[ " bash " , " -c " , " pkill - {} clickhouse " . format ( signal ) ] , user = " root "
)
2019-06-21 08:03:13 +00:00
retries = int ( stop_start_wait_sec / 0.5 )
local_counter = 0
# wait stop
while local_counter < retries :
if not self . get_process_pid ( " clickhouse server " ) :
break
time . sleep ( 0.5 )
local_counter + = 1
2019-12-17 18:07:13 +00:00
# force kill if server hangs
if self . get_process_pid ( " clickhouse server " ) :
2020-08-03 14:40:02 +00:00
# server can die before kill, so don't throw exception, it's expected
self . exec_in_container (
[ " bash " , " -c " , " pkill - {} clickhouse " . format ( 9 ) ] ,
nothrow = True ,
user = " root " ,
)
2019-12-17 18:07:13 +00:00
2019-06-20 16:25:32 +00:00
if callback_onstop :
callback_onstop ( self )
2021-08-20 15:58:43 +00:00
self . exec_in_container (
[ " bash " , " -c " , " cp /usr/bin/clickhouse /usr/share/clickhouse_original " ] ,
user = " root " ,
)
2019-11-20 11:56:38 +00:00
self . exec_in_container (
2022-03-22 16:39:58 +00:00
[
" bash " ,
" -c " ,
2019-11-20 11:56:38 +00:00
" cp /usr/share/clickhouse_fresh /usr/bin/clickhouse && chmod 777 /usr/bin/clickhouse " ,
2022-03-22 16:39:58 +00:00
] ,
user = " root " ,
)
2019-11-20 11:56:38 +00:00
self . exec_in_container (
[
" bash " ,
" -c " ,
2021-11-11 07:41:48 +00:00
" echo ' restart_with_latest_version: From version ' && /usr/share/clickhouse_original server --version && echo ' To version ' /usr/share/clickhouse_fresh server --version " ,
2022-03-22 16:39:58 +00:00
]
2021-11-11 07:41:48 +00:00
)
2022-06-23 19:38:43 +00:00
if fix_metadata :
# Versions older than 20.7 might not create .sql file for system and default database
# Create it manually if upgrading from older version
self . exec_in_container (
2022-06-23 19:40:05 +00:00
[
" bash " ,
" -c " ,
2023-11-23 16:22:51 +00:00
" if [ ! -f /var/lib/clickhouse/metadata/system.sql ]; then echo ' ATTACH DATABASE system ENGINE=Ordinary ' > /var/lib/clickhouse/metadata/system.sql; fi " ,
2022-07-07 20:42:41 +00:00
]
2022-06-23 19:38:43 +00:00
)
self . exec_in_container (
2022-06-23 19:40:05 +00:00
[
" bash " ,
" -c " ,
2023-11-23 16:22:51 +00:00
" if [ ! -f /var/lib/clickhouse/metadata/default.sql ]; then echo ' ATTACH DATABASE system ENGINE=Ordinary ' > /var/lib/clickhouse/metadata/default.sql; fi " ,
2022-07-07 20:42:41 +00:00
]
2022-06-23 19:38:43 +00:00
)
2021-05-21 18:56:22 +00:00
self . exec_in_container (
[ " bash " , " -c " , " {} --daemon " . format ( self . clickhouse_start_command ) ] ,
user = str ( os . getuid ( ) ) ,
)
2021-02-24 11:46:58 +00:00
2019-06-21 08:03:13 +00:00
# wait start
2021-11-11 07:41:48 +00:00
time_left = begin_time + stop_start_wait_sec - time . time ( )
if time_left < = 0 :
raise Exception ( f " No time left during restart " )
else :
self . wait_start ( time_left )
2017-08-02 14:42:35 +00:00
2017-05-30 11:49:17 +00:00
def get_docker_handle ( self ) :
2021-04-13 14:55:31 +00:00
return self . cluster . get_docker_handle ( self . docker_id )
2017-05-19 18:54:05 +00:00
2017-05-30 11:49:17 +00:00
def stop ( self ) :
2019-06-04 20:59:31 +00:00
self . get_docker_handle ( ) . stop ( )
2017-05-30 11:49:17 +00:00
def start ( self ) :
self . get_docker_handle ( ) . start ( )
2021-05-18 07:38:32 +00:00
def wait_for_start ( self , start_timeout = None , connection_timeout = None ) :
2021-06-16 12:31:19 +00:00
handle = self . get_docker_handle ( )
2017-05-30 11:49:17 +00:00
2021-05-18 07:38:32 +00:00
if start_timeout is None or start_timeout < = 0 :
raise Exception ( " Invalid timeout: {} " . format ( start_timeout ) )
if connection_timeout is not None and connection_timeout < start_timeout :
raise Exception (
" Connection timeout {} should be grater then start timeout {} " . format (
connection_timeout , start_timeout
)
2022-03-22 16:39:58 +00:00
)
2017-05-19 18:54:05 +00:00
2017-05-30 11:49:17 +00:00
start_time = time . time ( )
2021-05-17 17:04:43 +00:00
prev_rows_in_log = 0
2017-05-30 11:49:17 +00:00
2021-05-18 07:38:32 +00:00
def has_new_rows_in_log ( ) :
nonlocal prev_rows_in_log
try :
rows_in_log = int ( self . count_in_log ( " .* " ) . strip ( ) )
res = rows_in_log > prev_rows_in_log
prev_rows_in_log = rows_in_log
return res
except ValueError :
return False
2017-05-19 18:54:05 +00:00
while True :
2021-06-16 12:31:19 +00:00
handle . reload ( )
2019-11-20 11:56:38 +00:00
status = handle . status
2017-05-30 11:49:17 +00:00
if status == " exited " :
2021-06-16 12:31:19 +00:00
raise Exception (
f " Instance ` { self . name } ' failed to start. Container status: { status } , logs: { handle . logs ( ) . decode ( ' utf-8 ' ) } "
)
2017-05-30 11:49:17 +00:00
2021-05-18 07:38:32 +00:00
deadline = start_time + start_timeout
2021-05-17 17:04:43 +00:00
# It is possible that server starts slowly.
# If container is running, and there is some progress in log, check connection_timeout.
2021-05-18 07:38:32 +00:00
if connection_timeout and status == " running " and has_new_rows_in_log ( ) :
deadline = start_time + connection_timeout
2017-05-30 11:49:17 +00:00
current_time = time . time ( )
2021-05-18 07:38:32 +00:00
if current_time > = deadline :
2021-06-16 12:31:19 +00:00
raise Exception (
f " Timed out while waiting for instance ` { self . name } ' with ip address { self . ip_address } to start. "
f " Container status: { status } , logs: { handle . logs ( ) . decode ( ' utf-8 ' ) } "
)
2017-05-19 18:54:05 +00:00
2021-05-18 07:38:32 +00:00
socket_timeout = min ( start_timeout , deadline - current_time )
2017-05-19 18:54:05 +00:00
# Repeatedly poll the instance address until there is something that listens there.
# Usually it means that ClickHouse is ready to accept queries.
try :
sock = socket . socket ( socket . AF_INET , socket . SOCK_STREAM )
2021-05-18 07:38:32 +00:00
sock . settimeout ( socket_timeout )
2017-05-19 18:54:05 +00:00
sock . connect ( ( self . ip_address , 9000 ) )
2021-03-05 13:39:51 +00:00
self . is_up = True
2017-05-19 18:54:05 +00:00
return
2017-05-23 17:13:36 +00:00
except socket . timeout :
continue
2017-05-19 18:54:05 +00:00
except socket . error as e :
2020-09-07 16:43:35 +00:00
if (
e . errno == errno . ECONNREFUSED
or e . errno == errno . EHOSTUNREACH
or e . errno == errno . ENETUNREACH
) :
2017-05-19 18:54:05 +00:00
time . sleep ( 0.1 )
else :
raise
finally :
sock . close ( )
2021-09-22 15:00:08 +00:00
def dict_to_xml ( self , dictionary ) :
xml_str = dict2xml (
dictionary , wrap = self . config_root_name , indent = " " , newlines = True
)
2021-01-27 09:50:11 +00:00
return xml_str
2017-05-30 11:49:17 +00:00
2024-05-14 16:53:59 +00:00
def get_machine_name ( self ) :
return platform . machine ( )
2018-08-22 15:42:27 +00:00
@property
def odbc_drivers ( self ) :
if self . odbc_ini_path :
return {
" SQLite3 " : {
" DSN " : " sqlite3_odbc " ,
2019-11-20 11:56:38 +00:00
" Database " : " /tmp/sqliteodbc " ,
2024-05-14 16:53:59 +00:00
" Driver " : f " /usr/lib/ { self . get_machine_name ( ) } -linux-gnu/odbc/libsqlite3odbc.so " ,
" Setup " : f " /usr/lib/ { self . get_machine_name ( ) } -linux-gnu/odbc/libsqlite3odbc.so " ,
2018-08-22 15:42:27 +00:00
} ,
" MySQL " : {
" DSN " : " mysql_odbc " ,
2024-05-14 16:53:59 +00:00
" Driver " : f " /usr/lib/ { self . get_machine_name ( ) } -linux-gnu/odbc/libmyodbc.so " ,
2022-04-04 18:49:30 +00:00
" Database " : odbc_mysql_db ,
" Uid " : odbc_mysql_uid ,
" Pwd " : odbc_mysql_pass ,
2024-02-13 12:15:46 +00:00
" Server " : self . cluster . mysql8_host ,
2018-08-22 15:42:27 +00:00
} ,
" PostgreSQL " : {
" DSN " : " postgresql_odbc " ,
2022-04-04 18:49:30 +00:00
" Database " : odbc_psql_db ,
" UserName " : odbc_psql_user ,
" Password " : odbc_psql_pass ,
2021-04-13 10:52:22 +00:00
" Port " : str ( self . cluster . postgres_port ) ,
2021-04-13 11:27:52 +00:00
" Servername " : self . cluster . postgres_host ,
2018-10-15 14:49:23 +00:00
" Protocol " : " 9.3 " ,
" ReadOnly " : " No " ,
" RowVersioning " : " No " ,
" ShowSystemTables " : " No " ,
2024-05-14 16:53:59 +00:00
" Driver " : f " /usr/lib/ { self . get_machine_name ( ) } -linux-gnu/odbc/psqlodbca.so " ,
" Setup " : f " /usr/lib/ { self . get_machine_name ( ) } -linux-gnu/odbc/libodbcpsqlS.so " ,
2018-10-15 14:49:23 +00:00
" ConnSettings " : " " ,
2018-08-22 15:42:27 +00:00
} ,
}
else :
return { }
def _create_odbc_config_file ( self ) :
with open ( self . odbc_ini_path . split ( " : " ) [ 0 ] , " w " ) as f :
2020-10-02 16:54:07 +00:00
for driver_setup in list ( self . odbc_drivers . values ( ) ) :
2018-08-22 15:42:27 +00:00
f . write ( " [ {} ] \n " . format ( driver_setup [ " DSN " ] ) )
2020-10-02 16:54:07 +00:00
for key , value in list ( driver_setup . items ( ) ) :
2018-08-22 15:42:27 +00:00
if key != " DSN " :
f . write ( key + " = " + value + " \n " )
2017-05-30 11:49:17 +00:00
2019-06-20 16:25:32 +00:00
def replace_config ( self , path_to_config , replacement ) :
self . exec_in_container (
[ " bash " , " -c " , " echo ' {} ' > {} " . format ( replacement , path_to_config ) ]
)
2021-10-14 10:21:41 +00:00
def replace_in_config ( self , path_to_config , replace , replacement ) :
self . exec_in_container (
[ " bash " , " -c " , f " sed -i ' s/ { replace } / { replacement } /g ' { path_to_config } " ]
)
2022-07-07 20:19:15 +00:00
def create_dir ( self ) :
2017-05-19 18:54:05 +00:00
""" Create the instance directory and all the needed files there. """
2017-05-30 11:49:17 +00:00
os . makedirs ( self . path )
2017-05-19 18:54:05 +00:00
2020-08-12 08:55:04 +00:00
instance_config_dir = p . abspath ( p . join ( self . path , " configs " ) )
os . makedirs ( instance_config_dir )
2017-05-19 18:54:05 +00:00
2021-05-24 06:59:16 +00:00
print (
f " Copy common default production configuration from { self . base_config_dir } . Files: { self . main_config_name } , { self . users_config_name } "
)
2021-05-21 18:56:22 +00:00
shutil . copyfile (
p . join ( self . base_config_dir , self . main_config_name ) ,
p . join ( instance_config_dir , self . main_config_name ) ,
)
shutil . copyfile (
p . join ( self . base_config_dir , self . users_config_name ) ,
p . join ( instance_config_dir , self . users_config_name ) ,
)
2017-05-19 18:54:05 +00:00
2021-02-18 21:21:50 +00:00
logging . debug ( " Create directory for configuration generated in this helper " )
2018-09-28 14:53:20 +00:00
# used by all utils with any config
2020-08-12 08:55:04 +00:00
conf_d_dir = p . abspath ( p . join ( instance_config_dir , " conf.d " ) )
2018-09-28 14:53:20 +00:00
os . mkdir ( conf_d_dir )
2020-08-12 08:55:04 +00:00
2021-02-18 21:21:50 +00:00
logging . debug ( " Create directory for common tests configuration " )
2020-08-12 08:55:04 +00:00
# used by server with main config.xml
self . config_d_dir = p . abspath ( p . join ( instance_config_dir , " config.d " ) )
2020-02-25 02:59:02 +00:00
os . mkdir ( self . config_d_dir )
2020-08-12 08:55:04 +00:00
users_d_dir = p . abspath ( p . join ( instance_config_dir , " users.d " ) )
2017-07-11 11:44:16 +00:00
os . mkdir ( users_d_dir )
2020-08-12 08:55:04 +00:00
dictionaries_dir = p . abspath ( p . join ( instance_config_dir , " dictionaries " ) )
os . mkdir ( dictionaries_dir )
2021-11-19 01:58:34 +00:00
extra_conf_dir = p . abspath ( p . join ( instance_config_dir , " extra_conf.d " ) )
os . mkdir ( extra_conf_dir )
2017-05-19 18:54:05 +00:00
2021-10-07 19:05:51 +00:00
def write_embedded_config ( name , dest_dir , fix_log_level = False ) :
2021-09-22 15:00:08 +00:00
with open ( p . join ( HELPERS_DIR , name ) , " r " ) as f :
data = f . read ( )
2021-10-26 05:58:50 +00:00
data = data . replace ( " clickhouse " , self . config_root_name )
2021-10-07 19:05:51 +00:00
if fix_log_level :
data = data . replace ( " <level>test</level> " , " <level>trace</level> " )
2021-09-22 15:00:08 +00:00
with open ( p . join ( dest_dir , name ) , " w " ) as r :
r . write ( data )
2021-05-11 13:13:26 +00:00
2021-02-18 21:21:50 +00:00
logging . debug ( " Copy common configuration from helpers " )
2019-09-02 19:57:09 +00:00
# The file is named with 0_ prefix to be processed before other configuration overloads.
2021-05-21 18:56:22 +00:00
if self . copy_common_configs :
2021-10-07 19:05:51 +00:00
write_embedded_config (
2023-04-10 18:23:56 +00:00
" 0_common_instance_config.xml " ,
self . config_d_dir ,
self . with_installed_binary ,
2021-10-07 19:05:51 +00:00
)
2021-09-22 15:00:08 +00:00
write_embedded_config ( " 0_common_instance_users.xml " , users_d_dir )
2024-04-27 08:32:27 +00:00
use_old_analyzer = os . environ . get ( " CLICKHOUSE_USE_OLD_ANALYZER " ) is not None
2024-04-27 13:49:09 +00:00
# If specific version was used there can be no
2024-07-12 12:49:26 +00:00
# enable_analyzer setting, so do this only if it was
2024-04-27 13:49:09 +00:00
# explicitly requested.
if self . tag :
use_old_analyzer = False
2024-04-27 08:32:27 +00:00
# Prefer specified in the test option:
if self . use_old_analyzer is not None :
use_old_analyzer = self . use_old_analyzer
if use_old_analyzer :
2024-04-27 08:33:21 +00:00
write_embedded_config ( " 0_common_enable_old_analyzer.xml " , users_d_dir )
2021-05-21 18:56:22 +00:00
2020-08-12 08:55:04 +00:00
if len ( self . custom_dictionaries_paths ) :
2021-09-22 15:00:08 +00:00
write_embedded_config ( " 0_common_enable_dictionaries.xml " , self . config_d_dir )
2017-05-19 18:54:05 +00:00
2023-09-11 12:34:16 +00:00
version = None
version_parts = self . tag . split ( " . " )
if version_parts [ 0 ] . isdigit ( ) and version_parts [ 1 ] . isdigit ( ) :
version = { " major " : int ( version_parts [ 0 ] ) , " minor " : int ( version_parts [ 1 ] ) }
# async replication is only supported in version 23.9+
# for tags that don't specify a version we assume it has a version of ClickHouse
# that supports async replication if a test for it is present
if (
version == None
or version [ " major " ] > 23
or ( version [ " major " ] == 23 and version [ " minor " ] > = 9 )
) :
2023-09-01 07:39:36 +00:00
write_embedded_config (
" 0_common_enable_keeper_async_replication.xml " , self . config_d_dir
)
2023-08-09 06:25:47 +00:00
2021-02-18 21:21:50 +00:00
logging . debug ( " Generate and write macros file " )
2018-07-25 16:00:51 +00:00
macros = self . macros . copy ( )
macros [ " instance " ] = self . name
2020-08-12 08:55:04 +00:00
with open ( p . join ( conf_d_dir , " macros.xml " ) , " w " ) as macros_config :
2019-11-20 11:56:38 +00:00
macros_config . write ( self . dict_to_xml ( { " macros " : macros } ) )
2017-05-19 18:54:05 +00:00
2017-05-30 11:49:17 +00:00
# Put ZooKeeper config
2017-05-19 18:54:05 +00:00
if self . with_zookeeper :
2018-09-28 14:53:20 +00:00
shutil . copy ( self . zookeeper_config_path , conf_d_dir )
2017-05-19 18:54:05 +00:00
2023-04-29 21:23:55 +00:00
if self . with_secrets :
2022-12-19 11:28:49 +00:00
if self . with_kerberos_kdc :
base_secrets_dir = self . cluster . instances_dir
else :
base_secrets_dir = self . path
2023-05-08 20:49:43 +00:00
from_dir = self . secrets_dir
to_dir = p . abspath ( p . join ( base_secrets_dir , " secrets " ) )
2023-04-29 21:23:55 +00:00
logging . debug ( f " Copy secret from { from_dir } to { to_dir } " )
2020-09-29 08:56:37 +00:00
shutil . copytree (
2023-04-29 21:23:55 +00:00
self . secrets_dir ,
2022-12-19 11:28:49 +00:00
p . abspath ( p . join ( base_secrets_dir , " secrets " ) ) ,
dirs_exist_ok = True ,
2022-03-22 16:39:58 +00:00
)
2020-09-29 08:56:37 +00:00
2022-07-20 17:09:38 +00:00
if self . with_coredns :
shutil . copytree (
self . coredns_config_dir , p . abspath ( p . join ( self . path , " coredns_config " ) )
)
2017-05-30 11:49:17 +00:00
# Copy config.d configs
2021-02-24 11:46:58 +00:00
logging . debug (
f " Copy custom test config files { self . custom_main_config_paths } to { self . config_d_dir } "
)
2017-05-30 11:49:17 +00:00
for path in self . custom_main_config_paths :
2020-02-25 02:59:02 +00:00
shutil . copy ( path , self . config_d_dir )
2017-05-19 18:54:05 +00:00
2017-05-30 11:49:17 +00:00
# Copy users.d configs
for path in self . custom_user_config_paths :
shutil . copy ( path , users_d_dir )
2020-08-12 08:55:04 +00:00
# Copy dictionaries configs to configs/dictionaries
for path in self . custom_dictionaries_paths :
shutil . copy ( path , dictionaries_dir )
2021-11-19 01:58:34 +00:00
for path in self . custom_extra_config_paths :
shutil . copy ( path , extra_conf_dir )
2020-08-12 08:55:04 +00:00
2017-08-25 13:47:09 +00:00
db_dir = p . abspath ( p . join ( self . path , " database " ) )
2021-02-24 11:46:58 +00:00
logging . debug ( f " Setup database dir { db_dir } " )
2017-06-15 20:08:26 +00:00
if self . clickhouse_path_dir is not None :
2021-02-24 11:46:58 +00:00
logging . debug ( f " Database files taken from { self . clickhouse_path_dir } " )
2020-09-09 11:11:59 +00:00
shutil . copytree ( self . clickhouse_path_dir , db_dir )
2021-02-24 11:46:58 +00:00
logging . debug (
f " Database copied from { self . clickhouse_path_dir } to { db_dir } "
)
2020-09-14 07:01:20 +00:00
else :
os . mkdir ( db_dir )
2017-05-19 18:54:05 +00:00
2017-08-25 13:47:09 +00:00
logs_dir = p . abspath ( p . join ( self . path , " logs " ) )
2021-02-24 11:46:58 +00:00
logging . debug ( f " Setup logs dir { logs_dir } " )
2017-05-19 18:54:05 +00:00
os . mkdir ( logs_dir )
2021-08-19 19:13:12 +00:00
self . logs_dir = logs_dir
2017-05-19 18:54:05 +00:00
2018-05-14 11:10:07 +00:00
depends_on = [ ]
2021-04-13 14:55:31 +00:00
if self . with_mysql_client :
depends_on . append ( self . cluster . mysql_client_host )
2024-02-13 12:15:46 +00:00
if self . with_mysql57 :
2021-02-15 09:35:45 +00:00
depends_on . append ( " mysql57 " )
2018-05-14 11:10:07 +00:00
2021-02-16 07:10:01 +00:00
if self . with_mysql8 :
depends_on . append ( " mysql80 " )
2021-04-13 14:55:31 +00:00
if self . with_mysql_cluster :
2024-02-13 12:15:46 +00:00
depends_on . append ( " mysql80 " )
2021-04-13 14:55:31 +00:00
depends_on . append ( " mysql2 " )
depends_on . append ( " mysql3 " )
depends_on . append ( " mysql4 " )
if self . with_postgres_cluster :
depends_on . append ( " postgres2 " )
depends_on . append ( " postgres3 " )
depends_on . append ( " postgres4 " )
2021-06-13 12:56:22 +00:00
2018-07-18 05:22:01 +00:00
if self . with_kafka :
depends_on . append ( " kafka1 " )
2020-02-03 00:02:19 +00:00
depends_on . append ( " schema-registry " )
2018-07-18 05:22:01 +00:00
2020-09-29 08:56:37 +00:00
if self . with_kerberized_kafka :
depends_on . append ( " kerberized_kafka1 " )
2022-12-14 13:39:23 +00:00
if self . with_kerberos_kdc :
depends_on . append ( " kerberoskdc " )
2020-09-10 10:02:46 +00:00
if self . with_kerberized_hdfs :
depends_on . append ( " kerberizedhdfs1 " )
2020-05-20 06:22:12 +00:00
if self . with_rabbitmq :
depends_on . append ( " rabbitmq1 " )
2022-05-17 13:58:09 +00:00
if self . with_nats :
depends_on . append ( " nats1 " )
2017-05-19 18:54:05 +00:00
if self . with_zookeeper :
2018-05-14 11:10:07 +00:00
depends_on . append ( " zoo1 " )
depends_on . append ( " zoo2 " )
depends_on . append ( " zoo3 " )
2017-05-19 18:54:05 +00:00
2019-11-20 11:56:38 +00:00
if self . with_minio :
depends_on . append ( " minio1 " )
2021-11-26 14:04:53 +00:00
if self . with_azurite :
depends_on . append ( " azurite1 " )
2023-10-02 19:09:35 +00:00
# In case the environment variables are exclusive, we don't want it to be in the cluster's env file.
# Instead, a separate env file will be created for the instance and needs to be filled with cluster's env variables.
2023-10-03 16:53:38 +00:00
if self . instance_env_variables is True :
2023-10-03 13:33:55 +00:00
# Create a dictionary containing cluster & instance env variables.
# Instance env variables will override cluster's.
temp_env_variables = self . cluster . env_variables . copy ( )
temp_env_variables . update ( self . env_variables )
self . env_variables = temp_env_variables
2023-10-02 18:59:15 +00:00
else :
self . cluster . env_variables . update ( self . env_variables )
2018-07-28 14:38:08 +00:00
2018-08-22 15:42:27 +00:00
odbc_ini_path = " "
if self . odbc_ini_path :
self . _create_odbc_config_file ( )
odbc_ini_path = " - " + self . odbc_ini_path
2021-05-21 18:56:22 +00:00
entrypoint_cmd = self . clickhouse_start_command
2018-11-22 15:59:00 +00:00
if self . stay_alive :
2023-07-11 02:32:33 +00:00
entrypoint_cmd = self . clickhouse_stay_alive_command . replace (
2021-05-21 18:56:22 +00:00
" {main_config_file} " , self . main_config_name
)
2021-09-26 15:06:32 +00:00
else :
entrypoint_cmd = (
2022-03-22 16:39:58 +00:00
" [ "
2021-09-26 15:06:32 +00:00
+ " , " . join ( map ( lambda x : ' " ' + x + ' " ' , entrypoint_cmd . split ( ) ) )
+ " ] "
2022-03-22 16:39:58 +00:00
)
2018-11-22 15:59:00 +00:00
2021-02-18 21:21:50 +00:00
logging . debug ( " Entrypoint cmd: {} " . format ( entrypoint_cmd ) )
2020-08-12 08:55:04 +00:00
2019-12-18 13:42:39 +00:00
networks = app_net = ipv4_address = ipv6_address = net_aliases = net_alias1 = " "
if (
self . ipv4_address is not None
or self . ipv6_address is not None
or self . hostname != self . name
) :
2018-12-27 15:55:51 +00:00
networks = " networks: "
2019-06-04 20:59:31 +00:00
app_net = " default: "
2018-12-27 15:55:51 +00:00
if self . ipv4_address is not None :
ipv4_address = " ipv4_address: " + self . ipv4_address
if self . ipv6_address is not None :
ipv6_address = " ipv6_address: " + self . ipv6_address
2019-12-18 13:42:39 +00:00
if self . hostname != self . name :
net_aliases = " aliases: "
net_alias1 = " - " + self . hostname
2018-12-27 15:55:51 +00:00
2019-06-20 16:25:32 +00:00
if not self . with_installed_binary :
binary_volume = " - " + self . server_bin_path + " :/usr/bin/clickhouse "
odbc_bridge_volume = (
" - " + self . odbc_bridge_bin_path + " :/usr/bin/clickhouse-odbc-bridge "
2022-03-22 16:39:58 +00:00
)
2021-03-11 17:48:47 +00:00
library_bridge_volume = (
" - "
+ self . library_bridge_bin_path
+ " :/usr/bin/clickhouse-library-bridge "
2022-03-22 16:39:58 +00:00
)
2019-06-20 16:25:32 +00:00
else :
binary_volume = " - " + self . server_bin_path + " :/usr/share/clickhouse_fresh "
odbc_bridge_volume = (
" - "
+ self . odbc_bridge_bin_path
+ " :/usr/share/clickhouse-odbc-bridge_fresh "
2022-03-22 16:39:58 +00:00
)
2021-03-11 17:48:47 +00:00
library_bridge_volume = (
" - "
+ self . library_bridge_bin_path
+ " :/usr/share/clickhouse-library-bridge_fresh "
2022-03-22 16:39:58 +00:00
)
2019-06-20 16:25:32 +00:00
2021-11-09 08:34:37 +00:00
external_dirs_volumes = " "
if self . external_dirs :
for external_dir in self . external_dirs :
external_dir_abs_path = p . abspath (
2022-04-19 18:15:27 +00:00
p . join ( self . cluster . instances_dir , external_dir . lstrip ( " / " ) )
2022-03-22 16:39:58 +00:00
)
2021-11-09 08:34:37 +00:00
logging . info ( f " external_dir_abs_path= { external_dir_abs_path } " )
2022-04-19 18:15:27 +00:00
os . makedirs ( external_dir_abs_path , exist_ok = True )
2021-11-09 08:34:37 +00:00
external_dirs_volumes + = (
" - " + external_dir_abs_path + " : " + external_dir + " \n "
2022-03-22 16:39:58 +00:00
)
2023-10-02 19:04:02 +00:00
# The current implementation of `self.env_variables` is not exclusive. Meaning the variables
# are shared with all nodes within the same cluster, even if it is specified for a single node.
2023-10-03 16:53:38 +00:00
# In order not to break the existing tests, the `self.instance_env_variables` option was added as a workaround.
2023-10-09 13:00:02 +00:00
# IMHO, it would be better to make `self.env_variables` exclusive by default and remove the `self.instance_env_variables` option.
2023-10-03 16:53:38 +00:00
if self . instance_env_variables :
2023-10-02 18:59:15 +00:00
self . env_file = p . abspath ( p . join ( self . path , " .env " ) )
_create_env_file ( self . env_file , self . env_variables )
2017-05-19 18:54:05 +00:00
with open ( self . docker_compose_path , " w " ) as docker_compose :
docker_compose . write (
DOCKER_COMPOSE_TEMPLATE . format (
2018-08-22 15:42:27 +00:00
image = self . image ,
2020-09-01 06:38:23 +00:00
tag = self . tag ,
2017-05-19 18:54:05 +00:00
name = self . name ,
2017-08-02 14:42:35 +00:00
hostname = self . hostname ,
2019-06-20 16:25:32 +00:00
binary_volume = binary_volume ,
odbc_bridge_volume = odbc_bridge_volume ,
2021-03-11 17:48:47 +00:00
library_bridge_volume = library_bridge_volume ,
2020-08-12 08:55:04 +00:00
instance_config_dir = instance_config_dir ,
2020-02-25 02:59:02 +00:00
config_d_dir = self . config_d_dir ,
2018-05-14 11:14:49 +00:00
db_dir = db_dir ,
2021-11-09 08:34:37 +00:00
external_dirs_volumes = external_dirs_volumes ,
2019-05-22 16:40:30 +00:00
tmpfs = str ( self . tmpfs ) ,
2024-03-12 22:31:33 +00:00
mem_limit = self . mem_limit ,
2017-05-19 18:54:05 +00:00
logs_dir = logs_dir ,
2018-07-28 14:38:08 +00:00
depends_on = str ( depends_on ) ,
2019-02-21 17:34:19 +00:00
user = os . getuid ( ) ,
2021-02-12 15:51:21 +00:00
env_file = self . env_file ,
2018-08-22 15:42:27 +00:00
odbc_ini_path = odbc_ini_path ,
2020-09-29 08:56:37 +00:00
keytab_path = self . keytab_path ,
krb5_conf = self . krb5_conf ,
2018-11-22 15:59:00 +00:00
entrypoint_cmd = entrypoint_cmd ,
2018-12-27 15:55:51 +00:00
networks = networks ,
app_net = app_net ,
ipv4_address = ipv4_address ,
ipv6_address = ipv6_address ,
2020-09-09 11:11:59 +00:00
net_aliases = net_aliases ,
net_alias1 = net_alias1 ,
2022-03-22 16:39:58 +00:00
)
2018-08-22 15:42:27 +00:00
)
2017-05-19 18:54:05 +00:00
2021-11-29 16:07:00 +00:00
def wait_for_path_exists ( self , path , seconds ) :
while seconds > 0 :
seconds - = 1
if self . path_exists ( path ) :
return
time . sleep ( 1 )
2020-02-06 12:18:19 +00:00
2021-11-18 18:14:46 +00:00
def get_backuped_s3_objects ( self , disk , backup_name ) :
2021-11-29 16:07:00 +00:00
path = f " /var/lib/clickhouse/disks/ { disk } /shadow/ { backup_name } /store "
self . wait_for_path_exists ( path , 10 )
2022-09-28 11:09:48 +00:00
return self . get_s3_objects ( path )
def get_s3_objects ( self , path ) :
2021-11-29 16:07:00 +00:00
command = [
" find " ,
path ,
" -type " ,
2021-11-18 18:14:46 +00:00
" f " ,
" -exec " ,
" grep " ,
" -o " ,
" r[01] \\ { 64 \\ }-file-[[:lower:]] \\ { 32 \\ } " ,
" {} " ,
2022-03-22 16:39:58 +00:00
" ; " ,
2021-11-18 18:14:46 +00:00
]
2022-09-28 11:09:48 +00:00
2021-11-18 18:14:46 +00:00
return self . exec_in_container ( command ) . split ( " \n " )
2022-09-28 11:09:48 +00:00
def get_s3_data_objects ( self , path ) :
command = [
" find " ,
path ,
" -type " ,
" f " ,
" -name " ,
" *.bin " ,
" -exec " ,
" grep " ,
" -o " ,
" r[01] \\ { 64 \\ }-file-[[:lower:]] \\ { 32 \\ } " ,
" {} " ,
" ; " ,
]
return self . exec_in_container ( command ) . split ( " \n " )
def get_table_objects ( self , table , database = None ) :
objects = [ ]
database_query = " "
if database :
database_query = f " AND database= ' { database } ' "
data_paths = self . query (
f """
SELECT arrayJoin ( data_paths )
FROM system . tables
WHERE name = ' {table} '
{ database_query }
"""
)
paths = data_paths . split ( " \n " )
for path in paths :
if path :
objects = objects + self . get_s3_data_objects ( path )
return objects
2023-01-27 13:22:50 +00:00
def create_format_schema ( self , file_name , content ) :
self . exec_in_container (
2023-02-09 13:09:46 +00:00
[
" bash " ,
" -c " ,
" echo ' {} ' > {} " . format (
content , " /var/lib/clickhouse/format_schemas/ " + file_name
) ,
]
2023-01-27 13:22:50 +00:00
)
2020-02-06 12:18:19 +00:00
class ClickHouseKiller ( object ) :
def __init__ ( self , clickhouse_node ) :
self . clickhouse_node = clickhouse_node
def __enter__ ( self ) :
2021-01-13 12:05:32 +00:00
self . clickhouse_node . stop_clickhouse ( kill = True )
2020-02-06 12:18:19 +00:00
def __exit__ ( self , exc_type , exc_val , exc_tb ) :
2021-01-11 17:37:08 +00:00
self . clickhouse_node . start_clickhouse ( )
2024-03-20 09:46:07 +00:00
@cache
def is_arm ( ) :
return any ( arch in platform . processor ( ) . lower ( ) for arch in ( " arm, aarch " ) )