ClickHouse/tests/integration/helpers/keeper_utils.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

380 lines
11 KiB
Python
Raw Normal View History

2024-09-27 10:19:39 +00:00
import contextlib
import io
import logging
import re
2024-09-27 10:19:39 +00:00
import select
2022-09-06 10:58:14 +00:00
import socket
2024-09-27 10:19:39 +00:00
import subprocess
2022-09-06 10:58:14 +00:00
import time
from os import path as p
from typing import Iterable, List, Optional, Sequence, Union
2024-09-27 10:19:39 +00:00
2023-04-20 13:26:02 +00:00
from kazoo.client import KazooClient
2024-09-27 10:19:39 +00:00
from helpers.client import CommandRequest
2024-09-27 10:19:39 +00:00
from helpers.cluster import ClickHouseCluster, ClickHouseInstance
ss_established = [
"ss",
"--resolve",
"--tcp",
"--no-header",
"state",
"ESTABLISHED",
"( dport = 2181 or sport = 2181 )",
]
def get_active_zk_connections(node: ClickHouseInstance) -> List[str]:
return (
str(node.exec_in_container(ss_established, privileged=True, user="root"))
.strip()
.split("\n")
)
def get_zookeeper_which_node_connected_to(node: ClickHouseInstance) -> str:
line = str(
node.exec_in_container(ss_established, privileged=True, user="root")
).strip()
pattern = re.compile(r"zoo[0-9]+", re.IGNORECASE)
result = pattern.findall(line)
assert (
len(result) == 1
), "ClickHouse must be connected only to one Zookeeper at a time"
assert isinstance(result[0], str)
return result[0]
2023-09-07 00:44:54 +00:00
def execute_keeper_client_query(
cluster: ClickHouseCluster, node: ClickHouseInstance, query: str
) -> str:
request = CommandRequest(
[
cluster.server_bin_path,
"keeper-client",
"--host",
str(cluster.get_instance_ip(node.name)),
"--port",
str(cluster.zookeeper_port),
"-q",
query,
],
stdin="",
)
return request.get_answer()
class KeeperException(Exception):
pass
class KeeperClient(object):
2023-09-07 00:44:54 +00:00
SEPARATOR = b"\a\a\a\a\n"
def __init__(self, bin_path: str, host: str, port: int, connection_tries=30):
self.bin_path = bin_path
self.host = host
self.port = port
retry_count = 0
while True:
try:
self.proc = subprocess.Popen(
[
bin_path,
"keeper-client",
"--host",
host,
"--port",
str(port),
"--log-level",
"error",
"--tests-mode",
"--no-confirmation",
],
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
)
self.poller = select.epoll()
self.poller.register(self.proc.stdout)
self.poller.register(self.proc.stderr)
self._fd_nums = {
self.proc.stdout.fileno(): self.proc.stdout,
self.proc.stderr.fileno(): self.proc.stderr,
}
self.stopped = False
self.get("/keeper", 60.0)
break
except Exception as e:
retry_count += 1
if (
"All connection tries failed while connecting to ZooKeeper"
in str(e)
and retry_count < connection_tries
):
logging.debug(
"Got exception while connecting to Keeper: %s\nWill reconnect, reconnect count = %s",
e,
retry_count,
)
time.sleep(1)
else:
raise
def execute_query(self, query: str, timeout: float = 60.0) -> str:
output = io.BytesIO()
2023-09-07 00:44:54 +00:00
self.proc.stdin.write(query.encode() + b"\n")
self.proc.stdin.flush()
events = self.poller.poll(timeout)
2023-09-11 18:46:40 +00:00
if not events:
raise TimeoutError(f"Keeper client returned no output")
for fd_num, event in events:
if event & (select.EPOLLIN | select.EPOLLPRI):
file = self._fd_nums[fd_num]
if file == self.proc.stdout:
while True:
chunk = file.readline()
if chunk.endswith(self.SEPARATOR):
break
output.write(chunk)
elif file == self.proc.stderr:
self.proc.stdout.readline()
raise KeeperException(self.proc.stderr.readline().strip().decode())
else:
2023-09-07 00:44:54 +00:00
raise ValueError(f"Failed to read from pipe. Flag {event}")
data = output.getvalue().strip().decode()
return data
def cd(self, path: str, timeout: float = 60.0):
2024-06-20 15:40:01 +00:00
self.execute_query(f"cd '{path}'", timeout)
def ls(self, path: str, timeout: float = 60.0) -> list[str]:
2024-06-20 15:40:01 +00:00
return self.execute_query(f"ls '{path}'", timeout).split(" ")
def create(self, path: str, value: str, timeout: float = 60.0):
2024-06-20 15:40:01 +00:00
self.execute_query(f"create '{path}' '{value}'", timeout)
def get(self, path: str, timeout: float = 60.0) -> str:
2024-06-20 15:40:01 +00:00
return self.execute_query(f"get '{path}'", timeout)
def set(self, path: str, value: str, version: Optional[int] = None) -> None:
2023-09-16 00:32:32 +00:00
self.execute_query(
2024-06-23 00:46:58 +00:00
f"set '{path}' '{value}' {version if version is not None else ''}"
2023-09-16 00:32:32 +00:00
)
2023-09-16 00:07:54 +00:00
def rm(self, path: str, version: Optional[int] = None) -> None:
2024-06-20 15:40:01 +00:00
self.execute_query(f"rm '{path}' {version if version is not None else ''}")
2023-09-16 00:03:46 +00:00
def exists(self, path: str, timeout: float = 60.0) -> bool:
2024-06-20 15:40:01 +00:00
return bool(int(self.execute_query(f"exists '{path}'", timeout)))
def stop(self):
if not self.stopped:
self.stopped = True
2023-09-07 00:44:54 +00:00
self.proc.communicate(b"exit\n", timeout=10.0)
def sync(self, path: str, timeout: float = 60.0):
2024-06-20 15:40:01 +00:00
self.execute_query(f"sync '{path}'", timeout)
def touch(self, path: str, timeout: float = 60.0):
2024-06-20 15:40:01 +00:00
self.execute_query(f"touch '{path}'", timeout)
def find_big_family(self, path: str, n: int = 10, timeout: float = 60.0) -> str:
2024-06-20 15:40:01 +00:00
return self.execute_query(f"find_big_family '{path}' {n}", timeout)
def find_super_nodes(self, threshold: int, timeout: float = 60.0) -> str:
2023-09-07 00:44:54 +00:00
return self.execute_query(f"find_super_nodes {threshold}", timeout)
2023-10-12 14:01:37 +00:00
def get_direct_children_number(self, path: str, timeout: float = 60.0) -> str:
2024-06-20 15:40:01 +00:00
return self.execute_query(f"get_direct_children_number '{path}'", timeout)
2023-10-12 13:15:52 +00:00
def get_all_children_number(self, path: str, timeout: float = 60.0) -> str:
2024-06-20 15:40:01 +00:00
return self.execute_query(f"get_all_children_number '{path}'", timeout)
def delete_stale_backups(self, timeout: float = 60.0) -> str:
2023-09-07 00:44:54 +00:00
return self.execute_query("delete_stale_backups", timeout)
2023-10-12 14:01:37 +00:00
2023-09-07 00:44:54 +00:00
def reconfig(
self,
joining: Optional[str],
leaving: Optional[str],
new_members: Optional[str],
timeout: float = 60.0,
2023-09-07 00:44:54 +00:00
) -> str:
if bool(joining) + bool(leaving) + bool(new_members) != 1:
2023-09-07 00:44:54 +00:00
raise ValueError(
"Exactly one of joining, leaving or new_members must be specified"
)
if joining is not None:
2023-09-07 00:44:54 +00:00
operation = "add"
elif leaving is not None:
2023-09-07 00:44:54 +00:00
operation = "remove"
elif new_members is not None:
2023-09-07 00:44:54 +00:00
operation = "set"
else:
2023-09-07 00:44:54 +00:00
raise ValueError(
"At least one of joining, leaving or new_members must be specified"
)
2023-09-07 00:44:54 +00:00
return self.execute_query(
2024-06-23 00:46:58 +00:00
f"reconfig {operation} '{joining or leaving or new_members}'", timeout
2023-09-07 00:44:54 +00:00
)
@classmethod
@contextlib.contextmanager
2023-09-07 00:44:54 +00:00
def from_cluster(
cls, cluster: ClickHouseCluster, keeper_node: str, port: Optional[int] = None
2023-09-07 00:44:54 +00:00
) -> "KeeperClient":
client = cls(
cluster.server_bin_path,
cluster.get_instance_ip(keeper_node),
port or cluster.zookeeper_port,
)
try:
yield client
finally:
client.stop()
2022-11-08 11:56:20 +00:00
2022-09-06 10:58:14 +00:00
def get_keeper_socket(cluster, node, port=9181):
hosts = cluster.get_instance_ip(node.name)
client = socket.socket()
client.settimeout(10)
client.connect((hosts, port))
return client
def send_4lw_cmd(cluster, node, cmd="ruok", port=9181):
client = None
try:
client = get_keeper_socket(cluster, node, port)
client.send(cmd.encode())
data = client.recv(100_000)
data = data.decode()
return data
finally:
if client is not None:
client.close()
NOT_SERVING_REQUESTS_ERROR_MSG = "This instance is not currently serving requests"
2023-04-20 13:26:02 +00:00
def wait_until_connected(cluster, node, port=9181, timeout=30.0):
start = time.time()
2023-04-20 13:26:02 +00:00
2022-09-06 10:58:14 +00:00
while send_4lw_cmd(cluster, node, "mntr", port) == NOT_SERVING_REQUESTS_ERROR_MSG:
time.sleep(0.1)
2023-04-20 13:26:02 +00:00
if time.time() - start > timeout:
2023-04-20 13:26:02 +00:00
raise Exception(
f"{timeout}s timeout while waiting for {node.name} to start serving requests"
)
2022-09-06 10:58:14 +00:00
def wait_until_quorum_lost(cluster, node, port=9181):
while send_4lw_cmd(cluster, node, "mntr", port) != NOT_SERVING_REQUESTS_ERROR_MSG:
time.sleep(0.1)
def wait_nodes(cluster, nodes):
for node in nodes:
wait_until_connected(cluster, node)
2022-11-08 10:44:43 +00:00
def is_leader(cluster, node, port=9181):
stat = send_4lw_cmd(cluster, node, "stat", port)
2022-11-08 11:43:36 +00:00
return "Mode: leader" in stat
2022-11-16 01:33:46 +00:00
def is_follower(cluster, node, port=9181):
stat = send_4lw_cmd(cluster, node, "stat", port)
return "Mode: follower" in stat
2022-11-16 01:33:46 +00:00
def get_leader(cluster, nodes):
for node in nodes:
if is_leader(cluster, node):
return node
raise Exception("No leader in Keeper cluster.")
2023-04-20 13:26:02 +00:00
2023-12-07 06:49:10 +00:00
2023-12-12 08:32:16 +00:00
def get_any_follower(cluster, nodes):
2023-12-07 06:28:41 +00:00
for node in nodes:
if is_follower(cluster, node):
return node
raise Exception("No followers in Keeper cluster.")
2023-04-20 13:26:02 +00:00
2023-12-07 06:49:10 +00:00
2023-04-20 13:26:02 +00:00
def get_fake_zk(cluster, node, timeout: float = 30.0) -> KazooClient:
_fake = KazooClient(
hosts=cluster.get_instance_ip(node.name) + ":9181", timeout=timeout
)
_fake.start()
return _fake
def get_config_str(zk: KeeperClient) -> str:
2023-04-20 13:26:02 +00:00
"""
Return decoded contents of /keeper/config node
"""
return zk.get("/keeper/config")
2023-04-20 13:26:02 +00:00
def wait_configs_equal(left_config: str, right_zk: KeeperClient, timeout: float = 30.0):
2023-04-20 13:26:02 +00:00
"""
Check whether get /keeper/config result in left_config is equal
to get /keeper/config on right_zk ZK connection.
2023-04-20 13:26:02 +00:00
"""
start = time.time()
left_config = sorted(left_config.split("\n"))
while True:
right_config = sorted(get_config_str(right_zk).split("\n"))
if left_config == right_config:
return
time.sleep(1)
if time.time() - start > timeout:
raise Exception(
f"timeout while checking nodes configs to get equal. "
f"Left: {left_config}, right: {right_config}"
2023-07-03 17:52:57 +00:00
)
def replace_zookeeper_config(
nodes: Union[Sequence[ClickHouseInstance], ClickHouseInstance], new_config: str
) -> None:
if not isinstance(nodes, Sequence):
nodes = (nodes,)
for node in nodes:
node.replace_config("/etc/clickhouse-server/conf.d/zookeeper.xml", new_config)
node.query("SYSTEM RELOAD CONFIG")
def reset_zookeeper_config(
nodes: Union[Sequence[ClickHouseInstance], ClickHouseInstance],
file_path: str = p.join(p.dirname(p.realpath(__file__)), "zookeeper_config.xml"),
) -> None:
"""Resets the keeper config to default or to a given path on the disk"""
with open(file_path, "r", encoding="utf-8") as cf:
replace_zookeeper_config(nodes, cf.read())