ClickHouse/tests/integration/test_storage_redis/test.py
2023-06-02 10:05:54 +08:00

389 lines
9.2 KiB
Python

## sudo -H pip install redis
import redis
import pytest
import struct
import sys
from helpers.client import QueryRuntimeException
from helpers.cluster import ClickHouseCluster
from helpers.test_tools import TSV
cluster = ClickHouseCluster(__file__)
node = cluster.add_instance("node", with_redis=True)
@pytest.fixture(scope="module")
def started_cluster():
try:
cluster.start()
yield cluster
finally:
cluster.shutdown()
def get_redis_connection(db_id=0):
client = redis.Redis(
host="localhost", port=cluster.redis_port, password="clickhouse", db=db_id
)
return client
def get_address_for_ch():
return cluster.redis_host + ":6379"
def drop_table(table):
node.query(f"DROP TABLE IF EXISTS {table} SYNC")
# see SerializationString.serializeBinary
def serialize_binary_for_string(x):
var_uint_max = (1 << 63) - 1
buf = bytearray()
# write length
length = len(x)
# length = (length << 1) ^ (length >> 63)
if length > var_uint_max:
raise ValueError("Value too large for varint encoding")
for i in range(9):
byte = length & 0x7F
if length > 0x7F:
byte |= 0x80
buf += bytes([byte])
length >>= 7
if not length:
break
# write data
buf += x.encode("utf-8")
return bytes(buf)
# see SerializationNumber.serializeBinary
def serialize_binary_for_uint32(x):
buf = bytearray()
packed_num = struct.pack("I", x)
buf += packed_num
if sys.byteorder != "little":
buf.reverse()
return bytes(buf)
def test_simple_select(started_cluster):
client = get_redis_connection()
address = get_address_for_ch()
# clean all
client.flushall()
drop_table("test_simple_select")
data = {}
for i in range(100):
packed = serialize_binary_for_string(str(i))
data[packed] = packed
client.mset(data)
client.close()
# create table
node.query(
f"""
CREATE TABLE test_simple_select(
k String,
v String
) Engine=Redis('{address}', 0, 'clickhouse') PRIMARY KEY (k)
"""
)
response = TSV.toMat(
node.query("SELECT k, v FROM test_simple_select WHERE k='0' FORMAT TSV")
)
assert len(response) == 1
assert response[0] == ["0", "0"]
response = TSV.toMat(
node.query("SELECT * FROM test_simple_select ORDER BY k FORMAT TSV")
)
assert len(response) == 100
assert response[0] == ["0", "0"]
def test_select_int(started_cluster):
client = get_redis_connection()
address = get_address_for_ch()
# clean all
client.flushall()
drop_table("test_select_int")
data = {}
for i in range(100):
packed = serialize_binary_for_uint32(i)
data[packed] = packed
client.mset(data)
client.close()
# create table
node.query(
f"""
CREATE TABLE test_select_int(
k UInt32,
v UInt32
) Engine=Redis('{address}', 0, 'clickhouse') PRIMARY KEY (k)
"""
)
response = TSV.toMat(
node.query("SELECT k, v FROM test_select_int WHERE k=0 FORMAT TSV")
)
assert len(response) == 1
assert response[0] == ["0", "0"]
response = TSV.toMat(
node.query("SELECT * FROM test_select_int ORDER BY k FORMAT TSV")
)
assert len(response) == 100
assert response[0] == ["0", "0"]
def test_create_table(started_cluster):
address = get_address_for_ch()
# simple creation
drop_table("test_create_table")
node.query(
f"""
CREATE TABLE test_create_table(
k String,
v UInt32
) Engine=Redis('{address}') PRIMARY KEY (k)
"""
)
# simple creation with full engine args
drop_table("test_create_table")
node.query(
f"""
CREATE TABLE test_create_table(
k String,
v UInt32
) Engine=Redis('{address}', 0, 'clickhouse', 10) PRIMARY KEY (k)
"""
)
drop_table("test_create_table")
node.query(
f"""
CREATE TABLE test_create_table(
k String,
f String,
v UInt32
) Engine=Redis('{address}', 0, 'clickhouse', 10) PRIMARY KEY (k)
"""
)
drop_table("test_create_table")
with pytest.raises(QueryRuntimeException):
node.query(
f"""
CREATE TABLE test_create_table(
k String,
f String,
v UInt32
) Engine=Redis('{address}', 0, 'clickhouse', 10) PRIMARY KEY ()
"""
)
drop_table("test_create_table")
with pytest.raises(QueryRuntimeException):
node.query(
f"""
CREATE TABLE test_create_table(
k String,
f String,
v UInt32
) Engine=Redis('{address}', 0, 'clickhouse', 10)
"""
)
def test_simple_insert(started_cluster):
client = get_redis_connection()
address = get_address_for_ch()
# clean all
client.flushall()
drop_table("test_simple_insert")
node.query(
f"""
CREATE TABLE test_simple_insert(
k UInt32,
m DateTime,
n String
) Engine=Redis('{address}', 0, 'clickhouse') PRIMARY KEY (k)
"""
)
node.query(
"""
INSERT INTO test_simple_insert Values
(1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy')
"""
)
response = node.query("SELECT COUNT(*) FROM test_simple_insert FORMAT Values")
assert response == "(2)"
response = TSV.toMat(
node.query("SELECT k, m, n FROM test_simple_insert WHERE k=1 FORMAT TSV")
)
assert len(response) == 1
assert response[0] == ["1", "2023-06-01 00:00:00", "lili"]
response = TSV.toMat(
node.query(
"SELECT k, m, n FROM test_simple_insert WHERE m='2023-06-01 00:00:00' FORMAT TSV"
)
)
assert len(response) == 1
assert response[0] == ["1", "2023-06-01 00:00:00", "lili"]
response = TSV.toMat(
node.query("SELECT k, m, n FROM test_simple_insert WHERE n='lili' FORMAT TSV")
)
assert len(response) == 1
assert response[0] == ["1", "2023-06-01 00:00:00", "lili"]
def test_update(started_cluster):
client = get_redis_connection()
address = get_address_for_ch()
# clean all
client.flushall()
drop_table("test_update")
node.query(
f"""
CREATE TABLE test_update(
k UInt32,
m DateTime,
n String
) Engine=Redis('{address}', 0, 'clickhouse') PRIMARY KEY (k)
"""
)
node.query(
"""
INSERT INTO test_update Values
(1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy')
"""
)
response = node.query(
"""
ALTER TABLE test_update UPDATE m='2023-06-03 00:00:00' WHERE k=1
"""
)
print("update response: ", response)
response = TSV.toMat(
node.query("SELECT k, m, n FROM test_update WHERE k=1 FORMAT TSV")
)
assert len(response) == 1
assert response[0] == ["1", "2023-06-03 00:00:00", "lili"]
# can not update key
with pytest.raises(QueryRuntimeException):
node.query(
"""
ALTER TABLE test_update UPDATE k=2 WHERE k=1
"""
)
def test_delete(started_cluster):
client = get_redis_connection()
address = get_address_for_ch()
# clean all
client.flushall()
drop_table("test_delete")
node.query(
f"""
CREATE TABLE test_delete(
k UInt32,
m DateTime,
n String
) Engine=Redis('{address}', 0, 'clickhouse') PRIMARY KEY (k)
"""
)
node.query(
"""
INSERT INTO test_delete Values
(1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy')
"""
)
response = node.query(
"""
ALTER TABLE test_delete DELETE WHERE k=1
"""
)
print("delete response: ", response)
response = TSV.toMat(node.query("SELECT k, m, n FROM test_delete FORMAT TSV"))
assert len(response) == 1
assert response[0] == ["2", "2023-06-02 00:00:00", "lucy"]
response = node.query(
"""
ALTER TABLE test_delete DELETE WHERE m='2023-06-02 00:00:00'
"""
)
response = TSV.toMat(node.query("SELECT k, m, n FROM test_delete FORMAT TSV"))
assert len(response) == 0
def test_truncate(started_cluster):
client = get_redis_connection()
address = get_address_for_ch()
# clean all
client.flushall()
drop_table("test_truncate")
node.query(
f"""
CREATE TABLE test_truncate(
k UInt32,
m DateTime,
n String
) Engine=Redis('{address}', 0, 'clickhouse') PRIMARY KEY (k)
"""
)
node.query(
"""
INSERT INTO test_truncate Values
(1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy')
"""
)
response = node.query(
"""
TRUNCATE TABLE test_truncate
"""
)
print("truncate table response: ", response)
response = TSV.toMat(node.query("SELECT COUNT(*) FROM test_truncate FORMAT TSV"))
assert len(response) == 1
assert response[0] == ["0"]