From 4f273c60c3432f1e919bae87188563e2a078cec1 Mon Sep 17 00:00:00 2001 From: Max Kainov Date: Wed, 20 Mar 2024 09:46:07 +0000 Subject: [PATCH 1/5] CI: disable some kerberos and hdfs integration tests for arm #ci_set_arm #batch_2 #no_merge_commit --- tests/integration/helpers/cluster.py | 7 +++++++ tests/integration/test_kerberos_auth/test.py | 5 ++++- tests/integration/test_storage_kerberized_hdfs/test.py | 9 ++++++++- 3 files changed, 19 insertions(+), 2 deletions(-) diff --git a/tests/integration/helpers/cluster.py b/tests/integration/helpers/cluster.py index 152b0d066ee..b3088ba7c81 100644 --- a/tests/integration/helpers/cluster.py +++ b/tests/integration/helpers/cluster.py @@ -1,8 +1,10 @@ import base64 import errno +from functools import cache import http.client import logging import os +import platform import stat import os.path as p import pprint @@ -4743,3 +4745,8 @@ class ClickHouseKiller(object): def __exit__(self, exc_type, exc_val, exc_tb): self.clickhouse_node.start_clickhouse() + + +@cache +def is_arm(): + return any(arch in platform.processor().lower() for arch in ("arm, aarch")) diff --git a/tests/integration/test_kerberos_auth/test.py b/tests/integration/test_kerberos_auth/test.py index a41255cff3e..0025ca7c6b6 100644 --- a/tests/integration/test_kerberos_auth/test.py +++ b/tests/integration/test_kerberos_auth/test.py @@ -1,5 +1,5 @@ import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm cluster = ClickHouseCluster(__file__) instance1 = cluster.add_instance( @@ -62,10 +62,12 @@ def make_auth(instance): ) +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_kerberos_auth_with_keytab(kerberos_cluster): assert make_auth(instance1) == "kuser\n" +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_kerberos_auth_without_keytab(kerberos_cluster): assert ( "DB::Exception: : Authentication failed: password is incorrect, or there is no user with such name." @@ -73,6 +75,7 @@ def test_kerberos_auth_without_keytab(kerberos_cluster): ) +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_bad_path_to_keytab(kerberos_cluster): assert ( "DB::Exception: : Authentication failed: password is incorrect, or there is no user with such name." diff --git a/tests/integration/test_storage_kerberized_hdfs/test.py b/tests/integration/test_storage_kerberized_hdfs/test.py index 5ac8b4670f9..c33a9614945 100644 --- a/tests/integration/test_storage_kerberized_hdfs/test.py +++ b/tests/integration/test_storage_kerberized_hdfs/test.py @@ -3,7 +3,7 @@ import pytest import os -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm import subprocess cluster = ClickHouseCluster(__file__) @@ -29,6 +29,7 @@ def started_cluster(): cluster.shutdown() +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_read_table(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -44,6 +45,7 @@ def test_read_table(started_cluster): assert select_read == data +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_read_write_storage(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -59,6 +61,7 @@ def test_read_write_storage(started_cluster): assert select_read == "1\tMark\t72.53\n" +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_write_storage_not_expired(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -76,6 +79,7 @@ def test_write_storage_not_expired(started_cluster): assert select_read == "1\tMark\t72.53\n" +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_two_users(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -98,6 +102,7 @@ def test_two_users(started_cluster): ) +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_read_table_expired(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -118,6 +123,7 @@ def test_read_table_expired(started_cluster): started_cluster.unpause_container("hdfskerberos") +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_prohibited(started_cluster): node1.query( "create table HDFSStorTwoProhibited (id UInt32, name String, weight Float64) ENGINE = HDFS('hdfs://suser@kerberizedhdfs1:9010/storage_user_two_prohibited', 'TSV')" @@ -132,6 +138,7 @@ def test_prohibited(started_cluster): ) +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_cache_path(started_cluster): node1.query( "create table HDFSStorCachePath (id UInt32, name String, weight Float64) ENGINE = HDFS('hdfs://dedicatedcachepath@kerberizedhdfs1:9010/storage_dedicated_cache_path', 'TSV')" From 7b8c5f35b522c23a9abae735739496622cdef703 Mon Sep 17 00:00:00 2001 From: Max Kainov Date: Wed, 20 Mar 2024 14:05:06 +0000 Subject: [PATCH 2/5] disable test_strage_hdfs module #ci_set_arm #batch_2 #no_merge_commit --- tests/integration/test_storage_hdfs/test.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/tests/integration/test_storage_hdfs/test.py b/tests/integration/test_storage_hdfs/test.py index 121263fb622..9dec1954406 100644 --- a/tests/integration/test_storage_hdfs/test.py +++ b/tests/integration/test_storage_hdfs/test.py @@ -2,10 +2,13 @@ import os import pytest import time -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.test_tools import TSV from pyhdfs import HdfsClient +if is_arm(): + pytestmark = pytest.mark.skip + cluster = ClickHouseCluster(__file__) node1 = cluster.add_instance( "node1", From 984c7e69a0a58dfe1930a9e0ee18e8f19f52e9d1 Mon Sep 17 00:00:00 2001 From: Max K Date: Fri, 15 Mar 2024 15:18:45 +0100 Subject: [PATCH 3/5] disable test_allowed_url_with_config tests with HDFS #no_merge_commit #ci_set_arm --- tests/ci/ci_config.py | 2 +- .../test_allowed_url_from_config/test.py | 13 +++++++++---- 2 files changed, 10 insertions(+), 5 deletions(-) diff --git a/tests/ci/ci_config.py b/tests/ci/ci_config.py index 7c213da27ec..d2caf75f4bd 100644 --- a/tests/ci/ci_config.py +++ b/tests/ci/ci_config.py @@ -380,7 +380,7 @@ upgrade_check_digest = DigestConfig( docker=["clickhouse/upgrade-check"], ) integration_check_digest = DigestConfig( - include_paths=["./tests/ci/integration_test_check.py", "./tests/integration"], + include_paths=["./tests/ci/integration_test_check.py", "./tests/integration/"], exclude_files=[".md"], docker=IMAGES.copy(), ) diff --git a/tests/integration/test_allowed_url_from_config/test.py b/tests/integration/test_allowed_url_from_config/test.py index 3106cf12702..df8934aa69b 100644 --- a/tests/integration/test_allowed_url_from_config/test.py +++ b/tests/integration/test_allowed_url_from_config/test.py @@ -1,5 +1,5 @@ import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm cluster = ClickHouseCluster(__file__) node1 = cluster.add_instance("node1", main_configs=["configs/config_with_hosts.xml"]) @@ -16,9 +16,11 @@ node5 = cluster.add_instance( "node5", main_configs=["configs/config_without_allowed_hosts.xml"] ) node6 = cluster.add_instance("node6", main_configs=["configs/config_for_remote.xml"]) -node7 = cluster.add_instance( - "node7", main_configs=["configs/config_for_redirect.xml"], with_hdfs=True -) + +if not is_arm(): + node7 = cluster.add_instance( + "node7", main_configs=["configs/config_for_redirect.xml"], with_hdfs=True + ) @pytest.fixture(scope="module") @@ -270,6 +272,7 @@ def test_table_function_remote(start_cluster): ) +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_redirect(start_cluster): hdfs_api = start_cluster.hdfs_api @@ -284,6 +287,7 @@ def test_redirect(start_cluster): node7.query("DROP TABLE table_test_7_1") +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_HDFS(start_cluster): assert "not allowed" in node7.query_and_get_error( "CREATE TABLE table_test_7_2 (word String) ENGINE=HDFS('http://hdfs1:50075/webhdfs/v1/simple_storage?op=OPEN&namenoderpcaddress=hdfs1:9000&offset=0', 'CSV')" @@ -293,6 +297,7 @@ def test_HDFS(start_cluster): ) +@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_schema_inference(start_cluster): error = node7.query_and_get_error("desc url('http://test.com`, 'TSVRaw'')") assert error.find("ReadWriteBufferFromHTTPBase") == -1 From 55737de0e604760fadeacc5aae82b54d69d52df7 Mon Sep 17 00:00:00 2001 From: Max Kainov Date: Wed, 20 Mar 2024 15:48:53 +0000 Subject: [PATCH 4/5] disable more hdfs and kafka inttests #ci_set_arm --- tests/integration/test_disk_types/test.py | 9 ++++++--- .../test_endpoint_macro_substitution/test.py | 4 +++- tests/integration/test_format_avro_confluent/test.py | 6 +++++- tests/integration/test_jdbc_bridge/test.py | 5 ++++- tests/integration/test_kafka_bad_messages/test.py | 6 +++++- tests/integration/test_keeper_disks/test.py | 6 +++++- tests/integration/test_log_family_hdfs/test.py | 5 ++++- .../test_materialized_mysql_database/test.py | 5 +++++ tests/integration/test_merge_tree_hdfs/test.py | 6 +++++- tests/integration/test_mysql_protocol/test.py | 11 ++++++++++- tests/integration/test_redirect_url_storage/test.py | 7 ++++++- tests/integration/test_storage_kafka/test.py | 4 +++- 12 files changed, 61 insertions(+), 13 deletions(-) diff --git a/tests/integration/test_disk_types/test.py b/tests/integration/test_disk_types/test.py index b9b8ef2010d..3c4169be4de 100644 --- a/tests/integration/test_disk_types/test.py +++ b/tests/integration/test_disk_types/test.py @@ -1,14 +1,17 @@ import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.test_tools import TSV disk_types = { "default": "Local", "disk_s3": "S3", - "disk_hdfs": "HDFS", "disk_encrypted": "S3", } +# do not test HDFS on ARM +if not is_arm(): + disk_types["disk_hdfs"] = "HDFS" + @pytest.fixture(scope="module") def cluster(): @@ -18,7 +21,7 @@ def cluster(): "node", main_configs=["configs/storage.xml"], with_minio=True, - with_hdfs=True, + with_hdfs=not is_arm(), ) cluster.start() diff --git a/tests/integration/test_endpoint_macro_substitution/test.py b/tests/integration/test_endpoint_macro_substitution/test.py index ee72fb9b492..22a649e2225 100644 --- a/tests/integration/test_endpoint_macro_substitution/test.py +++ b/tests/integration/test_endpoint_macro_substitution/test.py @@ -1,5 +1,5 @@ import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.test_tools import TSV from pyhdfs import HdfsClient @@ -10,6 +10,8 @@ disk_types = { "disk_encrypted": "S3", } +if is_arm(): + pytestmark = pytest.mark.skip @pytest.fixture(scope="module") def cluster(): diff --git a/tests/integration/test_format_avro_confluent/test.py b/tests/integration/test_format_avro_confluent/test.py index 540f90ae05e..ccaaee83514 100644 --- a/tests/integration/test_format_avro_confluent/test.py +++ b/tests/integration/test_format_avro_confluent/test.py @@ -8,9 +8,13 @@ from confluent_kafka.avro.cached_schema_registry_client import ( CachedSchemaRegistryClient, ) from confluent_kafka.avro.serializer.message_serializer import MessageSerializer -from helpers.cluster import ClickHouseCluster, ClickHouseInstance +from helpers.cluster import ClickHouseCluster, ClickHouseInstance, is_arm from urllib import parse +# Skip on ARM due to Confluent/Kafka +if is_arm(): + pytestmark = pytest.mark.skip + @pytest.fixture(scope="module") def started_cluster(): diff --git a/tests/integration/test_jdbc_bridge/test.py b/tests/integration/test_jdbc_bridge/test.py index 0e41cc8c8b7..c4a0a525df3 100644 --- a/tests/integration/test_jdbc_bridge/test.py +++ b/tests/integration/test_jdbc_bridge/test.py @@ -3,7 +3,7 @@ import os.path as p import pytest import uuid -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.test_tools import TSV from string import Template @@ -14,6 +14,9 @@ instance = cluster.add_instance( datasource = "self" records = 1000 +if is_arm(): + pytestmark = pytest.mark.skip + @pytest.fixture(scope="module") def started_cluster(): diff --git a/tests/integration/test_kafka_bad_messages/test.py b/tests/integration/test_kafka_bad_messages/test.py index 954b6042305..0446ca5cb47 100644 --- a/tests/integration/test_kafka_bad_messages/test.py +++ b/tests/integration/test_kafka_bad_messages/test.py @@ -2,10 +2,14 @@ import time import logging import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from kafka import KafkaAdminClient, KafkaProducer, KafkaConsumer, BrokerConnection from kafka.admin import NewTopic +if is_arm(): + pytestmark = pytest.mark.skip + + cluster = ClickHouseCluster(__file__) instance = cluster.add_instance( "instance", diff --git a/tests/integration/test_keeper_disks/test.py b/tests/integration/test_keeper_disks/test.py index e41837b89b4..0c91aa03419 100644 --- a/tests/integration/test_keeper_disks/test.py +++ b/tests/integration/test_keeper_disks/test.py @@ -1,11 +1,15 @@ #!/usr/bin/env python3 import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm import helpers.keeper_utils as keeper_utils from minio.deleteobjects import DeleteObject import os +if is_arm(): + pytestmark = pytest.mark.skip + + CURRENT_TEST_DIR = os.path.dirname(os.path.abspath(__file__)) cluster = ClickHouseCluster(__file__) node_logs = cluster.add_instance( diff --git a/tests/integration/test_log_family_hdfs/test.py b/tests/integration/test_log_family_hdfs/test.py index e8afe364ec4..6c3d28d2e3c 100644 --- a/tests/integration/test_log_family_hdfs/test.py +++ b/tests/integration/test_log_family_hdfs/test.py @@ -2,10 +2,13 @@ import logging import sys import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from pyhdfs import HdfsClient +if is_arm(): + pytestmark = pytest.mark.skip + @pytest.fixture(scope="module") def started_cluster(): diff --git a/tests/integration/test_materialized_mysql_database/test.py b/tests/integration/test_materialized_mysql_database/test.py index 89c69c42adc..57e496fe737 100644 --- a/tests/integration/test_materialized_mysql_database/test.py +++ b/tests/integration/test_materialized_mysql_database/test.py @@ -6,6 +6,7 @@ from helpers.cluster import ( ClickHouseCluster, ClickHouseInstance, get_docker_compose_path, + is_arm, ) import logging @@ -13,6 +14,10 @@ from . import materialized_with_ddl DOCKER_COMPOSE_PATH = get_docker_compose_path() +# skip all test on arm due to no arm support in mysql57 +if is_arm(): + pytestmark = pytest.mark.skip + cluster = ClickHouseCluster(__file__) mysql_node = None mysql8_node = None diff --git a/tests/integration/test_merge_tree_hdfs/test.py b/tests/integration/test_merge_tree_hdfs/test.py index 95b63a5c8a3..5ca7dc5feb0 100644 --- a/tests/integration/test_merge_tree_hdfs/test.py +++ b/tests/integration/test_merge_tree_hdfs/test.py @@ -3,7 +3,7 @@ import time import os import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.utility import generate_values from helpers.wait_for_helpers import wait_for_delete_inactive_parts from helpers.wait_for_helpers import wait_for_delete_empty_parts @@ -16,6 +16,10 @@ CONFIG_PATH = os.path.join( ) +if is_arm(): + pytestmark = pytest.mark.skip + + def create_table(cluster, table_name, additional_settings=None): node = cluster.instances["node"] diff --git a/tests/integration/test_mysql_protocol/test.py b/tests/integration/test_mysql_protocol/test.py index 7a69d07633c..e641d4c2300 100644 --- a/tests/integration/test_mysql_protocol/test.py +++ b/tests/integration/test_mysql_protocol/test.py @@ -12,11 +12,20 @@ from typing import Literal import docker import pymysql.connections import pytest -from helpers.cluster import ClickHouseCluster, get_docker_compose_path, run_and_check +from helpers.cluster import ( + ClickHouseCluster, + get_docker_compose_path, + is_arm, + run_and_check, +) SCRIPT_DIR = os.path.dirname(os.path.realpath(__file__)) DOCKER_COMPOSE_PATH = get_docker_compose_path() +if is_arm(): + pytestmark = pytest.mark.skip + + cluster = ClickHouseCluster(__file__) node = cluster.add_instance( "node", diff --git a/tests/integration/test_redirect_url_storage/test.py b/tests/integration/test_redirect_url_storage/test.py index 033f02d7bde..ba3fb3e14ab 100644 --- a/tests/integration/test_redirect_url_storage/test.py +++ b/tests/integration/test_redirect_url_storage/test.py @@ -1,10 +1,15 @@ import pytest -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.network import PartitionManager import threading import time +# skip all tests in the module on ARM due to HDFS +if is_arm(): + pytestmark = pytest.mark.skip + + cluster = ClickHouseCluster(__file__) node1 = cluster.add_instance( "node1", diff --git a/tests/integration/test_storage_kafka/test.py b/tests/integration/test_storage_kafka/test.py index dea1ea49851..5bdea179449 100644 --- a/tests/integration/test_storage_kafka/test.py +++ b/tests/integration/test_storage_kafka/test.py @@ -22,7 +22,7 @@ import kafka.errors import pytest from google.protobuf.internal.encoder import _VarintBytes from helpers.client import QueryRuntimeException -from helpers.cluster import ClickHouseCluster +from helpers.cluster import ClickHouseCluster, is_arm from helpers.network import PartitionManager from helpers.test_tools import TSV from kafka import KafkaAdminClient, KafkaProducer, KafkaConsumer, BrokerConnection @@ -40,6 +40,8 @@ from . import kafka_pb2 from . import social_pb2 from . import message_with_repeated_pb2 +if is_arm(): + pytestmark = pytest.mark.skip # TODO: add test for run-time offset update in CH, if we manually update it on Kafka side. # TODO: add test for SELECT LIMIT is working. From 9f3c625ab5034cadb2805bcd2068f4341eab5829 Mon Sep 17 00:00:00 2001 From: Max Kainov Date: Wed, 20 Mar 2024 17:23:08 +0000 Subject: [PATCH 5/5] fix mysql client tests to use mysql8 image #ci_set_integration #ci_set_arm --- .../compose/docker_compose_mysql_client.yml | 2 +- .../test_endpoint_macro_substitution/test.py | 1 + tests/integration/test_kerberos_auth/test.py | 7 ++++--- tests/integration/test_mysql_protocol/test.py | 4 ---- tests/integration/test_storage_kerberized_hdfs/test.py | 10 +++------- 5 files changed, 9 insertions(+), 15 deletions(-) diff --git a/tests/integration/compose/docker_compose_mysql_client.yml b/tests/integration/compose/docker_compose_mysql_client.yml index 5b37b6e6c09..ee590118d4f 100644 --- a/tests/integration/compose/docker_compose_mysql_client.yml +++ b/tests/integration/compose/docker_compose_mysql_client.yml @@ -1,7 +1,7 @@ version: '2.3' services: mysql_client: - image: mysql:5.7 + image: mysql:8.0 restart: always environment: MYSQL_ALLOW_EMPTY_PASSWORD: 1 diff --git a/tests/integration/test_endpoint_macro_substitution/test.py b/tests/integration/test_endpoint_macro_substitution/test.py index 22a649e2225..7dc282a980f 100644 --- a/tests/integration/test_endpoint_macro_substitution/test.py +++ b/tests/integration/test_endpoint_macro_substitution/test.py @@ -13,6 +13,7 @@ disk_types = { if is_arm(): pytestmark = pytest.mark.skip + @pytest.fixture(scope="module") def cluster(): try: diff --git a/tests/integration/test_kerberos_auth/test.py b/tests/integration/test_kerberos_auth/test.py index 0025ca7c6b6..2d0dce33c14 100644 --- a/tests/integration/test_kerberos_auth/test.py +++ b/tests/integration/test_kerberos_auth/test.py @@ -1,6 +1,10 @@ import pytest from helpers.cluster import ClickHouseCluster, is_arm +if is_arm(): + pytestmark = pytest.mark.skip + + cluster = ClickHouseCluster(__file__) instance1 = cluster.add_instance( "instance1", @@ -62,12 +66,10 @@ def make_auth(instance): ) -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_kerberos_auth_with_keytab(kerberos_cluster): assert make_auth(instance1) == "kuser\n" -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_kerberos_auth_without_keytab(kerberos_cluster): assert ( "DB::Exception: : Authentication failed: password is incorrect, or there is no user with such name." @@ -75,7 +77,6 @@ def test_kerberos_auth_without_keytab(kerberos_cluster): ) -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_bad_path_to_keytab(kerberos_cluster): assert ( "DB::Exception: : Authentication failed: password is incorrect, or there is no user with such name." diff --git a/tests/integration/test_mysql_protocol/test.py b/tests/integration/test_mysql_protocol/test.py index e641d4c2300..094ae7b9fd0 100644 --- a/tests/integration/test_mysql_protocol/test.py +++ b/tests/integration/test_mysql_protocol/test.py @@ -15,16 +15,12 @@ import pytest from helpers.cluster import ( ClickHouseCluster, get_docker_compose_path, - is_arm, run_and_check, ) SCRIPT_DIR = os.path.dirname(os.path.realpath(__file__)) DOCKER_COMPOSE_PATH = get_docker_compose_path() -if is_arm(): - pytestmark = pytest.mark.skip - cluster = ClickHouseCluster(__file__) node = cluster.add_instance( diff --git a/tests/integration/test_storage_kerberized_hdfs/test.py b/tests/integration/test_storage_kerberized_hdfs/test.py index c33a9614945..c72152fa376 100644 --- a/tests/integration/test_storage_kerberized_hdfs/test.py +++ b/tests/integration/test_storage_kerberized_hdfs/test.py @@ -6,6 +6,9 @@ import os from helpers.cluster import ClickHouseCluster, is_arm import subprocess +if is_arm(): + pytestmark = pytest.mark.skip + cluster = ClickHouseCluster(__file__) node1 = cluster.add_instance( "node1", @@ -29,7 +32,6 @@ def started_cluster(): cluster.shutdown() -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_read_table(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -45,7 +47,6 @@ def test_read_table(started_cluster): assert select_read == data -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_read_write_storage(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -61,7 +62,6 @@ def test_read_write_storage(started_cluster): assert select_read == "1\tMark\t72.53\n" -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_write_storage_not_expired(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -79,7 +79,6 @@ def test_write_storage_not_expired(started_cluster): assert select_read == "1\tMark\t72.53\n" -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_two_users(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -102,7 +101,6 @@ def test_two_users(started_cluster): ) -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_read_table_expired(started_cluster): hdfs_api = started_cluster.hdfs_api @@ -123,7 +121,6 @@ def test_read_table_expired(started_cluster): started_cluster.unpause_container("hdfskerberos") -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_prohibited(started_cluster): node1.query( "create table HDFSStorTwoProhibited (id UInt32, name String, weight Float64) ENGINE = HDFS('hdfs://suser@kerberizedhdfs1:9010/storage_user_two_prohibited', 'TSV')" @@ -138,7 +135,6 @@ def test_prohibited(started_cluster): ) -@pytest.mark.skipif(is_arm(), reason="skip for ARM") def test_cache_path(started_cluster): node1.query( "create table HDFSStorCachePath (id UInt32, name String, weight Float64) ENGINE = HDFS('hdfs://dedicatedcachepath@kerberizedhdfs1:9010/storage_dedicated_cache_path', 'TSV')"