diff --git a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClickHouseClusterReadSuite.scala b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClickHouseClusterReadSuite.scala index 2c5c03c1..44fe1ff2 100644 --- a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClickHouseClusterReadSuite.scala +++ b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClickHouseClusterReadSuite.scala @@ -22,7 +22,6 @@ import org.apache.spark.sql.execution.datasources.v2.BatchScanExec class ClickHouseClusterReadSuite extends SparkClickHouseClusterTest { test("clickhouse metadata column - distributed table") { - assume(isOnPrem, "This test is only for on prem version") withSimpleDistTable("single_replica", "db_w", "t_dist", true) { (_, db, tbl_dist, _) => assert(READ_DISTRIBUTED_CONVERT_LOCAL.defaultValueString == "true") @@ -49,7 +48,6 @@ class ClickHouseClusterReadSuite extends SparkClickHouseClusterTest { } test("push down aggregation - distributed table") { - assume(isOnPrem, "This test is only for on prem version") withSimpleDistTable("single_replica", "db_agg_col", "t_dist", true) { (_, db, tbl_dist, _) => checkAnswer( spark.sql(s"SELECT COUNT(id) FROM $db.$tbl_dist"), @@ -89,7 +87,6 @@ class ClickHouseClusterReadSuite extends SparkClickHouseClusterTest { } test("runtime filter - distributed table") { - assume(isOnPrem, "This test is only for on prem version") withSimpleDistTable("single_replica", "runtime_db", "runtime_tbl", true) { (_, db, tbl_dist, _) => spark.sql("set spark.clickhouse.read.runtimeFilter.enabled=false") checkAnswer( diff --git a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterDeleteSuite.scala b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterDeleteSuite.scala index 665665b5..a5d7d0e4 100644 --- a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterDeleteSuite.scala +++ b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterDeleteSuite.scala @@ -17,7 +17,6 @@ package org.apache.spark.sql.clickhouse.cluster class ClusterDeleteSuite extends SparkClickHouseClusterTest { test("truncate distribute table") { - assume(isOnPrem, "This test is only for on prem version") withSimpleDistTable("single_replica", "db_truncate", "tbl_truncate", true) { (_, db, tbl_dist, _) => assert(spark.table(s"$db.$tbl_dist").count() === 4) spark.sql(s"TRUNCATE TABLE $db.$tbl_dist") @@ -26,7 +25,6 @@ class ClusterDeleteSuite extends SparkClickHouseClusterTest { } test("delete from distribute table") { - assume(isOnPrem, "This test is only for on prem version") withSimpleDistTable("single_replica", "db_delete", "tbl_delete", true) { (_, db, tbl_dist, _) => assert(spark.table(s"$db.$tbl_dist").count() === 4) spark.sql(s"DELETE FROM $db.$tbl_dist WHERE m = 1") diff --git a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterPartitionManagementSuite.scala b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterPartitionManagementSuite.scala index 061394e9..63da1075 100644 --- a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterPartitionManagementSuite.scala +++ b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterPartitionManagementSuite.scala @@ -19,7 +19,6 @@ import org.apache.spark.sql.Row class ClusterPartitionManagementSuite extends SparkClickHouseClusterTest { test("distribute table partition") { - assume(isOnPrem, "This test is only for on prem version") withSimpleDistTable("single_replica", "db_part", "tbl_part", true) { (_, db, tbl_dist, _) => checkAnswer( spark.sql(s"SHOW PARTITIONS $db.$tbl_dist"), diff --git a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterShardByRandSuite.scala b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterShardByRandSuite.scala index 6b27849c..bade6e91 100644 --- a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterShardByRandSuite.scala +++ b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterShardByRandSuite.scala @@ -19,7 +19,6 @@ import org.apache.spark.sql.Row class ClusterShardByRandSuite extends SparkClickHouseClusterTest { test("shard by rand()") { - assume(isOnPrem, "This test is only for on prem version") val cluster = "single_replica" val db = "db_rand_shard" val tbl_dist = "tbl_rand_shard" diff --git a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterTableManagementSuite.scala b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterTableManagementSuite.scala index 184967c1..7096160d 100644 --- a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterTableManagementSuite.scala +++ b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/ClusterTableManagementSuite.scala @@ -17,7 +17,6 @@ package org.apache.spark.sql.clickhouse.cluster class ClusterTableManagementSuite extends SparkClickHouseClusterTest { test("create or replace distribute table") { - assume(isOnPrem, "This test is only for on prem version") autoCleanupDistTable("single_replica", "db_cor", "tbl_cor_dist") { (cluster, db, _, tbl_local) => def createLocalTable(): Unit = spark.sql( s"""CREATE TABLE $db.$tbl_local ( diff --git a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/TPCDSClusterSuite.scala b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/TPCDSClusterSuite.scala index c50ada33..6f5686fe 100644 --- a/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/TPCDSClusterSuite.scala +++ b/spark-3.5/clickhouse-spark-it/src/test/scala/org/apache/spark/sql/clickhouse/cluster/TPCDSClusterSuite.scala @@ -34,7 +34,6 @@ class TPCDSClusterSuite extends SparkClickHouseClusterTest { .set("spark.clickhouse.write.format", "json") test("Cluster: TPC-DS sf1 write and count(*)") { - assume(isOnPrem, "This test is only for on prem version") withDatabase("tpcds_sf1_cluster") { spark.sql("CREATE DATABASE tpcds_sf1_cluster WITH DBPROPERTIES (cluster = 'single_replica')")