diff --git a/pom.xml b/pom.xml index b591224fe744..dcb98d73a398 100644 --- a/pom.xml +++ b/pom.xml @@ -189,7 +189,7 @@ 4.17.0 10.23.1 7.8.0 - 119 + 121 1.24 11.13.1 1.15.1 diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeAzure.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeAzure.java index c46ef68e7356..36822fef8ff9 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeAzure.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeAzure.java @@ -130,7 +130,8 @@ public void extendEnvironment(Environment.Builder builder) private DockerContainer createSpark() { - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, "spark") + // TODO Update the image to spark4-iceberg once "java.lang.NoClassDefFoundError: org/codehaus/jackson/map/ObjectMapper" is fixed + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:119", "spark") .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(getSparkConf()), diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeGcs.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeGcs.java index 19091983dd4d..421f04ee93c7 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeGcs.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvMultinodeGcs.java @@ -134,7 +134,7 @@ public void extendEnvironment(Environment.Builder builder) private DockerContainer createSpark(MountableFile credentialsFile) { String containerGcpCredentialsFile = "/spark/conf/gcp-credentials.json"; - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, "spark") + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + hadoopImagesVersion, "spark") .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(dockerFiles.getDockerFilesHostPath("conf/environment/multinode-gcs/spark-defaults.conf")), diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHive.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHive.java index 19bb0e5ff3cd..57dd0f431c7f 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHive.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHive.java @@ -81,8 +81,8 @@ public void extendEnvironment(Environment.Builder builder) @SuppressWarnings("resource") private DockerContainer createSpark() { - // TODO: Switch to pure Spark 3 image once it's available (https://github.com/trinodb/trino/issues/7063) - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, "spark") + // TODO: Switch to pure Spark 4 image once it's available (https://github.com/trinodb/trino/issues/7063) + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + hadoopImagesVersion, "spark") .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(dockerFiles.getDockerFilesHostPath("conf/environment/singlenode-spark-hive/spark-defaults.conf")), diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHiveNoStatsFallback.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHiveNoStatsFallback.java index cbdce8583475..54a1b8db744f 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHiveNoStatsFallback.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkHiveNoStatsFallback.java @@ -72,7 +72,7 @@ public void extendEnvironment(Environment.Builder builder) @SuppressWarnings("resource") private DockerContainer createSpark() { - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, "spark") + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + hadoopImagesVersion, "spark") .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer(forHostPath(configDir.getPath("spark-defaults.conf")), "/spark/conf/spark-defaults.conf") .withCommand( diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIceberg.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIceberg.java index 1064b3a9f6a8..4afe0d59cbac 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIceberg.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIceberg.java @@ -81,7 +81,7 @@ public void extendEnvironment(Environment.Builder builder) @SuppressWarnings("resource") private DockerContainer createSpark() { - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, "spark") + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + hadoopImagesVersion, "spark") .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(dockerFiles.getDockerFilesHostPath("conf/environment/singlenode-spark-iceberg/spark-defaults.conf")), diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergJdbcCatalog.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergJdbcCatalog.java index dc206847a339..45abb4e4043f 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergJdbcCatalog.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergJdbcCatalog.java @@ -105,7 +105,7 @@ public DockerContainer createPostgreSql() private DockerContainer createSpark() { try { - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + imagesVersion, "spark") + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + imagesVersion, "spark") .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(dockerFiles.getDockerFilesHostPath("conf/environment/singlenode-spark-iceberg-jdbc-catalog/spark-defaults.conf")), diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergNessie.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergNessie.java index a7f7df42d48b..20761a6f7c8d 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergNessie.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergNessie.java @@ -75,7 +75,7 @@ public void extendEnvironment(Environment.Builder builder) @SuppressWarnings("resource") private DockerContainer createSparkContainer() { - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, SPARK) + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + hadoopImagesVersion, SPARK) .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(dockerFiles.getDockerFilesHostPath("conf/environment/singlenode-spark-iceberg-nessie/spark-defaults.conf")), diff --git a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergRest.java b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergRest.java index 4854e212dfde..2da33f6db6a9 100644 --- a/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergRest.java +++ b/testing/trino-product-tests-launcher/src/main/java/io/trino/tests/product/launcher/env/environment/EnvSinglenodeSparkIcebergRest.java @@ -94,7 +94,7 @@ private DockerContainer createRESTContainer() @SuppressWarnings("resource") private DockerContainer createSparkContainer() { - DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark3-iceberg:" + hadoopImagesVersion, SPARK_CONTAINER_NAME) + DockerContainer container = new DockerContainer("ghcr.io/trinodb/testing/spark4-iceberg:" + hadoopImagesVersion, SPARK_CONTAINER_NAME) .withEnv("HADOOP_USER_NAME", "hive") .withCopyFileToContainer( forHostPath(dockerFiles.getDockerFilesHostPath(