diff --git a/.github/workflows/tests.yml b/.github/workflows/tests.yml index d29eb25f..3ee673f9 100644 --- a/.github/workflows/tests.yml +++ b/.github/workflows/tests.yml @@ -61,7 +61,7 @@ jobs: shell: bash -el {0} strategy: matrix: - spark_version: [3.4.1, 3.5.1] + spark_version: [3.4.1, 3.5.2] scala_version: [2.12.19] env: SPARK_VERSION: ${{ matrix.spark_version }} diff --git a/build.sbt b/build.sbt index f7bf0c13..759f9448 100644 --- a/build.sbt +++ b/build.sbt @@ -11,7 +11,7 @@ import sbt.nio.Keys._ lazy val scala212 = "2.12.19" lazy val scala213 = "2.13.14" -lazy val spark3 = "3.5.1" +lazy val spark3 = "3.5.2" lazy val spark4 = "4.0.0-SNAPSHOT" lazy val sparkVersion = settingKey[String]("sparkVersion") diff --git a/python/environment.yml b/python/environment.yml index 8f8db2e2..c3bdf4d9 100644 --- a/python/environment.yml +++ b/python/environment.yml @@ -24,7 +24,7 @@ dependencies: - typeguard - yapf=0.40.1 - pip: - - pyspark==3.5.1 + - pyspark==3.5.2 - databricks-cli==0.18 # Docs notebook source generation - databricks-sdk - setuptools==65.6.3 # Python packaging diff --git a/python/spark-4-environment.yml b/python/spark-4-environment.yml index 52a7c346..c06d76ab 100644 --- a/python/spark-4-environment.yml +++ b/python/spark-4-environment.yml @@ -27,7 +27,7 @@ dependencies: # Note: Spark 4 must currently be installed from source # We only install pyspark to pull in all the dependencies. It should be # uninstalled before testing. - - pyspark==3.5.1 + - pyspark==3.5.2 - databricks-cli==0.18 # Docs notebook source generation - databricks-sdk - setuptools==65.6.3 # Python packaging