Spark Release 3.5.7

Spark 3.5.7 is the seventh maintenance release containing security and correctness fixes. This release is based on the branch-3.5 maintenance branch of Spark. We strongly recommend all 3.5 users to upgrade to this stable release.

Notable changes

  • [SPARK-52721]: Wrong message parameter for CANNOT_PARSE_DATATYPE
  • [SPARK-52749]: Replace preview1 to dev1 in its PyPI package name
  • [SPARK-53518]: catalogString of User Defined Type should not be truncated
  • [SPARK-46941]: Can’t insert window group limit node for top-k computation if contains SizeBasedWindowFunction
  • [SPARK-49872]: Spark History UI – StreamConstraintsException: String length (20054016) exceeds the maximum length (20000000)
  • [SPARK-52023]: udaf returning Option can cause data corruption and crashes
  • [SPARK-52032]: ORC filter pushdown causes incorrect results with eqNullSafe (<=>) in DataFrame filter
  • [SPARK-52240]: VectorizedDeltaLengthByteArrayReader throws ParquetDecodingException: Failed to read X bytes
  • [SPARK-52339]: Relations may appear equal even though they are different
  • [SPARK-52516]: Memory Leak with coalesce foreachpartitions and v2 datasources
  • [SPARK-52611]: Fix SQLConf version for excludeSubqueryRefsFromRemoveRedundantAliases configuration
  • [SPARK-52684]: Make CACHE TABLE Commands atomic while encounting execution errors
  • [SPARK-52737]: SHS performance regression when visiting homepage
  • [SPARK-52776]: ProcfsMetricsGetter splits the comm field if it contains space characters
  • [SPARK-52791]: Inferring a UDT errors when first element is null
  • [SPARK-52809]: Don’t hold reader and iterator references for all partitions in task completion listeners for metric update
  • [SPARK-52873]: Hint causes semi join results to vary
  • [SPARK-53054]: The Scala Spark Connect DataFrameReader does not use the correct default format
  • [SPARK-53094]: Cube-related data quality problem
  • [SPARK-53155]: Global lower agggregation should not be removed
  • [SPARK-53435]: race condition in CachedRDDBuilder
  • [SPARK-53560]: Crash looping when retrying uncommitted batch in Kafka source and AvailableNow trigger
  • [SPARK-53581]: Potential thread-safety issue for mapTaskIds.add() in IndexShuffleBlockResolver

Dependency changes

While being a maintenance release we did still upgrade some dependencies in this release they are:

You can consult JIRA for the detailed changes.

We would like to acknowledge all community members for contributing patches to this release.


Spark News Archive