Spark Release 3.5.7
Spark 3.5.7 is the seventh maintenance release containing security and correctness fixes. This release is based on the branch-3.5 maintenance branch of Spark. We strongly recommend all 3.5 users to upgrade to this stable release.
Notable changes
- [SPARK-52721]: Wrong message parameter for CANNOT_PARSE_DATATYPE
- [SPARK-52749]: Replace preview1 to dev1 in its PyPI package name
- [SPARK-53518]: catalogString of User Defined Type should not be truncated
- [SPARK-46941]: Can’t insert window group limit node for top-k computation if contains SizeBasedWindowFunction
- [SPARK-49872]: Spark History UI – StreamConstraintsException: String length (20054016) exceeds the maximum length (20000000)
- [SPARK-52023]: udaf returning Option can cause data corruption and crashes
- [SPARK-52032]: ORC filter pushdown causes incorrect results with eqNullSafe (<=>) in DataFrame filter
- [SPARK-52240]: VectorizedDeltaLengthByteArrayReader throws ParquetDecodingException: Failed to read X bytes
- [SPARK-52339]: Relations may appear equal even though they are different
- [SPARK-52516]: Memory Leak with coalesce foreachpartitions and v2 datasources
- [SPARK-52611]: Fix SQLConf version for excludeSubqueryRefsFromRemoveRedundantAliases configuration
- [SPARK-52684]: Make CACHE TABLE Commands atomic while encounting execution errors
- [SPARK-52737]: SHS performance regression when visiting homepage
- [SPARK-52776]: ProcfsMetricsGetter splits the comm field if it contains space characters
- [SPARK-52791]: Inferring a UDT errors when first element is null
- [SPARK-52809]: Don’t hold reader and iterator references for all partitions in task completion listeners for metric update
- [SPARK-52873]: Hint causes semi join results to vary
- [SPARK-53054]: The Scala Spark Connect DataFrameReader does not use the correct default format
- [SPARK-53094]: Cube-related data quality problem
- [SPARK-53155]: Global lower agggregation should not be removed
- [SPARK-53435]: race condition in CachedRDDBuilder
- [SPARK-53560]: Crash looping when retrying uncommitted batch in Kafka source and AvailableNow trigger
- [SPARK-53581]: Potential thread-safety issue for mapTaskIds.add() in IndexShuffleBlockResolver
Dependency changes
While being a maintenance release we did still upgrade some dependencies in this release they are:
You can consult JIRA for the detailed changes.
We would like to acknowledge all community members for contributing patches to this release.
Spark News Archive