Hi, All.

Since Apache Spark 3.1.1 tag creation (Feb 21),
new 172 patches including 9 correctness patches and 4 K8s patches arrived
at branch-3.1.

Shall we make a new release, Apache Spark 3.1.2, as the second release at
3.1 line?
I'd like to volunteer for the release manager for Apache Spark 3.1.2.
I'm thinking about starting the first RC next week.

$ git log --oneline v3.1.1..HEAD | wc -l
     172

# Known correctness issues
SPARK-34534     New protocol FetchShuffleBlocks in OneForOneBlockFetcher
lead to data loss or correctness
SPARK-34545     PySpark Python UDF return inconsistent results when
applying 2 UDFs with different return type to 2 columns together
SPARK-34681     Full outer shuffled hash join when building left side
produces wrong result
SPARK-34719     fail if the view query has duplicated column names
SPARK-34794     Nested higher-order functions broken in DSL
SPARK-34829     transform_values return identical values when it's used
with udf that returns reference type
SPARK-34833     Apply right-padding correctly for correlated subqueries
SPARK-35381     Fix lambda variable name issues in nested DataFrame
functions in R APIs
SPARK-35382     Fix lambda variable name issues in nested DataFrame
functions in Python APIs

# Notable K8s patches since K8s GA
SPARK-34674    Close SparkContext after the Main method has finished
SPARK-34948    Add ownerReference to executor configmap to fix leakages
SPARK-34820    add apt-update before gnupg install
SPARK-34361    In case of downscaling avoid killing of executors already
known by the scheduler backend in the pod allocator

Bests,
Dongjoon.

Reply via email to