[ 
https://issues.apache.org/jira/browse/SPARK-25075?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17350073#comment-17350073
 ] 

Dongjoon Hyun commented on SPARK-25075:
---------------------------------------

It sounds like two issues.
- For the deprecation policy of Apache Kafka 3.0, it sounds reasonable to me 
(as a user).
- For the relation with Apache Spark 3.2 and Apache Kafka 3.0, let's see after 
Apache Kafka 3.0 is released.

You know, I'm a big fan of Kafka and I usually catch up the new release of 
Apache Kafka to bring the latest client-side bug fixes to Apache Spark. 
However, the followings are the current AS-IS status which Apache Spark 
community are trying to stabilize.
- The latest Apache Spark release (3.1.1) was shipped Apache Kafka 2.6.0. 
Apache Spark 3.1.2 RC1 vote starts in a few days.
- Apache Spark master branch has been using Apache Kafka 2.8.

> Build and test Spark against Scala 2.13
> ---------------------------------------
>
>                 Key: SPARK-25075
>                 URL: https://issues.apache.org/jira/browse/SPARK-25075
>             Project: Spark
>          Issue Type: Umbrella
>          Components: Build, MLlib, Project Infra, Spark Core, SQL
>    Affects Versions: 3.0.0
>            Reporter: Guillaume Massé
>            Priority: Major
>
> This umbrella JIRA tracks the requirements for building and testing Spark 
> against the current Scala 2.13 milestone.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to