[ 
https://issues.apache.org/jira/browse/CASSANALYTICS-58?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17979923#comment-17979923
 ] 

Jack edited comment on CASSANALYTICS-58 at 6/17/25 12:21 PM:
-------------------------------------------------------------

I have hit a blocking issue in relation to the join planner, spark session 
extensions require a type of org.apache.spark.sql.SparkSession while the Data 
source v2 API used for that optimisation requires a type of 
org.apache.spark.sql.classic.SparkSession. I have reached out to the mailing 
list and might raise it to spark dev list if the community list is silent for a 
while, as this is a real use case and one I suspect will be common.

To anyone else looking at this issue, a general question I have - should we 
look at alternatives to supporting only spark classic/datasource v2 for the 
join planner to be something we can carry forward?

I still see value in getting _something_ working against spark 4, but pulling 
that out will have user impact. Community guidance and input would be 
appreciated on this topic, I'll leave this comment to sit for a while in case 
anybody comes across it/is watching this, but I think this could be something 
we reasonably discuss on the dev mailing list given the number of potentially 
impacted users.


was (Author: jackbuggins):
I have hit a blocking issue in relation to the join planner, spark session 
extensions require a type of org.apache.spark.sql.SparkSession while the Data 
source v2 API used for that optimisation requires a type of 
org.apache.spark.sql.classic.SparkSession. I have reached out to the mailing 
list and might raise it to dev, as this is a real use case and one I suspect 
will be common.

To anyone else looking at this issue, a general question I have - should we 
look at alternatives to supporting only spark classic/datasource v2 for the 
join planner to be something we can carry forward?

I still see value in getting _something_ working against spark 4, but pulling 
that out will have user impact. Community guidance and input would be 
appreciated on this topic, I'll leave this comment to sit for a while in case 
anybody comes across it/is watching this, but I think this could be something 
we reasonably discuss on the dev mailing list given the number of potentially 
impacted users.

> Apache Spark 4.0 support for the Apache Spark Connector 
> --------------------------------------------------------
>
>                 Key: CASSANALYTICS-58
>                 URL: https://issues.apache.org/jira/browse/CASSANALYTICS-58
>             Project: Apache Cassandra Analytics
>          Issue Type: New Feature
>            Reporter: Jack
>            Priority: Normal
>
> Spark 4.0 recently passed voting and subsequently a release has been made, in 
> order for users to upgrade to Spark 4, we will need to ensure compatibility 
> with spark 4 and mark a stable release once work is completed for this.
> At a high level this will include ensuring compatibility at scala level, 
> updating dependencies to match those of the official spark distribution, as 
> well as making adjustments for other existing spark interfaces implemented by 
> the connector.
> This issue does not account for integration with any new interfaces 
> introduced during this release - if any changes of this nature are required, 
> these can be split to smaller focussed tasks once all existing 
> implementations are ported upward.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to