If I remember correctly, we landed on option 1, creating a v3.1 without the
extra reflection logic and then just deprecating 3.0 when the time comes.
If everyone agrees with that I can amend the notes to describe that more
explicitly.

-Sam

On Mon, Oct 25, 2021 at 11:30 AM Wing Yew Poon <wyp...@cloudera.com.invalid>
wrote:

> Adding v3.2 to Spark Build Refactoring
>>
>>    -
>>
>>    Russell and Anton will coordinate on dropping in a Spark 3.2 module
>>    -
>>
>>    We currently have 3.1 in the `spark3` module. We’ll move that out to
>>    its own module and mirror what we do with the 3.2 module. (This will 
>> enable
>>    cleaning up some mixed 3.0/3.1 code)
>>
>> Hi,
> I'm sorry I missed the last sync and only have these meeting minutes to go
> by.
> A Spark 3.2 module has now been added. Is the plan still to add a Spark
> 3.1 module. Will we have v3.0, v3.1 and v3.2 subdirectories under spark/ ?
> I think when we first started discussing the issue for Spark 3 support and
> how to organize the code, the proposal was to support two versions?
> IMO, for maintainability, we should only support two versions of Spark 3.
> However, in this transition period, I can see two approaches:
> 1. Create a v3.1 subdirectory, remove the reflection workarounds for its
> code, add explicit 3.1-specific modules, and build and test against 3.1. We
> then have 3 Spark 3 versions. At the next release, deprecate Spark 3.0
> support and remove the v3.0 directory and its modules.
> 2. Support Spark 3.1 and 3.0 from the common 3.0-based code. At the next
> release, deprecate Spark 3.0 support, rename v3.0 to v3.1, and update its
> code to remove the reflection workarounds.
> As I said, I missed the meeting. Perhaps 1 is the plan that was decided?
> (If it is, I'm willing to take on the work. I just need to know the plan.)
> Thanks,
> Wing Yew
>
>
>

Reply via email to