+1, it’s super messy without that. But great to see this running!
> On Aug 26, 2019, at 10:53 AM, Reynold Xin <r...@databricks.com> wrote:
>
> Exactly - I think it's important to be able to create a single binary build.
> Otherwise downstream users (the 99.99% won't be building their own Spark but
> just pull it from Maven) will have to deal with the mess, and it's even worse
> for libraries.
>
>
> On Mon, Aug 26, 2019 at 10:51 AM, Dongjoon Hyun <dongjoon.h...@gmail.com
> <mailto:dongjoon.h...@gmail.com>> wrote:
> Oh, right. If you want to publish something to Maven, it will inherit the
> situation.
> Thank you for feedback. :)
>
> On Mon, Aug 26, 2019 at 10:37 AM Michael Heuer <heue...@gmail.com
> <mailto:heue...@gmail.com>> wrote:
> That is not true for any downstream users who also provide a library.
> Whatever build mess you create in Apache Spark, we'll have to inherit it. ;)
>
> michael
>
>
>> On Aug 26, 2019, at 12:32 PM, Dongjoon Hyun <dongjoon.h...@gmail.com
>> <mailto:dongjoon.h...@gmail.com>> wrote:
>>
>> As Shane wrote, not yet.
>>
>> `one build for works for both` is our aspiration and the next step mentioned
>> in the first email.
>>
>> > The next step is `how to support JDK8/JDK11 together in a single artifact`.
>>
>> For the downstream users who build from the Apache Spark source, that will
>> not be a blocker because they will prefer a single JDK.
>>
>> Bests,
>> Dongjoon.
>>
>> On Mon, Aug 26, 2019 at 10:28 AM Shane Knapp <skn...@berkeley.edu
>> <mailto:skn...@berkeley.edu>> wrote:
>> maybe in the future, but not right now as the hadoop 2.7 build is broken.
>>
>> also, i busted dev/run-tests.py <http://dev/run-tests.py> in my changes to
>> support java11 in PRBs:
>> https://github.com/apache/spark/pull/25585
>> <https://github.com/apache/spark/pull/25585>
>>
>> quick fix, testing now.
>>
>> On Mon, Aug 26, 2019 at 10:23 AM Reynold Xin <r...@databricks.com
>> <mailto:r...@databricks.com>> wrote:
>> Would it be possible to have one build that works for both?
>