Oh, right. If you want to publish something to Maven, it will inherit the situation. Thank you for feedback. :)
On Mon, Aug 26, 2019 at 10:37 AM Michael Heuer <[email protected]> wrote: > That is not true for any downstream users who also provide a library. > Whatever build mess you create in Apache Spark, we'll have to inherit it. > ;) > > michael > > > On Aug 26, 2019, at 12:32 PM, Dongjoon Hyun <[email protected]> > wrote: > > As Shane wrote, not yet. > > `one build for works for both` is our aspiration and the next step > mentioned in the first email. > > > The next step is `how to support JDK8/JDK11 together in a single > artifact`. > > For the downstream users who build from the Apache Spark source, that will > not be a blocker because they will prefer a single JDK. > > Bests, > Dongjoon. > > On Mon, Aug 26, 2019 at 10:28 AM Shane Knapp <[email protected]> wrote: > >> maybe in the future, but not right now as the hadoop 2.7 build is broken. >> >> also, i busted dev/run-tests.py in my changes to support java11 in PRBs: >> https://github.com/apache/spark/pull/25585 >> >> quick fix, testing now. >> >> On Mon, Aug 26, 2019 at 10:23 AM Reynold Xin <[email protected]> wrote: >> >>> Would it be possible to have one build that works for both? >>> >>> >> >
