That is not true for any downstream users who also provide a library. Whatever build mess you create in Apache Spark, we'll have to inherit it. ;)
michael > On Aug 26, 2019, at 12:32 PM, Dongjoon Hyun <dongjoon.h...@gmail.com> wrote: > > As Shane wrote, not yet. > > `one build for works for both` is our aspiration and the next step mentioned > in the first email. > > > The next step is `how to support JDK8/JDK11 together in a single artifact`. > > For the downstream users who build from the Apache Spark source, that will > not be a blocker because they will prefer a single JDK. > > Bests, > Dongjoon. > > On Mon, Aug 26, 2019 at 10:28 AM Shane Knapp <skn...@berkeley.edu > <mailto:skn...@berkeley.edu>> wrote: > maybe in the future, but not right now as the hadoop 2.7 build is broken. > > also, i busted dev/run-tests.py in my changes to support java11 in PRBs: > https://github.com/apache/spark/pull/25585 > <https://github.com/apache/spark/pull/25585> > > quick fix, testing now. > > On Mon, Aug 26, 2019 at 10:23 AM Reynold Xin <r...@databricks.com > <mailto:r...@databricks.com>> wrote: > Would it be possible to have one build that works for both? > >