What if we remove the hadoop2 profile, making all of its setting just hard
coded default (it existed at the time bc Hadoop 1/Hadoop 2, but we haven't
supported hadoop 1 for a while.

Then, override those values in the Spark 2.2. profile with Hadoop2.6, and
specify Java8 with a plugin so it will fail on the build if compiled with
java7

My thought.



On Mon, Jul 17, 2017 at 11:02 AM, dustin vanstee <dustinvans...@gmail.com>
wrote:

> Hi, Trevor and I were able to get the latest version of Mahout to compile
> with Spark 2.2.  The main tweaks being that Spark 2.2 requires java8 and
> hadoop2.6 or greater.   This issue is that we have a hadoop2 profile that
> sets up hadoop.version= 2.4.1, and loads dependencies and this is only
> compatible with Spark 2.1 and below.   I would like to propose removing the
> hadoop2 profile, and just baking in the hadoop version and dependencies
> within each spark profile.  I wanted to run that by the community before I
> went to far with it and get some feedback if there would be a better
> alternative.  Trevor can you weigh in if I missed something?
>

Reply via email to