[
https://issues.apache.org/jira/browse/MAHOUT-1546?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13989777#comment-13989777
]
Dmitriy Lyubimov commented on MAHOUT-1546:
------------------------------------------
I'd say if "mahout -spark classpath" returns nothing, something is wrong with
your setup. Define MAHOUT_HOME and SCALA_HOME. Also, spark programs do not need
all Mahout artifacts and their dependendencies (e.g. core). More over, it may
cause (and is known to cause) dependency interference, especially with Hadoop
and its transitives. So using standard mahout classpath should be avoided.
> building spark context fails due to incorrect classpath query
> -------------------------------------------------------------
>
> Key: MAHOUT-1546
> URL: https://issues.apache.org/jira/browse/MAHOUT-1546
> Project: Mahout
> Issue Type: Bug
> Environment: Spark running locally
> Reporter: Pat Ferrel
> Assignee: Dmitriy Lyubimov
> Priority: Critical
>
> The classpath retrieval is using a "-spark" flag that returns nothing, using
> the default "mahout classpath" seems to get all needed jar paths so
> commenting out the "-spark" makes it work for me. Not sure this is the best
> fix though.
> This is in def mahoutSparkContext(...)
> {code}
> //val p = Runtime.getRuntime.exec(Array(exec.getAbsolutePath,
> "-spark", "classpath"))
> val p = Runtime.getRuntime.exec(Array(exec.getAbsolutePath,
> "classpath"))
> {code}
--
This message was sent by Atlassian JIRA
(v6.2#6252)