[
https://issues.apache.org/jira/browse/MAHOUT-1570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15193704#comment-15193704
]
ASF GitHub Bot commented on MAHOUT-1570:
----------------------------------------
Github user smarthi commented on a diff in the pull request:
https://github.com/apache/mahout/pull/187#discussion_r56041224
--- Diff:
spark/src/main/scala/org/apache/mahout/drivers/TrainNBDriver.scala ---
@@ -48,33 +47,33 @@ object TrainNBDriver extends MahoutSparkDriver {
// default trainComplementary is false
opts = opts + ("trainComplementary" -> false)
- opt[Unit]("trainComplementary") abbr ("c") action { (_, options) =>
+ opt[Unit]("trainComplementary") abbr "c" action { (_, options) =>
options + ("trainComplementary" -> true)
- } text ("Train a complementary model, Default: false.")
+ } text "Train a complementary model, Default: false."
// Laplace smoothing paramater default is 1.0
opts = opts + ("alphaI" -> 1.0)
- opt[Double]("alphaI") abbr ("a") action { (x, options) =>
+ opt[Double]("alphaI") abbr "a" action { (x, options) =>
options + ("alphaI" -> x)
- } text ("Laplace soothing factor default is 1.0") validate { x =>
+ } text "Laplace soothing factor default is 1.0" validate { x =>
--- End diff --
its 'smoothing' not 'soothing'
> Adding support for Apache Flink as a backend for the Mahout DSL
> ---------------------------------------------------------------
>
> Key: MAHOUT-1570
> URL: https://issues.apache.org/jira/browse/MAHOUT-1570
> Project: Mahout
> Issue Type: Improvement
> Reporter: Till Rohrmann
> Assignee: Suneel Marthi
> Labels: DSL, flink, scala
> Fix For: 0.12.0
>
>
> With the finalized abstraction of the Mahout DSL plans from the backend
> operations (MAHOUT-1529), it should be possible to integrate further backends
> for the Mahout DSL. Apache Flink would be a suitable candidate to act as a
> good execution backend.
> With respect to the implementation, the biggest difference between Spark and
> Flink at the moment is probably the incremental rollout of plans, which is
> triggered by Spark's actions and which is not supported by Flink yet.
> However, the Flink community is working on this issue. For the moment, it
> should be possible to circumvent this problem by writing intermediate results
> required by an action to HDFS and reading from there.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)