[ https://issues.apache.org/jira/browse/MAHOUT-1529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13988493#comment-13988493 ]
Dmitriy Lyubimov commented on MAHOUT-1529: ------------------------------------------ bq. I think we also need bq. (6) Rename mahout spark-shell (both command and source dir/files/variables) to "mahout shell" (or mahout console?) which only uses the logical layer and backend layer is selected at runtime/startup. No, we don't . Shell is in essense Spark's REPL. in that sense it is exactly and literally spark-shell. It includes byte code mechanisms to compile closures on-the-fly and pass them to the backend. How other engines would want to do that, i have no clue. Chances for a generic (and cheap) Mahout shell are very slim IMO. > Finalize abstraction of distributed logical plans from backend operations > ------------------------------------------------------------------------- > > Key: MAHOUT-1529 > URL: https://issues.apache.org/jira/browse/MAHOUT-1529 > Project: Mahout > Issue Type: Improvement > Reporter: Dmitriy Lyubimov > Fix For: 1.0 > > > We have a few situations when algorithm-facing API has Spark dependencies > creeping in. > In particular, we know of the following cases: > (1) checkpoint() accepts Spark constant StorageLevel directly; > (2) certain things in CheckpointedDRM; > (3) drmParallelize etc. routines in the "drm" and "sparkbindings" package. > (5) drmBroadcast returns a Spark-specific Broadcast object -- This message was sent by Atlassian JIRA (v6.2#6252)