[
https://issues.apache.org/jira/browse/MAHOUT-1529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14061469#comment-14061469
]
ASF GitHub Bot commented on MAHOUT-1529:
----------------------------------------
GitHub user avati opened a pull request:
https://github.com/apache/mahout/pull/29
MAHOUT-1529: Move dense/sparse matrix test in mapBlock into spark/
In h2o engine, the Matrix provided to mapBlock() is an instance of
"H2OBlockMatrix extends AbstractMatrix", and neither a DenseMatrix
nor SparseMatrix. H2OBlockMatrix is a 0-copy virtual Matrix exposing
just the partition's data (created at almost no expense), and creates
a copy-on-write Matrix only if modified by the blockmapfunction.
So these two tests are failing with h2obindings. Hence moving these two
tests into spark module.
Signed-off-by: Anand Avati <[email protected]>
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/avati/mahout MAHOUT-1529e
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/mahout/pull/29.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #29
----
commit 1e3cdb68198636c9f38f2d41d782d12edba7a2f7
Author: Anand Avati <[email protected]>
Date: 2014-07-15T00:20:09Z
MAHOUT-1529: Move dense/sparse matrix test in mapBlock into spark/
In h2o engine, the Matrix provided to mapBlock() is an instance of
"H2OBlockMatrix extends AbstractMatrix", and neither a DenseMatrix
nor SparseMatrix. H2OBlockMatrix is a 0-copy virtual Matrix exposing
just the partition's data (created at almost no expense), and creates
a copy-on-write Matrix only if modified by the blockmapfunction.
So these two tests are failing with h2obindings. Hence moving these two
tests into spark module.
Signed-off-by: Anand Avati <[email protected]>
----
> Finalize abstraction of distributed logical plans from backend operations
> -------------------------------------------------------------------------
>
> Key: MAHOUT-1529
> URL: https://issues.apache.org/jira/browse/MAHOUT-1529
> Project: Mahout
> Issue Type: Improvement
> Reporter: Dmitriy Lyubimov
> Assignee: Dmitriy Lyubimov
> Fix For: 1.0
>
>
> We have a few situations when algorithm-facing API has Spark dependencies
> creeping in.
> In particular, we know of the following cases:
> -(1) checkpoint() accepts Spark constant StorageLevel directly;-
> -(2) certain things in CheckpointedDRM;-
> -(3) drmParallelize etc. routines in the "drm" and "sparkbindings" package.-
> -(5) drmBroadcast returns a Spark-specific Broadcast object-
> (6) Stratosphere/Flink conceptual api changes.
> *Current tracker:* PR #1 https://github.com/apache/mahout/pull/1 - closed,
> need new PR for remaining things once ready.
> *Pull requests are welcome*.
--
This message was sent by Atlassian JIRA
(v6.2#6252)