[ 
https://issues.apache.org/jira/browse/MAHOUT-1529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14061469#comment-14061469
 ] 

ASF GitHub Bot commented on MAHOUT-1529:
----------------------------------------

GitHub user avati opened a pull request:

    https://github.com/apache/mahout/pull/29

    MAHOUT-1529: Move dense/sparse matrix test in mapBlock into spark/

    In h2o engine, the Matrix provided to mapBlock() is an instance of
    "H2OBlockMatrix extends AbstractMatrix", and neither a DenseMatrix
    nor SparseMatrix. H2OBlockMatrix is a 0-copy virtual Matrix exposing
    just the partition's data (created at almost no expense), and creates
    a copy-on-write Matrix only if modified by the blockmapfunction.
    
    So these two tests are failing with h2obindings. Hence moving these two
    tests into spark module.
    
    Signed-off-by: Anand Avati <[email protected]>

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/avati/mahout MAHOUT-1529e

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/mahout/pull/29.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #29
    
----
commit 1e3cdb68198636c9f38f2d41d782d12edba7a2f7
Author: Anand Avati <[email protected]>
Date:   2014-07-15T00:20:09Z

    MAHOUT-1529: Move dense/sparse matrix test in mapBlock into spark/
    
    In h2o engine, the Matrix provided to mapBlock() is an instance of
    "H2OBlockMatrix extends AbstractMatrix", and neither a DenseMatrix
    nor SparseMatrix. H2OBlockMatrix is a 0-copy virtual Matrix exposing
    just the partition's data (created at almost no expense), and creates
    a copy-on-write Matrix only if modified by the blockmapfunction.
    
    So these two tests are failing with h2obindings. Hence moving these two
    tests into spark module.
    
    Signed-off-by: Anand Avati <[email protected]>

----


> Finalize abstraction of distributed logical plans from backend operations
> -------------------------------------------------------------------------
>
>                 Key: MAHOUT-1529
>                 URL: https://issues.apache.org/jira/browse/MAHOUT-1529
>             Project: Mahout
>          Issue Type: Improvement
>            Reporter: Dmitriy Lyubimov
>            Assignee: Dmitriy Lyubimov
>             Fix For: 1.0
>
>
> We have a few situations when algorithm-facing API has Spark dependencies 
> creeping in. 
> In particular, we know of the following cases:
> -(1) checkpoint() accepts Spark constant StorageLevel directly;-
> -(2) certain things in CheckpointedDRM;-
> -(3) drmParallelize etc. routines in the "drm" and "sparkbindings" package.-
> -(5) drmBroadcast returns a Spark-specific Broadcast object-
> (6) Stratosphere/Flink conceptual api changes.
> *Current tracker:* PR #1 https://github.com/apache/mahout/pull/1 - closed, 
> need new PR for remaining things once ready.
> *Pull requests are welcome*.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to