[ 
https://issues.apache.org/jira/browse/PIG-4489?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14499286#comment-14499286
 ] 

Mohit Sabharwal commented on PIG-4489:
--------------------------------------

Thanks, [~kellyzly], the patch is to enable the 212 tests, which are not 
running for Spark engine.

For example, before this patch, unit tests like 
{{TestLoadStoreFunctionLifeCycle}} were not running for Spark engine because 
this tests creates {{PigServer}} using {code}        PigServer pigServer = new 
PigServer(Util.getLocalTestMode()); {code} and {{Util.getLocalTestMode()}} 
currently returns "local" for Spark unit tests -- which means "MapReduce local".

I think we'd like to run *unit tests* for Spark engine in two modes:
(1) {{ant test-spark-local}} would mean "local" (i.e. SPARK_MASTER="local")
(2) {{ant test-spark}} would mean "local cluster" (i.e. 
SPARK_MASTER="local-cluster [ 2 ,2, 1024 ]")

We can add support for "local cluster" in a separate jira. This is a mode 
supported in Spark that allows to unit test against a Spark cluster running on 
a local node, which multiple JVMs. With "local", unit tests are running against 
embedded Spark all in the same JVM.

I don't think we'd ever want to set SPARK_MASTER="spark://xxx:7077" when 
running unit tests.

Currently, it is confusing, because  {{ant test-spark}} means 
SPARK_MASTER="local"

We also need to add a file called {{test/spark-local-tests}} like we have 
{{test/spark-tests}} ( just like {{tets/tez-tests}} and 
{{test/tez-local-tests}}.) Any idea how we decide which tests go to which file ?




> Enable local mode tests for Spark engine
> ----------------------------------------
>
>                 Key: PIG-4489
>                 URL: https://issues.apache.org/jira/browse/PIG-4489
>             Project: Pig
>          Issue Type: Sub-task
>          Components: spark
>            Reporter: Mohit Sabharwal
>            Assignee: Mohit Sabharwal
>             Fix For: spark-branch
>
>         Attachments: PIG-4489.1.patch, PIG-4489.patch
>
>
> Util.getLocalTestMode() currently only returns "tez_local" or "local".
> I see that ~212 testcases do this check, and we are not running these tests 
> against Spark at this point.
> Currently all Spark tests run in local mode ("local" as a the Spark Cluster 
> URL passed to JavaSparkContext), so we should enable these tests as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to