GitHub user zjffdu opened a pull request:

    https://github.com/apache/spark/pull/15159

    [SPARK-17605][SPARK_SUBMIT] Add option spark.usePython and spark.useR for 
applications that use both pyspark and sparkr

    ## What changes were proposed in this pull request?
    
    It is mostly for yarn mode, standalone mode don't need to distribute 
resources (sparkr.zip, pyspark.zip and etc) if I understand correctly. Add 2 
options spark.usePython and spark.useR, so that any project using both sparkR 
and pyspark can leverage these 2 options. 
    
    ## How was this patch tested?
    Use the following command to launch SparkPi and notice pyspark.zip, 
sparkr.zip and py4j are all distirbuted to executors. 
    ```
    bin/spark-submit --master yarn-client --conf spark.useR=true --conf 
spark.usePython=true --class org.apache.spark.examples.SparkPi 
examples/target/original-spark-examples_2.11-2.1.0-SNAPSHOT.jar
    ```
    Client output
    ```
    16/09/20 16:25:15 INFO Client: Uploading resource 
file:/private/var/folders/dp/hmchg5dd3vbcvds26q91spdw0000gp/T/spark-1fa45f53-e75e-4671-bf75-1ef554a3dda5/__spark_libs__3260692671624418275.zip
 -> 
hdfs://localhost:9009/user/jzhang/.sparkStaging/application_1474162755082_0035/__spark_libs__3260692671624418275.zip
    16/09/20 16:25:16 INFO Client: Uploading resource 
file:/Users/jzhang/github/spark/R/lib/sparkr.zip#sparkr -> 
hdfs://localhost:9009/user/jzhang/.sparkStaging/application_1474162755082_0035/sparkr.zip
    16/09/20 16:25:17 INFO Client: Uploading resource 
file:/Users/jzhang/github/spark/python/lib/pyspark.zip -> 
hdfs://localhost:9009/user/jzhang/.sparkStaging/application_1474162755082_0035/pyspark.zip
    16/09/20 16:25:17 INFO Client: Uploading resource 
file:/Users/jzhang/github/spark/python/lib/py4j-0.10.3-src.zip -> 
hdfs://localhost:9009/user/jzhang/.sparkStaging/application_1474162755082_0035/py4j-0.10.3-src.zip
    16/09/20 16:25:17 INFO Client: Uploading resource 
file:/private/var/folders/dp/hmchg5dd3vbcvds26q91spdw0000gp/T/spark-1fa45f53-e75e-4671-bf75-1ef554a3dda5/__spark_conf__2718308972579262508.zip
 -> 
hdfs://localhost:9009/user/jzhang/.sparkStaging/application_1474162755082_0035/__spark_conf__.zip
    ```


You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/zjffdu/spark SPARK-17605

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/15159.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #15159
    
----
commit 2271887102c87f5fb0bf6b7bcb3c602135b90930
Author: Jeff Zhang <zjf...@apache.org>
Date:   2016-09-18T01:11:06Z

    [SPARK-17605][SPARK_SUBMIT] Add option spark.usePython and spark.useR for 
applications that use both pyspark and sparkr

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to