[ 
https://issues.apache.org/jira/browse/SPARK-1688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron Davidson resolved SPARK-1688.
-----------------------------------

    Resolution: Fixed

> PySpark throws unhelpful exception when pyspark cannot be loaded
> ----------------------------------------------------------------
>
>                 Key: SPARK-1688
>                 URL: https://issues.apache.org/jira/browse/SPARK-1688
>             Project: Spark
>          Issue Type: Improvement
>          Components: PySpark, Spark Core
>    Affects Versions: 0.9.1
>            Reporter: Andrew Or
>            Assignee: Andrew Or
>             Fix For: 1.0.0
>
>
> Currently, if pyspark cannot be loaded, this happens:
> java.io.EOFException
>         at java.io.DataInputStream.readInt(DataInputStream.java:392)
>         at 
> org.apache.spark.api.python.PythonWorkerFactory.startDaemon(PythonWorkerFactory.scala:183)
>         at 
> org.apache.spark.api.python.PythonWorkerFactory.createThroughDaemon(PythonWorkerFactory.scala:55)
>         at 
> org.apache.spark.api.python.PythonWorkerFactory.create(PythonWorkerFactory.scala:42)
>         at org.apache.spark.SparkEnv.createPythonWorker(SparkEnv.scala:97)
>         at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:57)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
> This can be caused by a few things:
>   (1) PYTHONPATH is not set
>   (2) PYTHONPATH does not contain the python directory (or jar, in the case 
> of YARN)
>   (3) The jar does not contain pyspark files (YARN)
>   (4) The jar does not contain py4j files (YARN)
> We should have explicit error messages for each one of them. For (2 - 4), we 
> should print out the PYTHONPATH so the user doesn't have to SSH into the 
> executor machines themselves to figure this out.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to