[ 
https://issues.apache.org/jira/browse/SPARK-1616?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13994044#comment-13994044
 ] 

Marcelo Vanzin commented on SPARK-1616:
---------------------------------------

Hi Prasad,

This doesn't really sound like a bug, but a mismatch between your expectations 
and Spark's.

When you tell a Spark job to read data from a file, Spark expects the file to 
be available to all the workers. This can be achieved in several ways:

* Using a distributed file system such as HDFS
* Using a networked file system such as NFS
* Using Spark's file distribution mechanism, which will copy the file to the 
workers for you (e.g. spark-submit's --files argument if you run 1.0)
* Manually copying the file like you did

But Spark will not automatically copy data to worker nodes on your behalf.

> input file not found issue 
> ---------------------------
>
>                 Key: SPARK-1616
>                 URL: https://issues.apache.org/jira/browse/SPARK-1616
>             Project: Spark
>          Issue Type: Bug
>          Components: Input/Output
>    Affects Versions: 0.9.0
>         Environment: Linux 2.6.18-348.3.1.el5 
>            Reporter: prasad potipireddi
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to