Can you please direct me to the right way of doing this.

On Mon, Sep 15, 2014 at 10:18 PM, rapelly kartheek <kartheek.m...@gmail.com>
wrote:

> I came across these APIs in one the scala tutorials over the net.
>
> On Mon, Sep 15, 2014 at 10:14 PM, Mohit Jaggi <mohitja...@gmail.com>
> wrote:
>
>> But the above APIs are not for HDFS.
>>
>> On Mon, Sep 15, 2014 at 9:40 AM, rapelly kartheek <
>> kartheek.m...@gmail.com> wrote:
>>
>>> Yes. I have HDFS. My cluster has 5 nodes. When I run the above commands,
>>> I see that the file gets created in the master node. But, there wont be any
>>> data written to it.
>>>
>>>
>>> On Mon, Sep 15, 2014 at 10:06 PM, Mohit Jaggi <mohitja...@gmail.com>
>>> wrote:
>>>
>>>> Is this code running in an executor? You need to make sure the file is
>>>> accessible on ALL executors. One way to do that is to use a distributed
>>>> filesystem like HDFS or GlusterFS.
>>>>
>>>> On Mon, Sep 15, 2014 at 8:51 AM, rapelly kartheek <
>>>> kartheek.m...@gmail.com> wrote:
>>>>
>>>>> Hi
>>>>>
>>>>> I am trying to perform some read/write file operations in spark.
>>>>> Somehow I am neither able to write to a file nor read.
>>>>>
>>>>> import java.io._
>>>>>
>>>>>       val writer = new PrintWriter(new File("test.txt" ))
>>>>>
>>>>>       writer.write("Hello Scala")
>>>>>
>>>>>
>>>>> Can someone please tell me how to perform file I/O in spark.
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to