So, just before running the job, if you run the HDFS command at a shell
prompt: "hdfs dfs -ls hdfs://172.31.42.10:54310/...../weblogReadResult".
Does it say the path doesn't exist?

Dean Wampler, Ph.D.
Author: Programming Scala, 2nd Edition
<http://shop.oreilly.com/product/0636920033073.do> (O'Reilly)
Typesafe <http://typesafe.com>
@deanwampler <http://twitter.com/deanwampler>
http://polyglotprogramming.com

On Mon, Aug 10, 2015 at 7:58 AM, Yasemin Kaya <godo...@gmail.com> wrote:

> Thanx Dean, i am giving unique output path and in every time i also delete
> the directory before i run the job.
>
> 2015-08-10 15:30 GMT+03:00 Dean Wampler <deanwamp...@gmail.com>:
>
>> Following Hadoop conventions, Spark won't overwrite an existing
>> directory. You need to provide a unique output path every time you run the
>> program, or delete or rename the target directory before you run the job.
>>
>> dean
>>
>> Dean Wampler, Ph.D.
>> Author: Programming Scala, 2nd Edition
>> <http://shop.oreilly.com/product/0636920033073.do> (O'Reilly)
>> Typesafe <http://typesafe.com>
>> @deanwampler <http://twitter.com/deanwampler>
>> http://polyglotprogramming.com
>>
>> On Mon, Aug 10, 2015 at 7:08 AM, Yasemin Kaya <godo...@gmail.com> wrote:
>>
>>> Hi,
>>>
>>> I have EC2 cluster, and am using spark 1.3, yarn and HDFS . When i
>>> submit at local there is no problem , but i run at cluster, saveAsTextFile
>>> doesn't work."*It says me User class threw exception: Output directory
>>> hdfs://172.31.42.10:54310/...../weblogReadResult
>>> <http://172.31.42.10:54310/...../weblogReadResult> already exists*"
>>>
>>> Is there anyone can help me about this issue ?
>>>
>>> Best,
>>> yasemin
>>>
>>>
>>>
>>> --
>>> hiç ender hiç
>>>
>>
>>
>
>
> --
> hiç ender hiç
>

Reply via email to