Thank you. HADOOP_CONF_DIR has been missing.

On Wed, Sep 24, 2014 at 4:48 PM, Matt Narrell <matt.narr...@gmail.com>
wrote:

> Yes, this works.  Make sure you have HADOOP_CONF_DIR set on your Spark
> machines
>
> mn
>
> On Sep 24, 2014, at 5:35 AM, Petr Novak <oss.mli...@gmail.com> wrote:
>
> Hello,
> if our Hadoop cluster is configured with HA and "fs.defaultFS" points to a
> namespace instead of a namenode hostname - hdfs://<namespace_name>/ - then
> our Spark job fails with exception. Is there anything to configure or it is
> not implemented?
>
>
> Exception in thread "main" org.apache.spark.SparkException: Job aborted
> due to stage failure: Task 0 in stage 1.0 failed 4 times, most recent
> failure: Lost task 0.3 in stage 1.0 (TID 4, <hostanme>):
>
>
> java.lang.IllegalArgumentException: java.net.UnknownHostException:
> *<namespace_name>*
>
>
> Many thanks,
> P.
>
>
>

Reply via email to