Yes, this works. Make sure you have HADOOP_CONF_DIR set on your Spark machines
mn On Sep 24, 2014, at 5:35 AM, Petr Novak <oss.mli...@gmail.com> wrote: > Hello, > if our Hadoop cluster is configured with HA and "fs.defaultFS" points to a > namespace instead of a namenode hostname - hdfs://<namespace_name>/ - then > our Spark job fails with exception. Is there anything to configure or it is > not implemented? > > > Exception in thread "main" org.apache.spark.SparkException: Job aborted due > to stage failure: Task 0 in stage 1.0 failed 4 times, most recent failure: > Lost task 0.3 in stage 1.0 (TID 4, <hostanme>): > > > > java.lang.IllegalArgumentException: java.net.UnknownHostException: > <namespace_name> > > > > Many thanks, > P. >