ok solved it. as it happened in spark/conf i also had a file called
core.site.xml (with some tachyone related stuff in it) so thats why it
ignored /etc/hadoop/conf/core-site.xml




On Fri, Jun 20, 2014 at 3:24 PM, Koert Kuipers <ko...@tresata.com> wrote:

> i put some logging statements in yarn.Client and that confirms its using
> local filesystem:
> 14/06/20 15:20:33 INFO Client: fs.defaultFS is file:///
>
> so somehow fs.defaultFS is not being picked up from
> /etc/hadoop/conf/core-site.xml, but spark does correctly pick up
> yarn.resourcemanager.hostname from /etc/hadoop/conf/yarn-site.xml
>
> strange!
>
>
> On Fri, Jun 20, 2014 at 1:26 PM, Koert Kuipers <ko...@tresata.com> wrote:
>
>> in /etc/hadoop/conf/core-site.xml:
>>   <property>
>>     <name>fs.defaultFS</name>
>>     <value>hdfs://cdh5-yarn.tresata.com:8020</value>
>>   </property>
>>
>>
>> also hdfs seems the default:
>> [koert@cdh5-yarn ~]$ hadoop fs -ls /
>> Found 5 items
>> drwxr-xr-x   - hdfs supergroup          0 2014-06-19 12:31 /data
>> drwxrwxrwt   - hdfs supergroup          0 2014-06-20 12:17 /lib
>> drwxrwxrwt   - hdfs supergroup          0 2014-06-18 14:58 /tmp
>> drwxr-xr-x   - hdfs supergroup          0 2014-06-18 15:02 /user
>> drwxr-xr-x   - hdfs supergroup          0 2014-06-18 14:59 /var
>>
>> and in my spark-site.env:
>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>>
>>
>>
>> On Fri, Jun 20, 2014 at 1:04 PM, bc Wong <bcwal...@cloudera.com> wrote:
>>
>>> Koert, is there any chance that your fs.defaultFS isn't setup right?
>>>
>>>
>>> On Fri, Jun 20, 2014 at 9:57 AM, Koert Kuipers <ko...@tresata.com>
>>> wrote:
>>>
>>>>  yeah sure see below. i strongly suspect its something i misconfigured
>>>> causing yarn to try to use local filesystem mistakenly.
>>>>
>>>> *********************
>>>>
>>>> [koert@cdh5-yarn ~]$ /usr/local/lib/spark/bin/spark-submit --class
>>>> org.apache.spark.examples.SparkPi --master yarn-cluster --num-executors 3
>>>> --executor-cores 1
>>>> hdfs://cdh5-yarn/lib/spark-examples-1.0.0-hadoop2.3.0-cdh5.0.2.jar 10
>>>> 14/06/20 12:54:40 WARN NativeCodeLoader: Unable to load native-hadoop
>>>> library for your platform... using builtin-java classes where applicable
>>>> 14/06/20 12:54:40 INFO RMProxy: Connecting to ResourceManager at
>>>> cdh5-yarn.tresata.com/192.168.1.85:8032
>>>> 14/06/20 12:54:41 INFO Client: Got Cluster metric info from
>>>> ApplicationsManager (ASM), number of NodeManagers: 1
>>>> 14/06/20 12:54:41 INFO Client: Queue info ... queueName: root.default,
>>>> queueCurrentCapacity: 0.0, queueMaxCapacity: -1.0,
>>>>       queueApplicationCount = 0, queueChildQueueCount = 0
>>>> 14/06/20 12:54:41 INFO Client: Max mem capabililty of a single resource
>>>> in this cluster 8192
>>>> 14/06/20 12:54:41 INFO Client: Preparing Local resources
>>>> 14/06/20 12:54:41 WARN BlockReaderLocal: The short-circuit local reads
>>>> feature cannot be used because libhadoop cannot be loaded.
>>>> 14/06/20 12:54:41 INFO Client: Uploading
>>>> hdfs://cdh5-yarn/lib/spark-examples-1.0.0-hadoop2.3.0-cdh5.0.2.jar to
>>>> file:/home/koert/.sparkStaging/application_1403201750110_0060/spark-examples-1.0.0-hadoop2.3.0-cdh5.0.2.jar
>>>> 14/06/20 12:54:43 INFO Client: Setting up the launch environment
>>>> 14/06/20 12:54:43 INFO Client: Setting up container launch context
>>>> 14/06/20 12:54:43 INFO Client: Command for starting the Spark
>>>> ApplicationMaster: List($JAVA_HOME/bin/java, -server, -Xmx512m,
>>>> -Djava.io.tmpdir=$PWD/tmp, -Dspark.akka.retry.wait=\"30000\",
>>>> -Dspark.storage.blockManagerTimeoutIntervalMs=\"120000\",
>>>> -Dspark.storage.blockManagerHeartBeatMs=\"120000\", 
>>>> -Dspark.app.name=\"org.apache.spark.examples.SparkPi\",
>>>> -Dspark.akka.frameSize=\"10000\", -Dspark.akka.timeout=\"30000\",
>>>> -Dspark.worker.timeout=\"30000\",
>>>> -Dspark.akka.logLifecycleEvents=\"true\",
>>>> -Dlog4j.configuration=log4j-spark-container.properties,
>>>> org.apache.spark.deploy.yarn.ApplicationMaster, --class,
>>>> org.apache.spark.examples.SparkPi, --jar ,
>>>> hdfs://cdh5-yarn/lib/spark-examples-1.0.0-hadoop2.3.0-cdh5.0.2.jar,
>>>> --args  '10' , --executor-memory, 1024, --executor-cores, 1,
>>>> --num-executors , 3, 1>, <LOG_DIR>/stdout, 2>, <LOG_DIR>/stderr)
>>>> 14/06/20 12:54:43 INFO Client: Submitting application to ASM
>>>> 14/06/20 12:54:43 INFO YarnClientImpl: Submitted application
>>>> application_1403201750110_0060
>>>> 14/06/20 12:54:44 INFO Client: Application report from ASM:
>>>>      application identifier: application_1403201750110_0060
>>>>      appId: 60
>>>>      clientToAMToken: null
>>>>      appDiagnostics:
>>>>      appMasterHost: N/A
>>>>      appQueue: root.koert
>>>>      appMasterRpcPort: -1
>>>>      appStartTime: 1403283283505
>>>>      yarnAppState: ACCEPTED
>>>>      distributedFinalState: UNDEFINED
>>>>      appTrackingUrl:
>>>> http://cdh5-yarn.tresata.com:8088/proxy/application_1403201750110_0060/
>>>>      appUser: koert
>>>> 14/06/20 12:54:45 INFO Client: Application report from ASM:
>>>>      application identifier: application_1403201750110_0060
>>>>      appId: 60
>>>>      clientToAMToken: null
>>>>      appDiagnostics:
>>>>      appMasterHost: N/A
>>>>      appQueue: root.koert
>>>>      appMasterRpcPort: -1
>>>>      appStartTime: 1403283283505
>>>>      yarnAppState: ACCEPTED
>>>>      distributedFinalState: UNDEFINED
>>>>      appTrackingUrl:
>>>> http://cdh5-yarn.tresata.com:8088/proxy/application_1403201750110_0060/
>>>>      appUser: koert
>>>> 14/06/20 12:54:46 INFO Client: Application report from ASM:
>>>>      application identifier: application_1403201750110_0060
>>>>      appId: 60
>>>>      clientToAMToken: null
>>>>      appDiagnostics:
>>>>      appMasterHost: N/A
>>>>      appQueue: root.koert
>>>>      appMasterRpcPort: -1
>>>>      appStartTime: 1403283283505
>>>>      yarnAppState: ACCEPTED
>>>>      distributedFinalState: UNDEFINED
>>>>      appTrackingUrl:
>>>> http://cdh5-yarn.tresata.com:8088/proxy/application_1403201750110_0060/
>>>>      appUser: koert
>>>> 14/06/20 12:54:47 INFO Client: Application report from ASM:
>>>>      application identifier: application_1403201750110_0060
>>>>      appId: 60
>>>>      clientToAMToken: null
>>>>      appDiagnostics: Application application_1403201750110_0060 failed
>>>> 2 times due to AM Container for appattempt_1403201750110_0060_000002 exited
>>>> with  exitCode: -1000 due to: File
>>>> file:/home/koert/.sparkStaging/application_1403201750110_0060/spark-examples-1.0.0-hadoop2.3.0-cdh5.0.2.jar
>>>> does not exist
>>>> .Failing this attempt.. Failing the application.
>>>>      appMasterHost: N/A
>>>>      appQueue: root.koert
>>>>      appMasterRpcPort: -1
>>>>      appStartTime: 1403283283505
>>>>      yarnAppState: FAILED
>>>>      distributedFinalState: FAILED
>>>>      appTrackingUrl:
>>>> cdh5-yarn.tresata.com:8088/cluster/app/application_1403201750110_0060
>>>>      appUser: koert
>>>>
>>>>
>>>>
>>>>
>>>> On Fri, Jun 20, 2014 at 12:42 PM, Marcelo Vanzin <van...@cloudera.com>
>>>> wrote:
>>>>
>>>>> Hi Koert,
>>>>>
>>>>> Could you provide more details? Job arguments, log messages, errors,
>>>>> etc.
>>>>>
>>>>> On Fri, Jun 20, 2014 at 9:40 AM, Koert Kuipers <ko...@tresata.com>
>>>>> wrote:
>>>>> > i noticed that when i submit a job to yarn it mistakenly tries to
>>>>> upload
>>>>> > files to local filesystem instead of hdfs. what could cause this?
>>>>> >
>>>>> > in spark-env.sh i have HADOOP_CONF_DIR set correctly (and
>>>>> spark-submit does
>>>>> > find yarn), and my core-site.xml has a fs.defaultFS that is hdfs,
>>>>> not local
>>>>> > filesystem.
>>>>> >
>>>>> > thanks! koert
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Marcelo
>>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to