bq. check application tracking
page:http://slave1:8088/proxy/application_1452763526769_0011/
Then <http://slave1:8088/proxy/application_1452763526769_0011/Then>, ...

Have you done the above to see what error was in each attempt ?

Which Spark / hadoop release are you using ?

Thanks

On Fri, Jan 15, 2016 at 5:58 AM, Siddharth Ubale <
siddharth.ub...@syncoms.com> wrote:

> Hi,
>
>
>
> I am trying to run a Spark streaming application in yarn-cluster mode.
> However I am facing an issue where the job ends asking for a particular
> Hadoop_conf_**.zip file in hdfs location.
>
> Can any one guide with this?
>
> The application works fine in local mode only it stops abruptly for want
> of memory.
>
>
>
> Below is the error stack trace:
>
>
>
> diagnostics: Application application_1452763526769_0011 failed 2 times due
> to AM Container for appattempt_1452763526769_0011_000002 exited with
> exitCode: -1000
>
> For more detailed output, check application tracking page:
> http://slave1:8088/proxy/application_1452763526769_0011/Then, click on
> links to logs of each attempt.
>
> Diagnostics: File does not exist:
> hdfs://slave1:9000/user/hduser/.sparkStaging/application_1452763526769_0011/__hadoop_conf__1057113228186399290.zip
>
> *java.io.FileNotFoundException: File does not exist:
> hdfs://slave1:9000/user/hduser/.sparkStaging/application_1452763526769_0011/__hadoop_conf__1057113228186399290.zip*
>
>                 at
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1122)
>
>                 at
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
>
>                 at
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>
>                 at
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
>
>                 at
> org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251)
>
>                 at
> org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61)
>
>                 at
> org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359)
>
>                 at
> org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357)
>
>                 at java.security.AccessController.doPrivileged(Native
> Method)
>
>                 at javax.security.auth.Subject.doAs(Subject.java:422)
>
>                 at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>
>                 at
> org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356)
>
>                 at
> org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
>
>                 at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>
>                 at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>
>                 at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>
>                 at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>
>                 at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>
>                 at java.lang.Thread.run(Thread.java:745)
>
>
>
> Failing this attempt. Failing the application.
>
>                 ApplicationMaster host: N/A
>
>                 ApplicationMaster RPC port: -1
>
>                 queue: default
>
>                 start time: 1452866026622
>
>                 final status: FAILED
>
>                 tracking URL:
> http://slave1:8088/cluster/app/application_1452763526769_0011
>
>                 user: hduser
>
> Exception in thread "main" org.apache.spark.SparkException: Application
> application_1452763526769_0011 finished with failed status
>
>                 at
> org.apache.spark.deploy.yarn.Client.run(Client.scala:841)
>
>                 at
> org.apache.spark.deploy.yarn.Client$.main(Client.scala:867)
>
>                 at org.apache.spark.deploy.yarn.Client.main(Client.scala)
>
>                 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
>
>                 at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>
>                 at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>
>                 at java.lang.reflect.Method.invoke(Method.java:497)
>
>                 at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:664)
>
>                 at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:169)
>
>                 at
> org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:192)
>
>                 at
> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:111)
>
>                 at
> org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>
> 16/01/15 19:23:53 INFO Utils: Shutdown hook called
>
> 16/01/15 19:23:53 INFO Utils: Deleting directory
> /tmp/spark-b6ebcb83-efff-432a-9a7a-b4764f482d81
>
> java.lang.UNIXProcess$ProcessPipeOutputStream@7a0a6f73  1
>
>
>
>
>
>
>
> Siddharth Ubale,
>
> *Synchronized Communications *
>
> *#43, Velankani Tech Park, Block No. II, *
>
> *3rd Floor, Electronic City Phase I,*
>
> *Bangalore – 560 100*
>
> *Tel : +91 80 3202 4060 <%2B91%2080%203202%204060>*
>
> *Web:* *www.syncoms.com* <http://www.syncoms.com/>
>
> *[image: LogoNEWmohLARGE]*
>
> *London*|*Bangalore*|*Orlando*
>
>
>
> *we innovate, plan, execute, and transform the business​*
>
>
>

Reply via email to