Hi, Below is the log form the worker.

15/07/14 17:18:56 ERROR FileAppender: Error writing stream to file
/spark/app-20150714171703-0004/5/stderr

java.io.IOException: Stream closed

at java.io.BufferedInputStream.getBufIfOpen(BufferedInputStream.java:170)

at java.io.BufferedInputStream.read1(BufferedInputStream.java:283)

at java.io.BufferedInputStream.read(BufferedInputStream.java:345)

at java.io.FilterInputStream.read(FilterInputStream.java:107)

at
org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:70)

at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)

at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)

at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1772)

at
org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)

15/07/14 17:18:57 INFO Worker: Executor app-20150714171703-0004/5 finished
with state KILLED exitStatus 143

15/07/14 17:18:57 INFO Worker: Cleaning up local directories for
application app-20150714171703-0004

15/07/14 17:18:57 WARN ReliableDeliverySupervisor: Association with remote
system [akka.tcp://sparkExecutor@10.10.10.1:52635] has failed, address is
now gated for [5000] ms. Reason is: [Disassociated].

Reply via email to