My users use Spark 1.5.1 in standalone mode on an HPC cluster, with a
smattering still using 1.4.0

I have been getting reports of errors like this: 

15/12/21 15:40:33 ERROR FileAppender: Error writing stream to file
/scratch/spark/work/app-20151221150645-0000/3/stdout
java.io.IOException: Stream closed
  at java.io.BufferedInputStream.getBufIfOpen(BufferedInputStream.java:162)
  at java.io.BufferedInputStream.read1(BufferedInputStream.java:272)
  at java.io.BufferedInputStream.read(BufferedInputStream.java:334)
  at java.io.FilterInputStream.read(FilterInputStream.java:107)
  at
org.apache.spark.util.logging.FileAppender.appendStreamToFile(FileAppender.scala:70)
  at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply$mcV$sp(FileAppender.scala:39)
  at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
  at
org.apache.spark.util.logging.FileAppender$$anon$1$$anonfun$run$1.apply(FileAppender.scala:39)
  at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
  at
org.apache.spark.util.logging.FileAppender$$anon$1.run(FileAppender.scala:38)
'

So far I have been unable to reproduce reliably, but does anyone have any
ideas? 



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/error-writing-to-stdout-tp25758.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to