It is giving OOM at 32GB ? Something looks wrong with that ... that is already on the higher side.
Regards, Mridul On Fri, Oct 30, 2015 at 11:28 AM, shane knapp <skn...@berkeley.edu> wrote: > here's the current heap settings on our workers: > InitialHeapSize == 2.1G > MaxHeapSize == 32G > > system ram: 128G > > we can bump it pretty easily... it's just a matter of deciding if we > want to do this globally (super easy, but will affect ALL maven builds > on our system -- not just spark) or on a per-job basis (this doesn't > scale that well). > > thoughts? > > On Fri, Oct 30, 2015 at 9:47 AM, Ted Yu <yuzhih...@gmail.com> wrote: >> This happened recently on Jenkins: >> >> https://amplab.cs.berkeley.edu/jenkins/job/Spark-Master-Maven-with-YARN/HADOOP_PROFILE=hadoop-2.3,label=spark-test/3964/console >> >> On Sun, Oct 18, 2015 at 7:54 AM, Ted Yu <yuzhih...@gmail.com> wrote: >>> >>> From >>> https://amplab.cs.berkeley.edu/jenkins/job/Spark-Master-Maven-with-YARN/HADOOP_PROFILE=hadoop-2.4,label=spark-test/3846/console >>> : >>> >>> SparkListenerSuite: >>> - basic creation and shutdown of LiveListenerBus >>> - bus.stop() waits for the event queue to completely drain >>> - basic creation of StageInfo >>> - basic creation of StageInfo with shuffle >>> - StageInfo with fewer tasks than partitions >>> - local metrics >>> - onTaskGettingResult() called when result fetched remotely *** FAILED *** >>> org.apache.spark.SparkException: Job aborted due to stage failure: Task >>> 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage >>> 0.0 (TID 0, localhost): java.lang.OutOfMemoryError: Java heap space >>> at java.util.Arrays.copyOf(Arrays.java:2271) >>> at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113) >>> at >>> java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93) >>> at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:140) >>> at >>> java.io.ObjectOutputStream$BlockDataOutputStream.write(ObjectOutputStream.java:1852) >>> at java.io.ObjectOutputStream.write(ObjectOutputStream.java:708) >>> at org.apache.spark.util.Utils$.writeByteBuffer(Utils.scala:182) >>> at >>> org.apache.spark.scheduler.DirectTaskResult$$anonfun$writeExternal$1.apply$mcV$sp(TaskResult.scala:52) >>> at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1160) >>> at >>> org.apache.spark.scheduler.DirectTaskResult.writeExternal(TaskResult.scala:49) >>> at >>> java.io.ObjectOutputStream.writeExternalData(ObjectOutputStream.java:1458) >>> at >>> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1429) >>> at >>> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177) >>> at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:347) >>> at >>> org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:44) >>> at >>> org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:101) >>> at >>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:256) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> >>> >>> Should more heap be given to test suite ? >>> >>> >>> Cheers >> >> > > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > For additional commands, e-mail: dev-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org