Hello everyone,
I have a question about running the Graph analytics benchmark on the Xeon
server.
I typed in these commands sequentially
docker create --name data cloudsuite/twitter-dataset-graph
docker network create spark-net
docker run -dP --net spark-net --hostname spark-master --name spark-master
cloudsuite/spark master
docker run -dP --net spark-net --volumes-from data --name spark-worker-01
cloudsuite/spark worker spark://spark-master:7077
docker run --rm --net spark-net --volumes-from data cloudsuite/graph-analytics
--driver-memory 1g --executor-memory 4g --master spark://spark-master:7077 2>&1
| tee graph_analytics_worker_mem_1-to-1.log
and it shows some errors which is indicated as OutOfMemoryError: Java heap
space.
17/08/02 00:19:18 INFO TaskSetManager: Starting task 29.1 in stage 0.0 (TID 40,
172.18.0.3, executor 0, partition 29, PROCESS_LOCAL, 5956 bytes)
17/08/02 00:19:34 ERROR TaskSchedulerImpl: Lost executor 0 on 172.18.0.3:
Remote RPC client disassociated. Likely due to containers exceeding thresholds,
or network issues. Check driver logs for WARN messages.
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.OutOfMemoryError: Java heap space
at
org.apache.spark.graphx.impl.EdgePartitionBuilder$mcI$sp.toEdgePartition$mcI$sp(EdgePartitionBuilder.scala:42)
at
org.apache.spark.graphx.GraphLoader$$anonfun$1.apply(GraphLoader.scala:92)
at
org.apache.spark.graphx.GraphLoader$$anonfun$1.apply(GraphLoader.scala:75)
at
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsWithIndex$1$$anonfun$apply$26.apply(RDD.scala:843)
at
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsWithIndex$1$$anonfun$apply$26.apply(RDD.scala:843)
at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
at org.apache.spark.rdd.RDD$$anonfun$8.apply(RDD.scala:336)
at org.apache.spark.rdd.RDD$$anonfun$8.apply(RDD.scala:334)
at
org.apache.spark.storage.BlockManager$$anonfun$doPutIterator$1.apply(BlockManager.scala:957)
at
org.apache.spark.storage.BlockManager$$anonfun$doPutIterator$1.apply(BlockManager.scala:948)
at org.apache.spark.storage.BlockManager.doPut(BlockManager.scala:888)
at
org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:948)
at
org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:694)
at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:334)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:285)
at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
at org.apache.spark.scheduler.Task.run(Task.scala:99)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
17/08/02 00:31:34 WARN JobProgressListener: Task start for unknown stage 2
17/08/02 00:31:34 INFO SparkContext: Invoking stop() from shutdown hook
17/08/02 00:31:34 INFO SparkUI: Stopped Spark web UI at http://172.18.0.4:4040
17/08/02 00:31:34 INFO StandaloneSchedulerBackend: Shutting down all executors
17/08/02 00:31:34 INFO CoarseGrainedSchedulerBackend$DriverEndpoint: Asking
each executor to shut down
17/08/02 00:31:34 INFO MapOutputTrackerMasterEndpoint:
MapOutputTrackerMasterEndpoint stopped!
17/08/02 00:31:34 INFO MemoryStore: MemoryStore cleared
17/08/02 00:31:34 INFO BlockManager: BlockManager stopped
17/08/02 00:31:34 INFO BlockManagerMaster: BlockManagerMaster stopped
17/08/02 00:31:34 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:
OutputCommitCoordinator stopped!
17/08/02 00:31:34 INFO SparkContext: Successfully stopped SparkContext
17/08/02 00:31:34 INFO ShutdownHookManager: Shutdown hook called
17/08/02 00:31:34 INFO ShutdownHookManager: Deleting directory
/tmp/spark-a1248aba-7976-40a6-9ef1-b5dd284e9472
My servers hardware specs are
CPU: Intel(R) Xeon(R) CPU E5-2630
MemTotal: 65810820 kB
Kernel: 4.9.20
OS: Ubuntu 14.04.5 LTS
So, if anybody knows the solution for that please let me know
Thanks in advance.
Best Regards
Ziho Shin (Joe) | Research Student
Cloud Computing Research Group | Infra/Core Software Technology Research
Division
Electronics and Telecommunications Research Institute(ETRI)
[email protected]<mailto:[email protected]> | +82 42 860 3919 | +82 10 9241 2276
218 Gajeongno, Yuseong-gu, Daejeon, 34129, S.Korea
CONFIDENTIALITY NOTICE: The contents of this email message and any attachments
are intended solely for the addressee(s) and may contain confidential and/or
privileged information and may be legally protected from disclosure. If you are
not the intended recipient of this message or their agent, or if this message
has been addressed to you in error, please immediately alert the sender by
reply email and then delete this message and any attachments. If you are not
the intended recipient, you are hereby notified that any use, dissemination,
copying, or storage of this message or its attachments is strictly prohibited.