There could be multiple of them, why its not being generated even after setting the ulimit appropriately.
Try out the options listed on this thread: http://stackoverflow.com/questions/7732983/core-dump-file-is-not-generated On Thu, Jun 30, 2016 at 2:25 AM, prateek arora <prateek.arora...@gmail.com> wrote: > Thanks for the information. My problem is resolved now . > > > > I have one more issue. > > > > I am not able to save core dump file. Always shows *“# Failed to write > core dump. Core dumps have been disabled. To enable core dumping, try > "ulimit -c unlimited" before starting Java again"* > > > > I set core dump limit to unlimited in all nodes. Using below settings > Edit /etc/security/limits.conf file and add " * soft core unlimited " > line. > > I rechecked using : $ ulimit -all > > core file size (blocks, -c) unlimited > data seg size (kbytes, -d) unlimited > scheduling priority (-e) 0 > file size (blocks, -f) unlimited > pending signals (-i) 241204 > max locked memory (kbytes, -l) 64 > max memory size (kbytes, -m) unlimited > open files (-n) 1024 > pipe size (512 bytes, -p) 8 > POSIX message queues (bytes, -q) 819200 > real-time priority (-r) 0 > stack size (kbytes, -s) 8192 > cpu time (seconds, -t) unlimited > max user processes (-u) 241204 > virtual memory (kbytes, -v) unlimited > file locks (-x) unlimited > > but when my spark application crash , show error " Failed to > write core dump. Core dumps have been disabled. To enablecore dumping, try > "ulimit -c unlimited" before starting Java again”. > > > Regards > > Prateek > > > > > > On Wed, Jun 29, 2016 at 9:30 PM, dhruve ashar <dhruveas...@gmail.com> > wrote: > >> You can look at the yarn-default configuration file. >> >> Check your log related settings to see if log aggregation is enabled or >> also the log retention duration to see if its too small and files are being >> deleted. >> >> On Wed, Jun 29, 2016 at 4:47 PM, prateek arora < >> prateek.arora...@gmail.com> wrote: >> >>> >>> Hi >>> >>> My Spark application was crashed and show information >>> >>> LogType:stdout >>> Log Upload Time:Wed Jun 29 14:38:03 -0700 2016 >>> LogLength:1096 >>> Log Contents: >>> # >>> # A fatal error has been detected by the Java Runtime Environment: >>> # >>> # SIGILL (0x4) at pc=0x00007f67baa0d221, pid=12207, tid=140083473176320 >>> # >>> # JRE version: Java(TM) SE Runtime Environment (7.0_67-b01) (build >>> 1.7.0_67-b01) >>> # Java VM: Java HotSpot(TM) 64-Bit Server VM (24.65-b04 mixed mode >>> linux-amd64 compressed oops) >>> # Problematic frame: >>> # C [libcaffe.so.1.0.0-rc3+0x786221] sgemm_kernel+0x21 >>> # >>> # Failed to write core dump. Core dumps have been disabled. To enable >>> core >>> dumping, try "ulimit -c unlimited" before starting Java again >>> # >>> # An error report file with more information is saved as: >>> # >>> >>> /yarn/nm/usercache/ubuntu/appcache/application_1467236060045_0001/container_1467236060045_0001_01_000003/hs_err_pid12207.log >>> >>> >>> >>> but I am not able to found >>> >>> "/yarn/nm/usercache/ubuntu/appcache/application_1467236060045_0001/container_1467236060045_0001_01_000003/hs_err_pid12207.log" >>> file . its deleted automatically after Spark application >>> finished >>> >>> >>> how to retain report file , i am running spark with yarn . >>> >>> Regards >>> Prateek >>> >>> >>> >>> -- >>> View this message in context: >>> http://apache-spark-user-list.1001560.n3.nabble.com/Error-report-file-is-deleted-automatically-after-spark-application-finished-tp27247.html >>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>> >>> --------------------------------------------------------------------- >>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >>> >>> >> >> >> -- >> -Dhruve Ashar >> >> > -- -Dhruve Ashar