Hi, I am running a Spark Job on the yarn cluster. The spark job is a spark streaming application which is reading JSON from a kafka topic , inserting the JSON values to hbase tables via Phoenix , ands then sending out certain messages to a websocket if the JSON satisfies a certain criteria.
My cluster is a 3 node cluster with 24GB ram and 24 cores in total. Now : 1. when I am submitting the job with 10GB memory, the application fails saying memory is insufficient to run the job 2. The job is submitted with 6G ram. However, it does not run successfully always.Common issues faced : a. Container exited with a non-zero exit code 1 , and after multiple such warning the job is finished. d. The failed job notifies that it was unable to find a file in HDFS which is something like _hadoop_conf_xxxxxx.zip Can someone pls let me know why am I seeing the above 2 issues. Thanks, Siddharth Ubale,