Hi, I am using Spark 1.5.2 with Scala 2.10 and my Spark job keeps failing with exit code 143 . except one job where I am using unionAll and groupBy operation on multiple columns .
Please advice me the options to optimize it . The one option which I am using it now --conf spark.executor.extraJavaOptions -XX:MaxPermSize=1024m -XX:PermSize=256m --conf spark.driver.extraJavaOptions -XX:MaxPermSize=1024m -XX:PermSize=256m --conf spark.yarn.executor.memoryOverhead=1024 Need to know the best practices/better ways to optimize code. Thanks, Divya