Hi , I am getting below error while I am trying to select data using SPARK SQL from a RDD table.
java.lang.OutOfMemoryError: GC overhead limit exceeded "Spark Context Cleaner" java.lang.InterruptedException The file or table size is around 113 GB and I am running SPARK 1.4 on a standalone cluster. Tried to extend the heap size but extending to 64GB also didn't help. I would really appreciate any help on this. Thanks, Debobrota