Re: GC overhead exceeded

2017-08-18 Thread KhajaAsmath Mohammed
you are using YARN? > > -Pat > > From: KhajaAsmath Mohammed <mdkhajaasm...@gmail.com> > Date: Friday, August 18, 2017 at 5:30 AM > To: Pralabh Kumar <pralabhku...@gmail.com> > Cc: "user @spark" <user@spark.apache.org> > Subject: Re: GC overhead ex

Re: GC overhead exceeded

2017-08-18 Thread Patrick Alwell
: KhajaAsmath Mohammed <mdkhajaasm...@gmail.com> Date: Friday, August 18, 2017 at 5:30 AM To: Pralabh Kumar <pralabhku...@gmail.com> Cc: "user @spark" <user@spark.apache.org> Subject: Re: GC overhead exceeded It is just a sql from hive table with transformation if adding

Re: GC overhead exceeded

2017-08-18 Thread KhajaAsmath Mohammed
It is just a sql from hive table with transformation if adding 10 more columns calculated for currency. Input size for this query is 2 months which has around 450gb data. I added persist but it didn't help. Also the executor memory is 8g . Any suggestions please ? Sent from my iPhone > On

Re: GC overhead exceeded

2017-08-17 Thread Pralabh Kumar
what's is your exector memory , please share the code also On Fri, Aug 18, 2017 at 10:06 AM, KhajaAsmath Mohammed < mdkhajaasm...@gmail.com> wrote: > > HI, > > I am getting below error when running spark sql jobs. This error is thrown > after running 80% of tasks. any solution? > >

GC overhead exceeded

2017-08-17 Thread KhajaAsmath Mohammed
HI, I am getting below error when running spark sql jobs. This error is thrown after running 80% of tasks. any solution? spark.storage.memoryFraction=0.4 spark.sql.shuffle.partitions=2000 spark.default.parallelism=100 #spark.eventLog.enabled=false #spark.scheduler.revive.interval=1s