Did you try increasing sql partitions? On Tue, Aug 25, 2015 at 11:06 AM, kundan kumar <iitr.kun...@gmail.com> wrote:
> I am running this query on a data size of 4 billion rows and > getting org.apache.spark.shuffle.FetchFailedException error. > > select adid,position,userid,price > from ( > select adid,position,userid,price, > dense_rank() OVER (PARTITION BY adlocationid ORDER BY price DESC) as rank > FROM trainInfo) as tmp > WHERE rank <= 2 > > > I have attached the error logs from spark-sql terminal. > > Please suggest what is the reason for these kind of errors and how can I > resolve them. > > > Regards, > Kundan > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org >