Re: Cancelled Key Exceptions on Massive Join

2014-11-16 Thread Akhil Das
This usually happens when one of the worker is stuck on GC Pause and it times out. Enable the following configurations while creating sparkContext: sc.set(spark.rdd.compress,true) sc.set(spark.storage.memoryFraction,1) sc.set(spark.core.connection.ack.wait.timeout,6000)

Cancelled Key Exceptions on Massive Join

2014-11-14 Thread Ganelin, Ilya
Hello all. I have been running a Spark Job that eventually needs to do a large join. 24 million x 150 million A broadcast join is infeasible in this instance clearly, so I am instead attempting to do it with Hash Partitioning by defining a custom partitioner as: class