I am using spark 2.1.0 On Fri, Feb 2, 2018 at 5:08 PM, Pralabh Kumar <pralabhku...@gmail.com> wrote:
> Hi > > I am performing broadcast join where my small table is 1 gb . I am > getting following error . > > I am using > > > org.apache.spark.SparkException: > . Available: 0, required: 28869232. To avoid this, increase > spark.kryoserializer.buffer.max value > > > > I increase the value to > > spark.conf.set("spark.kryoserializer.buffer.max","2g") > > > But I am still getting the error . > > Please help > > Thx >