Hi, In spark streaming job i had the following setting
this.jsc.getConf().set("spark.driver.maxResultSize", “0”); and i got the error in the job as below User class threw exception: Job aborted due to stage failure: Total size of serialized results of 120 tasks (1082.2 MB) is bigger than spark.driver.maxResultSize (1024.0 MB) Basically i realized that as default value is 1 GB. I changed the configuration as below. this.jsc.getConf().set("spark.driver.maxResultSize", “2g”); and when i ran the job it gave the error User class threw exception: Job aborted due to stage failure: Total size of serialized results of 120 tasks (1082.2 MB) is bigger than spark.driver.maxResultSize (1024.0 MB) So, basically the change i made is not been considered in the job. so my question is - "spark.driver.maxResultSize", “2g” is this the right way to change or any other way to do it. - Is this a bug in spark 1.3 or something or any one had this issue before?