Glad it worked! Actually I got similar issues even with Spark Streaming v1.2.x based drivers.
Think also that the default config in Spark on EMR is 512m ! Roberto On Thu, Jun 25, 2015 at 1:20 AM, Srikanth <srikanth...@gmail.com> wrote: > That worked. Thanks! > > I wonder what changed in 1.4 to cause this. It wouldn't work with anything > less than 256m for a simple piece of code. > 1.3.1 used to work with default(64m I think) > > Srikanth > > On Wed, Jun 24, 2015 at 12:47 PM, Roberto Coluccio < > roberto.coluc...@gmail.com> wrote: > >> Did you try to pass it with >> >> --driver-java-options -XX:MaxPermSize=256m >> >> as spark-shell input argument? >> >> >> Roberto >> >> >> On Wed, Jun 24, 2015 at 5:57 PM, stati <srikanth...@gmail.com> wrote: >> >>> Hello, >>> >>> I moved from 1.3.1 to 1.4.0 and started receiving >>> "java.lang.OutOfMemoryError: PermGen space" when I use spark-shell. >>> Same Scala code works fine in 1.3.1 spark-shell. I was loading same set >>> of >>> external JARs and have same imports in 1.3.1. >>> >>> I tried increasing perm size to 256m. I still got OOM. >>> >>> /SPARK_REPL_OPTS="-XX:MaxPermSize=256m" bin/spark-shell --master >>> spark://machu:7077 --total-executor-cores 12 --packages >>> com.databricks:spark-csv_2.10:1.0.3 --packages joda-time:joda-time:2.8.1 >>> / >>> >>> Spark UI "Environment" tab didn't show "-XX:MaxPermSize". I'm not sure if >>> this config was picked up. >>> This is standalone mode. >>> >>> Any pointers to next step? >>> >>> >>> >>> >>> >>> -- >>> View this message in context: >>> http://apache-spark-user-list.1001560.n3.nabble.com/java-lang-OutOfMemoryError-PermGen-space-tp23472.html >>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> For additional commands, e-mail: user-h...@spark.apache.org >>> >>> >> >