Or you can increase the driver heap space (export _JAVA_OPTIONS="-Xmx5g")

Thanks
Best Regards

On Wed, Sep 2, 2015 at 11:57 PM, Mike Hynes <91m...@gmail.com> wrote:

> Just a thought; this has worked for me before on standalone client
> with a similar OOM error in a driver thread. Try setting:
> export SPARK_DAEMON_MEMORY=4G #or whatever size you can afford on your
> machine
> in your environment/spark-env.sh before running spark-submit.
> Mike
>
> On 9/2/15, ankit tyagi <ankittyagi.mn...@gmail.com> wrote:
> > Hi All,
> >
> > I am using spark-sql 1.3.1 with hadoop 2.4.0 version.  I am running sql
> > query against parquet files and wanted to save result on s3 but looks
> like
> > https://issues.apache.org/jira/browse/SPARK-2984 problem still coming
> while
> > saving data to s3.
> >
> > Hence Now i am saving result on hdfs and with the help
> > of JavaSparkListener, copying file from hdfs to s3 with hadoop fileUtil
> > in onApplicationEnd method. But  my job is getting failed with OOM in
> spark
> > driver.
> >
> > *5/09/02 04:17:57 INFO cluster.YarnClusterSchedulerBackend: Asking each
> > executor to shut down*
> > *15/09/02 04:17:59 INFO
> > scheduler.OutputCommitCoordinator$OutputCommitCoordinatorActor:
> > OutputCommitCoordinator stopped!*
> > *Exception in thread "Reporter" *
> > *Exception: java.lang.OutOfMemoryError thrown from the
> > UncaughtExceptionHandler in thread "Reporter"*
> > *Exception in thread "SparkListenerBus" *
> > *Exception: java.lang.OutOfMemoryError thrown from the
> > UncaughtExceptionHandler in thread "SparkListenerBus"*
> > *Exception in thread "Driver" *
> > *Exception: java.lang.OutOfMemoryError thrown from the
> > UncaughtExceptionHandler in thread "Driver"*
> >
> >
> > Strage part is, result is getting saved on HDFS but while copying file
> job
> > is getting failed. size of file is under 1MB.
> >
> > Any help or leads would be appreciated.
> >
>
>
> --
> Thanks,
> Mike
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>
>

Reply via email to