Yes, I think in the spark-env.sh.template, it is listed in the comments (didn’t 
check….)  

Best,  

--  
Nan Zhu


On Sunday, June 15, 2014 at 5:21 PM, Surendranauth Hiraman wrote:

> Is SPARK_DAEMON_JAVA_OPTS valid in 1.0.0?
>  
>  
>  
> On Sun, Jun 15, 2014 at 4:59 PM, Nan Zhu <zhunanmcg...@gmail.com 
> (mailto:zhunanmcg...@gmail.com)> wrote:
> > SPARK_JAVA_OPTS is deprecated in 1.0, though it works fine if you don’t 
> > mind the WARNING in the logs
> >  
> > you can set spark.executor.extraJavaOpts in your SparkConf obj  
> >  
> > Best,
> >  
> > --  
> > Nan Zhu
> >  
> >  
> > On Sunday, June 15, 2014 at 12:13 PM, Hao Wang wrote:
> >  
> > > Hi, Wei
> > >  
> > > You may try to set JVM opts in spark-env.sh (http://spark-env.sh) as 
> > > follow to prevent or mitigate GC pause:  
> > >  
> > > export SPARK_JAVA_OPTS="-XX:-UseGCOverheadLimit -XX:+UseConcMarkSweepGC 
> > > -Xmx2g -XX:MaxPermSize=256m"
> > >  
> > > There are more options you could add, please just Google :)  
> > >  
> > >  
> > > Regards,
> > > Wang Hao(王灏)
> > >  
> > > CloudTeam | School of Software Engineering
> > > Shanghai Jiao Tong University
> > > Address:800 Dongchuan Road, Minhang District, Shanghai, 200240
> > > Email:wh.s...@gmail.com (mailto:wh.s...@gmail.com)
> > >  
> > >  
> > >  
> > >  
> > >  
> > >  
> > > On Sun, Jun 15, 2014 at 10:24 AM, Wei Tan <w...@us.ibm.com 
> > > (mailto:w...@us.ibm.com)> wrote:
> > > > Hi,  
> > > >  
> > > >   I have a single node (192G RAM) stand-alone spark, with memory 
> > > > configuration like this in spark-env.sh (http://spark-env.sh)  
> > > >  
> > > > SPARK_WORKER_MEMORY=180g  
> > > > SPARK_MEM=180g  
> > > >  
> > > >  
> > > >  In spark-shell I have a program like this:  
> > > >  
> > > > val file = sc.textFile("/localpath") //file size is 40G  
> > > > file.cache()  
> > > >  
> > > >  
> > > > val output = file.map(line => extract something from line)  
> > > >  
> > > > output.saveAsTextFile (...)  
> > > >  
> > > >  
> > > > When I run this program again and again, or keep trying 
> > > > file.unpersist() --> file.cache() --> output.saveAsTextFile(), the run 
> > > > time varies a lot, from 1 min to 3 min to 50+ min. Whenever the 
> > > > run-time is more than 1 min, from the stage monitoring GUI I observe 
> > > > big GC pause (some can be 10+ min). Of course when run-time is 
> > > > "normal", say ~1 min, no significant GC is observed. The behavior seems 
> > > > somewhat random.  
> > > >  
> > > > Is there any JVM tuning I should do to prevent this long GC pause from 
> > > > happening?  
> > > >  
> > > >  
> > > >  
> > > > I used java-1.6.0-openjdk.x86_64, and my spark-shell process is 
> > > > something like this:  
> > > >  
> > > > root     10994  1.7  0.6 196378000 1361496 pts/51 Sl+ 22:06   0:12 
> > > > /usr/lib/jvm/java-1.6.0-openjdk.x86_64/bin/java -cp 
> > > > ::/home/wtan/scala/spark-1.0.0-bin-hadoop1/conf:/home/wtan/scala/spark-1.0.0-bin-hadoop1/lib/spark-assembly-1.0.0-hadoop1.0.4.jar:/home/wtan/scala/spark-1.0.0-bin-hadoop1/lib/datanucleus-core-3.2.2.jar:/home/wtan/scala/spark-1.0.0-bin-hadoop1/lib/datanucleus-rdbms-3.2.1.jar:/home/wtan/scala/spark-1.0.0-bin-hadoop1/lib/datanucleus-api-jdo-3.2.1.jar
> > > >  -XX:MaxPermSize=128m -Djava.library.path= -Xms180g -Xmx180g 
> > > > org.apache.spark.deploy.SparkSubmit spark-shell --class 
> > > > org.apache.spark.repl.Main  
> > > >  
> > > > Best regards,  
> > > > Wei  
> > > >  
> > > > ---------------------------------  
> > > > Wei Tan, PhD  
> > > > Research Staff Member  
> > > > IBM T. J. Watson Research Center  
> > > > http://researcher.ibm.com/person/us-wtan
> >  
>  
>  
>  
> --  
>                                                              SUREN HIRAMAN, 
> VP TECHNOLOGY
> Velos
> Accelerating Machine Learning
>  
> 440 NINTH AVENUE, 11TH FLOOR
> NEW YORK, NY 10001
> O: (917) 525-2466 ext. 105
> F: 646.349.4063
> E: suren.hiraman@v (mailto:suren.hira...@sociocast.com)elos.io 
> (http://elos.io)
> W: www.velos.io (http://www.velos.io/)
>  

Reply via email to