@Crystal
You can use spark on yarn. Yarn have fair scheduler,modified yarn-site.xml.

发自我的 iPad

> 在 2014年8月11日,6:49,Matei Zaharia <matei.zaha...@gmail.com> 写道:
> 
> Hi Crystal,
> 
> The fair scheduler is only for jobs running concurrently within the same 
> SparkContext (i.e. within an application), not for separate applications on 
> the standalone cluster manager. It has no effect there. To run more of those 
> concurrently, you need to set a cap on how many cores they each grab with 
> spark.cores.max.
> 
> Matei
> 
> On August 10, 2014 at 12:13:08 PM, 李宜芳 (xuite...@gmail.com) wrote:
> 
> Hi  
> 
> I am trying to switch from FIFO to FAIR with standalone mode.  
> 
> my environment:  
> hadoop 1.2.1  
> spark 0.8.0 using stanalone mode  
> 
> and i modified the code..........  
> 
> ClusterScheduler.scala -> System.getProperty("spark.scheduler.mode",  
> "FAIR"))  
> SchedulerBuilder.scala ->  
> val DEFAULT_SCHEDULING_MODE = SchedulingMode.FAIR  
> 
> LocalScheduler.scala ->  
> System.getProperty("spark.scheduler.mode", "FAIR)  
> 
> spark-env.sh ->  
> export SPARK_JAVA_OPTS="-Dspark.scheduler.mode=FAIR"  
> export SPARK_JAVA_OPTS=" -Dspark.scheduler.mode=FAIR" ./run-example  
> org.apache.spark.examples.SparkPi spark://streaming1:7077  
> 
> 
> but it's not work  
> i want to switch from fifo to fair  
> how can i do??  
> 
> Regards  
> Crystal Lee  
> 


---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to