Hi Crystal,

The fair scheduler is only for jobs running concurrently within the same 
SparkContext (i.e. within an application), not for separate applications on the 
standalone cluster manager. It has no effect there. To run more of those 
concurrently, you need to set a cap on how many cores they each grab with 
spark.cores.max.

Matei

On August 10, 2014 at 12:13:08 PM, 李宜芳 (xuite...@gmail.com) wrote:

Hi  

I am trying to switch from FIFO to FAIR with standalone mode.  

my environment:  
hadoop 1.2.1  
spark 0.8.0 using stanalone mode  

and i modified the code..........  

ClusterScheduler.scala -> System.getProperty("spark.scheduler.mode",  
"FAIR"))  
SchedulerBuilder.scala ->  
val DEFAULT_SCHEDULING_MODE = SchedulingMode.FAIR  

LocalScheduler.scala ->  
System.getProperty("spark.scheduler.mode", "FAIR)  

spark-env.sh ->  
export SPARK_JAVA_OPTS="-Dspark.scheduler.mode=FAIR"  
export SPARK_JAVA_OPTS=" -Dspark.scheduler.mode=FAIR" ./run-example  
org.apache.spark.examples.SparkPi spark://streaming1:7077  


but it's not work  
i want to switch from fifo to fair  
how can i do??  

Regards  
Crystal Lee  

Reply via email to