Whoa! That worked! I was half afraid it wont, since I hadnt tried it myself.
TD On Wed, Jul 30, 2014 at 8:32 PM, liuwei <stupi...@126.com> wrote: > Hi, Tathagata Das: > > I followed your advice and solved this problem, thank you very much! > > > 在 2014年7月31日,上午3:07,Tathagata Das <tathagata.das1...@gmail.com> 写道: > >> This is because setLocalProperty makes all Spark jobs submitted using >> the current thread belong to the set pool. However, in Spark >> Streaming, all the jobs are actually launched in the background from a >> different thread. So this setting does not work. However, there is a >> work around. If you are doing any kind of output operations on >> DStreams, like DStream.foreachRDD(), you can set the property inside >> that >> >> dstream.foreachRDD(rdd => >> rdd.sparkContext.setLocalProperty(...) >> ) >> >> >> >> On Wed, Jul 30, 2014 at 1:43 AM, liuwei <stupi...@126.com> wrote: >>> In my spark streaming program, I set scheduler pool, just as follows: >>> >>> val myFairSchedulerFile = “xxx.xml” >>> val myStreamingPool = “xxx” >>> >>> System.setProperty(“spark.scheduler.allocation.file”, myFairSchedulerFile) >>> val conf = new SparkConf() >>> val ssc = new StreamingContext(conf, batchInterval) >>> ssc.sparkContext.setLocalProperty(“spark.scheduler.pool”, myStreamingPool) >>> …. >>> ssc.start() >>> ssc.awaitTermination() >>> >>> I submit my spark streaming job in my spark cluster, and I found stage’s >>> pool name is “default”, it seem >>> ssc.sparkContext.setLocalProperty(“spark.scheduler.pool”, myStreamingPool) >>> not work. > >