Thanks, Mark. Then, I guess JobServer can fundamentally solve my problem, so 
that jobs can be submitted at different time and still share RDDs.

Best Regards,
Jia


On Jan 17, 2016, at 3:44 PM, Mark Hamstra <m...@clearstorydata.com> wrote:

> There is a 1-to-1 relationship between Spark Applications and SparkContexts 
> -- fundamentally, a Spark Applications is a program that creates and uses a 
> SparkContext, and that SparkContext is destroyed when then Application ends.  
> A jobserver generically and the Spark JobServer specifically is an 
> Application that keeps a SparkContext open for a long time and allows many 
> Jobs to be be submitted and run using that shared SparkContext.
> 
> More than one Application/SparkContext unavoidably implies more than one JVM 
> process per Worker -- Applications/SparkContexts cannot share JVM processes.  
> 
> On Sun, Jan 17, 2016 at 1:15 PM, Jia <jacqueline...@gmail.com> wrote:
> Hi, Mark, sorry for the confusion.
> 
> Let me clarify, when an application is submitted, the master will tell each 
> Spark worker to spawn an executor JVM process. All the task sets  of the 
> application will be executed by the executor. After the application runs to 
> completion. The executor process will be killed.
> But I hope that all applications submitted can run in the same executor, can 
> JobServer do that? If so, it’s really good news!
> 
> Best Regards,
> Jia
> 
> On Jan 17, 2016, at 3:09 PM, Mark Hamstra <m...@clearstorydata.com> wrote:
> 
>> You've still got me confused.  The SparkContext exists at the Driver, not on 
>> an Executor.
>> 
>> Many Jobs can be run by a SparkContext -- it is a common pattern to use 
>> something like the Spark Jobserver where all Jobs are run through a shared 
>> SparkContext.
>> 
>> On Sun, Jan 17, 2016 at 12:57 PM, Jia Zou <jacqueline...@gmail.com> wrote:
>> Hi, Mark, sorry, I mean SparkContext.
>> I mean to change Spark into running all submitted jobs (SparkContexts) in 
>> one executor JVM.
>> 
>> Best Regards,
>> Jia
>> 
>> On Sun, Jan 17, 2016 at 2:21 PM, Mark Hamstra <m...@clearstorydata.com> 
>> wrote:
>> -dev
>> 
>> What do you mean by JobContext?  That is a Hadoop mapreduce concept, not 
>> Spark.
>> 
>> On Sun, Jan 17, 2016 at 7:29 AM, Jia Zou <jacqueline...@gmail.com> wrote:
>> Dear all,
>> 
>> Is there a way to reuse executor JVM across different JobContexts? Thanks.
>> 
>> Best Regards,
>> Jia
>> 
>> 
>> 
> 
> 

Reply via email to