Got it. Thanks!
On Nov 5, 2015 12:32 AM, "Sandy Ryza" <sandy.r...@cloudera.com> wrote:

> Hi Nisrina,
>
> The resources you specify are shared by all jobs that run inside the
> application.
>
> -Sandy
>
> On Wed, Nov 4, 2015 at 9:24 AM, Nisrina Luthfiyati <
> nisrina.luthfiy...@gmail.com> wrote:
>
>> Hi all,
>>
>> I'm running some spark jobs in java on top of YARN by submitting one
>> application jar that starts multiple jobs.
>> My question is, if I'm setting some resource configurations, either when
>> submitting the app or in spark-defaults.conf, would this configs apply to
>> each job or the entire application?
>>
>> For example if I lauch it with:
>>
>> spark-submit --class org.some.className \
>>     --master yarn-client \
>>     --num-executors 3 \
>>     --executor-memory 5g \
>>     someJar.jar \
>>
>> , would the 3 executor x 5G memory be allocated to each job or would all
>> jobs share the resources?
>>
>> Thank you!
>> Nisrina
>>
>>
>

Reply via email to