You can set your executor number with --num-executors. Also changing 
yarn-client save you one container for driver. Then check your yarn resource 
manager to make sure there are more containers available to serve your extra 
apps.

Thanks.

Zhan Zhang

On Oct 28, 2014, at 5:31 PM, Soumya Simanta <soumya.sima...@gmail.com> wrote:

> Maybe changing --master yarn-cluster to --master yarn-client help. 
> 
> 
> On Tue, Oct 28, 2014 at 7:25 PM, Josh J <joshjd...@gmail.com> wrote:
> Sorry, I should've included some stats with my email
> 
> I execute each job in the following manner
> 
> ./bin/spark-submit --class CLASSNAME --master yarn-cluster --driver-memory 1g 
> --executor-memory 1g --executor-cores 1 UBER.JAR ${ZK_PORT_2181_TCP_ADDR} 
> my-consumer-group1 1
> 
> 
> 
> The box has 
> 
> 24 CPUs, Intel(R) Xeon(R) CPU E5-2420 v2 @ 2.20GHz
> 
> 32 GB RAM
> 
> 
> 
> Thanks,
> 
> Josh
> 
> 
> On Tue, Oct 28, 2014 at 4:15 PM, Soumya Simanta <soumya.sima...@gmail.com> 
> wrote:
> Try reducing the resources (cores and memory) of each application.
> 
> 
> 
> > On Oct 28, 2014, at 7:05 PM, Josh J <joshjd...@gmail.com> wrote:
> >
> > Hi,
> >
> > How do I run multiple spark applications in parallel? I tried to run on 
> > yarn cluster, though the second application submitted does not run.
> >
> > Thanks,
> > Josh
> 
> 


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Reply via email to