Can a container have multiple JVMs running in YARN?

I am comparing Hadoop Mapreduce running on yarn vs spark running on yarn
here :

1.Is the difference is in Hadoop Mapreduce job - say I specify 20 reducers
and my job uses 10 map tasks then, it need total 30 containers or 30 vcores
? I guess 30 vcores and runs multiple max 4 jvms (set using max vcores in a
container) in  my case ?So in taotal 8 containers - 7 with 4 vcores and 1
with 2 ?

2.for each map/reduce tasks requested is equal to vcore requested  and it
launch a new JVM inside same container or per container 1 JVM ?

3.In spark jobs it launches task in same jvm running in a container and 1
container has max 1 JVM and max tasks = max vcore per container limit (4 in
my case).







On Wed, Jul 15, 2015 at 12:17 AM, Jong Wook Kim <jongw...@nyu.edu> wrote:

> it's probably because your YARN cluster has only 40 vCores available.
>
> Go to your resource manager and check if "VCores Total" and "Memory Total"
> exceeds what you have set. (40 cores and 5120 MB)
>
> If that looks fine, go to "Scheduler" page and find the queue on which
> your jobs run, and check the resources allocated for that queue.
>
> Hope this helps.
>
> Jong Wook
>
>
> > On Jul 15, 2015, at 01:57, Shushant Arora <shushantaror...@gmail.com>
> wrote:
> >
> > I am running spark application on yarn managed cluster.
> >
> > When I specify --executor-cores > 4 it fails to start the application.
> > I am starting the app as
> >
> > spark-submit --class classname --num-executors 10 --executor-cores 5
> --master masteradd jarname
> >
> > Exception in thread "main" org.apache.spark.SparkException: Yarn
> application has already ended! It might have been killed or unable to
> launch application master.
> >
> > When I give --executor-cores as 4 , it works fine.
> >
> > My Cluster has 10 nodes .
> > Why am I not able to specify more than 4 concurrent tasks. Is there any
> max limit yarn side or spark side which I can override to make use of more
> tasks ?
>
>

Reply via email to