Hello,

I'm currently using HDP 2.0 so it's Hadoop 2.2.0.
My cluster consist in 4 nodes, 16 coeurs 16 GB RAM 4*3To each.

Recently we passed from 2 users to 8. We need now a more appropriate
Scheduler.
We begin with Capacity Scheduler. There was some issues with the different
queues particularly when using some spark shell that used some resources
for a long time.
So we decide to try Fair Scheduler which seems to be a good solution.
The problem is that FairScheduler doesn't allow all available resources.
It's capped at 73% of the available memory for one jobs 63% for 2 jobs and
45% for 3 jobs. The problem could come from shells that take resources for
a long time.

We tried some configuration like
yarn.scheduler.fair.user-as-default-queue=false
or play with the minimum ressources allocated minResources in
fair-scheduler.xml but it doesn't seems to resolve the issue.

Any advices or good practices to held a good Fair Scheduler?

Regards,

Julien

Reply via email to