Hi Everyone,

I'm using Hadoop-2.2.0 with fair scheduler in my YARN cluster, but something is 
wrong with the fair scheduler.

Here is my fair-scheduler.xml looks like:

<allocations>
  <queue name="longrun">
    <maxResources>15360 mb, 5 vcores</maxResources>
    <weight>0.5</weight>
    <minMaps>2</minMaps>
    <maxMaps>5</maxMaps>
    <minReduces>1</minReduces>
  </queue>
</allocations>

I create a "longrun" queue to ensure that huge MR application can only use 5 
resources. My YARN setup for each resource memory is 3072 MB:

  <property>
    <name>mapreduce.map.memory.mb</name>
    <value>3072</value>
  </property>
  <property>
    <name>mapreduce.reduce.memory.mb</name>
    <value>3072</value>
  </property>

When the huge application started, it works just fine and scheduler restrict it 
to only run 5 maps in parallel.
But after running for some time, the application run 10 maps in parallel.
The scheduler page show that the "longrun" queue used 66%, exceed the fair 
share 30%.

Can anyone tell me why the application can get more than it deserved?
Is the problem with my configuration? Or there is a bug?

Best regards,
Henry Hung

________________________________
The privileged confidential information contained in this email is intended for 
use only by the addressees as indicated by the original sender of this email. 
If you are not the addressee indicated in this email or are not responsible for 
delivery of the email to such a person, please kindly reply to the sender 
indicating this fact and delete all copies of it from your computer and network 
server immediately. Your cooperation is highly appreciated. It is advised that 
any unauthorized use of confidential information of Winbond is strictly 
prohibited; and any information in this email irrelevant to the official 
business of Winbond shall be deemed as neither given nor endorsed by Winbond.

Reply via email to