This is our yarn related settings:

yarn.scheduler.fair.assignmultiple: "true"
yarn.scheduler.fair.dynamic.max.assign: "false"
yarn.scheduler.fair.max.assign: 1

any suggestions?

Best
Lu

On Wed, Sep 6, 2023 at 9:16 AM Lu Niu <qqib...@gmail.com> wrote:

> Hi, Thanks for all your help. Are there any other insights?
>
>
> Best
> Lu
>
> On Wed, Aug 30, 2023 at 11:29 AM Lu Niu <qqib...@gmail.com> wrote:
>
>> No. we don't use yarn.taskmanager.node-label
>>
>> Best
>> Lu
>>
>> On Tue, Aug 29, 2023 at 12:17 AM Geng Biao <biaoge...@gmail.com> wrote:
>>
>>> Maybe you can check if you have set yarn.taskmanager.node-label for some
>>> flink jobs?
>>>
>>> Best,
>>> Biao Geng
>>>
>>> 发送自 Outlook for iOS <https://aka.ms/o0ukef>
>>> ------------------------------
>>> *发件人:* Chen Zhanghao <zhanghao.c...@outlook.com>
>>> *发送时间:* Tuesday, August 29, 2023 12:14:53 PM
>>> *收件人:* Lu Niu <qqib...@gmail.com>; Weihua Hu <huweihua....@gmail.com>
>>> *抄送:* Kenan Kılıçtepe <kkilict...@gmail.com>; user <
>>> user@flink.apache.org>
>>> *主题:* Re: Uneven TM Distribution of Flink on YARN
>>>
>>> CCing @Weihua Hu <huweihua....@gmail.com> , who is an expert on this.
>>> Do you have any ideas on the phenomenon here?
>>>
>>> Best,
>>> Zhanghao Chen
>>> ------------------------------
>>> *From:* Lu Niu <qqib...@gmail.com>
>>> *Sent:* Tuesday, August 29, 2023 12:11:35 PM
>>> *To:* Chen Zhanghao <zhanghao.c...@outlook.com>
>>> *Cc:* Kenan Kılıçtepe <kkilict...@gmail.com>; user <
>>> user@flink.apache.org>
>>> *Subject:* Re: Uneven TM Distribution of Flink on YARN
>>>
>>> Thanks for your reply.
>>>
>>> The interesting fact is that we also managed spark on yarn. However.
>>> Only the flink cluster is having the issue. I am wondering whether there is
>>> a difference in the implementation on flink side.
>>>
>>> Best
>>> Lu
>>>
>>> On Mon, Aug 28, 2023 at 8:38 PM Chen Zhanghao <zhanghao.c...@outlook.com>
>>> wrote:
>>>
>>> Hi Lu Niu,
>>>
>>> TM distribution on YARN nodes is managed by YARN RM and is out of the
>>> scope of Flink. On the other hand, cluster.evenly-spread-out-slots forces
>>> even distribution of tasks among Flink TMs, and has nothing to do with your
>>> concerns. Also, the config currently only supports Standalone mode Flink
>>> clusters, and does not take effect on a Flink cluster on YARN.
>>>
>>> Best,
>>> Zhanghao Chen
>>> ------------------------------
>>> *发件人:* Lu Niu <qqib...@gmail.com>
>>> *发送时间:* 2023年8月29日 4:30
>>> *收件人:* Kenan Kılıçtepe <kkilict...@gmail.com>
>>> *抄送:* user <user@flink.apache.org>
>>> *主题:* Re: Uneven TM Distribution of Flink on YARN
>>>
>>> Thanks for the reply. We've already set cluster.evenly-spread-out-slots
>>> = true
>>>
>>> Best
>>> Lu
>>>
>>> On Mon, Aug 28, 2023 at 1:23 PM Kenan Kılıçtepe <kkilict...@gmail.com>
>>> wrote:
>>>
>>> Have you checked config param cluster.evenly-spread-out-slots ?
>>>
>>>
>>> On Mon, Aug 28, 2023 at 10:31 PM Lu Niu <qqib...@gmail.com> wrote:
>>>
>>> Hi, Flink users
>>>
>>> We have recently observed that the allocation of Flink TaskManagers in
>>> our YARN cluster is not evenly distributed. We would like to hear your
>>> thoughts on this matter.
>>>
>>> 1. Our setup includes Flink version 1.15.1 and Hadoop 2.10.0.
>>> 2. The uneven distribution is that out of a 370-node YARN cluster, there
>>> are 16 nodes with either 0 or 1 vCore available, while 110 nodes have more
>>> than 10 vCores available.
>>>
>>> Is such behavior expected? If not, is there a fix provided in Flink?
>>> Thanks!
>>>
>>> Best
>>> Lu
>>>
>>>

Reply via email to