On Tue, Dec 5, 2023 at 6:34 AM Xiubo Li wrote:
>
>
> On 12/4/23 16:25, zxcs wrote:
> > Thanks a lot, Xiubo!
> >
> > we already set ‘mds_bal_interval’ to 0. and the slow mds seems decrease.
> >
> > But somehow we still see mds complain slow request. and from mds log , can
> > see
> >
> > “slow req
On 12/4/23 16:25, zxcs wrote:
Thanks a lot, Xiubo!
we already set ‘mds_bal_interval’ to 0. and the slow mds seems decrease.
But somehow we still see mds complain slow request. and from mds log , can see
“slow request *** seconds old, received at 2023-12-04T…: internal op
exportdir:mds.* curr
Thanks a lot, Xiubo!
we already set ‘mds_bal_interval’ to 0. and the slow mds seems decrease.
But somehow we still see mds complain slow request. and from mds log , can see
“slow request *** seconds old, received at 2023-12-04T…: internal op
exportdir:mds.* currently acquired locks”
so our qu
On 11/27/23 13:12, zxcs wrote:
current, we using `ceph config set mds mds_bal_interval 3600` to set a
fixed time(1 hour).
we also have a question about how to set no balance for multi active mds.
means, we will enable multi active mds(to improve throughput) and no
balance for these mds.
an
current, we using `ceph config set mds mds_bal_interval 3600` to set a fixed
time(1 hour).
we also have a question about how to set no balance for multi active mds.
means, we will enable multi active mds(to improve throughput) and no balance
for these mds.
and if we set mds_bal_interval as big
with the same mds configuration, we see exactly the same(problem, log and
solution) with 17.2.5, constantly happening again and again in couples days
intervals. MDS servers are stuck somewhere, ceph status reports no issue
however. We need to restart some of the mds (if not all of them) to restore
On 11/23/23 11:25, zxcs wrote:
Thanks a ton, Xiubo!
it not disappear.
even we umount the ceph directory on these two old os node.
after dump ops flight , we can see some request, and the earliest complain “failed
to authpin, subtree is being exported"
And how to avoid this, would you please
Thanks a ton, Xiubo!
it not disappear.
even we umount the ceph directory on these two old os node.
after dump ops flight , we can see some request, and the earliest complain
“failed to authpin, subtree is being exported"
And how to avoid this, would you please help to shed some light here?
Th
From: Eugen Block
Sent: Wednesday, November 22, 2023 12:30 PM
To: ceph-users@ceph.io
Subject: [ceph-users] Re: mds slow request with “failed to authpin, subtree is
being exported"
Hi,
we've seen this a year ago in a Nautilus cluster with multi-active MDS
as well. It turned up only o
On 11/22/23 16:02, zxcs wrote:
HI, Experts,
we are using cephfs with 16.2.* with multi active mds, and recently, we have
two nodes mount with ceph-fuse due to the old os system.
and one nodes run a python script with `glob.glob(path)`, and another client
doing `cp` operation on the same pa
Hi,
we've seen this a year ago in a Nautilus cluster with multi-active MDS
as well. It turned up only once within several years and we decided
not to look too closely at that time. How often do you see it? Is it
reproducable? In that case I'd recommend to create a tracker issue.
Regards,
11 matches
Mail list logo