On 2021/7/12 15:52, Ulrich Windl wrote:
Hi!

can you give some details on what is necessary to trigger the problem?
There is a ABBA lock between reflink comand and ocfs2 ocfs2_complete_recovery routine(this routine will be triggered by timer, mount, node recovery), the dead lock is not always encountered.
For the more details, refer to the link as below,
https://oss.oracle.com/pipermail/ocfs2-devel/2021-July/015671.html

Thanks
Gang

(I/O load, CPU load, concurrent operations on one node or on multiple nodes,
using reflink snapshots, using ioctl(FS_IOC_FIEMAP), etc.)

Regards,
Ulrich

Gang He <g...@suse.com> schrieb am 11.07.2021 um 10:55 in Nachricht
<am6pr04mb648817316db7b124f414d60acf...@am6pr04mb6488.eurprd04.prod.outlook.com>

Hi Ulrich,

Thank for your update.
Based on some feedback from the upstream, there is a patch (ocfs2:
initialize ip_next_orphan), which should fix this problem.
I can comfirm the patch looks very similar with your problem.
I will verify it next week, then let you know the result.

Thanks
Gang

________________________________________
From: Users <users‑boun...@clusterlabs.org> on behalf of Ulrich Windl
<ulrich.wi...@rz.uni‑regensburg.de>
Sent: Friday, July 9, 2021 15:56
To: users@clusterlabs.org
Subject: [ClusterLabs] Antw: [EXT] Re: Antw: Hanging OCFS2 Filesystem any
one else?

Hi!

An update on the issue:
SUSE support found out that the reason for the hanging processes is a
deadlock caused by a race condition (Kernel 5.3.18‑24.64‑default). Support
is
working on a fix.
Today the cluster "fixed" the problem in an unusual way:

h19 kernel: Out of memory: Killed process 6838 (corosync) total‑vm:261212kB,

anon‑rss:31444kB, file‑rss:7700kB, shmem‑rss:121872kB

I doubt that was the best possible choice ;‑)

The dead corosync caused the DC (h18) to fence h19 (which was successful),
but the DC was fenced while it tried to recover resources, so the complete
cluster rebooted.

Regards,
Ulrich




_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/



_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/

Reply via email to