Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both
Ok i reconfigure server and do again test hope tommorow die again because i see in log he crash after 10 hours work with no problem. Thanks -Oryginalna wiadomość- From: srinivas eeda Sent: Thursday, December 22, 2011 9:12 PM To: Marek Królikowski Cc: ocfs2-users@oss.oracle.com Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both We need to know what happened to node 2. Was the node rebooted because of a network timeout or kernel panic? can you please configure netconsole, serial console and rerun the test? On 12/22/2011 8:08 AM, Marek Królikowski wrote: > Hello > After 24 hours i see TEST-MAIL2 reboot ( possible kernel panic) but > TEST-MAIL1 got in dmesg: > TEST-MAIL1 ~ #dmesg > [cut] > o2net: accepted connection from node TEST-MAIL2 (num 1) at > 172.17.1.252: > o2dlm: Node 1 joins domain B24C4493BBC74FEAA3371E2534BB3611 > o2dlm: Nodes in domain B24C4493BBC74FEAA3371E2534BB3611: 0 1 > o2net: connection to node TEST-MAIL2 (num 1) at 172.17.1.252: has been > idle for 60.0 seconds, shutting it down. > (swapper,0,0):o2net_idle_timer:1562 Here are some times that might help > debug the situation: (Timer: 33127732045, Now 33187808090, DataReady > 33127732039, Advance 33127732051-33127732051, Key 0xebb9cd47, Func 506, > FuncTime 33127732045-33127732048) > o2net: no longer connected to node TEST-MAIL2 (num 1) at 172.17.1.252: > (du,5099,12):dlm_do_master_request:1324 ERROR: link to 1 went down! > (du,5099,12):dlm_get_lock_resource:907 ERROR: status = -112 > (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: > B24C4493BBC74FEAA3371E2534BB3611: res M0cf023ef70, > error -112 send AST to node 1 > (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -112 > (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: > B24C4493BBC74FEAA3371E2534BB3611: res P00, > error -107 send AST to node 1 > (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -107 > (kworker/u:3,5071,0):o2net_connect_expired:1724 ERROR: no connection > established with node 1 after 60.0 seconds, giving up and returning > errors. > (o2hb-B24C4493BB,14310,0):o2dlm_eviction_cb:267 o2dlm has evicted node 1 > from group B24C4493BBC74FEAA3371E2534BB3611 > (ocfs2rec,5504,6):dlm_get_lock_resource:834 > B24C4493BBC74FEAA3371E2534BB3611:M15f023ef70: at least > one node (1) to recover before lock mastery can begin > (ocfs2rec,5504,6):dlm_get_lock_resource:888 > B24C4493BBC74FEAA3371E2534BB3611:M15f023ef70: at least > one node (1) to recover before lock mastery can begin > (du,5099,12):dlm_restart_lock_mastery:1213 ERROR: node down! 1 > (du,5099,12):dlm_wait_for_lock_mastery:1030 ERROR: status = -11 > (du,5099,12):dlm_get_lock_resource:888 > B24C4493BBC74FEAA3371E2534BB3611:N0020924f: at least one node (1) > to recover before lock mastery can begin > (dlm_reco_thread,14322,0):dlm_get_lock_resource:834 > B24C4493BBC74FEAA3371E2534BB3611:$RECOVERY: at least one node (1) to > recover before lock mastery can begin > (dlm_reco_thread,14322,0):dlm_get_lock_resource:868 > B24C4493BBC74FEAA3371E2534BB3611: recovery map is not empty, but must > master $RECOVERY lock now > (dlm_reco_thread,14322,0):dlm_do_recovery:523 (14322) Node 0 is the > Recovery Master for the Dead Node 1 for Domain > B24C4493BBC74FEAA3371E2534BB3611 > (ocfs2rec,5504,6):ocfs2_replay_journal:1549 Recovering node 1 from slot 1 > on device (253,0) > (ocfs2rec,5504,6):ocfs2_begin_quota_recovery:407 Beginning quota recovery > in slot 1 > (kworker/u:0,2909,0):ocfs2_finish_quota_recovery:599 Finishing quota > recovery in slot 1 > > And i try give this command: > debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP > allow > debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or directory > debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP off > debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or directory > > But not working > > > -Oryginalna wiadomość- From: Srinivas Eeda > Sent: Wednesday, December 21, 2011 8:43 PM > To: Marek Królikowski > Cc: ocfs2-users@oss.oracle.com > Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from > both > > Those numbers look good. Basically with the fixes backed out and another > fix I gave, you are not seeing that many orphans hanging around and > hence not seeing the process stuck kernel stacks. You can run the test > longer or if you are satisfied, please enable quotas and re-run the test > with the modified kernel. You might see a dead lock which needs to be > fixed(I was not able to reproduce this yet). If the system hangs, please > capture the following and provide me the output > > 1. echo t > /proc/sysrq-trigger > 2. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP > allow > 3. wait for 10 minutes > 4. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP >
Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both
We need to know what happened to node 2. Was the node rebooted because of a network timeout or kernel panic? can you please configure netconsole, serial console and rerun the test? On 12/22/2011 8:08 AM, Marek Królikowski wrote: > Hello > After 24 hours i see TEST-MAIL2 reboot ( possible kernel panic) but > TEST-MAIL1 got in dmesg: > TEST-MAIL1 ~ #dmesg > [cut] > o2net: accepted connection from node TEST-MAIL2 (num 1) at > 172.17.1.252: > o2dlm: Node 1 joins domain B24C4493BBC74FEAA3371E2534BB3611 > o2dlm: Nodes in domain B24C4493BBC74FEAA3371E2534BB3611: 0 1 > o2net: connection to node TEST-MAIL2 (num 1) at 172.17.1.252: has > been idle for 60.0 seconds, shutting it down. > (swapper,0,0):o2net_idle_timer:1562 Here are some times that might > help debug the situation: (Timer: 33127732045, Now 33187808090, > DataReady 33127732039, Advance 33127732051-33127732051, Key > 0xebb9cd47, Func 506, FuncTime 33127732045-33127732048) > o2net: no longer connected to node TEST-MAIL2 (num 1) at > 172.17.1.252: > (du,5099,12):dlm_do_master_request:1324 ERROR: link to 1 went down! > (du,5099,12):dlm_get_lock_resource:907 ERROR: status = -112 > (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: > B24C4493BBC74FEAA3371E2534BB3611: res M0cf023ef70, > error -112 send AST to node 1 > (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -112 > (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: > B24C4493BBC74FEAA3371E2534BB3611: res P00, > error -107 send AST to node 1 > (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -107 > (kworker/u:3,5071,0):o2net_connect_expired:1724 ERROR: no connection > established with node 1 after 60.0 seconds, giving up and returning > errors. > (o2hb-B24C4493BB,14310,0):o2dlm_eviction_cb:267 o2dlm has evicted node > 1 from group B24C4493BBC74FEAA3371E2534BB3611 > (ocfs2rec,5504,6):dlm_get_lock_resource:834 > B24C4493BBC74FEAA3371E2534BB3611:M15f023ef70: at > least one node (1) to recover before lock mastery can begin > (ocfs2rec,5504,6):dlm_get_lock_resource:888 > B24C4493BBC74FEAA3371E2534BB3611:M15f023ef70: at > least one node (1) to recover before lock mastery can begin > (du,5099,12):dlm_restart_lock_mastery:1213 ERROR: node down! 1 > (du,5099,12):dlm_wait_for_lock_mastery:1030 ERROR: status = -11 > (du,5099,12):dlm_get_lock_resource:888 > B24C4493BBC74FEAA3371E2534BB3611:N0020924f: at least one node > (1) to recover before lock mastery can begin > (dlm_reco_thread,14322,0):dlm_get_lock_resource:834 > B24C4493BBC74FEAA3371E2534BB3611:$RECOVERY: at least one node (1) to > recover before lock mastery can begin > (dlm_reco_thread,14322,0):dlm_get_lock_resource:868 > B24C4493BBC74FEAA3371E2534BB3611: recovery map is not empty, but must > master $RECOVERY lock now > (dlm_reco_thread,14322,0):dlm_do_recovery:523 (14322) Node 0 is the > Recovery Master for the Dead Node 1 for Domain > B24C4493BBC74FEAA3371E2534BB3611 > (ocfs2rec,5504,6):ocfs2_replay_journal:1549 Recovering node 1 from > slot 1 on device (253,0) > (ocfs2rec,5504,6):ocfs2_begin_quota_recovery:407 Beginning quota > recovery in slot 1 > (kworker/u:0,2909,0):ocfs2_finish_quota_recovery:599 Finishing quota > recovery in slot 1 > > And i try give this command: > debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP > allow > debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or > directory > debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP > off > debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or > directory > > But not working > > > -Oryginalna wiadomość- From: Srinivas Eeda > Sent: Wednesday, December 21, 2011 8:43 PM > To: Marek Królikowski > Cc: ocfs2-users@oss.oracle.com > Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read > from both > > Those numbers look good. Basically with the fixes backed out and another > fix I gave, you are not seeing that many orphans hanging around and > hence not seeing the process stuck kernel stacks. You can run the test > longer or if you are satisfied, please enable quotas and re-run the test > with the modified kernel. You might see a dead lock which needs to be > fixed(I was not able to reproduce this yet). If the system hangs, please > capture the following and provide me the output > > 1. echo t > /proc/sysrq-trigger > 2. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC > EXTENT_MAP allow > 3. wait for 10 minutes > 4. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC > EXTENT_MAP off > 5. echo t > /proc/sysrq-trigger > ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users
Re: [Ocfs2-users] One node, two clusters?
On 12/22/2011 10:39 AM, Kushnir, Michael (NIH/NLM/LHC) [C] wrote: > Is there a separate DLM instance for each ocfs2 volume? > > I have two "sub-clusters" in the same cluster... A 10 node Hadoop cluster > sharing a SATA RAID10 and a Two node web server cluster sharing a SSD RAID0. > One server mounts both volumes to move data between as necessary. > > This morning I got the following error (see end of message), and all nodes > lost access to all storage. I'm trying to mitigate risk of this happening > again. > > My hadoop nodes are used to generate search engine indexes, so they can go > down. But my web servers provide the search engine service so I need them to > not be tied to my hadoop nodes. I just feel safer that way. At the same time, > I need a "bridge" node to move data between the two. I can do it via NFS or > SCP, but I figured it'd be worth while to ask if one node can be in two > different clusters. > > Dec 22 09:15:42 lhce-imed-web1 kernel: > (updatedb,1832,1):dlm_get_lock_resource:898 > 042F68B6AF134E5C9A9EDF4D7BD7BE99:O0013d2ef94: at least > one node (11) to recover before lock mastery can begin > You should add ocfs2 to PRUNEFS in /etc/updatedb.conf. updatedb generates a lot of io and network traffic. And it will happen around the same time on all nodes. Yes, each volume has a different dlm domain (instance). ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users
Re: [Ocfs2-users] One node, two clusters?
Is there a separate DLM instance for each ocfs2 volume? I have two "sub-clusters" in the same cluster... A 10 node Hadoop cluster sharing a SATA RAID10 and a Two node web server cluster sharing a SSD RAID0. One server mounts both volumes to move data between as necessary. This morning I got the following error (see end of message), and all nodes lost access to all storage. I'm trying to mitigate risk of this happening again. My hadoop nodes are used to generate search engine indexes, so they can go down. But my web servers provide the search engine service so I need them to not be tied to my hadoop nodes. I just feel safer that way. At the same time, I need a "bridge" node to move data between the two. I can do it via NFS or SCP, but I figured it'd be worth while to ask if one node can be in two different clusters. Dec 22 09:15:42 lhce-imed-web1 kernel: (updatedb,1832,1):dlm_get_lock_resource:898 042F68B6AF134E5C9A9EDF4D7BD7BE99:O0013d2ef94: at least one node (11) to recover before lock mastery can begin Thanks, Mike -Original Message- From: Sunil Mushran [mailto:sunil.mush...@oracle.com] Sent: Thursday, December 22, 2011 1:21 PM To: Werner Flamme Cc: ocfs2-users ML Subject: Re: [Ocfs2-users] One node, two clusters? You don't need to have two clusters for this. This can be accomplished with one cluster with the default local heartbeat. Create one cluster.conf with all the nodes. All nodes, except the one machine, will mount from just one san. The common node will mount from both sans. If you look at the cluster membership, other than the common node, all nodes will be interacting (network connection, etc.) with nodes that they can see on the san. On 12/22/2011 09:40 AM, Werner Flamme wrote: > -BEGIN PGP SIGNED MESSAGE- > Hash: SHA1 > > Kushnir, Michael (NIH/NLM/LHC) [C] [22.12.2011 18:20]: >> Is it possible to have one machine be part of two different ocfs2 >> clusters with two different sans? Kind of to serve as a bridge for >> moving data between two clusters but without actually fully combining >> the two clusters? >> >> Thanks, Michael > Michael, > > I asked this two years ago and the answer was no. > > When I look at /etc/ocfs2/cluster.conf, I do not see a possibility to > configure a second cluster. Though the nodes must be assigned to a > cluster (and exactly one cluster, this is), there ist only one entry > "cluster:" in the file, and so there is no way to define a second one. > > We synced via rsync :-( > > HTH > Werner > > -BEGIN PGP SIGNATURE- > Version: GnuPG v2.0.18 (GNU/Linux) > Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/ > > iEYEARECAAYFAk7za4EACgkQk33Krq8b42MvSwCfQAXzqVQRPyhOdFrKM8PCPqbf > g0cAn20CV4rjzXNrTa/YGaUeNlO3+rmc > =CBmQ > -END PGP SIGNATURE- > > ___ > Ocfs2-users mailing list > Ocfs2-users@oss.oracle.com > http://oss.oracle.com/mailman/listinfo/ocfs2-users ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users
Re: [Ocfs2-users] One node, two clusters?
You don't need to have two clusters for this. This can be accomplished with one cluster with the default local heartbeat. Create one cluster.conf with all the nodes. All nodes, except the one machine, will mount from just one san. The common node will mount from both sans. If you look at the cluster membership, other than the common node, all nodes will be interacting (network connection, etc.) with nodes that they can see on the san. On 12/22/2011 09:40 AM, Werner Flamme wrote: > -BEGIN PGP SIGNED MESSAGE- > Hash: SHA1 > > Kushnir, Michael (NIH/NLM/LHC) [C] [22.12.2011 18:20]: >> Is it possible to have one machine be part of two different ocfs2 >> clusters with two different sans? Kind of to serve as a bridge for >> moving data between two clusters but without actually fully >> combining the two clusters? >> >> Thanks, Michael > Michael, > > I asked this two years ago and the answer was no. > > When I look at /etc/ocfs2/cluster.conf, I do not see a possibility to > configure a second cluster. Though the nodes must be assigned to a > cluster (and exactly one cluster, this is), there ist only one entry > "cluster:" in the file, and so there is no way to define a second one. > > We synced via rsync :-( > > HTH > Werner > > -BEGIN PGP SIGNATURE- > Version: GnuPG v2.0.18 (GNU/Linux) > Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/ > > iEYEARECAAYFAk7za4EACgkQk33Krq8b42MvSwCfQAXzqVQRPyhOdFrKM8PCPqbf > g0cAn20CV4rjzXNrTa/YGaUeNlO3+rmc > =CBmQ > -END PGP SIGNATURE- > > ___ > Ocfs2-users mailing list > Ocfs2-users@oss.oracle.com > http://oss.oracle.com/mailman/listinfo/ocfs2-users ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users
Re: [Ocfs2-users] One node, two clusters?
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Kushnir, Michael (NIH/NLM/LHC) [C] [22.12.2011 18:20]: > Is it possible to have one machine be part of two different ocfs2 > clusters with two different sans? Kind of to serve as a bridge for > moving data between two clusters but without actually fully > combining the two clusters? > > Thanks, Michael Michael, I asked this two years ago and the answer was no. When I look at /etc/ocfs2/cluster.conf, I do not see a possibility to configure a second cluster. Though the nodes must be assigned to a cluster (and exactly one cluster, this is), there ist only one entry "cluster:" in the file, and so there is no way to define a second one. We synced via rsync :-( HTH Werner -BEGIN PGP SIGNATURE- Version: GnuPG v2.0.18 (GNU/Linux) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/ iEYEARECAAYFAk7za4EACgkQk33Krq8b42MvSwCfQAXzqVQRPyhOdFrKM8PCPqbf g0cAn20CV4rjzXNrTa/YGaUeNlO3+rmc =CBmQ -END PGP SIGNATURE- ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users
[Ocfs2-users] One node, two clusters?
Is it possible to have one machine be part of two different ocfs2 clusters with two different sans? Kind of to serve as a bridge for moving data between two clusters but without actually fully combining the two clusters? Thanks, Michael ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users
Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both
Hello After 24 hours i see TEST-MAIL2 reboot ( possible kernel panic) but TEST-MAIL1 got in dmesg: TEST-MAIL1 ~ #dmesg [cut] o2net: accepted connection from node TEST-MAIL2 (num 1) at 172.17.1.252: o2dlm: Node 1 joins domain B24C4493BBC74FEAA3371E2534BB3611 o2dlm: Nodes in domain B24C4493BBC74FEAA3371E2534BB3611: 0 1 o2net: connection to node TEST-MAIL2 (num 1) at 172.17.1.252: has been idle for 60.0 seconds, shutting it down. (swapper,0,0):o2net_idle_timer:1562 Here are some times that might help debug the situation: (Timer: 33127732045, Now 33187808090, DataReady 33127732039, Advance 33127732051-33127732051, Key 0xebb9cd47, Func 506, FuncTime 33127732045-33127732048) o2net: no longer connected to node TEST-MAIL2 (num 1) at 172.17.1.252: (du,5099,12):dlm_do_master_request:1324 ERROR: link to 1 went down! (du,5099,12):dlm_get_lock_resource:907 ERROR: status = -112 (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: B24C4493BBC74FEAA3371E2534BB3611: res M0cf023ef70, error -112 send AST to node 1 (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -112 (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: B24C4493BBC74FEAA3371E2534BB3611: res P00, error -107 send AST to node 1 (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -107 (kworker/u:3,5071,0):o2net_connect_expired:1724 ERROR: no connection established with node 1 after 60.0 seconds, giving up and returning errors. (o2hb-B24C4493BB,14310,0):o2dlm_eviction_cb:267 o2dlm has evicted node 1 from group B24C4493BBC74FEAA3371E2534BB3611 (ocfs2rec,5504,6):dlm_get_lock_resource:834 B24C4493BBC74FEAA3371E2534BB3611:M15f023ef70: at least one node (1) to recover before lock mastery can begin (ocfs2rec,5504,6):dlm_get_lock_resource:888 B24C4493BBC74FEAA3371E2534BB3611:M15f023ef70: at least one node (1) to recover before lock mastery can begin (du,5099,12):dlm_restart_lock_mastery:1213 ERROR: node down! 1 (du,5099,12):dlm_wait_for_lock_mastery:1030 ERROR: status = -11 (du,5099,12):dlm_get_lock_resource:888 B24C4493BBC74FEAA3371E2534BB3611:N0020924f: at least one node (1) to recover before lock mastery can begin (dlm_reco_thread,14322,0):dlm_get_lock_resource:834 B24C4493BBC74FEAA3371E2534BB3611:$RECOVERY: at least one node (1) to recover before lock mastery can begin (dlm_reco_thread,14322,0):dlm_get_lock_resource:868 B24C4493BBC74FEAA3371E2534BB3611: recovery map is not empty, but must master $RECOVERY lock now (dlm_reco_thread,14322,0):dlm_do_recovery:523 (14322) Node 0 is the Recovery Master for the Dead Node 1 for Domain B24C4493BBC74FEAA3371E2534BB3611 (ocfs2rec,5504,6):ocfs2_replay_journal:1549 Recovering node 1 from slot 1 on device (253,0) (ocfs2rec,5504,6):ocfs2_begin_quota_recovery:407 Beginning quota recovery in slot 1 (kworker/u:0,2909,0):ocfs2_finish_quota_recovery:599 Finishing quota recovery in slot 1 And i try give this command: debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP allow debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or directory debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP off debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or directory But not working -Oryginalna wiadomość- From: Srinivas Eeda Sent: Wednesday, December 21, 2011 8:43 PM To: Marek Królikowski Cc: ocfs2-users@oss.oracle.com Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both Those numbers look good. Basically with the fixes backed out and another fix I gave, you are not seeing that many orphans hanging around and hence not seeing the process stuck kernel stacks. You can run the test longer or if you are satisfied, please enable quotas and re-run the test with the modified kernel. You might see a dead lock which needs to be fixed(I was not able to reproduce this yet). If the system hangs, please capture the following and provide me the output 1. echo t > /proc/sysrq-trigger 2. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP allow 3. wait for 10 minutes 4. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP off 5. echo t > /proc/sysrq-trigger ___ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users