On 2018-01-14 12:29 PM, Digimer wrote:
> I recently changed the host name of a cluster. It may or may not be
> related, but after I noticed that I can cleanly start gfs2 when the node
> boots. However, if the node is withdrawn and then I try to rejoin it
> without a reboot, it hangs with this in syslog;
> 
> ====
> Jan 14 12:21:34 kp-a10n01 kernel: Pid: 22580, comm: kslowd000 Not
> tainted 2.6.32-696.18.7.el6.x86_64 #1
> Jan 14 12:21:34 kp-a10n01 kernel: Call Trace:
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffffa0714308>] ?
> gfs2_lm_withdraw+0x128/0x160 [gfs2]
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffffa071451d>] ?
> gfs2_consist_inode_i+0x5d/0x60 [gfs2]
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffffa070b466>] ?
> find_good_lh+0x76/0x90 [gfs2]
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffffa070b509>] ?
> gfs2_find_jhead+0x89/0x170 [gfs2]
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff8107e0ee>] ?
> vprintk_default+0xe/0x10
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffffa070b6ee>] ?
> gfs2_recover_work+0xfe/0x790 [gfs2]
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff8106b73e>] ?
> perf_event_task_sched_out+0x2e/0x70
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff8100968b>] ?
> __switch_to+0x6b/0x320
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff8154b728>] ?
> schedule+0x458/0xc50
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff81063883>] ? __wake_up+0x53/0x70
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff81121be3>] ?
> slow_work_execute+0x233/0x310
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff81121e17>] ?
> slow_work_thread+0x157/0x360
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff810a71a0>] ?
> autoremove_wake_function+0x0/0x40
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff81121cc0>] ?
> slow_work_thread+0x0/0x360
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff810a6d0e>] ? kthread+0x9e/0xc0
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff81557afa>] ? child_rip+0xa/0x20
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff810a6c70>] ? kthread+0x0/0xc0
> Jan 14 12:21:34 kp-a10n01 kernel: [<ffffffff81557af0>] ? child_rip+0x0/0x20
> Jan 14 12:21:34 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
> jid=0: Failed
> Jan 14 12:21:34 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
> jid=1: Trying to acquire journal lock...
> Jan 14 12:21:34 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0: can't
> read in statfs inode: -5
> Jan 14 12:21:34 kp-a10n01 gfs_controld[20749]: recovery_uevent mg not
> found shared
> Jan 14 12:21:34 kp-a10n01 gfs_controld[20749]: recovery_uevent mg not
> found shared
> Jan 14 12:21:34 kp-a10n01 gfs_controld[20749]: recovery_uevent mg not
> found shared
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 2,19 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 5,19 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 5,805b err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 5,18 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 5,17 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 5,16 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 1,0 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 2,18 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 9,0 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 1,1 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 4,0 err=-22
> Jan 14 12:21:34 kp-a10n01 kernel: gdlm_unlock 2,17 err=-22
> ====
> 
> I have to fence the node to get the system back up. It happens on either
> node, and it happens regardless of the peer node being connected.
> 
> GFS2 on top of clvmd on an rhcs cluster on RHEL 6. Would configs help?
> 
> digimer
> 

Happened again (well, many times, but here's the log output from another
hang);

====
Jan 14 12:46:41 kp-a10n01 kernel: GFS2 (built Jan  4 2018 17:32:36)
installed
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=: Trying to join cluster
"lock_dlm", "kp-anvil-10:shared"
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
Joined cluster. Now mounting FS...
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
jid=0, already locked for use
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
jid=0: Looking at journal...
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
fatal: filesystem consistency error
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
inode = 4 25
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
function = find_good_lh, file = fs/gfs2/recovery.c, line = 205
Jan 14 12:46:41 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0: about
to withdraw this file system
Jan 14 12:46:42 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
telling LM to unmount
Jan 14 12:46:42 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0: withdrawn
Jan 14 12:46:42 kp-a10n01 kernel: Pid: 11668, comm: kslowd000 Not
tainted 2.6.32-696.18.7.el6.x86_64 #1
Jan 14 12:46:42 kp-a10n01 kernel: Call Trace:
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffffa06fb308>] ?
gfs2_lm_withdraw+0x128/0x160 [gfs2]
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffffa06fb51d>] ?
gfs2_consist_inode_i+0x5d/0x60 [gfs2]
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffffa06f2466>] ?
find_good_lh+0x76/0x90 [gfs2]
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffffa06f2509>] ?
gfs2_find_jhead+0x89/0x170 [gfs2]
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff8107e0ee>] ?
vprintk_default+0xe/0x10
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffffa06f26ee>] ?
gfs2_recover_work+0xfe/0x790 [gfs2]
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff8106b73e>] ?
perf_event_task_sched_out+0x2e/0x70
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff81074a83>] ?
dequeue_entity+0x113/0x2e0
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff8100968b>] ?
__switch_to+0x6b/0x320
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff8154b728>] ?
schedule+0x458/0xc50
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff8107543b>] ?
enqueue_task_fair+0xfb/0x100
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff81121be3>] ?
slow_work_execute+0x233/0x310
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff81121e17>] ?
slow_work_thread+0x157/0x360
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff810a71a0>] ?
autoremove_wake_function+0x0/0x40
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff81121cc0>] ?
slow_work_thread+0x0/0x360
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff810a6d0e>] ? kthread+0x9e/0xc0
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff81557afa>] ? child_rip+0xa/0x20
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff810a6c70>] ? kthread+0x0/0xc0
Jan 14 12:46:42 kp-a10n01 kernel: [<ffffffff81557af0>] ? child_rip+0x0/0x20
Jan 14 12:46:42 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
jid=0: Failed
Jan 14 12:46:42 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0:
jid=1: Trying to acquire journal lock...
Jan 14 12:46:42 kp-a10n01 kernel: GFS2: fsid=kp-anvil-10:shared.0: can't
read in statfs inode: -5
Jan 14 12:46:42 kp-a10n01 gfs_controld[5371]: recovery_uevent mg not
found shared
Jan 14 12:46:42 kp-a10n01 gfs_controld[5371]: recovery_uevent mg not
found shared
Jan 14 12:46:42 kp-a10n01 gfs_controld[5371]: recovery_uevent mg not
found shared
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 2,19 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 5,19 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 5,805b err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 5,18 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 5,17 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 5,16 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 1,0 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 4,0 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 2,17 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 2,18 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 9,0 err=-22
Jan 14 12:46:42 kp-a10n01 kernel: gdlm_unlock 1,1 err=-22
====


-- 
Digimer
Papers and Projects: https://alteeve.com/w/
"I am, somehow, less interested in the weight and convolutions of
Einstein’s brain than in the near certainty that people of equal talent
have lived and died in cotton fields and sweatshops." - Stephen Jay Gould

_______________________________________________
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to