2010/11/1 Tao Ma <tao...@oracle.com>:
> Hi Ronald,
>
> On 10/29/2010 06:03 PM, Ronald Moesbergen wrote:
>>
>> 2010/10/29 Ronald Moesbergen<intercom...@gmail.com>:
>>>
>>> 2010/10/29 Tao Ma<tao...@oracle.com>:
>>>>
>>>> Hi Ronald,
>>>
>>> Hi Tao,
>>>
>>> Thanks for looking into this.
>>>
>>>> On 10/29/2010 05:12 PM, Ronald Moesbergen wrote:
>>>>>
>>>>> Hello,
>>>>>
>>>>> I was testing kernel 2.6.36 (vanilla mainline) and encountered the
>>>>> following BUG():
>>>>>
>>>>> [157756.266000] o2net: no longer connected to node app01 (num 0) at
>>>>> 10.2.25.13:7777
>>>>> [157756.266077] (o2hb-5FA56B1D0A,2908,0):o2dlm_eviction_cb:267 o2dlm
>>>>> has evicted node 0 from group 5FA56B1D0A9249099CE58C82CFEC873A
>>>>> [157756.274443] (ocfs2rec,14060,0):dlm_get_lock_resource:836
>>>>> 5FA56B1D0A9249099CE58C82CFEC873A:M00000000000000000000186ba2b09b: at
>>>>> least one node (0) to recover before lock mastery can begin
>>>>> [157757.275776] (ocfs2rec,14060,0):dlm_get_lock_resource:890
>>>>> 5FA56B1D0A9249099CE58C82CFEC873A:M00000000000000000000186ba2b09b: at
>>>>> least one node (0) to recover before lock mastery can begin
>>>>> [157760.774045] (dlm_reco_thread,2920,2):dlm_get_lock_resource:836
>>>>> 5FA56B1D0A9249099CE58C82CFEC873A:$RECOVERY: at least one node (0) to
>>>>> recover before lock mastery can begin
>>>>> [157760.774124] (dlm_reco_thread,2920,2):dlm_get_lock_resource:870
>>>>> 5FA56B1D0A9249099CE58C82CFEC873A: recovery map is not empty, but must
>>>>> master $RECOVERY lock now
>>>>> [157760.774205] (dlm_reco_thread,2920,2):dlm_do_recovery:523 (2920)
>>>>> Node 1 is the Recovery Master for the Dead Node 0 for Domain
>>>>> 5FA56B1D0A9249099CE58C82CFEC873A
>>>>> [157768.261818] (ocfs2rec,14060,0):ocfs2_replay_journal:1605
>>>>> Recovering node 0 from slot 0 on device (8,32)
>>>>> [157772.850182] ------------[ cut here ]------------
>>>>> [157772.850211] kernel BUG at fs/ocfs2/journal.c:1700!
>>>>
>>>> Strange. the bug line is
>>>> BUG_ON(osb->node_num == node_num);
>>>> and it detects the same node number in the cluster.
>>
>> I just tried to reproduce it and succeeded. Here's what I did:
>> - unmount the filesystem on node app02
>> - shutdown the o2cb services on app02
>> - Do a halt -f on app01, which still has the OCFS2 volume mounted.
>> - Start o2cb services on app02
>> - Mount the OCFS2 filesystem ->  BUG
>>
>> Works everytime. So one of the 2 variables checked in that BUG_ON
>> statement must no be set correctly somewhere.
>
> I have tried several times in my local test env, but with no luck by now.
> And it seems to me quite strange at least from the code.
>
> So could you please file a bug in oss.oracle.com/bugzilla so that it is
> easier to track and discuss? Great thanks.

Ok, it's filed as:
http://oss.oracle.com/bugzilla/show_bug.cgi?id=1296

I've tried to reproduce on another cluster and there I don't see the
bug either, so it must be something specific to this setup.

Regards,
Ronald.

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users

Reply via email to