Re: [Ocfs2-users] BUG at fs/ocfs2/extent_map.c:476

2012-05-08 Thread richard -rw- weinberger
On Tue, May 8, 2012 at 2:38 AM, Joel Becker  wrote:
> On Mon, Apr 02, 2012 at 10:36:53AM +0200, richard -rw- weinberger wrote:
>> On Mon, Apr 2, 2012 at 10:25 AM, Joel Becker  wrote:
>> > On Sat, Mar 24, 2012 at 09:50:49AM +0100, richard -rw- weinberger wrote:
>> >> Hi!
>> >>
>> >> Today one of my kvm hosts ran into this BUG.
>> >> The host is member of y small ocfs2-cluster (3 hosts).
>> >> CentOS 6.2, Kernel 3.1.1 (vanilla).
>> > ...
>> >> Is this a known/fixed bug?
>> >
>> > Not that I know of.  Have you seen it more than once?
>> >
>>
>> No. All three cluster members had an uptime of 70 days.
>> Only one showed the BUG.
>> I've rebooted it and it seems to work fine now.
>>
>> I'm not an ocfs2 expert, what could cause this kind of BUG?
>> Internal logic error? Filesystem corruption?
>
>        No idea.  That's the problem.  More details would be nice, but I
> suspect you'd rather not crash again :-)

Sorry, that's all I know.
The crash happened only once.

-- 
Thanks,
//richard

___
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users


Re: [Ocfs2-users] BUG at fs/ocfs2/extent_map.c:476

2012-05-07 Thread Joel Becker
On Mon, Apr 02, 2012 at 10:36:53AM +0200, richard -rw- weinberger wrote:
> On Mon, Apr 2, 2012 at 10:25 AM, Joel Becker  wrote:
> > On Sat, Mar 24, 2012 at 09:50:49AM +0100, richard -rw- weinberger wrote:
> >> Hi!
> >>
> >> Today one of my kvm hosts ran into this BUG.
> >> The host is member of y small ocfs2-cluster (3 hosts).
> >> CentOS 6.2, Kernel 3.1.1 (vanilla).
> > ...
> >> Is this a known/fixed bug?
> >
> > Not that I know of.  Have you seen it more than once?
> >
> 
> No. All three cluster members had an uptime of 70 days.
> Only one showed the BUG.
> I've rebooted it and it seems to work fine now.
> 
> I'm not an ocfs2 expert, what could cause this kind of BUG?
> Internal logic error? Filesystem corruption?

No idea.  That's the problem.  More details would be nice, but I
suspect you'd rather not crash again :-)

Joel

-- 

"Soap and education are not as sudden as a massacre, but they are more
 deadly in the long run."
- Mark Twain


http://www.jlbec.org/
jl...@evilplan.org

___
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users


Re: [Ocfs2-users] BUG at fs/ocfs2/extent_map.c:476

2012-04-02 Thread richard -rw- weinberger
On Mon, Apr 2, 2012 at 10:25 AM, Joel Becker  wrote:
> On Sat, Mar 24, 2012 at 09:50:49AM +0100, richard -rw- weinberger wrote:
>> Hi!
>>
>> Today one of my kvm hosts ran into this BUG.
>> The host is member of y small ocfs2-cluster (3 hosts).
>> CentOS 6.2, Kernel 3.1.1 (vanilla).
> ...
>> Is this a known/fixed bug?
>
> Not that I know of.  Have you seen it more than once?
>

No. All three cluster members had an uptime of 70 days.
Only one showed the BUG.
I've rebooted it and it seems to work fine now.

I'm not an ocfs2 expert, what could cause this kind of BUG?
Internal logic error? Filesystem corruption?

-- 
Thanks,
//richard

___
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users


Re: [Ocfs2-users] BUG at fs/ocfs2/extent_map.c:476

2012-04-02 Thread Joel Becker
On Sat, Mar 24, 2012 at 09:50:49AM +0100, richard -rw- weinberger wrote:
> Hi!
> 
> Today one of my kvm hosts ran into this BUG.
> The host is member of y small ocfs2-cluster (3 hosts).
> CentOS 6.2, Kernel 3.1.1 (vanilla).
...
> Is this a known/fixed bug?

Not that I know of.  Have you seen it more than once?

Joel

-- 

None of our men are "experts."  We have most unfortunately found
it necessary to get rid of a man as soon as he thinks himself an
expert -- because no one ever considers himself expert if he really
knows his job.  A man who knows a job sees so much more to be done
than he has done, that he is always pressing forward and never
gives up an instant of thought to how good and how efficient he is.
Thinking always ahead, thinking always of trying to do more, brings
a state of mind in which nothing is impossible. The moment one gets
into the "expert" state of mind a great number of things become
impossible.
- From Henry Ford Sr., "My Life and Work"

http://www.jlbec.org/
jl...@evilplan.org

___
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users


Re: [Ocfs2-users] BUG at fs/ocfs2/extent_map.c:476

2012-03-29 Thread richard -rw- weinberger
On Sat, Mar 24, 2012 at 10:50 AM, richard -rw- weinberger
 wrote:
> Hi!
>
> Today one of my kvm hosts ran into this BUG.
> The host is member of y small ocfs2-cluster (3 hosts).
> CentOS 6.2, Kernel 3.1.1 (vanilla).
>
> [8138858.899066] [ cut here ]
> [8138858.899156] kernel BUG at fs/ocfs2/extent_map.c:476!
> [8138858.899232] invalid opcode:  [#1] SMP
> [8138858.899314] CPU 4
> [8138858.899324] Modules linked in: ip6table_filter ip6_tables
> iptable_filter ip_tables ebtable_nat ebtables netconsole ocfs2
> ocfs2_dlmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodemanager
> ocfs2_stackglue configfs sunrpc scsi_dh_rdac bridge 8021q garp stp llc
> be2iscsi iscsi_boot_sysfs bnx2i cnic uio cxgb3i libcxgbi cxgb3 mdio
> ib_iser rdma_cm ib_cm iw_cm ib_sa ib_mad ib_core ib_addr ipv6
> iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi dm_round_robin
> dm_multipath vhost_net macvtap macvlan tun kvm_intel kvm serio_raw
> pcspkr i2c_i801 i2c_core iTCO_wdt iTCO_vendor_support ioatdma
> i7core_edac edac_core sg ext4 mbcache jbd2 sd_mod crc_t10dif pata_acpi
> ata_generic ata_piix 3w_sas igb dca dm_mod [last unloaded:
> scsi_wait_scan]
> [8138858.900430]
> [8138858.900500] Pid: 3044, comm: qemu-kvm Not tainted 3.1.1 #1
> Supermicro X8DT3/X8DT3
> [8138858.900656] RIP: 0010:[]  []
> ocfs2_get_clusters_nocache+0x4d1/0x520 [ocfs2]
> [8138858.900836] RSP: 0018:8803efa01688  EFLAGS: 00010287
> [8138858.900911] RAX: 002f RBX: 8802ebe79030 RCX:
> 00062114
> [8138858.901048] RDX: 0300 RSI: 000620ec RDI:
> 8802ebe79320
> [8138858.901183] RBP: 8803efa01708 R08: 00062114 R09:
> 
> [8138858.901322] R10:  R11:  R12:
> 8803efa01738
> [8138858.901458] R13: 8804d6095000 R14:  R15:
> 0001
> [8138858.901595] FS:  7fa10d5f9700() GS:88063fc0()
> knlGS:
> [8138858.901734] CS:  0010 DS:  ES:  CR0: 8005003b
> [8138858.901809] CR2: 006bdc5c CR3: 0004cf67 CR4:
> 26e0
> [8138858.901942] DR0: 0001 DR1: 0002 DR2:
> 0001
> [8138858.902075] DR3: 000a DR6: 0ff0 DR7:
> 0400
> [8138858.902208] Process qemu-kvm (pid: 3044, threadinfo
> 8803efa0, task 880349986080)
> [8138858.902343] Stack:
> [8138858.902408]   8803 8802ebe79000
> 8803efa01754
> [8138858.902556]  880505f9acb0 000620ec 8803efa01874
> 8803efa017c4
> [8138858.902706]  880002557130 a04fdade 88062a8f2000
> 000620ec
> [8138858.902854] Call Trace:
> [8138858.902945]  [] ?
> ocfs2_read_inode_block_full+0x3e/0x60 [ocfs2]
> [8138858.903100]  [] ocfs2_get_clusters+0x1f3/0x380 [ocfs2]
> [8138858.903197]  []
> ocfs2_extent_map_get_blocks+0x5d/0x1a0 [ocfs2]
> [8138858.903337]  [] ? __css_put+0x3f/0x90
> [8138858.903429]  [] ocfs2_get_block+0xad/0x9a0 [ocfs2]
> [8138858.903512]  [] ? zone_statistics+0x99/0xc0
> [8138858.903595]  [] ? __mem_cgroup_commit_charge+0x6c/0xc0
> [8138858.903679]  [] do_mpage_readpage+0x492/0x630
> [8138858.903758]  [] ? __inc_zone_page_state+0x35/0x40
> [8138858.903841]  [] ? add_to_page_cache_locked+0xd1/0x140
> [8138858.903921]  [] mpage_readpages+0x102/0x160
> [8138858.904015]  [] ?
> ocfs2_map_page_blocks+0x3b0/0x3b0 [ocfs2]
> [8138858.904170]  [] ?
> ocfs2_map_page_blocks+0x3b0/0x3b0 [ocfs2]
> [8138858.904331]  [] ?
> ocfs2_inode_lock_full_nested+0x28a/0xad0 [ocfs2]
> [8138858.904484]  [] ? ocfs2_readpages+0x41/0xe0 [ocfs2]
> [8138858.904583]  [] ocfs2_readpages+0xc7/0xe0 [ocfs2]
> [8138858.904662]  [] __do_page_cache_readahead+0x1b2/0x260
> [8138858.904740]  [] ra_submit+0x21/0x30
> [8138858.904815]  [] ondemand_readahead+0x115/0x240
> [8138858.904894]  [] page_cache_async_readahead+0x88/0xb0
> [8138858.904974]  [] generic_file_aio_read+0x48b/0x740
> [8138858.905074]  [] ocfs2_file_aio_read+0xe3/0x380 [ocfs2]
> [8138858.905155]  [] do_sync_read+0xda/0x120
> [8138858.905233]  [] ? security_file_permission+0x8b/0x90
> [8138858.905315]  [] vfs_read+0xc5/0x190
> [8138858.905391]  [] sys_pread64+0xa2/0xb0
> [8138858.905469]  [] system_call_fastpath+0x16/0x1b
> [8138858.905545] Code: c0 90 56 a0 bb e2 ff ff ff 48 8b 42 28 48 8b 4a
> 40 48 c7 c2 78 2e 57 a0 48 89 c7 31 c0 e8 18 b1 04 00 48 8b 7d b8 e9
> 37 fe ff ff <0f> 0b eb fe 65 48 8b 34 25 40 c4 00 00 8b 96 8c 04 00 00
> 48 63
> [8138858.906011] RIP  []
> ocfs2_get_clusters_nocache+0x4d1/0x520 [ocfs2]
> [8138858.906174]  RSP 
> [8138858.906541] ---[ end trace 5b1e50f0d8b3c799 ]---
>
> Is this a known/fixed bug?

*ping*

-- 
Thanks,
//richard

___
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users


[Ocfs2-users] BUG at fs/ocfs2/extent_map.c:476

2012-03-24 Thread richard -rw- weinberger
Hi!

Today one of my kvm hosts ran into this BUG.
The host is member of y small ocfs2-cluster (3 hosts).
CentOS 6.2, Kernel 3.1.1 (vanilla).

[8138858.899066] [ cut here ]
[8138858.899156] kernel BUG at fs/ocfs2/extent_map.c:476!
[8138858.899232] invalid opcode:  [#1] SMP
[8138858.899314] CPU 4
[8138858.899324] Modules linked in: ip6table_filter ip6_tables
iptable_filter ip_tables ebtable_nat ebtables netconsole ocfs2
ocfs2_dlmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodemanager
ocfs2_stackglue configfs sunrpc scsi_dh_rdac bridge 8021q garp stp llc
be2iscsi iscsi_boot_sysfs bnx2i cnic uio cxgb3i libcxgbi cxgb3 mdio
ib_iser rdma_cm ib_cm iw_cm ib_sa ib_mad ib_core ib_addr ipv6
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi dm_round_robin
dm_multipath vhost_net macvtap macvlan tun kvm_intel kvm serio_raw
pcspkr i2c_i801 i2c_core iTCO_wdt iTCO_vendor_support ioatdma
i7core_edac edac_core sg ext4 mbcache jbd2 sd_mod crc_t10dif pata_acpi
ata_generic ata_piix 3w_sas igb dca dm_mod [last unloaded:
scsi_wait_scan]
[8138858.900430]
[8138858.900500] Pid: 3044, comm: qemu-kvm Not tainted 3.1.1 #1
Supermicro X8DT3/X8DT3
[8138858.900656] RIP: 0010:[]  []
ocfs2_get_clusters_nocache+0x4d1/0x520 [ocfs2]
[8138858.900836] RSP: 0018:8803efa01688  EFLAGS: 00010287
[8138858.900911] RAX: 002f RBX: 8802ebe79030 RCX:
00062114
[8138858.901048] RDX: 0300 RSI: 000620ec RDI:
8802ebe79320
[8138858.901183] RBP: 8803efa01708 R08: 00062114 R09:

[8138858.901322] R10:  R11:  R12:
8803efa01738
[8138858.901458] R13: 8804d6095000 R14:  R15:
0001
[8138858.901595] FS:  7fa10d5f9700() GS:88063fc0()
knlGS:
[8138858.901734] CS:  0010 DS:  ES:  CR0: 8005003b
[8138858.901809] CR2: 006bdc5c CR3: 0004cf67 CR4:
26e0
[8138858.901942] DR0: 0001 DR1: 0002 DR2:
0001
[8138858.902075] DR3: 000a DR6: 0ff0 DR7:
0400
[8138858.902208] Process qemu-kvm (pid: 3044, threadinfo
8803efa0, task 880349986080)
[8138858.902343] Stack:
[8138858.902408]   8803 8802ebe79000
8803efa01754
[8138858.902556]  880505f9acb0 000620ec 8803efa01874
8803efa017c4
[8138858.902706]  880002557130 a04fdade 88062a8f2000
000620ec
[8138858.902854] Call Trace:
[8138858.902945]  [] ?
ocfs2_read_inode_block_full+0x3e/0x60 [ocfs2]
[8138858.903100]  [] ocfs2_get_clusters+0x1f3/0x380 [ocfs2]
[8138858.903197]  []
ocfs2_extent_map_get_blocks+0x5d/0x1a0 [ocfs2]
[8138858.903337]  [] ? __css_put+0x3f/0x90
[8138858.903429]  [] ocfs2_get_block+0xad/0x9a0 [ocfs2]
[8138858.903512]  [] ? zone_statistics+0x99/0xc0
[8138858.903595]  [] ? __mem_cgroup_commit_charge+0x6c/0xc0
[8138858.903679]  [] do_mpage_readpage+0x492/0x630
[8138858.903758]  [] ? __inc_zone_page_state+0x35/0x40
[8138858.903841]  [] ? add_to_page_cache_locked+0xd1/0x140
[8138858.903921]  [] mpage_readpages+0x102/0x160
[8138858.904015]  [] ?
ocfs2_map_page_blocks+0x3b0/0x3b0 [ocfs2]
[8138858.904170]  [] ?
ocfs2_map_page_blocks+0x3b0/0x3b0 [ocfs2]
[8138858.904331]  [] ?
ocfs2_inode_lock_full_nested+0x28a/0xad0 [ocfs2]
[8138858.904484]  [] ? ocfs2_readpages+0x41/0xe0 [ocfs2]
[8138858.904583]  [] ocfs2_readpages+0xc7/0xe0 [ocfs2]
[8138858.904662]  [] __do_page_cache_readahead+0x1b2/0x260
[8138858.904740]  [] ra_submit+0x21/0x30
[8138858.904815]  [] ondemand_readahead+0x115/0x240
[8138858.904894]  [] page_cache_async_readahead+0x88/0xb0
[8138858.904974]  [] generic_file_aio_read+0x48b/0x740
[8138858.905074]  [] ocfs2_file_aio_read+0xe3/0x380 [ocfs2]
[8138858.905155]  [] do_sync_read+0xda/0x120
[8138858.905233]  [] ? security_file_permission+0x8b/0x90
[8138858.905315]  [] vfs_read+0xc5/0x190
[8138858.905391]  [] sys_pread64+0xa2/0xb0
[8138858.905469]  [] system_call_fastpath+0x16/0x1b
[8138858.905545] Code: c0 90 56 a0 bb e2 ff ff ff 48 8b 42 28 48 8b 4a
40 48 c7 c2 78 2e 57 a0 48 89 c7 31 c0 e8 18 b1 04 00 48 8b 7d b8 e9
37 fe ff ff <0f> 0b eb fe 65 48 8b 34 25 40 c4 00 00 8b 96 8c 04 00 00
48 63
[8138858.906011] RIP  []
ocfs2_get_clusters_nocache+0x4d1/0x520 [ocfs2]
[8138858.906174]  RSP 
[8138858.906541] ---[ end trace 5b1e50f0d8b3c799 ]---

Is this a known/fixed bug?

-- 
Thanks,
//richard

___
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users