i 'm waiting for any advice tank you very much in advance and my problem is below probably it'll be more nice to read at the mailing list archive for my posts ( sender svetljo ) and the subject should be smth similar to the one of this mail for the end of august ( i think 30 and 31) and for september http://lists.sistina.com/pipermail/linux-lvm/ >>>> it's me again and out there is no one >>>> so >>>> well linux-2.4.10-pre2-xfs LVM-1.0.1rc2 works with XFS and JFS on >>>> the LV >>>> ext2 and reiserfs segfaults as befor >>>> i've compiled the other kernel's with JFS 1.0.4 but i don't see the >>>> point to test them >>>> i'm shure that JFS will work on the LV and the story with ext2 and >>>> reiserfs will be the same >>>> but i'll probably try the ReiserFS patches for 2.4.9 ( after >>>> www.namesys.com they already sent them to Linus ) >>>> >>>> i'm using e2fsprogs-1.2.4 and reiserfsprogs-3.x.0j >>>> >>>> svetljo wrote: >>>> >>>> > well i did the tests >>>> > >>>> > clean Linus kernel-2.4.9 LVM-1.0.1rc1 ext2 and >>>> > reiserfs segfaults >>>> > clean linux-2.4.9-ac5 LVM-1.0 ext2 >>>> > and reiserfs segfaults >>>> > | linux-2.4.10-pre2-xfs LVM-1.0.1rc2 ext2 and >>>> > reiserfs segfaults, but xfs seems to work >>>> > | from SGI's cvs tree linux-2.4-xfs taken >>>> today in >>>> > the early morning >>>> > >>>> > i'll try to ad IBM's JFS and try it again and i'll try bonnie on >>>> them >>>> > can you give me ideas how to stres the FS to find out whether it >>>> realy >>>> > works with XFS and JFS >>>> > >>>> > what could be wrong with ext2 and reiserfs as it works with >>>> JFS-1.0.3 >>>> > and with the latest cvs XFS >>>> > svetljo wrote: >>>> > >>>> >> Hi >>>> >> >>>> >> a small adition i've tried to format the LV with ext2 and reiserfs, >>>> >> but it didn't worked : mkfs segfaults >>>> >> a strange one : i'm able to format with IBM JFS , and i can work >>>> >> without a problem with the LV everything just to be fine with JFS >>>> >> >>>> >> i'm currently building: >>>> >> clean 2.4.9-linus with LVM-1.0.1rc1 >>>> >> 2.4.9-ac5 with LVM-1.0 ( i couldn't do it with >>>> LVM-1.0.1rc1 & rc2) >>>> >> 2.4.10-pre2-xfs-cvs with LVM-1.0.1rc2 >>>> >> to find out what is going on with ext2 reiserfs XFS , >>>> >> is the problem coming from the XFS kernel changes >>>> >> >>>> >> >Hi >>>> >> >i'm having a serios trouble with creating >>>> >> >a LVM over software linear RAID >>>> >> >well i created it, formated it with XFS >>>> >> >but every time i try to mount the LV mount segfaults >>>> >> >and then i can not mount any other file system ( partition, CD, .. >>>> >> >until i reboot, when i try to mount smth mount simple stop to >>>> respond >>>> >> >without an error and blocks the console >>>> >> > >>>> >> >i'm using XFS cvs kernel 2.4.9 and LVM-1.0.1rc1 >>>> >> >on ABIT's BP6 2xCelleron 533 512Mb RAM >>>> >> >the drives are on onboard HPT366 controler 2xWD 30Gb 1xMaxtor 40Gb >>>> >> > >>>> >> >the LV is striped over the 3 devices of the VG >>>> >> >the VG is /dev/hdh10 /dev/md6 /dev/md7 >>>> >> >/dev/md6 is linear software RAID /dev/hde6 /dev/hde12 >>>> >> >/dev/md7 is linear software RAID /dev/hdg1 /dev/hdg5 /dev/hdg6 >>>> >> >dev/hdg11 >>>> >> > >>>> >> >i posted to the LVM-lists and there i was told >>>> >> >to try "dmesg | ksymoops" >>>> >> > >>>> >> >and i became the folowing answer >>>> >> > >>>> >> > > >>EIP; e29c0266 <[linear]linear_make_request+36/f0> <===== >>>> >> > > >>>> >> > >> Trace; c023fa12 <__make_request+412/6d0> >>>> >> > >> Trace; c0278dcd <md_make_request+4d/80> >>>> >> > >> Trace; c027fa0f <lvm_make_request_fn+f/20> >>>> >> > >> Trace; c023fd89 <generic_make_request+b9/120> >>>> >> > > >>>> >> > >OK, so the oops is inside the RAID layer, but it may be that >>>> it is >>>> >> > >being fed bogus data from a higher layer. Even so, it should >>>> not >>>> >> > >oops in this case. Since XFS changes a lot of the kernel >>>> code, I >>>> >> > >would either suggest asking the XFS folks to look at this oops, >>>> >> > >or maybe on the MD RAID mailing list, as they will know more >>>> about >>>> >> >it. >>>> >> > >>>> >> >this is the full "dmesg | ksymoops" , i'll try to use other FS >>>> to find >>>> >> >out whether it's a problem with XFS, but i wish me not to have >>>> to use >>>> >> >other FS, i realy love XFS >>>> >> >>>> >> >>>> >> Using defaults from ksymoops -t elf32-i386 -a i386 >>>> >> EFLAGS: 00010247 >>>> >> eax: 004ac1ab ebx: 004ac1ab ecx: 00000000 edx: 00000000 >>>> >> esi: d54eb320 edi: c188b928 ebp: 00958357 esp: d4eb3670 >>>> >> ds: 0018 es: 0018 ss: 0018 >>>> >> Process mount (pid: 5536, stackpage=d4eb3000) >>>> >> Stack: d54eb3e0 c023fa12 00000907 d54eb320 00000000 01c02000 >>>> c0278dcd >>>> >> dcec43c0 >>>> >> 00000000 d54eb320 d54eb320 00000000 01c02000 c027fa0f >>>> 00000001 >>>> >> d54eb320 >>>> >> c023fd89 c03a7254 00000000 d54eb320 00000282 00000021 >>>> 00000000 >>>> >> 00000000 >>>> >> Call Trace: [<c023fa12>] [<c0278dcd>] [<c027fa0f>] [<c023fd89>] >>>> >> [<c01a6814>] >>>> >> [<c01a6a85>] [<c01a6fc1>] [<c01a6c47>] [<c01a6990>] >>>> [<c0105dac>] >>>> >> [<c0105f1c>] >>>> >> [<c02e2140>] [<c021c10a>] [<c01fe5b8>] [<c01ff2a4>] >>>> [<c01a553e>] >>>> >> [<c01feb6f>] >>>> >> [<c01feed8>] [<c01fc322>] [<c0201f40>] [<c01fb8f3>] >>>> [<c0202fdf>] >>>> >> [<c02026bf>] >>>> >> [<c01a60be>] [<c02026eb>] [<c021e674>] [<c020b69c>] >>>> [<c020b843>] >>>> >> [<c020b871>] >>>> >> [<c021cf48>] [<c01294e0>] [<c0125f0e>] [<c0125d9d>] >>>> [<c013cd72>] >>>> >> [<c013d01b>] >>>> >> [<c013dafc>] [<c01131e0>] [<c010724c>] [<c013dd56>] >>>> [<c013dbfc>] >>>> >> [<c013de13>] >>>> >> [<c010715b>] >>>> >> Code: f7 f9 85 d2 74 24 55 51 68 c0 03 9c e2 e8 58 6c 75 dd 6a 00 >>>> >> >>>> >> >>EIP; e29c0266 <[linear]linear_make_request+36/f0> <===== >>>> >> Trace; c023fa12 <__make_request+412/6d0> >>>> >> Trace; c0278dcd <md_make_request+4d/80> >>>> >> Trace; c027fa0f <lvm_make_request_fn+f/20> >>>> >> Trace; c023fd89 <generic_make_request+b9/120> >>>> >> Trace; c01a6814 <_pagebuf_page_io+1f4/370> >>>> >> Trace; c01a6a85 <_page_buf_page_apply+f5/1c0> >>>> >> Trace; c01a6fc1 <pagebuf_segment_apply+b1/e0> >>>> >> Trace; c01a6c47 <pagebuf_iorequest+f7/160> >>>> >> Trace; c01a6990 <_page_buf_page_apply+0/1c0> >>>> >> Trace; c0105dac <__down+bc/d0> >>>> >> Trace; c0105f1c <__down_failed+8/c> >>>> >> Trace; c02e2140 <stext_lock+45b4/99d6> >>>> >> Trace; c021c10a <xfsbdstrat+3a/40> >>>> >> Trace; c01fe5b8 <xlog_bread+48/80> >>>> >> Trace; c01ff2a4 <xlog_find_zeroed+94/1e0> >>>> >> Trace; c01a553e <_pagebuf_get_object+3e/170> >>>> >> Trace; c01feb6f <xlog_find_head+1f/370> >>>> >> Trace; c01feed8 <xlog_find_tail+18/350> >>>> >> Trace; c01fc322 <xlog_alloc_log+2a2/2e0> >>>> >> Trace; c0201f40 <xlog_recover+20/c0> >>>> >> Trace; c01fb8f3 <xfs_log_mount+73/b0> >>>> >> Trace; c0202fdf <xfs_mountfs+55f/e20> >>>> >> Trace; c02026bf <xfs_readsb+af/f0> >>>> >> Trace; c01a60be <pagebuf_rele+3e/80> >>>> >> Trace; c02026eb <xfs_readsb+db/f0> >>>> >> Trace; c021e674 <kmem_alloc+e4/110> >>>> >> Trace; c020b69c <xfs_cmountfs+4bc/590> >>>> >> Trace; c020b843 <xfs_mount+63/70> >>>> >> Trace; c020b871 <xfs_vfsmount+21/40> >>>> >> Trace; c021cf48 <linvfs_read_super+188/270> >>>> >> Trace; c01294e0 <filemap_nopage+2c0/410> >>>> >> Trace; c0125f0e <handle_mm_fault+ce/e0> >>>> >> Trace; c0125d9d <do_no_page+4d/f0> >>>> >> Trace; c013cd72 <read_super+72/110> >>>> >> Trace; c013d01b <get_sb_bdev+18b/1e0> >>>> >> Trace; c013dafc <do_add_mount+1dc/290> >>>> >> Trace; c01131e0 <do_page_fault+0/4b0> >>>> >> Trace; c010724c <error_code+34/3c> >>>> >> Trace; c013dd56 <do_mount+106/120> >>>> >> Trace; c013dbfc <copy_mount_options+4c/a0> >>>> >> Trace; c013de13 <sys_mount+a3/130> >>>> >> Trace; c010715b <system_call+33/38> >>>> >> Code; e29c0266 <[linear]linear_make_request+36/f0> >>>> >> 00000000 <_EIP>: >>>> >> Code; e29c0266 <[linear]linear_make_request+36/f0> <===== >>>> >> 0: f7 f9 idiv %ecx,%eax <===== >>>> >> Code; e29c0268 <[linear]linear_make_request+38/f0> >>>> >> 2: 85 d2 test %edx,%edx >>>> >> Code; e29c026a <[linear]linear_make_request+3a/f0> >>>> >> 4: 74 24 je 2a <_EIP+0x2a> e29c0290 >>>> >> <[linear]linear_make_request+60/f0> >>>> >> Code; e29c026c <[linear]linear_make_request+3c/f0> >>>> >> 6: 55 push %ebp >>>> >> Code; e29c026d <[linear]linear_make_request+3d/f0> >>>> >> 7: 51 push %ecx >>>> >> Code; e29c026e <[linear]linear_make_request+3e/f0> >>>> >> 8: 68 c0 03 9c e2 push $0xe29c03c0 >>>> >> Code; e29c0273 <[linear]linear_make_request+43/f0> >>>> >> d: e8 58 6c 75 dd call dd756c6a >>>> <_EIP+0xdd756c6a> >>>> >> c0116ed0 <printk+0/1a0> >>>> >> Code; e29c0278 <[linear]linear_make_request+48/f0> >>>> >> 12: 6a 00 push $0x0 >>>> >> >>>> >> Andreas Dilger wrote: >>>> >> >>>> >> >On Aug 31, 2001 15:08 +0200, svetljo wrote: >>>> >> > >>>> >> >>[root@svetljo mnt]# mount -t xfs /dev/myData/Music music >>>> >> >>Segmentation fault >>>> >> >> >>>> >> > >>>> >> >Generally this is a bad sign. Either mount is segfaulting >>>> (unlikely) >>>> >> >or you are getting an oops in the kernel. You need do run >>>> something >>>> >> >like "dmesg | ksymoops" in order to get some useful data about >>>> where >>>> >> >the problem is (could be xfs, LVM, or elsewhere in the kernel). >>>> >> > >>>> >> >Once you have an oops, you are best off rebooting the system, >>>> because >>>> >> >your kernel memory may be corrupted, and cause more oopses >>>> which do >>>> >> >not mean anything. If you look in /var/log/messages (or >>>> >> /var/log/kern.log >>>> >> >or some other place, depending on where kernel messages go), >>>> you can >>>> >> >decode the FIRST oops in the log with ksymoops. All subsequent >>>> >> ones are >>>> >> >useless. >>>> >> > >>>> >> > >>>> >> >>the LV ( lvcreate -i3 -I4 -L26G -nMusic ) >>>> >> >> >>>> >> >>the VG -> myData /dev/hdh10 /dev/linVG1/linLV1 >>>> /dev/linVG2/linLV2 >>>> >> >> >>>> >> >>/dev/hdh10 normal partition 14G >>>> >> >>/dev/linVG1/linLV1 -> linear LV 14G /dev/hde6 /dev/hde12 >>>> >> >>/dev/linVg2/linLV2 -> linear LV 14G /dev/hdg1 /dev/hdg5 >>>> /dev/hdg6 >>>> >> /dev/hdg12 >>>> >> >> >>>> >> > >>>> >> >There is absolutely no point in doing this (not that it is >>>> possible >>>> >> to do >>>> >> >so anyways). First of all, striping is almost never needed "for >>>> >> performance" >>>> >> >unless you are normally doing very large sequential I/Os, and >>>> even >>>> >> so most >>>> >> >disks today have very good sequential I/O rates (e.g. 15-30MB/s). >>>> >> Secondly, >>>> >> >you _should_ be able to just create a single LV that is striped >>>> >> across all >>>> >> >of the PVs above. You would likely need to build it in steps, to >>>> >> ensure >>>> >> >that it is striped across the disks correctly. >>>> >> > >>>> >> >Cheers, Andreas >>>> >> > >>>> >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> > >>>> > >>>> > >>>> >>>> >>>> >>>> >>> >>> >>> >> >> >> > >