Re: 3.14.4 Kernel BUG at fs/namei.c : KDE's kate + ecryptfs + BTRFS + LVM2 + LUKS

2014-05-16 Thread Ritesh Raj Sarraf

On 05/16/2014 06:36 PM, Swâmi Petaramesh wrote:
> Underlying is ecryptfs over BTRFS over LVM2 over LUKS (what else ?)

I recently switched one of my drives to BTRFS because I wanted
transparent compression. Since I also needed encryption, I chose:

BTRFS => LUKS (Device Mapper) => SCSI Block.

Is there any value having LVM2 and ecryptfs in the way you have stacked
? Just curious.

-- 
Given the large number of mailing lists I follow, I request you to CC me
in replies for quicker response

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


3.14.4 Kernel BUG at fs/namei.c : KDE's kate + ecryptfs + BTRFS + LVM2 + LUKS

2014-05-16 Thread Swâmi Petaramesh
(I'm not subscribed to linux-kernel, pls copy me on the anwsers)

Hi there,

# uname -a
Linux zafu 3.14.4-1-ARCH #1 SMP PREEMPT Tue May 13 16:41:39 CEST 2014 x86_64 
GNU/Linux

I've seen the same thing happen several times in the last couple of months (so 
with said kernel version + at least 3.13)

Symptom: KDE's kate freezes while saving a small (arbitrary) text file.

(Yes, always KDE's kate and only KDE's kate, while saving a quite small text 
file... Besides this, machine's stability is rock-solid.)

Afterwards the machine remains more or less able to access its filesystems, 
however KDE's Dolphin file manager is usually dead as well.

Underlying is ecryptfs over BTRFS over LVM2 over LUKS (what else ?)

System logs :


kernel: [ cut here ]
kernel: kernel BUG at fs/namei.c:2404!
kernel: invalid opcode:  [#1] PREEMPT SMP 
kernel: Modules linked in: nls_iso8859_1 usb_storage ctr ccm md5 fuse ecb 
ecryptfs encrypted_keys sha1_ssse3 sha1_generic hmac trusted nls_utf8 
nls_cp437 vfat fat uvcvideo
kernel:  vboxdrv(O) sha256_ssse3 sha256_generic cbc dm_crypt dm_mod btrfs xor 
raid6_pq hid_multitouch hid_generic hid_logitech_dj usbhid hid sd_mod 
crc_t10dif atkbd libps2
kernel: CPU: 0 PID: 2142 Comm: kate Tainted: G   O 3.14.4-1-ARCH #1
kernel: Hardware name: ASUSTeK COMPUTER INC. X202EV/X202EV, BIOS X202EV.200 
02/27/2013
kernel: task: 8800a503bae0 ti: 880012dba000 task.ti: 880012dba000
kernel: RIP: 0010:[]  [] 
may_delete+0x138/0x150
kernel: RSP: 0018:880012dbbd78  EFLAGS: 00010283
kernel: RAX: 88006ae82cc0 RBX: 8800b0f23d80 RCX: 000d1616
kernel: RDX:  RSI: 8800b0f23d80 RDI: 8800ba841568
kernel: RBP: 880012dbbd98 R08: 7077732d6574616b R09: 8800b0f23c00
kernel: R10: 8080808080808080 R11: fefefefefefefeff R12: 8800a3836918
kernel: R13: 8800561501b8 R14: 8800b0f23d80 R15: 8800ba841568
kernel: FS:  7fe1e2648780() GS:88011ee0() 
knlGS:
kernel: CS:  0010 DS:  ES:  CR0: 80050033
kernel: CR2: 7f2875f7 CR3: 13ee5000 CR4: 001407f0
kernel: Stack:
kernel:  8800b0f23d80 8800ba841568 8800561501b8 8800b0f23d80
kernel:  880012dbbdd8 811ca6db  8800b0f23d80
kernel:  88006ae82cc0 880012f0e000 88008cc52000 8800ba841568
kernel: Call Trace:
kernel:  [] vfs_unlink+0x2b/0x160
kernel:  [] ecryptfs_do_unlink+0x65/0x120 [ecryptfs]
kernel:  [] ecryptfs_unlink+0x12/0x20 [ecryptfs]
kernel:  [] vfs_unlink+0xe6/0x160
kernel:  [] do_unlinkat+0x299/0x300
kernel:  [] ? fput+0xe/0x10
kernel:  [] ? task_work_run+0xb4/0xe0
kernel:  [] ? do_notify_resume+0x95/0xa0
kernel:  [] SyS_unlink+0x16/0x20
kernel:  [] system_call_fastpath+0x16/0x1b
kernel: Code: 1f 41 5d 83 e0 f0 41 5e 5d c3 66 0f 1f 84 00 00 00 00 00 b8 ff ff 
ff ff eb b1 66 0f 1f 84 00 00 00 00 00 0f 0b 66 0f 1f 44 00 00 <0f> 0b 66 0f 1f 
44 00 00 b8
kernel: RIP  [] may_delete+0x138/0x150
kernel:  RSP 
kernel: ---[ end trace 73914218811ba28d ]---

-- 
Swâmi Petaramesh  http://petaramesh.org PGP 9076E32E


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


AIX LVM support : block/partitions, kpartx, dmraid or lvm2 or ...

2013-05-11 Thread Philippe De Muyter
Hello,

I have recently written a read-only support for linux for AIX 3 & 4 lvm
disks, and submitted a patch for the linux block/partitions subsystem,
for which I have no feedback yet.

Now I wonder, as AIX LVM can describe mirrored and splitted logical
volumes, is block/partitions the right place for it ?  It seems that
only contiguous partitions can be described.  Am I mistaken ?
I looked also at kpartx, but it seems to me that kpartx has the same
limitations as block/partitions.

Should I rather make a patch for dmraid or lvm2 or some other tool I
am not aware of ?  What would be the best place for that support ?

I have working code that can generate tables for dmsetup, and I know
there is some demand for AIX LVM and AIX JFS support in linux, but where
should I send it ?

Thanks in advance

Philippe

PS: Of course, AIX LVM support does only make sense if AIX JFS file systems
are supported.  I have written read-only support for cwAIX JFSit also, and
I plan to submit it when I have a solution for AIX LVM, so that the AIX JFS
code can be tested by others.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Strange LVM2/DM data corruption with 2.6.11.12

2005-09-08 Thread Chris Wright
* Alexander Nyberg ([EMAIL PROTECTED]) wrote:
> Please upgrade to 2.6.12.6 (I don't remember exactly in which
> 2.6.12.x it went in), it contains a bugfix that should fix what
> you are seeing. 2.6.13 also has this.

Yep, that was 2.6.12.4, and here's the patch:

http://www.kernel.org/git/?p=linux/kernel/git/chrisw/stable-queue.git;a=blob_plain;h=6267a6b8da4b52eaf0fbddd9091a6e6ff2fe233c

thanks,
-chris
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Strange LVM2/DM data corruption with 2.6.11.12

2005-09-08 Thread Alexander Nyberg
On Thu, Sep 08, 2005 at 11:58:54AM +0200 Ludovic Drolez wrote:

> Hi !
> 
> We are developing (GPLed) disk cloning software similar to partimage: it's 
> an intelligent 'dd' which backups only used sectors.
> 
> Recently I added LVM1/2 support to it, and sometimes we saw LVM 
> restorations failing randomly (Disk images are not corrupted, but the 
> result of the restoration can be lead to a corrupted filesystem). If a 
> restoration fails, just try another one and it will work...
> 

Please upgrade to 2.6.12.6 (I don't remember exactly in which
2.6.12.x it went in), it contains a bugfix that should fix what
you are seeing. 2.6.13 also has this.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Strange LVM2/DM data corruption with 2.6.11.12

2005-09-08 Thread Ludovic Drolez

Hi !

We are developing (GPLed) disk cloning software similar to partimage: it's an 
intelligent 'dd' which backups only used sectors.


Recently I added LVM1/2 support to it, and sometimes we saw LVM restorations 
failing randomly (Disk images are not corrupted, but the result of the 
restoration can be lead to a corrupted filesystem). If a restoration fails, just 
try another one and it will work...


How the restoration program works:
- I restore the LVM2 administrative data (384 sectors, most of the time),
- I 'vgscan', 'vgchange',
- open for writing the '/dev/dm-',
- read a compressed file over NFS,
- and put the sectors in place, so it's a succession of '_llseek()' and 
'write()' to the DM device.


But, *sometimes*, for example, the current seek position is at 9GB, and some 
data is written to sector 0 ! It happens randomly.


Here is a typical strace of a restoration:

write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
_llseek(5, 20963328, [37604032512], SEEK_CUR) = 0
_llseek(5, 0, [37604032512], SEEK_CUR)  = 0
_llseek(5, 2097152, [37606129664], SEEK_CUR) = 0
write(5, "\1\335E\0\f\0\1\2.\0\0\0\2\0\0\0\364\17\2\2..\0\0\0\0\0"..., 512) = 51
2
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
write(5, "\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 512) = 512
_llseek(5, 88076288, [37694210048], SEEK_CUR) = 0
_llseek(5, 0, [37694210048], SEEK_CUR)  = 0
_llseek(5, 20971520, [37715181568], SEEK_CUR) = 0
write(5, "\377\377\377\377\377\377\377\377\377\377\377\377\377\377"..., 512) = 5
12



As you can see, there are no seeks to sector 0, but something randomly write 
some data to sector 0 !

I could reproduce these random problems on different kind of PCs.

But, the strace above comes from an improved version, which aggregates 
'_llseek's. A previous version, which did *many* 512 bytes seeks had much more 
problems. Aggregating seeks made the corruption to appears very rarely... And I 
more likely to happen, for a 40GB restoration than for a 10GB one.


So less system calls to the DM/LVM2 layer seems to give less corruption 
probability.


Any ideas ? Newer kernel releases could have fixed such a problem ?


--
Ludovic DROLEZ  Linbox / Free&ALter Soft
http://lrs.linbox.org - Linbox Rescue Server GPL edition
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: File corruption on LVM2 on top of software RAID1

2005-08-05 Thread Andrew Morton
"Simon Matter" <[EMAIL PROTECTED]> wrote:
>
> While looking at some data corruption vulnerability reports on
>  Securityfocus I wonder why this issue does not get any attention from
>  distributors. I have an open bugzilla report with RedHat, have an open
>  customer service request with RedHat, have mailed peoples directly. No
>  real feedback.
>  I'm now in the process of restoring intergrity of my data with the help of
>  backups and mirrored data. Maybe I just care too much about other peoples
>  data, but I know that this bug will corrupt files on hundreds or thousands
>  of servers today and most people simply don't know it. Did I miss
>  something?

I guess the bug hit really rarely and maybe the reports were lost in the
permanent background noise of dodgy hardware.

We only found and fixed it last week, due to much sleuthing by Matthew
Stapleton.  I assume vendor updates are in the pipeline.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: File corruption on LVM2 on top of software RAID1

2005-08-05 Thread Simon Matter
> "Simon Matter" <[EMAIL PROTECTED]> wrote:
>>
>> Hi,
>>
>> Please CC me as I'm not subscribed to the kernel list.
>>
>> I had a hard time identifying a serious problem in the 2.6 linux kernel.
>
> Yes, it is a serious problem.
>
>> It all started while evaluating RHEL4 for new servers. My data integrity
>> tests gave me bad results - which I couldn't believe - and my first idea
>> was - of course - bad hardware. I ordered new SCSI disks instead of the
>> IDE disks, took another server, spent some money again, tried again and
>> again. That's all long ago now...
>>
>> In my tests I get corrupt files on LVM2 which is on top of software
>> raid1.
>> (This is a common setup even mentioned in the software RAID HOWTO and
>> has
>> worked for me on RedHat 9 / kernel 2.4 for a long time now and it's my
>> favourite configuration). Now, I tested two different distributions,
>> three
>> kernels, three different filesystems and three different hardware. I can
>> always reproduce it with the following easy scripts:
>>
>
> Thanks for doing that.
>
> There's one fix against 2.6.12.3 which is needed, but 2.6.9 didn't have
> the
> bug which this fix addresses.  So unless 2.6.9 has a different problem,
> this won't help.

Sorry for the confusion, the 2.6.9 in question is a "RedHat Enterprise
Kernel" which includes lots of backported patches and features. So it is
no plain 2.6.9.

The real problem is that most current distributions include this bug in
one or the other way. They either have one of the affected versions in the
kernel package or have an older kernel with the bug backported. What I
know for sure is that the following distributions are affected:
RedHat Enterprise 4
Fedora Core 4
SuSE 9.3
Debian with the latest "unstable" 2.6 kernel
I couldn't verify Mandrake, Gentoo and others, but I expect them to be at
risk too.

While looking at some data corruption vulnerability reports on
Securityfocus I wonder why this issue does not get any attention from
distributors. I have an open bugzilla report with RedHat, have an open
customer service request with RedHat, have mailed peoples directly. No
real feedback.
I'm now in the process of restoring intergrity of my data with the help of
backups and mirrored data. Maybe I just care too much about other peoples
data, but I know that this bug will corrupt files on hundreds or thousands
of servers today and most people simply don't know it. Did I miss
something?

Simon

>
> But still, you should include this in testing please.
>
>
> diff -puN fs/bio.c~bio_clone-fix fs/bio.c
> --- devel/fs/bio.c~bio_clone-fix  2005-07-28 00:39:40.0 -0700
> +++ devel-akpm/fs/bio.c   2005-07-28 01:02:34.0 -0700
> @@ -261,6 +261,7 @@ inline void __bio_clone(struct bio *bio,
>*/
>   bio->bi_vcnt = bio_src->bi_vcnt;
>   bio->bi_size = bio_src->bi_size;
> + bio->bi_idx = bio_src->bi_idx;
>   bio_phys_segments(q, bio);
>   bio_hw_segments(q, bio);
>  }
> _
>
>

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: File corruption on LVM2 on top of software RAID1

2005-08-04 Thread Andrew Morton
Andrew Morton <[EMAIL PROTECTED]> wrote:
>
> There's one fix against 2.6.12.3 which is needed, but 2.6.9 didn't have the
>  bug which this fix addresses.

aargh, I see that it did fix it.

Don't blame me.  Blame people who screw up list threading by reading a
mail->news gateway and hitting "reply".

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: File corruption on LVM2 on top of software RAID1

2005-08-04 Thread Andrew Morton
"Simon Matter" <[EMAIL PROTECTED]> wrote:
>
> Hi,
> 
> Please CC me as I'm not subscribed to the kernel list.
> 
> I had a hard time identifying a serious problem in the 2.6 linux kernel.

Yes, it is a serious problem.

> It all started while evaluating RHEL4 for new servers. My data integrity
> tests gave me bad results - which I couldn't believe - and my first idea
> was - of course - bad hardware. I ordered new SCSI disks instead of the
> IDE disks, took another server, spent some money again, tried again and
> again. That's all long ago now...
> 
> In my tests I get corrupt files on LVM2 which is on top of software raid1.
> (This is a common setup even mentioned in the software RAID HOWTO and has
> worked for me on RedHat 9 / kernel 2.4 for a long time now and it's my
> favourite configuration). Now, I tested two different distributions, three
> kernels, three different filesystems and three different hardware. I can
> always reproduce it with the following easy scripts:
> 
> LOGF=/root/diff.log
> while true; do
>   rm -rf /home/XXX2
>   rsync -a /home/XXX/ /home/XXX2
>   date >> $LOGF
>   diff -r /home/XXX /home/XXX2 >> $LOGF
> done
> 
> the files in /home/XXX are ~15G of ISO images and rpms.
> 
> diff.log looks like this:
> Wed Aug  3 13:45:57 CEST 2005
> Binary files /home/XXX/ES3-U3/rhel-3-U3-i386-es-disc3.iso and
> /home/XXX2/ES3-U3/rhel-3-U3-i386-es-disc3.iso differ
> Wed Aug  3 14:09:14 CEST 2005
> Binary files /home/XXX/8.0/psyche-i386-disc1.iso and
> /home/XXX2/8.0/psyche-i386-disc1.iso differ
> Wed Aug  3 14:44:17 CEST 2005
> Binary files /home/XXX/7.3/valhalla-i386-disc3.iso and
> /home/XXX2/7.3/valhalla-i386-disc3.iso differ
> Wed Aug  3 15:15:05 CEST 2005
> Wed Aug  3 15:45:40 CEST 2005
> 
> 
> Tested software:
> 1) RedHat EL4
> kernel-2.6.9-11.EL
> vanilla 2.6.12.3 kernel
> filesystems: EXT2, EXT3, XFS
> 
> 2) NOVELL/SUSE 9.3
> kernel-default-2.6.11.4-21.7
> filesystem: EXT3
> 
> Tested Hardware:
> 1)
> - ASUS P2B-S board
> - CPU PIII 450MHz
> - Intel 440BX/ZX/DX Chipset
> - 4x128M memory (ECC enabled)
> - 2x IDE disks Seagate Barracuda 400G, connected to onboard "Intel PIIX4
> Ultra 33"
> - Promise Ultra100TX2 adapter for additional tests
> 
> 2)
> - DELL PowerEdge 1400
> - CPU PIII 800MHz
> - ServerWorks OSB4 Chipset
> - 4x256M memory (ECC enabled)
> - 2x U320 SCSI disks Maxtor Atlas 10K 146G
> - onboard Adaptec aic7899 Ultra160 SCSI adapter
> 
> 3)
> - DELL PowerEdge 1850
> - CPU P4 XEON 2.8GHz
> - 2G memory
> - 2x U320 SCSI disks Maxtor Atlas 10K 73G SCA
> - onboard LSI53C1030 SCSI adapter
> 
> I've put some files toghether from the last test on the PE1850 server:
> http://www.invoca.ch/bugs/linux-2.6-corruption-on-lvm2-on-raid1/
> 
> I've also filed a bug with RedHat:
> https://bugzilla.redhat.com/bugzilla/show_bug.cgi?id=164696
> 
> I have spent a lot of time on this bug because I consider it very serious.
> I'm not a kernel hacker but if there is anything I can do to fix this, let
> me know.
> 

Thanks for doing that.

There's one fix against 2.6.12.3 which is needed, but 2.6.9 didn't have the
bug which this fix addresses.  So unless 2.6.9 has a different problem,
this won't help.

But still, you should include this in testing please.



diff -puN fs/bio.c~bio_clone-fix fs/bio.c
--- devel/fs/bio.c~bio_clone-fix2005-07-28 00:39:40.0 -0700
+++ devel-akpm/fs/bio.c 2005-07-28 01:02:34.0 -0700
@@ -261,6 +261,7 @@ inline void __bio_clone(struct bio *bio,
 */
bio->bi_vcnt = bio_src->bi_vcnt;
bio->bi_size = bio_src->bi_size;
+   bio->bi_idx = bio_src->bi_idx;
bio_phys_segments(q, bio);
bio_hw_segments(q, bio);
 }
_

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: File corruption on LVM2 on top of software RAID1

2005-08-04 Thread Simon Matter
> Once upon a time, "Simon Matter" <[EMAIL PROTECTED]> said:
>>In my tests I get corrupt files on LVM2 which is on top of software
>> raid1.
>>(This is a common setup even mentioned in the software RAID HOWTO and has
>>worked for me on RedHat 9 / kernel 2.4 for a long time now and it's my
>>favourite configuration). Now, I tested two different distributions,
>> three
>>kernels, three different filesystems and three different hardware. I can
>>always reproduce it with the following easy scripts:
>
> See:
>
> http://bugzilla.kernel.org/show_bug.cgi?id=4946
> https://bugzilla.redhat.com/bugzilla/show_bug.cgi?id=152162
>
> There's a one-line patch in there; see if that fixes the problem for
> you.

Hi Chris,

Thank you for your help, indeed this oneliner fixes the corruption. I have
tested with 2.6.12.3 + bio_clone fix and also built an updated RedHat EL4
kernel with the fix included. No corruption anymore.
This fix should be applied to most kernels shipped with the latest Linux
distributions. At least those products called Enterprise something should
hurry in the interest of their customers. Do you think this will happen
anytime soon?

-- 
Simon Matter
Invoca Systems

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: File corruption on LVM2 on top of software RAID1

2005-08-03 Thread Chris Adams
Once upon a time, "Simon Matter" <[EMAIL PROTECTED]> said:
>In my tests I get corrupt files on LVM2 which is on top of software raid1.
>(This is a common setup even mentioned in the software RAID HOWTO and has
>worked for me on RedHat 9 / kernel 2.4 for a long time now and it's my
>favourite configuration). Now, I tested two different distributions, three
>kernels, three different filesystems and three different hardware. I can
>always reproduce it with the following easy scripts:

See:

http://bugzilla.kernel.org/show_bug.cgi?id=4946
https://bugzilla.redhat.com/bugzilla/show_bug.cgi?id=152162

There's a one-line patch in there; see if that fixes the problem for
you.

-- 
Chris Adams <[EMAIL PROTECTED]>
Systems and Network Administrator - HiWAAY Internet Services
I don't speak for anybody but myself - that's enough trouble.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


File corruption on LVM2 on top of software RAID1

2005-08-03 Thread Simon Matter
Hi,

Please CC me as I'm not subscribed to the kernel list.

I had a hard time identifying a serious problem in the 2.6 linux kernel.
It all started while evaluating RHEL4 for new servers. My data integrity
tests gave me bad results - which I couldn't believe - and my first idea
was - of course - bad hardware. I ordered new SCSI disks instead of the
IDE disks, took another server, spent some money again, tried again and
again. That's all long ago now...

In my tests I get corrupt files on LVM2 which is on top of software raid1.
(This is a common setup even mentioned in the software RAID HOWTO and has
worked for me on RedHat 9 / kernel 2.4 for a long time now and it's my
favourite configuration). Now, I tested two different distributions, three
kernels, three different filesystems and three different hardware. I can
always reproduce it with the following easy scripts:

LOGF=/root/diff.log
while true; do
  rm -rf /home/XXX2
  rsync -a /home/XXX/ /home/XXX2
  date >> $LOGF
  diff -r /home/XXX /home/XXX2 >> $LOGF
done

the files in /home/XXX are ~15G of ISO images and rpms.

diff.log looks like this:
Wed Aug  3 13:45:57 CEST 2005
Binary files /home/XXX/ES3-U3/rhel-3-U3-i386-es-disc3.iso and
/home/XXX2/ES3-U3/rhel-3-U3-i386-es-disc3.iso differ
Wed Aug  3 14:09:14 CEST 2005
Binary files /home/XXX/8.0/psyche-i386-disc1.iso and
/home/XXX2/8.0/psyche-i386-disc1.iso differ
Wed Aug  3 14:44:17 CEST 2005
Binary files /home/XXX/7.3/valhalla-i386-disc3.iso and
/home/XXX2/7.3/valhalla-i386-disc3.iso differ
Wed Aug  3 15:15:05 CEST 2005
Wed Aug  3 15:45:40 CEST 2005


Tested software:
1) RedHat EL4
kernel-2.6.9-11.EL
vanilla 2.6.12.3 kernel
filesystems: EXT2, EXT3, XFS

2) NOVELL/SUSE 9.3
kernel-default-2.6.11.4-21.7
filesystem: EXT3

Tested Hardware:
1)
- ASUS P2B-S board
- CPU PIII 450MHz
- Intel 440BX/ZX/DX Chipset
- 4x128M memory (ECC enabled)
- 2x IDE disks Seagate Barracuda 400G, connected to onboard "Intel PIIX4
Ultra 33"
- Promise Ultra100TX2 adapter for additional tests

2)
- DELL PowerEdge 1400
- CPU PIII 800MHz
- ServerWorks OSB4 Chipset
- 4x256M memory (ECC enabled)
- 2x U320 SCSI disks Maxtor Atlas 10K 146G
- onboard Adaptec aic7899 Ultra160 SCSI adapter

3)
- DELL PowerEdge 1850
- CPU P4 XEON 2.8GHz
- 2G memory
- 2x U320 SCSI disks Maxtor Atlas 10K 73G SCA
- onboard LSI53C1030 SCSI adapter

I've put some files toghether from the last test on the PE1850 server:
http://www.invoca.ch/bugs/linux-2.6-corruption-on-lvm2-on-raid1/

I've also filed a bug with RedHat:
https://bugzilla.redhat.com/bugzilla/show_bug.cgi?id=164696

I have spent a lot of time on this bug because I consider it very serious.
I'm not a kernel hacker but if there is anything I can do to fix this, let
me know.

Simon

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


[patch 016/198] oom-killer disable for iscsi/lvm2/multipath userland critical sections

2005-04-12 Thread akpm

From: Andrea Arcangeli <[EMAIL PROTECTED]>

iscsi/lvm2/multipath needs guaranteed protection from the oom-killer, so
make the magical value of -17 in /proc//oom_adj defeat the oom-killer
altogether.

(akpm: we still need to document oom_adj and friends in
Documentation/filesystems/proc.txt!)

Signed-off-by: Andrea Arcangeli <[EMAIL PROTECTED]>
Signed-off-by: Andrew Morton <[EMAIL PROTECTED]>
---

 25-akpm/fs/proc/base.c |2 +-
 25-akpm/include/linux/mm.h |3 +++
 25-akpm/mm/oom_kill.c  |2 +-
 3 files changed, 5 insertions(+), 2 deletions(-)

diff -puN 
fs/proc/base.c~oom-killer-disable-for-iscsi-lvm2-multipath-userland-critical-sections
 fs/proc/base.c
--- 
25/fs/proc/base.c~oom-killer-disable-for-iscsi-lvm2-multipath-userland-critical-sections
2005-04-12 03:21:07.242035944 -0700
+++ 25-akpm/fs/proc/base.c  2005-04-12 03:21:07.249034880 -0700
@@ -751,7 +751,7 @@ static ssize_t oom_adjust_write(struct f
if (copy_from_user(buffer, buf, count))
return -EFAULT;
oom_adjust = simple_strtol(buffer, &end, 0);
-   if (oom_adjust < -16 || oom_adjust > 15)
+   if ((oom_adjust < -16 || oom_adjust > 15) && oom_adjust != OOM_DISABLE)
return -EINVAL;
if (*end == '\n')
end++;
diff -puN 
include/linux/mm.h~oom-killer-disable-for-iscsi-lvm2-multipath-userland-critical-sections
 include/linux/mm.h
--- 
25/include/linux/mm.h~oom-killer-disable-for-iscsi-lvm2-multipath-userland-critical-sections
2005-04-12 03:21:07.243035792 -0700
+++ 25-akpm/include/linux/mm.h  2005-04-12 03:21:07.250034728 -0700
@@ -857,5 +857,8 @@ int in_gate_area_no_task(unsigned long a
 #define in_gate_area(task, addr) ({(void)task; in_gate_area_no_task(addr);})
 #endif /* __HAVE_ARCH_GATE_AREA */
 
+/* /proc//oom_adj set to -17 protects from the oom-killer */
+#define OOM_DISABLE -17
+
 #endif /* __KERNEL__ */
 #endif /* _LINUX_MM_H */
diff -puN 
mm/oom_kill.c~oom-killer-disable-for-iscsi-lvm2-multipath-userland-critical-sections
 mm/oom_kill.c
--- 
25/mm/oom_kill.c~oom-killer-disable-for-iscsi-lvm2-multipath-userland-critical-sections
 2005-04-12 03:21:07.245035488 -0700
+++ 25-akpm/mm/oom_kill.c   2005-04-12 03:21:07.251034576 -0700
@@ -145,7 +145,7 @@ static struct task_struct * select_bad_p
do_posix_clock_monotonic_gettime(&uptime);
do_each_thread(g, p)
/* skip the init task with pid == 1 */
-   if (p->pid > 1) {
+   if (p->pid > 1 && p->oomkilladj != OOM_DISABLE) {
unsigned long points;
 
/*
_
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: oom-killer disable for iscsi/lvm2/multipath userland critical sections

2005-04-01 Thread Dmitry Yusupov
Andrea,

I just successfully tested the patch on my environment. It actually
resolved OOM-killer problem for my iscsid.

Important note: daemon's parent must be init.

In my test, OOM-killer killed everything around but iscsid, and iscsid
successfully finished registration of new SCSI host in the middle of
crazy OOM-killer :)

Thanks!
Dima

On Sat, 2005-04-02 at 00:14 +0200, Andrea Arcangeli wrote:
> Hello,
> 
> some private discussion (that was continuing some kernel-summit-discuss
> thread) ended in the below patch. I also liked a textual "disable"
> instead of value "-17" (internally to the kernel it could be represented
> the same way, but the /proc parsing would be more complicated). If you
> prefer textual "disable" we can change this of course.
> 
> Comments welcome.
> 
> From: Andrea Arcangeli <[EMAIL PROTECTED]>
> Subject: oom killer protection
> 
> iscsi/lvm2/multipath needs guaranteed protection from the oom-killer.
> 
> Signed-off-by: Andrea Arcangeli <[EMAIL PROTECTED]>
> 
> --- 2.6.12-seccomp/fs/proc/base.c.~1~ 2005-03-25 05:13:28.0 +0100
> +++ 2.6.12-seccomp/fs/proc/base.c 2005-04-01 23:47:22.0 +0200
> @@ -751,7 +751,7 @@ static ssize_t oom_adjust_write(struct f
>   if (copy_from_user(buffer, buf, count))
>   return -EFAULT;
>   oom_adjust = simple_strtol(buffer, &end, 0);
> - if (oom_adjust < -16 || oom_adjust > 15)
> + if ((oom_adjust < -16 || oom_adjust > 15) && oom_adjust != OOM_DISABLE)
>   return -EINVAL;
>   if (*end == '\n')
>   end++;
> --- 2.6.12-seccomp/include/linux/mm.h.~1~ 2005-03-25 05:13:28.0 
> +0100
> +++ 2.6.12-seccomp/include/linux/mm.h 2005-04-01 23:53:11.0 +0200
> @@ -856,5 +856,8 @@ int in_gate_area_no_task(unsigned long a
>  #define in_gate_area(task, addr) ({(void)task; in_gate_area_no_task(addr);})
>  #endif   /* __HAVE_ARCH_GATE_AREA */
>  
> +/* /proc//oom_adj set to -17 protects from the oom-killer */
> +#define OOM_DISABLE -17
> +
>  #endif /* __KERNEL__ */
>  #endif /* _LINUX_MM_H */
> --- 2.6.12-seccomp/mm/oom_kill.c.~1~  2005-03-08 01:02:30.0 +0100
> +++ 2.6.12-seccomp/mm/oom_kill.c  2005-04-01 23:46:18.0 +0200
> @@ -145,7 +145,7 @@ static struct task_struct * select_bad_p
>   do_posix_clock_monotonic_gettime(&uptime);
>   do_each_thread(g, p)
>   /* skip the init task with pid == 1 */
> - if (p->pid > 1) {
> + if (p->pid > 1 && p->oomkilladj != OOM_DISABLE) {
>   unsigned long points;
>  
>   /*
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


oom-killer disable for iscsi/lvm2/multipath userland critical sections

2005-04-01 Thread Andrea Arcangeli
Hello,

some private discussion (that was continuing some kernel-summit-discuss
thread) ended in the below patch. I also liked a textual "disable"
instead of value "-17" (internally to the kernel it could be represented
the same way, but the /proc parsing would be more complicated). If you
prefer textual "disable" we can change this of course.

Comments welcome.

From: Andrea Arcangeli <[EMAIL PROTECTED]>
Subject: oom killer protection

iscsi/lvm2/multipath needs guaranteed protection from the oom-killer.

Signed-off-by: Andrea Arcangeli <[EMAIL PROTECTED]>

--- 2.6.12-seccomp/fs/proc/base.c.~1~   2005-03-25 05:13:28.0 +0100
+++ 2.6.12-seccomp/fs/proc/base.c   2005-04-01 23:47:22.0 +0200
@@ -751,7 +751,7 @@ static ssize_t oom_adjust_write(struct f
if (copy_from_user(buffer, buf, count))
return -EFAULT;
oom_adjust = simple_strtol(buffer, &end, 0);
-   if (oom_adjust < -16 || oom_adjust > 15)
+   if ((oom_adjust < -16 || oom_adjust > 15) && oom_adjust != OOM_DISABLE)
return -EINVAL;
if (*end == '\n')
end++;
--- 2.6.12-seccomp/include/linux/mm.h.~1~   2005-03-25 05:13:28.0 
+0100
+++ 2.6.12-seccomp/include/linux/mm.h   2005-04-01 23:53:11.0 +0200
@@ -856,5 +856,8 @@ int in_gate_area_no_task(unsigned long a
 #define in_gate_area(task, addr) ({(void)task; in_gate_area_no_task(addr);})
 #endif /* __HAVE_ARCH_GATE_AREA */
 
+/* /proc//oom_adj set to -17 protects from the oom-killer */
+#define OOM_DISABLE -17
+
 #endif /* __KERNEL__ */
 #endif /* _LINUX_MM_H */
--- 2.6.12-seccomp/mm/oom_kill.c.~1~2005-03-08 01:02:30.0 +0100
+++ 2.6.12-seccomp/mm/oom_kill.c2005-04-01 23:46:18.0 +0200
@@ -145,7 +145,7 @@ static struct task_struct * select_bad_p
do_posix_clock_monotonic_gettime(&uptime);
do_each_thread(g, p)
/* skip the init task with pid == 1 */
-   if (p->pid > 1) {
+   if (p->pid > 1 && p->oomkilladj != OOM_DISABLE) {
unsigned long points;
 
/*
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: md and RAID 5 [was Re: LVM2]

2005-01-23 Thread Neil Brown
On Friday January 21, [EMAIL PROTECTED] wrote:
> Thank you all for having been so kind in your responses and help.
> 
> However, there is one more set of questions I have.
> 
> Does the md (software raid) have disk size or raid volume limits?

2^31 sectors for individual disks.  Arrays do not have this limit.

> 
> If I am using such things as USB or 1394 disks, is there a way to use
> labels in /etc/raidtab and with the tools so that when the disks, if
> they do, get renumbered in /dev that all works fine. I am aware that the
> kernel will autodetect these devices, but that the raidtab needs to be
> consistent. This is what I am trying to figure out how to do.

Scrap raidtools and /etc/raidtab.  Explore "mdadm" and /etc/mdadm.conf

NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-23 Thread Kyle Moffett
On Jan 20, 2005, at 16:40, Norbert van Nobelen wrote:
RAID5 in software works pretty good (survived a failed disk, and 
recovered
another failing raid in 1 month). Hardware is better since you don't 
have a
boot partition left which is usually just present on one disk (you can 
mirror
that yourself ofcourse).
Err, you _can_ boot completely from a software RAID, it just takes a 
bit more
work.  I have an old PowerMac G4 400MHz with a Promise 20268 controller 
and 3
80GB drives booting from a software RAID.  You just set up a 250-500MB 
boot
partition mirrored with RAID 1 across all drives, then set up a RAID 5 
swap
partition and a RAID 5 LVM partition on each drive.  Once LVM is 
configured
with each remaining filesystem, install your distro (The new 
Debian-installer
does very well) and set up Yaboot/GRUB/whatever to install a boot 
sector on
each drive.  Then set up a RAID+LVM initrd (Debian does this mostly
automatically too), and reboot. This computer boots a custom 2.6.8.1 
kernel,
has 896MB RAM, and a 400MHz CPU, but it reads 41.5MiByte/sec from its 
RAID 5
partition with a 1MiByte blocksize, and has 16.8MiByte/sec over LVM over
RAID 5 with the same blocksize.  I've been following the discussions on 
2.6
instability and "New development model" problems, but AFAICT, 2.6 has 
been
rock stable on this box, which acts as an IPv4/IPv6 
router/firewall/server.

Cheers,
Kyle Moffett
-BEGIN GEEK CODE BLOCK-
Version: 3.12
GCM/CS/IT/U d- s++: a18 C>$ UB/L/X/*(+)>$ P+++()>$
L(+++) E W++(+) N+++(++) o? K? w--- O? M++ V? PS+() PE+(-) Y+
PGP+++ t+(+++) 5 X R? tv-(--) b(++) DI+ D+ G e->$ h!*()>++$ r  
!y?(-)
--END GEEK CODE BLOCK--

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: md and RAID 5 [was Re: LVM2]

2005-01-21 Thread Wakko Warner
Trever L. Adams wrote:
> Thank you all for having been so kind in your responses and help.
> 
> However, there is one more set of questions I have.
> 
> Does the md (software raid) have disk size or raid volume limits?
> 
> If I am using such things as USB or 1394 disks, is there a way to use
> labels in /etc/raidtab and with the tools so that when the disks, if
> they do, get renumbered in /dev that all works fine. I am aware that the
> kernel will autodetect these devices, but that the raidtab needs to be
> consistent. This is what I am trying to figure out how to do.

EVMS can handle this for you.  I've used it with a raid set I made with some
of the drives being on usb and some on ide.

If you use evms, you might want to consider the activation after all disks
have been discovered.

-- 
 Lab tests show that use of micro$oft causes cancer in lab animals
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


md and RAID 5 [was Re: LVM2]

2005-01-21 Thread Trever L. Adams
Thank you all for having been so kind in your responses and help.

However, there is one more set of questions I have.

Does the md (software raid) have disk size or raid volume limits?

If I am using such things as USB or 1394 disks, is there a way to use
labels in /etc/raidtab and with the tools so that when the disks, if
they do, get renumbered in /dev that all works fine. I am aware that the
kernel will autodetect these devices, but that the raidtab needs to be
consistent. This is what I am trying to figure out how to do.

Thank you,
Trever Adams
--
"A modest woman, dressed out in all her finery, is the most tremendous
object in the whole creation." -- Goldsmith 

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-21 Thread Norbert van Nobelen
With using RAID5 you can choose yourself howmany hot standby/failover disks 
you want to use. The number (or ratio) of disks used for failover of your 
raid will determine the chance that you have when one disk fails and complete 
failure of a raid.
It is still pretty safe just to have 7 active disks and 1 hot standby disk, 
thus getting you 3500 Manufacturers GB (3.34TB)

About LVM on a RAID: 
With hardware RAID it will work for sure.
With software RAID:
Make the RAID partitions and build your RAID.
On that RAID, set the partition type to LVM
And it should work
On LVM you can then use ext3 (your preferred fs)

On Thursday 20 January 2005 23:17, you wrote:
> It is for a group. For the most part it is data access/retention. Writes
> and such would be more similar to a desktop. I would use SATA if they
> were (nearly) equally priced and there were awesome 1394 to SATA bridge
> chips that worked well with Linux. So, right now, I am looking at ATA to
> 1394.
>
> So, to get 2TB of RAID5 you have 6 500 GB disks right? So, will this
> work within on LV? Or is it 2TB of diskspace total? So, are volume
> groups pretty fault tolerant if you have a bunch of RAID5 LVs below
> them? This is my one worry about this.
>
> Second, you mentioned file systems. We were talking about ext3. I have
> never used any others in Linux (barring ext2, minixfs, and fat). I had
> heard XFS from IBM was pretty good. I would rather not use reiserfs.
>
> Any recommendations.
>
> Trever
>
> P.S. Why won't an LV support over 2TB?
>
> S.P.S. I am not really worried about the boot and programs drive. They
> will be spun down most of the time I am sure.
>
> On Thu, 2005-01-20 at 22:40 +0100, Norbert van Nobelen wrote:
> > A logical volume in LVM will not handle more than 2TB. You can tie
> > together the LVs in a volume group, thus going over the 2TB limit. Choose
> > your filesystem well though, some have a 2TB limit too.
> >
> > Disk size: What are you doing with it. 500GB disks are ATA (maybe SATA).
> > ATA is good for low end servers or near line storage, SATA can be used
> > equally to SCSI (I am going to suffer for this remark).
> >
> > RAID5 in software works pretty good (survived a failed disk, and
> > recovered another failing raid in 1 month). Hardware is better since you
> > don't have a boot partition left which is usually just present on one
> > disk (you can mirror that yourself ofcourse).
> >
> > Regards,
> >
> > Norbert van Nobelen
> >
> > On Thursday 20 January 2005 20:51, you wrote:
> > > I recently saw Alan Cox say on this list that LVM won't handle more
> > > than 2 terabytes. Is this LVM2 or LVM? What is the maximum amount of
> > > disk space LVM2 (or any other RAID/MIRROR capable technology that is in
> > > Linus's kernel) handle? I am talking with various people and we are
> > > looking at Samba on Linux to do several different namespaces (obviously
> > > one tree), most averaging about 3 terabytes, but one would have in
> > > excess of 20 terabytes. We are looking at using 320 to 500 gigabyte
> > > drives in these arrays. (How? IEEE-1394. Which brings a question I will
> > > ask in a second email.)
> > >
> > > Is RAID 5 all that bad using this software method? Is RAID 5 available?
> > >
> > > Trever Adams
> > > --
> > > "They that can give up essential liberty to obtain a little temporary
> > > safety deserve neither liberty nor safety." -- Benjamin Franklin, 1759
> > >
> > > -
> > > To unsubscribe from this list: send the line "unsubscribe linux-kernel"
> > > in the body of a message to [EMAIL PROTECTED]
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > Please read the FAQ at  http://www.tux.org/lkml/
> >
> > -
> > To unsubscribe from this list: send the line "unsubscribe linux-kernel"
> > in the body of a message to [EMAIL PROTECTED]
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > Please read the FAQ at  http://www.tux.org/lkml/
>
> --
> "Assassination is the extreme form of censorship." -- George Bernard
> Shaw (1856-1950)

-- 
http://www.edusupport.nl";>EduSupport: Linux Desktop for schools and 
small to medium business in The Netherlands and Belgium
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-21 Thread Norbert van Nobelen
Even as LVM user, guess what I used before answering (-:
On Thursday 20 January 2005 23:34, Alasdair G Kergon wrote:
> On Thu, Jan 20, 2005 at 03:22:14PM -0700, Trever L. Adams wrote:
> > PV = the device
> > VG = groups of them (the RAID5 array?)
> > LV = what? the file system?
>
> http://www.tldp.org/HOWTO/LVM-HOWTO/anatomy.html
> http://www.novell.com/products/linuxenterpriseserver8/whitepapers/LVM.pdf
>   [Out-of-date now, but descriptions of concepts still useful.]
>
> LVM mailing list for LVM questions:
>   https://www.redhat.com/mailman/listinfo/linux-lvm
>
> Alasdair
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Steve Lord
Trever L. Adams wrote:
It is for a group. For the most part it is data access/retention. Writes
and such would be more similar to a desktop. I would use SATA if they
were (nearly) equally priced and there were awesome 1394 to SATA bridge
chips that worked well with Linux. So, right now, I am looking at ATA to
1394.
So, to get 2TB of RAID5 you have 6 500 GB disks right? So, will this
work within on LV? Or is it 2TB of diskspace total? So, are volume
groups pretty fault tolerant if you have a bunch of RAID5 LVs below
them? This is my one worry about this.
Second, you mentioned file systems. We were talking about ext3. I have
never used any others in Linux (barring ext2, minixfs, and fat). I had
heard XFS from IBM was pretty good. I would rather not use reiserfs.
Any recommendations.
Trever
They all forgot to mention one more limitation, the maximum filesystem
size supported by the address_space structure in linux. If you are running
on ia32, then you get stuck with 2^32 filesystem blocks, or 16 Tbytes in
one filesystem because of the way an address space structure is used to
cache the metadata. If you use an Athlon 64 that limitation goes away.
Steve
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Alasdair G Kergon
On Thu, Jan 20, 2005 at 03:22:14PM -0700, Trever L. Adams wrote:
> PV = the device
> VG = groups of them (the RAID5 array?)
> LV = what? the file system?
 
http://www.tldp.org/HOWTO/LVM-HOWTO/anatomy.html
http://www.novell.com/products/linuxenterpriseserver8/whitepapers/LVM.pdf
  [Out-of-date now, but descriptions of concepts still useful.]

LVM mailing list for LVM questions:
  https://www.redhat.com/mailman/listinfo/linux-lvm

Alasdair
-- 
[EMAIL PROTECTED]
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Jeffrey E. Hundstad
XFS is an SGI project.
http://oss.sgi.com/
I've been using it for quite a while and am quite happy with it; it is 
very fast and very fault tolerant.  The only warning I'd like to give 
about it is it seems that some Linux developers seem to have a bad taste 
in their mouth when it comes to XFS; go figure.

--
jeffrey hundstad
Trever L. Adams wrote:
It is for a group. For the most part it is data access/retention. Writes
and such would be more similar to a desktop. I would use SATA if they
were (nearly) equally priced and there were awesome 1394 to SATA bridge
chips that worked well with Linux. So, right now, I am looking at ATA to
1394.
So, to get 2TB of RAID5 you have 6 500 GB disks right? So, will this
work within on LV? Or is it 2TB of diskspace total? So, are volume
groups pretty fault tolerant if you have a bunch of RAID5 LVs below
them? This is my one worry about this.
Second, you mentioned file systems. We were talking about ext3. I have
never used any others in Linux (barring ext2, minixfs, and fat). I had
heard XFS from IBM was pretty good. I would rather not use reiserfs.
Any recommendations.
Trever
P.S. Why won't an LV support over 2TB?
S.P.S. I am not really worried about the boot and programs drive. They
will be spun down most of the time I am sure.
On Thu, 2005-01-20 at 22:40 +0100, Norbert van Nobelen wrote:
 

A logical volume in LVM will not handle more than 2TB. You can tie together 
the LVs in a volume group, thus going over the 2TB limit. Choose your 
filesystem well though, some have a 2TB limit too.

Disk size: What are you doing with it. 500GB disks are ATA (maybe SATA). ATA 
is good for low end servers or near line storage, SATA can be used equally to 
SCSI (I am going to suffer for this remark).

RAID5 in software works pretty good (survived a failed disk, and recovered 
another failing raid in 1 month). Hardware is better since you don't have a 
boot partition left which is usually just present on one disk (you can mirror 
that yourself ofcourse).

Regards,
Norbert van Nobelen
On Thursday 20 January 2005 20:51, you wrote:
   

I recently saw Alan Cox say on this list that LVM won't handle more than
2 terabytes. Is this LVM2 or LVM? What is the maximum amount of disk
space LVM2 (or any other RAID/MIRROR capable technology that is in
Linus's kernel) handle? I am talking with various people and we are
looking at Samba on Linux to do several different namespaces (obviously
one tree), most averaging about 3 terabytes, but one would have in
excess of 20 terabytes. We are looking at using 320 to 500 gigabyte
drives in these arrays. (How? IEEE-1394. Which brings a question I will
ask in a second email.)
Is RAID 5 all that bad using this software method? Is RAID 5 available?
Trever Adams
--
"They that can give up essential liberty to obtain a little temporary
safety deserve neither liberty nor safety." -- Benjamin Franklin, 1759
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/
 

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/
   

--
"Assassination is the extreme form of censorship." -- George Bernard
Shaw (1856-1950)
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/
 

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread William Lee Irwin III
On Thu, Jan 20, 2005 at 03:17:37PM -0700, Trever L. Adams wrote:
> Second, you mentioned file systems. We were talking about ext3. I have
> never used any others in Linux (barring ext2, minixfs, and fat). I had
> heard XFS from IBM was pretty good. I would rather not use reiserfs.

XFS is from SGI. JFS is from IBM.


-- wli
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Trever L. Adams
PV = the device
VG = groups of them (the RAID5 array?)
LV = what? the file system?

So, from what you are telling me, and the man page, 2.6.x with LVM2 can
have basically any size of PV, VG, and LV I want.

Am I flawed in my understanding?

Thank you,
Trever

On Thu, 2005-01-20 at 22:02 +, Alasdair G Kergon wrote:
> On Thu, Jan 20, 2005 at 10:40:02PM +0100, Norbert van Nobelen wrote:
> > A logical volume in LVM will not handle more than 2TB. You can tie together 
> > the LVs in a volume group, thus going over the 2TB limit. 
> 
> Confused over terminology?
> Tie PVs together to form a VG, then divide VG up into LVs.
> 
> Size limit depends on metadata format and the kernel: old LVM1 format has 
> lower size limits - see the vgcreate man page.
> 
> New LVM2 metadata format relaxes those limits and lets you have LVs > 2TB
> with a 2.6 kernel.
> 
> Alasdair
--
"Assassination is the extreme form of censorship." -- George Bernard
Shaw (1856-1950)

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Trever L. Adams
It is for a group. For the most part it is data access/retention. Writes
and such would be more similar to a desktop. I would use SATA if they
were (nearly) equally priced and there were awesome 1394 to SATA bridge
chips that worked well with Linux. So, right now, I am looking at ATA to
1394.

So, to get 2TB of RAID5 you have 6 500 GB disks right? So, will this
work within on LV? Or is it 2TB of diskspace total? So, are volume
groups pretty fault tolerant if you have a bunch of RAID5 LVs below
them? This is my one worry about this.

Second, you mentioned file systems. We were talking about ext3. I have
never used any others in Linux (barring ext2, minixfs, and fat). I had
heard XFS from IBM was pretty good. I would rather not use reiserfs.

Any recommendations.

Trever

P.S. Why won't an LV support over 2TB?

S.P.S. I am not really worried about the boot and programs drive. They
will be spun down most of the time I am sure.

On Thu, 2005-01-20 at 22:40 +0100, Norbert van Nobelen wrote:
> A logical volume in LVM will not handle more than 2TB. You can tie together 
> the LVs in a volume group, thus going over the 2TB limit. Choose your 
> filesystem well though, some have a 2TB limit too.
> 
> Disk size: What are you doing with it. 500GB disks are ATA (maybe SATA). ATA 
> is good for low end servers or near line storage, SATA can be used equally to 
> SCSI (I am going to suffer for this remark).
> 
> RAID5 in software works pretty good (survived a failed disk, and recovered 
> another failing raid in 1 month). Hardware is better since you don't have a 
> boot partition left which is usually just present on one disk (you can mirror 
> that yourself ofcourse).
> 
> Regards,
> 
> Norbert van Nobelen
> 
> On Thursday 20 January 2005 20:51, you wrote:
> > I recently saw Alan Cox say on this list that LVM won't handle more than
> > 2 terabytes. Is this LVM2 or LVM? What is the maximum amount of disk
> > space LVM2 (or any other RAID/MIRROR capable technology that is in
> > Linus's kernel) handle? I am talking with various people and we are
> > looking at Samba on Linux to do several different namespaces (obviously
> > one tree), most averaging about 3 terabytes, but one would have in
> > excess of 20 terabytes. We are looking at using 320 to 500 gigabyte
> > drives in these arrays. (How? IEEE-1394. Which brings a question I will
> > ask in a second email.)
> >
> > Is RAID 5 all that bad using this software method? Is RAID 5 available?
> >
> > Trever Adams
> > --
> > "They that can give up essential liberty to obtain a little temporary
> > safety deserve neither liberty nor safety." -- Benjamin Franklin, 1759
> >
> > -
> > To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> > the body of a message to [EMAIL PROTECTED]
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > Please read the FAQ at  http://www.tux.org/lkml/
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/
> 
--
"Assassination is the extreme form of censorship." -- George Bernard
Shaw (1856-1950)

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Alasdair G Kergon
On Thu, Jan 20, 2005 at 10:40:02PM +0100, Norbert van Nobelen wrote:
> A logical volume in LVM will not handle more than 2TB. You can tie together 
> the LVs in a volume group, thus going over the 2TB limit. 

Confused over terminology?
Tie PVs together to form a VG, then divide VG up into LVs.

Size limit depends on metadata format and the kernel: old LVM1 format has 
lower size limits - see the vgcreate man page.

New LVM2 metadata format relaxes those limits and lets you have LVs > 2TB
with a 2.6 kernel.

Alasdair
-- 
[EMAIL PROTECTED]
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: LVM2

2005-01-20 Thread Norbert van Nobelen
A logical volume in LVM will not handle more than 2TB. You can tie together 
the LVs in a volume group, thus going over the 2TB limit. Choose your 
filesystem well though, some have a 2TB limit too.

Disk size: What are you doing with it. 500GB disks are ATA (maybe SATA). ATA 
is good for low end servers or near line storage, SATA can be used equally to 
SCSI (I am going to suffer for this remark).

RAID5 in software works pretty good (survived a failed disk, and recovered 
another failing raid in 1 month). Hardware is better since you don't have a 
boot partition left which is usually just present on one disk (you can mirror 
that yourself ofcourse).

Regards,

Norbert van Nobelen

On Thursday 20 January 2005 20:51, you wrote:
> I recently saw Alan Cox say on this list that LVM won't handle more than
> 2 terabytes. Is this LVM2 or LVM? What is the maximum amount of disk
> space LVM2 (or any other RAID/MIRROR capable technology that is in
> Linus's kernel) handle? I am talking with various people and we are
> looking at Samba on Linux to do several different namespaces (obviously
> one tree), most averaging about 3 terabytes, but one would have in
> excess of 20 terabytes. We are looking at using 320 to 500 gigabyte
> drives in these arrays. (How? IEEE-1394. Which brings a question I will
> ask in a second email.)
>
> Is RAID 5 all that bad using this software method? Is RAID 5 available?
>
> Trever Adams
> --
> "They that can give up essential liberty to obtain a little temporary
> safety deserve neither liberty nor safety." -- Benjamin Franklin, 1759
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


LVM2

2005-01-20 Thread Trever L. Adams
I recently saw Alan Cox say on this list that LVM won't handle more than
2 terabytes. Is this LVM2 or LVM? What is the maximum amount of disk
space LVM2 (or any other RAID/MIRROR capable technology that is in
Linus's kernel) handle? I am talking with various people and we are
looking at Samba on Linux to do several different namespaces (obviously
one tree), most averaging about 3 terabytes, but one would have in
excess of 20 terabytes. We are looking at using 320 to 500 gigabyte
drives in these arrays. (How? IEEE-1394. Which brings a question I will
ask in a second email.)

Is RAID 5 all that bad using this software method? Is RAID 5 available?

Trever Adams
--
"They that can give up essential liberty to obtain a little temporary
safety deserve neither liberty nor safety." -- Benjamin Franklin, 1759

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/