Re: Possible Raid Bug

2016-03-28 Thread Patrik Lundquist
On 28 March 2016 at 05:54, Anand Jain wrote: > > On 03/26/2016 07:51 PM, Patrik Lundquist wrote: >> >> # btrfs device stats /mnt >> >> [/dev/sde].write_io_errs 11 >> [/dev/sde].read_io_errs0 >> [/dev/sde].flush_io_errs 2 >> [/dev/sde].corruption_errs 0 >>

Re: Possible Raid Bug

2016-03-27 Thread Anand Jain
Hi Patrik, Thanks for posting a test case. more below. On 03/26/2016 07:51 PM, Patrik Lundquist wrote: So with the lessons learned: # mkfs.btrfs -m raid10 -d raid10 /dev/sdb /dev/sdc /dev/sdd /dev/sde # mount /dev/sdb /mnt; dmesg | tail # touch /mnt/test1; sync; btrfs device usage /mnt

Re: Possible Raid Bug

2016-03-27 Thread Stephen Williams
Yeah I think the Gotchas page would be a good place to give people a heads up. -- Stephen Williams steph...@veryfast.biz On Sat, Mar 26, 2016, at 09:58 PM, Chris Murphy wrote: > On Sat, Mar 26, 2016 at 8:00 AM, Stephen Williams > wrote: > > > I know this is quite a

Re: Possible Raid Bug

2016-03-26 Thread Chris Murphy
On Sat, Mar 26, 2016 at 8:00 AM, Stephen Williams wrote: > I know this is quite a rare occurrence for home use but for Data center > use this is something that will happen A LOT. > This really should be placed in the wiki while we wait for a fix. I can > see a lot of sys

Re: Possible Raid Bug

2016-03-26 Thread Chris Murphy
On Sat, Mar 26, 2016 at 5:51 AM, Patrik Lundquist wrote: > # btrfs replace start -B 4 /dev/sde /mnt; dmesg | tail > > # btrfs device stats /mnt > > [/dev/sde].write_io_errs 0 > [/dev/sde].read_io_errs0 > [/dev/sde].flush_io_errs 0 > [/dev/sde].corruption_errs

Re: Possible Raid Bug

2016-03-26 Thread Stephen Williams
Can confirm that you only get one chance to fix the problem before the array is dead. I know this is quite a rare occurrence for home use but for Data center use this is something that will happen A LOT. This really should be placed in the wiki while we wait for a fix. I can see a lot of sys

Re: Possible Raid Bug

2016-03-26 Thread Patrik Lundquist
So with the lessons learned: # mkfs.btrfs -m raid10 -d raid10 /dev/sdb /dev/sdc /dev/sdd /dev/sde # mount /dev/sdb /mnt; dmesg | tail # touch /mnt/test1; sync; btrfs device usage /mnt Only raid10 profiles. # echo 1 >/sys/block/sde/device/delete We lost a disk. # touch /mnt/test2; sync; dmesg

Re: Possible Raid Bug

2016-03-25 Thread Duncan
Chris Murphy posted on Fri, 25 Mar 2016 15:34:11 -0600 as excerpted: > Basically you get one chance to mount rw,degraded and you have to fix > the problem at that time. And you have to balance away any phantom > single chunks that have appeared. For what it's worth it's not the > reboot that

Re: Possible Raid Bug

2016-03-25 Thread Anand Jain
On 03/26/2016 04:09 AM, Alexander Fougner wrote: 2016-03-25 20:57 GMT+01:00 Patrik Lundquist : On 25 March 2016 at 18:20, Stephen Williams wrote: Your information below was very helpful and I was able to recreate the Raid array. However my

Re: Possible Raid Bug

2016-03-25 Thread Chris Murphy
On Fri, Mar 25, 2016 at 1:57 PM, Patrik Lundquist wrote: > > Only errors on the device formerly known as /dev/sde, so why won't it > mount degraded,rw? Now I'm stuck like Stephen. > > # btrfs device usage /mnt > /dev/sdb, ID: 1 >Device size: 2.00GiB >

Re: Possible Raid Bug

2016-03-25 Thread Alexander Fougner
2016-03-25 20:57 GMT+01:00 Patrik Lundquist : > On 25 March 2016 at 18:20, Stephen Williams wrote: >> >> Your information below was very helpful and I was able to recreate the >> Raid array. However my initial question still stands - What if the

Re: Possible Raid Bug

2016-03-25 Thread Patrik Lundquist
On 25 March 2016 at 18:20, Stephen Williams wrote: > > Your information below was very helpful and I was able to recreate the > Raid array. However my initial question still stands - What if the > drives dies completely? I work in a Data center and we see this quite a > lot

Re: Possible Raid Bug

2016-03-25 Thread Stephen Williams
Hi Patrik, [root@Xen ~]# uname -r 4.4.5-1-ARCH [root@Xen ~]# pacman -Q btrfs-progs btrfs-progs 4.4.1-1 Your information below was very helpful and I was able to recreate the Raid array. However my initial question still stands - What if the drives dies completely? I work in a Data center and we

Re: Possible Raid Bug

2016-03-25 Thread Patrik Lundquist
On Debian Stretch with Linux 4.4.6, btrfs-progs 4.4 in VirtualBox 5.0.16 with 4*2GB VDIs: # mkfs.btrfs -m raid10 -d raid10 /dev/sdb /dev/sdc /dev/sdd /dev/sdbe # mount /dev/sdb /mnt # touch /mnt/test # umount /mnt Everything fine so far. # wipefs -a /dev/sde *reboot* # mount /dev/sdb /mnt

Re: Possible Raid Bug

2016-03-25 Thread Duncan
Patrik Lundquist posted on Fri, 25 Mar 2016 13:48:08 +0100 as excerpted: > On 25 March 2016 at 12:49, Stephen Williams > wrote: >> >> So catch 22, you need all the drives otherwise it won't let you mount, >> But what happens if a drive dies and the OS doesn't detect it?

Re: Possible Raid Bug

2016-03-25 Thread Patrik Lundquist
On 25 March 2016 at 12:49, Stephen Williams wrote: > > So catch 22, you need all the drives otherwise it won't let you mount, > But what happens if a drive dies and the OS doesn't detect it? BTRFS > wont allow you to mount the raid volume to remove the bad disk! Version of

Possible Raid Bug

2016-03-25 Thread Stephen Williams
Hi, Find instructions on how to recreate below - I have a BTRFS raid 10 setup in Virtualbox (I'm getting to grips with the Filesystem) I have the raid mounted to /mnt like so - [root@Xen ~]# btrfs filesystem show /mnt/ Label: none uuid: ad1d95ee-5cdc-420f-ad30-bd16158ad8cb Total