On 28 March 2016 at 05:54, Anand Jain wrote:
>
> On 03/26/2016 07:51 PM, Patrik Lundquist wrote:
>>
>> # btrfs device stats /mnt
>>
>> [/dev/sde].write_io_errs 11
>> [/dev/sde].read_io_errs0
>> [/dev/sde].flush_io_errs 2
>> [/dev/sde].corruption_errs 0
>>
Hi Patrik,
Thanks for posting a test case. more below.
On 03/26/2016 07:51 PM, Patrik Lundquist wrote:
So with the lessons learned:
# mkfs.btrfs -m raid10 -d raid10 /dev/sdb /dev/sdc /dev/sdd /dev/sde
# mount /dev/sdb /mnt; dmesg | tail
# touch /mnt/test1; sync; btrfs device usage /mnt
Yeah I think the Gotchas page would be a good place to give people a
heads up.
--
Stephen Williams
steph...@veryfast.biz
On Sat, Mar 26, 2016, at 09:58 PM, Chris Murphy wrote:
> On Sat, Mar 26, 2016 at 8:00 AM, Stephen Williams
> wrote:
>
> > I know this is quite a
On Sat, Mar 26, 2016 at 8:00 AM, Stephen Williams wrote:
> I know this is quite a rare occurrence for home use but for Data center
> use this is something that will happen A LOT.
> This really should be placed in the wiki while we wait for a fix. I can
> see a lot of sys
On Sat, Mar 26, 2016 at 5:51 AM, Patrik Lundquist
wrote:
> # btrfs replace start -B 4 /dev/sde /mnt; dmesg | tail
>
> # btrfs device stats /mnt
>
> [/dev/sde].write_io_errs 0
> [/dev/sde].read_io_errs0
> [/dev/sde].flush_io_errs 0
> [/dev/sde].corruption_errs
Can confirm that you only get one chance to fix the problem before the
array is dead.
I know this is quite a rare occurrence for home use but for Data center
use this is something that will happen A LOT.
This really should be placed in the wiki while we wait for a fix. I can
see a lot of sys
So with the lessons learned:
# mkfs.btrfs -m raid10 -d raid10 /dev/sdb /dev/sdc /dev/sdd /dev/sde
# mount /dev/sdb /mnt; dmesg | tail
# touch /mnt/test1; sync; btrfs device usage /mnt
Only raid10 profiles.
# echo 1 >/sys/block/sde/device/delete
We lost a disk.
# touch /mnt/test2; sync; dmesg
Chris Murphy posted on Fri, 25 Mar 2016 15:34:11 -0600 as excerpted:
> Basically you get one chance to mount rw,degraded and you have to fix
> the problem at that time. And you have to balance away any phantom
> single chunks that have appeared. For what it's worth it's not the
> reboot that
On 03/26/2016 04:09 AM, Alexander Fougner wrote:
2016-03-25 20:57 GMT+01:00 Patrik Lundquist :
On 25 March 2016 at 18:20, Stephen Williams wrote:
Your information below was very helpful and I was able to recreate the
Raid array. However my
On Fri, Mar 25, 2016 at 1:57 PM, Patrik Lundquist
wrote:
>
> Only errors on the device formerly known as /dev/sde, so why won't it
> mount degraded,rw? Now I'm stuck like Stephen.
>
> # btrfs device usage /mnt
> /dev/sdb, ID: 1
>Device size: 2.00GiB
>
2016-03-25 20:57 GMT+01:00 Patrik Lundquist :
> On 25 March 2016 at 18:20, Stephen Williams wrote:
>>
>> Your information below was very helpful and I was able to recreate the
>> Raid array. However my initial question still stands - What if the
On 25 March 2016 at 18:20, Stephen Williams wrote:
>
> Your information below was very helpful and I was able to recreate the
> Raid array. However my initial question still stands - What if the
> drives dies completely? I work in a Data center and we see this quite a
> lot
Hi Patrik,
[root@Xen ~]# uname -r
4.4.5-1-ARCH
[root@Xen ~]# pacman -Q btrfs-progs
btrfs-progs 4.4.1-1
Your information below was very helpful and I was able to recreate the
Raid array. However my initial question still stands - What if the
drives dies completely? I work in a Data center and we
On Debian Stretch with Linux 4.4.6, btrfs-progs 4.4 in VirtualBox
5.0.16 with 4*2GB VDIs:
# mkfs.btrfs -m raid10 -d raid10 /dev/sdb /dev/sdc /dev/sdd /dev/sdbe
# mount /dev/sdb /mnt
# touch /mnt/test
# umount /mnt
Everything fine so far.
# wipefs -a /dev/sde
*reboot*
# mount /dev/sdb /mnt
Patrik Lundquist posted on Fri, 25 Mar 2016 13:48:08 +0100 as excerpted:
> On 25 March 2016 at 12:49, Stephen Williams
> wrote:
>>
>> So catch 22, you need all the drives otherwise it won't let you mount,
>> But what happens if a drive dies and the OS doesn't detect it?
On 25 March 2016 at 12:49, Stephen Williams wrote:
>
> So catch 22, you need all the drives otherwise it won't let you mount,
> But what happens if a drive dies and the OS doesn't detect it? BTRFS
> wont allow you to mount the raid volume to remove the bad disk!
Version of
Hi,
Find instructions on how to recreate below -
I have a BTRFS raid 10 setup in Virtualbox (I'm getting to grips with
the Filesystem)
I have the raid mounted to /mnt like so -
[root@Xen ~]# btrfs filesystem show /mnt/
Label: none uuid: ad1d95ee-5cdc-420f-ad30-bd16158ad8cb
Total
17 matches
Mail list logo