I had progs 3.12 and updated to the latest from git(3.16). With this
update, btrfs fi show reports there is a missing device immediately
after i pull it out. Thanks!

I am using virtualbox to test this. So, I am detaching the drive like so:

vboxmanage storageattach <vm> --storagectl <controller> --port <port>
--device <device> --medium none

Next I am going to try and test a more realistic scenario where a
harddrive is not pulled out, but is damaged.

Can/does btrfs mark a filesystem(say, 2 drive raid1) degraded or
unhealthy automatically when one drive is damaged badly enough that it
cannot be written to or read from reliably?

Suman

On Sun, Oct 12, 2014 at 7:21 PM, Anand Jain <anand.j...@oracle.com> wrote:
>
> Suman,
>
>> To simulate the failure, I detached one of the drives from the system.
>> After that, I see no sign of a problem except for these errors:
>
>  Are you physically pulling out the device ? I wonder if lsblk or blkid
>  shows the error ? reporting device missing logic is in the progs (so
>  have that latest) and it works provided user script such as blkid/lsblk
>  also reports the problem. OR for soft-detach tests you could use
>  devmgt at http://github.com/anajain/devmgt
>
>  Also I am trying to get the device management framework for the btrfs
>  with a more better device management and reporting.
>
> Thanks,  Anand
>
>
>
> On 10/13/14 07:50, Suman C wrote:
>>
>> Hi,
>>
>> I am testing some disk failure scenarios in a 2 drive raid1 mirror.
>> They are 4GB each, virtual SATA drives inside virtualbox.
>>
>> To simulate the failure, I detached one of the drives from the system.
>> After that, I see no sign of a problem except for these errors:
>>
>> Oct 12 15:37:14 rock-dev kernel: btrfs: bdev /dev/sdb errs: wr 0, rd
>> 0, flush 1, corrupt 0, gen 0
>> Oct 12 15:37:14 rock-dev kernel: lost page write due to I/O error on
>> /dev/sdb
>>
>> /dev/sdb is gone from the system, but btrfs fi show still lists it.
>>
>> Label: raid1pool  uuid: 4e5d8b43-1d34-4672-8057-99c51649b7c6
>>      Total devices 2 FS bytes used 1.46GiB
>>      devid    1 size 4.00GiB used 2.45GiB path /dev/sdb
>>      devid    2 size 4.00GiB used 2.43GiB path /dev/sdc
>>
>> I am able to read and write just fine, but do see the above errors in
>> dmesg.
>>
>> What is the best way to find out that one of the drives has gone bad?
>>
>> Suman
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
>> the body of a message to majord...@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to