Bob Friesenhahn wrote:
On Mon, 28 Sep 2009, Richard Elling wrote:

Scrub could be faster, but you can try
    tar cf - . > /dev/null

If you think about it, validating checksums requires reading the data.
So you simply need to read the data.

This should work but it does not verify the redundant metadata. For example, the duplicate metadata copy might be corrupt but the problem is not detected since it did not happen to be used.


Not only that - it won't also read all the copies of data if zfs has redundancy configured at a pool level. Scrubbing the pool will. And that's the main reason behind the scrub - to be able to detect and repair checksum errors (if any) while a redundant copy is still fine.

--
Robert Milkowski
http://milek.blogspot.com


_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to