Hello,
I've been getting warnings that my zfs pool is degraded. At first it was
complaining about a few corrupt files, which were listed as hex numbers instead
of filenames, i.e.
VOL1:<0x0>
After a scrub, a couple of the filenames appeared - turns out they were in
snapshots I don't really need, so I destroyed those snapshots and started a new
scrub. Subsequently, I typed " zpool status -v VOL1" ... and the machine
rebooted. When I could log on again, I looked at /var/log/messages, but found
nothing interesting prior to the reboot. I typed " zpool status -v VOL1" again,
whereupon the machine rebooted. When the machine was back up, I stopped the
scrub, waited a while, then typed "zpool status -v VOL1" again, and this time
got:
r...@nexenta1:~# zpool status -v VOL1
pool: VOL1
state: DEGRADED
status: One or more devices has experienced an unrecoverable error. An
attempt was made to correct the error. Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
using 'zpool clear' or replace the device with 'zpool replace'.
see: http://www.sun.com/msg/ZFS-8000-9P
scan: scrub canceled on Wed Aug 11 11:03:15 2010
config:
NAME STATE READ WRITE CKSUM
VOL1 DEGRADED 0 0 0
raidz1-0 DEGRADED 0 0 0
c2d0 DEGRADED 0 0 0 too many errors
c3d0 DEGRADED 0 0 0 too many errors
c4d0 DEGRADED 0 0 0 too many errors
c5d0 DEGRADED 0 0 0 too many errors
So, I have the following questions:
1) How do I find out which file is corrupt, when I only get something like
"VOL1:<0x0>"
2) What could be causing these reboots?
3) How can I fix my pool?
Thanks!
--
This message posted from opensolaris.org
_______________________________________________
opensolaris-help mailing list
[email protected]