I also posted this in zfs discuss but got no responses:

I have a 10 drive raidz, recently one of the disks appeared to be generating 
errors (this later turned out to be a cable), I removed the disk from the 
array, ran vendor diagnostics (which zeroed it). Upon reinstalling the disk 
however zfs will not resilver it, it gets referred to numerically instead of by 
device name, and when i try to replace it, i get:

# zpool replace data 17096229131581286394 c0t2d0

cannot replace 17096229131581286394 with c0t2d0: cannot replace a replacing 
device

(this also happens with the other device 11342560969745958696)
if i try to detach it i get:

# zpool detach data 17096229131581286394

cannot detach 17096229131581286394: no valid replicas


current zpool output looks like:

# zpool status -v
pool: data
state: DEGRADED
scrub: none requested
config:

NAME STATE READ WRITE CKSUM
data DEGRADED 0 0 0
- raidz1 DEGRADED 0 0 0
--- c0t0d0 ONLINE 0 0 0
--- c0t1d0 ONLINE 0 0 0
--- replacing UNAVAIL 0 543 0 insufficient replicas
------ 17096229131581286394 FAULTED 0 581 0 was /dev/dsk/c0t2d0s0/old
------ 11342560969745958696 FAULTED 0 582 0 was /dev/dsk/c0t2d0s0
--- c0t3d0 ONLINE 0 0 0
--- c0t4d0 ONLINE 0 0 0
--- c0t5d0 ONLINE 0 0 0
--- c0t6d0 ONLINE 0 0 0
--- c0t7d0 ONLINE 0 0 0
--- c2t2d0 ONLINE 0 0 0
--- c2t3d0 ONLINE 0 0 0

errors: No known data errors

i have also tried exporting and reimporting the pool, any help would greatly 
appreciated.
-- 
This message posted from opensolaris.org

Reply via email to