I'm in the process of standing up a couple of t5440's, of which the config will eventually end up in another data center 6k miles from the original config, and I'm supposed to send disks to the data center and we'll start from there (yes, I know how to flar and jumpstart. When the boss says do something, sometimes you *just* have to do it)
As I've already run into the boot failsafe when moving a root disk from one sparc host to another, I recently found out that a sys-unconfig'd disk does not suffer from the same problem. While I am probably going to be told, I shouldn't be doing this, I ran into an interesting "semantics" issue that I think zfs should at least be able to avoid (and which I have seen in other non-abusive configurations..... ;-) 2 zfs disk, root mirrored. c2t0 and c2t1. hot unplug c2t0, (and I should probably have removed the busted mirror from c2t1, but I didn't) sys-unconfig disk in c2t1 move disk to new t5440 boot disk, and it enumerates everything correctly and then I notice zpool thinks it's degraded. I had added the mirror after I realized I wanted to run this by the list.... pool: rpool state: DEGRADED status: One or more devices could not be used because the label is missing or invalid. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Replace the device using 'zpool replace'. see: http://www.sun.com/msg/ZFS-8000-4J scrub: resilver completed after 0h7m with 0 errors on Thu Jan 7 12:10:03 2010 config: NAME STATE READ WRITE CKSUM rpool DEGRADED 0 0 0 mirror DEGRADED 0 0 0 c2t0d0s0 ONLINE 0 0 0 c2t0d0s0 FAULTED 0 0 0 corrupted data c2t3d0s0 ONLINE 0 0 0 13.8G resilvered Anyway, should zfs report a faulted drive of the same ctd# which is already active? I understand why this happened, but from a logistics perspective, shouldn't zfs be smart enough to ignore a faulted disk like this? And this is not the first time I've had this scenario happen (I had an x4500 that had suffered through months of marvell driver bugs and corruption, and we probably had 2 or 3 of these types of things happen while trying to "soft" fix the problems). This also happened with hot-spares, which caused support to spend some time with back-line to figure out a procedure to clear those fauled disks which had the same ctd# as a working hot-spare....... Ben -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss