Can you send the output of the attached D script when running 'zpool status'?
- Eric On Thu, Dec 04, 2008 at 02:58:54PM -0800, Brett wrote: > As a result of a power spike during a thunder storm I lost a sata controller > card. This card supported my zfs pool called newsan which is a 4 x samsung > 1Tb sata2 disk raid-z. I replaced the card and the devices have the same > controller/disk numbers, but now have the following issue. > > -bash-3.2$ zpool status > pool: newsan > state: FAULTED > status: The pool metadata is corrupted and the pool cannot be opened. > action: Destroy and re-create the pool from a backup source. > see: http://www.sun.com/msg/ZFS-8000-72 > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > newsan FAULTED 1 0 0 corrupted data > raidz1 ONLINE 6 0 0 > c10d1 ONLINE 17 0 0 > c10d0 ONLINE 17 0 0 > c9d1 ONLINE 24 0 0 > c9d0 ONLINE 24 0 0 > > Something majorly weird is going on as when i run format i see this :- > -bash-3.2$ pfexec format > Searching for disks...done > > > AVAILABLE DISK SELECTIONS: > 0. c3d0 <DEFAULT cyl 19454 alt 2 hd 255 sec 63> > /[EMAIL PROTECTED],0/pci8086,[EMAIL PROTECTED],4/[EMAIL > PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 > 1. c3d1 <DEFAULT cyl 19454 alt 2 hd 255 sec 63> > /[EMAIL PROTECTED],0/pci8086,[EMAIL PROTECTED],4/[EMAIL > PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 > 2. c9d0 <SAMSUNG-S13PJ1BQ60312-0001-31.50MB> > /[EMAIL PROTECTED],0/pci8086,[EMAIL PROTECTED]/[EMAIL > PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 > 3. c9d1 <SAMSUNG-S13PJ1BQ60311-0001-31.50MB> > /[EMAIL PROTECTED],0/pci8086,[EMAIL PROTECTED]/[EMAIL > PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 > 4. c10d0 <SAMSUNG-S13PJ1BQ60311-0001-31.50MB> > /[EMAIL PROTECTED],0/pci8086,[EMAIL PROTECTED]/[EMAIL > PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 > 5. c10d1 <SAMSUNG-S13PJ1BQ60312-0001-31.50MB> > /[EMAIL PROTECTED],0/pci8086,[EMAIL PROTECTED]/[EMAIL > PROTECTED]/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 > > ??? 31.50 MB ??? they all used to show as 1Tb i believe (or 931Mb or whatever) > > Specify disk (enter its number): 2 > selecting c9d0 > NO Alt slice > No defect list found > [disk formatted, no defect list found] > /dev/dsk/c9d0s0 is part of active ZFS pool newsan. Please see zpool(1M). > format> p > partition> p > Current partition table (original): > Total disk sectors available: 1953503710 + 16384 (reserved sectors) > > Part Tag Flag First Sector Size Last Sector > 0 usr wm 256 931.50GB 1953503710 > 1 unassigned wm 0 0 0 > 2 unassigned wm 0 0 0 > 3 unassigned wm 0 0 0 > 4 unassigned wm 0 0 0 > 5 unassigned wm 0 0 0 > 6 unassigned wm 0 0 0 > 8 reserved wm 1953503711 8.00MB 1953520094 > > So the partition table is looking correct. I dont believe all 4 disks died > concurrently. > > Any thoughts on how to recover? I dont particularly want to restore the > couple of terabytes of data if i dont have to. > > analyze> read > Ready to analyze (won't harm SunOS). This takes a long time, > but is interruptable with CTRL-C. Continue? y > Current Defect List must be initialized to do automatic repair. > > Oh and whats this defect list thing? I havnt seen that before > > defect> print > No working list defined. > defect> create > Controller does not support creating manufacturer's defect list. > defect> extract > Ready to extract working list. This cannot be interrupted > and may take a long while. Continue? y > NO Alt slice > NO Alt slice > Extracting defect list...No defect list found > Extraction failed. > defect> commit > Ready to update Current Defect List, continue? y > Current Defect List updated, total of 0 defects. > Disk must be reformatted for changes to take effect. > analyze> read > Ready to analyze (won't harm SunOS). This takes a long time, > but is interruptable with CTRL-C. Continue? y > > pass 0 > 64386 > > pass 1 > 64386 > > Total of 0 defective blocks repaired. > > So the read test seemed to work fine. > > Any suggestions on how to proceed? Thoughts on why the disks are showing > weirdly in format? Any way to recover/rebuild the zpool metadata? > > Any help would be appreciated > > Regards Rep > -- > This message posted from opensolaris.org > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Eric Schrock, Fishworks http://blogs.sun.com/eschrock
#!/sbin/dtrace -s #pragma D option quiet BEGIN { printf("run 'zpool import' to generate trace\n\n"); } vdev_raidz_open:entry { printf("%d BEGIN RAIDZ OPEN\n", timestamp); printf("%d config asize = %d\n", timestamp, args[0]->vdev_asize); printf("%d config ashift = %d\n", timestamp, args[0]->vdev_top->vdev_ashift); self->child = 1; self->asize = args[1]; self->ashift = args[2]; } vdev_disk_open:entry /self->child/ { self->disk_asize = args[1]; self->disk_ashift = args[2]; } vdev_disk_open:return /self->child/ { printf("%d child[%d]: asize = %d, ashift = %d\n", timestamp, self->child - 1, *self->disk_asize, *self->disk_ashift); self->disk_asize = 0; self->disk_ashift = 0; self->child++; } vdev_raidz_open:return { printf("%d asize = %d\n", timestamp, *self->asize); printf("%d ashift = %d\n", timestamp, *self->ashift); printf("%d END RAIDZ OPEN\n", timestamp); self->child = 0; self->asize = 0; self->ashift = 0; }
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss