I have a FBSD 8 system that won't boot. Console messages suggest I lost my ad6 drive.

When I created the system, built root on a zfs mirror and a raidz1 pool for data. I've managed to boot into an 8.2 Fixit shell and been able to mount root on /mnt/root. However my raidz1 pool (data) can't be imported. I'm *really* hoping there is some way to fix this.

From the fixit shell, here is what I get when I run 'zpool import':

# zpool import
  pool: data
    id: 10923029512613632741
 state: UNAVAIL
status: The pool was last accessed by another system.
action: The pool cannot be imported due to damaged devices or data.
   see: http://www.sun.com/msg/ZFS-8000-EY
config:

data UNAVAIL insufficient replicas raidz1 UNAVAIL insufficient replicas
            gptid/1be11998-6c47-11de-ae82-001b21361de7  ONLINE
ad6p4 UNAVAIL cannot open
            gptid/fc514567-6c46-11de-ae82-001b21361de7  ONLINE
gptid/0f45dd02-6c47-11de-ae82-001b21361de7 UNAVAIL cannot open

However I'm not exactly sure what this is telling me other than something "bad". I did capture this information when I upgraded to version 15:

# 11/26/2010 - DLT
# http://www.freebsddiary.org/zfs-upgrade.php
#
# After upgrading to zpool v 15 and learning about upgrading zfs file systems, # I decided to take a snapshot of disks. It's interesting in the zpool history
# commands, it shows different physical disks than what are currently shown.
# I don't know why.

ad6: 715404MB <Seagate ST3750330AS SD15> at ata3-master UDMA100 SATA 3Gb/s
ad10: 476940MB <Hitachi HDP725050GLA360 GM4OA52A> at ata5-master UDMA100 SATA 3Gb/s
acd0: DVDR <LITE-ON DVDRW LH-20A1L/BL05> at ata6-master UDMA100 SATA 1.5Gb/s
ad14: 476940MB <Hitachi HDP725050GLA360 GM4OA52A> at ata7-master UDMA100 SATA 3Gb/s ad16: 476940MB <Hitachi HDP725050GLA360 GM4OA52A> at ata8-master UDMA100 SATA 3Gb/s

vm# zpool history data
History for 'data':
2009-07-12.08:56:22 zpool create data raidz1 ad8p4 ad6p4 ad12p2 ad14p2
2009-07-12.08:57:37 zfs create data/usr
2009-07-12.08:57:41 zfs create data/var
2009-07-12.08:57:59 zfs create data/home
2009-07-12.09:00:46 zfs set mountpoint=none data
2009-07-12.09:19:27 zfs set mountpoint=/usr data/usr
2009-07-12.09:19:37 zfs set mountpoint=/home data/home
2009-07-12.09:19:46 zfs set mountpoint=/var data/var
2009-07-12.10:00:29 zpool import -f data
2009-07-12.10:19:52 zpool import -f data
2009-07-12.10:35:33 zpool import -f data
2009-07-12.11:34:30 zpool import -f data
2009-07-15.13:02:10 zpool import -f data
2009-07-15.13:24:10 zpool import -f data
2009-07-17.12:48:39 zpool import -f data
2009-07-17.12:49:49 zpool upgrade data
2009-07-17.13:17:44 zpool import -f data
2009-07-17.13:48:19 zpool export data
2009-07-17.13:48:24 zpool import data
2009-07-22.16:05:37 zfs create data/archive
2009-07-22.16:07:15 zfs set mountpoint=/archive data/archive
2009-07-24.23:18:18 zfs create -V 9G data/test
2009-07-24.23:35:54 zfs destroy data/test
2010-06-05.16:39:26 zpool upgrade data
2010-07-05.06:20:09 zpool import -f data
2010-07-05.06:27:09 zfs set mountpoint=/mnt data
2010-11-28.07:39:01 zpool upgrade -a

From dmesg output, these are the drives as seen after booting to Fixit:

ad6: 715404MB <Seagate ST3750330AS SD15> at ata3-master UDMA100 SATA 3Gb/s
ad10: 476940MB <Hitachi HDP725050GLA360 GM4OA52A> at ata5-master UDMA100 SATA 3Gb/s ad14: 476940MB <Hitachi HDP725050GLA360 GM4OA52A> at ata7-master UDMA100 SATA 3Gb/s

Thus it appears I am missing ad16 that I used to have. My data zpool was the bulk of my system with over 600 gig of files and things I'd like to have back. I thought that by creating a raidz1 I could avoid having to back up the huge drive and avoid this grief. However it appears I have lost 2 disks at the same time. :(

Any thoughts before I just give up on recovering my data pool?

And regarding my root pool, my system can't mount root and start. What do I need to do to boot from my degraded root pool. Here's the current status:

# zpool status
  pool: root
 state: DEGRADED
status: One or more devices could not be opened. Sufficient replicas exist for
        the pool to continue functioning in a degraded state.
action: Attach the missing device and online it using 'zpool online'.
   see: http://www.sun.com/msg/ZFS-8000-2Q
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM root DEGRADED 0 0 0 mirror DEGRADED 0 0 0 gptid/5b623854-6c46-11de-ae82-001b21361de7 ONLINE 0 0 0 12032653780322685599 UNAVAIL 0 0 0 was /dev/ad6p3

Do I just need to do a 'zpool detach root /dev/ad6p3' to remove it from the pool and get it to boot? And then once I replace the disk a 'zpool attach root <new partition>' to fix?

Thanks for your time.

Cheers,

Drew

--
Like card tricks?

Visit The Alchemist's Warehouse to
learn card magic secrets for free!

http://alchemistswarehouse.com


_______________________________________________
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "freebsd-questions-unsubscr...@freebsd.org"

Reply via email to