Hmmm,

just upgraded some servers to U7. Unfortunately one server's primary disk
died during the upgrade, so that luactivate was not able to activate the
s10u7 BE ("Unable to determine the configuration ..."). Since the rpool
is a 2-way mirror, the boot-device=/p...@1f,700000/s...@2/d...@1,0:a was
simply set to /p...@1f,700000/s...@2/d...@0,0:a and checked, whether the
machine still reboots unattended. As expected - no problem.

At the evening the faulty disk was replaced and the mirror resilvered via
'zpool replace rpool c1t1d0s0' (see below).  Since there was no error and
everything stated to be healthy, the s10u7 BE was luactivated (no error
here message as well) and 'init 6'.

Unfortunately, now the server was gone and no known recipe helped to
revive it (I guess, LU damaged the zpool.cache?) :((((

Any hints, how to get the rpool back? 

Regards,
jel.


What has been tried 'til now:
########################################
{3} ok boot
Boot device: /p...@1f,700000/s...@2/d...@1,0:a  File and args:
Bad magic number in disk label
Can't open disk label package

Can't open boot device

{3} ok
########################################
{3} ok boot /p...@1f,700000/s...@2/d...@0,0:a
Boot device: /p...@1f,700000/s...@2/d...@0,0:a  File and args:
SunOS Release 5.10 Version Generic_138888-08 64-bit
Copyright 1983-2009 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
NOTICE:
spa_import_rootpool: error 22

Cannot mount root on /p...@1f,700000/s...@2/d...@0,0:a fstype zfs

panic[cpu3]/thread=180e000: vfs_mountroot: cannot mount root

000000000180b950 genunix:vfs_mountroot+358 (800, 200, 0, 1875c00,
189f800, 18ca000)
  %l0-3: 00000000010ba000 00000000010ba208 000000000187bba8
  %00000000011e8400
  %l4-7: 00000000011e8400 00000000018cc400 0000000000000600
  %0000000000000200
000000000180ba10 genunix:main+a0 (1815178, 180c000, 18397b0, 18c6800,
181b578, 1815000)
  %l0-3: 0000000001015400 0000000000000001 0000000070002000
  %0000000000000000
  %l4-7: 000000000183ec00 0000000000000003 000000000180c000
  %0000000000000000

skipping system dump - no dump device configured
rebooting...

SC Alert: Host System has Reset
########################################
{3} ok boot net -s
Boot device: /p...@1c,600000/netw...@2  File and args: -s
1000 Mbps FDX Link up
Timeout waiting for ARP/RARP packet
3a000 1000 Mbps FDX Link up
SunOS Release 5.10 Version Generic_137137-09 64-bit
Copyright 1983-2008 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
Hardware watchdog enabled
Booting to milestone "milestone/single-user:default".
Configuring devices.
Using RPC Bootparams for network configuration information.
Attempting to configure interface ce1...
Skipped interface ce1
Attempting to configure interface ce0...
Configured interface ce0
Requesting System Maintenance Mode
SINGLE USER MODE
# mount -F zfs /dev/dsk/c1t1d0s0 /mnt
cannot open '/dev/dsk/c1t1d0s0': invalid dataset name
# mount -F zfs /dev/dsk/c1t0d0s0 /mnt
cannot open '/dev/dsk/c1t0d0s0': invalid dataset name
# zpool import
  pool: pool1
    id: 5088500955966129017
 state: ONLINE
action: The pool can be imported using its name or numeric identifier.
config:

        pool1       ONLINE
          mirror    ONLINE
            c1t2d0  ONLINE
            c1t3d0  ONLINE

  pool: rpool
    id: 5910200402071733373
 state: UNAVAIL
action: The pool cannot be imported due to damaged devices or data.
config:

        rpool         UNAVAIL  insufficient replicas
          mirror      UNAVAIL  corrupted data
            c1t1d0s0  ONLINE
            c1t0d0s0  ONLINE

# dd if=/dev/rdsk/c1t1d0s0 of=/tmp/bb bs=1b iseek=1 count=15
15+0 records in
15+0 records out
# dd if=/dev/rdsk/c1t1d0s0 of=/tmp/bb bs=1b iseek=1024 oseek=15 count=16
16+0 records in
16+0 records out
# cmp /tmp/bb /usr/platform/`uname -i`/lib/fs/zfs/bootblk
# echo $?
0
# dd if=/dev/rdsk/c1t0d0s0 of=/tmp/ab bs=1b iseek=1 count=15
15+0 records in
15+0 records out
# dd if=/dev/rdsk/c1t0d0s0 of=/tmp/ab bs=1b iseek=1024 oseek=15 count=16
16+0 records in
16+0 records out
# cmp /tmp/ab /usr/platform/`uname -i`/lib/fs/zfs/bootblk
# echo $?
0
#

pre-history:
============
admin.tpol ~ # zpool status -xv
  pool: rpool
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
        continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
 scrub: resilver in progress for 0h28m, 98.10% done, 0h0m to go
config:

        NAME                STATE     READ WRITE CKSUM
        rpool               DEGRADED     0     0     0
          mirror            DEGRADED     0     0     0
            replacing       DEGRADED     0     0     0
              c1t1d0s0/old  FAULTED      0     0     0  corrupted data
              c1t1d0s0      ONLINE       0     0     0
            c1t0d0s0        ONLINE       0     0     0

errors: No known data errors

admin.tpol ~ # zpool status -xv
all pools are healthy
admin.tpol ~ # zpool status 
  pool: pool1
 state: ONLINE
 scrub: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        pool1       ONLINE       0     0     0
          mirror    ONLINE       0     0     0
            c1t2d0  ONLINE       0     0     0
            c1t3d0  ONLINE       0     0     0

errors: No known data errors

  pool: rpool
 state: ONLINE
 scrub: resilver completed after 0h28m with 0 errors on Tue Jun 16 19:13:57 2009
config:

        NAME          STATE     READ WRITE CKSUM
        rpool         ONLINE       0     0     0
          mirror      ONLINE       0     0     0
            c1t1d0s0  ONLINE       0     0     0
            c1t0d0s0  ONLINE       0     0     0

errors: No known data errors

admin.tpol / # luactivate -l $ERR -o $OUT S10u7
...
In case of a failure while booting ...
mount -Fzfs /dev/dsk/c1t1d0s0 /mnt
...
Modifying boot archive service
Activation of boot environment <s10u7> successful.
8.32u 16.74s 1:25.15 29.4%
admin.tpol / # init 6
-- 
Otto-von-Guericke University     http://www.cs.uni-magdeburg.de/
Department of Computer Science   Geb. 29 R 027, Universitaetsplatz 2
39106 Magdeburg, Germany         Tel: +49 391 67 12768
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to