Re: [zfs-discuss] bogus zfs error message on boot

2006-11-15 Thread Eric Schrock
This is likely a variation of:

6401126 FM reports 'pool data unavailable' because of timing between FM and 
mounting of file systems

Basically, what's happening is that ZFS is trying to open the pool
before the underlying device backing the vdev is available.  My guess is
that your new hardware is loading later in boot for some reason.  When
you actually come all the way up, the device is available and your pool
is fine.

The precise fix is a little complicated.  As a workaround, you may be
able to add a forceload directive to /etc/system to force the driver
associated with your hardware to attach earlier in boot.

- Eric

On Tue, Nov 14, 2006 at 09:54:50PM -0800, Frank Cusack wrote:
> After swapping some hardware and rebooting:
> 
> SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major
> EVENT-TIME: Tue Nov 14 21:37:55 PST 2006
> PLATFORM: SUNW,Sun-Fire-T1000, CSN: -, HOSTNAME:
> SOURCE: zfs-diagnosis, REV: 1.0
> EVENT-ID: 60b31acc-0de8-c1f3-84ec-935574615804
> DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS 
> for more information.
> AUTO-RESPONSE: No automated response will occur.
> IMPACT: The pool data is unavailable
> REC-ACTION: Run 'zpool status -x' and either attach the missing device or
>restore from backup.
> 
> # zpool status -x
> all pools are healthy
> 
> And in fact they are.  What gives?  This message occurs on every boot now.
> It didn't occur before I changed the hardware.
> 
> I had replaced the FC card with a fw800 card, then I changed it back.
> (the fw800 card didn't work)
> 
> -frank
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bogus zfs error message on boot

2006-11-15 Thread Frank Cusack
On November 16, 2006 1:18:22 AM +1100 James McPherson 
<[EMAIL PROTECTED]> wrote:

On 11/15/06, Frank Cusack <[EMAIL PROTECTED]> wrote:

After swapping some hardware and rebooting:

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major
EVENT-TIME: Tue Nov 14 21:37:55 PST 2006
PLATFORM: SUNW,Sun-Fire-T1000, CSN: -, HOSTNAME:
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 60b31acc-0de8-c1f3-84ec-935574615804
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS
for more information.
AUTO-RESPONSE: No automated response will occur.
IMPACT: The pool data is unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device or
restore from backup.

# zpool status -x
all pools are healthy

And in fact they are.  What gives?  This message occurs on every boot
now. It didn't occur before I changed the hardware.


Sounds like an opportunity for enhancement. At the
very least the ZFS :: FMA interaction should include the
component (pool in this case) which was noted to be
marginal/faulty/dead.


Does zpool status -xv show anything that zpool status -x
doesn't?


Nope.

But I see that my raid array (3511) is now beeping like crazy, playing
a song really.  I think there must be some delay that is causing the
disks not to be available early in the boot?  Then they become available
and get imported?  (I do notice that unlike scsi disks, if I add a disk
to the 3511 it is noticed immediately on the host.)

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bogus zfs error message on boot

2006-11-15 Thread Asif Iqbal

On 11/15/06, Frank Cusack <[EMAIL PROTECTED]> wrote:

After swapping some hardware and rebooting:

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major
EVENT-TIME: Tue Nov 14 21:37:55 PST 2006
PLATFORM: SUNW,Sun-Fire-T1000, CSN: -, HOSTNAME:
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 60b31acc-0de8-c1f3-84ec-935574615804
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS
for more information.
AUTO-RESPONSE: No automated response will occur.
IMPACT: The pool data is unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device or
restore from backup.

# zpool status -x
all pools are healthy


How about this?

zpool export 
zpool import -f 




And in fact they are.  What gives?  This message occurs on every boot now.
It didn't occur before I changed the hardware.

I had replaced the FC card with a fw800 card, then I changed it back.
(the fw800 card didn't work)

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




--
Asif Iqbal
PGP Key: 0xE62693C5 KeyServer: pgp.mit.edu
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bogus zfs error message on boot

2006-11-15 Thread James McPherson

On 11/15/06, Frank Cusack <[EMAIL PROTECTED]> wrote:

After swapping some hardware and rebooting:

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major
EVENT-TIME: Tue Nov 14 21:37:55 PST 2006
PLATFORM: SUNW,Sun-Fire-T1000, CSN: -, HOSTNAME:
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 60b31acc-0de8-c1f3-84ec-935574615804
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS
for more information.
AUTO-RESPONSE: No automated response will occur.
IMPACT: The pool data is unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device or
restore from backup.

# zpool status -x
all pools are healthy

And in fact they are.  What gives?  This message occurs on every boot now.
It didn't occur before I changed the hardware.


Sounds like an opportunity for enhancement. At the
very least the ZFS :: FMA interaction should include the
component (pool in this case) which was noted to be
marginal/faulty/dead.


Does zpool status -xv show anything that zpool status -x
doesn't?

James C. McPherson
--
Solaris kernel software engineer, system admin and troubleshooter
 http://www.jmcp.homeunix.com/blog
Find me on LinkedIn @ http://www.linkedin.com/in/jamescmcpherson
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] bogus zfs error message on boot

2006-11-14 Thread Frank Cusack

After swapping some hardware and rebooting:

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major
EVENT-TIME: Tue Nov 14 21:37:55 PST 2006
PLATFORM: SUNW,Sun-Fire-T1000, CSN: -, HOSTNAME:
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 60b31acc-0de8-c1f3-84ec-935574615804
DESC: A ZFS pool failed to open.  Refer to http://sun.com/msg/ZFS-8000-CS 
for more information.

AUTO-RESPONSE: No automated response will occur.
IMPACT: The pool data is unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device or
   restore from backup.

# zpool status -x
all pools are healthy

And in fact they are.  What gives?  This message occurs on every boot now.
It didn't occur before I changed the hardware.

I had replaced the FC card with a fw800 card, then I changed it back.
(the fw800 card didn't work)

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss