One of our Solaris 10 update 3 servers paniced today with the following error:

Sep 18 00:34:53 m2000ef savecore: [ID 570001 auth.error] reboot after
panic: assertion failed: ss != NULL, file:
../../common/fs/zfs/space_map.c, line: 125

The server saved a core file, and the resulting backtrace is listed below:

$ mdb unix.0 vmcore.0
> $c
vpanic()
0xfffffffffb9b49f3()
space_map_remove+0x239()
space_map_load+0x17d()
metaslab_activate+0x6f()
metaslab_group_alloc+0x187()
metaslab_alloc_dva+0xab()
metaslab_alloc+0x51()
zio_dva_allocate+0x3f()
zio_next_stage+0x72()
zio_checksum_generate+0x5f()
zio_next_stage+0x72()
zio_write_compress+0x136()
zio_next_stage+0x72()
zio_wait_for_children+0x49()
zio_wait_children_ready+0x15()
zio_next_stage_async+0xae()
zio_wait+0x2d()
arc_write+0xcc()
dmu_objset_sync+0x141()
dsl_dataset_sync+0x23()
dsl_pool_sync+0x7b()
spa_sync+0x116()
txg_sync_thread+0x115()
thread_start+8()

It appears ZFS is still able to read the labels from the drive:

$ zdb -lv  /dev/rdsk/c3t50002AC00039040Bd0p0
--------------------------------------------
LABEL 0
--------------------------------------------
    version=3
    name='fpool0'
    state=0
    txg=4
    pool_guid=10406529929620343615
    top_guid=3365726235666077346
    guid=3365726235666077346
    vdev_tree
        type='disk'
        id=0
        guid=3365726235666077346
        path='/dev/dsk/c3t50002AC00039040Bd0p0'
        devid='id1,[EMAIL PROTECTED]/q'
        whole_disk=0
        metaslab_array=13
        metaslab_shift=31
        ashift=9
        asize=322117566464
--------------------------------------------
LABEL 1
--------------------------------------------
    version=3
    name='fpool0'
    state=0
    txg=4
    pool_guid=10406529929620343615
    top_guid=3365726235666077346
    guid=3365726235666077346
    vdev_tree
        type='disk'
        id=0
        guid=3365726235666077346
        path='/dev/dsk/c3t50002AC00039040Bd0p0'
        devid='id1,[EMAIL PROTECTED]/q'
        whole_disk=0
        metaslab_array=13
        metaslab_shift=31
        ashift=9
        asize=322117566464
--------------------------------------------
LABEL 2
--------------------------------------------
    version=3
    name='fpool0'
    state=0
    txg=4
    pool_guid=10406529929620343615
    top_guid=3365726235666077346
    guid=3365726235666077346
    vdev_tree
        type='disk'
        id=0
        guid=3365726235666077346
        path='/dev/dsk/c3t50002AC00039040Bd0p0'
        devid='id1,[EMAIL PROTECTED]/q'
        whole_disk=0
        metaslab_array=13
        metaslab_shift=31
        ashift=9
        asize=322117566464
--------------------------------------------
LABEL 3
--------------------------------------------
    version=3
    name='fpool0'
    state=0
    txg=4
    pool_guid=10406529929620343615
    top_guid=3365726235666077346
    guid=3365726235666077346
    vdev_tree
        type='disk'
        id=0
        guid=3365726235666077346
        path='/dev/dsk/c3t50002AC00039040Bd0p0'
        devid='id1,[EMAIL PROTECTED]/q'
        whole_disk=0
        metaslab_array=13
        metaslab_shift=31
        ashift=9
        asize=322117566464

But for some reason it is unable to open the pool:

$ zdb -c fpool0
zdb: can't open fpool0: error 2

I saw several bugs related to space_map.c, but the stack traces listed
in the bug reports were different than the one listed above.  Has
anyone seen this bug before? Is there anyway to recover from it?

Thanks for any insight,
- Ryan
-- 
UNIX Administrator
http://prefetch.net
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to