Hi Victor, Thanks for the prompt reply. Here are the results from your suggestions.
> Panic stack would be useful. I'm sorry I don't have this available and I don't want to cause another panic :) > > It is apparently blocked somewhere in kernel. Try to do something like this > from another window to get better idea: > > echo "<pid of zpool>::pid2proc|::walk thread|::findtsack -v" | mdb -k > echo "::threadlist -v" | mdb -k > window 1: bash-3.2# zpool import -f data1 window 2: bash-3.2# ps -fel | grep zpool 0 S root 897 874 0 40 20 ? 1262 ? 15:03:23 pts/3 0:00 zpool import -f data1 bash-3.2# echo "0t897::pid2proc|::walk thread|::findstack -v" | mdb -k stack pointer for thread ffffff01b57ab840: ffffff00083709f0 [ ffffff00083709f0 _resume_from_idle+0xf1() ] ffffff0008370a30 swtch+0x17f() ffffff0008370a60 cv_wait+0x61(ffffff01b3bd71ca, ffffff01b3bd7188) ffffff0008370ab0 txg_wait_synced+0x81(ffffff01b3bd7000, 299ee597) ffffff0008370b10 spa_config_update_common+0x79(ffffff01b42a8a80, 0, 0) ffffff0008370bb0 spa_import_common+0x36e(ffffff01b5ad4000, ffffff01b5325310, 0 , 0, 0) ffffff0008370be0 spa_import+0x1e(ffffff01b5ad4000, ffffff01b5325310, 0) ffffff0008370c30 zfs_ioc_pool_import+0xad(ffffff01b5ad4000) ffffff0008370cb0 zfsdev_ioctl+0x10d(b600000000, 5a02, 80424f0, 100003, ffffff01b3f181a0, ffffff0008370e9c) ffffff0008370cf0 cdev_ioctl+0x48(b600000000, 5a02, 80424f0, 100003, ffffff01b3f181a0, ffffff0008370e9c) ffffff0008370d30 spec_ioctl+0x86(ffffff01adf41d00, 5a02, 80424f0, 100003, ffffff01b3f181a0, ffffff0008370e9c, 0) ffffff0008370db0 fop_ioctl+0x7b(ffffff01adf41d00, 5a02, 80424f0, 100003, ffffff01b3f181a0, ffffff0008370e9c, 0) ffffff0008370ec0 ioctl+0x174(3, 5a02, 80424f0) ffffff0008370f10 _sys_sysenter_post_swapgs+0x14b() bash-3.2# echo "::threadlist -v" | mdb -k Output a bit too long to post here. Is there anything in particular i should look for in this output? > > Some useful information may be logged in FMA, try to see what is available > there with > > fmdump -eV - I get a few ereport.fs.zfs.checksum reports like this bash-3.2# fmdump -eV Aug 22 2008 15:03:23.203687016 ereport.fs.zfs.checksum nvlist version: 0 class = ereport.fs.zfs.checksum ena = 0x1a77b8287a00001 detector = (embedded nvlist) nvlist version: 0 version = 0x0 scheme = zfs pool = 0xe2bba51ab8c26b53 vdev = 0x79871af1e1f39de1 (end detector) pool = data1 pool_guid = 0xe2bba51ab8c26b53 pool_context = 0 pool_failmode = wait vdev_guid = 0x79871af1e1f39de1 vdev_type = disk vdev_path = /dev/dsk/c2t2d0s0 vdev_devid = id1,[EMAIL PROTECTED]/a parent_guid = 0xe2bba51ab8c26b53 parent_type = root zio_err = 50 zio_offset = 0x1e800416c00 zio_size = 0x400 zio_objset = 0x0 zio_object = 0xb zio_level = 0 zio_blkid = 0x0 __ttl = 0x1 Aug 22 2008 15:03:23.203687247 ereport.fs.zfs.data nvlist version: 0 class = ereport.fs.zfs.data ena = 0x1a77b8287a00001 detector = (embedded nvlist) nvlist version: 0 version = 0x0 scheme = zfs pool = 0xe2bba51ab8c26b53 (end detector) pool = data1 pool_guid = 0xe2bba51ab8c26b53 pool_context = 0 pool_failmode = wait zio_err = 50 zio_objset = 0x0 zio_object = 0xb zio_level = 0 zio_blkid = 0x0 __ttl = 0x1 __tod = 0x48aeb91b 0xc24054f Aug 22 2008 15:03:23.207225717 ereport.fs.zfs.io_failure nvlist version: 0 class = ereport.fs.zfs.io_failure ena = 0x1a77ee27cb00001 detector = (embedded nvlist) nvlist version: 0 version = 0x0 scheme = zfs pool = 0xe2bba51ab8c26b53 (end detector) pool = data1 pool_guid = 0xe2bba51ab8c26b53 pool_context = 0 pool_failmode = wait __ttl = 0x1 __tod = 0x48aeb91b 0xc5a0375 > > On Nevada you can try the following to (same option repeated several times > increases verbosity): > > zdb -e -bb data1 > zdb -e -dddd data1 > bash-3.2# zdb -e -bb data1 Traversing all blocks to verify nothing leaked ... out of memory -- generating core dump Abort Seems i need to get a machine with more ram to do this :) This can be arranged on monday. bash-3.2# zdb -e -dddd data1 Dataset mos [META], ID 0, cr_txg 4, 210M, 189 objects, rootbp [L0 DMU objset] 400L/200P DVA[0]=<0:56800000200:200> DVA[1]=<0:3000020200:200> DVA[2]=<0:48800001800:200> fletcher4 lzjb LE contiguous birth=698279317 fill=189 cksum=89744d6d8:36e7cf71f81:b1d06b2acd36:1850b4cc5621f3 Object lvl iblk dblk lsize asize type 0 2 16K 16K 96.0K 94.5K DMU dnode Object lvl iblk dblk lsize asize type 1 1 16K 512 512 1K object directory microzap: 512 bytes, 6 entries sync_bplist = 12 config = 11 root_dataset = 2 errlog_scrub = 0 deflate = 1 errlog_last = 0 Object lvl iblk dblk lsize asize type 2 1 16K 512 512 0 DSL directory 256 bonus DSL directory creation_time = Tue Mar 18 12:03:05 2008 head_dataset_obj = 5 parent_dir_obj = 0 origin_obj = 0 child_dir_zapobj = 4 used_bytes = 4.41T compressed_bytes = 4.40T uncompressed_bytes = 4.40T quota = 0 reserved = 0 props_zapobj = 3 deleg_zapobj = 0 Object lvl iblk dblk lsize asize type 3 1 16K 16K 32K 7.00K DSL props Fat ZAP stats: Pointer table: 1024 elements zt_blk: 0 zt_numblks: 0 zt_shift: 10 zt_blks_copied: 0 zt_nextblk: 0 ZAP entries: 3 Leaf blocks: 1 Total blocks: 2 zap_block_type: 0x8000000000000001 zap_magic: 0x2f52ab2ab zap_salt: 0x1d0e3ebe7 Leafs with 2^n pointers: 9: 1 * Blocks with n*5 entries: 0: 1 * Blocks n/10 full: 1: 1 * Entries with n chunks: 3: 3 *** Buckets with n entries: 0: 509 **************************************** 1: 3 * atime = 0 exec = 0 mountpoint = legacy Object lvl iblk dblk lsize asize type 4 1 16K 512 512 1K DSL directory child map microzap: 512 bytes, 1 entries $MOS = 8 Object lvl iblk dblk lsize asize type 5 1 16K 512 512 0 DSL dataset 320 bonus DSL dataset dir_obj = 2 prev_snap_obj = 0 prev_snap_txg = 0 next_snap_obj = 0 snapnames_zapobj = 6 num_children = 0 creation_time = Tue Mar 18 12:03:05 2008 creation_txg = 4 deadlist_obj = 7 used_bytes = 4.41T compressed_bytes = 4.40T uncompressed_bytes = 4.40T unique = 4.41T fsid_guid = 63153422916496006 guid = 3076616381815835747 flags = 0 next_clones_obj = 0 bp = [L0 DMU objset] 400L/200P DVA[0]=<0:56800000000:200> DVA[1]=<0:3000020000:200> DVA[2]=<0:48800000a00:200> fletcher4 lzjb LE contiguous birth=698279317 fill=6040564 cksum=f9e53509d:5f89f3d9404:12e97014634a8:294b54e7bf25c2 Object lvl iblk dblk lsize asize type 6 1 16K 512 512 1K DSL dataset snap map microzap: 512 bytes, 0 entries Object lvl iblk dblk lsize asize type 7 1 16K 128K 128K 0 bplist 32 bonus bplist header Object lvl iblk dblk lsize asize type 8 1 16K 512 512 0 DSL directory 256 bonus DSL directory creation_time = Tue Mar 18 12:03:05 2008 head_dataset_obj = 0 parent_dir_obj = 2 origin_obj = 0 child_dir_zapobj = 10 used_bytes = 210M compressed_bytes = 104M uncompressed_bytes = 104M quota = 0 reserved = 0 props_zapobj = 9 deleg_zapobj = 0 Object lvl iblk dblk lsize asize type 9 1 16K 512 512 1K DSL props microzap: 512 bytes, 0 entries Object lvl iblk dblk lsize asize type 10 1 16K 512 512 1K DSL directory child map microzap: 512 bytes, 0 entries Object lvl iblk dblk lsize asize type 11 1 16K 16K 16K 2K packed nvlist 8 bonus packed nvlist size Assertion failed: 0 == dmu_read(os, object, 0, nvsize, packed), file ../zdb.c, line 216, function dump_packed_nvlist > Btw, why does timestamp on your uberblock show July 1? Well, this is about the time when the crash happened. The clock on the server is correct. Once again, thank you for taking the time to look into this. cheers Erik _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss