On 29.07.09 14:42, Pavel Kovalenko wrote:
fortunately, after several hours terminal went back -->
# zdb -e data1
Uberblock
magic = 0000000000bab10c
version = 6
txg = 2682808
guid_sum = 14250651627001887594
timestamp = 1247866318 UTC = Sat Jul 18 01:31:58 2009
Dataset mos [META], ID 0, cr_txg 4, 27.1M, 3050 objects
Dataset data1 [ZPL], ID 5, cr_txg 4, 5.74T, 52987 objects
capacity operations bandwidth ---- errors ----
description used avail read write read write read write cksum
data1 5.74T 6.99T 772 0 96.0M 0 0 0 91
/dev/dsk/c14t0d0 5.74T 6.99T 772 0 96.0M 0 0 0 223
#
So we know that there are some checksum errors there but at least zdb
was able to open pool in read-only mode.
i've tried to run zdb -e -t 2682807 data1
and
#echo "0t::pid2proc|::walk thread|::findstack -v" | mdb -k
This is wrong - you need to put PID of the 'zpool import data1' process
right after '0t'.
and
#fmdump -eV
shows checksum errors, such as
Jul 28 2009 11:17:35.386268381 ereport.fs.zfs.checksum
nvlist version: 0
class = ereport.fs.zfs.checksum
ena = 0x1baa23c52ce01c01
detector = (embedded nvlist)
nvlist version: 0
version = 0x0
scheme = zfs
pool = 0x578154df5f3260c0
vdev = 0x6e4327476e17daaa
(end detector)
pool = data1
pool_guid = 0x578154df5f3260c0
pool_context = 2
pool_failmode = wait
vdev_guid = 0x6e4327476e17daaa
vdev_type = disk
vdev_path = /dev/dsk/c14t0d0p0
vdev_devid = id1,s...@n2661000612646364/q
parent_guid = 0x578154df5f3260c0
parent_type = root
zio_err = 50
zio_offset = 0x2313d58000
zio_size = 0x4000
zio_objset = 0x0
zio_object = 0xc
zio_level = 0
zio_blkid = 0x0
__ttl = 0x1
__tod = 0x4a6ea60f 0x1705fcdd
This tells us that object 0xc in metabjset (objset 0x0) is corrupted.
So to get more details you can do the following:
zdb -e -dddd data1
zdb -e -bbcs data1
victor
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss