I've been struggling to fully understand why disk space seems to vanish. I've dug through bits of code and reviewed all the mails on the subject that I can find, but I still don't have a proper understanding of whats going on.
I did a test with a local zpool on snv_97... zfs list, zpool list, and zdb all seem to disagree on how much space is available. In this case its only a discrepancy of about 20G or so, but I've got Thumpers that have a discrepancy of over 6TB! Can someone give a really detailed explanation about whats going on? block traversal size 670225837056 != alloc 720394438144 (leaked 50168601088) bp count: 15182232 bp logical: 672332631040 avg: 44284 bp physical: 669020836352 avg: 44066 compression: 1.00 bp allocated: 670225837056 avg: 44145 compression: 1.00 SPA allocated: 720394438144 used: 96.40% Blocks LSIZE PSIZE ASIZE avg comp %Total Type 12 120K 26.5K 79.5K 6.62K 4.53 0.00 deferred free 1 512 512 1.50K 1.50K 1.00 0.00 object directory 3 1.50K 1.50K 4.50K 1.50K 1.00 0.00 object array 1 16K 1.50K 4.50K 4.50K 10.67 0.00 packed nvlist - - - - - - - packed nvlist size 72 8.45M 889K 2.60M 37.0K 9.74 0.00 bplist - - - - - - - bplist header - - - - - - - SPA space map header 974 4.48M 2.65M 7.94M 8.34K 1.70 0.00 SPA space map - - - - - - - ZIL intent log 96.7K 1.51G 389M 777M 8.04K 3.98 0.12 DMU dnode 17 17.0K 8.50K 17.5K 1.03K 2.00 0.00 DMU objset - - - - - - - DSL directory 13 6.50K 6.50K 19.5K 1.50K 1.00 0.00 DSL directory child map 12 6.00K 6.00K 18.0K 1.50K 1.00 0.00 DSL dataset snap map 14 38.0K 10.0K 30.0K 2.14K 3.80 0.00 DSL props - - - - - - - DSL dataset - - - - - - - ZFS znode 2 1K 1K 2K 1K 1.00 0.00 ZFS V0 ACL 5.81M 558G 557G 557G 95.8K 1.00 89.27 ZFS plain file 382K 301M 200M 401M 1.05K 1.50 0.06 ZFS directory 9 4.50K 4.50K 9.00K 1K 1.00 0.00 ZFS master node 12 482K 20.0K 40.0K 3.33K 24.10 0.00 ZFS delete queue 8.20M 66.1G 65.4G 65.8G 8.03K 1.01 10.54 zvol object 1 512 512 1K 1K 1.00 0.00 zvol prop - - - - - - - other uint8[] - - - - - - - other uint64[] - - - - - - - other ZAP - - - - - - - persistent error log 1 128K 10.5K 31.5K 31.5K 12.19 0.00 SPA history - - - - - - - SPA history offsets - - - - - - - Pool properties - - - - - - - DSL permissions - - - - - - - ZFS ACL - - - - - - - ZFS SYSACL - - - - - - - FUID table - - - - - - - FUID table size 5 3.00K 2.50K 7.50K 1.50K 1.20 0.00 DSL dataset next clones - - - - - - - scrub work queue 14.5M 626G 623G 624G 43.1K 1.00 100.00 Total real 21m16.862s user 0m36.984s sys 0m5.757s ======================================================= Looking at the data: [EMAIL PROTECTED] ~$ zfs list backup && zpool list backup NAME USED AVAIL REFER MOUNTPOINT backup 685G 237K 27K /backup NAME SIZE USED AVAIL CAP HEALTH ALTROOT backup 696G 671G 25.1G 96% ONLINE - So zdb says 626GB is used, zfs list says 685GB is used, and zpool list says 671GB is used. The pool was filled to 100% capacity via dd, this is confirmed, I can't write data, but yet zpool list says its only 96%. benr. -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss