Do we have enormous fragmentation here on our X4500 with Solaris 10, ZFS Version 10?
What except zfs send/receive can be done to free the fragmented space? One ZFS was used for some month to store some large disk images (each 50GByte large) which are copied there with rsync. This ZFS then reports 6.39TByte usage with zfs list and only 2TByte usage with du. The other ZFS was used for similar sized disk images, this time copied via NFS as whole files. On this ZFS du and zfs report exactly the same usage of 3.7TByte. bash-3.00# zfs list -r zpool1/vmwarersync NAME USED AVAIL REFER MOUNTPOINT zpool1/vmwarersync 6.39T 985G 6.39T /export/archiv/VMs/rsync bash-3.00# du -hs /export/archiv/VMs/rsync 2.0T /export/archiv/VMs/rsync bash-3.00# zfs list -r zpool1/vmwarevcb NAME USED AVAIL REFER MOUNTPOINT zpool1/vmwarevcb 3.75T 985G 3.75T /export/archiv/VMs/vcb bash-3.00# du -hs /export/archiv/VMs/vcb 3.7T /export/archiv/VMs/vcb bash-3.00# zpool upgrade This system is currently running ZFS pool version 10. bash-3.00# zpool status zpool1 pool: zpool1 state: ONLINE scrub: scrub completed after 14h2m with 0 errors on Thu Mar 4 10:22:47 2010 config: bash-3.00# zpool list zpool1 NAME SIZE USED AVAIL CAP HEALTH ALTROOT zpool1 20.8T 19.3T 1.53T 92% ONLINE - -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss