> See below. Right around 400,000 blocks, dedup is suddenly an order of > magnitude slower than without dedup. > > 400000 10.7sec 136.7sec 143 MB 195 MB > 800000 21.0sec 465.6sec 287 MB 391 MB
The interesting thing is - In all these cases, the complete DDT and the complete data file itself should fit entirely in ARC comfortably. So it makes no sense for performance to be so terrible at this level. So I need to start figuring out exactly what's going on. Unfortunately I don't know how to do that very well. I'm looking for advice from anyone - how to poke around and see how much memory is being consumed for what purposes. I know how to lookup c_min and c and c_max... But that didn't do me much good. The actual value for c barely changes at all over time... Even when I rm the file, c does not change immediately. All the other metrics from kstat ... have less than obvious names ... so I don't know what to look for... _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss