On 03/25/2013 09:48 PM, Dan McDonald wrote: > On Mon, Mar 25, 2013 at 09:43:31PM +0100, Sa?o Kiselkov wrote: >> On 03/25/2013 06:39 PM, Franz Schober wrote: >>> Hi, >> >> A bunch more things came to my mind, see comments below: >> >>> Test: Writing a 1 GB File >>> time dd if=/dev/zero of=/tmp/testfile1 bs=128k count=8k >>> Mean value of 3 tests on every system. >>> (I also tried IOZone but the simple dd test seems to show the problem) >> >> Make sure you write a lot more data per test run. Give it like 50-60GB >> at least. Btw: "/tmp" is tmpfs, which resides on your swap device, which >> is decided not a high-performance file system store. > > /tmp is memory first, and if there's no swap device, it uses available system > memory. > > I *thought* that's what the original poster was decrying, his memory > throughput, as measured via in-memory tmpfs.
I understand that that's what his issue was, however, I'm not aware of tmpfs having been designed with hundreds of MB/s of throughput in mind... If he wants to test raw memory throughput, why not simply created a ramdisk (ramdiskadm) and do a dd to it? Heck of a lot thinner in number of layers you need to traverse before you hit the VM layer. Plus, tmpfs isn't pre-allocated, it eats pages from the disk pagelist, possibly contending with the ZFS ARC. In short, if a VM test is needed, ramdisk is a lot simpler to manage than tmpfs. -- Saso ------------------------------------------- illumos-discuss Archives: https://www.listbox.com/member/archive/182180/=now RSS Feed: https://www.listbox.com/member/archive/rss/182180/21175430-2e6923be Modify Your Subscription: https://www.listbox.com/member/?member_id=21175430&id_secret=21175430-6a77cda4 Powered by Listbox: http://www.listbox.com