We were getting less perf with nfsv4 so we stuck with v3, but perhaps
you have other reasons for using v4...

On Mon, Aug 27, 2012 at 11:39 AM, Sabuj Pattanayek <sab...@gmail.com> wrote:
> Yes, that's what'll happen if you use sync. One way around using sync
> and getting decent write performance is to use rsize and wsize >=
> 128kb, but then you'll lose small file performance since you're moving
> larger blocks over the wire. The problem I've seen with all the EMC
> solutions I've tested is that NFSv3 client (read) data cache usage is
> broken unless you use sync or you set the EMC server to force NFS
> unstable writes to filesync or datasync rather than the default of an
> unstable NFS write (async).
>
> What I mean by broken client (read) cache usage is that if you write a
> file out to the server which fits in memory on the client (since the
> client buffers the write to local memory before sending it to the
> server), then if you try to read the file back and it's still in local
> client memory, the EMC server will do something with the client that
> causes it to read the entire file back again over the wire, rather
> than just getting it from local memory (really fast). Between Linux
> NFS servers and clients or other high end NAS solutions I've tested
> the client will read the file out of cache if it exists and if it's
> unchanged, which is the correct behavior.
>
> dd is actually a pretty good way of determining whether several NFS
> stacks are "broken" in this regard, I've written a nice wrapper around
> it https://code.google.com/p/nfsspeedtest/
>
> On Mon, Aug 27, 2012 at 9:19 AM, Gerhardus Geldenhuis
> <gerhardus.geldenh...@gmail.com> wrote:
>> Hi
>> I am debugging NFS performance problems at the moment and going through a
>> steep learning curve with regards to all its inner workings.
>>
>> In short we are mounting EMC provided NFS storage and seeing a massive
>> difference when mounting with and without the sync option.
>>
>> Without the sync option I can achieve 90MB/s without any tuning, with sync
>> option turned on that falls to 4MB/s. That seems to slow... any thoughts and
>> experience on this would be appreciated.
>> These values were obtained by doing a dd if=/dev/zero of=<file onnfs share>
>> bs=1M count=1024. I am aware that I need to optimize for specific use case
>> with regards to concurrency, file size, writes, rewrites, reads, rereads
>> etc. However the performance penalty for sync seems excessive and I wanted
>> to know if that is a shared experience.
>>
>> My understanding of sync is basically that in the case of NFS, it will wait
>> for a confirmation from the NFS server that the data has been written to
>> disk before acknowledging the write locally, so being "very certain". My
>> understanding could be flawed and I would appreciate anyone correcting me on
>> that and/or pointing me to more reading which I would happily do.
>>
>> I am a bit confused however about what the man pages are saying:
>> man mount says:
>> the sync option today has effect only for ext2, ext3, fat, vfat and ufs
>> but when I change it I can see a big difference in performance so the man
>> page is out of date... or is there something I am misreading?
>>
>> man 5 nfs, does not make mention of sync.
>>
>> OS: RHEL 5.8
>> Kernel: 2.6.18-308.4.1.0.1.el5
>> NFS:
>> nfs-utils-lib-1.0.8-7.9.el5
>> nfs-utils-1.0.9-60.0.1.el5
>> mount parameters:
>> nfs4
>> rw,bg,hard,nointr,rsize={16384,32768,65536},wsize={16384,32768,65536},suid,timeo=600,_netdev
>> 0 0
>>
>> I am currently running iozone and changing various mount parameters to see
>> how they effect performance. iozone is arguably a better test of performance
>> than dd
>>
>> Regards
>>
>> --
>> Gerhardus Geldenhuis
>>
>> _______________________________________________
>> rhelv5-list mailing list
>> rhelv5-list@redhat.com
>> https://www.redhat.com/mailman/listinfo/rhelv5-list

_______________________________________________
rhelv5-list mailing list
rhelv5-list@redhat.com
https://www.redhat.com/mailman/listinfo/rhelv5-list

Reply via email to