> On Jun 1, 2016, at 1:25 PM, Gandalf Corvotempesta > <gandalf.corvotempe...@gmail.com> wrote: > > Il 01 giu 2016 22:06, "Gmail" <b.s.mikh...@gmail.com > <mailto:b.s.mikh...@gmail.com>> ha scritto: > > stat() on NFS, is just a single stat() from the client to the storage node, > > then all the storage nodes in the same replica group talk to each other > > using libgfapi (no FUSE overhead) > > > > conclusion, I’d prefer NFS over FUSE with small files. > > drawback, NFS HA is more complicated to setup and maintain than FUSE. > > NFS HA with ganesha should be easier than kernel NFS > > Skipping the whole fuse stack should be good also for big files > with big files, i don’t notice much difference in performance for NFS over FUSE > with nfs replication is made directly by gluster servers with no client > involved? > correct > In this case would be possibile to split the gluster networks with 10gb used > for replication and multiple 1gb bonded for clients. > don’t forget the complication of Ganesha HA setup, pacemaker is pain in the butt. > I can see only advantage for nfs over native gluster > > One question: with no gluster client that always know on which node a single > file is located, who is telling nfs where to find the required file? Is nfs > totally distributed with no "gateway"/"proxy" or any centralized server? > the NFS client talks to only one NFS server (the one which it mounts), the NFS HA setup is only to failover a virtual IP to another healthy node. so the NFS client will just do 3 minor timeouts then it will do a major timeout, when that happens, the virtual IP failover will be already done.
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users