We are running RHCS2.3 (jewel) with ganesha 2.4.2 and cephfs fsal, compiled
from srpm. experimenting with CTDB for controlling ganesha HA since we run
samba on same servers.

Haven't done much functionality/stress testing but on face value basic
stuff seems to work well (file operations).

In terms of performance, last time I tested ganesha it seemed comparable to
ceph-fuse (RHCS2.x/jewel, i think luminous ceph-fuse is better). Though I
haven't done rigorous metadata tests or multiple client tests. Also our
ganesha servers are quite small, as we are thus far only serving cephfs
natively. eg 4G ram 1 core. Here are some FIO results:

jobs in order are:
1. async 1M
2. sync 1M
3. async 4k
4. sync 4k
5. seq read 1M
6. rand read 4k

Ceph cluster is RHCS 2.3 (10.2.7)

CEPH-FUSE (10.2.x)
  WRITE: io=143652MB, aggrb=490328KB/s, minb=490328KB/s, maxb=490328KB/s,
mint=300002msec, maxt=300002msec
  WRITE: io=14341MB, aggrb=48947KB/s, minb=48947KB/s, maxb=48947KB/s,
mint=300018msec, maxt=300018msec
  WRITE: io=9808.2MB, aggrb=33478KB/s, minb=33478KB/s, maxb=33478KB/s,
mint=300001msec, maxt=300001msec
  WRITE: io=424476KB, aggrb=1414KB/s, minb=1414KB/s, maxb=1414KB/s,
mint=300003msec, maxt=300003ms
   READ: io=158069MB, aggrb=539527KB/s, minb=539527KB/s, maxb=539527KB/s,
mint=300008msec, maxt=300008msec
   READ: io=1881.2MB, aggrb=6420KB/s, minb=6420KB/s, maxb=6420KB/s,
mint=300001msec, maxt=300001msec

ganesha (nfs3)
  WRITE: io=157891MB, aggrb=538923KB/s, minb=538923KB/s, maxb=538923KB/s,
mint=300006msec, maxt=300006msec
  WRITE: io=38700MB, aggrb=132093KB/s, minb=132093KB/s, maxb=132093KB/s,
mint=300006msec, maxt=300006msec
  WRITE: io=3072.0MB, aggrb=10148KB/s, minb=10148KB/s, maxb=10148KB/s,
mint=309957msec, maxt=309957msec
  WRITE: io=397516KB, aggrb=1325KB/s, minb=1325KB/s, maxb=1325KB/s,
mint=300001msec, maxt=300001msec
   READ: io=82521MB, aggrb=281669KB/s, minb=281669KB/s, maxb=281669KB/s,
mint=300002msec, maxt=300002msec
   READ: io=1322.2MB, aggrb=4513KB/s, minb=4513KB/s, maxb=4513KB/s,
mint=300001msec, maxt=300001msec

cephfs kernel client
  WRITE: io=471041MB, aggrb=1568.8MB/s, minb=1568.8MB/s, maxb=1568.8MB/s,
mint=300394msec, maxt=300394msec
  WRITE: io=50005MB, aggrb=170680KB/s, minb=170680KB/s, maxb=170680KB/s,
mint=300006msec, maxt=300006msec
  WRITE: io=169092MB, aggrb=577166KB/s, minb=577166KB/s, maxb=577166KB/s,
mint=300000msec, maxt=300000msec
  WRITE: io=530548KB, aggrb=1768KB/s, minb=1768KB/s, maxb=1768KB/s,
mint=300003msec, maxt=300003msec
   READ: io=121501MB, aggrb=414720KB/s, minb=414720KB/s, maxb=414720KB/s,
mint=300002msec, maxt=300002msec
   READ: io=3264.6MB, aggrb=11142KB/s, minb=11142KB/s, maxb=11142KB/s,
mint=300001msec, maxt=300001msec

happy to share fio job file if anyone wants it.


On 9 November 2017 at 08:41, Sage Weil <sw...@redhat.com> wrote:

> Who is running nfs-ganesha's FSAL to export CephFS?  What has your
> experience been?
>
> (We are working on building proper testing and support for this into
> Mimic, but the ganesha FSAL has been around for years.)
>
> Thanks!
> sage
>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majord...@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>



-- 
*Rafael Lopez*
Research Devops Engineer
Monash University eResearch Centre

T: +61 3 9905 9118
M: +61 (0)427682670 <%2B61%204%2027682%20670>
E: rafael.lo...@monash.edu
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to