Hi Stijn, Yes, on my cluster I am running; CentOS 7, ZoL 0.6.3, Ceph 80.5.
Cheers -----Original Message----- From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Stijn De Weirdt Sent: October-29-14 3:49 PM To: ceph-users@lists.ceph.com Subject: Re: [ceph-users] use ZFS for OSDs hi michal, thanks for the info. we will certainly try it and see if we come to the same conclusions ;) one small detail: since you were using centos7, i'm assuming you were using ZoL 0.6.3? stijn On 10/29/2014 08:03 PM, Michal Kozanecki wrote: > Forgot to mention, when you create the ZFS/ZPOOL datasets, make sure > to set the xattar setting to sa > > e.g. > > zpool create osd01 -O xattr=sa -O compression=lz4 sdb > > OR if zpool/zfs dataset already created > > zfs set xattr=sa osd01 > > Cheers > > > > -----Original Message----- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf > Of Michal Kozanecki > Sent: October-29-14 11:33 AM > To: Kenneth Waegeman; ceph-users > Subject: Re: [ceph-users] use ZFS for OSDs > > Hi Kenneth, > > I run a small ceph test cluster using ZoL (ZFS on Linux) ontop of > CentOS 7, so I'll try and answer any questions. :) > > Yes, ZFS writeparallel support is there, but NOT compiled in by > default. You'll need to compile it with --with-zlib, but that by > itself will fail to compile the ZFS support as I found out. You need > to ensure you have ZoL installed and working, and then pass the > location of libzfs to ceph at compile time. Personally I just set my > environment variables before compiling like so; > > ldconfig > export LIBZFS_LIBS="/usr/include/libzfs/" > export LIBZFS_CFLAGS="-I/usr/include/libzfs -I/usr/include/libspl" > > However, the writeparallel performance isn't all that great. The > writeparallel mode makes heavy use of ZFS's (and BtrFS's for that matter) > snapshotting capability, and the snap performance on ZoL, at least when I > last tested it, is pretty terrible. You lose any performance benefits you > gain with writeparallel to the poor snap performance. > > If you decide that you don't need writeparallel mode you, can use the > prebuilt packages (or compile with default options) without issue. Ceph > (without zlib support compiled in) will detect ZFS as a generic/ext4 file > system and work accordingly. > > As far as performance tweaking, ZIL, write journals and etc, I found that the > performance difference between using a ZIL vs ceph write journal is about the > same. I also found that doing both (ZIL AND writejournal) didn't give me much > of a performance benefit. In my small test cluster I decided after testing to > forego the ZIL and only use a SSD backed ceph write journal on each OSD, with > each OSD being a single ZFS dataset/vdev(no zraid or mirroring). With Ceph > handling the redundancy at the OSD level I saw no need for using ZFS > mirroring or zraid, instead if ZFS detects corruption instead of self-healing > it sends a read failure of the pg file to ceph, and then ceph's scrub > mechanisms should then repair/replace the pg file using a good replica > elsewhere on the cluster. ZFS + ceph are a beautiful bitrot fighting match! > > Let me know if there's anything else I can answer. > > Cheers > > -----Original Message----- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf > Of Kenneth Waegeman > Sent: October-29-14 6:09 AM > To: ceph-users > Subject: [ceph-users] use ZFS for OSDs > > Hi, > > We are looking to use ZFS for our OSD backend, but I have some questions. > > My main question is: Does Ceph already supports the writeparallel mode for > ZFS ? (as described here: > http://www.sebastien-han.fr/blog/2013/12/02/ceph-performance-interesti > ng-things-going-on/) I've found this, but I suppose it is outdated: > https://wiki.ceph.com/Planning/Blueprints/Emperor/osd%3A_ceph_on_zfs > > Should Ceph be build with ZFS support? I found a --with-zfslib option > somewhere, but can someone verify this, or better has instructions for > it?:-) > > What parameters should be tuned to use this? > I found these : > filestore zfs_snap = 1 > journal_aio = 0 > journal_dio = 0 > > Are there other things we need for it? > > Many thanks!! > Kenneth > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com