On 2/06/2016 6:10 AM, Jeremy McCoy wrote:
I am new here and am working on designing a Proxmox cluster. Just wondering if
anyone has tried doing Ceph on ZFS (instead of XFS on LVM or whatever pveceph
sets up) and, if so, how you went about implementing it. I have 4 hosts that
each have 1 spinner and 1 SSD to offer to the cluster.


Been there ... ceph does not like zfs, apparently COW filestems perform badly for cephs workload. Its recommended against (as is ext4 btw).

ceph does not do well on small setups, with only one osd/disk per node you will get *terrible* performance.

The 9.x versions of ceph have a latency bug that triggers under memory pressure (common on compute/storage nodes) and 10.x IMO is even less friendly to small setups. Also as you've probabluy noiced, its a maintenance headache for small shops, especially when things go wrong.



Are there any pitfalls to be aware of here? My goal is to mainly run LXC
containers (plus a few KVM VMs) on distributed storage, and I was hoping to take
advantage of ZFS's caching, compression, and data integrity features. I am also
open to doing GlusterFS or something else, but it looked like Proxmox does not
support LXC containers running on that yet.

Probably because lxc doesn't support the native gluster api (gfapi), I imagine the same problem with ceph/rbd.

However there is also the gluster fuse mount that proxmox automatic creates (/mnt/pve/<glusterid>), you should be able to set that up as shared directory storage and use that with lxc. I'll have a test of that myself later today.


gluster works pretty with with zfs, I get excellent performance, maxing out my network for writes and much better iops than I was getting with ceph. Enabling lz4 compression gave me a 33% saving on space with no noticeable impact on performance.

If you can afford it I recommend using ZFS RAID1 or better yet, ZFS RAID10 per node. Apart from the extra redundancy it had much better read/write performance than a single disk. And its much easier to replace a failed disk on a zfs mirror than it is to replace a failed gluster or ceph node.


--
Lindsay Mathieson

_______________________________________________
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user

Reply via email to