Running a single node Proxmox "cluster", with Ceph on top. 1 Mon. Same node. I have 24 HDD (no dedicated journal) and 8 SSD split via "custom crush location hook". Cache-Tier (SSD-OSD) for a EC-pool (HDD-OSD) providing access for proxmox via krdb. 15 TB Capacity (Assortment of Disk sizes/speeds). Vdisks are Virtio and XFS. OSD's are XFS as well.
While setting up a virtual OpenmediaVault (VM) the following Question arose regarding vDisks (virtio) and their best practice. Q1: How does the amount and size of vDisks affect Write/Read performance? Do i bottleneck myself with overhead (single Mon)? Or does it maybe not matter at all? Values are academic examples. 120x 100GB vDisks - In OMV as Raid0 120x 100GB vDisks - In OMV as JBOD 12x 1TB vDisk - In OMV as Raid0 12x 1TB vDisk - In OMV as JBOD 2x 6TB vDisk - In OMV as Raid0 2x 6TB vDisk - In OMV as JBOD Q2: How does this best practice change if i add 2 more nodes (same config) and by implication 2 more mons? Not been able to find much on this topic. kind regards, Wolf F. _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com