--- Begin Message ---
Hi Leandro,
El 13/9/21 a las 13:32, Leandro Roggerone escribió:
hi guys , your responses were very useful.
Lets suppose I have my 3 nodes running and forming a cluster.
Please confirm:
a -Can I add the ceph storage at any time ?
Yes
b- All nodes should be running the same pve version ?
Generally speaking this is advisable. What versions do you have right now?
c- All nodes should have 1 or more non used storages with no hardware raid
to be included in the ceph ?
It is advisable to have OSDs in at least 3 nodes yes (some may say 4 is
better).
Those storages (c) should be exactly same in capacity , speed , and so ...
?
Roughly speaking, Ceph will perform as well as the worst disk configured
for Ceph. If you plan to use SSD disks, use Enteprise SSD disk, not
consumer/client SSDs.
What can goes wrong if dont have 10 but 1 gbps ports ?
Latency and overall performance of Ceph storage will be worse/slower. If
you plan using 1G, consider setting up separate "cluster" ports for Ceph
(1G for VM traffic, 1G for ceph public, 1G for ceph cluster/private)
We have clusters with both 10G and 1G (3x1G) networks. All of them work
well but 10G network is quite noticeable, specially with SSD disks.
Cheers
Regards.
Leandro
<https://www.avast.com/sig-email?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=webmail>
Libre
de virus. www.avast.com
<https://www.avast.com/sig-email?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=webmail>
<#DAB4FAD8-2DD7-40BB-A1B8-4E2AA1F9FDF2>
El mié, 8 sept 2021 a las 19:21, ic (<[email protected]>) escribió:
Hi there,
On 8 Sep 2021, at 14:46, Leandro Roggerone <[email protected]>
wrote:
I would like to know the goods that a ceph storage can bring to my
existing
cluster.
What is an easy / recommended way to implement it ?
Wich hardware should I consider to use ?
First, HW.
Get two Cisco Nexus 3064PQ (they typically go for $600-700 for 48 10G
ports) and two Intel X520-DA2 per server.
Hook up each port of the Intel cards to each of the Nexuses, getting a
full redundancy between network cards and switches.
Add 4x40G DAC cables between the switches, setup 2 as VPC peer-links, 2 as
a simple L2 trunk (can provide more details as why if needed).
Use ports 0 from both NICs for ceph, ports 1 for VM traffic. This way you
get 2x10 Gbps for Ceph only and 2x10 Gbps for everything else, and if you
loose one card or one switch, you still have 10 Gbps for each.
The benefits? With default configuration, your data lives in 3 places.
Also, scale out. You know the expensive stuff, hyperconverged servers
(nutanix and such) ? You get that with this.
The performance is wild, just moved my customers from a proxmox cluster
backed by a TrueNAS server (full flash, 4x10Gbps) to a 3 node cluster of
AMD EPYC nodes with Ceph on local SATA SSDs and the VMs started flying.
Keep your old storage infrastructure, whatever that is, for backups with
PBS.
YMMV
Regards, ic
_______________________________________________
pve-user mailing list
[email protected]
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
_______________________________________________
pve-user mailing list
[email protected]
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user
Eneko Lacunza
Zuzendari teknikoa | Director técnico
Binovo IT Human Project
Tel. +34 943 569 206 | https://www.binovo.es
Astigarragako Bidea, 2 - 2º izda. Oficina 10-11, 20180 Oiartzun
https://www.youtube.com/user/CANALBINOVO
https://www.linkedin.com/company/37269706/
--- End Message ---
_______________________________________________
pve-user mailing list
[email protected]
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user