Thanks for the feedback ! Leandro. El lun., 2 mar. 2020 a las 5:23, Eneko Lacunza (<elacu...@binovo.es>) escribió:
> Hola Leandro, > > El 28/2/20 a las 15:25, Leandro Roggerone escribió: > > When you talk about redundacy and availability, do you want... > > - HA? (automatic restart of VMs in the other node in case one server > fails) > > - Be able to move VMs from one server to the other "fast" (without > > copying the disks)? > > Yes , this is what I want. > > So regarding my original layout from my 5.5Tb storage: > > Im using 1T for LVM , 1TB for LVM-thin and 3.5 TB unassigned space, is > it > > ok to use this unassigned space for a ceph ? > > Can I set it later ? with server on production ? > You can configure Ceph later, with the server in production, yes. > > But for Ceph you want single disks, no RAID. So I'd build the server > with a 2x2TB RAID1 and leave the other 2x2TB disks for adding to Ceph. > > Also, read a bit about Ceph before deploying, it is not "sinple" :-) > https://docs.ceph.com/docs/master/start/intro/ > > > Other: > > Using an NFS system, means to have an external server running a file > sistem? > > So you should have at least two servers for the cluster and one for the > > file system? > > It seems to me that using ceph has a better redundancy plan an it is > easier > > to deploy since I only need two servers. (am i right?). > For Ceph you need at least 3 servers; one of them can be a simple PC but > you need it for Ceph monitor's quorum. It is recommended too for Proxmox > cluster, for the same reason. > > Really a NFS based solution is simpler, but then NFS server is a simple > point of failure. Ceph will be more resilient, but you have to > understand how it works. You may find that having only two servers with > Ceph storage can be risky when performing maintenance on one of the > servers. > > Saludos > Eneko > > > Thanks! > > > > El vie., 28 feb. 2020 a las 11:06, Eneko Lacunza (<elacu...@binovo.es>) > > escribió: > > > >> Hola Leandro, > >> > >> El 28/2/20 a las 14:43, Leandro Roggerone escribió: > >>> Regarding your question , what is the tarjet use for this server. > >>> I have a dell R610 with 6 drive bays. > >>> Today I have 4 (2TB) drives in Raid5 , resulting a 5.5TB capacity. > >>> I will add 2 ssd drives later in raid1 for applications that need more > >> read > >>> speed. > >>> The purpose for this server is to run proxmox with some VMs for > external > >>> and internal access. > >>> Im planning to build a second server and create a cluster just to have > >> more > >>> redundancy and availability. > >>> > >>> I would like to set all I can now that server is not in production and > >>> minimize risk later. > >>> Thats why im asking so many questions. > >> Asking is good, but we need info to be able to help you ;) > >> > >> When you talk about redundacy and availability, do you want... > >> - HA? (automatic restart of VMs in the other node in case one server > fails) > >> - Be able to move VMs from one server to the other "fast" (without > >> copying the disks)? > >> > >> If your answer is yes to any of the previous questions, you have to look > >> at using a NFS server or deploying Ceph. > >> > >> If it's no, then we can talk about local storage in your servers. What > >> RAID card do you have in that server? Does it have write cache (non > >> volatile of battery-backed) If it doesn't have such, RAID5 could prove > >> slow (eat quite CPU), I suggest you use 2xRAID1 or a RAID10 setup. Also, > >> please bear in mind that RAID5 with "big" disks is considered quite > >> unsecure (risk of having a second disk failure during recovery is high). > >> > >> Saludos > >> Eneko > >>> Regards. > >>> Leandro. > >>> > >>> > >>> El vie., 28 feb. 2020 a las 5:49, Eneko Lacunza (<elacu...@binovo.es>) > >>> escribió: > >>> > >>>> Hola Leandro, > >>>> > >>>> El 27/2/20 a las 17:29, Leandro Roggerone escribió: > >>>>> Hi guys , i'm still tunning my 5.5 Tb server. > >>>>> While setting storage options during install process, I set 2000 for > hd > >>>>> size, so I have 3.5 TB free to assign later. > >>>>> > >>>>> my layout is as follows: > >>>>> root@pve:~# lsblk > >>>>> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT > >>>>> sda 8:0 0 5.5T 0 disk > >>>>> ├─sda1 8:1 0 1007K 0 part > >>>>> ├─sda2 8:2 0 512M 0 part > >>>>> └─sda3 8:3 0 2T 0 part > >>>>> ├─pve-swap 253:0 0 8G 0 lvm [SWAP] > >>>>> ├─pve-root 253:1 0 1T 0 lvm / > >>>>> ├─pve-data_tmeta 253:2 0 9G 0 lvm > >>>>> │ └─pve-data 253:4 0 949.6G 0 lvm > >>>>> └─pve-data_tdata 253:3 0 949.6G 0 lvm > >>>>> └─pve-data 253:4 0 949.6G 0 lvm > >>>>> sr0 11:0 1 1024M 0 rom > >>>>> > >>>>> My question is: > >>>>> Is it possible to expand sda3 partition later without service outage > ? > >>>>> Is it possible to expand pve group on sda3 partition ? > >>>> You don't need to expand sda3 really. You can just create a new > >>>> partition, create a new PV with it and add the new PV to pve VG. > >>>> > >>>>> In case to create a proxmox cluster, what should I do with that 3.5 > TB > >>>> free > >>>>> ? > >>>> I don't know really how to reply to this. If you're building a > cluster, > >>>> I suggest you configure some kind of shared storage; NFS server or > Ceph > >>>> cluster for example. > >>>> > >>>>> Is there a best partition type suited for this ? Can I do it without > >>>>> service outage? > >>>> For what? > >>>> > >>>>> I have not any service running yet , so I can experiment what it > takes. > >>>>> Any thought about this would be great. > >>>> Maybe you can start telling us your target use for this > server/cluster. > >>>> Also some detailed spec of the server would help; for example does it > >>>> have a RAID card with more than one disk, or you're using a 6TB single > >>>> disk? > >>>> > >>>> Cheers > >>>> Eneko > >>>> > >>>> -- > >>>> Zuzendari Teknikoa / Director Técnico > >>>> Binovo IT Human Project, S.L. > >>>> Telf. 943569206 > >>>> Astigarragako bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa) > >>>> www.binovo.es > >>>> > >>>> _______________________________________________ > >>>> pve-user mailing list > >>>> pve-user@pve.proxmox.com > >>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > >>>> > >>> _______________________________________________ > >>> pve-user mailing list > >>> pve-user@pve.proxmox.com > >>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > >> > >> -- > >> Zuzendari Teknikoa / Director Técnico > >> Binovo IT Human Project, S.L. > >> Telf. 943569206 > >> Astigarragako bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa) > >> www.binovo.es > >> > >> _______________________________________________ > >> pve-user mailing list > >> pve-user@pve.proxmox.com > >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > >> > > _______________________________________________ > > pve-user mailing list > > pve-user@pve.proxmox.com > > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > > > -- > Zuzendari Teknikoa / Director Técnico > Binovo IT Human Project, S.L. > Telf. 943569206 > Astigarragako bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa) > www.binovo.es > > _______________________________________________ > pve-user mailing list > pve-user@pve.proxmox.com > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > _______________________________________________ pve-user mailing list pve-user@pve.proxmox.com https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user