Op 08/08/2024 om 09:53 schreef Rohit Yadav:
I think in CloudStack ceph storage pool you'll need to update its capacity if/when you increase the storage capacity after adding it initially. Wido and other Ceph gurus can advise other best practices.



Libvirt will automatically detect the increased capacity of a Ceph cluster when you add capacity.

There is no limit on how large a pool can be in Ceph and thus CloudStack. You can store multiple PB in a single Ceph pool. No need to split into smaller pools.

Wido

Regards.


------------------------------------------------------------------------
*From:* Muhammad Hanis Irfan Mohd Zaid <[email protected]>
*Sent:* Thursday, August 8, 2024 11:43
*To:* [email protected] <[email protected]>
*Cc:* Rohit Yadav <[email protected]>
*Subject:* Re: Best steps to deploy a working KVM cluster with RHEL
The gist that you shared plus your blog really helps me to set up CS on our Rocky Linux servers. It's now running great along with Ceph RBD (the other email) for primary and Ceph NFS for secondary.  Thanks Rohit!

Getting back to the pool size. Say I configured a Ceph RBD pool with no quota, this means I can expand it indefinitely without the need to split my Ceph storage into multiple pools of 6 TB based on the CS docs (best practice section).

I'll look at both Ceph and NFS limitations before actually designing the production cluster.

On Fri, 2 Aug 2024 at 20:46, Rohit Yadav <[email protected] <mailto:[email protected]>> wrote:

    Except the way you configure Linux bridges, using nmcli, on EL9 more
    or less all steps apply as earlier EL distros.

    I've my old notes here -
    
https://gist.github.com/rohityadavcloud/fc401a0fe8e8ea16b4b3a4e3d149ce0c#file-el9-or-rhel9-acs
 
<https://gist.github.com/rohityadavcloud/fc401a0fe8e8ea16b4b3a4e3d149ce0c#file-el9-or-rhel9-acs>

    CloudStack-Ceph users have easily used 100-1000s of TB of storage
    (RBD) in production, so that's fine. For NFS you can refer to the
    Ceph-NFS specific limitations (if any).


    Regards.




    ________________________________
    From: Muhammad Hanis Irfan Mohd Zaid <[email protected]
    <mailto:[email protected]>>
    Sent: Friday, August 2, 2024 07:54
    To: [email protected] <mailto:[email protected]>
    <[email protected] <mailto:[email protected]>>
    Subject: Re: Best steps to deploy a working KVM cluster with RHEL

    Oh say we're using Rocky Linux 9 or AlmaLinux 9, is there any workable
    steps that can be shared that works in production?

    We're going to be working mostly with 25G LACP bonded interfaces. And,
    we're planning to use Ceph RBD for primary and Ceph NFS for secondary
    storage. Does this mean provisioning more than 10 TB for both are
    okay with
    CloudStack?

    On Thu, 1 Aug 2024 at 14:50, Rohit Yadav <[email protected]
    <mailto:[email protected]>> wrote:

     > Hi Hanis,
     >
     > The docs may be a bit outdated and were originally written in
    scope for
     > XenServer - thanks for sharing that. It appears you're using KVM,
    so you
     > should look at the max-limitations and specific recommendations
    of your KVM
     > distro and NFS vendor.
     >
     > Majority of NFS datastore (both primary & secondary storage
    pools) there
     > days are in the 10s of TB in size/range, with even 100s of TBs
    also seen in
     > production usage.
     >
     > While using NFS, it's equally important to also consider networking
     > aspects such as switching capacity in the (KVM) cluster, the switch &
     > host-nic capabilities such as 1G, 10G, teaming/bond, LACP etc.
     >
     >
     > Regards.
     >
     >
     >
     >
     > ________________________________
     > From: Muhammad Hanis Irfan Mohd Zaid <[email protected]
    <mailto:[email protected]>>
     > Sent: Thursday, August 1, 2024 07:01
     > To: [email protected]
    <mailto:[email protected]> <[email protected]
    <mailto:[email protected]>>
     > Subject: Re: Best steps to deploy a working KVM cluster with RHEL
     >
     > Does anyone have thoughts on this?
     >
     >
     >
    
https://docs.cloudstack.apache.org/en/4.19.1.0/conceptsandterminology/choosing_deployment_architecture.html#best-practices
 
<https://docs.cloudstack.apache.org/en/4.19.1.0/conceptsandterminology/choosing_deployment_architecture.html#best-practices>
     >
     > Btw after reading that page, it looks like for primary storage
    the size
     > should be < 6 TB. What about secondary storage? Assumes both are
    using NFS.
     >
     > On Wed, 31 Jul 2024 at 16:52, Muhammad Hanis Irfan Mohd Zaid <
     > [email protected] <mailto:[email protected]>> wrote:
     >
     > > Hi CloudStack community!
     > >
     > > I'm currently testing out a POC with VLAN on our current
    vSphere cluster.
     > > As someone with a mostly VMware background, setting up each
    individual
     > KVM
     > > host and adding it to the CS management server is a bit of a
    hard task
     > for
     > > me. I've hit a few roadblocks and am hoping the community can
    assist me
     > in
     > > my journey. You can refer to the steps that I took to configure
    a KVM
     > node
     > > here: https://pastebin.com/MpSUq5mF <https://pastebin.com/MpSUq5mF>
     > >
     > > One of the issues that I'm having is that after the setup that
    I ran on
     > > the pastebin, an error occurred which I'm sure should be
    resolved with
     > > libvirtd sockets masking, which proved it's not. I've to reboot
    the host
     > > while the UI is still adding the host so it can be successfully
    added.
     > >
     > >
     > >
     > >
     > >
     > >
     > >
     > >
     > >
     > >
     > > *2024-07-30 03:56:37,871 INFO  [kvm.resource.LibvirtConnection]
     > > (main:null) (logid:) No existing libvirtd connection found.
    Opening a new
     > > one2024-07-30 03:56:38,109 ERROR [cloud.agent.AgentShell]
    (main:null)
     > > (logid:) Unable to start
     > > agent:com.cloud.utils.exception.CloudRuntimeException: Failed
    to connect
     > > socket to '/var/run/libvirt/virtqemud-sock': Connection
    refused        at
     > >
     >
    
com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.configure(LibvirtComputingResource.java:1153)
     > >       at com.cloud.agent.Agent.<init>(Agent.java:193)        at
> > com.cloud.agent.AgentShell.launchNewAgent(AgentShell.java:452)       at
     > >
    com.cloud.agent.AgentShell.launchAgentFromClassInfo(AgentShell.java:431)
     > >     at com.cloud.agent.AgentShell.launchAgent(AgentShell.java:415)
     > > at com.cloud.agent.AgentShell.start(AgentShell.java:511)        at
     > > com.cloud.agent.AgentShell.main(AgentShell.java:541)*
     > >
     > > Another issue that I'm having is that VNC doesn't work the
    first time.
     > > I've to do these steps to get VNC working for newly added hosts:
     > >
     > >    - Need to migrate a VM to a newly added host.
     > >    - Try to use VNC (doesn't work).
     > >    - Migrate it back out.
     > >    - Reboot the new host.
     > >    - Migrate the VM back into the new host.
     > >    - Try to use VNC (now it works).
     > >
     > >
     > > I humbly request, is there anyone that can share any steps that
    I can
     > > follow to deploy a POC or even production capable cluster for
    KVM running
     > > on RHEL-based OS or even Ubuntu. Thanks :)
     > >
     > >
     > >
     >

Reply via email to