Thanks to all guys for your feedback.

We think that the problem is hard to solve without damage a VM. We were trying with Gluster+NFS Ganesha+PaceMaker+Corosync and when the NFS goes down de IP floats to other node but it takes few seconds, and all VMs goes down and in this scenario the VM be Damaged. And the performance with gluster is not really good.

Now we want to test with ACS 4.16 and linstor, have someone any references about this?.

Regards

Mauro


El 20/10/2021 a las 05:44, Piotr Pisz escribió:
Hi,
I experienced this problem myself, in the KVM, Ceph, NFS-Ganesha environment at 
full Ceph load, the Ganesha NFS server was able to hang.
Servers were able to randomly restart due to lack of NFS access.
Which magnified the problem and there was a cascade and restart of the entire 
environment.
We currently have the restart line removed in kvmheartbeat, instead we report 
the restart attempt via prometheus.

Regards,
Piotr


-----Original Message-----
From: Sina Kashipazha <s.kashipa...@protonmail.com.INVALID>
Sent: Wednesday, October 20, 2021 10:35 AM
To: users@cloudstack.apache.org
Subject: Re: All cluster reboot when a Primary storage fails

Hey Daniel,

PR #4586 (https://github.com/apache/cloudstack/pull/4586) addressed your issue, 
as well. I'm currently working on it. Could you share with me how I can 
reproduce your reboot problem?

Kind regards,
Sina

‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐

On Saturday, October 16th, 2021 at 05:40, Daniel Augusto Veronezi Salvador 
<dvsalvador...@gmail.com> wrote:

Hi Mauro,

On KVM's monitor, when there is an inconsistency on the heartbeat's file

or heartbeat timeout is extrapolated several times, by default, the host

is restarted.

The PR 4586 (https://github.com/apache/cloudstack/pull/4586) already

addressed this issue by externalizing a property, which allows the

operator to decide if the host must be restarted or not (default is

'true', meaning that the host will be restarted). However, this feature

will be available only after release 4.16.

Best regards,

Daniel Salvador

On 15/10/2021 20:43, Mauro Ferraro - G2K Hosting wrote:

Hi guys, how are you?.

We are having this problems with ACS when a primary storages fails.

We have several primary storage with Linux and NFS server serving KVM

images. So every hosts have been mounted all the NFS servers because

in one Host can be running VMs from different storages. The main

problem of this, is when some storage fails because any reason all the

cluster gets crazy and start rebooting the hosts to reconnect with

this storage and all the VMs on the cluster, (including the VMs that

were working good) goes down becuase the conection to one storage fails.

If the problem with storage is permanent, the cluster never start

again and hosts will reboot indefinitely.

When this problem appears, the logs say this:

host heartbeat: kvmheartbeat.sh will reboot system because it was

unable to write the heartbeat to the storage.

Many users, edit the script kvmheartbeat.shto avoid the hosts reboot

or restart the agent on the host but i really not be sure that this is

the real solution.

Can someone help to propose a best solution at this high risk problem?.

Regards,

Mauro

Reply via email to