Re: [ClusterLabs] crm services not getting started after upgrading to snmp40

2024-05-22 Thread ., Anoop
What is the "certain filesystem"? If cluster services require it, that would 
explain why they can't start.
 -  Here we have btrfs and xfs filesystems. Yes cluster services require these 
filesystem to be mounted.
What do the systemd journal logs say about the filesystem and cluster services? 
Did it try to start them at all?
 -  journalctl doesn't give you much input, it does try to start, but it fails. 
Another thing is if you boot with downgraded version of snmp (snmp30) all works 
fine



-Original Message-
From: Users  On Behalf Of Ken Gaillot
Sent: Wednesday, May 22, 2024 7:32 PM
To: Cluster Labs - All topics related to open-source clustering welcomed 

Subject: Re: [ClusterLabs] crm services not getting started after upgrading to 
snmp40

On Wed, 2024-05-22 at 07:33 +, ., Anoop wrote:
> Hello,
>  
> We have HA setup with 2 node cluster using CRM. OS is Suse 15sp3.
> After upgrading to snmp40, cluster services are not getting started 
> like pacemaker , corosync etc. After booting we have to manually mount 
> certain filesystem and start the crm services like pacemaker etc. We 
> have a SharedFileSystem group as the resource with 5 fileystems , but 
> not getting mounted while booting.  Let me know any other info 
> required.
>  
> Regards
> Anoop
> 

What is the "certain filesystem"? If cluster services require it, that would 
explain why they can't start.

What do the systemd journal logs say about the filesystem and cluster services? 
Did it try to start them at all?
--
Ken Gaillot 

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users 

ClusterLabs home: https://www.clusterlabs.org/ 
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] crm services not getting started after upgrading to snmp40

2024-05-22 Thread Ken Gaillot
On Wed, 2024-05-22 at 07:33 +, ., Anoop wrote:
> Hello,
>  
> We have HA setup with 2 node cluster using CRM. OS is Suse 15sp3.
> After upgrading to snmp40, cluster services are not getting started
> like pacemaker , corosync etc. After booting we have to manually
> mount certain filesystem and start the crm services like pacemaker
> etc. We have a SharedFileSystem group as the resource with 5
> fileystems , but not getting mounted while booting.  Let me know any
> other info required.
>  
> Regards
> Anoop
> 

What is the "certain filesystem"? If cluster services require it, that
would explain why they can't start.

What do the systemd journal logs say about the filesystem and cluster
services? Did it try to start them at all?
-- 
Ken Gaillot 

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] Disabled resources after parallel removing of group

2024-05-22 Thread Miroslav Lisik

Hi,
see comments inline.

On 5/17/24 17:46, Александр Руденко wrote:

Miroslav, thank you!

It helps me understand that it's not a configuration issue.

BTW, is it okay to create new resources in parallel?


Same as with parallel 'remove' operations it is not safe to do parallel
'create' operations, although it may work in some cases.

The 'pcs resource create' updates CIB by using CIB diffs and cibadmin's
'--patch' option, which is different from 'pcs resource remove', where
combination of '--replace' and '--delete' is used.

There is still risk that cib patch will not apply or something will
break due to parallel actions.

Do not use pcs command in parallel on live cluster, rather modify cib
file using pcs '-f' option and then push cib configuration to a cluster:
pcs cluster cib-push 
OR
pcs cluster cib-ppush  diff-against=

The difference in this two commands is in a method how cib update is
applied. The first command uses cibadmin's '--replace' option and the
second uses '--patch' option.


On timeline it looks like:

pcs resource create resA1  --group groupA
pcs resource create resB1  --group groupB
resA1 Started
pcs resource create resA2  --group groupA
res B1 Started
pcs resource create resB2  --group groupB
res A2 Started
res B2 Started

For now, it works okay)

In our case, cluster events like 'create' and 'remove' are generated by 
users, and for now we don't have any queue for operations. But now, I 
realized that we need a queue for 'remove' operations. Maybe we need a 
queue for 'create' operations to?


Yes, it is better to prevent users from doing modify operations at the
same time.



пт, 17 мая 2024 г. в 17:49, Miroslav Lisik >:


Hi Aleksandr!

It is not safe to use `pcs resource remove` command in parallel because
you run into the same issues as you already described. Processes run by
remove command are not synchronized.

Unfortunately, remove command does not support more than one resource
yet.

If you really need to remove resources at once you can use this method:
1. get the current cib configuration:
pcs cluster cib > original.xml

2. create a new copy of the file:
cp original.xml new.xml

3. disable all to be removed resources using -f option and new
configuration file:
pcs -f new.xml resource disable ...

4. remove resources using -f option and new configuration file:
pcs -f new.xml resource remove 
...

5. push new cib configuration to the cluster
pcs cluster cib-push new.xml diff-against=original.xml


On 5/17/24 13:47, Александр Руденко wrote:
 > Hi!
 >
 > I am new in the pacemaker world, and I, unfortunately, have problems
 > with simple actions like group removal. Please, help me
understand when
 > I'm wrong.
 >
 > For simplicity I will use standard resources like IPaddr2 (but we
have
 > this problem on any type of our custom resources).
 >
 > I have 5 groups like this:
 >
 > Full List of Resources:
 >    * Resource Group: group-1:
 >      * ip-11 (ocf::heartbeat:IPaddr2): Started vdc16
 >      * ip-12 (ocf::heartbeat:IPaddr2): Started vdc16
 >    * Resource Group: group-2:
 >      * ip-21 (ocf::heartbeat:IPaddr2): Started vdc17
 >      * ip-22 (ocf::heartbeat:IPaddr2): Started vdc17
 >    * Resource Group: group-3:
 >      * ip-31 (ocf::heartbeat:IPaddr2): Started vdc18
 >      * ip-32 (ocf::heartbeat:IPaddr2): Started vdc18
 >    * Resource Group: group-4:
 >      * ip-41 (ocf::heartbeat:IPaddr2): Started vdc16
 >      * ip-42 (ocf::heartbeat:IPaddr2): Started vdc16
 >
 > Groups were created by next simple script:
 > cat groups.sh
 > pcs resource create ip-11 ocf:heartbeat:IPaddr2 ip=10.7.1.11
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-1
 > pcs resource create ip-12 ocf:heartbeat:IPaddr2 ip=10.7.1.12
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-1
 >
 > pcs resource create ip-21 ocf:heartbeat:IPaddr2 ip=10.7.1.21
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-2
 > pcs resource create ip-22 ocf:heartbeat:IPaddr2 ip=10.7.1.22
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-2
 >
 > pcs resource create ip-31 ocf:heartbeat:IPaddr2 ip=10.7.1.31
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-3
 > pcs resource create ip-32 ocf:heartbeat:IPaddr2 ip=10.7.1.32
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-3
 >
 > pcs resource create ip-41 ocf:heartbeat:IPaddr2 ip=10.7.1.41
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-4
 > pcs resource create ip-42 ocf:heartbeat:IPaddr2 ip=10.7.1.42
 > cidr_netmask=24 nic=lo op monitor interval=10s --group group-4
 >
 > Next, i try to remove all of these group in 'parallel':
 > cat remove.sh
 > pcs 

[ClusterLabs] crm services not getting started after upgrading to snmp40

2024-05-22 Thread ., Anoop
Hello,

We have HA setup with 2 node cluster using CRM. OS is Suse 15sp3. After 
upgrading to snmp40, cluster services are not getting started like pacemaker , 
corosync etc. After booting we have to manually mount certain filesystem and 
start the crm services like pacemaker etc. We have a SharedFileSystem group as 
the resource with 5 fileystems , but not getting mounted while booting.  Let me 
know any other info required.

Regards
Anoop
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/