[ClusterLabs] Problem with a ping resource on CentOS 8

2021-08-27 Thread fatcharly
17 with HA-IP_211 (score:INFINITY) 
(id:colocation-HA-IP_217-HA-IP_211-INFINITY)
  HA-IP_218 with HA-IP_211 (score:INFINITY) 
(id:colocation-HA-IP_218-HA-IP_211-INFINITY)
  HA-IP_219 with HA-IP_211 (score:INFINITY) 
(id:colocation-HA-IP_219-HA-IP_211-INFINITY)
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
  No defaults set
Operations Defaults:
  No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: loadbalancer
 dc-version: 2.0.5-9.el8_4.1-ba59be7122
 have-watchdog: false
 stonith-enabled: false

Tags:
 No tags defined

Quorum:
  Options:


Any suggestions are welcome

Best regards and stay safe

fatcharly


___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?

2021-05-18 Thread fatcharly



> Gesendet: Dienstag, 18. Mai 2021 um 14:49 Uhr
> Von: fatcha...@gmx.de
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
>
> Hi Andrei,Hi everybody,
>
> ...
> > > and it works great Thanks for the hint.
> > > But the thing I still don't understand is why the cluster demotes is 
> > > active node for a short time when I reenable a node from standby back to 
> > > unstandby ? Is it not possible to join the drbd as secondary without 
> > > demote the primary for a short moment ?
> >
> > Try adding interleave=true to your clones.
>
> I tried this but it get me an error msg, what is wrong ?
>
>  pcs resource update database_drbd ocf:linbit:drbd drbd_resource=drbd1 
> promotable promoted-max=1 promoted-node-max=1 clone-max=2 clone-node-max=1 
> notify=true interleave=true
>
> Error: invalid resource options: 'clone-max', 'clone-node-max', 'interleave', 
> 'notify', 'promoted-max', 'promoted-node-max', allowed options are: 
> 'adjust_master_score', 'connect_only_after_promote', 'drbd_resource', 
> 'drbdconf', 'fail_promote_early_if_peer_primary', 
> 'ignore_missing_notifications', 'remove_master_score_if_peer_primary', 
> 'require_drbd_module_version_ge', 'require_drbd_module_version_lt', 
> 'stop_outdates_secondary', 'unfence_extra_args', 'unfence_if_all_uptodate', 
> 'wfc_timeout', use --force to override

or is it simply:
pcs resource update database_drbd-clone interleave=true ?


>
> Any suggestions are welcome
>
> Stay safe and take care
>
> fatcharly
>
>
>
>
> > Gesendet: Mittwoch, 12. Mai 2021 um 19:04 Uhr
> > Von: "Andrei Borzenkov" 
> > An: users@clusterlabs.org
> > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
> >
> > On 12.05.2021 17:34, fatcha...@gmx.de wrote:
> > > Hi Andrei, Hi everybody,
> > >
> > >
> > >> Gesendet: Mittwoch, 12. Mai 2021 um 16:01 Uhr
> > >> Von: fatcha...@gmx.de
> > >> An: users@clusterlabs.org
> > >> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
> > >>
> > >> Hi Andrei, Hi everybody,
> > >>
> > >>
> > >>> You need order fs_database after promote operation; and as I just found
> > >>> pacemaker also does not reverse it correctly and executes fs stop and
> > >>> drbd demote concurrently. So you need additional order constraint to
> > >>> first stop fs then demote drbd.
> > >>
> > >> is there so good doku about this, I don't know how to archive a "after 
> > >> promote operation" and how can I tell the pcs to first dismount the 
> > >> filesystem mountpoint and then demote the drbd-device.
> > >>
> > > ok, so I found something and used this:
> > >
> > > pcs constraint order stop fs_logfiles then demote drbd_logsfiles-clone
> > > pcs constraint order stop fs_database then demote database_drbd-clone
> > >
> > > and it works great Thanks for the hint.
> > > But the thing I still don't understand is why the cluster demotes is 
> > > active node for a short time when I reenable a node from standby back to 
> > > unstandby ? Is it not possible to join the drbd as secondary without 
> > > demote the primary for a short moment ?
> >
> > Try adding interleave=true to your clones.
> >
> > >
> > > Best regards and take care
> > >
> > > fatcharly
> > >
> > >
> > >
> > >> Sorry but this is new for me.
> > >>
> > >> Best regards and take care
> > >>
> > >> fatcharly
> > >>
> > >>
> > >>
> > >>
> > >>> Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr
> > >>> Von: "Andrei Borzenkov" 
> > >>> An: users@clusterlabs.org
> > >>> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems 
> > >>> ?
> > >>>
> > >>> On 11.05.2021 17:43, fatcha...@gmx.de wrote:
> > >>>> Hi,
> > >>>>
> > >>>> I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + 
> > >>>> corosync-3.0.3-4.el8.x86_64 and 
> > >>>> kmod-drbd90-9.0.25-2.el8_3.elrepo

Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?

2021-05-18 Thread fatcharly
Hi Andrei,Hi everybody,

...
> > and it works great Thanks for the hint.
> > But the thing I still don't understand is why the cluster demotes is active 
> > node for a short time when I reenable a node from standby back to unstandby 
> > ? Is it not possible to join the drbd as secondary without demote the 
> > primary for a short moment ?
>
> Try adding interleave=true to your clones.

I tried this but it get me an error msg, what is wrong ?

 pcs resource update database_drbd ocf:linbit:drbd drbd_resource=drbd1 
promotable promoted-max=1 promoted-node-max=1 clone-max=2 clone-node-max=1 
notify=true interleave=true

Error: invalid resource options: 'clone-max', 'clone-node-max', 'interleave', 
'notify', 'promoted-max', 'promoted-node-max', allowed options are: 
'adjust_master_score', 'connect_only_after_promote', 'drbd_resource', 
'drbdconf', 'fail_promote_early_if_peer_primary', 
'ignore_missing_notifications', 'remove_master_score_if_peer_primary', 
'require_drbd_module_version_ge', 'require_drbd_module_version_lt', 
'stop_outdates_secondary', 'unfence_extra_args', 'unfence_if_all_uptodate', 
'wfc_timeout', use --force to override

Any suggestions are welcome

Stay safe and take care

fatcharly




> Gesendet: Mittwoch, 12. Mai 2021 um 19:04 Uhr
> Von: "Andrei Borzenkov" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
>
> On 12.05.2021 17:34, fatcha...@gmx.de wrote:
> > Hi Andrei, Hi everybody,
> >
> >
> >> Gesendet: Mittwoch, 12. Mai 2021 um 16:01 Uhr
> >> Von: fatcha...@gmx.de
> >> An: users@clusterlabs.org
> >> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
> >>
> >> Hi Andrei, Hi everybody,
> >>
> >>
> >>> You need order fs_database after promote operation; and as I just found
> >>> pacemaker also does not reverse it correctly and executes fs stop and
> >>> drbd demote concurrently. So you need additional order constraint to
> >>> first stop fs then demote drbd.
> >>
> >> is there so good doku about this, I don't know how to archive a "after 
> >> promote operation" and how can I tell the pcs to first dismount the 
> >> filesystem mountpoint and then demote the drbd-device.
> >>
> > ok, so I found something and used this:
> >
> > pcs constraint order stop fs_logfiles then demote drbd_logsfiles-clone
> > pcs constraint order stop fs_database then demote database_drbd-clone
> >
> > and it works great Thanks for the hint.
> > But the thing I still don't understand is why the cluster demotes is active 
> > node for a short time when I reenable a node from standby back to unstandby 
> > ? Is it not possible to join the drbd as secondary without demote the 
> > primary for a short moment ?
>
> Try adding interleave=true to your clones.
>
> >
> > Best regards and take care
> >
> > fatcharly
> >
> >
> >
> >> Sorry but this is new for me.
> >>
> >> Best regards and take care
> >>
> >> fatcharly
> >>
> >>
> >>
> >>
> >>> Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr
> >>> Von: "Andrei Borzenkov" 
> >>> An: users@clusterlabs.org
> >>> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
> >>>
> >>> On 11.05.2021 17:43, fatcha...@gmx.de wrote:
> >>>> Hi,
> >>>>
> >>>> I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + 
> >>>> corosync-3.0.3-4.el8.x86_64 and kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64.
> >>>> The cluster consists of two nodes which are providing a ha-mariadb with 
> >>>> the help of two drbd devices for the database and the logfiles. The 
> >>>> corosync is working over two rings and both machines are virtual 
> >>>> kvm-guests.
> >>>>
> >>>> Problem:
> >>>> Node susanne is the active node and lisbon is changing from standby to 
> >>>> active, susanna is trying to demote one drbd-device but is failling to. 
> >>>> The cluster is working on properly, but the error stays.
> >>>> This is the what happens:
> >>>>
> >>>> Cluster Summary:
> >>>>   * Stack: corosync
> >>>>   * Current DC:

Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?

2021-05-12 Thread fatcharly
Hi Andrei, Hi everybody,


> Gesendet: Mittwoch, 12. Mai 2021 um 16:01 Uhr
> Von: fatcha...@gmx.de
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
>
> Hi Andrei, Hi everybody,
>
>
> > You need order fs_database after promote operation; and as I just found
> > pacemaker also does not reverse it correctly and executes fs stop and
> > drbd demote concurrently. So you need additional order constraint to
> > first stop fs then demote drbd.
>
> is there so good doku about this, I don't know how to archive a "after 
> promote operation" and how can I tell the pcs to first dismount the 
> filesystem mountpoint and then demote the drbd-device.
>
ok, so I found something and used this:

pcs constraint order stop fs_logfiles then demote drbd_logsfiles-clone
pcs constraint order stop fs_database then demote database_drbd-clone

and it works great Thanks for the hint.
But the thing I still don't understand is why the cluster demotes is active 
node for a short time when I reenable a node from standby back to unstandby ? 
Is it not possible to join the drbd as secondary without demote the primary for 
a short moment ?

Best regards and take care

fatcharly



> Sorry but this is new for me.
>
> Best regards and take care
>
> fatcharly
>
>
>
>
> > Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr
> > Von: "Andrei Borzenkov" 
> > An: users@clusterlabs.org
> > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
> >
> > On 11.05.2021 17:43, fatcha...@gmx.de wrote:
> > > Hi,
> > >
> > > I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + 
> > > corosync-3.0.3-4.el8.x86_64 and kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64.
> > > The cluster consists of two nodes which are providing a ha-mariadb with 
> > > the help of two drbd devices for the database and the logfiles. The 
> > > corosync is working over two rings and both machines are virtual 
> > > kvm-guests.
> > >
> > > Problem:
> > > Node susanne is the active node and lisbon is changing from standby to 
> > > active, susanna is trying to demote one drbd-device but is failling to. 
> > > The cluster is working on properly, but the error stays.
> > > This is the what happens:
> > >
> > > Cluster Summary:
> > >   * Stack: corosync
> > >   * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition 
> > > with quo rum
> > >   * Last updated: Tue May 11 16:15:54 2021
> > >   * Last change:  Tue May 11 16:15:42 2021 by root via cibadmin on susanne
> > >   * 2 nodes configured
> > >   * 11 resource instances configured
> > >
> > > Node List:
> > >   * Online: [ lisbon susanne ]
> > >
> > > Active Resources:
> > >   * HA_IP   (ocf::heartbeat:IPaddr2):Started susanne
> > >   * Clone Set: database_drbd-clone [database_drbd] (promotable):
> > > * Masters: [ susanne ]
> > > * Slaves: [ lisbon ]
> > >   * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable):
> > > * drbd_logsfiles(ocf::linbit:drbd):  Demoting susanne
> > >   * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
> >
> > Presumably fs_logfiles is located on drbd_logfiles, so how comes it is
> > active while drbd_logfiles is being demoted? Then drbdadm fails to
> > change status to secondary and RA simply loops forever until timeout.
> >
> > >   * fs_database (ocf::heartbeat:Filesystem): Started susanne
> > >   * mysql-server(ocf::heartbeat:mysql):  Started susanne
> > >   * Clone Set: ping_fw-clone [ping_fw]:
> > > * Started: [ lisbon susanne ]
> > >
> > > ---
> > > after a few seconds it switches over:
> > >
> > > Cluster Summary:
> > >   * Stack: corosync
> > >   * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition 
> > > with quo rum
> > >   * Last updated: Tue May 11 16:17:59 2021
> > >   * Last change:  Tue May 11 16:15:42 2021 by root via cibadmin on susanne
> > >   * 2 nodes configured
> > >   * 11 resource instances configured
> > >
> > > Node List:
> > >   * Online: [ lisbon susanne ]
> > >
> > > Active Resources:
> > >   * HA_IP   (ocf::heartbeat:IPaddr2):Started susanne
> > >   * Clone Set: database_drbd-clone [database_drbd] (prom

Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?

2021-05-12 Thread fatcharly
Hi Andrei, Hi everybody,


> You need order fs_database after promote operation; and as I just found
> pacemaker also does not reverse it correctly and executes fs stop and
> drbd demote concurrently. So you need additional order constraint to
> first stop fs then demote drbd.

is there so good doku about this, I don't know how to archive a "after promote 
operation" and how can I tell the pcs to first dismount the filesystem 
mountpoint and then demote the drbd-device.

Sorry but this is new for me.

Best regards and take care

fatcharly




> Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr
> Von: "Andrei Borzenkov" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
>
> On 11.05.2021 17:43, fatcha...@gmx.de wrote:
> > Hi,
> >
> > I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + 
> > corosync-3.0.3-4.el8.x86_64 and kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64.
> > The cluster consists of two nodes which are providing a ha-mariadb with the 
> > help of two drbd devices for the database and the logfiles. The corosync is 
> > working over two rings and both machines are virtual kvm-guests.
> >
> > Problem:
> > Node susanne is the active node and lisbon is changing from standby to 
> > active, susanna is trying to demote one drbd-device but is failling to. The 
> > cluster is working on properly, but the error stays.
> > This is the what happens:
> >
> > Cluster Summary:
> >   * Stack: corosync
> >   * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition 
> > with quo rum
> >   * Last updated: Tue May 11 16:15:54 2021
> >   * Last change:  Tue May 11 16:15:42 2021 by root via cibadmin on susanne
> >   * 2 nodes configured
> >   * 11 resource instances configured
> >
> > Node List:
> >   * Online: [ lisbon susanne ]
> >
> > Active Resources:
> >   * HA_IP   (ocf::heartbeat:IPaddr2):Started susanne
> >   * Clone Set: database_drbd-clone [database_drbd] (promotable):
> > * Masters: [ susanne ]
> > * Slaves: [ lisbon ]
> >   * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable):
> > * drbd_logsfiles(ocf::linbit:drbd):  Demoting susanne
> >   * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
>
> Presumably fs_logfiles is located on drbd_logfiles, so how comes it is
> active while drbd_logfiles is being demoted? Then drbdadm fails to
> change status to secondary and RA simply loops forever until timeout.
>
> >   * fs_database (ocf::heartbeat:Filesystem): Started susanne
> >   * mysql-server(ocf::heartbeat:mysql):  Started susanne
> >   * Clone Set: ping_fw-clone [ping_fw]:
> > * Started: [ lisbon susanne ]
> >
> > ---
> > after a few seconds it switches over:
> >
> > Cluster Summary:
> >   * Stack: corosync
> >   * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition 
> > with quo rum
> >   * Last updated: Tue May 11 16:17:59 2021
> >   * Last change:  Tue May 11 16:15:42 2021 by root via cibadmin on susanne
> >   * 2 nodes configured
> >   * 11 resource instances configured
> >
> > Node List:
> >   * Online: [ lisbon susanne ]
> >
> > Active Resources:
> >   * HA_IP   (ocf::heartbeat:IPaddr2):Started susanne
> >   * Clone Set: database_drbd-clone [database_drbd] (promotable):
> > * Masters: [ susanne ]
> > * Slaves: [ lisbon ]
> >   * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable):
> > * Masters: [ susanne ]
> > * Slaves: [ lisbon ]
> >   * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
> >   * fs_database (ocf::heartbeat:Filesystem): Started susanne
> >   * mysql-server(ocf::heartbeat:mysql):  Started susanne
> >   * Resource Group: apache:
> > * httpd_srv (ocf::heartbeat:apache): Started susanne
> >   * Clone Set: ping_fw-clone [ping_fw]:
> > * Started: [ lisbon susanne ]
> >
> > Failed Resource Actions:
> >   * drbd_logsfiles_demote_0 on susanne 'error' (1): call=736, status='Timed 
> > Out'
> > , exitreason='', last-rc-change='2021-05-11 16:15:42 +02:00', queued=0ms, 
> > exec=9 0001ms
> > --
> >
>
> And what you see in logs?
>
> > I think it is a constraint-problem, but I can't find it.
> > 

[ClusterLabs] 2 node mariadb-cluster - constraint-problems ?

2021-05-11 Thread fatcharly
l-10)
   notify interval=0s timeout=90 (drbd_logsfiles-notify-interval-0s)
   promote interval=0s timeout=90 
(drbd_logsfiles-promote-interval-0s)
   reload interval=0s timeout=30 (drbd_logsfiles-reload-interval-0s)
   start interval=0s timeout=240 (drbd_logsfiles-start-interval-0s)
   stop interval=0s timeout=100 (drbd_logsfiles-stop-interval-0s)
 Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem)
  Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4
  Operations: monitor interval=20s timeout=40s 
(fs_logfiles-monitor-interval-20s)
  start interval=0s timeout=60s (fs_logfiles-start-interval-0s)
  stop interval=0s timeout=60s (fs_logfiles-stop-interval-0s)
 Resource: fs_database (class=ocf provider=heartbeat type=Filesystem)
  Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4
  Operations: monitor interval=20s timeout=40s 
(fs_database-monitor-interval-20s)
  start interval=0s timeout=60s (fs_database-start-interval-0s)
  stop interval=0s timeout=60s (fs_database-stop-interval-0s)
 Resource: mysql-server (class=ocf provider=heartbeat type=mysql)
  Attributes: additional_parameters=--bind-address=0.0.0.0 
binary=/usr/bin/mysqld_safe config=/etc/my.cnf datadir=/mnt/clusterfs1/mysql 
pid=/var/lib/mysql/run/mariadb.pid socket=/var/lib/mysql/mysql.sock
  Operations: demote interval=0s timeout=120s (mysql-server-demote-interval-0s)
  monitor interval=20s timeout=30s 
(mysql-server-monitor-interval-20s)
  notify interval=0s timeout=90s (mysql-server-notify-interval-0s)
  promote interval=0s timeout=120s 
(mysql-server-promote-interval-0s)
  start interval=0s timeout=60s (mysql-server-start-interval-0s)
  stop interval=0s timeout=60s (mysql-server-stop-interval-0s)
 Group: apache
  Resource: httpd_srv (class=ocf provider=heartbeat type=apache)
   Attributes: configfile=/etc/httpd/conf/httpd.conf 
statusurl=http://127.0.0.1/server-status
   Operations: monitor interval=10s timeout=20s (httpd_srv-monitor-interval-10s)
   start interval=0s timeout=40s (httpd_srv-start-interval-0s)
   stop interval=0s timeout=60s (httpd_srv-stop-interval-0s)
 Clone: ping_fw-clone
  Resource: ping_fw (class=ocf provider=pacemaker type=ping)
   Attributes: dampen=10s host_list=192.168.18.1 multiplier=1000
   Operations: monitor interval=10s timeout=60s (ping_fw-monitor-interval-10s)
   start interval=0s timeout=60s (ping_fw-start-interval-0s)
   stop interval=0s timeout=20s (ping_fw-stop-interval-0s)

Stonith Devices:
Fencing Levels:

Location Constraints:
  Resource: mysql-server
Constraint: location-mysql-server
  Rule: boolean-op=or score=-INFINITY (id:location-mysql-server-rule)
Expression: pingd lt 1 (id:location-mysql-server-rule-expr)
Expression: not_defined pingd (id:location-mysql-server-rule-expr-1)
Ordering Constraints:
  start mysql-server then start httpd_srv (kind:Mandatory) 
(id:order-mysql-server-httpd_srv-mandatory)
  start database_drbd-clone then start drbd_logsfiles-clone (kind:Mandatory) 
(id:order-database_drbd-clone-drbd_logsfiles-clone-mandatory)
  start drbd_logsfiles-clone then start fs_database (kind:Mandatory) 
(id:order-drbd_logsfiles-clone-fs_database-mandatory)
  start fs_database then start fs_logfiles (kind:Mandatory) 
(id:order-fs_database-fs_logfiles-mandatory)
  start fs_logfiles then start mysql-server (kind:Mandatory) 
(id:order-fs_logfiles-mysql-server-mandatory)
Colocation Constraints:
  fs_logfiles with drbd_logsfiles-clone (score:INFINITY) (with-rsc-role:Master) 
(id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY)
  fs_database with database_drbd-clone (score:INFINITY) (with-rsc-role:Master) 
(id:colocation-fs_database-database_drbd-clone-INFINITY)
  drbd_logsfiles-clone with database_drbd-clone (score:INFINITY) 
(rsc-role:Master) (with-rsc-role:Master) 
(id:colocation-drbd_logsfiles-clone-database_drbd-clone-INFINITY)
  HA_IP with database_drbd-clone (score:INFINITY) (rsc-role:Started) 
(with-rsc-role:Master) (id:colocation-HA_IP-database_drbd-clone-INFINITY)
  mysql-server with fs_database (score:INFINITY) 
(id:colocation-mysql-server-fs_database-INFINITY)
  httpd_srv with mysql-server (score:INFINITY) 
(id:colocation-httpd_srv-mysql-server-INFINITY)
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
  No defaults set
Operations Defaults:
  No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: mysql_cluster
 dc-version: 2.0.4-6.el8_3.1-2deceaa3ae
 have-watchdog: false
 last-lrm-refresh: 1620742514
 stonith-enabled: FALSE

Tags:
 No tags defined

Quorum:
  Options:





Any suggestions are welcome

best regards stay safe, take care

fatcharly

___
Manage your subscription:
https://lists.clusterlabs.o

Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started

2021-01-19 Thread fatcharly
Hi Brent,

I now use the drbd without starting the drbd service. All is managed by the 
pcs, everything gets mounted and unmounted as expected.
In the beginning I was dealing with a lot of problems caused by the selinux and 
a wrong used constrain.
Now everything works as I know it from my other cluster which are build on 
CentOS 5/6/7.
Now in the beginning I also work with a disabled firewall because the system is 
in a secured dmz.
Is there anything I can do for you ?

best regards and stay healthy

fatcharly




> Gesendet: Montag, 18. Januar 2021 um 22:16 Uhr
> Von: "Brent Jensen" 
> An: fatcha...@gmx.de
> Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started
>
> Are you getting the the cluster to switch over when doing a 'pcs node
> standby' on the promoted node? I have a super basic config w/ your same
> software versions and I cannot get the slave to promote (constantly
> looping errors such as "Refusing to be Primary while peer is not
> outdated", until I do a drbdadm up  on the standby node). I
> appreciate your input. Brent
>
> On 1/18/2021 9:49 AM, fatcha...@gmx.de wrote:
> > Sorry Guys,
> >
> > problem found, it was a colocation constrain of the drbd-devices.
> >
> > best regards
> >
> > fatcharly
> >
> >
> >
> >> Gesendet: Montag, 18. Januar 2021 um 16:52 Uhr
> >> Von: fatcha...@gmx.de
> >> An: "clusterlabs" 
> >> Betreff: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started
> >>
> >> Hi,
> >>
> >> I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm 
> >> using this versions:
> >>
> >> kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64
> >> drbd90-utils-9.13.1-1.el8.elrepo.x86_64
> >>
> >> pacemaker-cluster-libs-2.0.4-6.el8.x86_64
> >> pacemaker-cli-2.0.4-6.el8.x86_64
> >> pacemaker-schemas-2.0.4-6.el8.noarch
> >> pacemaker-2.0.4-6.el8.x86_64
> >> pacemaker-libs-2.0.4-6.el8.x86_64
> >>
> >> clusternode-names are lisbon and susanne
> >>
> >> There are two drbd-resources configured which are paired with 
> >> filesystem-resources. Both are working in simple master/slave 
> >> configuration.
> >> When I start up the cluster one resource is promoted to a master and a 
> >> slave, the other is just getting on one the master started, but the slave 
> >> is not starting up.
> >>
> >> Status of the cluster:
> >> Cluster Summary:
> >>* Stack: corosync
> >>* Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - partition with 
> >> quorum
> >>* Last updated: Mon Jan 18 16:30:21 2021
> >>* Last change:  Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon
> >>* 2 nodes configured
> >>* 7 resource instances configured
> >>
> >> Node List:
> >>* Online: [ lisbon susanne ]
> >>
> >> Active Resources:
> >>* HA-IP_1 (ocf::heartbeat:IPaddr2):Started susanne
> >>* Clone Set: drbd_database-clone [drbd_database] (promotable):
> >>  * Masters: [ susanne ]
> >>  * Slaves: [ lisbon ]
> >>* fs_database (ocf::heartbeat:Filesystem): Started susanne
> >>* Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable):
> >>  * Masters: [ susanne ]
> >>* fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
> >>
> >> drbdadm status
> >>
> >> [root@susanne ~]# drbdadm status
> >> drbd1 role:Primary
> >>disk:UpToDate
> >>lisbon role:Secondary
> >>  peer-disk:UpToDate
> >>
> >> drbd2 role:Primary
> >>disk:UpToDate
> >>lisbon connection:Connecting
> >>
> >>   [root@lisbon ~]# drbdadm status
> >> drbd1 role:Secondary
> >>disk:UpToDate
> >>susanne role:Primary
> >>  peer-disk:UpToDate
> >>
> >>
> >> cluster-config:
> >> Cluster Name: mysql_cluster
> >> Corosync Nodes:
> >>   susanne lisbon
> >> Pacemaker Nodes:
> >>   lisbon susanne
> >>
> >> Resources:
> >>   Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2)
> >>Attributes: cidr_netmask=24 ip=192.168.18.150
> >>Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s)
> >>start interval=0s timeout=20s (HA-IP_1-start-interval-0s)
> >>stop interval=0s timeout=20s (HA-IP_1-stop-int

Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation

2021-01-19 Thread fatcharly
Thanks Ken, I will give it a try.

best regards and stay healthy

fatcharly



> Gesendet: Montag, 18. Januar 2021 um 22:49 Uhr
> Von: "Ken Gaillot" 
> An: "Cluster Labs - All topics related to open-source clustering welcomed" 
> 
> Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
>
> On Mon, 2021-01-18 at 18:43 +0100, fatcha...@gmx.de wrote:
> > Hi again,
> >
> > I need some help to figure out how to let a two node cluster with two
> > drbd-devices start the master-devices on the same node.
> > How can I configure colocation to work that way ? I tried to bond one
> > drbd-device with the other but that didn't work out quite well.
>
> Once you have the DRBD itself working, you can create a colocation
> constraint specifying just the master role. In pcs it's "pcs constraint
> colocation add Master  with Master ".
>
> Keep in mind that the dependent resource ('s master role in this
> case) will not be able to start if the other resource is not active.
>
> > This is my config:
> > > > I'm installing a 2 node pacemaker/drbd cluster on a CentOS 8.3.
> > > > I'm using this versions:
> > > >
> > > > kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64
> > > > drbd90-utils-9.13.1-1.el8.elrepo.x86_64
> > > >
> > > > pacemaker-cluster-libs-2.0.4-6.el8.x86_64
> > > > pacemaker-cli-2.0.4-6.el8.x86_64
> > > > pacemaker-schemas-2.0.4-6.el8.noarch
> > > > pacemaker-2.0.4-6.el8.x86_64
> > > > pacemaker-libs-2.0.4-6.el8.x86_64
> > > >
> > > > clusternode-names are lisbon and susanne
> > > >
> > > > Status of the cluster:
> > > > Cluster Summary:
> > > >   * Stack: corosync
> > > >   * Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) -
> > > > partition with quorum
> > > >   * Last updated: Mon Jan 18 16:30:21 2021
> > > >   * Last change:  Mon Jan 18 16:30:17 2021 by root via cibadmin
> > > > on lisbon
> > > >   * 2 nodes configured
> > > >   * 7 resource instances configured
> > > >
> > > > Node List:
> > > >   * Online: [ lisbon susanne ]
> > > >
> > > > Active Resources:
> > > >   * HA-IP_1 (ocf::heartbeat:IPaddr2):Started susanne
> > > >   * Clone Set: drbd_database-clone [drbd_database] (promotable):
> > > > * Masters: [ susanne ]
> > > > * Slaves: [ lisbon ]
> > > >   * fs_database (ocf::heartbeat:Filesystem): Started susanne
> > > >   * Clone Set: drbd_logsfiles-clone [drbd_logsfiles]
> > > > (promotable):
> > > > * Masters: [ susanne ]
> > > >   * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
> > > >
> > > > drbdadm status
> > > >
> > > > [root@susanne ~]# drbdadm status
> > > > drbd1 role:Primary
> > > >   disk:UpToDate
> > > >   lisbon role:Secondary
> > > > peer-disk:UpToDate
> > > >
> > > > drbd2 role:Primary
> > > >   disk:UpToDate
> > > >   lisbon connection:Connecting
> > > >
> > > >  [root@lisbon ~]# drbdadm status
> > > > drbd1 role:Secondary
> > > >   disk:UpToDate
> > > >   susanne role:Primary
> > > > peer-disk:UpToDate
> > > >
> > > >
> > > > cluster-config:
> > > > Cluster Name: mysql_cluster
> > > > Corosync Nodes:
> > > >  susanne lisbon
> > > > Pacemaker Nodes:
> > > >  lisbon susanne
> > > >
> > > > Resources:
> > > >  Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2)
> > > >   Attributes: cidr_netmask=24 ip=192.168.18.150
> > > >   Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s)
> > > >   start interval=0s timeout=20s (HA-IP_1-start-
> > > > interval-0s)
> > > >   stop interval=0s timeout=20s (HA-IP_1-stop-
> > > > interval-0s)
> > > >  Clone: drbd_database-clone
> > > >   Meta Attrs: clone-max=2 clone-node-max=1 notify=true
> > > > promotable=true promoted-max=1 promoted-node-max=1
> > > >   Resource: drbd_database (class=ocf provider=linbit type=drbd)
> > > >Attributes: drbd_resource=drbd1
> > > >Operations: demote interval=0s timeout=90 (drbd_database-
> > > > demote-interval-0s

Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation

2021-01-19 Thread fatcharly
Hi Brent,

this was an old status messages, pleasse ignore this, problem was already 
solved.

Kind regards and stay healthy 

fatcharly



> Gesendet: Montag, 18. Januar 2021 um 22:53 Uhr
> Von: "Brent Jensen" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
>
> Yes all works fine outside of the cluster. No firewall running nor any 
> selinux.
> 
> On 1/18/2021 11:53 AM, Strahil Nikolov wrote:
> >>> [root@susanne ~]# drbdadm status
> >>> drbd1 role:Primary
> >>>    disk:UpToDate
> >>>    lisbon role:Secondary
> >>>      peer-disk:UpToDate
> >>>
> >>> drbd2 role:Primary
> >>>    disk:UpToDate
> >>>    lisbon connection:Connecting
> > Are you sure you got the firewall open ? This node should be connected.
> >
> > Try to verify that each drbd is up and running and promoting any of the 2 
> > nodes is possible before proceeding with the cluster setup.
> >
> >
> > Best Regards,
> > Strahil Nikolov
> > ___
> > Manage your subscription:
> > https://lists.clusterlabs.org/mailman/listinfo/users
> >
> > ClusterLabs home: https://www.clusterlabs.org/
> 
> -- 
> This email has been checked for viruses by Avast antivirus software.
> https://www.avast.com/antivirus
> 
> ___
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> ClusterLabs home: https://www.clusterlabs.org/
>
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation

2021-01-19 Thread fatcharly
Hi Nikolov,

sorry, this is an old status, status is right now uptodate. 

Please ignore this messages.

Best regards and stay healthy

fatcharly



> Gesendet: Dienstag, 19. Januar 2021 um 05:23 Uhr
> Von: "Strahil Nikolov" 
> An: "users@clusterlabs.org" 
> Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
>
> So why it is saying 'connecting' ?
> 
> Best Regards,
> Strahil Nikolov
> 
> 
> 
> 
> 
> 
> В понеделник, 18 януари 2021 г., 23:54:02 Гринуич+2, Brent Jensen 
>  написа: 
> 
> 
> 
> 
> 
> Yes all works fine outside of the cluster. No firewall running nor any 
> selinux.
> 
> On 1/18/2021 11:53 AM, Strahil Nikolov wrote:
> >>> [root@susanne ~]# drbdadm status
> >>> drbd1 role:Primary
> >>>    disk:UpToDate
> >>>    lisbon role:Secondary
> >>>      peer-disk:UpToDate
> >>>
> >>> drbd2 role:Primary
> >>>    disk:UpToDate
> >>>    lisbon connection:Connecting
> > Are you sure you got the firewall open ? This node should be connected.
> >
> > Try to verify that each drbd is up and running and promoting any of the 2 
> > nodes is possible before proceeding with the cluster setup.
> >
> >
> > Best Regards,
> > Strahil Nikolov
> 
> > ___
> > Manage your subscription:
> > https://lists.clusterlabs.org/mailman/listinfo/users
> >
> > ClusterLabs home: https://www.clusterlabs.org/
> 
> -- 
> This email has been checked for viruses by Avast antivirus software.
> https://www.avast.com/antivirus
> 
> ___
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> ClusterLabs home: https://www.clusterlabs.org/
> 
> ___
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> ClusterLabs home: https://www.clusterlabs.org/
>
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation

2021-01-18 Thread fatcharly
rval=10 role=Master timeout=20 
> > (drbd_logsfiles-monitor-interval-10)
> >notify interval=0s timeout=90 
> > (drbd_logsfiles-notify-interval-0s)
> >promote interval=0s timeout=90 
> > (drbd_logsfiles-promote-interval-0s)
> >reload interval=0s timeout=30 
> > (drbd_logsfiles-reload-interval-0s)
> >start interval=0s timeout=240 
> > (drbd_logsfiles-start-interval-0s)
> >stop interval=0s timeout=100 
> > (drbd_logsfiles-stop-interval-0s)
> >  Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem)
> >   Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4
> >   Operations: monitor interval=20s timeout=40s 
> > (fs_logfiles-monitor-interval-20s)
> >   start interval=0s timeout=60s (fs_logfiles-start-interval-0s)
> >   stop interval=0s timeout=60s (fs_logfiles-stop-interval-0s)
> >
> > Stonith Devices:
> > Fencing Levels:
> >
> > Location Constraints:
> > Ordering Constraints:
> >   start drbd_database-clone then start fs_database (kind:Mandatory) 
> > (id:order-drbd_database-clone-fs_database-mandatory)
> >   start drbd_logsfiles-clone then start fs_logfiles (kind:Mandatory) 
> > (id:order-drbd_logsfiles-clone-fs_logfiles-mandatory)
> > Colocation Constraints:
> >   fs_database with drbd_database-clone (score:INFINITY) 
> > (with-rsc-role:Master) 
> > (id:colocation-fs_database-drbd_database-clone-INFINITY)
> >   fs_logfiles with drbd_logsfiles-clone (score:INFINITY) 
> > (with-rsc-role:Master) 
> > (id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY)
ERROR:>>>>>  drbd_logsfiles-clone with drbd_database-clone (score:INFINITY) 
(with-rsc-role:Master) 
(id:colocation-drbd_logsfiles-clone-drbd_database-clone-INFINITY)
> > Ticket Constraints:
> >
> > Alerts:
> >  No alerts defined
> >
> > Resources Defaults:
> >   No defaults set
> > Operations Defaults:
> >   No defaults set
> >
> > Cluster Properties:
> >  cluster-infrastructure: corosync
> >  cluster-name: mysql_cluster
> >  dc-version: 2.0.4-6.el8-2deceaa3ae
> >  have-watchdog: false
> >  last-lrm-refresh: 1610382881
> >  stonith-enabled: false
> >
> > Tags:
> >  No tags defined
> >
> > Quorum:
> >   Options:



Any suggestions are welcome

stay safe and healty

fatcharly

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started

2021-01-18 Thread fatcharly
Sorry Guys,

problem found, it was a colocation constrain of the drbd-devices.

best regards

fatcharly



> Gesendet: Montag, 18. Januar 2021 um 16:52 Uhr
> Von: fatcha...@gmx.de
> An: "clusterlabs" 
> Betreff: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started
>
> Hi,
>
> I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm 
> using this versions:
>
> kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64
> drbd90-utils-9.13.1-1.el8.elrepo.x86_64
>
> pacemaker-cluster-libs-2.0.4-6.el8.x86_64
> pacemaker-cli-2.0.4-6.el8.x86_64
> pacemaker-schemas-2.0.4-6.el8.noarch
> pacemaker-2.0.4-6.el8.x86_64
> pacemaker-libs-2.0.4-6.el8.x86_64
>
> clusternode-names are lisbon and susanne
>
> There are two drbd-resources configured which are paired with 
> filesystem-resources. Both are working in simple master/slave configuration.
> When I start up the cluster one resource is promoted to a master and a slave, 
> the other is just getting on one the master started, but the slave is not 
> starting up.
>
> Status of the cluster:
> Cluster Summary:
>   * Stack: corosync
>   * Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - partition with 
> quorum
>   * Last updated: Mon Jan 18 16:30:21 2021
>   * Last change:  Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon
>   * 2 nodes configured
>   * 7 resource instances configured
>
> Node List:
>   * Online: [ lisbon susanne ]
>
> Active Resources:
>   * HA-IP_1 (ocf::heartbeat:IPaddr2):Started susanne
>   * Clone Set: drbd_database-clone [drbd_database] (promotable):
> * Masters: [ susanne ]
> * Slaves: [ lisbon ]
>   * fs_database (ocf::heartbeat:Filesystem): Started susanne
>   * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable):
> * Masters: [ susanne ]
>   * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
>
> drbdadm status
>
> [root@susanne ~]# drbdadm status
> drbd1 role:Primary
>   disk:UpToDate
>   lisbon role:Secondary
> peer-disk:UpToDate
>
> drbd2 role:Primary
>   disk:UpToDate
>   lisbon connection:Connecting
>
>  [root@lisbon ~]# drbdadm status
> drbd1 role:Secondary
>   disk:UpToDate
>   susanne role:Primary
> peer-disk:UpToDate
>
>
> cluster-config:
> Cluster Name: mysql_cluster
> Corosync Nodes:
>  susanne lisbon
> Pacemaker Nodes:
>  lisbon susanne
>
> Resources:
>  Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2)
>   Attributes: cidr_netmask=24 ip=192.168.18.150
>   Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s)
>   start interval=0s timeout=20s (HA-IP_1-start-interval-0s)
>   stop interval=0s timeout=20s (HA-IP_1-stop-interval-0s)
>  Clone: drbd_database-clone
>   Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true 
> promoted-max=1 promoted-node-max=1
>   Resource: drbd_database (class=ocf provider=linbit type=drbd)
>Attributes: drbd_resource=drbd1
>Operations: demote interval=0s timeout=90 
> (drbd_database-demote-interval-0s)
>monitor interval=20 role=Slave timeout=20 
> (drbd_database-monitor-interval-20)
>monitor interval=10 role=Master timeout=20 
> (drbd_database-monitor-interval-10)
>notify interval=0s timeout=90 
> (drbd_database-notify-interval-0s)
>promote interval=0s timeout=90 
> (drbd_database-promote-interval-0s)
>reload interval=0s timeout=30 
> (drbd_database-reload-interval-0s)
>start interval=0s timeout=240 (drbd_database-start-interval-0s)
>stop interval=0s timeout=100 (drbd_database-stop-interval-0s)
>  Resource: fs_database (class=ocf provider=heartbeat type=Filesystem)
>   Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4
>   Operations: monitor interval=20s timeout=40s 
> (fs_database-monitor-interval-20s)
>   start interval=0s timeout=60s (fs_database-start-interval-0s)
>   stop interval=0s timeout=60s (fs_database-stop-interval-0s)
>  Clone: drbd_logsfiles-clone
>   Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true 
> promoted-max=1 promoted-node-max=1
>   Resource: drbd_logsfiles (class=ocf provider=linbit type=drbd)
>Attributes: drbd_resource=drbd2
>Operations: demote interval=0s timeout=90 
> (drbd_logsfiles-demote-interval-0s)
>monitor interval=20 role=Slave timeout=20 
> (drbd_logsfiles-monitor-interval-20)
>monitor interval=10 role=Master timeout=20 
> (drbd_logsfiles-monitor-interval-10)
>notify interval=0s timeout=90 
> (drbd_logsfiles-notify-interval-0s)

[ClusterLabs] CentOS 8 & drbd 9 second slave is not started

2021-01-18 Thread fatcharly
base-clone-fs_database-mandatory)
  start drbd_logsfiles-clone then start fs_logfiles (kind:Mandatory) 
(id:order-drbd_logsfiles-clone-fs_logfiles-mandatory)
Colocation Constraints:
  fs_database with drbd_database-clone (score:INFINITY) (with-rsc-role:Master) 
(id:colocation-fs_database-drbd_database-clone-INFINITY)
  fs_logfiles with drbd_logsfiles-clone (score:INFINITY) (with-rsc-role:Master) 
(id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY)
  drbd_logsfiles-clone with drbd_database-clone (score:INFINITY) 
(with-rsc-role:Master) 
(id:colocation-drbd_logsfiles-clone-drbd_database-clone-INFINITY)
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
  No defaults set
Operations Defaults:
  No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: mysql_cluster
 dc-version: 2.0.4-6.el8-2deceaa3ae
 have-watchdog: false
 last-lrm-refresh: 1610382881
 stonith-enabled: false

Tags:
 No tags defined

Quorum:
  Options:


Any suggestions are welcome

stay safe and healty

fatcharly
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] drbd9 and Centos 8.3

2021-01-08 Thread fatcharly
Hi,

I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm 
using this versions:

kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64
drbd90-utils-9.13.1-1.el8.elrepo.x86_64

pacemaker-cluster-libs-2.0.4-6.el8.x86_64
pacemaker-cli-2.0.4-6.el8.x86_64
pacemaker-schemas-2.0.4-6.el8.noarch
pacemaker-2.0.4-6.el8.x86_64
pacemaker-libs-2.0.4-6.el8.x86_64

clusternode-names are lisbon and susanne

when I try to install my drbd resource with
pcs resource create drbd_database ocf:linbit:drbd drbd_resource=drbd1 
promotable promoted-max=1 promoted-node-max=1 clone-max=2 clone-node-max=1 
notify=true

I get this information in the cluster information:
* drbd_database_monitor_0 on lisbon 'not installed' (5): call=20, 
status='complete', exitreason='DRBD kernel (module) not available?', 
last-rc-change='2021-01-08 18:03:10 +01:00', queued=0ms,
exec=96ms

The drbd-service is not started, shouldn't it be started by the clustermanager ?

Any suggestions are welcome

Have a nice weekand and stay healthy

fatcharly

___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] cluster problems after let's encrypt

2020-07-06 Thread fatcharly
Hi,

I'm running a two node corosync httpd-cluster on a CentOS 7.
corosync-2.4.5-4.el7.x86_64
pcs-0.9.168-4.el7.centos.x86_64
Today I used lets encrypt to installt https for two domains on that system.
After that the node with the new https-domains is not longer able to hold the 
apache resource:

The resource is configured like this:
Resource: apache (class=ocf provider=heartbeat type=apache)
Attributes: configfile=/etc/httpd/conf/httpd.conf 
statusurl=http://127.0.0.1:8089/server-status
Operations: start interval=0s timeout=40s (apache-start-interval-0s)
stop interval=0s timeout=60s (apache-stop-interval-0s)
monitor interval=1min (apache-monitor-interval-1min)

The status-page is configured like this:
Listen 127.0.0.1:8089

SetHandler server-status
Order deny,allow
Deny from all
Allow from 127.0.0.1

And the log shows this:
Jul 6 16:55:18 bachi2 apache(apache)[7182]: INFO: waiting for apache 
/etc/httpd/conf/httpd.conf to come up
Jul 6 16:55:19 bachi2 apache(apache)[7182]: INFO: apache not running
Jul 6 16:55:19 bachi2 apache(apache)[7182]: INFO: waiting for apache 
/etc/httpd/conf/httpd.conf to come up
Jul 6 16:55:20 bachi2 apache(apache)[7182]: INFO: apache not running
Jul 6 16:55:20 bachi2 apache(apache)[7182]: INFO: waiting for apache 
/etc/httpd/conf/httpd.conf to come up
Jul 6 16:55:21 bachi2 lrmd[8756]: warning: apache_start_0 process (PID 7182) 
timed out
Jul 6 16:55:21 bachi2 lrmd[8756]: warning: apache_start_0:7182 - timed out 
after 4ms
Jul 6 16:55:21 bachi2 crmd[8759]: error: Result of start operation for apache 
on bachi2: Timed Out
Jul 6 16:55:21 bachi2 apache(apache)[8416]: INFO: apache is not running.

Any suggestions are welcome

Best regards

fatcharly
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/


[ClusterLabs] failure-timeout

2019-04-11 Thread fatcharly
Hi,

I´m running a pcs/corosync two node cluster on CentOS 7.6 
I use a cloned ping resource and I'd like to add a failure-timeout to this.
How do I do this ?

Any suggestions are welcome

Kind regards

fatcharly
___
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/

Re: [ClusterLabs] pacemaker apache and umask on CentOS 7

2016-04-22 Thread fatcharly


> Gesendet: Mittwoch, 20. April 2016 um 19:35 Uhr
> Von: "Ken Gaillot" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] pacemaker apache and umask on CentOS 7
>
> On 04/20/2016 12:20 PM, Klaus Wenninger wrote:
> > On 04/20/2016 05:35 PM, fatcha...@gmx.de wrote:
> >>
> >>> Gesendet: Mittwoch, 20. April 2016 um 16:31 Uhr
> >>> Von: "Klaus Wenninger" 
> >>> An: users@clusterlabs.org
> >>> Betreff: Re: [ClusterLabs] pacemaker apache and umask on CentOS 7
> >>>
> >>> On 04/20/2016 04:11 PM, fatcha...@gmx.de wrote:
> >>>> Hi,
> >>>>
> >>>> I´m running a 2-node apache webcluster on a fully patched CentOS 7 
> >>>> (pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64).
> >>>> Some files which are generated by the apache are created with a umask 
> >>>> 137 but I need this files created with a umask of 117.
> >>>> To change this I first tried to add a umask 117 to /etc/sysconfig/httpd 
> >>>> & rebooted the system. This had no effekt.
> >>>> So I found out (after some research) that this is not working under 
> >>>> CentOS 7 and that this had to be changed via systemd.
> >>>> So I created a directory "/etc/systemd/system/httpd.service.d" and put 
> >>>> there a "umask.conf"-File with this content: 
> >>>> [Service]
> >>>> UMask=0117
> >>>>
> >>>> Again I rebooted the system but no effekt.
> >>>> Is the pacemaker really starting the apache over the systemd ? And how 
> >>>> can I solve the problem ?
> >>> Didn't check with CentOS7 but on RHEL7 there is a
> >>> /usr/lib/ocf/resource.d/heartbeat/apache.
> >>> So it depends on how you defined the resource starting apache if systemd
> >>> is used or if it being done by the ocf-ra.
> >> MY configuration is:
> >> Resource: apache (class=ocf provider=heartbeat type=apache)
> >>   Attributes: configfile=/etc/httpd/conf/httpd.conf 
> >> statusurl=http://127.0.0.1:8089/server-status
> >>   Operations: start interval=0s timeout=40s (apache-start-timeout-40s)
> >>   stop interval=0s timeout=60s (apache-stop-timeout-60s)
> >>   monitor interval=1min (apache-monitor-interval-1min)
> >>
> >> So I quess it is ocf. But what will be the right way to do it ? I lack a 
> >> bit of understandig about this /usr/lib/ocf/resource.d/heartbeat/apache 
> >> file.  
> >>
> > There are the ocf-Resource-Agents (if there is none you can always
> > create one for your service) which usually
> > give you a little bit more control of the service from the cib. (You can
> > set a couple of variables like in this example
> > the pointer to the config-file)
> > And of course you can always create resources referring the native
> > services of your distro (systemd-units in
> > this case).
> >>
> >>
> >>
> >>>> Any suggestions are welcome
> 
> If you add envfiles="/etc/sysconfig/httpd" to your apache resource, it
> should work.

worked like a charm. Thanks to everybody for your support.

Kind regards

fatcharly

> 
> >>>> Kind regards
> >>>>
> >>>> fatcharly
> 
> ___
> Users mailing list: Users@clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] pacemaker apache and umask on CentOS 7

2016-04-20 Thread fatcharly


> Gesendet: Mittwoch, 20. April 2016 um 16:31 Uhr
> Von: "Klaus Wenninger" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] pacemaker apache and umask on CentOS 7
>
> On 04/20/2016 04:11 PM, fatcha...@gmx.de wrote:
> > Hi,
> >
> > I´m running a 2-node apache webcluster on a fully patched CentOS 7 
> > (pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64).
> > Some files which are generated by the apache are created with a umask 137 
> > but I need this files created with a umask of 117.
> > To change this I first tried to add a umask 117 to /etc/sysconfig/httpd & 
> > rebooted the system. This had no effekt.
> > So I found out (after some research) that this is not working under CentOS 
> > 7 and that this had to be changed via systemd.
> > So I created a directory "/etc/systemd/system/httpd.service.d" and put 
> > there a "umask.conf"-File with this content: 
> > [Service]
> > UMask=0117
> >
> > Again I rebooted the system but no effekt.
> > Is the pacemaker really starting the apache over the systemd ? And how can 
> > I solve the problem ?
> Didn't check with CentOS7 but on RHEL7 there is a
> /usr/lib/ocf/resource.d/heartbeat/apache.
> So it depends on how you defined the resource starting apache if systemd
> is used or if it being done by the ocf-ra.
MY configuration is:
Resource: apache (class=ocf provider=heartbeat type=apache)
  Attributes: configfile=/etc/httpd/conf/httpd.conf 
statusurl=http://127.0.0.1:8089/server-status
  Operations: start interval=0s timeout=40s (apache-start-timeout-40s)
  stop interval=0s timeout=60s (apache-stop-timeout-60s)
  monitor interval=1min (apache-monitor-interval-1min)

So I quess it is ocf. But what will be the right way to do it ? I lack a bit of 
understandig about this /usr/lib/ocf/resource.d/heartbeat/apache file.  





> >
> > Any suggestions are welcome
> >
> > Kind regards
> >
> > fatcharly
> >  
> >
> >
> > ___
> > Users mailing list: Users@clusterlabs.org
> > http://clusterlabs.org/mailman/listinfo/users
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> 
> 
> ___
> Users mailing list: Users@clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] pacemaker apache and umask on CentOS 7

2016-04-20 Thread fatcharly
Hi,

I´m running a 2-node apache webcluster on a fully patched CentOS 7 
(pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64).
Some files which are generated by the apache are created with a umask 137 but I 
need this files created with a umask of 117.
To change this I first tried to add a umask 117 to /etc/sysconfig/httpd & 
rebooted the system. This had no effekt.
So I found out (after some research) that this is not working under CentOS 7 
and that this had to be changed via systemd.
So I created a directory "/etc/systemd/system/httpd.service.d" and put there a 
"umask.conf"-File with this content: 
[Service]
UMask=0117

Again I rebooted the system but no effekt.
Is the pacemaker really starting the apache over the systemd ? And how can I 
solve the problem ?

Any suggestions are welcome

Kind regards

fatcharly
 
   

___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] cloned pingd resource problem

2016-03-30 Thread fatcharly
Hi,

I`m running a two node cluster on a fully updated CentOS 7 
(pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64) . I see on one 
of our nodes a lot of this in the logfiles:

Mar 30 12:32:13 localhost crmd[12986]:  notice: State transition S_IDLE -> 
S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Mar 30 12:32:13 localhost pengine[12985]:  notice: On loss of CCM Quorum: Ignore
Mar 30 12:32:13 localhost pengine[12985]: warning: Processing failed op monitor 
for ping_fw:0 on kathie2: unknown error (1)
Mar 30 12:32:13 localhost pengine[12985]: warning: Processing failed op start 
for ping_fw:1 on stacy2: unknown error (1)
Mar 30 12:32:13 localhost pengine[12985]: warning: Forcing ping_fw-clone away 
from stacy2 after 100 failures (max=100)
Mar 30 12:32:13 localhost pengine[12985]: warning: Forcing ping_fw-clone away 
from stacy2 after 100 failures (max=100)
Mar 30 12:32:13 localhost pengine[12985]:  notice: Calculated Transition 1823: 
/var/lib/pacemaker/pengine/pe-input-355.bz2
Mar 30 12:32:13 localhost crmd[12986]:  notice: Transition 1823 (Complete=0, 
Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-input-355.bz2): Complete
Mar 30 12:32:13 localhost crmd[12986]:  notice: State transition 
S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL 
origin=notify_crmd ]


The configuration looks like this:

Clone: ping_fw-clone
  Resource: ping_fw (class=ocf provider=pacemaker type=ping)
   Attributes: dampen=5s multiplier=1000 host_list=192.168.16.1 timeout=60
   Operations: start interval=0s timeout=60 (ping_fw-start-interval-0s)
   stop interval=0s timeout=20 (ping_fw-stop-interval-0s)
   monitor interval=45 (ping_fw-monitor-interval-45)


What can I do to resolve the problem ? 

Any suggestions are welcome

Kind regards

fatcharly



___
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org