[ClusterLabs] Problem with a ping resource on CentOS 8
17 with HA-IP_211 (score:INFINITY) (id:colocation-HA-IP_217-HA-IP_211-INFINITY) HA-IP_218 with HA-IP_211 (score:INFINITY) (id:colocation-HA-IP_218-HA-IP_211-INFINITY) HA-IP_219 with HA-IP_211 (score:INFINITY) (id:colocation-HA-IP_219-HA-IP_211-INFINITY) Ticket Constraints: Alerts: No alerts defined Resources Defaults: No defaults set Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: loadbalancer dc-version: 2.0.5-9.el8_4.1-ba59be7122 have-watchdog: false stonith-enabled: false Tags: No tags defined Quorum: Options: Any suggestions are welcome Best regards and stay safe fatcharly ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
> Gesendet: Dienstag, 18. Mai 2021 um 14:49 Uhr > Von: fatcha...@gmx.de > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > Hi Andrei,Hi everybody, > > ... > > > and it works great Thanks for the hint. > > > But the thing I still don't understand is why the cluster demotes is > > > active node for a short time when I reenable a node from standby back to > > > unstandby ? Is it not possible to join the drbd as secondary without > > > demote the primary for a short moment ? > > > > Try adding interleave=true to your clones. > > I tried this but it get me an error msg, what is wrong ? > > pcs resource update database_drbd ocf:linbit:drbd drbd_resource=drbd1 > promotable promoted-max=1 promoted-node-max=1 clone-max=2 clone-node-max=1 > notify=true interleave=true > > Error: invalid resource options: 'clone-max', 'clone-node-max', 'interleave', > 'notify', 'promoted-max', 'promoted-node-max', allowed options are: > 'adjust_master_score', 'connect_only_after_promote', 'drbd_resource', > 'drbdconf', 'fail_promote_early_if_peer_primary', > 'ignore_missing_notifications', 'remove_master_score_if_peer_primary', > 'require_drbd_module_version_ge', 'require_drbd_module_version_lt', > 'stop_outdates_secondary', 'unfence_extra_args', 'unfence_if_all_uptodate', > 'wfc_timeout', use --force to override or is it simply: pcs resource update database_drbd-clone interleave=true ? > > Any suggestions are welcome > > Stay safe and take care > > fatcharly > > > > > > Gesendet: Mittwoch, 12. Mai 2021 um 19:04 Uhr > > Von: "Andrei Borzenkov" > > An: users@clusterlabs.org > > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > > > On 12.05.2021 17:34, fatcha...@gmx.de wrote: > > > Hi Andrei, Hi everybody, > > > > > > > > >> Gesendet: Mittwoch, 12. Mai 2021 um 16:01 Uhr > > >> Von: fatcha...@gmx.de > > >> An: users@clusterlabs.org > > >> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > >> > > >> Hi Andrei, Hi everybody, > > >> > > >> > > >>> You need order fs_database after promote operation; and as I just found > > >>> pacemaker also does not reverse it correctly and executes fs stop and > > >>> drbd demote concurrently. So you need additional order constraint to > > >>> first stop fs then demote drbd. > > >> > > >> is there so good doku about this, I don't know how to archive a "after > > >> promote operation" and how can I tell the pcs to first dismount the > > >> filesystem mountpoint and then demote the drbd-device. > > >> > > > ok, so I found something and used this: > > > > > > pcs constraint order stop fs_logfiles then demote drbd_logsfiles-clone > > > pcs constraint order stop fs_database then demote database_drbd-clone > > > > > > and it works great Thanks for the hint. > > > But the thing I still don't understand is why the cluster demotes is > > > active node for a short time when I reenable a node from standby back to > > > unstandby ? Is it not possible to join the drbd as secondary without > > > demote the primary for a short moment ? > > > > Try adding interleave=true to your clones. > > > > > > > > Best regards and take care > > > > > > fatcharly > > > > > > > > > > > >> Sorry but this is new for me. > > >> > > >> Best regards and take care > > >> > > >> fatcharly > > >> > > >> > > >> > > >> > > >>> Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr > > >>> Von: "Andrei Borzenkov" > > >>> An: users@clusterlabs.org > > >>> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems > > >>> ? > > >>> > > >>> On 11.05.2021 17:43, fatcha...@gmx.de wrote: > > >>>> Hi, > > >>>> > > >>>> I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + > > >>>> corosync-3.0.3-4.el8.x86_64 and > > >>>> kmod-drbd90-9.0.25-2.el8_3.elrepo
Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
Hi Andrei,Hi everybody, ... > > and it works great Thanks for the hint. > > But the thing I still don't understand is why the cluster demotes is active > > node for a short time when I reenable a node from standby back to unstandby > > ? Is it not possible to join the drbd as secondary without demote the > > primary for a short moment ? > > Try adding interleave=true to your clones. I tried this but it get me an error msg, what is wrong ? pcs resource update database_drbd ocf:linbit:drbd drbd_resource=drbd1 promotable promoted-max=1 promoted-node-max=1 clone-max=2 clone-node-max=1 notify=true interleave=true Error: invalid resource options: 'clone-max', 'clone-node-max', 'interleave', 'notify', 'promoted-max', 'promoted-node-max', allowed options are: 'adjust_master_score', 'connect_only_after_promote', 'drbd_resource', 'drbdconf', 'fail_promote_early_if_peer_primary', 'ignore_missing_notifications', 'remove_master_score_if_peer_primary', 'require_drbd_module_version_ge', 'require_drbd_module_version_lt', 'stop_outdates_secondary', 'unfence_extra_args', 'unfence_if_all_uptodate', 'wfc_timeout', use --force to override Any suggestions are welcome Stay safe and take care fatcharly > Gesendet: Mittwoch, 12. Mai 2021 um 19:04 Uhr > Von: "Andrei Borzenkov" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > On 12.05.2021 17:34, fatcha...@gmx.de wrote: > > Hi Andrei, Hi everybody, > > > > > >> Gesendet: Mittwoch, 12. Mai 2021 um 16:01 Uhr > >> Von: fatcha...@gmx.de > >> An: users@clusterlabs.org > >> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > >> > >> Hi Andrei, Hi everybody, > >> > >> > >>> You need order fs_database after promote operation; and as I just found > >>> pacemaker also does not reverse it correctly and executes fs stop and > >>> drbd demote concurrently. So you need additional order constraint to > >>> first stop fs then demote drbd. > >> > >> is there so good doku about this, I don't know how to archive a "after > >> promote operation" and how can I tell the pcs to first dismount the > >> filesystem mountpoint and then demote the drbd-device. > >> > > ok, so I found something and used this: > > > > pcs constraint order stop fs_logfiles then demote drbd_logsfiles-clone > > pcs constraint order stop fs_database then demote database_drbd-clone > > > > and it works great Thanks for the hint. > > But the thing I still don't understand is why the cluster demotes is active > > node for a short time when I reenable a node from standby back to unstandby > > ? Is it not possible to join the drbd as secondary without demote the > > primary for a short moment ? > > Try adding interleave=true to your clones. > > > > > Best regards and take care > > > > fatcharly > > > > > > > >> Sorry but this is new for me. > >> > >> Best regards and take care > >> > >> fatcharly > >> > >> > >> > >> > >>> Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr > >>> Von: "Andrei Borzenkov" > >>> An: users@clusterlabs.org > >>> Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > >>> > >>> On 11.05.2021 17:43, fatcha...@gmx.de wrote: > >>>> Hi, > >>>> > >>>> I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + > >>>> corosync-3.0.3-4.el8.x86_64 and kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64. > >>>> The cluster consists of two nodes which are providing a ha-mariadb with > >>>> the help of two drbd devices for the database and the logfiles. The > >>>> corosync is working over two rings and both machines are virtual > >>>> kvm-guests. > >>>> > >>>> Problem: > >>>> Node susanne is the active node and lisbon is changing from standby to > >>>> active, susanna is trying to demote one drbd-device but is failling to. > >>>> The cluster is working on properly, but the error stays. > >>>> This is the what happens: > >>>> > >>>> Cluster Summary: > >>>> * Stack: corosync > >>>> * Current DC:
Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
Hi Andrei, Hi everybody, > Gesendet: Mittwoch, 12. Mai 2021 um 16:01 Uhr > Von: fatcha...@gmx.de > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > Hi Andrei, Hi everybody, > > > > You need order fs_database after promote operation; and as I just found > > pacemaker also does not reverse it correctly and executes fs stop and > > drbd demote concurrently. So you need additional order constraint to > > first stop fs then demote drbd. > > is there so good doku about this, I don't know how to archive a "after > promote operation" and how can I tell the pcs to first dismount the > filesystem mountpoint and then demote the drbd-device. > ok, so I found something and used this: pcs constraint order stop fs_logfiles then demote drbd_logsfiles-clone pcs constraint order stop fs_database then demote database_drbd-clone and it works great Thanks for the hint. But the thing I still don't understand is why the cluster demotes is active node for a short time when I reenable a node from standby back to unstandby ? Is it not possible to join the drbd as secondary without demote the primary for a short moment ? Best regards and take care fatcharly > Sorry but this is new for me. > > Best regards and take care > > fatcharly > > > > > > Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr > > Von: "Andrei Borzenkov" > > An: users@clusterlabs.org > > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > > > On 11.05.2021 17:43, fatcha...@gmx.de wrote: > > > Hi, > > > > > > I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + > > > corosync-3.0.3-4.el8.x86_64 and kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64. > > > The cluster consists of two nodes which are providing a ha-mariadb with > > > the help of two drbd devices for the database and the logfiles. The > > > corosync is working over two rings and both machines are virtual > > > kvm-guests. > > > > > > Problem: > > > Node susanne is the active node and lisbon is changing from standby to > > > active, susanna is trying to demote one drbd-device but is failling to. > > > The cluster is working on properly, but the error stays. > > > This is the what happens: > > > > > > Cluster Summary: > > > * Stack: corosync > > > * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition > > > with quo rum > > > * Last updated: Tue May 11 16:15:54 2021 > > > * Last change: Tue May 11 16:15:42 2021 by root via cibadmin on susanne > > > * 2 nodes configured > > > * 11 resource instances configured > > > > > > Node List: > > > * Online: [ lisbon susanne ] > > > > > > Active Resources: > > > * HA_IP (ocf::heartbeat:IPaddr2):Started susanne > > > * Clone Set: database_drbd-clone [database_drbd] (promotable): > > > * Masters: [ susanne ] > > > * Slaves: [ lisbon ] > > > * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable): > > > * drbd_logsfiles(ocf::linbit:drbd): Demoting susanne > > > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > > > > Presumably fs_logfiles is located on drbd_logfiles, so how comes it is > > active while drbd_logfiles is being demoted? Then drbdadm fails to > > change status to secondary and RA simply loops forever until timeout. > > > > > * fs_database (ocf::heartbeat:Filesystem): Started susanne > > > * mysql-server(ocf::heartbeat:mysql): Started susanne > > > * Clone Set: ping_fw-clone [ping_fw]: > > > * Started: [ lisbon susanne ] > > > > > > --- > > > after a few seconds it switches over: > > > > > > Cluster Summary: > > > * Stack: corosync > > > * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition > > > with quo rum > > > * Last updated: Tue May 11 16:17:59 2021 > > > * Last change: Tue May 11 16:15:42 2021 by root via cibadmin on susanne > > > * 2 nodes configured > > > * 11 resource instances configured > > > > > > Node List: > > > * Online: [ lisbon susanne ] > > > > > > Active Resources: > > > * HA_IP (ocf::heartbeat:IPaddr2):Started susanne > > > * Clone Set: database_drbd-clone [database_drbd] (prom
Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
Hi Andrei, Hi everybody, > You need order fs_database after promote operation; and as I just found > pacemaker also does not reverse it correctly and executes fs stop and > drbd demote concurrently. So you need additional order constraint to > first stop fs then demote drbd. is there so good doku about this, I don't know how to archive a "after promote operation" and how can I tell the pcs to first dismount the filesystem mountpoint and then demote the drbd-device. Sorry but this is new for me. Best regards and take care fatcharly > Gesendet: Dienstag, 11. Mai 2021 um 17:19 Uhr > Von: "Andrei Borzenkov" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] 2 node mariadb-cluster - constraint-problems ? > > On 11.05.2021 17:43, fatcha...@gmx.de wrote: > > Hi, > > > > I'm using a CentOS 8.3.2011 with a pacemaker-2.0.4-6.el8_3.1.x86_64 + > > corosync-3.0.3-4.el8.x86_64 and kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64. > > The cluster consists of two nodes which are providing a ha-mariadb with the > > help of two drbd devices for the database and the logfiles. The corosync is > > working over two rings and both machines are virtual kvm-guests. > > > > Problem: > > Node susanne is the active node and lisbon is changing from standby to > > active, susanna is trying to demote one drbd-device but is failling to. The > > cluster is working on properly, but the error stays. > > This is the what happens: > > > > Cluster Summary: > > * Stack: corosync > > * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition > > with quo rum > > * Last updated: Tue May 11 16:15:54 2021 > > * Last change: Tue May 11 16:15:42 2021 by root via cibadmin on susanne > > * 2 nodes configured > > * 11 resource instances configured > > > > Node List: > > * Online: [ lisbon susanne ] > > > > Active Resources: > > * HA_IP (ocf::heartbeat:IPaddr2):Started susanne > > * Clone Set: database_drbd-clone [database_drbd] (promotable): > > * Masters: [ susanne ] > > * Slaves: [ lisbon ] > > * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable): > > * drbd_logsfiles(ocf::linbit:drbd): Demoting susanne > > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > > Presumably fs_logfiles is located on drbd_logfiles, so how comes it is > active while drbd_logfiles is being demoted? Then drbdadm fails to > change status to secondary and RA simply loops forever until timeout. > > > * fs_database (ocf::heartbeat:Filesystem): Started susanne > > * mysql-server(ocf::heartbeat:mysql): Started susanne > > * Clone Set: ping_fw-clone [ping_fw]: > > * Started: [ lisbon susanne ] > > > > --- > > after a few seconds it switches over: > > > > Cluster Summary: > > * Stack: corosync > > * Current DC: lisbon (version 2.0.4-6.el8_3.1-2deceaa3ae) - partition > > with quo rum > > * Last updated: Tue May 11 16:17:59 2021 > > * Last change: Tue May 11 16:15:42 2021 by root via cibadmin on susanne > > * 2 nodes configured > > * 11 resource instances configured > > > > Node List: > > * Online: [ lisbon susanne ] > > > > Active Resources: > > * HA_IP (ocf::heartbeat:IPaddr2):Started susanne > > * Clone Set: database_drbd-clone [database_drbd] (promotable): > > * Masters: [ susanne ] > > * Slaves: [ lisbon ] > > * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable): > > * Masters: [ susanne ] > > * Slaves: [ lisbon ] > > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > > * fs_database (ocf::heartbeat:Filesystem): Started susanne > > * mysql-server(ocf::heartbeat:mysql): Started susanne > > * Resource Group: apache: > > * httpd_srv (ocf::heartbeat:apache): Started susanne > > * Clone Set: ping_fw-clone [ping_fw]: > > * Started: [ lisbon susanne ] > > > > Failed Resource Actions: > > * drbd_logsfiles_demote_0 on susanne 'error' (1): call=736, status='Timed > > Out' > > , exitreason='', last-rc-change='2021-05-11 16:15:42 +02:00', queued=0ms, > > exec=9 0001ms > > -- > > > > And what you see in logs? > > > I think it is a constraint-problem, but I can't find it. > >
[ClusterLabs] 2 node mariadb-cluster - constraint-problems ?
l-10) notify interval=0s timeout=90 (drbd_logsfiles-notify-interval-0s) promote interval=0s timeout=90 (drbd_logsfiles-promote-interval-0s) reload interval=0s timeout=30 (drbd_logsfiles-reload-interval-0s) start interval=0s timeout=240 (drbd_logsfiles-start-interval-0s) stop interval=0s timeout=100 (drbd_logsfiles-stop-interval-0s) Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4 Operations: monitor interval=20s timeout=40s (fs_logfiles-monitor-interval-20s) start interval=0s timeout=60s (fs_logfiles-start-interval-0s) stop interval=0s timeout=60s (fs_logfiles-stop-interval-0s) Resource: fs_database (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4 Operations: monitor interval=20s timeout=40s (fs_database-monitor-interval-20s) start interval=0s timeout=60s (fs_database-start-interval-0s) stop interval=0s timeout=60s (fs_database-stop-interval-0s) Resource: mysql-server (class=ocf provider=heartbeat type=mysql) Attributes: additional_parameters=--bind-address=0.0.0.0 binary=/usr/bin/mysqld_safe config=/etc/my.cnf datadir=/mnt/clusterfs1/mysql pid=/var/lib/mysql/run/mariadb.pid socket=/var/lib/mysql/mysql.sock Operations: demote interval=0s timeout=120s (mysql-server-demote-interval-0s) monitor interval=20s timeout=30s (mysql-server-monitor-interval-20s) notify interval=0s timeout=90s (mysql-server-notify-interval-0s) promote interval=0s timeout=120s (mysql-server-promote-interval-0s) start interval=0s timeout=60s (mysql-server-start-interval-0s) stop interval=0s timeout=60s (mysql-server-stop-interval-0s) Group: apache Resource: httpd_srv (class=ocf provider=heartbeat type=apache) Attributes: configfile=/etc/httpd/conf/httpd.conf statusurl=http://127.0.0.1/server-status Operations: monitor interval=10s timeout=20s (httpd_srv-monitor-interval-10s) start interval=0s timeout=40s (httpd_srv-start-interval-0s) stop interval=0s timeout=60s (httpd_srv-stop-interval-0s) Clone: ping_fw-clone Resource: ping_fw (class=ocf provider=pacemaker type=ping) Attributes: dampen=10s host_list=192.168.18.1 multiplier=1000 Operations: monitor interval=10s timeout=60s (ping_fw-monitor-interval-10s) start interval=0s timeout=60s (ping_fw-start-interval-0s) stop interval=0s timeout=20s (ping_fw-stop-interval-0s) Stonith Devices: Fencing Levels: Location Constraints: Resource: mysql-server Constraint: location-mysql-server Rule: boolean-op=or score=-INFINITY (id:location-mysql-server-rule) Expression: pingd lt 1 (id:location-mysql-server-rule-expr) Expression: not_defined pingd (id:location-mysql-server-rule-expr-1) Ordering Constraints: start mysql-server then start httpd_srv (kind:Mandatory) (id:order-mysql-server-httpd_srv-mandatory) start database_drbd-clone then start drbd_logsfiles-clone (kind:Mandatory) (id:order-database_drbd-clone-drbd_logsfiles-clone-mandatory) start drbd_logsfiles-clone then start fs_database (kind:Mandatory) (id:order-drbd_logsfiles-clone-fs_database-mandatory) start fs_database then start fs_logfiles (kind:Mandatory) (id:order-fs_database-fs_logfiles-mandatory) start fs_logfiles then start mysql-server (kind:Mandatory) (id:order-fs_logfiles-mysql-server-mandatory) Colocation Constraints: fs_logfiles with drbd_logsfiles-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY) fs_database with database_drbd-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-fs_database-database_drbd-clone-INFINITY) drbd_logsfiles-clone with database_drbd-clone (score:INFINITY) (rsc-role:Master) (with-rsc-role:Master) (id:colocation-drbd_logsfiles-clone-database_drbd-clone-INFINITY) HA_IP with database_drbd-clone (score:INFINITY) (rsc-role:Started) (with-rsc-role:Master) (id:colocation-HA_IP-database_drbd-clone-INFINITY) mysql-server with fs_database (score:INFINITY) (id:colocation-mysql-server-fs_database-INFINITY) httpd_srv with mysql-server (score:INFINITY) (id:colocation-httpd_srv-mysql-server-INFINITY) Ticket Constraints: Alerts: No alerts defined Resources Defaults: No defaults set Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: mysql_cluster dc-version: 2.0.4-6.el8_3.1-2deceaa3ae have-watchdog: false last-lrm-refresh: 1620742514 stonith-enabled: FALSE Tags: No tags defined Quorum: Options: Any suggestions are welcome best regards stay safe, take care fatcharly ___ Manage your subscription: https://lists.clusterlabs.o
Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started
Hi Brent, I now use the drbd without starting the drbd service. All is managed by the pcs, everything gets mounted and unmounted as expected. In the beginning I was dealing with a lot of problems caused by the selinux and a wrong used constrain. Now everything works as I know it from my other cluster which are build on CentOS 5/6/7. Now in the beginning I also work with a disabled firewall because the system is in a secured dmz. Is there anything I can do for you ? best regards and stay healthy fatcharly > Gesendet: Montag, 18. Januar 2021 um 22:16 Uhr > Von: "Brent Jensen" > An: fatcha...@gmx.de > Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started > > Are you getting the the cluster to switch over when doing a 'pcs node > standby' on the promoted node? I have a super basic config w/ your same > software versions and I cannot get the slave to promote (constantly > looping errors such as "Refusing to be Primary while peer is not > outdated", until I do a drbdadm up on the standby node). I > appreciate your input. Brent > > On 1/18/2021 9:49 AM, fatcha...@gmx.de wrote: > > Sorry Guys, > > > > problem found, it was a colocation constrain of the drbd-devices. > > > > best regards > > > > fatcharly > > > > > > > >> Gesendet: Montag, 18. Januar 2021 um 16:52 Uhr > >> Von: fatcha...@gmx.de > >> An: "clusterlabs" > >> Betreff: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started > >> > >> Hi, > >> > >> I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm > >> using this versions: > >> > >> kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64 > >> drbd90-utils-9.13.1-1.el8.elrepo.x86_64 > >> > >> pacemaker-cluster-libs-2.0.4-6.el8.x86_64 > >> pacemaker-cli-2.0.4-6.el8.x86_64 > >> pacemaker-schemas-2.0.4-6.el8.noarch > >> pacemaker-2.0.4-6.el8.x86_64 > >> pacemaker-libs-2.0.4-6.el8.x86_64 > >> > >> clusternode-names are lisbon and susanne > >> > >> There are two drbd-resources configured which are paired with > >> filesystem-resources. Both are working in simple master/slave > >> configuration. > >> When I start up the cluster one resource is promoted to a master and a > >> slave, the other is just getting on one the master started, but the slave > >> is not starting up. > >> > >> Status of the cluster: > >> Cluster Summary: > >>* Stack: corosync > >>* Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - partition with > >> quorum > >>* Last updated: Mon Jan 18 16:30:21 2021 > >>* Last change: Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon > >>* 2 nodes configured > >>* 7 resource instances configured > >> > >> Node List: > >>* Online: [ lisbon susanne ] > >> > >> Active Resources: > >>* HA-IP_1 (ocf::heartbeat:IPaddr2):Started susanne > >>* Clone Set: drbd_database-clone [drbd_database] (promotable): > >> * Masters: [ susanne ] > >> * Slaves: [ lisbon ] > >>* fs_database (ocf::heartbeat:Filesystem): Started susanne > >>* Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable): > >> * Masters: [ susanne ] > >>* fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > >> > >> drbdadm status > >> > >> [root@susanne ~]# drbdadm status > >> drbd1 role:Primary > >>disk:UpToDate > >>lisbon role:Secondary > >> peer-disk:UpToDate > >> > >> drbd2 role:Primary > >>disk:UpToDate > >>lisbon connection:Connecting > >> > >> [root@lisbon ~]# drbdadm status > >> drbd1 role:Secondary > >>disk:UpToDate > >>susanne role:Primary > >> peer-disk:UpToDate > >> > >> > >> cluster-config: > >> Cluster Name: mysql_cluster > >> Corosync Nodes: > >> susanne lisbon > >> Pacemaker Nodes: > >> lisbon susanne > >> > >> Resources: > >> Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2) > >>Attributes: cidr_netmask=24 ip=192.168.18.150 > >>Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s) > >>start interval=0s timeout=20s (HA-IP_1-start-interval-0s) > >>stop interval=0s timeout=20s (HA-IP_1-stop-int
Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
Thanks Ken, I will give it a try. best regards and stay healthy fatcharly > Gesendet: Montag, 18. Januar 2021 um 22:49 Uhr > Von: "Ken Gaillot" > An: "Cluster Labs - All topics related to open-source clustering welcomed" > > Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation > > On Mon, 2021-01-18 at 18:43 +0100, fatcha...@gmx.de wrote: > > Hi again, > > > > I need some help to figure out how to let a two node cluster with two > > drbd-devices start the master-devices on the same node. > > How can I configure colocation to work that way ? I tried to bond one > > drbd-device with the other but that didn't work out quite well. > > Once you have the DRBD itself working, you can create a colocation > constraint specifying just the master role. In pcs it's "pcs constraint > colocation add Master with Master ". > > Keep in mind that the dependent resource ('s master role in this > case) will not be able to start if the other resource is not active. > > > This is my config: > > > > I'm installing a 2 node pacemaker/drbd cluster on a CentOS 8.3. > > > > I'm using this versions: > > > > > > > > kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64 > > > > drbd90-utils-9.13.1-1.el8.elrepo.x86_64 > > > > > > > > pacemaker-cluster-libs-2.0.4-6.el8.x86_64 > > > > pacemaker-cli-2.0.4-6.el8.x86_64 > > > > pacemaker-schemas-2.0.4-6.el8.noarch > > > > pacemaker-2.0.4-6.el8.x86_64 > > > > pacemaker-libs-2.0.4-6.el8.x86_64 > > > > > > > > clusternode-names are lisbon and susanne > > > > > > > > Status of the cluster: > > > > Cluster Summary: > > > > * Stack: corosync > > > > * Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - > > > > partition with quorum > > > > * Last updated: Mon Jan 18 16:30:21 2021 > > > > * Last change: Mon Jan 18 16:30:17 2021 by root via cibadmin > > > > on lisbon > > > > * 2 nodes configured > > > > * 7 resource instances configured > > > > > > > > Node List: > > > > * Online: [ lisbon susanne ] > > > > > > > > Active Resources: > > > > * HA-IP_1 (ocf::heartbeat:IPaddr2):Started susanne > > > > * Clone Set: drbd_database-clone [drbd_database] (promotable): > > > > * Masters: [ susanne ] > > > > * Slaves: [ lisbon ] > > > > * fs_database (ocf::heartbeat:Filesystem): Started susanne > > > > * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] > > > > (promotable): > > > > * Masters: [ susanne ] > > > > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > > > > > > > > drbdadm status > > > > > > > > [root@susanne ~]# drbdadm status > > > > drbd1 role:Primary > > > > disk:UpToDate > > > > lisbon role:Secondary > > > > peer-disk:UpToDate > > > > > > > > drbd2 role:Primary > > > > disk:UpToDate > > > > lisbon connection:Connecting > > > > > > > > [root@lisbon ~]# drbdadm status > > > > drbd1 role:Secondary > > > > disk:UpToDate > > > > susanne role:Primary > > > > peer-disk:UpToDate > > > > > > > > > > > > cluster-config: > > > > Cluster Name: mysql_cluster > > > > Corosync Nodes: > > > > susanne lisbon > > > > Pacemaker Nodes: > > > > lisbon susanne > > > > > > > > Resources: > > > > Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2) > > > > Attributes: cidr_netmask=24 ip=192.168.18.150 > > > > Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s) > > > > start interval=0s timeout=20s (HA-IP_1-start- > > > > interval-0s) > > > > stop interval=0s timeout=20s (HA-IP_1-stop- > > > > interval-0s) > > > > Clone: drbd_database-clone > > > > Meta Attrs: clone-max=2 clone-node-max=1 notify=true > > > > promotable=true promoted-max=1 promoted-node-max=1 > > > > Resource: drbd_database (class=ocf provider=linbit type=drbd) > > > >Attributes: drbd_resource=drbd1 > > > >Operations: demote interval=0s timeout=90 (drbd_database- > > > > demote-interval-0s
Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
Hi Brent, this was an old status messages, pleasse ignore this, problem was already solved. Kind regards and stay healthy fatcharly > Gesendet: Montag, 18. Januar 2021 um 22:53 Uhr > Von: "Brent Jensen" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation > > Yes all works fine outside of the cluster. No firewall running nor any > selinux. > > On 1/18/2021 11:53 AM, Strahil Nikolov wrote: > >>> [root@susanne ~]# drbdadm status > >>> drbd1 role:Primary > >>> disk:UpToDate > >>> lisbon role:Secondary > >>> peer-disk:UpToDate > >>> > >>> drbd2 role:Primary > >>> disk:UpToDate > >>> lisbon connection:Connecting > > Are you sure you got the firewall open ? This node should be connected. > > > > Try to verify that each drbd is up and running and promoting any of the 2 > > nodes is possible before proceeding with the cluster setup. > > > > > > Best Regards, > > Strahil Nikolov > > ___ > > Manage your subscription: > > https://lists.clusterlabs.org/mailman/listinfo/users > > > > ClusterLabs home: https://www.clusterlabs.org/ > > -- > This email has been checked for viruses by Avast antivirus software. > https://www.avast.com/antivirus > > ___ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ > ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
Hi Nikolov, sorry, this is an old status, status is right now uptodate. Please ignore this messages. Best regards and stay healthy fatcharly > Gesendet: Dienstag, 19. Januar 2021 um 05:23 Uhr > Von: "Strahil Nikolov" > An: "users@clusterlabs.org" > Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation > > So why it is saying 'connecting' ? > > Best Regards, > Strahil Nikolov > > > > > > > В понеделник, 18 януари 2021 г., 23:54:02 Гринуич+2, Brent Jensen > написа: > > > > > > Yes all works fine outside of the cluster. No firewall running nor any > selinux. > > On 1/18/2021 11:53 AM, Strahil Nikolov wrote: > >>> [root@susanne ~]# drbdadm status > >>> drbd1 role:Primary > >>> disk:UpToDate > >>> lisbon role:Secondary > >>> peer-disk:UpToDate > >>> > >>> drbd2 role:Primary > >>> disk:UpToDate > >>> lisbon connection:Connecting > > Are you sure you got the firewall open ? This node should be connected. > > > > Try to verify that each drbd is up and running and promoting any of the 2 > > nodes is possible before proceeding with the cluster setup. > > > > > > Best Regards, > > Strahil Nikolov > > > ___ > > Manage your subscription: > > https://lists.clusterlabs.org/mailman/listinfo/users > > > > ClusterLabs home: https://www.clusterlabs.org/ > > -- > This email has been checked for viruses by Avast antivirus software. > https://www.avast.com/antivirus > > ___ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ > > ___ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ > ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] CentOS 8 & drbd 9, two drbd devices and colocation
rval=10 role=Master timeout=20 > > (drbd_logsfiles-monitor-interval-10) > >notify interval=0s timeout=90 > > (drbd_logsfiles-notify-interval-0s) > >promote interval=0s timeout=90 > > (drbd_logsfiles-promote-interval-0s) > >reload interval=0s timeout=30 > > (drbd_logsfiles-reload-interval-0s) > >start interval=0s timeout=240 > > (drbd_logsfiles-start-interval-0s) > >stop interval=0s timeout=100 > > (drbd_logsfiles-stop-interval-0s) > > Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem) > > Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4 > > Operations: monitor interval=20s timeout=40s > > (fs_logfiles-monitor-interval-20s) > > start interval=0s timeout=60s (fs_logfiles-start-interval-0s) > > stop interval=0s timeout=60s (fs_logfiles-stop-interval-0s) > > > > Stonith Devices: > > Fencing Levels: > > > > Location Constraints: > > Ordering Constraints: > > start drbd_database-clone then start fs_database (kind:Mandatory) > > (id:order-drbd_database-clone-fs_database-mandatory) > > start drbd_logsfiles-clone then start fs_logfiles (kind:Mandatory) > > (id:order-drbd_logsfiles-clone-fs_logfiles-mandatory) > > Colocation Constraints: > > fs_database with drbd_database-clone (score:INFINITY) > > (with-rsc-role:Master) > > (id:colocation-fs_database-drbd_database-clone-INFINITY) > > fs_logfiles with drbd_logsfiles-clone (score:INFINITY) > > (with-rsc-role:Master) > > (id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY) ERROR:>>>>> drbd_logsfiles-clone with drbd_database-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-drbd_logsfiles-clone-drbd_database-clone-INFINITY) > > Ticket Constraints: > > > > Alerts: > > No alerts defined > > > > Resources Defaults: > > No defaults set > > Operations Defaults: > > No defaults set > > > > Cluster Properties: > > cluster-infrastructure: corosync > > cluster-name: mysql_cluster > > dc-version: 2.0.4-6.el8-2deceaa3ae > > have-watchdog: false > > last-lrm-refresh: 1610382881 > > stonith-enabled: false > > > > Tags: > > No tags defined > > > > Quorum: > > Options: Any suggestions are welcome stay safe and healty fatcharly ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started
Sorry Guys, problem found, it was a colocation constrain of the drbd-devices. best regards fatcharly > Gesendet: Montag, 18. Januar 2021 um 16:52 Uhr > Von: fatcha...@gmx.de > An: "clusterlabs" > Betreff: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started > > Hi, > > I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm > using this versions: > > kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64 > drbd90-utils-9.13.1-1.el8.elrepo.x86_64 > > pacemaker-cluster-libs-2.0.4-6.el8.x86_64 > pacemaker-cli-2.0.4-6.el8.x86_64 > pacemaker-schemas-2.0.4-6.el8.noarch > pacemaker-2.0.4-6.el8.x86_64 > pacemaker-libs-2.0.4-6.el8.x86_64 > > clusternode-names are lisbon and susanne > > There are two drbd-resources configured which are paired with > filesystem-resources. Both are working in simple master/slave configuration. > When I start up the cluster one resource is promoted to a master and a slave, > the other is just getting on one the master started, but the slave is not > starting up. > > Status of the cluster: > Cluster Summary: > * Stack: corosync > * Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - partition with > quorum > * Last updated: Mon Jan 18 16:30:21 2021 > * Last change: Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon > * 2 nodes configured > * 7 resource instances configured > > Node List: > * Online: [ lisbon susanne ] > > Active Resources: > * HA-IP_1 (ocf::heartbeat:IPaddr2):Started susanne > * Clone Set: drbd_database-clone [drbd_database] (promotable): > * Masters: [ susanne ] > * Slaves: [ lisbon ] > * fs_database (ocf::heartbeat:Filesystem): Started susanne > * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable): > * Masters: [ susanne ] > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > > drbdadm status > > [root@susanne ~]# drbdadm status > drbd1 role:Primary > disk:UpToDate > lisbon role:Secondary > peer-disk:UpToDate > > drbd2 role:Primary > disk:UpToDate > lisbon connection:Connecting > > [root@lisbon ~]# drbdadm status > drbd1 role:Secondary > disk:UpToDate > susanne role:Primary > peer-disk:UpToDate > > > cluster-config: > Cluster Name: mysql_cluster > Corosync Nodes: > susanne lisbon > Pacemaker Nodes: > lisbon susanne > > Resources: > Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2) > Attributes: cidr_netmask=24 ip=192.168.18.150 > Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s) > start interval=0s timeout=20s (HA-IP_1-start-interval-0s) > stop interval=0s timeout=20s (HA-IP_1-stop-interval-0s) > Clone: drbd_database-clone > Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true > promoted-max=1 promoted-node-max=1 > Resource: drbd_database (class=ocf provider=linbit type=drbd) >Attributes: drbd_resource=drbd1 >Operations: demote interval=0s timeout=90 > (drbd_database-demote-interval-0s) >monitor interval=20 role=Slave timeout=20 > (drbd_database-monitor-interval-20) >monitor interval=10 role=Master timeout=20 > (drbd_database-monitor-interval-10) >notify interval=0s timeout=90 > (drbd_database-notify-interval-0s) >promote interval=0s timeout=90 > (drbd_database-promote-interval-0s) >reload interval=0s timeout=30 > (drbd_database-reload-interval-0s) >start interval=0s timeout=240 (drbd_database-start-interval-0s) >stop interval=0s timeout=100 (drbd_database-stop-interval-0s) > Resource: fs_database (class=ocf provider=heartbeat type=Filesystem) > Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4 > Operations: monitor interval=20s timeout=40s > (fs_database-monitor-interval-20s) > start interval=0s timeout=60s (fs_database-start-interval-0s) > stop interval=0s timeout=60s (fs_database-stop-interval-0s) > Clone: drbd_logsfiles-clone > Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true > promoted-max=1 promoted-node-max=1 > Resource: drbd_logsfiles (class=ocf provider=linbit type=drbd) >Attributes: drbd_resource=drbd2 >Operations: demote interval=0s timeout=90 > (drbd_logsfiles-demote-interval-0s) >monitor interval=20 role=Slave timeout=20 > (drbd_logsfiles-monitor-interval-20) >monitor interval=10 role=Master timeout=20 > (drbd_logsfiles-monitor-interval-10) >notify interval=0s timeout=90 > (drbd_logsfiles-notify-interval-0s)
[ClusterLabs] CentOS 8 & drbd 9 second slave is not started
base-clone-fs_database-mandatory) start drbd_logsfiles-clone then start fs_logfiles (kind:Mandatory) (id:order-drbd_logsfiles-clone-fs_logfiles-mandatory) Colocation Constraints: fs_database with drbd_database-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-fs_database-drbd_database-clone-INFINITY) fs_logfiles with drbd_logsfiles-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY) drbd_logsfiles-clone with drbd_database-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-drbd_logsfiles-clone-drbd_database-clone-INFINITY) Ticket Constraints: Alerts: No alerts defined Resources Defaults: No defaults set Operations Defaults: No defaults set Cluster Properties: cluster-infrastructure: corosync cluster-name: mysql_cluster dc-version: 2.0.4-6.el8-2deceaa3ae have-watchdog: false last-lrm-refresh: 1610382881 stonith-enabled: false Tags: No tags defined Quorum: Options: Any suggestions are welcome stay safe and healty fatcharly ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] drbd9 and Centos 8.3
Hi, I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm using this versions: kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64 drbd90-utils-9.13.1-1.el8.elrepo.x86_64 pacemaker-cluster-libs-2.0.4-6.el8.x86_64 pacemaker-cli-2.0.4-6.el8.x86_64 pacemaker-schemas-2.0.4-6.el8.noarch pacemaker-2.0.4-6.el8.x86_64 pacemaker-libs-2.0.4-6.el8.x86_64 clusternode-names are lisbon and susanne when I try to install my drbd resource with pcs resource create drbd_database ocf:linbit:drbd drbd_resource=drbd1 promotable promoted-max=1 promoted-node-max=1 clone-max=2 clone-node-max=1 notify=true I get this information in the cluster information: * drbd_database_monitor_0 on lisbon 'not installed' (5): call=20, status='complete', exitreason='DRBD kernel (module) not available?', last-rc-change='2021-01-08 18:03:10 +01:00', queued=0ms, exec=96ms The drbd-service is not started, shouldn't it be started by the clustermanager ? Any suggestions are welcome Have a nice weekand and stay healthy fatcharly ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] cluster problems after let's encrypt
Hi, I'm running a two node corosync httpd-cluster on a CentOS 7. corosync-2.4.5-4.el7.x86_64 pcs-0.9.168-4.el7.centos.x86_64 Today I used lets encrypt to installt https for two domains on that system. After that the node with the new https-domains is not longer able to hold the apache resource: The resource is configured like this: Resource: apache (class=ocf provider=heartbeat type=apache) Attributes: configfile=/etc/httpd/conf/httpd.conf statusurl=http://127.0.0.1:8089/server-status Operations: start interval=0s timeout=40s (apache-start-interval-0s) stop interval=0s timeout=60s (apache-stop-interval-0s) monitor interval=1min (apache-monitor-interval-1min) The status-page is configured like this: Listen 127.0.0.1:8089 SetHandler server-status Order deny,allow Deny from all Allow from 127.0.0.1 And the log shows this: Jul 6 16:55:18 bachi2 apache(apache)[7182]: INFO: waiting for apache /etc/httpd/conf/httpd.conf to come up Jul 6 16:55:19 bachi2 apache(apache)[7182]: INFO: apache not running Jul 6 16:55:19 bachi2 apache(apache)[7182]: INFO: waiting for apache /etc/httpd/conf/httpd.conf to come up Jul 6 16:55:20 bachi2 apache(apache)[7182]: INFO: apache not running Jul 6 16:55:20 bachi2 apache(apache)[7182]: INFO: waiting for apache /etc/httpd/conf/httpd.conf to come up Jul 6 16:55:21 bachi2 lrmd[8756]: warning: apache_start_0 process (PID 7182) timed out Jul 6 16:55:21 bachi2 lrmd[8756]: warning: apache_start_0:7182 - timed out after 4ms Jul 6 16:55:21 bachi2 crmd[8759]: error: Result of start operation for apache on bachi2: Timed Out Jul 6 16:55:21 bachi2 apache(apache)[8416]: INFO: apache is not running. Any suggestions are welcome Best regards fatcharly ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
[ClusterLabs] failure-timeout
Hi, I´m running a pcs/corosync two node cluster on CentOS 7.6 I use a cloned ping resource and I'd like to add a failure-timeout to this. How do I do this ? Any suggestions are welcome Kind regards fatcharly ___ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
Re: [ClusterLabs] pacemaker apache and umask on CentOS 7
> Gesendet: Mittwoch, 20. April 2016 um 19:35 Uhr > Von: "Ken Gaillot" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] pacemaker apache and umask on CentOS 7 > > On 04/20/2016 12:20 PM, Klaus Wenninger wrote: > > On 04/20/2016 05:35 PM, fatcha...@gmx.de wrote: > >> > >>> Gesendet: Mittwoch, 20. April 2016 um 16:31 Uhr > >>> Von: "Klaus Wenninger" > >>> An: users@clusterlabs.org > >>> Betreff: Re: [ClusterLabs] pacemaker apache and umask on CentOS 7 > >>> > >>> On 04/20/2016 04:11 PM, fatcha...@gmx.de wrote: > >>>> Hi, > >>>> > >>>> I´m running a 2-node apache webcluster on a fully patched CentOS 7 > >>>> (pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64). > >>>> Some files which are generated by the apache are created with a umask > >>>> 137 but I need this files created with a umask of 117. > >>>> To change this I first tried to add a umask 117 to /etc/sysconfig/httpd > >>>> & rebooted the system. This had no effekt. > >>>> So I found out (after some research) that this is not working under > >>>> CentOS 7 and that this had to be changed via systemd. > >>>> So I created a directory "/etc/systemd/system/httpd.service.d" and put > >>>> there a "umask.conf"-File with this content: > >>>> [Service] > >>>> UMask=0117 > >>>> > >>>> Again I rebooted the system but no effekt. > >>>> Is the pacemaker really starting the apache over the systemd ? And how > >>>> can I solve the problem ? > >>> Didn't check with CentOS7 but on RHEL7 there is a > >>> /usr/lib/ocf/resource.d/heartbeat/apache. > >>> So it depends on how you defined the resource starting apache if systemd > >>> is used or if it being done by the ocf-ra. > >> MY configuration is: > >> Resource: apache (class=ocf provider=heartbeat type=apache) > >> Attributes: configfile=/etc/httpd/conf/httpd.conf > >> statusurl=http://127.0.0.1:8089/server-status > >> Operations: start interval=0s timeout=40s (apache-start-timeout-40s) > >> stop interval=0s timeout=60s (apache-stop-timeout-60s) > >> monitor interval=1min (apache-monitor-interval-1min) > >> > >> So I quess it is ocf. But what will be the right way to do it ? I lack a > >> bit of understandig about this /usr/lib/ocf/resource.d/heartbeat/apache > >> file. > >> > > There are the ocf-Resource-Agents (if there is none you can always > > create one for your service) which usually > > give you a little bit more control of the service from the cib. (You can > > set a couple of variables like in this example > > the pointer to the config-file) > > And of course you can always create resources referring the native > > services of your distro (systemd-units in > > this case). > >> > >> > >> > >>>> Any suggestions are welcome > > If you add envfiles="/etc/sysconfig/httpd" to your apache resource, it > should work. worked like a charm. Thanks to everybody for your support. Kind regards fatcharly > > >>>> Kind regards > >>>> > >>>> fatcharly > > ___ > Users mailing list: Users@clusterlabs.org > http://clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > ___ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [ClusterLabs] pacemaker apache and umask on CentOS 7
> Gesendet: Mittwoch, 20. April 2016 um 16:31 Uhr > Von: "Klaus Wenninger" > An: users@clusterlabs.org > Betreff: Re: [ClusterLabs] pacemaker apache and umask on CentOS 7 > > On 04/20/2016 04:11 PM, fatcha...@gmx.de wrote: > > Hi, > > > > I´m running a 2-node apache webcluster on a fully patched CentOS 7 > > (pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64). > > Some files which are generated by the apache are created with a umask 137 > > but I need this files created with a umask of 117. > > To change this I first tried to add a umask 117 to /etc/sysconfig/httpd & > > rebooted the system. This had no effekt. > > So I found out (after some research) that this is not working under CentOS > > 7 and that this had to be changed via systemd. > > So I created a directory "/etc/systemd/system/httpd.service.d" and put > > there a "umask.conf"-File with this content: > > [Service] > > UMask=0117 > > > > Again I rebooted the system but no effekt. > > Is the pacemaker really starting the apache over the systemd ? And how can > > I solve the problem ? > Didn't check with CentOS7 but on RHEL7 there is a > /usr/lib/ocf/resource.d/heartbeat/apache. > So it depends on how you defined the resource starting apache if systemd > is used or if it being done by the ocf-ra. MY configuration is: Resource: apache (class=ocf provider=heartbeat type=apache) Attributes: configfile=/etc/httpd/conf/httpd.conf statusurl=http://127.0.0.1:8089/server-status Operations: start interval=0s timeout=40s (apache-start-timeout-40s) stop interval=0s timeout=60s (apache-stop-timeout-60s) monitor interval=1min (apache-monitor-interval-1min) So I quess it is ocf. But what will be the right way to do it ? I lack a bit of understandig about this /usr/lib/ocf/resource.d/heartbeat/apache file. > > > > Any suggestions are welcome > > > > Kind regards > > > > fatcharly > > > > > > > > ___ > > Users mailing list: Users@clusterlabs.org > > http://clusterlabs.org/mailman/listinfo/users > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > ___ > Users mailing list: Users@clusterlabs.org > http://clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > ___ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] pacemaker apache and umask on CentOS 7
Hi, I´m running a 2-node apache webcluster on a fully patched CentOS 7 (pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64). Some files which are generated by the apache are created with a umask 137 but I need this files created with a umask of 117. To change this I first tried to add a umask 117 to /etc/sysconfig/httpd & rebooted the system. This had no effekt. So I found out (after some research) that this is not working under CentOS 7 and that this had to be changed via systemd. So I created a directory "/etc/systemd/system/httpd.service.d" and put there a "umask.conf"-File with this content: [Service] UMask=0117 Again I rebooted the system but no effekt. Is the pacemaker really starting the apache over the systemd ? And how can I solve the problem ? Any suggestions are welcome Kind regards fatcharly ___ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[ClusterLabs] cloned pingd resource problem
Hi, I`m running a two node cluster on a fully updated CentOS 7 (pacemaker-1.1.13-10.el7_2.2.x86_64 pcs-0.9.143-15.el7.x86_64) . I see on one of our nodes a lot of this in the logfiles: Mar 30 12:32:13 localhost crmd[12986]: notice: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ] Mar 30 12:32:13 localhost pengine[12985]: notice: On loss of CCM Quorum: Ignore Mar 30 12:32:13 localhost pengine[12985]: warning: Processing failed op monitor for ping_fw:0 on kathie2: unknown error (1) Mar 30 12:32:13 localhost pengine[12985]: warning: Processing failed op start for ping_fw:1 on stacy2: unknown error (1) Mar 30 12:32:13 localhost pengine[12985]: warning: Forcing ping_fw-clone away from stacy2 after 100 failures (max=100) Mar 30 12:32:13 localhost pengine[12985]: warning: Forcing ping_fw-clone away from stacy2 after 100 failures (max=100) Mar 30 12:32:13 localhost pengine[12985]: notice: Calculated Transition 1823: /var/lib/pacemaker/pengine/pe-input-355.bz2 Mar 30 12:32:13 localhost crmd[12986]: notice: Transition 1823 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-355.bz2): Complete Mar 30 12:32:13 localhost crmd[12986]: notice: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ] The configuration looks like this: Clone: ping_fw-clone Resource: ping_fw (class=ocf provider=pacemaker type=ping) Attributes: dampen=5s multiplier=1000 host_list=192.168.16.1 timeout=60 Operations: start interval=0s timeout=60 (ping_fw-start-interval-0s) stop interval=0s timeout=20 (ping_fw-stop-interval-0s) monitor interval=45 (ping_fw-monitor-interval-45) What can I do to resolve the problem ? Any suggestions are welcome Kind regards fatcharly ___ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org