You've added the high-availability service to the public zone. Can you verify that the interface you're using for pcsd is bound to the public zone?
Check whether the https_proxy environment variable is set. You may need to unset it or to add it to /etc/sysconfig/pcsd. Just an FYI, the `--name` option isn't used with `pcs cluster auth`. It looks like it gets ignored though. On Mon, Nov 30, 2020 at 5:47 AM John Karippery <john.karipp...@yahoo.com> wrote: > > I have problem while setup pacemaker on debian 9 servers. i have 3 server and > I installed > > apt install pacemaker, corosync, pcsdfirewalld fence-agents > > > pcs status > > pcs status > Cluster name: vipcluster > Stack: corosync > Current DC: server1 (version 1.1.16-94ff4df) - partition with quorum > Last updated: Mon Nov 30 14:43:36 2020 > Last change: Mon Nov 30 13:03:09 2020 by root via cibadmin on server1 > > 3 nodes configured > 2 resources configured > > Online: [ server1 server2 server3 ] > > Full list of resources: > > MasterVip (ocf::heartbeat:IPaddr2): Started server1 > Apache (ocf::heartbeat:apache): Started server1 > > Daemon Status: > corosync: active/enabled > pacemaker: active/enabled > pcsd: active/enabled > > > > > pcs cluster status > Cluster Status: > Stack: corosync > Current DC: server1 (version 1.1.16-94ff4df) - partition with quorum > Last updated: Mon Nov 30 14:44:55 2020 > Last change: Mon Nov 30 13:03:09 2020 by root via cibadmin on server1 > 3 nodes configured > 2 resources configured > > PCSD Status: > server3: Offline > server1: Offline > server2: Offline > > > > error is showing while pcs auth > > pcs cluster auth server1 server2 server3 --name vipcluster -u hacluster -p > 12345678 --debug > Running: /usr/bin/ruby -I/usr/share/pcsd/ /usr/share/pcsd/pcsd-cli.rb auth > --Debug Input Start-- > {"username": "hacluster", "local": false, "nodes": ["server1", "server2", > "server3"], "password": "12345678", "force": false} > --Debug Input End-- > Return Value: 0 > --Debug Output Start-- > { > "status": "ok", > "data": { > "auth_responses": { > "server1": { > "status": "noresponse" > }, > "server2": { > "status": "noresponse" > }, > "server3": { > "status": "noresponse" > } > }, > "sync_successful": true, > "sync_nodes_err": [ > > ], > "sync_responses": { > } > }, > "log": [ > "I, [2020-11-30T14:46:24.758862 #9677] INFO -- : PCSD Debugging > enabled\n", > "D, [2020-11-30T14:46:24.758900 #9677] DEBUG -- : Did not detect RHEL > 6\n", > "I, [2020-11-30T14:46:24.758919 #9677] INFO -- : Running: > /usr/sbin/corosync-cmapctl totem.cluster_name\n", > "I, [2020-11-30T14:46:24.758931 #9677] INFO -- : CIB USER: hacluster, > groups: \n", > "D, [2020-11-30T14:46:24.770175 #9677] DEBUG -- : [\"totem.cluster_name > (str) = vipcluster\\n\"]\n", > "D, [2020-11-30T14:46:24.770373 #9677] DEBUG -- : []\n", > "D, [2020-11-30T14:46:24.770444 #9677] DEBUG -- : Duration: > 0.011215661s\n", > "I, [2020-11-30T14:46:24.770585 #9677] INFO -- : Return Value: 0\n", > "I, [2020-11-30T14:46:24.772514 #9677] INFO -- : SRWT Node: server1 > Request: check_auth\n", > "E, [2020-11-30T14:46:24.772628 #9677] ERROR -- : Unable to connect to > node server1, no token available\n", > "I, [2020-11-30T14:46:24.772943 #9677] INFO -- : SRWT Node: server2 > Request: check_auth\n", > "E, [2020-11-30T14:46:24.773032 #9677] ERROR -- : Unable to connect to > node server2, no token available\n", > "I, [2020-11-30T14:46:24.773202 #9677] INFO -- : SRWT Node: server3 > Request: check_auth\n", > "E, [2020-11-30T14:46:24.773278 #9677] ERROR -- : Unable to connect to > node server3, no token available\n" > ] > } > --Debug Output End-- > > Error: Unable to communicate with server1 > Error: Unable to communicate with server2 > Error: Unable to communicate with server3 > > > firewall settings > > > # firewall-cmd --permanent --add-service=high-availability > Warning: ALREADY_ENABLED: high-availability > success > ~# firewall-cmd --add-service=high-availability > Warning: ALREADY_ENABLED: 'high-availability' already in 'public' > success > > > > > > > > _______________________________________________ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ -- Regards, Reid Wahl, RHCA Senior Software Maintenance Engineer, Red Hat CEE - Platform Support Delivery - ClusterHA _______________________________________________ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/