I did not disable Selinux. I disabled and i tried its working okay. On Mon, Aug 3, 2015 at 6:36 PM, emmanuel segura <[email protected]> wrote:
> so you are using cman+pacemaker, can you provide your config and tell > us if you are using fencing. > > Thanks > Emmanuel > > 2015-08-03 14:37 GMT+02:00 Vijay Partha <[email protected]>: > > service cman start > > Starting cluster: > > Checking if cluster has been disabled at boot... [ OK ] > > Checking Network Manager... [ OK ] > > Global setup... [ OK ] > > Loading kernel modules... [ OK ] > > Mounting configfs... [ OK ] > > Starting cman... [ OK ] > > Waiting for quorum... [ OK ] > > Starting fenced... [ OK ] > > Starting dlm_controld... [ OK ] > > Tuning DLM kernel config... [ OK ] > > Starting gfs_controld... [ OK ] > > Unfencing self... [ OK ] > > Joining fence domain... > > > > It doesnt go beyond this. > > > > This is my pacemaker log > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > crm_ipc_connect: > > Could not establish pacemakerd connection: Connection refused (111) > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > config_find_next: > > Processing additional service options... > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'corosync_quorum' for option: name > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > config_find_next: > > Processing additional service options... > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'corosync_cman' for option: name > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > config_find_next: > > Processing additional service options... > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'openais_ckpt' for option: name > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > config_find_next: > > No additional configuration supplied for: service > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > config_find_next: > > Processing additional quorum options... > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'quorum_cman' for option: provider > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > get_cluster_type: > > Detected an active 'cman' cluster > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > mcp_read_config: > > Reading configure for stack: cman > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > config_find_next: > > Processing additional logging options... > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Defaulting to 'off' for option: debug > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'yes' for option: to_logfile > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found '/var/log/cluster/corosync.log' for option: logfile > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: notice: > crm_add_logfile: > > Additional logging available in /var/log/cluster/corosync.log > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'yes' for option: to_syslog > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: get_config_opt: > > Found 'local4' for option: syslog_facility > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: notice: main: > > Starting Pacemaker 1.1.11 (Build: 97629de): generated-manpages > > agent-manpages ascii-docs ncurses libqb-logging libqb-ipc nagios > > corosync-plugin cman acls > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: main: > > Maximum core file size is: 18446744073709551615 > > Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: > > qb_ipcs_us_publish: server name: pacemakerd > > Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd: error: > > cluster_connect_cpg: Could not join the CPG group 'pacemakerd': 6 > > Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd: error: main: > > Couldn't connect to Corosync's CPG service > > Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd: info: > crm_xml_cleanup: > > Cleaning up memory from libxml2 > > > > this is my message log > > > > Aug 3 14:35:31 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error > > retrying > > Aug 3 14:35:33 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my > > cluster] failed to contact node 10.61.40.194 > > Aug 3 14:35:33 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no > > answer from any [my cluster] datasource > > Aug 3 14:35:34 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying > > Aug 3 14:35:37 vmx-occ-004 /usr/sbin/gmond[4590]: Error creating > multicast > > server mcast_join=10.61.40.194 port=8649 mcast_if=NULL family='inet4'. > Will > > try again...#012 > > Aug 3 14:35:38 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error > > retrying > > Aug 3 14:35:41 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error > > retrying > > Aug 3 14:35:44 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying > > Aug 3 14:35:48 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my > > cluster] failed to contact node 10.61.40.194 > > Aug 3 14:35:48 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no > > answer from any [my cluster] datasource > > Aug 3 14:35:48 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error > > retrying > > Aug 3 14:35:51 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error > > retrying > > Aug 3 14:35:54 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying > > Aug 3 14:35:58 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error > > retrying > > Aug 3 14:36:01 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error > > retrying > > Aug 3 14:36:03 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my > > cluster] failed to contact node 10.61.40.194 > > Aug 3 14:36:03 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no > > answer from any [my cluster] datasource > > Aug 3 14:36:04 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying > > Aug 3 14:36:08 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error > > retrying > > Aug 3 14:36:11 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error > > retrying > > Aug 3 14:36:14 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying > > Aug 3 14:36:18 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error > > retrying > > Aug 3 14:36:18 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my > > cluster] failed to contact node 10.61.40.194 > > Aug 3 14:36:18 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no > > answer from any [my cluster] datasource > > Aug 3 14:36:21 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error > > retrying > > Aug 3 14:36:24 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying > > > > > > On Mon, Aug 3, 2015 at 6:02 PM, emmanuel segura <[email protected]> > wrote: > >> > >> Sorry, but I think is more easy to help you, If you provide more > >> information about your problem. > >> > >> 2015-08-03 14:14 GMT+02:00 Vijay Partha <[email protected]>: > >> > Hi > >> > > >> > When i start cman it hangs in joining fence domain. > >> > > >> > this is my message log. > >> > > >> > > >> > Aug 3 14:12:16 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:19 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:24 vmx-occ-005 fenced[2098]: daemon cpg_join error > retrying > >> > Aug 3 14:12:27 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:29 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:34 vmx-occ-005 fenced[2098]: daemon cpg_join error > retrying > >> > Aug 3 14:12:37 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:39 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:44 vmx-occ-005 fenced[2098]: daemon cpg_join error > retrying > >> > Aug 3 14:12:47 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:49 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error > >> > retrying > >> > Aug 3 14:12:54 vmx-occ-005 fenced[2098]: daemon cpg_join error > retrying > >> > > >> > How to solve this issue? > >> > -- > >> > With Regards > >> > P.Vijay > >> > > >> > _______________________________________________ > >> > Users mailing list: [email protected] > >> > http://clusterlabs.org/mailman/listinfo/users > >> > > >> > Project Home: http://www.clusterlabs.org > >> > Getting started: > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >> > Bugs: http://bugs.clusterlabs.org > >> > > >> > >> > >> > >> -- > >> .~. > >> /V\ > >> // \\ > >> /( )\ > >> ^`~'^ > >> > >> _______________________________________________ > >> Users mailing list: [email protected] > >> http://clusterlabs.org/mailman/listinfo/users > >> > >> Project Home: http://www.clusterlabs.org > >> Getting started: > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >> Bugs: http://bugs.clusterlabs.org > > > > > > > > > > -- > > With Regards > > P.Vijay > > > > _______________________________________________ > > Users mailing list: [email protected] > > http://clusterlabs.org/mailman/listinfo/users > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > > > > -- > .~. > /V\ > // \\ > /( )\ > ^`~'^ > > _______________________________________________ > Users mailing list: [email protected] > http://clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > -- With Regards P.Vijay
_______________________________________________ Users mailing list: [email protected] http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
