04.09.2013, 01:06, "Andrew Beekhof" <and...@beekhof.net>:
> On 03/09/2013, at 11:46 PM, Andrey Groshev <gre...@yandex.ru> wrote:
>
>>  03.09.2013, 08:27, "Andrew Beekhof" <and...@beekhof.net>:
>>>  On 02/09/2013, at 5:27 PM, Andrey Groshev <gre...@yandex.ru> wrote:
>>>>   30.08.2013, 07:18, "Andrew Beekhof" <and...@beekhof.net>:
>>>>>   On 29/08/2013, at 7:31 PM, Andrey Groshev <gre...@yandex.ru> wrote:
>>>>>>    29.08.2013, 12:25, "Andrey Groshev" <gre...@yandex.ru>:
>>>>>>>    29.08.2013, 02:55, "Andrew Beekhof" <and...@beekhof.net>:
>>>>>>>>     On 28/08/2013, at 5:38 PM, Andrey Groshev <gre...@yandex.ru> wrote:
>>>>>>>>>      28.08.2013, 04:06, "Andrew Beekhof" <and...@beekhof.net>:
>>>>>>>>>>      On 27/08/2013, at 1:13 PM, Andrey Groshev <gre...@yandex.ru> 
>>>>>>>>>> wrote:
>>>>>>>>>>>       27.08.2013, 05:39, "Andrew Beekhof" <and...@beekhof.net>:
>>>>>>>>>>>>       On 26/08/2013, at 3:09 PM, Andrey Groshev <gre...@yandex.ru> 
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>        26.08.2013, 03:34, "Andrew Beekhof" <and...@beekhof.net>:
>>>>>>>>>>>>>>        On 23/08/2013, at 9:39 PM, Andrey Groshev 
>>>>>>>>>>>>>> <gre...@yandex.ru> wrote:
>>>>>>>>>>>>>>>         Hello,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         Today I try remake my test cluster from cman to 
>>>>>>>>>>>>>>> corosync2.
>>>>>>>>>>>>>>>         I drew attention to the following:
>>>>>>>>>>>>>>>         If I reset cluster with cman through cibadmin --erase 
>>>>>>>>>>>>>>> --force
>>>>>>>>>>>>>>>         In cib is still there exist names of nodes.
>>>>>>>>>>>>>>        Yes, the cluster puts back entries for all the nodes it 
>>>>>>>>>>>>>> know about automagically.
>>>>>>>>>>>>>>>         cibadmin -Ql
>>>>>>>>>>>>>>>         .....
>>>>>>>>>>>>>>>            <nodes>
>>>>>>>>>>>>>>>              <node id="dev-cluster2-node2.unix.tensor.ru" 
>>>>>>>>>>>>>>> uname="dev-cluster2-node2"/>
>>>>>>>>>>>>>>>              <node id="dev-cluster2-node4.unix.tensor.ru" 
>>>>>>>>>>>>>>> uname="dev-cluster2-node4"/>
>>>>>>>>>>>>>>>              <node id="dev-cluster2-node3.unix.tensor.ru" 
>>>>>>>>>>>>>>> uname="dev-cluster2-node3"/>
>>>>>>>>>>>>>>>            </nodes>
>>>>>>>>>>>>>>>         ....
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         Even if cman and pacemaker running only one node.
>>>>>>>>>>>>>>        I'm assuming all three are configured in cluster.conf?
>>>>>>>>>>>>>        Yes, there exist list nodes.
>>>>>>>>>>>>>>>         And if I do too on cluster with corosync2
>>>>>>>>>>>>>>>         I see only names of nodes which run corosync and 
>>>>>>>>>>>>>>> pacemaker.
>>>>>>>>>>>>>>        Since you're not included your config, I can only guess 
>>>>>>>>>>>>>> that your corosync.conf does not have a nodelist.
>>>>>>>>>>>>>>        If it did, you should get the same behaviour.
>>>>>>>>>>>>>        I try and expected_node and nodelist.
>>>>>>>>>>>>       And it didn't work? What version of pacemaker?
>>>>>>>>>>>       It does not work as I expected.
>>>>>>>>>>      Thats because you've used IP addresses in the node list.
>>>>>>>>>>      ie.
>>>>>>>>>>
>>>>>>>>>>      node {
>>>>>>>>>>        ring0_addr: 10.76.157.17
>>>>>>>>>>      }
>>>>>>>>>>
>>>>>>>>>>      try including the node name as well, eg.
>>>>>>>>>>
>>>>>>>>>>      node {
>>>>>>>>>>        name: dev-cluster2-node2
>>>>>>>>>>        ring0_addr: 10.76.157.17
>>>>>>>>>>      }
>>>>>>>>>      The same thing.
>>>>>>>>     I don't know what to say.  I tested it here yesterday and it 
>>>>>>>> worked as expected.
>>>>>>>    I found that the reason that You and I have different results - I 
>>>>>>> did not have reverse DNS zone for these nodes.
>>>>>>>    I know what it should be, but (PACEMAKER + CMAN) worked without a 
>>>>>>> reverse area!
>>>>>>    Hasty. Deleted all. Reinstalled. Configured. Not working again. Damn!
>>>>>   It would have surprised me... pacemaker 1.1.11 doesn't do any dns 
>>>>> lookups - reverse or otherwise.
>>>>>   Can you set
>>>>>
>>>>>    PCMK_trace_files=corosync.c
>>>>>
>>>>>   in your environment and retest?
>>>>>
>>>>>   On RHEL6 that means putting the following in /etc/sysconfig/pacemaker
>>>>>     export PCMK_trace_files=corosync.c
>>>>>
>>>>>   It should produce additional logging[1] that will help diagnose the 
>>>>> issue.
>>>>>
>>>>>   [1] http://blog.clusterlabs.org/blog/2013/pacemaker-logging/
>>>>   Hello, Andrew.
>>>>
>>>>   You are a little misunderstood me.
>>>  No, I understood you fine.
>>>>   I wrote that I rushed to judgment.
>>>>   After I did the reverse DNS zone, the cluster behaved correctly.
>>>>   BUT after I took apart the cluster dropped configs and restarted on the 
>>>> new cluster,
>>>>   cluster again don't showed all the nodes in the nodes (only node with 
>>>> running pacemaker).
>>>>
>>>>   A small portion of the log. Full log
>>>>   In which (I thought) there is something interesting.
>>>>
>>>>   Aug 30 12:31:11 [9986] dev-cluster2-node4        cib: (  corosync.c:423  
>>>>  )   trace: check_message_sanity:      Verfied message 4: (dest=<all>:cib, 
>>>> from=dev-cluster2-node4:cib.9986, compressed=0, size=1551, total=2143)
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (  corosync.c:96   
>>>>  )   trace: corosync_node_name:        Checking 172793107 vs 0 from 
>>>> nodelist.node.0.nodeid
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (      ipcc.c:378  
>>>>  )   debug: qb_ipcc_disconnect:        qb_ipcc_disconnect()
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (ringbuffer.c:294  
>>>>  )   debug: qb_rb_close:       Closing ringbuffer: 
>>>> /dev/shm/qb-cmap-request-9616-9989-27-header
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (ringbuffer.c:294  
>>>>  )   debug: qb_rb_close:       Closing ringbuffer: 
>>>> /dev/shm/qb-cmap-response-9616-9989-27-header
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (ringbuffer.c:294  
>>>>  )   debug: qb_rb_close:       Closing ringbuffer: 
>>>> /dev/shm/qb-cmap-event-9616-9989-27-header
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (  corosync.c:134  
>>>>  )  notice: corosync_node_name:        Unable to get node name for nodeid 
>>>> 172793107
>>>  I wonder if you need to be including the nodeid too. ie.
>>>
>>>  node {
>>>   name: dev-cluster2-node2
>>>   ring0_addr: 10.76.157.17
>>>   nodeid: 2
>>>  }
>>>
>>>  I _thought_ that was implicit.
>>>  Chrissie: is "nodelist.node.%d.nodeid" always available for corosync2 or 
>>> only if explicitly defined in the config?
>>  I begin from end.
>>  "nodelist.node.%d.nodeid" available if defined in 
>> /etc/corosync/corosync.conf.
>>  If I use only "expected_votes", corosync-cmapctl don't show nodelist.
>>
>>  And now... interesting garbage turned out:
>>  I try change config as you advised.
>>
>>  ....
>>  quorum {
>>         provider: corosync_votequorum
>>  }
>>
>>  nodelist {
>>  node {
>>   ring0_addr: 10.76.157.17
>>   name: dev-cluster2-node2
>>   nodeid: 0
>
> 0 is not allowed here. it has a special meaning

Usually, this tells the parser. :)

>
>>  }
>>  node {
>>   ring0_addr: 10.76.157.18
>>   name: dev-cluster2-node3
>>   nodeid: 1
>>  }
>>  node {
>>   ring0_addr: 10.76.157.19
>>   name: dev-cluster2-node4
>>   nodeid: 2
>>  }
>>  }
>>  ......
>>  And!
>>  # service pacemaker start
>>  Starting Pacemaker Cluster Manager:                        [FAILED]
>>
>>  I see logs. (full log: http://pastebin.com/dW7g6URj)
>>  .... snip ...
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:501   
>> )   trace: corosync_initialize_nodelist:     Initializing corosync nodelist
>>  Sep 03 17:28:16 [22111] dev-cluster2-node4 corosync debug   [MAIN  ] 
>> connection created
>>  Sep 03 17:28:16 [22111] dev-cluster2-node4 corosync debug   [CMAP  ] 
>> lib_init_fn: conn=0x7f4de13c9870
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:96    
>> )   trace: corosync_node_name:       Checking 2 vs 0 from 
>> nodelist.node.0.nodeid
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:96    
>> )   trace: corosync_node_name:       Checking 2 vs 1 from 
>> nodelist.node.1.nodeid
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:96    
>> )   trace: corosync_node_name:       Checking 2 vs 2 from 
>> nodelist.node.2.nodeid
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:104   
>> )   trace: corosync_node_name:       Searching for node name for 2 in 
>> nodelist.node.2 (null)
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:108   
>> )   trace: corosync_node_name:       nodelist.node.2.ring0_addr = 
>> 10.76.157.19
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:120   
>> )   trace: corosync_node_name:       nodelist.node.2.name = 
>> dev-cluster2-node4 1
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (  corosync.c:525   
>> )    crit: corosync_initialize_nodelist:     Nodes 2 and 0 share the same 
>> name 'dev-cluster2-node4': shutting down
>>  Sep 03 17:28:16 [22140] dev-cluster2-node4       crmd: (       xml.c:2637  
>> )    info: crm_xml_cleanup:  Cleaning up memory from libxml2
>>  Sep 03 17:28:16 [22135] dev-cluster2-node4        cib: (      ipcs.c:736   
>> )   debug: qb_ipcs_dispatch_connection_request:      HUP conn 
>> (22135-22140-10)
>>  Sep 03 17:28:16 [22135] dev-cluster2-node4        cib: (      ipcs.c:578   
>> )   debug: qb_ipcs_disconnect:       qb_ipcs_disconnect(22135-22140-10) 
>> state:2
>>  Sep 03 17:28:16 [22135] dev-cluster2-node4        cib: (       ipc.c:341   
>> )    info: crm_client_destroy:       Destroying 0 events
>>  ..... snip ....
>>
>>  Ok, try renumbering and received the following: pacemaker starеed, but 
>> nodelist:
>>  .....
>>     <nodes>
>>       <node id="172793107" uname="dev-cluster2-node4.unix.tensor.ru"/>
>>       <node id="3" uname="dev-cluster2-node2.unix.tensor.ru"/>
>>       <node id="4" uname="dev-cluster2-node3.unix.tensor.ru"/>
>>       <node id="5" uname="dev-cluster2-node4.unix.tensor.ru"/>
>>     </nodes>
>>  .....
>>
>>  Ok, I send cibadmin --erase --force.
>>  .....
>>     <nodes>
>>       <node id="4" uname="dev-cluster2-node3.unix.tensor.ru"/>
>>       <node id="5" uname="dev-cluster2-node4.unix.tensor.ru"/>
>>       <node id="3" uname="dev-cluster2-node2.unix.tensor.ru"/>
>>     </nodes>
>>  .....
>>
>>  I think something, somewhere is not initialized :)
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (   cluster.c:338  
>>>>  )  notice: get_node_name:     Defaulting to uname -n for the local 
>>>> corosync node name
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (     attrd.c:651  
>>>>  )   debug: attrd_cib_callback:        Update 4 for probe_complete=true 
>>>> passed
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [QB    ] HUP 
>>>> conn (9616-9989-27)
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [QB    ] 
>>>> qb_ipcs_disconnect(9616-9989-27) state:2
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [QB    ] 
>>>> epoll_ctl(del): Bad file descriptor (9)
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [MAIN  ] 
>>>> cs_ipcs_connection_closed()
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [CMAP  ] 
>>>> exit_fn for conn=0x7fa96bcb31b0
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [MAIN  ] 
>>>> cs_ipcs_connection_destroyed()
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [QB    ] 
>>>> Free'ing ringbuffer: /dev/shm/qb-cmap-response-9616-9989-27-header
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [QB    ] 
>>>> Free'ing ringbuffer: /dev/shm/qb-cmap-event-9616-9989-27-header
>>>>   Aug 30 12:31:11 [9615] dev-cluster2-node4 corosync debug   [QB    ] 
>>>> Free'ing ringbuffer: /dev/shm/qb-cmap-request-9616-9989-27-header
>>>>   Aug 30 12:31:11 [9989] dev-cluster2-node4      attrd: (  corosync.c:423  
>>>>  )   trace: check_message_sanity:      Verfied message 1: 
>>>> (dest=<all>:attrd, from=dev-cluster2-node4:attrd.9989, compressed=0, 
>>>> size=181, total=773)
>>>>   Aug 30 12:31:42 [9984] dev-cluster2-node4 pacemakerd: (  mainloop.c:270  
>>>>  )    info: crm_signal_dispatch:       Invoking handler for signal 10: 
>>>> User defined signal 1
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: (       ipc.c:307  
>>>>  )    info: crm_client_new:    Connecting 0x16c98e0 for uid=0 gid=0 
>>>> pid=10007 id=f2f15044-8f76-4ea7-a714-984660619ae7
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: ( ipc_setup.c:476  
>>>>  )   debug: handle_new_connection:     IPC credentials authenticated 
>>>> (9986-10007-13)
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: (   ipc_shm.c:294  
>>>>  )   debug: qb_ipcs_shm_connect:       connecting to client [10007]
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: (ringbuffer.c:227  
>>>>  )   debug: qb_rb_open_2:      shm size:524288; real_size:524288; 
>>>> rb->word_size:131072
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: (ringbuffer.c:227  
>>>>  )   debug: qb_rb_open_2:      shm size:524288; real_size:524288; 
>>>> rb->word_size:131072
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: (ringbuffer.c:227  
>>>>  )   debug: qb_rb_open_2:      shm size:524288; real_size:524288; 
>>>> rb->word_size:131072
>>>>   Aug 30 12:31:59 [9986] dev-cluster2-node4        cib: (        io.c:579  
>>>>  )   debug: activateCibXml:    Triggering CIB write for cib_erase op
>>>>   Aug 30 12:31:59 [9991] dev-cluster2-node4       crmd: (te_callbacks:122  
>>>>  )   debug: te_update_diff:    Processing diff (cib_erase): 0.9.3 -> 
>>>> 0.11.1 (S_IDLE)
>>>>   Aug 30 12:31:59 [9991] dev-cluster2-node4       crmd: (  te_utils.c:423  
>>>>  )    info: abort_transition_graph:    te_update_diff:126 - Triggered 
>>>> transition abort (complete=1, node=, tag=diff, id=(null), magic=NA, 
>>>> cib=0.11.1) : Non-status change
>>>>>>>>>      # corosync-cmapctl |grep nodelist
>>>>>>>>>      nodelist.local_node_pos (u32) = 2
>>>>>>>>>      nodelist.node.0.name (str) = dev-cluster2-node2
>>>>>>>>>      nodelist.node.0.ring0_addr (str) = 10.76.157.17
>>>>>>>>>      nodelist.node.1.name (str) = dev-cluster2-node3
>>>>>>>>>      nodelist.node.1.ring0_addr (str) = 10.76.157.18
>>>>>>>>>      nodelist.node.2.name (str) = dev-cluster2-node4
>>>>>>>>>      nodelist.node.2.ring0_addr (str) = 10.76.157.19
>>>>>>>>>
>>>>>>>>>      # corosync-quorumtool -s
>>>>>>>>>      Quorum information
>>>>>>>>>      ------------------
>>>>>>>>>      Date:             Wed Aug 28 11:29:49 2013
>>>>>>>>>      Quorum provider:  corosync_votequorum
>>>>>>>>>      Nodes:            1
>>>>>>>>>      Node ID:          172793107
>>>>>>>>>      Ring ID:          52
>>>>>>>>>      Quorate:          No
>>>>>>>>>
>>>>>>>>>      Votequorum information
>>>>>>>>>      ----------------------
>>>>>>>>>      Expected votes:   3
>>>>>>>>>      Highest expected: 3
>>>>>>>>>      Total votes:      1
>>>>>>>>>      Quorum:           2 Activity blocked
>>>>>>>>>      Flags:
>>>>>>>>>
>>>>>>>>>      Membership information
>>>>>>>>>      ----------------------
>>>>>>>>>         Nodeid      Votes Name
>>>>>>>>>      172793107          1 dev-cluster2-node4 (local)
>>>>>>>>>
>>>>>>>>>      # cibadmin -Q
>>>>>>>>>      <cib epoch="25" num_updates="3" admin_epoch="0" 
>>>>>>>>> validate-with="pacemaker-1.2" crm_feature_set="3.0.7" 
>>>>>>>>> cib-last-written="Wed Aug 28 11:24:06 2013" 
>>>>>>>>> update-origin="dev-cluster2-node4" update-client="crmd" 
>>>>>>>>> have-quorum="0" dc-uuid="172793107">
>>>>>>>>>       <configuration>
>>>>>>>>>         <crm_config>
>>>>>>>>>           <cluster_property_set id="cib-bootstrap-options">
>>>>>>>>>             <nvpair id="cib-bootstrap-options-dc-version" 
>>>>>>>>> name="dc-version" value="1.1.11-1.el6-4f672bc"/>
>>>>>>>>>             <nvpair id="cib-bootstrap-options-cluster-infrastructure" 
>>>>>>>>> name="cluster-infrastructure" value="corosync"/>
>>>>>>>>>           </cluster_property_set>
>>>>>>>>>         </crm_config>
>>>>>>>>>         <nodes>
>>>>>>>>>           <node id="172793107" uname="dev-cluster2-node4"/>
>>>>>>>>>         </nodes>
>>>>>>>>>         <resources/>
>>>>>>>>>         <constraints/>
>>>>>>>>>       </configuration>
>>>>>>>>>       <status>
>>>>>>>>>         <node_state id="172793107" uname="dev-cluster2-node4" 
>>>>>>>>> in_ccm="true" crmd="online" crm-debug-origin="do_state_transition" 
>>>>>>>>> join="member" expected="member">
>>>>>>>>>           <lrm id="172793107">
>>>>>>>>>             <lrm_resources/>
>>>>>>>>>           </lrm>
>>>>>>>>>           <transient_attributes id="172793107">
>>>>>>>>>             <instance_attributes id="status-172793107">
>>>>>>>>>               <nvpair id="status-172793107-probe_complete" 
>>>>>>>>> name="probe_complete" value="true"/>
>>>>>>>>>             </instance_attributes>
>>>>>>>>>           </transient_attributes>
>>>>>>>>>         </node_state>
>>>>>>>>>       </status>
>>>>>>>>>      </cib>
>>>>>>>>>>>       I figured out a way get around this, but it would be easier 
>>>>>>>>>>> to do if the CIB has worked as a with CMAN.
>>>>>>>>>>>       I just do not start the main resource if the attribute is not 
>>>>>>>>>>> defined or it is not true.
>>>>>>>>>>>       This slightly changes the logic of the cluster.
>>>>>>>>>>>       But I'm not sure what the correct behavior.
>>>>>>>>>>>
>>>>>>>>>>>       libqb 0.14.4
>>>>>>>>>>>       corosync 2.3.1
>>>>>>>>>>>       pacemaker 1.1.11
>>>>>>>>>>>
>>>>>>>>>>>       All build from source in previews week.
>>>>>>>>>>>>>        Now in corosync.conf:
>>>>>>>>>>>>>
>>>>>>>>>>>>>        totem {
>>>>>>>>>>>>>               version: 2
>>>>>>>>>>>>>               crypto_cipher: none
>>>>>>>>>>>>>               crypto_hash: none
>>>>>>>>>>>>>               interface {
>>>>>>>>>>>>>                       ringnumber: 0
>>>>>>>>>>>>>        bindnetaddr: 10.76.157.18
>>>>>>>>>>>>>        mcastaddr: 239.94.1.56
>>>>>>>>>>>>>                       mcastport: 5405
>>>>>>>>>>>>>                       ttl: 1
>>>>>>>>>>>>>               }
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>        logging {
>>>>>>>>>>>>>               fileline: off
>>>>>>>>>>>>>               to_stderr: no
>>>>>>>>>>>>>               to_logfile: yes
>>>>>>>>>>>>>               logfile: /var/log/cluster/corosync.log
>>>>>>>>>>>>>               to_syslog: yes
>>>>>>>>>>>>>               debug: on
>>>>>>>>>>>>>               timestamp: on
>>>>>>>>>>>>>               logger_subsys {
>>>>>>>>>>>>>                       subsys: QUORUM
>>>>>>>>>>>>>                       debug: on
>>>>>>>>>>>>>               }
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>        quorum {
>>>>>>>>>>>>>               provider: corosync_votequorum
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>        nodelist {
>>>>>>>>>>>>>        node {
>>>>>>>>>>>>>        ring0_addr: 10.76.157.17
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>        node {
>>>>>>>>>>>>>        ring0_addr: 10.76.157.18
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>        node {
>>>>>>>>>>>>>        ring0_addr: 10.76.157.19
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>        }
>>>>>>>>>>>>>
>>>>>>>>>>>>>        _______________________________________________
>>>>>>>>>>>>>        Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>>>>>>>        http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>>>>>
>>>>>>>>>>>>>        Project Home: http://www.clusterlabs.org
>>>>>>>>>>>>>        Getting started: 
>>>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>>>>        Bugs: http://bugs.clusterlabs.org
>>>>>>>>>>>>       ,
>>>>>>>>>>>>       _______________________________________________
>>>>>>>>>>>>       Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>>>>>>       http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>>>>
>>>>>>>>>>>>       Project Home: http://www.clusterlabs.org
>>>>>>>>>>>>       Getting started: 
>>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>>>       Bugs: http://bugs.clusterlabs.org
>>>>>>>>>>>       _______________________________________________
>>>>>>>>>>>       Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>>>>>       http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>>>
>>>>>>>>>>>       Project Home: http://www.clusterlabs.org
>>>>>>>>>>>       Getting started: 
>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>>       Bugs: http://bugs.clusterlabs.org
>>>>>>>>>>      ,
>>>>>>>>>>      _______________________________________________
>>>>>>>>>>      Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>>>>      http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>>
>>>>>>>>>>      Project Home: http://www.clusterlabs.org
>>>>>>>>>>      Getting started: 
>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>      Bugs: http://bugs.clusterlabs.org
>>>>>>>>>      _______________________________________________
>>>>>>>>>      Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>>>      http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>
>>>>>>>>>      Project Home: http://www.clusterlabs.org
>>>>>>>>>      Getting started: 
>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>      Bugs: http://bugs.clusterlabs.org
>>>>>>>>     ,
>>>>>>>>     _______________________________________________
>>>>>>>>     Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>>     http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>
>>>>>>>>     Project Home: http://www.clusterlabs.org
>>>>>>>>     Getting started: 
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>     Bugs: http://bugs.clusterlabs.org
>>>>>>>    _______________________________________________
>>>>>>>    Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>>    http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>
>>>>>>>    Project Home: http://www.clusterlabs.org
>>>>>>>    Getting started: 
>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>    Bugs: http://bugs.clusterlabs.org
>>>>>>    _______________________________________________
>>>>>>    Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>>    http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>
>>>>>>    Project Home: http://www.clusterlabs.org
>>>>>>    Getting started: 
>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>    Bugs: http://bugs.clusterlabs.org
>>>>>   ,
>>>>>   _______________________________________________
>>>>>   Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>   http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>
>>>>>   Project Home: http://www.clusterlabs.org
>>>>>   Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>   Bugs: http://bugs.clusterlabs.org
>>>>   _______________________________________________
>>>>   Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>   http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>
>>>>   Project Home: http://www.clusterlabs.org
>>>>   Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>   Bugs: http://bugs.clusterlabs.org
>>>  ,
>>>  _______________________________________________
>>>  Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>  http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>>  Project Home: http://www.clusterlabs.org
>>>  Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>  Bugs: http://bugs.clusterlabs.org
>>  _______________________________________________
>>  Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>  http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>>  Project Home: http://www.clusterlabs.org
>>  Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>  Bugs: http://bugs.clusterlabs.org
>
> ,
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to