On 10/04/2013, at 3:20 PM, Oriol Mula-Valls <oriol.mula-va...@ic3.cat> wrote:

> On 10/04/13 02:10, Andrew Beekhof wrote:
>> 
>> On 09/04/2013, at 7:31 PM, Oriol Mula-Valls<oriol.mula-va...@ic3.cat>  wrote:
>> 
>>> Thanks Andrew I've managed to set up the system and currently I have it 
>>> working but still on testing.
>>> 
>>> I have configure external/ipmi as fencing device and then I force a reboot 
>>> doing a echo b>  /proc/sysrq-trigger. The fencing is working properly as 
>>> the node is shut off and the VM migrated. However, as soon as I turn on the 
>>> fenced now and the OS has started the surviving is shut down. Is it normal 
>>> or am I doing something wrong?
>> 
>> Can you clarify "turn on the fenced"?
>> 
> 
> To restart the fenced node I do either a power on with ipmitool or I power it 
> on using the iRMC web interface.

Oh, "fenced now" was meant to be "fenced node".  That makes more sense now :)

To answer your question, I would not expect the surviving node to be fenced 
when the previous node returns.
The network between the two is still functional?

> 
>>> 
>>> On the other hand I've seen that in case I completely lose power fencing 
>>> obviously fails. Would SBD stonith solve this issue?
>>> 
>>> Kind regards,
>>> Oriol
>>> 
>>> On 08/04/13 04:11, Andrew Beekhof wrote:
>>>> 
>>>> On 03/04/2013, at 9:15 PM, Oriol Mula-Valls<oriol.mula-va...@ic3.cat>   
>>>> wrote:
>>>> 
>>>>> Hi,
>>>>> 
>>>>> I've started with Linux HA about one year ago. Currently I'm facing a new 
>>>>> project in which I have to set up two nodes with high available virtual 
>>>>> machines. I have used as a starting point the Digimer's tutorial 
>>>>> (https://alteeve.ca/w/2-Node_Red_Hat_KVM_Cluster_Tutorial).
>>>>> 
>>>>> To deploy this new infrastructure I have two Fujitsu Primergy Rx100S7. 
>>>>> Both machines have 8GB of RAM and 2x500GB HD. I started creating a 
>>>>> software RAID1 with the internal drives and installing Debian 7.0 
>>>>> (Wheezy). Apart from the O.S. partition I have created 3 more partitions, 
>>>>> one for the shared storage between both machines with OCFS2 and the two 
>>>>> other will be used as PVs to create LVs to support the VMs (one for the 
>>>>> VMs that will be primary on node1 an the other for primary machines on 
>>>>> node2). These 3 partitions are replicated using DRBD.
>>>>> 
>>>>> The shared storage folder contains:
>>>>> * ISO images needed when provisioning VMs
>>>>> * scripts used to call virt-install which handles the creation of our VMs.
>>>>> * XML definition files which define the emulated hardware backing the VMs
>>>>> * old copies of the XML definition files.
>>>>> 
>>>>> I have more or less done the configuration for the OCFS2 fs and I was 
>>>>> about to start the configuration of cLVM for one of the VGs but I have 
>>>>> some doubts. I have one dlm for the OCFS2 filesystem, should I create 
>>>>> another for cLVM RA?
>>>> 
>>>> No, there should only ever be one dlm resource (cloned like you have it)
>>>> 
>>>>> 
>>>>> This is the current configuration:
>>>>> node node1
>>>>> node node2
>>>>> primitive p_dlm_controld ocf:pacemaker:controld \
>>>>>   op start interval="0" timeout="90" \
>>>>>   op stop interval="0" timeout="100" \
>>>>>   op monitor interval="10"
>>>>> primitive p_drbd_shared ocf:linbit:drbd \
>>>>>   params drbd_resource="shared" \
>>>>>   op monitor interval="10" role="Master" timeout="20" \
>>>>>   op monitor interval="20" role="Slave" timeout="20" \
>>>>>   op start interval="0" timeout="240s" \
>>>>>   op stop interval="0" timeout="120s"
>>>>> primitive p_drbd_vm_1 ocf:linbit:drbd \
>>>>>   params drbd_resource="vm_1" \
>>>>>   op monitor interval="10" role="Master" timeout="20" \
>>>>>   op monitor interval="20" role="Slave" timeout="20" \
>>>>>   op start interval="0" timeout="240s" \
>>>>>   op stop interval="0" timeout="120s"
>>>>> primitive p_fs_shared ocf:heartbeat:Filesystem \
>>>>>   params device="/dev/drbd/by-res/shared" directory="/shared" 
>>>>> fstype="ocfs2" \
>>>>>   meta target-role="Started" \
>>>>>   op monitor interval="10"
>>>>> primitive p_ipmi_node1 stonith:external/ipmi \
>>>>>   params hostname="node1" userid="admin" passwd="xxx" ipaddr="10.0.0.2" 
>>>>> interface="lanplus"
>>>>> primitive p_ipmi_node2 stonith:external/ipmi \
>>>>>   params hostname="node2" userid="admin" passwd="xxx" ipaddr="10.0.0.3" 
>>>>> interface="lanplus"
>>>>> primitive p_libvirtd lsb:libvirt-bin \
>>>>>   op monitor interval="120s" \
>>>>>   op start interval="0" \
>>>>>   op stop interval="0"
>>>>> primitive p_o2cb ocf:pacemaker:o2cb \
>>>>>   op start interval="0" timeout="90" \
>>>>>   op stop interval="0" timeout="100" \
>>>>>   op monitor interval="10" \
>>>>>   meta target-role="Started"
>>>>> group g_shared p_dlm_controld p_o2cb p_fs_shared
>>>>> ms ms_drbd_shared p_drbd_shared \
>>>>>   meta master-max="2" clone-max="2" notify="true"
>>>>> ms ms_drbd_vm_1 p_drbd_vm_1 \
>>>>>   meta master-max="2" clone-max="2" notify="true"
>>>>> clone cl_libvirtd p_libvirtd \
>>>>>   meta globally-unique="false" interlave="true"
>>>>> clone cl_shared g_shared \
>>>>>   meta interleave="true"
>>>>> location l_ipmi_node1 p_ipmi_node1 -inf: node1
>>>>> location l_ipmi_node2 p_ipmi_node2 -inf: node2
>>>>> order o_drbd_before_shared inf: ms_drbd_shared:promote cl_shared:start
>>>>> 
>>>>> Packages' versions:
>>>>> clvm                               2.02.95-7
>>>>> corosync                           1.4.2-3
>>>>> dlm-pcmk                           3.0.12-3.2+deb7u2
>>>>> drbd8-utils                        2:8.3.13-2
>>>>> libdlm3                            3.0.12-3.2+deb7u2
>>>>> libdlmcontrol3                     3.0.12-3.2+deb7u2
>>>>> ocfs2-tools                        1.6.4-1+deb7u1
>>>>> ocfs2-tools-pacemaker              1.6.4-1+deb7u1
>>>>> openais                            1.1.4-4.1
>>>>> pacemaker                          1.1.7-1
>>>>> 
>>>>> As this is my first serious set up suggestions are more than welcome.
>>>>> 
>>>>> Thanks for your help.
>>>>> 
>>>>> Oriol
>>>>> --
>>>>> Oriol Mula Valls
>>>>> Institut Català de Ciències del Clima (IC3)
>>>>> Doctor Trueta 203 - 08005 Barcelona
>>>>> Tel:+34 93 567 99 77
>>>>> 
>>>>> _______________________________________________
>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>> 
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>> Bugs: http://bugs.clusterlabs.org
>>>> 
>>>> 
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>> 
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>> 
>>> 
>>> 
>>> --
>>> Oriol Mula Valls
>>> Institut Català de Ciències del Clima (IC3)
>>> Doctor Trueta 203 - 08005 Barcelona
>>> Tel:+34 93 567 99 77
>>> 
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>> 
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>> 
>> 
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> 
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>> 
> 
> 
> -- 
> Oriol Mula Valls
> Institut Català de Ciències del Clima (IC3)
> Doctor Trueta 203 - 08005 Barcelona
> Tel:+34 93 567 99 77
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org


_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to