Thank you very much. Because I am new to pacemaker, and I have checked the docs that additional devices are needed when configing stonith, but now I does not have it in my environment.
I will see how to config it afterward. Now I want to know how the cluster LVM works. Thank you for your patience explanation. The scene is: controller node + compute1 node I mount a SAN to both controller and compute1 node. Then I run a pacemaker + corosync + clvmd cluster: [root@controller ~]# pcs status --full Cluster name: mycluster Last updated: Tue Dec 6 14:09:59 2016 Last change: Mon Dec 5 21:26:02 2016 by root via cibadmin on controller Stack: corosync Current DC: compute1 (2) (version 1.1.13-10.el7_2.4-44eb2dd) - partition with quorum 2 nodes and 4 resources configured Online: [ compute1 (2) controller (1) ] Full list of resources: Clone Set: dlm-clone [dlm] dlm (ocf::pacemaker:controld): Started compute1 dlm (ocf::pacemaker:controld): Started controller Started: [ compute1 controller ] Clone Set: clvmd-clone [clvmd] clvmd (ocf::heartbeat:clvm): Started compute1 clvmd (ocf::heartbeat:clvm): Started controller Started: [ compute1 controller ] Node Attributes: * Node compute1 (2): * Node controller (1): Migration Summary: * Node compute1 (2): * Node controller (1): PCSD Status: controller: Online compute1: Online Daemon Status: corosync: active/disabled pacemaker: active/disabled pcsd: active/enabled step 2: I create a cluster VG:cinder-volumes: [root@controller ~]# vgdisplay --- Volume group --- VG Name cinder-volumes System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 44 VG Access read/write VG Status resizable Clustered yes Shared no MAX LV 0 Cur LV 0 Open LV 0 Max PV 0 Cur PV 1 Act PV 1 VG Size 1000.00 GiB PE Size 4.00 MiB Total PE 255999 Alloc PE / Size 0 / 0 Free PE / Size 255999 / 1000.00 GiB VG UUID aLamHi-mMcI-2NsC-Spjm-QWZr-MzHx-pPYSTt [root@controller ~]# Step 3 : I create a LV and I want it can be seen and accessed on the compute1 node but it is failed: [root@controller ~]# lvcreate --name test001 --size 1024m cinder-volumes Logical volume "test001" created. [root@controller ~]# [root@controller ~]# [root@controller ~]# lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert test001 cinder-volumes -wi-a----- 1.00g [root@controller ~]# [root@controller ~]# [root@controller ~]# ll /dev/cinder-volumes/test001 lrwxrwxrwx 1 root root 7 Dec 6 14:13 /dev/cinder-volumes/test001 -> ../dm-0 I can access it on the contrller node, but on the comput1 node, I can see it with lvs command .but cant access it with ls command, because it is not exists on the /dev/cinder-volumes directory: [root@compute1 ~]# lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert test001 cinder-volumes -wi------- 1.00g [root@compute1 ~]# [root@compute1 ~]# [root@compute1 ~]# ll /dev/cinder-volumes ls: cannot access /dev/cinder-volumes: No such file or directory [root@compute1 ~]# [root@compute1 ~]# [root@compute1 ~]# lvscan inactive '/dev/cinder-volumes/test001' [1.00 GiB] inherit [root@compute1 ~]# Is something error with my configuration besides stonith? Could you help me? thank you very much. 2016-12-06 11:37 GMT+08:00 Digimer <li...@alteeve.ca>: > On 05/12/16 10:32 PM, su liu wrote: > > Digimer, thank you very much! > > > > I do not need to have the data accessible on both nodes at once. I want > > to use the clvm+pacemaker+corosync in OpenStack Cinder. > > I'm not sure what "cinder" is, so I don't know what it needs to work. > > > then only a VM need access the LV at once. But the Cinder service which > > runs on the controller node is responsible for snapshotting the LVs > > which are attaching on the VMs runs on other Compute nodes(such as > > compute1 node). > > If you don't need to access an LV on more than one node at a time, then > don't add clustered LVM and keep things simple. If you are using DRBD, > keep the backup secondary. If you are using LUNs, only connect the LUN > to the host that needs it at a given time. > > In HA, you always want to keep things as simple as possible. > > > Need I active the LVs in /exclusively mode all the time? to supoort > > snapping it while attaching on the VM./ > > If you use clustered LVM, yes, but then you can't access the LV on any > other nodes... If you don't need clustered LVM, then no, you continue to > use it as simple LVM. > > Note; Snapshoting VMs is NOT SAFE unless you have a way to be certain > that the guest VM has flushed it's caches and is made crash safe before > the snapshot is made. Otherwise, your snapshot might be corrupted. > > > /The following is the result when execute lvscan command on compute1 > node:/ > > / > > / > > / > > [root@compute1 ~]# lvs > > LV VG Attr > > LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert > > volume-1b0ea468-37c8-4b47-a6fa-6cce65b068b5 cinder-volumes -wi------- > > 1.00g > > > > > > > > and on the controller node: > > > > [root@controller ~]# lvscan ACTIVE > > '/dev/cinder-volumes/volume-1b0ea468-37c8-4b47-a6fa-6cce65b068b5' [1.00 > > GiB] inherit > > > > > > > > thank you very much! > > Did you setup stonith? If not, things will go bad. Not "if", only > "when". Even in a test environment, you _must_ setup stonith. > > -- > Digimer > Papers and Projects: https://alteeve.ca/w/ > What if the cure for cancer is trapped in the mind of a person without > access to education? > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > http://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org >
_______________________________________________ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org