Yes! I have tried it and succeeded.
Clustered LVM can take snapshot only on exclusively lock mode. 2016-12-06 16:23 GMT+08:00 Ulrich Windl <ulrich.wi...@rz.uni-regensburg.de>: > >>> su liu <liusu8...@gmail.com> schrieb am 06.12.2016 um 02:16 in > Nachricht > <CAN2gjWR_nSOGeZoc7RK_6WicA2L=KM_1=ar9k5--qi7q3vf...@mail.gmail.com>: > > *Hi all,* > > > > > > *I am new to pacemaker and I have some questions about the clvmd + > > pacemaker + corosync. I wish you could explain it for me if you are free. > > thank you very much!* > > *I have 2 nodes and the pacemaker's status is as follows:* > > > > [root@controller ~]# pcs status --full > > Cluster name: mycluster > > Last updated: Mon Dec 5 18:15:12 2016 Last change: Fri Dec 2 > > 15:01:03 2016 by root via cibadmin on compute1 > > Stack: corosync > > Current DC: compute1 (2) (version 1.1.13-10.el7_2.4-44eb2dd) - partition > > with quorum > > 2 nodes and 4 resources configured > > > > Online: [ compute1 (2) controller (1) ] > > > > Full list of resources: > > > > Clone Set: dlm-clone [dlm] > > dlm (ocf::pacemaker:controld): Started compute1 > > dlm (ocf::pacemaker:controld): Started controller > > Started: [ compute1 controller ] > > Clone Set: clvmd-clone [clvmd] > > clvmd (ocf::heartbeat:clvm): Started compute1 > > clvmd (ocf::heartbeat:clvm): Started controller > > Started: [ compute1 controller ] > > > > Node Attributes: > > * Node compute1 (2): > > * Node controller (1): > > > > Migration Summary: > > * Node compute1 (2): > > * Node controller (1): > > > > PCSD Status: > > controller: Online > > compute1: Online > > > > Daemon Status: > > corosync: active/disabled > > pacemaker: active/disabled > > pcsd: active/enabled > > > > > > > > *I create a lvm on controller node and it can be seen on the compute1 > > node immediately with 'lvs' command. but the lvm it not activate on > > compute1.* > > *then i want to create a snapshot of the lvm, but failed with the error > > message:* > > > > > > > > *### volume-4fad87bb-3d4c-4a96-bef1-8799980050d1 must be active > exclusively > > to create snapshot ###* > > *Can someone tell me how to snapshot a lvm in the cluster lvm > environment? > > thank you very much。* > > Did you try "vgchange -a e ..."? > > > > > > > Additional information: > > > > [root@controller ~]# vgdisplay > > --- Volume group --- > > VG Name cinder-volumes > > System ID > > Format lvm2 > > Metadata Areas 1 > > Metadata Sequence No 19 > > VG Access read/write > > VG Status resizable > > Clustered yes > > Shared no > > MAX LV 0 > > Cur LV 1 > > Open LV 0 > > Max PV 0 > > Cur PV 1 > > Act PV 1 > > VG Size 1000.00 GiB > > PE Size 4.00 MiB > > Total PE 255999 > > Alloc PE / Size 256 / 1.00 GiB > > Free PE / Size 255743 / 999.00 GiB > > VG UUID aLamHi-mMcI-2NsC-Spjm-QWZr-MzHx-pPYSTt > > > > [root@controller ~]# rpm -qa |grep pacem > > pacemaker-cli-1.1.13-10.el7_2.4.x86_64 > > pacemaker-libs-1.1.13-10.el7_2.4.x86_64 > > pacemaker-1.1.13-10.el7_2.4.x86_64 > > pacemaker-cluster-libs-1.1.13-10.el7_2.4.x86_64 > > > > > > [root@controller ~]# lvs > > LV VG Attr > > LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert > > volume-1b0ea468-37c8-4b47-a6fa-6cce65b068b5 cinder-volumes -wi-a----- > > 1.00g > > > > > > [root@compute1 ~]# lvs > > LV VG Attr > > LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert > > volume-1b0ea468-37c8-4b47-a6fa-6cce65b068b5 cinder-volumes -wi------- > > 1.00g > > > > > > thank you very much! > > > > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > http://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org >
_______________________________________________ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org