On Sat, 2017-09-02 at 01:21 +0200, Oscar Segarra wrote:
> Hi, 
> 
> I have updated the known_hosts:
> 
> Now, I get the following error:
> 
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_perform_op: +
>  /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resou
> rce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']:
>  @operation_key=vm-vdicdb01_migrate_to_0, @operation=migrate_to,
> @crm-debug-origin=cib_action_update, @transition-key=6:27:0:a7fef266-
> 46c3-429e-ab00-c1a0aab24da5, @transition-magic=-
> 1:193;6:27:0:a7fef266-46c3-429e-ab00-c1a0aab24da5, @call-id=-1, @rc-
> code=193, @op-status=-1, @last-run=1504307021, @last-rc-c
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_process_request:    Completed cib_modify operation for section
> status: OK (rc=0, origin=vdicnode01/crmd/77, version=0.169.1)
> VirtualDomain(vm-vdicdb01)[13085]:      2017/09/02_01:03:41 INFO:
> vdicdb01: Starting live migration to vdicnode02 (using: virsh --
> connect=qemu:///system --quiet migrate --live  vdicdb01
> qemu+ssh://vdicnode02/system ).
> VirtualDomain(vm-vdicdb01)[13085]:      2017/09/02_01:03:41 ERROR:
> vdicdb01: live migration to vdicnode02 failed: 1
>  ]p 02 01:03:41 [1537] vdicnode01       lrmd:   notice:
> operation_finished:     vm-vdicdb01_migrate_to_0:13085:stderr [
> error: Cannot recv data: Permission denied, please try again.
>  ]p 02 01:03:41 [1537] vdicnode01       lrmd:   notice:
> operation_finished:     vm-vdicdb01_migrate_to_0:13085:stderr [
> Permission denied, please try again.
> Sep 02 01:03:41 [1537] vdicnode01       lrmd:   notice:
> operation_finished:     vm-vdicdb01_migrate_to_0:13085:stderr [
> Permission denied (publickey,gssapi-keyex,gssapi-with-mic,password).: 
> Connection reset by peer ]
> Sep 02 01:03:41 [1537] vdicnode01       lrmd:   notice:
> operation_finished:     vm-vdicdb01_migrate_to_0:13085:stderr [ ocf-
> exit-reason:vdicdb01: live migration to vdicnode02 failed: 1 ]
> Sep 02 01:03:41 [1537] vdicnode01       lrmd:     info: log_finished:
>   finished - rsc:vm-vdicdb01 action:migrate_to call_id:16 pid:13085
> exit-code:1 exec-time:119ms queue-time:0ms
> Sep 02 01:03:41 [1540] vdicnode01       crmd:   notice:
> process_lrm_event:      Result of migrate_to operation for vm-
> vdicdb01 on vdicnode01: 1 (unknown error) | call=16 key=vm-
> vdicdb01_migrate_to_0 confirmed=true cib-update=78
> Sep 02 01:03:41 [1540] vdicnode01       crmd:   notice:
> process_lrm_event:      vdicnode01-vm-vdicdb01_migrate_to_0:16 [
> error: Cannot recv data: Permission denied, please try
> again.\r\nPermission denied, please try again.\r\nPermission denied
> (publickey,gssapi-keyex,gssapi-with-mic,password).: Connection reset
> by peer\nocf-exit-reason:vdicdb01: live migration to vdicnode02
> failed: 1\n ]
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_process_request:    Forwarding cib_modify operation for section
> status to all (origin=local/crmd/78)
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_perform_op: Diff: --- 0.169.1 2
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_perform_op: Diff: +++ 0.169.2 (null)
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_perform_op: +  /cib:  @num_updates=2
> Sep 02 01:03:41 [1535] vdicnode01        cib:     info:
> cib_perform_op: +
>  /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resou
> rce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']:  @crm-
> debug-origin=do_update_resource, @transition-
> magic=0:1;6:27:0:a7fef266-46c3-429e-ab00-c1a0aab24da5, @call-id=16,
> @rc-code=1, @op-status=0, @exec-time=119, @exit-reason=vdicdb01: live
> migration to vdicnode02 failed: 1
> Sep 02 01:03:4
> 
> as root <-- system prompts the password
> [root@vdicnode01 .ssh]# virsh --connect=qemu:///system --quiet
> migrate --live  vdicdb01 qemu+ssh://vdicnode02/system
> root@vdicnode02's password:
> 
> as oneadmin (the user that executes the qemu-kvm) <-- does not prompt
> the password
> virsh --connect=qemu:///system --quiet migrate --live  vdicdb01
> qemu+ssh://vdicnode02/system
> 
> Must I configure passwordless connection with root in order to make
> live migration work?
> 
> Or maybe is there any way to instruct pacemaker to use my oneadmin
> user for migrations inestad of root?

Pacemaker calls the VirtualDomain resource agent as root, but it's up
to the agent what to do from there. I don't see any user options in
VirtualDomain or virsh, so I don't think there is currently.

I see two options: configure passwordless ssh for root, or copy the
VirtualDomain resource and modify it to use "sudo -u oneadmin" when it
calls virsh.

We've discussed adding the capability to tell pacemaker to execute a
resource agent as a particular user. We've already put the plumbing in
for it, so that lrmd can execute alert agents as the hacluster user.
All that would be needed would be a new resource meta-attribute and the
IPC API to use it. It's low priority due to a large backlog at the
moment, but we'd be happy to take a pull request for it. The resource
agent would obviously have to be able to work as that user.

> 
> Thanks a lot:
> 
> 
> 2017-09-01 23:14 GMT+02:00 Ken Gaillot <kgail...@redhat.com>:
> > On Fri, 2017-09-01 at 00:26 +0200, Oscar Segarra wrote:
> > > Hi,
> > >
> > >
> > > Yes, it is....
> > >
> > >
> > > The qemu-kvm process is executed by the oneadmin user.
> > >
> > >
> > > When I cluster tries the live migration, what users do play?
> > >
> > >
> > > Oneadmin
> > > Root
> > > Hacluster
> > >
> > >
> > > I have just configured pasworless ssh connection with oneadmin.
> > >
> > >
> > > Do I need to configure any other passwordless ssh connection with
> > any
> > > other user?
> > >
> > >
> > > What user executes the virsh migrate - - live?
> > 
> > The cluster executes resource actions as root.
> > 
> > > Is there any way to check ssk keys?
> > 
> > I'd just login once to the host as root from the cluster nodes, to
> > make
> > it sure it works, and accept the host when asked.
> > 
> > >
> > > Sorry for all theese questions.
> > >
> > >
> > > Thanks a lot
> > >
> > >
> > >
> > >
> > >
> > >
> > > El 1 sept. 2017 0:12, "Ken Gaillot" <kgail...@redhat.com>
> > escribió:
> > >         On Thu, 2017-08-31 at 23:45 +0200, Oscar Segarra wrote:
> > >         > Hi Ken,
> > >         >
> > >         >
> > >         > Thanks a lot for you quick answer:
> > >         >
> > >         >
> > >         > Regarding to selinux, it is disabled. The FW is
> > disabled as
> > >         well.
> > >         >
> > >         >
> > >         > [root@vdicnode01 ~]# sestatus
> > >         > SELinux status:                 disabled
> > >         >
> > >         >
> > >         > [root@vdicnode01 ~]# service firewalld status
> > >         > Redirecting to /bin/systemctl status  firewalld.service
> > >         > ● firewalld.service - firewalld - dynamic firewall
> > daemon
> > >         >    Loaded: loaded
> > >         (/usr/lib/systemd/system/firewalld.service;
> > >         > disabled; vendor preset: enabled)
> > >         >    Active: inactive (dead)
> > >         >      Docs: man:firewalld(1)
> > >         >
> > >         >
> > >         > On migration, it performs a gracefully shutdown and a
> > start
> > >         on the new
> > >         > node.
> > >         >
> > >         >
> > >         > I attach the logs when trying to migrate from
> > vdicnode02 to
> > >         > vdicnode01:
> > >         >
> > >         >
> > >         > vdicnode02 corosync.log:
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.161.2 2
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.162.0 (null)
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op:
> > >         >
> > >         --
> > /cib/configuration/constraints/rsc_location[@id='location-vm-
> > vdicdb01-vdicnode01--INFINITY']
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @epoch=162, @num_updates=0
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_replace operation
> > for
> > >         section
> > >         > configuration: OK (rc=0, origin=vdicnode01/cibadmin/2,
> > >         > version=0.162.0)
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_file_backup:        Archived previous version
> > >         > as /var/lib/pacemaker/cib/cib-65.raw
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_file_write_with_digest:     Wrote version 0.162.0
> > of the
> > >         CIB to
> > >         > disk (digest: 1f87611b60cd7c48b95b6b788b47f65f)
> > >         > Aug 31 23:38:17 [1521] vdicnode02        cib:     info:
> > >         > cib_file_write_with_digest:     Reading cluster
> > >         configuration
> > >         > file /var/lib/pacemaker/cib/cib.jt2KPw
> > >         > (digest: /var/lib/pacemaker/cib/cib.Kwqfpl)
> > >         > Aug 31 23:38:22 [1521] vdicnode02        cib:     info:
> > >         > cib_process_ping:       Reporting our current digest to
> > >         vdicnode01:
> > >         > dace3a23264934279d439420d5a716cc for 0.162.0
> > (0x7f96bb26c5c0
> > >         0)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.162.0 2
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.0 (null)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @epoch=163
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: ++ /cib/configuration/constraints:
> > >         <rsc_location
> > >         > id="location-vm-vdicdb01-vdicnode02--INFINITY"
> > >         node="vdicnode02"
> > >         > rsc="vm-vdicdb01" score="-INFINITY"/>
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_replace operation
> > for
> > >         section
> > >         > configuration: OK (rc=0, origin=vdicnode01/cibadmin/2,
> > >         > version=0.163.0)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_file_backup:        Archived previous version
> > >         > as /var/lib/pacemaker/cib/cib-66.raw
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_file_write_with_digest:     Wrote version 0.163.0
> > of the
> > >         CIB to
> > >         > disk (digest: 47a548b36746de9275d66cc6aeb0fdc4)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_file_write_with_digest:     Reading cluster
> > >         configuration
> > >         > file /var/lib/pacemaker/cib/cib.rcgXiT
> > >         > (digest: /var/lib/pacemaker/cib/cib.7geMfi)
> > >         > Aug 31 23:38:27 [1523] vdicnode02       lrmd:     info:
> > >         > cancel_recurring_action:        Cancelling ocf
> > operation
> > >         > vm-vdicdb01_monitor_10000
> > >         > Aug 31 23:38:27 [1526] vdicnode02       crmd:     info:
> > >         > do_lrm_rsc_op:  Performing
> > >         > key=6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         > op=vm-vdicdb01_migrate_to_0
> > >         > Aug 31 23:38:27 [1523] vdicnode02       lrmd:     info:
> > >         log_execute:
> > >         >    executing - rsc:vm-vdicdb01 action:migrate_to
> > call_id:9
> > >         > Aug 31 23:38:27 [1526] vdicnode02       crmd:     info:
> > >         > process_lrm_event:      Result of monitor operation for
> > >         vm-vdicdb01 on
> > >         > vdicnode02: Cancelled | call=7 key=vm-
> > vdicdb01_monitor_10000
> > >         > confirmed=true
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.0 2
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.1 (null)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=1
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_migrate_to_0, @operation=migrate_to,
> > @crm-debug-origin=cib_action_update, @transition-
> > key=6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=-
> > 1:193;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=-1, @rc-
> > code=193, @op-status=-1, @last-run=1504215507, @last-rc-cha
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/41,
> > >         version=0.163.1)
> > >         > VirtualDomain(vm-vdicdb01)[5241]:     
> >  2017/08/31_23:38:27
> > >         INFO:
> > >         > vdicdb01: Starting live migration to vdicnode01 (using:
> > >         virsh
> > >         > --connect=qemu:///system --quiet migrate --live 
> > vdicdb01
> > >         qemu
> > >         > +ssh://vdicnode01/system ).
> > >         > VirtualDomain(vm-vdicdb01)[5241]:     
> >  2017/08/31_23:38:27
> > >         ERROR:
> > >         > vdicdb01: live migration to vdicnode01 failed: 1
> > >         > Aug 31 23:38:27 [1523] vdicnode02       lrmd:   notice:
> > >         > operation_finished:     vm-
> > vdicdb01_migrate_to_0:5241:stderr
> > >         [ error:
> > >         > Cannot recv data: Host key verification failed.:
> > Connection
> > >         reset by
> > >         > peer ]
> > >
> > >
> > >         ^^^ There you go. Sounds like the ssh key isn't being
> > >         accepted. No idea
> > >         why though.
> > >
> > >
> > >
> > >         > Aug 31 23:38:27 [1523] vdicnode02       lrmd:   notice:
> > >         > operation_finished:     vm-
> > vdicdb01_migrate_to_0:5241:stderr
> > >         > [ ocf-exit-reason:vdicdb01: live migration to
> > vdicnode01
> > >         failed: 1 ]
> > >         > Aug 31 23:38:27 [1523] vdicnode02       lrmd:     info:
> > >         log_finished:
> > >         > finished - rsc:vm-vdicdb01 action:migrate_to call_id:9
> > >         pid:5241
> > >         > exit-code:1 exec-time:78ms queue-time:0ms
> > >         > Aug 31 23:38:27 [1526] vdicnode02       crmd:   notice:
> > >         > process_lrm_event:      Result of migrate_to operation
> > for
> > >         vm-vdicdb01
> > >         > on vdicnode02: 1 (unknown error) | call=9
> > >         key=vm-vdicdb01_migrate_to_0
> > >         > confirmed=true cib-update=14
> > >         > Aug 31 23:38:27 [1526] vdicnode02       crmd:   notice:
> > >         > process_lrm_event:
> > >         vdicnode02-vm-vdicdb01_migrate_to_0:9 [ error:
> > >         > Cannot recv data: Host key verification failed.:
> > Connection
> > >         reset by
> > >         > peer\nocf-exit-reason:vdicdb01: live migration to
> > vdicnode01
> > >         failed: 1
> > >         > \n ]
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Forwarding cib_modify operation
> > for
> > >         section
> > >         > status to all (origin=local/crmd/14)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.1 2
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.2 (null)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=2
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @crm-debug-origin=do_update_resource, @transition-
> > magic=0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=9,
> > @rc-code=1, @op-status=0, @exec-time=78, @exit-reason=vdicdb01:
> > live migration to vdicnode01 failed: 1
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op:
> > >         >
> > >         ++
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']:  <lrm_rsc_op id="vm-
> > vdicdb01_last_failure_0" operation_key="vm-vdicdb01_migrate_to_0"
> > operation="migrate_to" crm-debug-origin="do_update_resource"
> > crm_feature_set="3.0.10" transition-key="6:6:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" transition-magic="0:1;6:6:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" exit-reason="vdicdb01: live migration to vdicn
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode02/crmd/14,
> > >         version=0.163.2)
> > >         > Aug 31 23:38:27 [1526] vdicnode02       crmd:     info:
> > >         > do_lrm_rsc_op:  Performing
> > >         > key=2:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         op=vm-vdicdb01_stop_0
> > >         > Aug 31 23:38:27 [1523] vdicnode02       lrmd:     info:
> > >         log_execute:
> > >         >    executing - rsc:vm-vdicdb01 action:stop call_id:10
> > >         > VirtualDomain(vm-vdicdb01)[5285]:     
> >  2017/08/31_23:38:27
> > >         INFO:
> > >         > Issuing graceful shutdown request for domain vdicdb01.
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.2 2
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.3 (null)
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=3
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-
> > key=4:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-
> > magic=0:0;4:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=6,
> > @rc-code=0, @last-run=1504215507, @last-rc-change=1504215507,
> > @exec-time=57
> > >         > Aug 31 23:38:27 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/43,
> > >         version=0.163.3)
> > >         > Aug 31 23:38:30 [1523] vdicnode02       lrmd:     info:
> > >         log_finished:
> > >         > finished - rsc:vm-vdicdb01 action:stop call_id:10
> > pid:5285
> > >         exit-code:0
> > >         > exec-time:3159ms queue-time:0ms
> > >         > Aug 31 23:38:30 [1526] vdicnode02       crmd:   notice:
> > >         > process_lrm_event:      Result of stop operation for
> > >         vm-vdicdb01 on
> > >         > vdicnode02: 0 (ok) | call=10 key=vm-vdicdb01_stop_0
> > >         confirmed=true
> > >         > cib-update=15
> > >         > Aug 31 23:38:30 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Forwarding cib_modify operation
> > for
> > >         section
> > >         > status to all (origin=local/crmd/15)
> > >         > Aug 31 23:38:30 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.3 2
> > >         > Aug 31 23:38:30 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.4 (null)
> > >         > Aug 31 23:38:30 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=4
> > >         > Aug 31 23:38:30 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-
> > key=2:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-
> > magic=0:0;2:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=10,
> > @rc-code=0, @exec-time=3159
> > >         > Aug 31 23:38:30 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode02/crmd/15,
> > >         version=0.163.4)
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.4 2
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.5 (null)
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=5
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_start_0, @operation=start, @transition-
> > key=5:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-
> > magic=0:0;5:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=7,
> > @last-run=1504215510, @last-rc-change=1504215510, @exec-time=528
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/44,
> > >         version=0.163.5)
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.5 2
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.6 (null)
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=6
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_perform_op:
> > >         >
> > >         ++
> > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']:  <lrm_rsc_op id="vm-
> > vdicdb01_monitor_10000" operation_key="vm-vdicdb01_monitor_10000"
> > operation="monitor" crm-debug-origin="do_update_resource"
> > crm_feature_set="3.0.10" transition-key="6:7:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" transition-magic="0:0;6:7:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" on_node="vdicnode01" call-id="8" rc-code="0" op-
> > s
> > >         > Aug 31 23:38:31 [1521] vdicnode02        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/45,
> > >         version=0.163.6)
> > >         > Aug 31 23:38:36 [1521] vdicnode02        cib:     info:
> > >         > cib_process_ping:       Reporting our current digest to
> > >         vdicnode01:
> > >         > 9141ea9880f5a44b133003982d863bc8 for 0.163.6
> > (0x7f96bb2625a0
> > >         0)
> > >         >
> > >         >
> > >         >
> > >         >
> > >         >
> > >         >
> > >         > vdicnode01 - corosync.log
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Forwarding cib_replace
> > operation for
> > >         section
> > >         > configuration to all (origin=local/cibadmin/2)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.162.0 2
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.0 (null)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @epoch=163
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: ++ /cib/configuration/constraints:
> > >         <rsc_location
> > >         > id="location-vm-vdicdb01-vdicnode02--INFINITY"
> > >         node="vdicnode02"
> > >         > rsc="vm-vdicdb01" score="-INFINITY"/>
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_replace operation
> > for
> > >         section
> > >         > configuration: OK (rc=0, origin=vdicnode01/cibadmin/2,
> > >         > version=0.163.0)
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > abort_transition_graph: Transition aborted by
> > >         > rsc_location.location-vm-vdicdb01-vdicnode02--INFINITY
> > >         'create':
> > >         > Non-status change | cib=0.163.0
> > source=te_update_diff:436
> > >         > path=/cib/configuration/constraints complete=true
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         > do_state_transition:    State transition S_IDLE ->
> > >         S_POLICY_ENGINE |
> > >         > input=I_PE_CALC cause=C_FSA_INTERNAL
> > >         origin=abort_transition_graph
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_file_backup:        Archived previous version
> > >         > as /var/lib/pacemaker/cib/cib-85.raw
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_file_write_with_digest:     Wrote version 0.163.0
> > of the
> > >         CIB to
> > >         > disk (digest: 47a548b36746de9275d66cc6aeb0fdc4)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_file_write_with_digest:     Reading cluster
> > >         configuration
> > >         > file /var/lib/pacemaker/cib/cib.npBIW2
> > >         > (digest: /var/lib/pacemaker/cib/cib.bDogoB)
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         > determine_online_status:        Node vdicnode02 is
> > online
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         > determine_online_status:        Node vdicnode01 is
> > online
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         native_print:
> > >         > vm-vdicdb01     (ocf::heartbeat:VirtualDomain): Started
> > >         vdicnode02
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         RecurringOp:
> > >         > Start recurring monitor (10s) for vm-vdicdb01 on
> > vdicnode01
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:   notice:
> > >         LogActions:
> > >         > Migrate vm-vdicdb01     (Started vdicnode02 ->
> > vdicnode01)
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:   notice:
> > >         > process_pe_message:     Calculated transition 6, saving
> > >         inputs
> > >         > in /var/lib/pacemaker/pengine/pe-input-96.bz2
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > do_state_transition:    State transition
> > S_POLICY_ENGINE ->
> > >         > S_TRANSITION_ENGINE | input=I_PE_SUCCESS
> > cause=C_IPC_MESSAGE
> > >         > origin=handle_response
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         do_te_invoke:
> > >         > Processing graph 6 (ref=pe_calc-dc-1504215507-24)
> > derived
> > >         > from /var/lib/pacemaker/pengine/pe-input-96.bz2
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         > te_rsc_command: Initiating migrate_to operation
> > >         > vm-vdicdb01_migrate_to_0 on vdicnode02 | action 6
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > create_operation_update:        cib_action_update:
> > Updating
> > >         resource
> > >         > vm-vdicdb01 after migrate_to op pending (interval=0)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Forwarding cib_modify operation
> > for
> > >         section
> > >         > status to all (origin=local/crmd/41)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.0 2
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.1 (null)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=1
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_migrate_to_0, @operation=migrate_to,
> > @crm-debug-origin=cib_action_update, @transition-
> > key=6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=-
> > 1:193;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=-1, @rc-
> > code=193, @op-status=-1, @last-run=1504215507, @last-rc-cha
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/41,
> > >         version=0.163.1)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.1 2
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.2 (null)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=2
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @crm-debug-origin=do_update_resource, @transition-
> > magic=0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=9,
> > @rc-code=1, @op-status=0, @exec-time=78, @exit-reason=vdicdb01:
> > live migration to vdicnode01 failed: 1
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op:
> > >         >
> > >         ++
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']:  <lrm_rsc_op id="vm-
> > vdicdb01_last_failure_0" operation_key="vm-vdicdb01_migrate_to_0"
> > operation="migrate_to" crm-debug-origin="do_update_resource"
> > crm_feature_set="3.0.10" transition-key="6:6:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" transition-magic="0:1;6:6:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" exit-reason="vdicdb01: live migration to vdicn
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode02/crmd/14,
> > >         version=0.163.2)
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:  warning:
> > >         > status_from_rc: Action 6 (vm-vdicdb01_migrate_to_0) on
> > >         vdicnode02
> > >         > failed (target: 0 vs. rc: 1): Error
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         > abort_transition_graph: Transition aborted by operation
> > >         > vm-vdicdb01_migrate_to_0 'modify' on vdicnode02: Event
> > >         failed |
> > >         > magic=0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         cib=0.163.2
> > >         > source=match_graph_event:310 complete=false
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > match_graph_event:      Action vm-vdicdb01_migrate_to_0 
> > (6)
> > >         confirmed
> > >         > on vdicnode02 (rc=1)
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > process_graph_event:    Detected action (6.6)
> > >         > vm-vdicdb01_migrate_to_0.9=unknown error: failed
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:  warning:
> > >         > status_from_rc: Action 6 (vm-vdicdb01_migrate_to_0) on
> > >         vdicnode02
> > >         > failed (target: 0 vs. rc: 1): Error
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > abort_transition_graph: Transition aborted by operation
> > >         > vm-vdicdb01_migrate_to_0 'create' on vdicnode02: Event
> > >         failed |
> > >         > magic=0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         cib=0.163.2
> > >         > source=match_graph_event:310 complete=false
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > match_graph_event:      Action vm-vdicdb01_migrate_to_0 
> > (6)
> > >         confirmed
> > >         > on vdicnode02 (rc=1)
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > process_graph_event:    Detected action (6.6)
> > >         > vm-vdicdb01_migrate_to_0.9=unknown error: failed
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         run_graph:
> > >         >    Transition 6 (Complete=1, Pending=0, Fired=0,
> > Skipped=0,
> > >         > Incomplete=5,
> > >         Source=/var/lib/pacemaker/pengine/pe-input-96.bz2):
> > >         > Complete
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > do_state_transition:    State transition
> > S_TRANSITION_ENGINE
> > >         ->
> > >         > S_POLICY_ENGINE | input=I_PE_CALC cause=C_FSA_INTERNAL
> > >         > origin=notify_crmd
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         > determine_online_status:        Node vdicnode02 is
> > online
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         > determine_online_status:        Node vdicnode01 is
> > online
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:  warning:
> > >         > unpack_rsc_op_failure:  Processing failed op migrate_to
> > for
> > >         > vm-vdicdb01 on vdicnode02: unknown error (1)
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:  warning:
> > >         > unpack_rsc_op_failure:  Processing failed op migrate_to
> > for
> > >         > vm-vdicdb01 on vdicnode02: unknown error (1)
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         native_print:
> > >         > vm-vdicdb01     (ocf::heartbeat:VirtualDomain): FAILED
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         native_print:
> > >         > 1 : vdicnode01
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         native_print:
> > >         > 2 : vdicnode02
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:    error:
> > >         > native_create_actions:  Resource vm-vdicdb01
> > >         (ocf::VirtualDomain) is
> > >         > active on 2 nodes attempting recovery
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:  warning:
> > >         > native_create_actions:  See
> > >         > http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active
> > for
> > >         more
> > >         > information.
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:     info:
> > >         RecurringOp:
> > >         > Start recurring monitor (10s) for vm-vdicdb01 on
> > vdicnode01
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:   notice:
> > >         LogActions:
> > >         > Recover vm-vdicdb01     (Started vdicnode01)
> > >         > Aug 31 23:38:27 [1535] vdicnode01    pengine:    error:
> > >         > process_pe_message:     Calculated transition 7 (with
> > >         errors), saving
> > >         > inputs in /var/lib/pacemaker/pengine/pe-error-8.bz2
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > do_state_transition:    State transition
> > S_POLICY_ENGINE ->
> > >         > S_TRANSITION_ENGINE | input=I_PE_SUCCESS
> > cause=C_IPC_MESSAGE
> > >         > origin=handle_response
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         do_te_invoke:
> > >         > Processing graph 7 (ref=pe_calc-dc-1504215507-26)
> > derived
> > >         > from /var/lib/pacemaker/pengine/pe-error-8.bz2
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         > te_rsc_command: Initiating stop operation vm-
> > vdicdb01_stop_0
> > >         locally
> > >         > on vdicnode01 | action 4
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > do_lrm_rsc_op:  Performing
> > >         > key=4:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         op=vm-vdicdb01_stop_0
> > >         > Aug 31 23:38:27 [1533] vdicnode01       lrmd:     info:
> > >         log_execute:
> > >         >    executing - rsc:vm-vdicdb01 action:stop call_id:6
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         > te_rsc_command: Initiating stop operation vm-
> > vdicdb01_stop_0
> > >         on
> > >         > vdicnode02 | action 2
> > >         > VirtualDomain(vm-vdicdb01)[5268]:     
> >  2017/08/31_23:38:27
> > >         INFO:
> > >         > Domain vdicdb01 already stopped.
> > >         > Aug 31 23:38:27 [1533] vdicnode01       lrmd:     info:
> > >         log_finished:
> > >         > finished - rsc:vm-vdicdb01 action:stop call_id:6
> > pid:5268
> > >         exit-code:0
> > >         > exec-time:57ms queue-time:0ms
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:   notice:
> > >         > process_lrm_event:      Result of stop operation for
> > >         vm-vdicdb01 on
> > >         > vdicnode01: 0 (ok) | call=6 key=vm-vdicdb01_stop_0
> > >         confirmed=true
> > >         > cib-update=43
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Forwarding cib_modify operation
> > for
> > >         section
> > >         > status to all (origin=local/crmd/43)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.2 2
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.3 (null)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=3
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-
> > key=4:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-
> > magic=0:0;4:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=6,
> > @rc-code=0, @last-run=1504215507, @last-rc-change=1504215507,
> > @exec-time=57
> > >         > Aug 31 23:38:27 [1536] vdicnode01       crmd:     info:
> > >         > match_graph_event:      Action vm-vdicdb01_stop_0 (4)
> > >         confirmed on
> > >         > vdicnode01 (rc=0)
> > >         > Aug 31 23:38:27 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/43,
> > >         version=0.163.3)
> > >         > Aug 31 23:38:30 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.3 2
> > >         > Aug 31 23:38:30 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.4 (null)
> > >         > Aug 31 23:38:30 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=4
> > >         > Aug 31 23:38:30 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-
> > key=2:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-
> > magic=0:0;2:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=10,
> > @rc-code=0, @exec-time=3159
> > >         > Aug 31 23:38:30 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode02/crmd/15,
> > >         version=0.163.4)
> > >         > Aug 31 23:38:30 [1536] vdicnode01       crmd:     info:
> > >         > match_graph_event:      Action vm-vdicdb01_stop_0 (2)
> > >         confirmed on
> > >         > vdicnode02 (rc=0)
> > >         > Aug 31 23:38:30 [1536] vdicnode01       crmd:   notice:
> > >         > te_rsc_command: Initiating start operation
> > >         vm-vdicdb01_start_0 locally
> > >         > on vdicnode01 | action 5
> > >         > Aug 31 23:38:30 [1536] vdicnode01       crmd:     info:
> > >         > do_lrm_rsc_op:  Performing
> > >         > key=5:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         op=vm-vdicdb01_start_0
> > >         > Aug 31 23:38:30 [1533] vdicnode01       lrmd:     info:
> > >         log_execute:
> > >         >    executing - rsc:vm-vdicdb01 action:start call_id:7
> > >         > Aug 31 23:38:31 [1533] vdicnode01       lrmd:     info:
> > >         log_finished:
> > >         > finished - rsc:vm-vdicdb01 action:start call_id:7
> > pid:5401
> > >         exit-code:0
> > >         > exec-time:528ms queue-time:0ms
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:     info:
> > >         > action_synced_wait:     Managed VirtualDomain_meta-
> > data_0
> > >         process 5486
> > >         > exited with rc=0
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:   notice:
> > >         > process_lrm_event:      Result of start operation for
> > >         vm-vdicdb01 on
> > >         > vdicnode01: 0 (ok) | call=7 key=vm-vdicdb01_start_0
> > >         confirmed=true
> > >         > cib-update=44
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Forwarding cib_modify operation
> > for
> > >         section
> > >         > status to all (origin=local/crmd/44)
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.4 2
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.5 (null)
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=5
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +
> > >         > 
> > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: 
> > @operation_key=vm-vdicdb01_start_0, @operation=start, @transition-
> > key=5:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-
> > magic=0:0;5:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=7,
> > @last-run=1504215510, @last-rc-change=1504215510, @exec-time=528
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/44,
> > >         version=0.163.5)
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:     info:
> > >         > match_graph_event:      Action vm-vdicdb01_start_0 (5)
> > >         confirmed on
> > >         > vdicnode01 (rc=0)
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:   notice:
> > >         > te_rsc_command: Initiating monitor operation
> > >         vm-vdicdb01_monitor_10000
> > >         > locally on vdicnode01 | action 6
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:     info:
> > >         > do_lrm_rsc_op:  Performing
> > >         > key=6:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a
> > >         > op=vm-vdicdb01_monitor_10000
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:     info:
> > >         > process_lrm_event:      Result of monitor operation for
> > >         vm-vdicdb01 on
> > >         > vdicnode01: 0 (ok) | call=8 key=vm-
> > vdicdb01_monitor_10000
> > >         > confirmed=false cib-update=45
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Forwarding cib_modify operation
> > for
> > >         section
> > >         > status to all (origin=local/crmd/45)
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: --- 0.163.5 2
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: Diff: +++ 0.163.6 (null)
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op: +  /cib:  @num_updates=6
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_perform_op:
> > >         >
> > >         ++
> > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_reso
> > urce[@id='vm-vdicdb01']:  <lrm_rsc_op id="vm-
> > vdicdb01_monitor_10000" operation_key="vm-vdicdb01_monitor_10000"
> > operation="monitor" crm-debug-origin="do_update_resource"
> > crm_feature_set="3.0.10" transition-key="6:7:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" transition-magic="0:0;6:7:0:fe1a9b0a-816c-4b97-
> > 96cb-b90dbf71417a" on_node="vdicnode01" call-id="8" rc-code="0" op-
> > s
> > >         > Aug 31 23:38:31 [1531] vdicnode01        cib:     info:
> > >         > cib_process_request:    Completed cib_modify operation
> > for
> > >         section
> > >         > status: OK (rc=0, origin=vdicnode01/crmd/45,
> > >         version=0.163.6)
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:     info:
> > >         > match_graph_event:      Action vm-
> > vdicdb01_monitor_10000 (6)
> > >         confirmed
> > >         > on vdicnode01 (rc=0)
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:   notice:
> > >         run_graph:
> > >         >    Transition 7 (Complete=5, Pending=0, Fired=0,
> > Skipped=0,
> > >         > Incomplete=0,
> > >         Source=/var/lib/pacemaker/pengine/pe-error-8.bz2):
> > >         > Complete
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:     info:
> > >         do_log: Input
> > >         > I_TE_SUCCESS received in state S_TRANSITION_ENGINE from
> > >         notify_crmd
> > >         > Aug 31 23:38:31 [1536] vdicnode01       crmd:   notice:
> > >         > do_state_transition:    State transition
> > S_TRANSITION_ENGINE
> > >         -> S_IDLE
> > >         > | input=I_TE_SUCCESS cause=C_FSA_INTERNAL
> > origin=notify_crmd
> > >         > Aug 31 23:38:36 [1531] vdicnode01        cib:     info:
> > >         > cib_process_ping:       Reporting our current digest to
> > >         vdicnode01:
> > >         > 9141ea9880f5a44b133003982d863bc8 for 0.163.6
> > (0x7f61cec09270
> > >         0)
> > >         >
> > >         >
> > >         > Thanks a lot
> > >         >
> > >         > 2017-08-31 16:20 GMT+02:00 Ken Gaillot
> > >         <kgail...@redhat.com>:
> > >         >         On Thu, 2017-08-31 at 01:13 +0200, Oscar
> > Segarra
> > >         wrote:
> > >         >         > Hi,
> > >         >         >
> > >         >         >
> > >         >         > In my environment, I have just two hosts,
> > where
> > >         qemu-kvm
> > >         >         process is
> > >         >         > launched by a regular user (oneadmin) - open
> > >         nebula -
> > >         >         >
> > >         >         >
> > >         >         > I have created a VirtualDomain resource that
> > >         starts and
> > >         >         stops the VM
> > >         >         > perfectly. Nevertheless, when I change the
> > >         location weight
> > >         >         in order to
> > >         >         > force the migration, It raises a migration
> > failure
> > >         "error:
> > >         >         1"
> > >         >         >
> > >         >         >
> > >         >         > If I execute the virsh migrate command (that
> > >         appears in
> > >         >         corosync.log)
> > >         >         > from command line, it works perfectly.
> > >         >         >
> > >         >         >
> > >         >         > Anybody has experienced the same issue?
> > >         >         >
> > >         >         >
> > >         >         > Thanks in advance for your help
> > >         >
> > >         >
> > >         >         If something works from the command line but
> > not
> > >         when run by a
> > >         >         daemon,
> > >         >         my first suspicion is SELinux. Check the audit
> > log
> > >         for denials
> > >         >         around
> > >         >         that time.
> > >         >
> > >         >         I'd also check the system log and Pacemaker
> > detail
> > >         log around
> > >         >         that time
> > >         >         to see if there is any more information.
> > >         >         --
> > >         >         Ken Gaillot <kgail...@redhat.com>
> > >         >
> > >         >
> > >         >
> > >         >
> > >         >
> > >         >         _______________________________________________
> > >         >         Users mailing list: Users@clusterlabs.org
> > >         >         http://lists.clusterlabs.org/mailman/listinfo/u
> > sers
> > >         >
> > >         >         Project Home: http://www.clusterlabs.org
> > >         >         Getting started:
> > >         >
> > >          http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > >         >         Bugs: http://bugs.clusterlabs.org
> > >         >
> > >         >
> > >
> > >
> > >         --
> > >         Ken Gaillot <kgail...@redhat.com>
> > >
> > >
> > >
> > >
> > >
> > >
> > 
> > --
> > Ken Gaillot <kgail...@redhat.com>
> > 
> > 
> > 
> > 
> > 
> 
> 
-- 
Ken Gaillot <kgail...@redhat.com>

_______________________________________________
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to