Hi, Yes, it is....
The qemu-kvm process is executed by the oneadmin user. When I cluster tries the live migration, what users do play? Oneadmin Root Hacluster I have just configured pasworless ssh connection with oneadmin. Do I need to configure any other passwordless ssh connection with any other user? What user executes the virsh migrate - - live? Is there any way to check ssk keys? Sorry for all theese questions. Thanks a lot El 1 sept. 2017 0:12, "Ken Gaillot" <kgail...@redhat.com> escribió: On Thu, 2017-08-31 at 23:45 +0200, Oscar Segarra wrote: > Hi Ken, > > > Thanks a lot for you quick answer: > > > Regarding to selinux, it is disabled. The FW is disabled as well. > > > [root@vdicnode01 ~]# sestatus > SELinux status: disabled > > > [root@vdicnode01 ~]# service firewalld status > Redirecting to /bin/systemctl status firewalld.service > ● firewalld.service - firewalld - dynamic firewall daemon > Loaded: loaded (/usr/lib/systemd/system/firewalld.service; > disabled; vendor preset: enabled) > Active: inactive (dead) > Docs: man:firewalld(1) > > > On migration, it performs a gracefully shutdown and a start on the new > node. > > > I attach the logs when trying to migrate from vdicnode02 to > vdicnode01: > > > vdicnode02 corosync.log: > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.161.2 2 > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.162.0 (null) > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_perform_op: > -- /cib/configuration/constraints/rsc_location[@id='location-vm-vdicdb01- vdicnode01--INFINITY'] > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @epoch=162, @num_updates=0 > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_replace operation for section > configuration: OK (rc=0, origin=vdicnode01/cibadmin/2, > version=0.162.0) > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_file_backup: Archived previous version > as /var/lib/pacemaker/cib/cib-65.raw > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_file_write_with_digest: Wrote version 0.162.0 of the CIB to > disk (digest: 1f87611b60cd7c48b95b6b788b47f65f) > Aug 31 23:38:17 [1521] vdicnode02 cib: info: > cib_file_write_with_digest: Reading cluster configuration > file /var/lib/pacemaker/cib/cib.jt2KPw > (digest: /var/lib/pacemaker/cib/cib.Kwqfpl) > Aug 31 23:38:22 [1521] vdicnode02 cib: info: > cib_process_ping: Reporting our current digest to vdicnode01: > dace3a23264934279d439420d5a716cc for 0.162.0 (0x7f96bb26c5c0 0) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.162.0 2 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.0 (null) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @epoch=163 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: ++ /cib/configuration/constraints: <rsc_location > id="location-vm-vdicdb01-vdicnode02--INFINITY" node="vdicnode02" > rsc="vm-vdicdb01" score="-INFINITY"/> > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_replace operation for section > configuration: OK (rc=0, origin=vdicnode01/cibadmin/2, > version=0.163.0) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_file_backup: Archived previous version > as /var/lib/pacemaker/cib/cib-66.raw > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_file_write_with_digest: Wrote version 0.163.0 of the CIB to > disk (digest: 47a548b36746de9275d66cc6aeb0fdc4) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_file_write_with_digest: Reading cluster configuration > file /var/lib/pacemaker/cib/cib.rcgXiT > (digest: /var/lib/pacemaker/cib/cib.7geMfi) > Aug 31 23:38:27 [1523] vdicnode02 lrmd: info: > cancel_recurring_action: Cancelling ocf operation > vm-vdicdb01_monitor_10000 > Aug 31 23:38:27 [1526] vdicnode02 crmd: info: > do_lrm_rsc_op: Performing > key=6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a > op=vm-vdicdb01_migrate_to_0 > Aug 31 23:38:27 [1523] vdicnode02 lrmd: info: log_execute: > executing - rsc:vm-vdicdb01 action:migrate_to call_id:9 > Aug 31 23:38:27 [1526] vdicnode02 crmd: info: > process_lrm_event: Result of monitor operation for vm-vdicdb01 on > vdicnode02: Cancelled | call=7 key=vm-vdicdb01_monitor_10000 > confirmed=true > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.163.0 2 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.1 (null) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @num_updates=1 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_migrate_to_0, @operation=migrate_to, @crm-debug-origin=cib_action_update, @transition-key=6:6:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=-1:193;6:6: 0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=-1, @rc-code=193, @op-status=-1, @last-run=1504215507, @last-rc-cha > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/41, version=0.163.1) > VirtualDomain(vm-vdicdb01)[5241]: 2017/08/31_23:38:27 INFO: > vdicdb01: Starting live migration to vdicnode01 (using: virsh > --connect=qemu:///system --quiet migrate --live vdicdb01 qemu > +ssh://vdicnode01/system ). > VirtualDomain(vm-vdicdb01)[5241]: 2017/08/31_23:38:27 ERROR: > vdicdb01: live migration to vdicnode01 failed: 1 > Aug 31 23:38:27 [1523] vdicnode02 lrmd: notice: > operation_finished: vm-vdicdb01_migrate_to_0:5241:stderr [ error: > Cannot recv data: Host key verification failed.: Connection reset by > peer ] ^^^ There you go. Sounds like the ssh key isn't being accepted. No idea why though. > Aug 31 23:38:27 [1523] vdicnode02 lrmd: notice: > operation_finished: vm-vdicdb01_migrate_to_0:5241:stderr > [ ocf-exit-reason:vdicdb01: live migration to vdicnode01 failed: 1 ] > Aug 31 23:38:27 [1523] vdicnode02 lrmd: info: log_finished: > finished - rsc:vm-vdicdb01 action:migrate_to call_id:9 pid:5241 > exit-code:1 exec-time:78ms queue-time:0ms > Aug 31 23:38:27 [1526] vdicnode02 crmd: notice: > process_lrm_event: Result of migrate_to operation for vm-vdicdb01 > on vdicnode02: 1 (unknown error) | call=9 key=vm-vdicdb01_migrate_to_0 > confirmed=true cib-update=14 > Aug 31 23:38:27 [1526] vdicnode02 crmd: notice: > process_lrm_event: vdicnode02-vm-vdicdb01_migrate_to_0:9 [ error: > Cannot recv data: Host key verification failed.: Connection reset by > peer\nocf-exit-reason:vdicdb01: live migration to vdicnode01 failed: 1 > \n ] > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_process_request: Forwarding cib_modify operation for section > status to all (origin=local/crmd/14) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.163.1 2 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.2 (null) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @num_updates=2 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @crm-debug-origin=do_update_resource, @transition-magic=0:1;6:6:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=9, @rc-code=1, @op-status=0, @exec-time=78, @exit-reason=vdicdb01: live migration to vdicnode01 failed: 1 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: > ++ /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']: <lrm_rsc_op id="vm-vdicdb01_last_failure_0" operation_key="vm-vdicdb01_migrate_to_0" operation="migrate_to" crm-debug-origin="do_update_resource" crm_feature_set="3.0.10" transition-key="6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" transition-magic="0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" exit-reason="vdicdb01: live migration to vdicn > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode02/crmd/14, version=0.163.2) > Aug 31 23:38:27 [1526] vdicnode02 crmd: info: > do_lrm_rsc_op: Performing > key=2:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a op=vm-vdicdb01_stop_0 > Aug 31 23:38:27 [1523] vdicnode02 lrmd: info: log_execute: > executing - rsc:vm-vdicdb01 action:stop call_id:10 > VirtualDomain(vm-vdicdb01)[5285]: 2017/08/31_23:38:27 INFO: > Issuing graceful shutdown request for domain vdicdb01. > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.163.2 2 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.3 (null) > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @num_updates=3 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-key=4:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=0:0;4:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=6, @rc-code=0, @last-run=1504215507, @last-rc-change=1504215507, @exec-time=57 > Aug 31 23:38:27 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/43, version=0.163.3) > Aug 31 23:38:30 [1523] vdicnode02 lrmd: info: log_finished: > finished - rsc:vm-vdicdb01 action:stop call_id:10 pid:5285 exit-code:0 > exec-time:3159ms queue-time:0ms > Aug 31 23:38:30 [1526] vdicnode02 crmd: notice: > process_lrm_event: Result of stop operation for vm-vdicdb01 on > vdicnode02: 0 (ok) | call=10 key=vm-vdicdb01_stop_0 confirmed=true > cib-update=15 > Aug 31 23:38:30 [1521] vdicnode02 cib: info: > cib_process_request: Forwarding cib_modify operation for section > status to all (origin=local/crmd/15) > Aug 31 23:38:30 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.163.3 2 > Aug 31 23:38:30 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.4 (null) > Aug 31 23:38:30 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @num_updates=4 > Aug 31 23:38:30 [1521] vdicnode02 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-key=2:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=0:0;2:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=10, @rc-code=0, @exec-time=3159 > Aug 31 23:38:30 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode02/crmd/15, version=0.163.4) > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.163.4 2 > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.5 (null) > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @num_updates=5 > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_start_0, @operation=start, @transition-key=5:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=0:0;5:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=7, @last-run=1504215510, @last-rc-change=1504215510, @exec-time=528 > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/44, version=0.163.5) > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: --- 0.163.5 2 > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: Diff: +++ 0.163.6 (null) > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: + /cib: @num_updates=6 > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_perform_op: > ++ /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']: <lrm_rsc_op id="vm-vdicdb01_monitor_10000" operation_key="vm-vdicdb01_monitor_10000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.10" transition-key="6:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" transition-magic="0:0;6:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" on_node="vdicnode01" call-id="8" rc-code="0" op-s > Aug 31 23:38:31 [1521] vdicnode02 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/45, version=0.163.6) > Aug 31 23:38:36 [1521] vdicnode02 cib: info: > cib_process_ping: Reporting our current digest to vdicnode01: > 9141ea9880f5a44b133003982d863bc8 for 0.163.6 (0x7f96bb2625a0 0) > > > > > > > vdicnode01 - corosync.log > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Forwarding cib_replace operation for section > configuration to all (origin=local/cibadmin/2) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.162.0 2 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.0 (null) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @epoch=163 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: ++ /cib/configuration/constraints: <rsc_location > id="location-vm-vdicdb01-vdicnode02--INFINITY" node="vdicnode02" > rsc="vm-vdicdb01" score="-INFINITY"/> > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_replace operation for section > configuration: OK (rc=0, origin=vdicnode01/cibadmin/2, > version=0.163.0) > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > abort_transition_graph: Transition aborted by > rsc_location.location-vm-vdicdb01-vdicnode02--INFINITY 'create': > Non-status change | cib=0.163.0 source=te_update_diff:436 > path=/cib/configuration/constraints complete=true > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: > do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE | > input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_file_backup: Archived previous version > as /var/lib/pacemaker/cib/cib-85.raw > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_file_write_with_digest: Wrote version 0.163.0 of the CIB to > disk (digest: 47a548b36746de9275d66cc6aeb0fdc4) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_file_write_with_digest: Reading cluster configuration > file /var/lib/pacemaker/cib/cib.npBIW2 > (digest: /var/lib/pacemaker/cib/cib.bDogoB) > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: > determine_online_status: Node vdicnode02 is online > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: > determine_online_status: Node vdicnode01 is online > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: native_print: > vm-vdicdb01 (ocf::heartbeat:VirtualDomain): Started vdicnode02 > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: RecurringOp: > Start recurring monitor (10s) for vm-vdicdb01 on vdicnode01 > Aug 31 23:38:27 [1535] vdicnode01 pengine: notice: LogActions: > Migrate vm-vdicdb01 (Started vdicnode02 -> vdicnode01) > Aug 31 23:38:27 [1535] vdicnode01 pengine: notice: > process_pe_message: Calculated transition 6, saving inputs > in /var/lib/pacemaker/pengine/pe-input-96.bz2 > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > do_state_transition: State transition S_POLICY_ENGINE -> > S_TRANSITION_ENGINE | input=I_PE_SUCCESS cause=C_IPC_MESSAGE > origin=handle_response > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: do_te_invoke: > Processing graph 6 (ref=pe_calc-dc-1504215507-24) derived > from /var/lib/pacemaker/pengine/pe-input-96.bz2 > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: > te_rsc_command: Initiating migrate_to operation > vm-vdicdb01_migrate_to_0 on vdicnode02 | action 6 > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > create_operation_update: cib_action_update: Updating resource > vm-vdicdb01 after migrate_to op pending (interval=0) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Forwarding cib_modify operation for section > status to all (origin=local/crmd/41) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.163.0 2 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.1 (null) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @num_updates=1 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_migrate_to_0, @operation=migrate_to, @crm-debug-origin=cib_action_update, @transition-key=6:6:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=-1:193;6:6: 0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=-1, @rc-code=193, @op-status=-1, @last-run=1504215507, @last-rc-cha > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/41, version=0.163.1) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.163.1 2 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.2 (null) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @num_updates=2 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @crm-debug-origin=do_update_resource, @transition-magic=0:1;6:6:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=9, @rc-code=1, @op-status=0, @exec-time=78, @exit-reason=vdicdb01: live migration to vdicnode01 failed: 1 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: > ++ /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']: <lrm_rsc_op id="vm-vdicdb01_last_failure_0" operation_key="vm-vdicdb01_migrate_to_0" operation="migrate_to" crm-debug-origin="do_update_resource" crm_feature_set="3.0.10" transition-key="6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" transition-magic="0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" exit-reason="vdicdb01: live migration to vdicn > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode02/crmd/14, version=0.163.2) > Aug 31 23:38:27 [1536] vdicnode01 crmd: warning: > status_from_rc: Action 6 (vm-vdicdb01_migrate_to_0) on vdicnode02 > failed (target: 0 vs. rc: 1): Error > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: > abort_transition_graph: Transition aborted by operation > vm-vdicdb01_migrate_to_0 'modify' on vdicnode02: Event failed | > magic=0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a cib=0.163.2 > source=match_graph_event:310 complete=false > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > match_graph_event: Action vm-vdicdb01_migrate_to_0 (6) confirmed > on vdicnode02 (rc=1) > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > process_graph_event: Detected action (6.6) > vm-vdicdb01_migrate_to_0.9=unknown error: failed > Aug 31 23:38:27 [1536] vdicnode01 crmd: warning: > status_from_rc: Action 6 (vm-vdicdb01_migrate_to_0) on vdicnode02 > failed (target: 0 vs. rc: 1): Error > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > abort_transition_graph: Transition aborted by operation > vm-vdicdb01_migrate_to_0 'create' on vdicnode02: Event failed | > magic=0:1;6:6:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a cib=0.163.2 > source=match_graph_event:310 complete=false > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > match_graph_event: Action vm-vdicdb01_migrate_to_0 (6) confirmed > on vdicnode02 (rc=1) > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > process_graph_event: Detected action (6.6) > vm-vdicdb01_migrate_to_0.9=unknown error: failed > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: run_graph: > Transition 6 (Complete=1, Pending=0, Fired=0, Skipped=0, > Incomplete=5, Source=/var/lib/pacemaker/pengine/pe-input-96.bz2): > Complete > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > do_state_transition: State transition S_TRANSITION_ENGINE -> > S_POLICY_ENGINE | input=I_PE_CALC cause=C_FSA_INTERNAL > origin=notify_crmd > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: > determine_online_status: Node vdicnode02 is online > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: > determine_online_status: Node vdicnode01 is online > Aug 31 23:38:27 [1535] vdicnode01 pengine: warning: > unpack_rsc_op_failure: Processing failed op migrate_to for > vm-vdicdb01 on vdicnode02: unknown error (1) > Aug 31 23:38:27 [1535] vdicnode01 pengine: warning: > unpack_rsc_op_failure: Processing failed op migrate_to for > vm-vdicdb01 on vdicnode02: unknown error (1) > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: native_print: > vm-vdicdb01 (ocf::heartbeat:VirtualDomain): FAILED > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: native_print: > 1 : vdicnode01 > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: native_print: > 2 : vdicnode02 > Aug 31 23:38:27 [1535] vdicnode01 pengine: error: > native_create_actions: Resource vm-vdicdb01 (ocf::VirtualDomain) is > active on 2 nodes attempting recovery > Aug 31 23:38:27 [1535] vdicnode01 pengine: warning: > native_create_actions: See > http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active for more > information. > Aug 31 23:38:27 [1535] vdicnode01 pengine: info: RecurringOp: > Start recurring monitor (10s) for vm-vdicdb01 on vdicnode01 > Aug 31 23:38:27 [1535] vdicnode01 pengine: notice: LogActions: > Recover vm-vdicdb01 (Started vdicnode01) > Aug 31 23:38:27 [1535] vdicnode01 pengine: error: > process_pe_message: Calculated transition 7 (with errors), saving > inputs in /var/lib/pacemaker/pengine/pe-error-8.bz2 > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > do_state_transition: State transition S_POLICY_ENGINE -> > S_TRANSITION_ENGINE | input=I_PE_SUCCESS cause=C_IPC_MESSAGE > origin=handle_response > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: do_te_invoke: > Processing graph 7 (ref=pe_calc-dc-1504215507-26) derived > from /var/lib/pacemaker/pengine/pe-error-8.bz2 > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: > te_rsc_command: Initiating stop operation vm-vdicdb01_stop_0 locally > on vdicnode01 | action 4 > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > do_lrm_rsc_op: Performing > key=4:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a op=vm-vdicdb01_stop_0 > Aug 31 23:38:27 [1533] vdicnode01 lrmd: info: log_execute: > executing - rsc:vm-vdicdb01 action:stop call_id:6 > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: > te_rsc_command: Initiating stop operation vm-vdicdb01_stop_0 on > vdicnode02 | action 2 > VirtualDomain(vm-vdicdb01)[5268]: 2017/08/31_23:38:27 INFO: > Domain vdicdb01 already stopped. > Aug 31 23:38:27 [1533] vdicnode01 lrmd: info: log_finished: > finished - rsc:vm-vdicdb01 action:stop call_id:6 pid:5268 exit-code:0 > exec-time:57ms queue-time:0ms > Aug 31 23:38:27 [1536] vdicnode01 crmd: notice: > process_lrm_event: Result of stop operation for vm-vdicdb01 on > vdicnode01: 0 (ok) | call=6 key=vm-vdicdb01_stop_0 confirmed=true > cib-update=43 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Forwarding cib_modify operation for section > status to all (origin=local/crmd/43) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.163.2 2 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.3 (null) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @num_updates=3 > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-key=4:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=0:0;4:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=6, @rc-code=0, @last-run=1504215507, @last-rc-change=1504215507, @exec-time=57 > Aug 31 23:38:27 [1536] vdicnode01 crmd: info: > match_graph_event: Action vm-vdicdb01_stop_0 (4) confirmed on > vdicnode01 (rc=0) > Aug 31 23:38:27 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/43, version=0.163.3) > Aug 31 23:38:30 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.163.3 2 > Aug 31 23:38:30 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.4 (null) > Aug 31 23:38:30 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @num_updates=4 > Aug 31 23:38:30 [1531] vdicnode01 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_stop_0, @operation=stop, @transition-key=2:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=0:0;2:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=10, @rc-code=0, @exec-time=3159 > Aug 31 23:38:30 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode02/crmd/15, version=0.163.4) > Aug 31 23:38:30 [1536] vdicnode01 crmd: info: > match_graph_event: Action vm-vdicdb01_stop_0 (2) confirmed on > vdicnode02 (rc=0) > Aug 31 23:38:30 [1536] vdicnode01 crmd: notice: > te_rsc_command: Initiating start operation vm-vdicdb01_start_0 locally > on vdicnode01 | action 5 > Aug 31 23:38:30 [1536] vdicnode01 crmd: info: > do_lrm_rsc_op: Performing > key=5:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a op=vm-vdicdb01_start_0 > Aug 31 23:38:30 [1533] vdicnode01 lrmd: info: log_execute: > executing - rsc:vm-vdicdb01 action:start call_id:7 > Aug 31 23:38:31 [1533] vdicnode01 lrmd: info: log_finished: > finished - rsc:vm-vdicdb01 action:start call_id:7 pid:5401 exit-code:0 > exec-time:528ms queue-time:0ms > Aug 31 23:38:31 [1536] vdicnode01 crmd: info: > action_synced_wait: Managed VirtualDomain_meta-data_0 process 5486 > exited with rc=0 > Aug 31 23:38:31 [1536] vdicnode01 crmd: notice: > process_lrm_event: Result of start operation for vm-vdicdb01 on > vdicnode01: 0 (ok) | call=7 key=vm-vdicdb01_start_0 confirmed=true > cib-update=44 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_process_request: Forwarding cib_modify operation for section > status to all (origin=local/crmd/44) > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.163.4 2 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.5 (null) > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @num_updates=5 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: + > /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']/lrm_rsc_op[@id='vm-vdicdb01_last_0']: @operation_key=vm-vdicdb01_start_0, @operation=start, @transition-key=5:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @transition-magic=0:0;5:7:0: fe1a9b0a-816c-4b97-96cb-b90dbf71417a, @call-id=7, @last-run=1504215510, @last-rc-change=1504215510, @exec-time=528 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/44, version=0.163.5) > Aug 31 23:38:31 [1536] vdicnode01 crmd: info: > match_graph_event: Action vm-vdicdb01_start_0 (5) confirmed on > vdicnode01 (rc=0) > Aug 31 23:38:31 [1536] vdicnode01 crmd: notice: > te_rsc_command: Initiating monitor operation vm-vdicdb01_monitor_10000 > locally on vdicnode01 | action 6 > Aug 31 23:38:31 [1536] vdicnode01 crmd: info: > do_lrm_rsc_op: Performing > key=6:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a > op=vm-vdicdb01_monitor_10000 > Aug 31 23:38:31 [1536] vdicnode01 crmd: info: > process_lrm_event: Result of monitor operation for vm-vdicdb01 on > vdicnode01: 0 (ok) | call=8 key=vm-vdicdb01_monitor_10000 > confirmed=false cib-update=45 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_process_request: Forwarding cib_modify operation for section > status to all (origin=local/crmd/45) > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: --- 0.163.5 2 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: Diff: +++ 0.163.6 (null) > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: + /cib: @num_updates=6 > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_perform_op: > ++ /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/ lrm_resource[@id='vm-vdicdb01']: <lrm_rsc_op id="vm-vdicdb01_monitor_10000" operation_key="vm-vdicdb01_monitor_10000" operation="monitor" crm-debug-origin="do_update_resource" crm_feature_set="3.0.10" transition-key="6:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" transition-magic="0:0;6:7:0:fe1a9b0a-816c-4b97-96cb-b90dbf71417a" on_node="vdicnode01" call-id="8" rc-code="0" op-s > Aug 31 23:38:31 [1531] vdicnode01 cib: info: > cib_process_request: Completed cib_modify operation for section > status: OK (rc=0, origin=vdicnode01/crmd/45, version=0.163.6) > Aug 31 23:38:31 [1536] vdicnode01 crmd: info: > match_graph_event: Action vm-vdicdb01_monitor_10000 (6) confirmed > on vdicnode01 (rc=0) > Aug 31 23:38:31 [1536] vdicnode01 crmd: notice: run_graph: > Transition 7 (Complete=5, Pending=0, Fired=0, Skipped=0, > Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-error-8.bz2): > Complete > Aug 31 23:38:31 [1536] vdicnode01 crmd: info: do_log: Input > I_TE_SUCCESS received in state S_TRANSITION_ENGINE from notify_crmd > Aug 31 23:38:31 [1536] vdicnode01 crmd: notice: > do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE > | input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd > Aug 31 23:38:36 [1531] vdicnode01 cib: info: > cib_process_ping: Reporting our current digest to vdicnode01: > 9141ea9880f5a44b133003982d863bc8 for 0.163.6 (0x7f61cec09270 0) > > > Thanks a lot > > 2017-08-31 16:20 GMT+02:00 Ken Gaillot <kgail...@redhat.com>: > On Thu, 2017-08-31 at 01:13 +0200, Oscar Segarra wrote: > > Hi, > > > > > > In my environment, I have just two hosts, where qemu-kvm > process is > > launched by a regular user (oneadmin) - open nebula - > > > > > > I have created a VirtualDomain resource that starts and > stops the VM > > perfectly. Nevertheless, when I change the location weight > in order to > > force the migration, It raises a migration failure "error: > 1" > > > > > > If I execute the virsh migrate command (that appears in > corosync.log) > > from command line, it works perfectly. > > > > > > Anybody has experienced the same issue? > > > > > > Thanks in advance for your help > > > If something works from the command line but not when run by a > daemon, > my first suspicion is SELinux. Check the audit log for denials > around > that time. > > I'd also check the system log and Pacemaker detail log around > that time > to see if there is any more information. > -- > Ken Gaillot <kgail...@redhat.com> > > > > > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > http://lists.clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > -- Ken Gaillot <kgail...@redhat.com>
_______________________________________________ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org