If I turn off the VM, migrate goes on. But make offline migration is out of the question!!!
2017-08-25 9:28 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: > Hi again > > I try remove all replication jobs and image files from target node... > Still get critical error: > > qm migrate 100 prox02 --online > 2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02' > (10.1.1.20) > 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in current VM > config) > 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't > live migrate attached local disks without with-local-disks option > 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM - check > log > 2017-08-25 09:24:44 aborting phase 1 - cleanup resources > 2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): Failed > to sync data - can't migrate VM - check log > migration aborted > prox01:~# qm migrate 100 prox02 --online --with-local-disks > 2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02' > (10.1.1.20) > 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in current VM > config) > 2017-08-25 09:24:58 copying disk images > 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate VM > with replicated volumes > 2017-08-25 09:24:58 aborting phase 1 - cleanup resources > 2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): Failed > to sync data - can't live migrate VM with replicated volumes > migration aborted > prox01:~# pvesr status > JobID Enabled Target LastSync > NextSync Duration FailCount State > 100-0 Yes local/prox02 2017-08-25_09:25:01 > 2017-08-25_12:00:00 15.200315 0 OK > > Somebody help me! > > Cheers > > > > > 2017-08-24 9:55 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: > >> Well... >> I will try it >> >> Thanks >> >> >> >> >> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csa...@proxmox.com>: >> >>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote: >>> >>>> more info: >>>> >>>> >>>> pvesr status >>>> JobID Enabled Target LastSync >>>> NextSync Duration FailCount State >>>> 100-0 Yes local/prox01 - >>>> 2017-08-23_15:55:04 3.151884 1 command 'set -o pipefail && >>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot >>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 | >>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01' root@10.1.1.10 >>>> -- >>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed: >>>> exit >>>> code 255 >>>> 100-1 Yes local/prox02 - >>>> 2017-08-23_15:55:01 3.089044 1 command 'set -o pipefail && >>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot >>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 | >>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02' root@10.1.1.20 >>>> -- >>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed: >>>> exit >>>> code 255 >>>> >>>> >>>> >>> according to this output, no lastsync was completed, so i guess the >>> replication did never work, so the migration will also not worK? >>> >>> i would remove all replication jobs (maybe with -force, via commandline) >>> delete all images of this vm from all nodes where the vm *not* is at the >>> moment (afaics from prox01 and prox02, as the vm is currently on prox03) >>> >>> then add the replication again wait for it to complete (verify with >>> pvesr status) and try again to migrate >>> >>> _______________________________________________ >>> pve-user mailing list >>> pve-user@pve.proxmox.com >>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >>> >> >> > _______________________________________________ pve-user mailing list pve-user@pve.proxmox.com https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user