Hi again I try remove all replication jobs and image files from target node... Still get critical error:
qm migrate 100 prox02 --online 2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02' (10.1.1.20) 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in current VM config) 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't live migrate attached local disks without with-local-disks option 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM - check log 2017-08-25 09:24:44 aborting phase 1 - cleanup resources 2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): Failed to sync data - can't migrate VM - check log migration aborted prox01:~# qm migrate 100 prox02 --online --with-local-disks 2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02' (10.1.1.20) 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in current VM config) 2017-08-25 09:24:58 copying disk images 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate VM with replicated volumes 2017-08-25 09:24:58 aborting phase 1 - cleanup resources 2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): Failed to sync data - can't live migrate VM with replicated volumes migration aborted prox01:~# pvesr status JobID Enabled Target LastSync NextSync Duration FailCount State 100-0 Yes local/prox02 2017-08-25_09:25:01 2017-08-25_12:00:00 15.200315 0 OK Somebody help me! Cheers 2017-08-24 9:55 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: > Well... > I will try it > > Thanks > > > > > 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csa...@proxmox.com>: > >> On 08/23/2017 08:50 PM, Gilberto Nunes wrote: >> >>> more info: >>> >>> >>> pvesr status >>> JobID Enabled Target LastSync >>> NextSync Duration FailCount State >>> 100-0 Yes local/prox01 - >>> 2017-08-23_15:55:04 3.151884 1 command 'set -o pipefail && >>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot >>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 | >>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01' root@10.1.1.10 >>> -- >>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed: >>> exit >>> code 255 >>> 100-1 Yes local/prox02 - >>> 2017-08-23_15:55:01 3.089044 1 command 'set -o pipefail && >>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 -snapshot >>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 | >>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02' root@10.1.1.20 >>> -- >>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' failed: >>> exit >>> code 255 >>> >>> >>> >> according to this output, no lastsync was completed, so i guess the >> replication did never work, so the migration will also not worK? >> >> i would remove all replication jobs (maybe with -force, via commandline) >> delete all images of this vm from all nodes where the vm *not* is at the >> moment (afaics from prox01 and prox02, as the vm is currently on prox03) >> >> then add the replication again wait for it to complete (verify with pvesr >> status) and try again to migrate >> >> _______________________________________________ >> pve-user mailing list >> pve-user@pve.proxmox.com >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> > > _______________________________________________ pve-user mailing list pve-user@pve.proxmox.com https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user