According to the docs: It replicates guest volumes to another node so that all data is available without using shared storage.
2017-08-25 10:11 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: > So.. One of the premise of the ZFS Replication volume, is to replicate > local volume to another node. > Or am I wrong? > > > Obrigado > > Cordialmente > > > Gilberto Ferreira > > Consultor TI Linux | IaaS Proxmox, CloudStack, KVM | Zentyal Server | > Zimbra Mail Server > > (47) 3025-5907 > (47) 99676-7530 > > Skype: gilberto.nunes36 > > > konnectati.com.br <http://www.konnectati.com.br/> > > > https://www.youtube.com/watch?v=dsiTPeNWcSE > > > 2017-08-25 10:07 GMT-03:00 Fábio Rabelo <fa...@fabiorabelo.wiki.br>: > >> this entry : >> >> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't >> live migrate attached local disks without with-local-disks option >> >> Seems to be the responsable . >> >> Local disk ? >> >> where this image are stored ? >> >> >> Fábio Rabelo >> >> 2017-08-25 9:36 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: >> > If I turn off the VM, migrate goes on. >> > But make offline migration is out of the question!!! >> > >> > >> > >> > 2017-08-25 9:28 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: >> > >> >> Hi again >> >> >> >> I try remove all replication jobs and image files from target node... >> >> Still get critical error: >> >> >> >> qm migrate 100 prox02 --online >> >> 2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02' >> >> (10.1.1.20) >> >> 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in current VM >> >> config) >> >> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't >> >> live migrate attached local disks without with-local-disks option >> >> 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM - >> check >> >> log >> >> 2017-08-25 09:24:44 aborting phase 1 - cleanup resources >> >> 2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): >> Failed >> >> to sync data - can't migrate VM - check log >> >> migration aborted >> >> prox01:~# qm migrate 100 prox02 --online --with-local-disks >> >> 2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02' >> >> (10.1.1.20) >> >> 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in current VM >> >> config) >> >> 2017-08-25 09:24:58 copying disk images >> >> 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate VM >> >> with replicated volumes >> >> 2017-08-25 09:24:58 aborting phase 1 - cleanup resources >> >> 2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): >> Failed >> >> to sync data - can't live migrate VM with replicated volumes >> >> migration aborted >> >> prox01:~# pvesr status >> >> JobID Enabled Target LastSync >> >> NextSync Duration FailCount State >> >> 100-0 Yes local/prox02 2017-08-25_09:25:01 >> >> 2017-08-25_12:00:00 15.200315 0 OK >> >> >> >> Somebody help me! >> >> >> >> Cheers >> >> >> >> >> >> >> >> >> >> 2017-08-24 9:55 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: >> >> >> >>> Well... >> >>> I will try it >> >>> >> >>> Thanks >> >>> >> >>> >> >>> >> >>> >> >>> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csa...@proxmox.com>: >> >>> >> >>>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote: >> >>>> >> >>>>> more info: >> >>>>> >> >>>>> >> >>>>> pvesr status >> >>>>> JobID Enabled Target LastSync >> >>>>> NextSync Duration FailCount State >> >>>>> 100-0 Yes local/prox01 - >> >>>>> 2017-08-23_15:55:04 3.151884 1 command 'set -o >> pipefail && >> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 >> -snapshot >> >>>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 | >> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01' >> root@10.1.1.10 >> >>>>> -- >> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' >> failed: >> >>>>> exit >> >>>>> code 255 >> >>>>> 100-1 Yes local/prox02 - >> >>>>> 2017-08-23_15:55:01 3.089044 1 command 'set -o >> pipefail && >> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 >> -snapshot >> >>>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 | >> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02' >> root@10.1.1.20 >> >>>>> -- >> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' >> failed: >> >>>>> exit >> >>>>> code 255 >> >>>>> >> >>>>> >> >>>>> >> >>>> according to this output, no lastsync was completed, so i guess the >> >>>> replication did never work, so the migration will also not worK? >> >>>> >> >>>> i would remove all replication jobs (maybe with -force, via >> commandline) >> >>>> delete all images of this vm from all nodes where the vm *not* is at >> the >> >>>> moment (afaics from prox01 and prox02, as the vm is currently on >> prox03) >> >>>> >> >>>> then add the replication again wait for it to complete (verify with >> >>>> pvesr status) and try again to migrate >> >>>> >> >>>> _______________________________________________ >> >>>> pve-user mailing list >> >>>> pve-user@pve.proxmox.com >> >>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> >>>> >> >>> >> >>> >> >> >> > _______________________________________________ >> > pve-user mailing list >> > pve-user@pve.proxmox.com >> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> _______________________________________________ >> pve-user mailing list >> pve-user@pve.proxmox.com >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> > > _______________________________________________ pve-user mailing list pve-user@pve.proxmox.com https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user