Sorry: "and wait for finish!" ****
Obrigado Cordialmente Gilberto Ferreira Consultor TI Linux | IaaS Proxmox, CloudStack, KVM | Zentyal Server | Zimbra Mail Server (47) 3025-5907 (47) 99676-7530 Skype: gilberto.nunes36 konnectati.com.br <http://www.konnectati.com.br/> https://www.youtube.com/watch?v=dsiTPeNWcSE 2017-08-25 10:22 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: > According to the design model of Proxmox Storage Replication, there is a > schedule to make the sync. > And of course, I set up the VM, I have scheduled the sync and for finish. > But still stuck! > > > > > > 2017-08-25 10:19 GMT-03:00 Fábio Rabelo <fa...@fabiorabelo.wiki.br>: > >> I never used zfs on Linux . >> >> But, in the Solaris OS family, this replication must be set up beforehand >> ... >> >> Someone with some milestone with zfs on linux can confirm or deny that ?? >> >> >> Fábio Rabelo >> >> 2017-08-25 10:11 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: >> > So.. One of the premise of the ZFS Replication volume, is to replicate >> > local volume to another node. >> > Or am I wrong? >> > >> > >> > Obrigado >> > >> > Cordialmente >> > >> > >> > Gilberto Ferreira >> > >> > Consultor TI Linux | IaaS Proxmox, CloudStack, KVM | Zentyal Server | >> > Zimbra Mail Server >> > >> > (47) 3025-5907 >> > (47) 99676-7530 >> > >> > Skype: gilberto.nunes36 >> > >> > >> > konnectati.com.br <http://www.konnectati.com.br/> >> > >> > >> > https://www.youtube.com/watch?v=dsiTPeNWcSE >> > >> > >> > 2017-08-25 10:07 GMT-03:00 Fábio Rabelo <fa...@fabiorabelo.wiki.br>: >> > >> >> this entry : >> >> >> >> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': can't >> >> live migrate attached local disks without with-local-disks option >> >> >> >> Seems to be the responsable . >> >> >> >> Local disk ? >> >> >> >> where this image are stored ? >> >> >> >> >> >> Fábio Rabelo >> >> >> >> 2017-08-25 9:36 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com>: >> >> > If I turn off the VM, migrate goes on. >> >> > But make offline migration is out of the question!!! >> >> > >> >> > >> >> > >> >> > 2017-08-25 9:28 GMT-03:00 Gilberto Nunes <gilberto.nune...@gmail.com >> >: >> >> > >> >> >> Hi again >> >> >> >> >> >> I try remove all replication jobs and image files from target >> node... >> >> >> Still get critical error: >> >> >> >> >> >> qm migrate 100 prox02 --online >> >> >> 2017-08-25 09:24:43 starting migration of VM 100 to node 'prox02' >> >> >> (10.1.1.20) >> >> >> 2017-08-25 09:24:44 found local disk 'stg:vm-100-disk-1' (in >> current VM >> >> >> config) >> >> >> 2017-08-25 09:24:44 can't migrate local disk 'stg:vm-100-disk-1': >> can't >> >> >> live migrate attached local disks without with-local-disks option >> >> >> 2017-08-25 09:24:44 ERROR: Failed to sync data - can't migrate VM - >> >> check >> >> >> log >> >> >> 2017-08-25 09:24:44 aborting phase 1 - cleanup resources >> >> >> 2017-08-25 09:24:44 ERROR: migration aborted (duration 00:00:02): >> Failed >> >> >> to sync data - can't migrate VM - check log >> >> >> migration aborted >> >> >> prox01:~# qm migrate 100 prox02 --online --with-local-disks >> >> >> 2017-08-25 09:24:58 starting migration of VM 100 to node 'prox02' >> >> >> (10.1.1.20) >> >> >> 2017-08-25 09:24:58 found local disk 'stg:vm-100-disk-1' (in >> current VM >> >> >> config) >> >> >> 2017-08-25 09:24:58 copying disk images >> >> >> 2017-08-25 09:24:58 ERROR: Failed to sync data - can't live migrate >> VM >> >> >> with replicated volumes >> >> >> 2017-08-25 09:24:58 aborting phase 1 - cleanup resources >> >> >> 2017-08-25 09:24:58 ERROR: migration aborted (duration 00:00:01): >> Failed >> >> >> to sync data - can't live migrate VM with replicated volumes >> >> >> migration aborted >> >> >> prox01:~# pvesr status >> >> >> JobID Enabled Target LastSync >> >> >> NextSync Duration FailCount State >> >> >> 100-0 Yes local/prox02 2017-08-25_09:25:01 >> >> >> 2017-08-25_12:00:00 15.200315 0 OK >> >> >> >> >> >> Somebody help me! >> >> >> >> >> >> Cheers >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> 2017-08-24 9:55 GMT-03:00 Gilberto Nunes < >> gilberto.nune...@gmail.com>: >> >> >> >> >> >>> Well... >> >> >>> I will try it >> >> >>> >> >> >>> Thanks >> >> >>> >> >> >>> >> >> >>> >> >> >>> >> >> >>> 2017-08-24 4:37 GMT-03:00 Dominik Csapak <d.csa...@proxmox.com>: >> >> >>> >> >> >>>> On 08/23/2017 08:50 PM, Gilberto Nunes wrote: >> >> >>>> >> >> >>>>> more info: >> >> >>>>> >> >> >>>>> >> >> >>>>> pvesr status >> >> >>>>> JobID Enabled Target LastSync >> >> >>>>> NextSync Duration FailCount State >> >> >>>>> 100-0 Yes local/prox01 - >> >> >>>>> 2017-08-23_15:55:04 3.151884 1 command 'set -o >> pipefail >> >> && >> >> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 >> >> -snapshot >> >> >>>>> __replicate_100-0_1503514204__ | /usr/bin/cstream -t 1024000000 | >> >> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox01' >> >> root@10.1.1.10 >> >> >>>>> -- >> >> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' >> failed: >> >> >>>>> exit >> >> >>>>> code 255 >> >> >>>>> 100-1 Yes local/prox02 - >> >> >>>>> 2017-08-23_15:55:01 3.089044 1 command 'set -o >> pipefail >> >> && >> >> >>>>> pvesm export local-zfs:vm-100-disk-1 zfs - -with-snapshots 1 >> >> -snapshot >> >> >>>>> __replicate_100-1_1503514201__ | /usr/bin/cstream -t 1024000000 | >> >> >>>>> /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=prox02' >> >> root@10.1.1.20 >> >> >>>>> -- >> >> >>>>> pvesm import local-zfs:vm-100-disk-1 zfs - -with-snapshots 1' >> failed: >> >> >>>>> exit >> >> >>>>> code 255 >> >> >>>>> >> >> >>>>> >> >> >>>>> >> >> >>>> according to this output, no lastsync was completed, so i guess >> the >> >> >>>> replication did never work, so the migration will also not worK? >> >> >>>> >> >> >>>> i would remove all replication jobs (maybe with -force, via >> >> commandline) >> >> >>>> delete all images of this vm from all nodes where the vm *not* is >> at >> >> the >> >> >>>> moment (afaics from prox01 and prox02, as the vm is currently on >> >> prox03) >> >> >>>> >> >> >>>> then add the replication again wait for it to complete (verify >> with >> >> >>>> pvesr status) and try again to migrate >> >> >>>> >> >> >>>> _______________________________________________ >> >> >>>> pve-user mailing list >> >> >>>> pve-user@pve.proxmox.com >> >> >>>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> >> >>>> >> >> >>> >> >> >>> >> >> >> >> >> > _______________________________________________ >> >> > pve-user mailing list >> >> > pve-user@pve.proxmox.com >> >> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> >> _______________________________________________ >> >> pve-user mailing list >> >> pve-user@pve.proxmox.com >> >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> >> >> > _______________________________________________ >> > pve-user mailing list >> > pve-user@pve.proxmox.com >> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> _______________________________________________ >> pve-user mailing list >> pve-user@pve.proxmox.com >> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user >> > > _______________________________________________ pve-user mailing list pve-user@pve.proxmox.com https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user