Theoretically, all the data for the snapshot is both in the Engine's DB and on 
storage (OVF file).

I was afraid to migrate from 4.3  to  4.4  ,  as I was planning  to  wipe the 
engine  and  just import the VMs..., but  I was  not sure  about the snapshots.

Most probably this is a bug.

@Sandro,

can you assist with this one  ?

Best Regards,
Strahil Nikolov

На 31 юли 2020 г. 10:01:17 GMT+03:00, Alex K <rightkickt...@gmail.com> написа:
>Has anyone been able to import a storage domain and still have access
>to VM
>snapshots or this might be a missing feature/bug that needs to be
>reported?
>Reading the redhat docs about the storage domain import it seems there
>is
>no mention of VM snapshots if they should be accessible following the
>import.
>Thanx
>
>On Thu, Jul 30, 2020 at 3:58 PM Alex K <rightkickt...@gmail.com> wrote:
>
>> Hi all,
>>
>> I have a dual node self hosted cluster v4.3 using gluster as storage
>so as
>> to test an actual scenario which will need to be followed at
>production.
>> The purpose is to rename the cluster FQDN to a new one, wiping out
>any
>> reference to the old previous FQDN. I was not successful in using the
>> engine-rename tool or other means as there are leftovers from
>previous FQDN
>> that cause issues.
>>
>> The cluster has a data storage domain with one guest VM running on it
>> which has one snapshot.
>> I am testing a destructive scenario as below and I find out that when
>> importing the storage domain to the newly configured cluster, while
>the
>> guest VM is imported fine, I do not see the guest VM disk snapshots.
>>
>> Steps that I follow for this scenario:
>>
>> *Initial status: *
>> I have an ovirt cluster with two hosts named v0 and v1.
>> The gluster storage domain is configured at a separate network where
>the
>> hosts are named gluster0 and gluster1.
>> The cluster has an engine and data storage domain named "engine" and
>"vms"
>> respectively.
>> The "vms" storage domain hosts one guest VM with one guest VM disk
>> snapshot.
>> All are configured with fqdn *localdomain.local*
>>
>> *# Steps to rename all cluster to new fqdn lab.local and import "vms"
>> storage domain*
>> 1. Set v1 ovirt host at maintenance then remove it from GUI.
>> 2.  At v1 install fresh CentOS7 using the new FQDN lab.local
>> 3.  at v0 set global maintenance and shutdown engine. Remove the
>engine
>> storage data. (complete wipe of any engine related data. What is
>important
>> is only VM guests and their snapshots).
>> 4.  at v0, remove bricks belonging to "engine" and "vms" gluster
>volumes
>> of v1 and detach gluster peer v1.
>>
>> gluster volume remove-brick engine replica 1
>> gluster1:/gluster/engine/brick force
>> gluster volume remove-brick vms replica 1 gluster1:/gluster/vms/brick
>force
>> gluster peer detach gluster1
>>
>> 5.  On v1, prepare gluster service, reattach peer and add bricks from
>v0.
>> At this phase all data from vms gluster volume will be synced to the
>new
>> host. Verify with `gluster heal info vms`.
>> from v0 server run:
>>
>> gluster peer probe gluster1
>> gluster volume add-brick engine replica 2
>gluster1:/gluster/engine/brick
>> gluster volume add-brick vms replica 2 gluster1:/gluster/vms/brick
>>
>> At this state all gluster volume are up and in sync. We confirm "vms"
>sync
>> with
>> gluster volume heal info vms
>>
>> 6.  At freshly installed v1 install engine using the same clean
>gluster
>> engine volume:
>> hosted-engine --deploy --config-append=/root/storage.conf
>> --config-append=answers.conf (use new FQDN!)
>>
>> 7.  Upon completion of engine deployment and after having ensured the
>vms
>> gluster volume is synced (step 5) remove bricks of v0 host (v0 now
>should
>> not be visible at ovirt GUI) and detach gluster peer v0.
>> at v1 host run:
>> gluster volume remove-brick engine replica 1
>> gluster0:/gluster/engine/brick force
>> gluster volume remove-brick vms replica 1 gluster0:/gluster/vms/brick
>force
>> gluster peer detach gluster0
>>
>> 8. Install fresh CentOS7 on v0 and prepare it with ovirt node
>packages,
>> networking and gluster.
>> 9. At v0, attach gluster bricks from v1. Confirm sync with gluster
>volume
>> heal info.
>> at v1 host:
>> gluster peer probe gluster0
>> gluster volume add-brick engine replica 2
>gluster0:/gluster/engine/brick
>> gluster volume add-brick vms replica 2 gluster0:/gluster/vms/brick
>>
>> 10. at engine, add entry for v0 host at /etc/hosts. At ovirt GUI, add
>v0.
>> /etc/hosts:
>> 10.10.10.220 node0 v0.lab.local
>> 10.10.10.221 node1 v1.lab.local
>> 10.10.10.222 engine.lab.local engine
>>
>> 10.100.100.1 gluster0
>> 10.100.100.2 gluster1
>>
>> 11. At ovirt GUI import vms gluster volume as vms storage domain.
>> At this step I have to approve operation:
>> [image: image.png]
>>
>>
>> 12. At ovirt GUI, import VMs from vms storage domain.
>> At this step the VM is found and imported from the imported storage
>domain
>> "vms", but the VM does not show the previously available disk
>snapshot.
>>
>> The import of the storage domain should have retained the guest VM
>> snapshot.
>> How can this be troubleshooted? Do I have to keep some type of engine
>DB
>> backup so as to make the snapshots visible? If yes, is it possible to
>> restore this backup to a fresh engine that has a new FQDN?
>> Thanx very much for any advise and hint.
>>
>> Alex
>>
>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/P32K3NJHGCQMOEVKR7FCWKSJPLIZVQJD/

Reply via email to