Public bug reported: Reproduced on juno version (actually tested on a fork from 2014.2.3, sorry in advance if invalid but i think the legacy version is also concerned by it)
not tested on younger versions, but looking at the code they seem impacted too For Rbd imagebackend only, when unrescuing an instance the disk.rescue file is actually not deleted on remote storage (only the rbd session is destroyed) Consequence: when rescuing instance once again, it simply ignores the new rescue image and takes the old _disk.rescue image Reproduce: 1. nova rescue instance (take care that you are booted to the vda rescue disk -> when rescuing an instance from the same image it was spawned from (case by default), since fs uuid is the same, according to your image fstab (if entry UUID=) you can actually boot from the image you are actually trying to rescue, but this is another matter that concerns template building) edit rescue image disk 2. nova unrescue instance 3. nova rescue instance -> you get back the disk.rescue spawned in 1 if confirmed, fix coming soon Concerning fix several possibilities: - nova.virt.libvirt.driver :LibvirtDriver-> unrescue method, not deleting the correct files or - nova.virt.libvirt.imagebackend:Rbd -> erase disk.rescue in create image method if already existing Rebuild not concerned by issue, delete instance correctly deletes files on remote storage ** Affects: nova Importance: Undecided Status: New ** Description changed: - Reproduced on juno version (actually tested on a fork from 2014.2.3, sorry in advance if invalid but i think the legacy version is also concerned by it) - - not tested on younger versions, but looking at the code they seem impacted too + Reproduced on juno version (actually tested on a fork from 2014.2.3, + sorry in advance if invalid but i think the legacy version is also + concerned by it) + + not tested on younger versions, but looking at the code they seem + impacted too For Rbd imagebackend only, when unrescuing an instance the disk.rescue file is actually not deleted on remote storage (only the rbd session is destroyed) Consequence: when rescuing instance once again, it simply ignores the new rescue image and takes the old _disk.rescue image - Reproduce: 1. nova rescue instance - (take care that you are booted to the vda rescue disk -> when rescuing an instance from the same image it was spawned from (case by default), since fs uuid is the same according to fsstab of the template (entry UUID=) you can actually boot from the image you are actually trying to rescue, but this is another matter that concerns template building) - + (take care that you are booted to the vda rescue disk -> when rescuing + an instance from the same image it was spawned from (case by default), + since fs uuid is the same, according to your image fstab (if entry + UUID=) you can actually boot from the image you are actually trying to + rescue, but this is another matter that concerns template building) + edit rescue image disk 2. nova unrescue instance 3. nova rescue instance -> you get back the disk.rescue spawned in 1 - if confirmed, fix coming soon - Concerning fix several possibilities: - - nova.virt.libvirt.driver :LibvirtDriver-> unrescue method, not deleting the correct files + Concerning fix several possibilities: + - nova.virt.libvirt.driver :LibvirtDriver-> unrescue method, not deleting the correct files or - nova.virt.libvirt.imagebackend:Rbd -> erase disk.rescue in create image method if already existing Rebuild not concerned by issue, delete instance correctly deletes files on remote storage -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1475652 Title: libvirt, rbd imagebackend, disk.rescue not deleted when unrescued Status in OpenStack Compute (nova): New Bug description: Reproduced on juno version (actually tested on a fork from 2014.2.3, sorry in advance if invalid but i think the legacy version is also concerned by it) not tested on younger versions, but looking at the code they seem impacted too For Rbd imagebackend only, when unrescuing an instance the disk.rescue file is actually not deleted on remote storage (only the rbd session is destroyed) Consequence: when rescuing instance once again, it simply ignores the new rescue image and takes the old _disk.rescue image Reproduce: 1. nova rescue instance (take care that you are booted to the vda rescue disk -> when rescuing an instance from the same image it was spawned from (case by default), since fs uuid is the same, according to your image fstab (if entry UUID=) you can actually boot from the image you are actually trying to rescue, but this is another matter that concerns template building) edit rescue image disk 2. nova unrescue instance 3. nova rescue instance -> you get back the disk.rescue spawned in 1 if confirmed, fix coming soon Concerning fix several possibilities: - nova.virt.libvirt.driver :LibvirtDriver-> unrescue method, not deleting the correct files or - nova.virt.libvirt.imagebackend:Rbd -> erase disk.rescue in create image method if already existing Rebuild not concerned by issue, delete instance correctly deletes files on remote storage To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1475652/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp