it's because the VM is down, you can manually activate using
$ lvchange -a y vgname/lvname

remember to deactivate after


On Tue, Feb 26, 2019 at 6:15 PM Alan G <alan+ov...@griff.me.uk> wrote:

> I tried that initially but I'm not sure how to access the image on block
> storage? The lv is marked as NOT available in lvdisplay.
>
> --- Logical volume ---
>   LV Path
> /dev/70205101-c6b1-4034-a9a2-e559897273bc/74d27dd2-3887-4833-9ce3-5925dbd551cc
>   LV Name                74d27dd2-3887-4833-9ce3-5925dbd551cc
>   VG Name                70205101-c6b1-4034-a9a2-e559897273bc
>   LV UUID                svAB48-Rgnd-0V2A-2O07-Z2Ic-4zfO-XyJiFo
>   LV Write Access        read/write
>   LV Creation host, time nyc-ovirt-01.redacted.com, 2018-05-15 12:02:41
> +0000
>   LV Status              NOT available
>   LV Size                14.00 GiB
>   Current LE             112
>   Segments               9
>   Allocation             inherit
>   Read ahead sectors     auto
>
>
> ---- On Tue, 26 Feb 2019 15:57:47 +0000 *Benny Zlotnik
> <bzlot...@redhat.com <bzlot...@redhat.com>>* wrote ----
>
> I haven't found anything other the leaks issue, you can try to run
> $ qemu-img check -r leaks <img>
> (make sure to have it backed up)
>
> On Tue, Feb 26, 2019 at 5:40 PM Alan G <alan+ov...@griff.me.uk> wrote:
>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/JCTVJ4K7QNA653MA7YZZSYNU5OG2SNQE/
>
>>
>> Logs are attached. The first error from snapshot deletion is
>> at 2019-02-26 13:27:11,877Z in the engine log.
>>
>>
>> ---- On Tue, 26 Feb 2019 15:11:39 +0000 *Benny Zlotnik
>> <bzlot...@redhat.com <bzlot...@redhat.com>>* wrote ----
>>
>> Can you provide full vdsm & engine logs?
>>
>> On Tue, Feb 26, 2019 at 5:10 PM Alan G <alan+ov...@griff.me.uk> wrote:
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/XKTHCAZENAB2WLPJGDLGDDUWKPCOJCFV/
>>
>>>
>>> Hi,
>>>
>>>
>>> I performed the following: -
>>>
>>> 1. Shutdown VM.
>>> 2. Take a snapshot
>>> 3. Create a clone from snapshot.
>>> 4. Start the clone. Clone starts fine.
>>> 5. Attempt to delete snapshot from original VM, fails.
>>> 6. Attempt to start original VM, fails with "Bad volume specification".
>>>
>>> This was logged in VDSM during the snapshot deletion attempt.
>>>
>>> 2019-02-26 13:27:10,907+0000 ERROR (tasks/3) [storage.TaskManager.Task]
>>> (Task='67577e64-f29d-4c47-a38f-e54b905cae03') Unexpected error (task:872)
>>> Traceback (most recent call last):
>>>   File "/usr/share/vdsm/storage/task.py", line 879, in _run
>>>     return fn(*args, **kargs)
>>>   File "/usr/share/vdsm/storage/task.py", line 333, in run
>>>     return self.cmd(*self.argslist, **self.argsdict)
>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py",
>>> line 79, in wrapper
>>>     return method(self, *args, **kwargs)
>>>   File "/usr/share/vdsm/storage/sp.py", line 1892, in finalizeMerge
>>>     merge.finalize(subchainInfo)
>>>   File "/usr/share/vdsm/storage/merge.py", line 271, in finalize
>>>     optimal_size = subchain.base_vol.optimal_size()
>>>   File "/usr/share/vdsm/storage/blockVolume.py", line 440, in
>>> optimal_size
>>>     check = qemuimg.check(self.getVolumePath(), qemuimg.FORMAT.QCOW2)
>>>   File "/usr/lib/python2.7/site-packages/vdsm/qemuimg.py", line 157, in
>>> check
>>>     out = _run_cmd(cmd)
>>>   File "/usr/lib/python2.7/site-packages/vdsm/qemuimg.py", line 426, in
>>> _run_cmd
>>>     raise QImgError(cmd, rc, out, err)
>>> QImgError: cmd=['/usr/bin/qemu-img', 'check', '--output', 'json', '-f',
>>> 'qcow2',
>>> '/rhev/data-center/mnt/blockSD/024109d5-ea84-47ed-87e5-1c8681fdd177/images/f7dea7bd-04
>>> 6c-4923-b5a5-d0c1201607fc/ac540314-989d-42c2-9e7e-3907eedbe27f'],
>>> ecode=3, stdout={
>>>     "image-end-offset": 52210892800,
>>>     "total-clusters": 1638400,
>>>     "check-errors": 0,
>>>     "leaks": 323,
>>>     "leaks-fixed": 0,
>>>     "allocated-clusters": 795890,
>>>     "filename":
>>> "/rhev/data-center/mnt/blockSD/024109d5-ea84-47ed-87e5-1c8681fdd177/images/f7dea7bd-046c-4923-b5a5-d0c1201607fc/ac540314-989d-42c2-9e7e-3907eedbe27f",
>>>     "format": "qcow2",
>>>     "fragmented-clusters": 692941
>>> }
>>> , stderr=Leaked cluster 81919 refcount=1 reference=0
>>> Leaked cluster 81920 refcount=1 reference=0
>>> Leaked cluster 81921 refcount=1 reference=0
>>> etc..
>>>
>>> Is there any way to fix these leaked clusters?
>>>
>>> Running oVirt 4.1.9 with FC block storage.
>>>
>>> Thanks,
>>>
>>> Alan
>>>
>>> _______________________________________________
>>> Users mailing list -- users@ovirt.org
>>> To unsubscribe send an email to users-le...@ovirt.org
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct:
>>> https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives:
>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/RHYFOHEYPK33FEKKLNHJPJBY2KXYKE56/
>>>
>>
>>
>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HUCHWNLOLPY7SBOBZ5DJQPFKXTRCPCCN/

Reply via email to