Hello Nir

Thanks for the answer. 
The output of the commands is below.


*********************************************************************************************************************************************
> 1. Please share the output of this command on one of the hosts:
> lvs -o vg_name,lv_name,tags | grep cdf1751b-64d3-42bc-b9ef-b0174c7ea068
*********************************************************************************************************************************************
# lvs -o vg_name,lv_name,tags | grep cdf1751b-64d3-42bc-b9ef-b0174c7ea068

  VG                                   LV                                   LV 
Tags
  ...
  6db73566-0f7f-4438-a9ef-6815075f45ea 208ece15-1c71-46f2-a019-6a9fce4309b2 
IU_cdf1751b-64d3-42bc-b9ef-b0174c7ea068,MD_23,PU_00000000-0000-0000-0000-000000000000
  6db73566-0f7f-4438-a9ef-6815075f45ea 4974a4cc-b388-456f-b98e-19d2158f0d58 
IU_cdf1751b-64d3-42bc-b9ef-b0174c7ea068,MD_15,PU_00000000-0000-0000-0000-000000000000
  6db73566-0f7f-4438-a9ef-6815075f45ea 8c66f617-7add-410c-b546-5214b0200832 
IU_cdf1751b-64d3-42bc-b9ef-b0174c7ea068,MD_16,PU_208ece15-1c71-46f2-a019-6a9fce4309b2
  ...


*********************************************************************************************************************************************
> 2. For every volume, share the output of qemu-img info:
> If the lv is not active, activate it:
>     lvchange -ay vg_name/lv_name
*********************************************************************************************************************************************

# lvdisplay 
6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2

  --- Logical volume ---
  LV Path                
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2
  LV Name                208ece15-1c71-46f2-a019-6a9fce4309b2
  VG Name                6db73566-0f7f-4438-a9ef-6815075f45ea
  LV UUID                k28hUo-Z6t7-wKdO-x7kz-ceYL-Vuzx-f9jLWi
  LV Write Access        read/write
  LV Creation host, time VM32.sub.holding.com, 2017-12-05 14:46:42 +0300
  LV Status              NOT available
  LV Size                33.00 GiB
  Current LE             264
  Segments               4
  Allocation             inherit
  Read ahead sectors     auto



# lvdisplay 
6db73566-0f7f-4438-a9ef-6815075f45ea/4974a4cc-b388-456f-b98e-19d2158f0d58

  --- Logical volume ---
  LV Path                
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/4974a4cc-b388-456f-b98e-19d2158f0d58
  LV Name                4974a4cc-b388-456f-b98e-19d2158f0d58
  VG Name                6db73566-0f7f-4438-a9ef-6815075f45ea
  LV UUID                HnnP01-JGxU-9zne-HB6n-BcaE-2lrM-qr9KPI
  LV Write Access        read/write
  LV Creation host, time VM12.sub.holding.com, 2018-07-31 03:35:20 +0300
  LV Status              NOT available
  LV Size                2.00 GiB
  Current LE             16
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto



# lvdisplay 
6db73566-0f7f-4438-a9ef-6815075f45ea/8c66f617-7add-410c-b546-5214b0200832

  --- Logical volume ---
  LV Path                
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/8c66f617-7add-410c-b546-5214b0200832
  LV Name                8c66f617-7add-410c-b546-5214b0200832
  VG Name                6db73566-0f7f-4438-a9ef-6815075f45ea
  LV UUID                MG1VRN-IqRn-mOGm-F4ul-ufbZ-Dywb-M3V14P
  LV Write Access        read/write
  LV Creation host, time VM12.sub.holding.com, 2018-08-01 03:34:31 +0300
  LV Status              NOT available
  LV Size                1.00 GiB
  Current LE             8
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto


# lvchange -ay 
6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2
# lvchange -ay 
6db73566-0f7f-4438-a9ef-6815075f45ea/4974a4cc-b388-456f-b98e-19d2158f0d58
# lvchange -ay 
6db73566-0f7f-4438-a9ef-6815075f45ea/8c66f617-7add-410c-b546-5214b0200832


*********************************************************************************************************************************************
> qemu-img info --backing /dev/vg_name/lv_name
*********************************************************************************************************************************************


# qemu-img info --backing 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2

image: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2
file format: qcow2
virtual size: 30G (32212254720 bytes)
disk size: 0
cluster_size: 65536
Format specific information:
    compat: 1.1
    lazy refcounts: false
    refcount bits: 16
    corrupt: false



# qemu-img info --backing 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/4974a4cc-b388-456f-b98e-19d2158f0d58

image: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/4974a4cc-b388-456f-b98e-19d2158f0d58
file format: qcow2
virtual size: 30G (32212254720 bytes)
disk size: 0
cluster_size: 65536
backing file: 208ece15-1c71-46f2-a019-6a9fce4309b2 (actual path: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2)
backing file format: qcow2
Format specific information:
    compat: 1.1
    lazy refcounts: false
    refcount bits: 16
    corrupt: false

image: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2
file format: qcow2
virtual size: 30G (32212254720 bytes)
disk size: 0
cluster_size: 65536
Format specific information:
    compat: 1.1
    lazy refcounts: false
    refcount bits: 16
    corrupt: false



# qemu-img info --backing 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/8c66f617-7add-410c-b546-5214b0200832

image: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/8c66f617-7add-410c-b546-5214b0200832
file format: qcow2
virtual size: 30G (32212254720 bytes)
disk size: 0
cluster_size: 65536
backing file: 208ece15-1c71-46f2-a019-6a9fce4309b2 (actual path: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2)
backing file format: qcow2
Format specific information:
    compat: 1.1
    lazy refcounts: false
    refcount bits: 16
    corrupt: false

image: 
/dev/6db73566-0f7f-4438-a9ef-6815075f45ea/208ece15-1c71-46f2-a019-6a9fce4309b2
file format: qcow2
virtual size: 30G (32212254720 bytes)
disk size: 0
cluster_size: 65536
Format specific information:
    compat: 1.1
    lazy refcounts: false
    refcount bits: 16
    corrupt: false



Do not quite understand. What should I do next?


15.08.2018, 15:33, "Nir Soffer" <nsof...@redhat.com>:
> On Tue, Aug 14, 2018 at 6:03 PM Алексей Максимов 
> <aleksey.i.maksi...@yandex.ru> wrote:
>> Hello, Nir
>>
>> Log in attachment.
>
> In the log we can see both createVolume and deleteVolume fail for this disk 
> uuid:
> cdf1751b-64d3-42bc-b9ef-b0174c7ea068
>
> 1. Please share the output of this command on one of the hosts:
>
>     lvs -o vg_name,lv_name,tags | grep cdf1751b-64d3-42bc-b9ef-b0174c7ea068
>
> This will show all the volumes belonging to this disk.
>
> 2. For every volume, share the output of qemu-img info:
>
> If the lv is not active, activate it:
>
>     lvchange -ay vg_name/lv_name
>
> Then run qemu-img info to find the actual chain:
>
>     qemu-img info --backing /dev/vg_name/lv_name
>
> If the lv was not active, deactivate it - we don't want to leave unused lvs 
> active.
>
>     lvchange -an vg_name/lv_name
> 3. On of the volume in the chain will not be part of the chain.
>
> No other volume will use it as backing file, and it may not have a backing
> file, or it may point to another volume in the chain.
>
> Once we found this volume, please check engine logs for this volume uuid. You 
> will probaly
> find that the volume was deleted in the past. Maybe you will not find it 
> since it was deleted
> months or years ago.
>
> 4. To verify that this volume does not have metadata, check the volume MD_N 
> tag.
> N is the offset in 512 bytes blocks from the start of the metadata volume.
>
> This will read the volume metadata block:
>
>     dd if=dev/vg_name/metadata bs=512 count=1 skip=N iflag=direct
>
> We expect to see:
>
>     NONE=####################################################...
>
> 5. To remove this volume use:
>
>     lvremove vg_name/lv_name
>
> Once the volume is removed, you will be able to create snapshot.
>
>> 14.08.2018, 01:30, "Nir Soffer" <nsof...@redhat.com>:
>>> On Mon, Aug 13, 2018 at 1:45 PM Aleksey Maksimov 
>>> <aleksey.i.maksi...@yandex.ru> wrote:
>>>> We use oVirt 4.2.5.2-1.el7 (Hosted engine / 4 hosts in cluster / about 
>>>> twenty virtual machines)
>>>> Virtual machine disks are located on the Data Domain from FC SAN.
>>>> Snapshots of all virtual machines are created normally. But for one 
>>>> virtual machine, we can not create a snapshot.
>>>>
>>>> When we try to create a snapshot in the oVirt web console, we see such 
>>>> errors:
>>>>
>>>> Aug 13, 2018, 1:05:06 PM Failed to complete snapshot 'KOM-APP14_BACKUP01' 
>>>> creation for VM 'KOM-APP14'.
>>>> Aug 13, 2018, 1:05:01 PM VDSM KOM-VM14 command HSMGetAllTasksStatusesVDS 
>>>> failed: Could not acquire resource. Probably resource factory threw an 
>>>> exception.: ()
>>>> Aug 13, 2018, 1:05:00 PM Snapshot 'KOM-APP14_BACKUP01' creation for VM 
>>>> 'KOM-APP14' was initiated by pe...@sub.holding.com@sub.holding.com-authz.
>>>>
>>>> At this time on the server with the role of "SPM" in the vdsm.log we see 
>>>> this:
>>>>
>>>> ...
>>>> 2018-08-13 05:05:06,471-0500 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC 
>>>> call VM.getStats succeeded in 0.00 seconds (__init__:573)
>>>> 2018-08-13 05:05:06,478-0500 INFO  (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC 
>>>> call Image.deleteVolumes succeeded in 0.05 seconds (__init__:573)
>>>> 2018-08-13 05:05:06,478-0500 INFO  (tasks/3) 
>>>> [storage.ThreadPool.WorkerThread] START task 
>>>> bb45ae7e-77e9-4fec-9ee2-8e1f0ad3d589 (cmd=<bound method Task.commit of 
>>>> <vdsm.storage.task.Task instance at 0x7f06b85a2128>>, args=None) 
>>>> (threadPool:208)
>>>> 2018-08-13 05:05:07,009-0500 WARN  (tasks/3) [storage.ResourceManager] 
>>>> Resource factory failed to create resource 
>>>> '01_img_6db73566-0f7f-4438-a9ef-6815075f45ea.cdf1751b-64d3-42bc-b9ef-b0174c7ea068'.
>>>>  Canceling request. (resourceManager:543)
>>>> Traceback (most recent call last):
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", 
>>>> line 539, in registerResource
>>>>     obj = namespaceObj.factory.createResource(name, lockType)
>>>>   File 
>>>> "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", line 
>>>> 193, in createResource
>>>>     lockType)
>>>>   File 
>>>> "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", line 
>>>> 122, in __getResourceCandidatesList
>>>>     imgUUID=resourceName)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/image.py", line 213, 
>>>> in getChain
>>>>     if srcVol.isLeaf():
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 
>>>> 1430, in isLeaf
>>>>     return self._manifest.isLeaf()
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 
>>>> 139, in isLeaf
>>>>     return self.getVolType() == sc.type2name(sc.LEAF_VOL)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 
>>>> 135, in getVolType
>>>>     self.voltype = self.getMetaParam(sc.VOLTYPE)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 
>>>> 119, in getMetaParam
>>>>     meta = self.getMetadata()
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/blockVolume.py", 
>>>> line 112, in getMetadata
>>>>     md = VolumeMetadata.from_lines(lines)
>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/volumemetadata.py", 
>>>> line 103, in from_lines
>>>>     "Missing metadata key: %s: found: %s" % (e, md))
>>>> MetaDataKeyNotFoundError: Meta Data key not found error: ("Missing 
>>>> metadata key: 'DOMAIN': found: {'NONE': 
>>>> '######################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################################'}",)
>>>
>>> Looks like you have a volume without metadata in the chain.
>>>
>>> This may happen in the past when deleting a volume failed, but we
>>> cleared the volume metadata. In current 4.2, this cannot happen, since
>>> we clear the metadata only if deleting the volume succeeded.
>>>
>>> Can you post complete vdsm log with this error?
>>>
>>> Once we find the volume without metadata, we can delete the LV
>>> using lvremove. This will fix the issue.
>>>
>>> Shani, do you remember the bug we have with this error?
>>> this probably the same issue.
>>>
>>> Ala, I think we need to add a tool to check and repair such chains.
>>>
>>> Nir


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VL2A7DK7H2AZ3R7CBZDWAN7RUQK47D6E/

Reply via email to