From: "Blaster" <blas...@556nato.com>
To: users@ovirt.org
Sent: Wednesday, January 8, 2014 8:53:57 PM
Subject: [Users] virt-io SCSI duplicate disk ID
So twice now under oVirt 3.3.2 I have added 2 virtio-scsi devices to a single
virtual host.
After doing so, the VM would fail to boot due to duplicate disk ID. The first
time I thought it a fluke, second time’s a bug?
Fortunately they were empty data disks and I was able to get around the
problem by deleting one and recreating it.
VDSM log:
Thread-32154::INFO::2014-01-08
11:54:39,717::clientIF::350::vds::(prepareVolumePath) prepared volume path:
/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2
Thread-32154::DEBUG::2014-01-08 11:54:39,740::vm::2984::vm.Vm::(_run)
vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::<?xml version="1.0"
encoding="utf-8"?>
<domain type="kvm">
<name>cobra</name>
<uuid>c2aff4cc-0de6-4342-a565-669b1825838c</uuid>
<memory>4194304</memory>
<currentMemory>4194304</currentMemory>
<vcpu>3</vcpu>
<memtune>
<min_guarantee>4194304</min_guarantee>
</memtune>
<devices>
<channel type="unix">
<target name="com.redhat.rhevm.vdsm" type="virtio"/>
<source mode="bind"
path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.com.redhat.rhevm.vdsm"/>
</channel>
<channel type="unix">
<target name="org.qemu.guest_agent.0" type="virtio"/>
<source mode="bind"
path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.org.qemu.guest_agent.0"/>
</channel>
<input bus="ps2" type="mouse"/>
<channel type="spicevmc">
<target name="com.redhat.spice.0" type="virtio"/>
</channel>
<graphics autoport="yes" keymap="en-us" listen="0" passwd="*****"
passwdValidTo="1970-01-01T00:00:01" port="-1" tlsPort="-1" type="spice">
<channel mode="secure" name="main"/>
<channel mode="secure" name="inputs"/>
<channel mode="secure" name="cursor"/>
<channel mode="secure" name="playback"/>
<channel mode="secure" name="record"/>
<channel mode="secure" name="display"/>
<channel mode="secure" name="usbredir"/>
<channel mode="secure" name="display"/>
<channel mode="secure" name="usbredir"/>
<channel mode="secure" name="smartcard"/>
</graphics>
<controller model="virtio-scsi" type="scsi">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x05" type="pci"/>
</controller>
<video>
<address bus="0x00" domain="0x0000" function="0x0" slot="0x02" type="pci"/>
<model heads="1" type="qxl" vram="32768"/>
</video>
<interface type="bridge">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x03" type="pci"/>
<mac address="00:1a:4a:5b:9f:02"/>
<model type="virtio"/>
<source bridge="ovirtmgmt"/>
<filterref filter="vdsm-no-mac-spoofing"/>
<link state="up"/>
</interface>
<disk device="cdrom" snapshot="no" type="file">
<address bus="1" controller="0" target="0" type="drive" unit="0"/>
<source file="" startupPolicy="optional"/>
<target bus="ide" dev="hdc"/>
<readonly/>
<serial/>
</disk>
<disk device="disk" snapshot="no" type="file">
<source
file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a"/>
<target bus="scsi" dev="sda"/>
<serial>94a298cb-87a5-48cd-b78f-770582b50639</serial>
<boot order="1"/>
<driver cache="none" error_policy="stop" io="threads" name="qemu"
type="raw"/>
</disk>
<disk device="disk" snapshot="no" type="file">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x07" type="pci"/>
<source
file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d"/>
<target bus="virtio" dev="vda"/>
<serial>8df43d38-c4c7-4711-bc87-55f35d1550e5</serial>
<driver cache="none" error_policy="stop" io="threads" name="qemu"
type="raw"/>
</disk>
<disk device="disk" snapshot="no" type="file">
<address bus="0" controller="0" target="0" type="drive" unit="0"/>
<source
file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2"/>
<target bus="scsi" dev="sdb"/>
<serial>e1886286-3d81-47d0-ae8d-77454e464078</serial>
<driver cache="none" error_policy="stop" io="threads" name="qemu"
type="raw"/>
</disk>
<sound model="ich6">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x04" type="pci"/>
</sound>
<memballoon model="virtio"/>
</devices>
<os>
<type arch="x86_64" machine="pc-1.0">hvm</type>
<smbios mode="sysinfo"/>
</os>
<sysinfo type="smbios">
<system>
<entry name="manufacturer">oVirt</entry>
<entry name="product">oVirt Node</entry>
<entry name="version">19-5</entry>
<entry name="serial">2061001F-C600-0006-E1BC-BCAEC518BA45</entry>
<entry name="uuid">c2aff4cc-0de6-4342-a565-669b1825838c</entry>
</system>
</sysinfo>
<clock adjustment="-21600" offset="variable">
<timer name="rtc" tickpolicy="catchup"/>
</clock>
<features>
<acpi/>
</features>
<cpu match="exact">
<model>Nehalem</model>
<topology cores="1" sockets="3" threads="1"/>
</cpu>
</domain>
Thread-32154::DEBUG::2014-01-08
11:54:40,218::libvirtconnection::108::libvirtconnection::(wrapper) Unknown
libvirterror: ecode: 1 edom: 10 level: 2 message: internal error process
exited while connecting to monitor: qemu-system-x86_64: -drive
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32154::DEBUG::2014-01-08
11:54:40,218::vm::2109::vm.Vm::(_startUnderlyingVm)
vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::_ongoingCreations released
Thread-32154::ERROR::2014-01-08
11:54:40,218::vm::2135::vm.Vm::(_startUnderlyingVm)
vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::The vm start process failed
Traceback (most recent call last):
File "/usr/share/vdsm/vm.py", line 2095, in _startUnderlyingVm
self._run()
File "/usr/share/vdsm/vm.py", line 3018, in _run
self._connection.createXML(domxml, flags),
File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", line 76,
in wrapper
ret = f(*args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 2805, in createXML
if ret is None:raise libvirtError('virDomainCreateXML() failed', conn=self)
libvirtError: internal error process exited while connecting to monitor:
qemu-system-x86_64: -drive
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32154::DEBUG::2014-01-08
11:54:40,223::vm::2577::vm.Vm::(setDownStatus)
vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::Changed state to Down: internal
error process exited while connecting to monitor: qemu-system-x86_64: -drive
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32158::WARNING::2014-01-08
11:54:42,185::clientIF::362::vds::(teardownVolumePath) Drive is not a vdsm
image: VOLWM_CHUNK_MB:1024 VOLWM_CHUNK_REPLICATE_MULT:2 VOLWM_FREE_PCT:50
_blockDev:False _checkIoTuneCategories:<bound method
Drive._checkIoTuneCategories of <vm.Drive object at 0x7f1150113a90>>
_customize:<bound method Drive._customize of <vm.Drive object at
0x7f1150113a90>> _deviceXML:<disk device="cdrom" snapshot="no"
type="file"><address bus="1" controller="0" target="0" type="drive"
unit="0"/><source file="" startupPolicy="optional"/><target bus="ide"
dev="hdc"/><readonly/><serial></serial></disk> _makeName:<bound method
Drive._makeName of <vm.Drive object at 0x7f1150113a90>>
_setExtSharedState:<bound method Drive._setExtSharedState of <vm.Drive
object at 0x7f1150113a90>> _validateIoTuneParams:<bound method
Drive._validateIoTuneParams of <vm.Drive object at 0x7f1150113a90>>
address:{' controller': '0', ' target': '0', 'unit': '0', ' bus': '1', '
type': 'drive'} apparentsize:0 blockDev:False cache:none conf:{'status':
'Down', 'acpiEnable': 'true', 'emulatedMachine': 'pc-1.0', 'vmId':
'c2aff4cc-0de6-4342-a565-669b1825838c', 'pid': '0', 'memGuaranteedSize':
4096, 'timeOffset': '-21600', 'keyboardLayout': 'en-us', 'displayPort':
'-1', 'displaySecurePort': '-1', 'spiceSslCipherSuite': 'DEFAULT',
'cpuType': 'Nehalem', 'custom':
{'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8device_a2e6354f-4ad6-475f-bd18-754fcedf505f':
'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
deviceId=a2e6354f-4ad6-475f-bd18-754fcedf505f, device=unix, type=CHANNEL,
bootOrder=0, specParams={}, address={port=2, bus=0, controller=0,
type=virtio-serial}, managed=false, plugged=true, readOnly=false,
deviceAlias=channel1, customProperties={}, snapshotId=null}',
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40': 'VmDevice
{vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
deviceId=142f948d-f916-4f42-bd28-cb4f0b8ebb40, device=virtio-serial,
type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00,
domain=0x0000, type=pci, slot=0x06, function=0x0}, managed=false,
plugged=true, readOnly=false, deviceAlias=virtio-serial0,
customProperties={}, snapshotId=null}',
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5a':
'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
deviceId=861eb290-19bc-4ace-b2cb-85cbb2e0eb5a, device=ide, type=CONTROLLER,
bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci,
slot=0x01, function=0x1}, managed=false, plugged=true, readOnly=false,
deviceAlias=ide0, customProperties={}, snapshotId=null}',
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8':
'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
deviceId=94db7fa0-071d-4181-bac6-826ecfca3dd8, device=unix, type=CHANNEL,
bootOrder=0, specParams={}, address={port=1, bus=0, controller=0,
type=virtio-serial}, managed=false, plugged=true, readOnly=false,
deviceAlias=channel0, customProperties={}, snapshotId=null}',
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8':
'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
deviceId=615c1466-850e-4362-a4fb-60df1aaee1e8, device=spicevmc,
type=CHANNEL, bootOrder=0, specParams={}, address={port=3, bus=0,
controller=0, type=virtio-serial}, managed=false, plugged=true,
readOnly=false, deviceAlias=channel2, customProperties={},
snapshotId=null}'}, 'clientIp': '', 'exitCode': 1, 'nicModel': 'rtl8139,pv',
'smartcardEnable': 'false', 'kvmEnable': 'true', 'exitMessage': "internal
error process exited while connecting to monitor: qemu-system-x86_64: -drive
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
Duplicate ID 'drive-scsi0-0-0-0' for drive\n", 'transparentHugePages':
'true', 'devices': [{'specParams': {}, 'deviceId':
'db6166cb-e977-485e-8c82-fa48ca75e709', 'address': {'bus': '0x00', ' slot':
'0x05', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'device':
'scsi', 'model': 'virtio-scsi', 'type': 'controller'}, {'device': 'qxl',
'specParams': {'vram': '32768', 'heads': '1'}, 'type': 'video', 'deviceId':
'8b0e3dbc-27c6-4eae-ba6b-201c3e1736aa', 'address': {'bus': '0x00', ' slot':
'0x02', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}},
{'nicModel': 'pv', 'macAddr': '00:1a:4a:5b:9f:02', 'linkActive': 'true',
'network': 'ovirtmgmt', 'filter': 'vdsm-no-mac-spoofing', 'specParams': {},
'deviceId': '738c8ebe-b014-4d65-8c78-942aaf12bfb5', 'address': {'bus':
'0x00', ' slot': '0x03', ' domain': '0x0000', ' type': 'pci', ' function':
'0x0'}, 'device': 'bridge', 'type': 'interface'}, {'index': '2', 'iface':
'ide', 'address': {' controller': '0', ' target': '0', 'unit': '0', ' bus':
'1', ' type': 'drive'}, 'specParams': {'path': ''}, 'readonly': 'true',
'deviceId': '5611019a-948e-41b3-8ffd-75790ebcdf84', 'path': '', 'device':
'cdrom', 'shared': 'false', 'type': 'disk'}, {'volumeInfo': {'domainID':
'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0,
'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a',
'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'leasePath':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease',
'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}, 'index': 0, 'iface':
'scsi', 'apparentsize': '162135015424', 'imageID':
'94a298cb-87a5-48cd-b78f-770582b50639', 'readonly': 'false', 'shared':
'false', 'truesize': '107119386624', 'type': 'disk', 'domainID':
'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw',
'deviceId': '94a298cb-87a5-48cd-b78f-770582b50639', 'poolID':
'18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a',
'propagateErrors': 'off', 'optional': 'false', 'bootOrder': '1', 'volumeID':
'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'specParams': {}, 'volumeChain':
[{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path',
'leaseOffset': 0, 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a',
'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'leasePath':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease',
'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}]}, {'address': {'bus':
'0x00', ' slot': '0x07', ' domain': '0x0000', ' type': 'pci', ' function':
'0x0'}, 'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a',
'volType': 'path', 'leaseOffset': 0, 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d',
'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'leasePath':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease',
'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}, 'index': '0', 'iface':
'virtio', 'apparentsize': '1073741824', 'imageID':
'8df43d38-c4c7-4711-bc87-55f35d1550e5', 'readonly': 'false', 'shared':
'false', 'truesize': '0', 'type': 'disk', 'domainID':
'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw',
'deviceId': '8df43d38-c4c7-4711-bc87-55f35d1550e5', 'poolID':
'18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d',
'propagateErrors': 'off', 'optional': 'false', 'volumeID':
'42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'specParams': {}, 'volumeChain':
[{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path',
'leaseOffset': 0, 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d',
'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'leasePath':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease',
'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}]}, {'address': {'
controller': '0', ' target': '0', 'unit': '0', ' bus': '0', ' type':
'drive'}, 'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a',
'volType': 'path', 'leaseOffset': 0, 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'leasePath':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease',
'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}, 'index': '1', 'iface':
'scsi', 'apparentsize': '1073741824', 'imageID':
'e1886286-3d81-47d0-ae8d-77454e464078', 'readonly': 'false', 'shared':
'false', 'truesize': '0', 'type': 'disk', 'domainID':
'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw',
'deviceId': 'e1886286-3d81-47d0-ae8d-77454e464078', 'poolID':
'18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
'propagateErrors': 'off', 'optional': 'false', 'volumeID':
'1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'specParams': {}, 'volumeChain':
[{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path',
'leaseOffset': 0, 'path':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'leasePath':
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease',
'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}]}, {'device': 'ich6',
'specParams': {}, 'type': 'sound', 'deviceId':
'a1e596e9-218f-46ba-9f32-b9c966e11d73', 'address': {'bus': '0x00', ' slot':
'0x04', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}},
{'device': 'memballoon', 'specParams': {'model': 'virtio'}, 'type':
'balloon', 'deviceId': '5c04fd7e-7249-4e3a-b8eb-91cce72d5b60', 'target':
4194304}], 'smp': '3', 'vmType': 'kvm', 'memSize': 4096, 'displayIp': '0',
'spiceSecureChannels':
'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard',
'smpCoresPerSocket': '1', 'vmName': 'cobra', 'display': 'qxl', 'nice': '0'}
createXmlElem:<bound method Drive.createXmlElem of <vm.Drive object at
0x7f1150113a90>> device:cdrom deviceId:5611019a-948e-41b3-8ffd-75790ebcdf84
extSharedState:none getLeasesXML:<bound method Drive.getLeasesXML of
<vm.Drive object at 0x7f1150113a90>> getNextVolumeSize:<bound method
Drive.getNextVolumeSize of <vm.Drive object at 0x7f1150113a90>>
getXML:<bound method Drive.getXML of <vm.Drive object at 0x7f1150113a90>>
hasVolumeLeases:False iface:ide index:2 isDiskReplicationInProgress:<bound
method Drive.isDiskReplicationInProgress of <vm.Drive object at
0x7f1150113a90>> isVdsmImage:<bound method Drive.isVdsmImage of <vm.Drive
object at 0x7f1150113a90>> log:<logUtils.SimpleLogAdapter object at
0x7f111838af90> name:hdc networkDev:False path: readonly:true reqsize:0
serial: shared:false specParams:{'path': ''} truesize:0 type:disk
volExtensionChunk:1024 watermarkLimit:536870912
Traceback (most recent call last):
File "/usr/share/vdsm/clientIF.py", line 356, in teardownVolumePath
res = self.irs.teardownImage(drive['domainID'],
File "/usr/share/vdsm/vm.py", line 1389, in __getitem__
raise KeyError(key)
KeyError: 'domainID'
Thread-32158::DEBUG::2014-01-08
11:54:42,186::task::579::TaskManager.Task::(_updateState)
Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state init -> state
preparing
Thread-32158::INFO::2014-01-08
11:54:42,187::logUtils::44::dispatcher::(wrapper) Run and protect:
teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a',
spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9',
imgUUID='94a298cb-87a5-48cd-b78f-770582b50639', volUUID=None)
Thread-32158::DEBUG::2014-01-08
11:54:42,187::resourceManager::197::ResourceManager.Request::(__init__)
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage'
Thread-32158::DEBUG::2014-01-08
11:54:42,187::resourceManager::541::ResourceManager::(registerResource)
Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
for lock type 'shared'
Thread-32158::DEBUG::2014-01-08
11:54:42,187::resourceManager::600::ResourceManager::(registerResource)
Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking
as 'shared' (1 active user)
Thread-32158::DEBUG::2014-01-08
11:54:42,187::resourceManager::237::ResourceManager.Request::(grant)
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Granted
request
Thread-32158::DEBUG::2014-01-08
11:54:42,188::task::811::TaskManager.Task::(resourceAcquired)
Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::_resourcesAcquired:
Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
Thread-32158::DEBUG::2014-01-08
11:54:42,188::task::974::TaskManager.Task::(_decref)
Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 1 aborting False
Thread-32158::INFO::2014-01-08
11:54:42,188::logUtils::47::dispatcher::(wrapper) Run and protect:
teardownImage, Return response: None
Thread-32158::DEBUG::2014-01-08
11:54:42,188::task::1168::TaskManager.Task::(prepare)
Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::finished: None
Thread-32158::DEBUG::2014-01-08
11:54:42,188::task::579::TaskManager.Task::(_updateState)
Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state preparing ->
state finished
Thread-32158::DEBUG::2014-01-08
11:54:42,188::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources
{'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef
'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj:
'None'>}
Thread-32158::DEBUG::2014-01-08
11:54:42,188::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-32158::DEBUG::2014-01-08
11:54:42,189::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
Thread-32158::DEBUG::2014-01-08
11:54:42,189::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active
users)
Thread-32158::DEBUG::2014-01-08
11:54:42,189::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out
if anyone is waiting for it.
Thread-32158::DEBUG::2014-01-08
11:54:42,189::resourceManager::648::ResourceManager::(releaseResource) No
one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a',
Clearing records.
Thread-32158::DEBUG::2014-01-08
11:54:42,189::task::974::TaskManager.Task::(_decref)
Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 0 aborting False
Thread-32158::DEBUG::2014-01-08
11:54:42,189::task::579::TaskManager.Task::(_updateState)
Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state init -> state
preparing
Thread-32158::INFO::2014-01-08
11:54:42,190::logUtils::44::dispatcher::(wrapper) Run and protect:
teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a',
spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9',
imgUUID='8df43d38-c4c7-4711-bc87-55f35d1550e5', volUUID=None)
Thread-32158::DEBUG::2014-01-08
11:54:42,190::resourceManager::197::ResourceManager.Request::(__init__)
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage'
Thread-32158::DEBUG::2014-01-08
11:54:42,190::resourceManager::541::ResourceManager::(registerResource)
Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
for lock type 'shared'
Thread-32158::DEBUG::2014-01-08
11:54:42,190::resourceManager::600::ResourceManager::(registerResource)
Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking
as 'shared' (1 active user)
Thread-32158::DEBUG::2014-01-08
11:54:42,190::resourceManager::237::ResourceManager.Request::(grant)
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Granted
request
Thread-32158::DEBUG::2014-01-08
11:54:42,190::task::811::TaskManager.Task::(resourceAcquired)
Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::_resourcesAcquired:
Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
Thread-32158::DEBUG::2014-01-08
11:54:42,191::task::974::TaskManager.Task::(_decref)
Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 1 aborting False
Thread-32158::INFO::2014-01-08
11:54:42,191::logUtils::47::dispatcher::(wrapper) Run and protect:
teardownImage, Return response: None
Thread-32158::DEBUG::2014-01-08
11:54:42,191::task::1168::TaskManager.Task::(prepare)
Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::finished: None
Thread-32158::DEBUG::2014-01-08
11:54:42,191::task::579::TaskManager.Task::(_updateState)
Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state preparing ->
state finished
Thread-32158::DEBUG::2014-01-08
11:54:42,191::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources
{'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef
'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj:
'None'>}
Thread-32158::DEBUG::2014-01-08
11:54:42,191::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-32158::DEBUG::2014-01-08
11:54:42,191::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
Thread-32158::DEBUG::2014-01-08
11:54:42,192::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active
users)
Thread-32158::DEBUG::2014-01-08
11:54:42,192::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out
if anyone is waiting for it.
Thread-32158::DEBUG::2014-01-08
11:54:42,192::resourceManager::648::ResourceManager::(releaseResource) No
one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a',
Clearing records.
Thread-32158::DEBUG::2014-01-08
11:54:42,192::task::974::TaskManager.Task::(_decref)
Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 0 aborting False
Thread-32158::DEBUG::2014-01-08
11:54:42,192::task::579::TaskManager.Task::(_updateState)
Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state init -> state
preparing
Thread-32158::INFO::2014-01-08
11:54:42,192::logUtils::44::dispatcher::(wrapper) Run and protect:
teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a',
spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9',
imgUUID='e1886286-3d81-47d0-ae8d-77454e464078', volUUID=None)
Thread-32158::DEBUG::2014-01-08
11:54:42,193::resourceManager::197::ResourceManager.Request::(__init__)
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage'
Thread-32158::DEBUG::2014-01-08
11:54:42,193::resourceManager::541::ResourceManager::(registerResource)
Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
for lock type 'shared'
Thread-32158::DEBUG::2014-01-08
11:54:42,196::resourceManager::600::ResourceManager::(registerResource)
Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking
as 'shared' (1 active user)
Thread-32158::DEBUG::2014-01-08
11:54:42,196::resourceManager::237::ResourceManager.Request::(grant)
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Granted
request
Thread-32158::DEBUG::2014-01-08
11:54:42,196::task::811::TaskManager.Task::(resourceAcquired)
Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::_resourcesAcquired:
Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
Thread-32158::DEBUG::2014-01-08
11:54:42,196::task::974::TaskManager.Task::(_decref)
Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::ref 1 aborting False
Thread-32158::INFO::2014-01-08
11:54:42,196::logUtils::47::dispatcher::(wrapper) Run and protect:
teardownImage, Return response: None
Thread-32158::DEBUG::2014-01-08
11:54:42,196::task::1168::TaskManager.Task::(prepare)
Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::finished: None
Thread-32158::DEBUG::2014-01-08
11:54:42,197::task::579::TaskManager.Task::(_updateState)
Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state preparing ->
state finished
Thread-32158::DEBUG::2014-01-08
11:54:42,197::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources
{'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef
'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj:
'None'>}
Thread-32158::DEBUG::2014-01-08
11:54:42,197::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-32158::DEBUG::2014-01-08
11:54:42,197::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
Thread-32158::DEBUG::2014-01-08
11:54:42,197::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active
users)
Thread-32158::DEBUG::2014-01-08
11:54:42,197::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out
if anyone is waiting for it.
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users