[Users] virt-io SCSI duplicate disk ID
Blaster
blaster at 556nato.com
Wed Jan 8 18:53:57 UTC 2014
So twice now under oVirt 3.3.2 I have added 2 virtio-scsi devices to a single virtual host.
After doing so, the VM would fail to boot due to duplicate disk ID. The first time I thought it a fluke, second time’s a bug?
Fortunately they were empty data disks and I was able to get around the problem by deleting one and recreating it.
VDSM log:
Thread-32154::INFO::2014-01-08 11:54:39,717::clientIF::350::vds::(prepareVolumePath) prepared volume path: /rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2
Thread-32154::DEBUG::2014-01-08 11:54:39,740::vm::2984::vm.Vm::(_run) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::<?xml version="1.0" encoding="utf-8"?>
<domain type="kvm">
<name>cobra</name>
<uuid>c2aff4cc-0de6-4342-a565-669b1825838c</uuid>
<memory>4194304</memory>
<currentMemory>4194304</currentMemory>
<vcpu>3</vcpu>
<memtune>
<min_guarantee>4194304</min_guarantee>
</memtune>
<devices>
<channel type="unix">
<target name="com.redhat.rhevm.vdsm" type="virtio"/>
<source mode="bind" path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.com.redhat.rhevm.vdsm"/>
</channel>
<channel type="unix">
<target name="org.qemu.guest_agent.0" type="virtio"/>
<source mode="bind" path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.org.qemu.guest_agent.0"/>
</channel>
<input bus="ps2" type="mouse"/>
<channel type="spicevmc">
<target name="com.redhat.spice.0" type="virtio"/>
</channel>
<graphics autoport="yes" keymap="en-us" listen="0" passwd="*****" passwdValidTo="1970-01-01T00:00:01" port="-1" tlsPort="-1" type="spice">
<channel mode="secure" name="main"/>
<channel mode="secure" name="inputs"/>
<channel mode="secure" name="cursor"/>
<channel mode="secure" name="playback"/>
<channel mode="secure" name="record"/>
<channel mode="secure" name="display"/>
<channel mode="secure" name="usbredir"/>
<channel mode="secure" name="display"/>
<channel mode="secure" name="usbredir"/>
<channel mode="secure" name="smartcard"/>
</graphics>
<controller model="virtio-scsi" type="scsi">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x05" type="pci"/>
</controller>
<video>
<address bus="0x00" domain="0x0000" function="0x0" slot="0x02" type="pci"/>
<model heads="1" type="qxl" vram="32768"/>
</video>
<interface type="bridge">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x03" type="pci"/>
<mac address="00:1a:4a:5b:9f:02"/>
<model type="virtio"/>
<source bridge="ovirtmgmt"/>
<filterref filter="vdsm-no-mac-spoofing"/>
<link state="up"/>
</interface>
<disk device="cdrom" snapshot="no" type="file">
<address bus="1" controller="0" target="0" type="drive" unit="0"/>
<source file="" startupPolicy="optional"/>
<target bus="ide" dev="hdc"/>
<readonly/>
<serial/>
</disk>
<disk device="disk" snapshot="no" type="file">
<source file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a"/>
<target bus="scsi" dev="sda"/>
<serial>94a298cb-87a5-48cd-b78f-770582b50639</serial>
<boot order="1"/>
<driver cache="none" error_policy="stop" io="threads" name="qemu" type="raw"/>
</disk>
<disk device="disk" snapshot="no" type="file">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x07" type="pci"/>
<source file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d"/>
<target bus="virtio" dev="vda"/>
<serial>8df43d38-c4c7-4711-bc87-55f35d1550e5</serial>
<driver cache="none" error_policy="stop" io="threads" name="qemu" type="raw"/>
</disk>
<disk device="disk" snapshot="no" type="file">
<address bus="0" controller="0" target="0" type="drive" unit="0"/>
<source file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2"/>
<target bus="scsi" dev="sdb"/>
<serial>e1886286-3d81-47d0-ae8d-77454e464078</serial>
<driver cache="none" error_policy="stop" io="threads" name="qemu" type="raw"/>
</disk>
<sound model="ich6">
<address bus="0x00" domain="0x0000" function="0x0" slot="0x04" type="pci"/>
</sound>
<memballoon model="virtio"/>
</devices>
<os>
<type arch="x86_64" machine="pc-1.0">hvm</type>
<smbios mode="sysinfo"/>
</os>
<sysinfo type="smbios">
<system>
<entry name="manufacturer">oVirt</entry>
<entry name="product">oVirt Node</entry>
<entry name="version">19-5</entry>
<entry name="serial">2061001F-C600-0006-E1BC-BCAEC518BA45</entry>
<entry name="uuid">c2aff4cc-0de6-4342-a565-669b1825838c</entry>
</system>
</sysinfo>
<clock adjustment="-21600" offset="variable">
<timer name="rtc" tickpolicy="catchup"/>
</clock>
<features>
<acpi/>
</features>
<cpu match="exact">
<model>Nehalem</model>
<topology cores="1" sockets="3" threads="1"/>
</cpu>
</domain>
Thread-32154::DEBUG::2014-01-08 11:54:40,218::libvirtconnection::108::libvirtconnection::(wrapper) Unknown libvirterror: ecode: 1 edom: 10 level: 2 message: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32154::DEBUG::2014-01-08 11:54:40,218::vm::2109::vm.Vm::(_startUnderlyingVm) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::_ongoingCreations released
Thread-32154::ERROR::2014-01-08 11:54:40,218::vm::2135::vm.Vm::(_startUnderlyingVm) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::The vm start process failed
Traceback (most recent call last):
File "/usr/share/vdsm/vm.py", line 2095, in _startUnderlyingVm
self._run()
File "/usr/share/vdsm/vm.py", line 3018, in _run
self._connection.createXML(domxml, flags),
File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", line 76, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 2805, in createXML
if ret is None:raise libvirtError('virDomainCreateXML() failed', conn=self)
libvirtError: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32154::DEBUG::2014-01-08 11:54:40,223::vm::2577::vm.Vm::(setDownStatus) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::Changed state to Down: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32158::WARNING::2014-01-08 11:54:42,185::clientIF::362::vds::(teardownVolumePath) Drive is not a vdsm image: VOLWM_CHUNK_MB:1024 VOLWM_CHUNK_REPLICATE_MULT:2 VOLWM_FREE_PCT:50 _blockDev:False _checkIoTuneCategories:<bound method Drive._checkIoTuneCategories of <vm.Drive object at 0x7f1150113a90>> _customize:<bound method Drive._customize of <vm.Drive object at 0x7f1150113a90>> _deviceXML:<disk device="cdrom" snapshot="no" type="file"><address bus="1" controller="0" target="0" type="drive" unit="0"/><source file="" startupPolicy="optional"/><target bus="ide" dev="hdc"/><readonly/><serial></serial></disk> _makeName:<bound method Drive._makeName of <vm.Drive object at 0x7f1150113a90>> _setExtSharedState:<bound method Drive._setExtSharedState of <vm.Drive object at 0x7f1150113a90>> _validateIoTuneParams:<bound method Drive._validateIoTuneParams of <vm.Drive object at 0x7f1150113a90>> address:{' controller': '0', ' target': '0', 'unit': '0', ' bus': '1', ' type': 'drive'} apparentsize:0 blockDev:False cache:none conf:{'status': 'Down', 'acpiEnable': 'true', 'emulatedMachine': 'pc-1.0', 'vmId': 'c2aff4cc-0de6-4342-a565-669b1825838c', 'pid': '0', 'memGuaranteedSize': 4096, 'timeOffset': '-21600', 'keyboardLayout': 'en-us', 'displayPort': '-1', 'displaySecurePort': '-1', 'spiceSslCipherSuite': 'DEFAULT', 'cpuType': 'Nehalem', 'custom': {'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8device_a2e6354f-4ad6-475f-bd18-754fcedf505f': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=a2e6354f-4ad6-475f-bd18-754fcedf505f, device=unix, type=CHANNEL, bootOrder=0, specParams={}, address={port=2, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel1, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=142f948d-f916-4f42-bd28-cb4f0b8ebb40, device=virtio-serial, type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci, slot=0x06, function=0x0}, managed=false, plugged=true, readOnly=false, deviceAlias=virtio-serial0, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5a': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=861eb290-19bc-4ace-b2cb-85cbb2e0eb5a, device=ide, type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci, slot=0x01, function=0x1}, managed=false, plugged=true, readOnly=false, deviceAlias=ide0, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=94db7fa0-071d-4181-bac6-826ecfca3dd8, device=unix, type=CHANNEL, bootOrder=0, specParams={}, address={port=1, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel0, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=615c1466-850e-4362-a4fb-60df1aaee1e8, device=spicevmc, type=CHANNEL, bootOrder=0, specParams={}, address={port=3, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel2, customProperties={}, snapshotId=null}'}, 'clientIp': '', 'exitCode': 1, 'nicModel': 'rtl8139,pv', 'smartcardEnable': 'false', 'kvmEnable': 'true', 'exitMessage': "internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive\n", 'transparentHugePages': 'true', 'devices': [{'specParams': {}, 'deviceId': 'db6166cb-e977-485e-8c82-fa48ca75e709', 'address': {'bus': '0x00', ' slot': '0x05', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'device': 'scsi', 'model': 'virtio-scsi', 'type': 'controller'}, {'device': 'qxl', 'specParams': {'vram': '32768', 'heads': '1'}, 'type': 'video', 'deviceId': '8b0e3dbc-27c6-4eae-ba6b-201c3e1736aa', 'address': {'bus': '0x00', ' slot': '0x02', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}}, {'nicModel': 'pv', 'macAddr': '00:1a:4a:5b:9f:02', 'linkActive': 'true', 'network': 'ovirtmgmt', 'filter': 'vdsm-no-mac-spoofing', 'specParams': {}, 'deviceId': '738c8ebe-b014-4d65-8c78-942aaf12bfb5', 'address': {'bus': '0x00', ' slot': '0x03', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'device': 'bridge', 'type': 'interface'}, {'index': '2', 'iface': 'ide', 'address': {' controller': '0', ' target': '0', 'unit': '0', ' bus': '1', ' type': 'drive'}, 'specParams': {'path': ''}, 'readonly': 'true', 'deviceId': '5611019a-948e-41b3-8ffd-75790ebcdf84', 'path': '', 'device': 'cdrom', 'shared': 'false', 'type': 'disk'}, {'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a', 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease', 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}, 'index': 0, 'iface': 'scsi', 'apparentsize': '162135015424', 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639', 'readonly': 'false', 'shared': 'false', 'truesize': '107119386624', 'type': 'disk', 'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw', 'deviceId': '94a298cb-87a5-48cd-b78f-770582b50639', 'poolID': '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a', 'propagateErrors': 'off', 'optional': 'false', 'bootOrder': '1', 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'specParams': {}, 'volumeChain': [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a', 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease', 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}]}, {'address': {'bus': '0x00', ' slot': '0x07', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease', 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}, 'index': '0', 'iface': 'virtio', 'apparentsize': '1073741824', 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5', 'readonly': 'false', 'shared': 'false', 'truesize': '0', 'type': 'disk', 'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw', 'deviceId': '8df43d38-c4c7-4711-bc87-55f35d1550e5', 'poolID': '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'propagateErrors': 'off', 'optional': 'false', 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'specParams': {}, 'volumeChain': [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease', 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}]}, {'address': {' controller': '0', ' target': '0', 'unit': '0', ' bus': '0', ' type': 'drive'}, 'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease', 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}, 'index': '1', 'iface': 'scsi', 'apparentsize': '1073741824', 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078', 'readonly': 'false', 'shared': 'false', 'truesize': '0', 'type': 'disk', 'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw', 'deviceId': 'e1886286-3d81-47d0-ae8d-77454e464078', 'poolID': '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'propagateErrors': 'off', 'optional': 'false', 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'specParams': {}, 'volumeChain': [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease', 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}]}, {'device': 'ich6', 'specParams': {}, 'type': 'sound', 'deviceId': 'a1e596e9-218f-46ba-9f32-b9c966e11d73', 'address': {'bus': '0x00', ' slot': '0x04', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}}, {'device': 'memballoon', 'specParams': {'model': 'virtio'}, 'type': 'balloon', 'deviceId': '5c04fd7e-7249-4e3a-b8eb-91cce72d5b60', 'target': 4194304}], 'smp': '3', 'vmType': 'kvm', 'memSize': 4096, 'displayIp': '0', 'spiceSecureChannels': 'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard', 'smpCoresPerSocket': '1', 'vmName': 'cobra', 'display': 'qxl', 'nice': '0'} createXmlElem:<bound method Drive.createXmlElem of <vm.Drive object at 0x7f1150113a90>> device:cdrom deviceId:5611019a-948e-41b3-8ffd-75790ebcdf84 extSharedState:none getLeasesXML:<bound method Drive.getLeasesXML of <vm.Drive object at 0x7f1150113a90>> getNextVolumeSize:<bound method Drive.getNextVolumeSize of <vm.Drive object at 0x7f1150113a90>> getXML:<bound method Drive.getXML of <vm.Drive object at 0x7f1150113a90>> hasVolumeLeases:False iface:ide index:2 isDiskReplicationInProgress:<bound method Drive.isDiskReplicationInProgress of <vm.Drive object at 0x7f1150113a90>> isVdsmImage:<bound method Drive.isVdsmImage of <vm.Drive object at 0x7f1150113a90>> log:<logUtils.SimpleLogAdapter object at 0x7f111838af90> name:hdc networkDev:False path: readonly:true reqsize:0 serial: shared:false specParams:{'path': ''} truesize:0 type:disk volExtensionChunk:1024 watermarkLimit:536870912
Traceback (most recent call last):
File "/usr/share/vdsm/clientIF.py", line 356, in teardownVolumePath
res = self.irs.teardownImage(drive['domainID'],
File "/usr/share/vdsm/vm.py", line 1389, in __getitem__
raise KeyError(key)
KeyError: 'domainID'
Thread-32158::DEBUG::2014-01-08 11:54:42,186::task::579::TaskManager.Task::(_updateState) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state init -> state preparing
Thread-32158::INFO::2014-01-08 11:54:42,187::logUtils::44::dispatcher::(wrapper) Run and protect: teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a', spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', imgUUID='94a298cb-87a5-48cd-b78f-770582b50639', volUUID=None)
Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::197::ResourceManager.Request::(__init__) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage'
Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::541::ResourceManager::(registerResource) Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' for lock type 'shared'
Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::600::ResourceManager::(registerResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking as 'shared' (1 active user)
Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::237::ResourceManager.Request::(grant) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Granted request
Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::811::TaskManager.Task::(resourceAcquired) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::_resourcesAcquired: Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::974::TaskManager.Task::(_decref) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 1 aborting False
Thread-32158::INFO::2014-01-08 11:54:42,188::logUtils::47::dispatcher::(wrapper) Run and protect: teardownImage, Return response: None
Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::1168::TaskManager.Task::(prepare) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::finished: None
Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::579::TaskManager.Task::(_updateState) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state preparing -> state finished
Thread-32158::DEBUG::2014-01-08 11:54:42,188::resourceManager::939::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj: 'None'>}
Thread-32158::DEBUG::2014-01-08 11:54:42,188::resourceManager::976::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}
Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::615::ResourceManager::(releaseResource) Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::634::ResourceManager::(releaseResource) Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active users)
Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::640::ResourceManager::(releaseResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out if anyone is waiting for it.
Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::648::ResourceManager::(releaseResource) No one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', Clearing records.
Thread-32158::DEBUG::2014-01-08 11:54:42,189::task::974::TaskManager.Task::(_decref) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 0 aborting False
Thread-32158::DEBUG::2014-01-08 11:54:42,189::task::579::TaskManager.Task::(_updateState) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state init -> state preparing
Thread-32158::INFO::2014-01-08 11:54:42,190::logUtils::44::dispatcher::(wrapper) Run and protect: teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a', spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', imgUUID='8df43d38-c4c7-4711-bc87-55f35d1550e5', volUUID=None)
Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::197::ResourceManager.Request::(__init__) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage'
Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::541::ResourceManager::(registerResource) Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' for lock type 'shared'
Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::600::ResourceManager::(registerResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking as 'shared' (1 active user)
Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::237::ResourceManager.Request::(grant) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Granted request
Thread-32158::DEBUG::2014-01-08 11:54:42,190::task::811::TaskManager.Task::(resourceAcquired) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::_resourcesAcquired: Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
Thread-32158::DEBUG::2014-01-08 11:54:42,191::task::974::TaskManager.Task::(_decref) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 1 aborting False
Thread-32158::INFO::2014-01-08 11:54:42,191::logUtils::47::dispatcher::(wrapper) Run and protect: teardownImage, Return response: None
Thread-32158::DEBUG::2014-01-08 11:54:42,191::task::1168::TaskManager.Task::(prepare) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::finished: None
Thread-32158::DEBUG::2014-01-08 11:54:42,191::task::579::TaskManager.Task::(_updateState) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state preparing -> state finished
Thread-32158::DEBUG::2014-01-08 11:54:42,191::resourceManager::939::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj: 'None'>}
Thread-32158::DEBUG::2014-01-08 11:54:42,191::resourceManager::976::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}
Thread-32158::DEBUG::2014-01-08 11:54:42,191::resourceManager::615::ResourceManager::(releaseResource) Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
Thread-32158::DEBUG::2014-01-08 11:54:42,192::resourceManager::634::ResourceManager::(releaseResource) Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active users)
Thread-32158::DEBUG::2014-01-08 11:54:42,192::resourceManager::640::ResourceManager::(releaseResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out if anyone is waiting for it.
Thread-32158::DEBUG::2014-01-08 11:54:42,192::resourceManager::648::ResourceManager::(releaseResource) No one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', Clearing records.
Thread-32158::DEBUG::2014-01-08 11:54:42,192::task::974::TaskManager.Task::(_decref) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 0 aborting False
Thread-32158::DEBUG::2014-01-08 11:54:42,192::task::579::TaskManager.Task::(_updateState) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state init -> state preparing
Thread-32158::INFO::2014-01-08 11:54:42,192::logUtils::44::dispatcher::(wrapper) Run and protect: teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a', spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', imgUUID='e1886286-3d81-47d0-ae8d-77454e464078', volUUID=None)
Thread-32158::DEBUG::2014-01-08 11:54:42,193::resourceManager::197::ResourceManager.Request::(__init__) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage'
Thread-32158::DEBUG::2014-01-08 11:54:42,193::resourceManager::541::ResourceManager::(registerResource) Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' for lock type 'shared'
Thread-32158::DEBUG::2014-01-08 11:54:42,196::resourceManager::600::ResourceManager::(registerResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking as 'shared' (1 active user)
Thread-32158::DEBUG::2014-01-08 11:54:42,196::resourceManager::237::ResourceManager.Request::(grant) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Granted request
Thread-32158::DEBUG::2014-01-08 11:54:42,196::task::811::TaskManager.Task::(resourceAcquired) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::_resourcesAcquired: Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
Thread-32158::DEBUG::2014-01-08 11:54:42,196::task::974::TaskManager.Task::(_decref) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::ref 1 aborting False
Thread-32158::INFO::2014-01-08 11:54:42,196::logUtils::47::dispatcher::(wrapper) Run and protect: teardownImage, Return response: None
Thread-32158::DEBUG::2014-01-08 11:54:42,196::task::1168::TaskManager.Task::(prepare) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::finished: None
Thread-32158::DEBUG::2014-01-08 11:54:42,197::task::579::TaskManager.Task::(_updateState) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state preparing -> state finished
Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::939::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj: 'None'>}
Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::976::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}
Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::615::ResourceManager::(releaseResource) Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::634::ResourceManager::(releaseResource) Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active users)
Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::640::ResourceManager::(releaseResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out if anyone is waiting for it.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20140108/9e6e9122/attachment-0001.html>
More information about the Users
mailing list