
Hi Daniel, Both times were on the same hypervisor which was a fresh 3.3.2 install, not an upgrade. One time was using disk images and the other time was using direct LUN. I will send log files to you directly. On 1/8/2014 3:15 PM, Daniel Erez wrote:
Hi Blaster,
Have you added the second disk after upgrading oVirt version? An explicit address setting mechanism has been introduced recently, which might cause such problems between minor versions. Can you please attach the full engine/vdsm logs?
Thanks, Daniel
----- Original Message -----
From: "Blaster" <blaster@556nato.com> To: users@ovirt.org Sent: Wednesday, January 8, 2014 8:53:57 PM Subject: [Users] virt-io SCSI duplicate disk ID
So twice now under oVirt 3.3.2 I have added 2 virtio-scsi devices to a single virtual host.
After doing so, the VM would fail to boot due to duplicate disk ID. The first time I thought it a fluke, second time’s a bug?
Fortunately they were empty data disks and I was able to get around the problem by deleting one and recreating it.
VDSM log:
Thread-32154::INFO::2014-01-08 11:54:39,717::clientIF::350::vds::(prepareVolumePath) prepared volume path: /rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2 Thread-32154::DEBUG::2014-01-08 11:54:39,740::vm::2984::vm.Vm::(_run) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::<?xml version="1.0" encoding="utf-8"?> <domain type="kvm"> <name>cobra</name> <uuid>c2aff4cc-0de6-4342-a565-669b1825838c</uuid> <memory>4194304</memory> <currentMemory>4194304</currentMemory> <vcpu>3</vcpu> <memtune> <min_guarantee>4194304</min_guarantee> </memtune> <devices> <channel type="unix"> <target name="com.redhat.rhevm.vdsm" type="virtio"/> <source mode="bind" path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.com.redhat.rhevm.vdsm"/> </channel> <channel type="unix"> <target name="org.qemu.guest_agent.0" type="virtio"/> <source mode="bind" path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.org.qemu.guest_agent.0"/> </channel> <input bus="ps2" type="mouse"/> <channel type="spicevmc"> <target name="com.redhat.spice.0" type="virtio"/> </channel> <graphics autoport="yes" keymap="en-us" listen="0" passwd="*****" passwdValidTo="1970-01-01T00:00:01" port="-1" tlsPort="-1" type="spice"> <channel mode="secure" name="main"/> <channel mode="secure" name="inputs"/> <channel mode="secure" name="cursor"/> <channel mode="secure" name="playback"/> <channel mode="secure" name="record"/> <channel mode="secure" name="display"/> <channel mode="secure" name="usbredir"/>
<channel mode="secure" name="display"/> <channel mode="secure" name="usbredir"/> <channel mode="secure" name="smartcard"/> </graphics> <controller model="virtio-scsi" type="scsi"> <address bus="0x00" domain="0x0000" function="0x0" slot="0x05" type="pci"/> </controller> <video> <address bus="0x00" domain="0x0000" function="0x0" slot="0x02" type="pci"/> <model heads="1" type="qxl" vram="32768"/> </video> <interface type="bridge"> <address bus="0x00" domain="0x0000" function="0x0" slot="0x03" type="pci"/> <mac address="00:1a:4a:5b:9f:02"/> <model type="virtio"/> <source bridge="ovirtmgmt"/> <filterref filter="vdsm-no-mac-spoofing"/> <link state="up"/> </interface> <disk device="cdrom" snapshot="no" type="file"> <address bus="1" controller="0" target="0" type="drive" unit="0"/> <source file="" startupPolicy="optional"/> <target bus="ide" dev="hdc"/> <readonly/> <serial/> </disk> <disk device="disk" snapshot="no" type="file"> <source file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a"/> <target bus="scsi" dev="sda"/> <serial>94a298cb-87a5-48cd-b78f-770582b50639</serial> <boot order="1"/> <driver cache="none" error_policy="stop" io="threads" name="qemu" type="raw"/> </disk> <disk device="disk" snapshot="no" type="file"> <address bus="0x00" domain="0x0000" function="0x0" slot="0x07" type="pci"/> <source file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d"/> <target bus="virtio" dev="vda"/> <serial>8df43d38-c4c7-4711-bc87-55f35d1550e5</serial> <driver cache="none" error_policy="stop" io="threads" name="qemu" type="raw"/> </disk> <disk device="disk" snapshot="no" type="file">
<address bus="0" controller="0" target="0" type="drive" unit="0"/> <source file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2"/> <target bus="scsi" dev="sdb"/> <serial>e1886286-3d81-47d0-ae8d-77454e464078</serial> <driver cache="none" error_policy="stop" io="threads" name="qemu" type="raw"/> </disk> <sound model="ich6"> <address bus="0x00" domain="0x0000" function="0x0" slot="0x04" type="pci"/> </sound> <memballoon model="virtio"/> </devices> <os> <type arch="x86_64" machine="pc-1.0">hvm</type> <smbios mode="sysinfo"/> </os> <sysinfo type="smbios"> <system> <entry name="manufacturer">oVirt</entry> <entry name="product">oVirt Node</entry> <entry name="version">19-5</entry> <entry name="serial">2061001F-C600-0006-E1BC-BCAEC518BA45</entry> <entry name="uuid">c2aff4cc-0de6-4342-a565-669b1825838c</entry> </system> </sysinfo> <clock adjustment="-21600" offset="variable"> <timer name="rtc" tickpolicy="catchup"/> </clock> <features> <acpi/> </features> <cpu match="exact"> <model>Nehalem</model> <topology cores="1" sockets="3" threads="1"/> </cpu> </domain> Thread-32154::DEBUG::2014-01-08 11:54:40,218::libvirtconnection::108::libvirtconnection::(wrapper) Unknown libvirterror: ecode: 1 edom: 10 level: 2 message: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32154::DEBUG::2014-01-08 11:54:40,218::vm::2109::vm.Vm::(_startUnderlyingVm) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::_ongoingCreations released Thread-32154::ERROR::2014-01-08 11:54:40,218::vm::2135::vm.Vm::(_startUnderlyingVm) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::The vm start process failed Traceback (most recent call last): File "/usr/share/vdsm/vm.py", line 2095, in _startUnderlyingVm self._run() File "/usr/share/vdsm/vm.py", line 3018, in _run self._connection.createXML(domxml, flags), File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", line 76, in wrapper ret = f(*args, **kwargs) File "/usr/lib64/python2.7/site-packages/libvirt.py", line 2805, in createXML if ret is None:raise libvirtError('virDomainCreateXML() failed', conn=self) libvirtError: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32154::DEBUG::2014-01-08 11:54:40,223::vm::2577::vm.Vm::(setDownStatus) vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::Changed state to Down: internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive
Thread-32158::WARNING::2014-01-08 11:54:42,185::clientIF::362::vds::(teardownVolumePath) Drive is not a vdsm image: VOLWM_CHUNK_MB:1024 VOLWM_CHUNK_REPLICATE_MULT:2 VOLWM_FREE_PCT:50 _blockDev:False _checkIoTuneCategories:<bound method Drive._checkIoTuneCategories of <vm.Drive object at 0x7f1150113a90>> _customize:<bound method Drive._customize of <vm.Drive object at 0x7f1150113a90>> _deviceXML:<disk device="cdrom" snapshot="no" type="file"><address bus="1" controller="0" target="0" type="drive" unit="0"/><source file="" startupPolicy="optional"/><target bus="ide" dev="hdc"/><readonly/><serial></serial></disk> _makeName:<bound method Drive._makeName of <vm.Drive object at 0x7f1150113a90>> _setExtSharedState:<bound method Drive._setExtSharedState of <vm.Drive object at 0x7f1150113a90>> _validateIoTuneParams:<bound method Drive._validateIoTuneParams of <vm.Drive object at 0x7f1150113a90>> address:{' controller': '0', ' target': '0', 'unit': '0', ' bus': '1', ' type': 'drive'} apparentsize:0 blockDev:False cache:none conf:{'status': 'Down', 'acpiEnable': 'true', 'emulatedMachine': 'pc-1.0', 'vmId': 'c2aff4cc-0de6-4342-a565-669b1825838c', 'pid': '0', 'memGuaranteedSize': 4096, 'timeOffset': '-21600', 'keyboardLayout': 'en-us', 'displayPort': '-1', 'displaySecurePort': '-1', 'spiceSslCipherSuite': 'DEFAULT', 'cpuType': 'Nehalem', 'custom': {'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8device_a2e6354f-4ad6-475f-bd18-754fcedf505f': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=a2e6354f-4ad6-475f-bd18-754fcedf505f, device=unix, type=CHANNEL, bootOrder=0, specParams={}, address={port=2, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel1, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=142f948d-f916-4f42-bd28-cb4f0b8ebb40, device=virtio-serial, type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci, slot=0x06, function=0x0}, managed=false, plugged=true, readOnly=false, deviceAlias=virtio-serial0, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5a': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=861eb290-19bc-4ace-b2cb-85cbb2e0eb5a, device=ide, type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci, slot=0x01, function=0x1}, managed=false, plugged=true, readOnly=false, deviceAlias=ide0, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=94db7fa0-071d-4181-bac6-826ecfca3dd8, device=unix, type=CHANNEL, bootOrder=0, specParams={}, address={port=1, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel0, customProperties={}, snapshotId=null}', 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8': 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c, deviceId=615c1466-850e-4362-a4fb-60df1aaee1e8, device=spicevmc, type=CHANNEL, bootOrder=0, specParams={}, address={port=3, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel2, customProperties={}, snapshotId=null}'}, 'clientIp': '', 'exitCode': 1, 'nicModel': 'rtl8139,pv', 'smartcardEnable': 'false', 'kvmEnable': 'true', 'exitMessage': "internal error process exited while connecting to monitor: qemu-system-x86_64: -drive file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads: Duplicate ID 'drive-scsi0-0-0-0' for drive\n", 'transparentHugePages': 'true', 'devices': [{'specParams': {}, 'deviceId': 'db6166cb-e977-485e-8c82-fa48ca75e709', 'address': {'bus': '0x00', ' slot': '0x05', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'device': 'scsi', 'model': 'virtio-scsi', 'type': 'controller'}, {'device': 'qxl', 'specParams': {'vram': '32768', 'heads': '1'}, 'type': 'video', 'deviceId': '8b0e3dbc-27c6-4eae-ba6b-201c3e1736aa', 'address': {'bus': '0x00', ' slot': '0x02', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}}, {'nicModel': 'pv', 'macAddr': '00:1a:4a:5b:9f:02', 'linkActive': 'true', 'network': 'ovirtmgmt', 'filter': 'vdsm-no-mac-spoofing', 'specParams': {}, 'deviceId': '738c8ebe-b014-4d65-8c78-942aaf12bfb5', 'address': {'bus': '0x00', ' slot': '0x03', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'device': 'bridge', 'type': 'interface'}, {'index': '2', 'iface': 'ide', 'address': {' controller': '0', ' target': '0', 'unit': '0', ' bus': '1', ' type': 'drive'}, 'specParams': {'path': ''}, 'readonly': 'true', 'deviceId': '5611019a-948e-41b3-8ffd-75790ebcdf84', 'path': '', 'device': 'cdrom', 'shared': 'false', 'type': 'disk'}, {'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a', 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease', 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}, 'index': 0, 'iface': 'scsi', 'apparentsize': '162135015424', 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639', 'readonly': 'false', 'shared': 'false', 'truesize': '107119386624', 'type': 'disk', 'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw', 'deviceId': '94a298cb-87a5-48cd-b78f-770582b50639', 'poolID': '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a', 'propagateErrors': 'off', 'optional': 'false', 'bootOrder': '1', 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'specParams': {}, 'volumeChain': [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a', 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease', 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}]}, {'address': {'bus': '0x00', ' slot': '0x07', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}, 'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease', 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}, 'index': '0', 'iface': 'virtio', 'apparentsize': '1073741824', 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5', 'readonly': 'false', 'shared': 'false', 'truesize': '0', 'type': 'disk', 'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw', 'deviceId': '8df43d38-c4c7-4711-bc87-55f35d1550e5', 'poolID': '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'propagateErrors': 'off', 'optional': 'false', 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'specParams': {}, 'volumeChain': [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease', 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}]}, {'address': {' controller': '0', ' target': '0', 'unit': '0', ' bus': '0', ' type': 'drive'}, 'volumeInfo': {'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease', 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}, 'index': '1', 'iface': 'scsi', 'apparentsize': '1073741824', 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078', 'readonly': 'false', 'shared': 'false', 'truesize': '0', 'type': 'disk', 'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0', 'format': 'raw', 'deviceId': 'e1886286-3d81-47d0-ae8d-77454e464078', 'poolID': '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk', 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'propagateErrors': 'off', 'optional': 'false', 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'specParams': {}, 'volumeChain': [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path', 'leaseOffset': 0, 'path': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'leasePath': '/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease', 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}]}, {'device': 'ich6', 'specParams': {}, 'type': 'sound', 'deviceId': 'a1e596e9-218f-46ba-9f32-b9c966e11d73', 'address': {'bus': '0x00', ' slot': '0x04', ' domain': '0x0000', ' type': 'pci', ' function': '0x0'}}, {'device': 'memballoon', 'specParams': {'model': 'virtio'}, 'type': 'balloon', 'deviceId': '5c04fd7e-7249-4e3a-b8eb-91cce72d5b60', 'target': 4194304}], 'smp': '3', 'vmType': 'kvm', 'memSize': 4096, 'displayIp': '0', 'spiceSecureChannels': 'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard', 'smpCoresPerSocket': '1', 'vmName': 'cobra', 'display': 'qxl', 'nice': '0'} createXmlElem:<bound method Drive.createXmlElem of <vm.Drive object at 0x7f1150113a90>> device:cdrom deviceId:5611019a-948e-41b3-8ffd-75790ebcdf84 extSharedState:none getLeasesXML:<bound method Drive.getLeasesXML of <vm.Drive object at 0x7f1150113a90>> getNextVolumeSize:<bound method Drive.getNextVolumeSize of <vm.Drive object at 0x7f1150113a90>> getXML:<bound method Drive.getXML of <vm.Drive object at 0x7f1150113a90>> hasVolumeLeases:False iface:ide index:2 isDiskReplicationInProgress:<bound method Drive.isDiskReplicationInProgress of <vm.Drive object at 0x7f1150113a90>> isVdsmImage:<bound method Drive.isVdsmImage of <vm.Drive object at 0x7f1150113a90>> log:<logUtils.SimpleLogAdapter object at 0x7f111838af90> name:hdc networkDev:False path: readonly:true reqsize:0 serial: shared:false specParams:{'path': ''} truesize:0 type:disk volExtensionChunk:1024 watermarkLimit:536870912 Traceback (most recent call last): File "/usr/share/vdsm/clientIF.py", line 356, in teardownVolumePath res = self.irs.teardownImage(drive['domainID'], File "/usr/share/vdsm/vm.py", line 1389, in __getitem__ raise KeyError(key) KeyError: 'domainID' Thread-32158::DEBUG::2014-01-08 11:54:42,186::task::579::TaskManager.Task::(_updateState) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state init -> state preparing Thread-32158::INFO::2014-01-08 11:54:42,187::logUtils::44::dispatcher::(wrapper) Run and protect: teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a', spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', imgUUID='94a298cb-87a5-48cd-b78f-770582b50639', volUUID=None) Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::197::ResourceManager.Request::(__init__) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage' Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::541::ResourceManager::(registerResource) Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' for lock type 'shared' Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::600::ResourceManager::(registerResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking as 'shared' (1 active user) Thread-32158::DEBUG::2014-01-08 11:54:42,187::resourceManager::237::ResourceManager.Request::(grant) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Granted request Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::811::TaskManager.Task::(resourceAcquired) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::_resourcesAcquired: Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared) Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::974::TaskManager.Task::(_decref) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 1 aborting False Thread-32158::INFO::2014-01-08 11:54:42,188::logUtils::47::dispatcher::(wrapper) Run and protect: teardownImage, Return response: None Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::1168::TaskManager.Task::(prepare) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::finished: None Thread-32158::DEBUG::2014-01-08 11:54:42,188::task::579::TaskManager.Task::(_updateState) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state preparing -> state finished Thread-32158::DEBUG::2014-01-08 11:54:42,188::resourceManager::939::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj: 'None'>} Thread-32158::DEBUG::2014-01-08 11:54:42,188::resourceManager::976::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::615::ResourceManager::(releaseResource) Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::634::ResourceManager::(releaseResource) Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active users) Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::640::ResourceManager::(releaseResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out if anyone is waiting for it. Thread-32158::DEBUG::2014-01-08 11:54:42,189::resourceManager::648::ResourceManager::(releaseResource) No one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', Clearing records. Thread-32158::DEBUG::2014-01-08 11:54:42,189::task::974::TaskManager.Task::(_decref) Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 0 aborting False Thread-32158::DEBUG::2014-01-08 11:54:42,189::task::579::TaskManager.Task::(_updateState) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state init -> state preparing Thread-32158::INFO::2014-01-08 11:54:42,190::logUtils::44::dispatcher::(wrapper) Run and protect: teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a', spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', imgUUID='8df43d38-c4c7-4711-bc87-55f35d1550e5', volUUID=None) Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::197::ResourceManager.Request::(__init__) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage' Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::541::ResourceManager::(registerResource) Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' for lock type 'shared' Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::600::ResourceManager::(registerResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking as 'shared' (1 active user) Thread-32158::DEBUG::2014-01-08 11:54:42,190::resourceManager::237::ResourceManager.Request::(grant) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Granted request Thread-32158::DEBUG::2014-01-08 11:54:42,190::task::811::TaskManager.Task::(resourceAcquired) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::_resourcesAcquired: Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared) Thread-32158::DEBUG::2014-01-08 11:54:42,191::task::974::TaskManager.Task::(_decref) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 1 aborting False Thread-32158::INFO::2014-01-08 11:54:42,191::logUtils::47::dispatcher::(wrapper) Run and protect: teardownImage, Return response: None Thread-32158::DEBUG::2014-01-08 11:54:42,191::task::1168::TaskManager.Task::(prepare) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::finished: None Thread-32158::DEBUG::2014-01-08 11:54:42,191::task::579::TaskManager.Task::(_updateState) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state preparing -> state finished Thread-32158::DEBUG::2014-01-08 11:54:42,191::resourceManager::939::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj: 'None'>} Thread-32158::DEBUG::2014-01-08 11:54:42,191::resourceManager::976::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-32158::DEBUG::2014-01-08 11:54:42,191::resourceManager::615::ResourceManager::(releaseResource) Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' Thread-32158::DEBUG::2014-01-08 11:54:42,192::resourceManager::634::ResourceManager::(releaseResource) Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active users) Thread-32158::DEBUG::2014-01-08 11:54:42,192::resourceManager::640::ResourceManager::(releaseResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out if anyone is waiting for it. Thread-32158::DEBUG::2014-01-08 11:54:42,192::resourceManager::648::ResourceManager::(releaseResource) No one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', Clearing records. Thread-32158::DEBUG::2014-01-08 11:54:42,192::task::974::TaskManager.Task::(_decref) Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 0 aborting False Thread-32158::DEBUG::2014-01-08 11:54:42,192::task::579::TaskManager.Task::(_updateState) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state init -> state preparing Thread-32158::INFO::2014-01-08 11:54:42,192::logUtils::44::dispatcher::(wrapper) Run and protect: teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a', spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', imgUUID='e1886286-3d81-47d0-ae8d-77454e464078', volUUID=None) Thread-32158::DEBUG::2014-01-08 11:54:42,193::resourceManager::197::ResourceManager.Request::(__init__) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at 'teardownImage' Thread-32158::DEBUG::2014-01-08 11:54:42,193::resourceManager::541::ResourceManager::(registerResource) Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' for lock type 'shared' Thread-32158::DEBUG::2014-01-08 11:54:42,196::resourceManager::600::ResourceManager::(registerResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking as 'shared' (1 active user) Thread-32158::DEBUG::2014-01-08 11:54:42,196::resourceManager::237::ResourceManager.Request::(grant) ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Granted request Thread-32158::DEBUG::2014-01-08 11:54:42,196::task::811::TaskManager.Task::(resourceAcquired) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::_resourcesAcquired: Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared) Thread-32158::DEBUG::2014-01-08 11:54:42,196::task::974::TaskManager.Task::(_decref) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::ref 1 aborting False Thread-32158::INFO::2014-01-08 11:54:42,196::logUtils::47::dispatcher::(wrapper) Run and protect: teardownImage, Return response: None Thread-32158::DEBUG::2014-01-08 11:54:42,196::task::1168::TaskManager.Task::(prepare) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::finished: None Thread-32158::DEBUG::2014-01-08 11:54:42,197::task::579::TaskManager.Task::(_updateState) Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state preparing -> state finished Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::939::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj: 'None'>} Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::976::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::615::ResourceManager::(releaseResource) Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::634::ResourceManager::(releaseResource) Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active users) Thread-32158::DEBUG::2014-01-08 11:54:42,197::resourceManager::640::ResourceManager::(releaseResource) Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out if anyone is waiting for it.
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users