Hi Daniel,
Both times were on the same hypervisor which was a fresh 3.3.2 install,
not an upgrade. One time was using disk images and the other time was
using direct LUN.
I will send log files to you directly.
On 1/8/2014 3:15 PM, Daniel Erez wrote:
Hi Blaster,
Have you added the second disk after upgrading oVirt version?
An explicit address setting mechanism has been introduced recently,
which might cause such problems between minor versions.
Can you please attach the full engine/vdsm logs?
Thanks,
Daniel
----- Original Message -----
> From: "Blaster" <blaster(a)556nato.com>
> To: users(a)ovirt.org
> Sent: Wednesday, January 8, 2014 8:53:57 PM
> Subject: [Users] virt-io SCSI duplicate disk ID
>
> So twice now under oVirt 3.3.2 I have added 2 virtio-scsi devices to a single
> virtual host.
>
> After doing so, the VM would fail to boot due to duplicate disk ID. The first
> time I thought it a fluke, second time’s a bug?
>
> Fortunately they were empty data disks and I was able to get around the
> problem by deleting one and recreating it.
>
> VDSM log:
>
> Thread-32154::INFO::2014-01-08
> 11:54:39,717::clientIF::350::vds::(prepareVolumePath) prepared volume path:
>
/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2
> Thread-32154::DEBUG::2014-01-08 11:54:39,740::vm::2984::vm.Vm::(_run)
> vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::<?xml version="1.0"
> encoding="utf-8"?>
> <domain type="kvm">
> <name>cobra</name>
> <uuid>c2aff4cc-0de6-4342-a565-669b1825838c</uuid>
> <memory>4194304</memory>
> <currentMemory>4194304</currentMemory>
> <vcpu>3</vcpu>
> <memtune>
> <min_guarantee>4194304</min_guarantee>
> </memtune>
> <devices>
> <channel type="unix">
> <target name="com.redhat.rhevm.vdsm" type="virtio"/>
> <source mode="bind"
>
path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.com.redhat.rhevm.vdsm"/>
> </channel>
> <channel type="unix">
> <target name="org.qemu.guest_agent.0" type="virtio"/>
> <source mode="bind"
>
path="/var/lib/libvirt/qemu/channels/c2aff4cc-0de6-4342-a565-669b1825838c.org.qemu.guest_agent.0"/>
> </channel>
> <input bus="ps2" type="mouse"/>
> <channel type="spicevmc">
> <target name="com.redhat.spice.0" type="virtio"/>
> </channel>
> <graphics autoport="yes" keymap="en-us" listen="0"
passwd="*****"
> passwdValidTo="1970-01-01T00:00:01" port="-1"
tlsPort="-1" type="spice">
> <channel mode="secure" name="main"/>
> <channel mode="secure" name="inputs"/>
> <channel mode="secure" name="cursor"/>
> <channel mode="secure" name="playback"/>
> <channel mode="secure" name="record"/>
> <channel mode="secure" name="display"/>
> <channel mode="secure" name="usbredir"/>
>
> <channel mode="secure" name="display"/>
> <channel mode="secure" name="usbredir"/>
> <channel mode="secure" name="smartcard"/>
> </graphics>
> <controller model="virtio-scsi" type="scsi">
> <address bus="0x00" domain="0x0000" function="0x0"
slot="0x05" type="pci"/>
> </controller>
> <video>
> <address bus="0x00" domain="0x0000" function="0x0"
slot="0x02" type="pci"/>
> <model heads="1" type="qxl" vram="32768"/>
> </video>
> <interface type="bridge">
> <address bus="0x00" domain="0x0000" function="0x0"
slot="0x03" type="pci"/>
> <mac address="00:1a:4a:5b:9f:02"/>
> <model type="virtio"/>
> <source bridge="ovirtmgmt"/>
> <filterref filter="vdsm-no-mac-spoofing"/>
> <link state="up"/>
> </interface>
> <disk device="cdrom" snapshot="no" type="file">
> <address bus="1" controller="0" target="0"
type="drive" unit="0"/>
> <source file="" startupPolicy="optional"/>
> <target bus="ide" dev="hdc"/>
> <readonly/>
> <serial/>
> </disk>
> <disk device="disk" snapshot="no" type="file">
> <source
>
file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a"/>
> <target bus="scsi" dev="sda"/>
> <serial>94a298cb-87a5-48cd-b78f-770582b50639</serial>
> <boot order="1"/>
> <driver cache="none" error_policy="stop"
io="threads" name="qemu"
> type="raw"/>
> </disk>
> <disk device="disk" snapshot="no" type="file">
> <address bus="0x00" domain="0x0000" function="0x0"
slot="0x07" type="pci"/>
> <source
>
file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d"/>
> <target bus="virtio" dev="vda"/>
> <serial>8df43d38-c4c7-4711-bc87-55f35d1550e5</serial>
> <driver cache="none" error_policy="stop"
io="threads" name="qemu"
> type="raw"/>
> </disk>
> <disk device="disk" snapshot="no" type="file">
>
> <address bus="0" controller="0" target="0"
type="drive" unit="0"/>
> <source
>
file="/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2"/>
> <target bus="scsi" dev="sdb"/>
> <serial>e1886286-3d81-47d0-ae8d-77454e464078</serial>
> <driver cache="none" error_policy="stop"
io="threads" name="qemu"
> type="raw"/>
> </disk>
> <sound model="ich6">
> <address bus="0x00" domain="0x0000" function="0x0"
slot="0x04" type="pci"/>
> </sound>
> <memballoon model="virtio"/>
> </devices>
> <os>
> <type arch="x86_64" machine="pc-1.0">hvm</type>
> <smbios mode="sysinfo"/>
> </os>
> <sysinfo type="smbios">
> <system>
> <entry name="manufacturer">oVirt</entry>
> <entry name="product">oVirt Node</entry>
> <entry name="version">19-5</entry>
> <entry
name="serial">2061001F-C600-0006-E1BC-BCAEC518BA45</entry>
> <entry
name="uuid">c2aff4cc-0de6-4342-a565-669b1825838c</entry>
> </system>
> </sysinfo>
> <clock adjustment="-21600" offset="variable">
> <timer name="rtc" tickpolicy="catchup"/>
> </clock>
> <features>
> <acpi/>
> </features>
> <cpu match="exact">
> <model>Nehalem</model>
> <topology cores="1" sockets="3" threads="1"/>
> </cpu>
> </domain>
> Thread-32154::DEBUG::2014-01-08
> 11:54:40,218::libvirtconnection::108::libvirtconnection::(wrapper) Unknown
> libvirterror: ecode: 1 edom: 10 level: 2 message: internal error process
> exited while connecting to monitor: qemu-system-x86_64: -drive
>
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
> Duplicate ID 'drive-scsi0-0-0-0' for drive
>
> Thread-32154::DEBUG::2014-01-08
> 11:54:40,218::vm::2109::vm.Vm::(_startUnderlyingVm)
> vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::_ongoingCreations released
> Thread-32154::ERROR::2014-01-08
> 11:54:40,218::vm::2135::vm.Vm::(_startUnderlyingVm)
> vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::The vm start process failed
> Traceback (most recent call last):
> File "/usr/share/vdsm/vm.py", line 2095, in _startUnderlyingVm
> self._run()
> File "/usr/share/vdsm/vm.py", line 3018, in _run
> self._connection.createXML(domxml, flags),
> File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", line
76,
> in wrapper
> ret = f(*args, **kwargs)
> File "/usr/lib64/python2.7/site-packages/libvirt.py", line 2805, in
createXML
> if ret is None:raise libvirtError('virDomainCreateXML() failed', conn=self)
> libvirtError: internal error process exited while connecting to monitor:
> qemu-system-x86_64: -drive
>
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
> Duplicate ID 'drive-scsi0-0-0-0' for drive
>
> Thread-32154::DEBUG::2014-01-08
> 11:54:40,223::vm::2577::vm.Vm::(setDownStatus)
> vmId=`c2aff4cc-0de6-4342-a565-669b1825838c`::Changed state to Down: internal
> error process exited while connecting to monitor: qemu-system-x86_64: -drive
>
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
> Duplicate ID 'drive-scsi0-0-0-0' for drive
>
> Thread-32158::WARNING::2014-01-08
> 11:54:42,185::clientIF::362::vds::(teardownVolumePath) Drive is not a vdsm
> image: VOLWM_CHUNK_MB:1024 VOLWM_CHUNK_REPLICATE_MULT:2 VOLWM_FREE_PCT:50
> _blockDev:False _checkIoTuneCategories:<bound method
> Drive._checkIoTuneCategories of <vm.Drive object at 0x7f1150113a90>>
> _customize:<bound method Drive._customize of <vm.Drive object at
> 0x7f1150113a90>> _deviceXML:<disk device="cdrom"
snapshot="no"
> type="file"><address bus="1" controller="0"
target="0" type="drive"
> unit="0"/><source file=""
startupPolicy="optional"/><target bus="ide"
> dev="hdc"/><readonly/><serial></serial></disk>
_makeName:<bound method
> Drive._makeName of <vm.Drive object at 0x7f1150113a90>>
> _setExtSharedState:<bound method Drive._setExtSharedState of <vm.Drive
> object at 0x7f1150113a90>> _validateIoTuneParams:<bound method
> Drive._validateIoTuneParams of <vm.Drive object at 0x7f1150113a90>>
> address:{' controller': '0', ' target': '0',
'unit': '0', ' bus': '1', '
> type': 'drive'} apparentsize:0 blockDev:False cache:none
conf:{'status':
> 'Down', 'acpiEnable': 'true', 'emulatedMachine':
'pc-1.0', 'vmId':
> 'c2aff4cc-0de6-4342-a565-669b1825838c', 'pid': '0',
'memGuaranteedSize':
> 4096, 'timeOffset': '-21600', 'keyboardLayout':
'en-us', 'displayPort':
> '-1', 'displaySecurePort': '-1',
'spiceSslCipherSuite': 'DEFAULT',
> 'cpuType': 'Nehalem', 'custom':
>
{'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8device_a2e6354f-4ad6-475f-bd18-754fcedf505f':
> 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
> deviceId=a2e6354f-4ad6-475f-bd18-754fcedf505f, device=unix, type=CHANNEL,
> bootOrder=0, specParams={}, address={port=2, bus=0, controller=0,
> type=virtio-serial}, managed=false, plugged=true, readOnly=false,
> deviceAlias=channel1, customProperties={}, snapshotId=null}',
> 'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40': 'VmDevice
> {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
> deviceId=142f948d-f916-4f42-bd28-cb4f0b8ebb40, device=virtio-serial,
> type=CONTROLLER, bootOrder=0, specParams={}, address={bus=0x00,
> domain=0x0000, type=pci, slot=0x06, function=0x0}, managed=false,
> plugged=true, readOnly=false, deviceAlias=virtio-serial0,
> customProperties={}, snapshotId=null}',
>
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5a':
> 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
> deviceId=861eb290-19bc-4ace-b2cb-85cbb2e0eb5a, device=ide, type=CONTROLLER,
> bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci,
> slot=0x01, function=0x1}, managed=false, plugged=true, readOnly=false,
> deviceAlias=ide0, customProperties={}, snapshotId=null}',
>
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8device_861eb290-19bc-4ace-b2cb-85cbb2e0eb5adevice_94db7fa0-071d-4181-bac6-826ecfca3dd8':
> 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
> deviceId=94db7fa0-071d-4181-bac6-826ecfca3dd8, device=unix, type=CHANNEL,
> bootOrder=0, specParams={}, address={port=1, bus=0, controller=0,
> type=virtio-serial}, managed=false, plugged=true, readOnly=false,
> deviceAlias=channel0, customProperties={}, snapshotId=null}',
>
'device_142f948d-f916-4f42-bd28-cb4f0b8ebb40device_615c1466-850e-4362-a4fb-60df1aaee1e8':
> 'VmDevice {vmId=c2aff4cc-0de6-4342-a565-669b1825838c,
> deviceId=615c1466-850e-4362-a4fb-60df1aaee1e8, device=spicevmc,
> type=CHANNEL, bootOrder=0, specParams={}, address={port=3, bus=0,
> controller=0, type=virtio-serial}, managed=false, plugged=true,
> readOnly=false, deviceAlias=channel2, customProperties={},
> snapshotId=null}'}, 'clientIp': '', 'exitCode': 1,
'nicModel': 'rtl8139,pv',
> 'smartcardEnable': 'false', 'kvmEnable': 'true',
'exitMessage': "internal
> error process exited while connecting to monitor: qemu-system-x86_64: -drive
>
file=/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2,if=none,id=drive-scsi0-0-0-0,format=raw,serial=e1886286-3d81-47d0-ae8d-77454e464078,cache=none,werror=stop,rerror=stop,aio=threads:
> Duplicate ID 'drive-scsi0-0-0-0' for drive\n",
'transparentHugePages':
> 'true', 'devices': [{'specParams': {}, 'deviceId':
> 'db6166cb-e977-485e-8c82-fa48ca75e709', 'address': {'bus':
'0x00', ' slot':
> '0x05', ' domain': '0x0000', ' type': 'pci',
' function': '0x0'}, 'device':
> 'scsi', 'model': 'virtio-scsi', 'type':
'controller'}, {'device': 'qxl',
> 'specParams': {'vram': '32768', 'heads':
'1'}, 'type': 'video', 'deviceId':
> '8b0e3dbc-27c6-4eae-ba6b-201c3e1736aa', 'address': {'bus':
'0x00', ' slot':
> '0x02', ' domain': '0x0000', ' type': 'pci',
' function': '0x0'}},
> {'nicModel': 'pv', 'macAddr': '00:1a:4a:5b:9f:02',
'linkActive': 'true',
> 'network': 'ovirtmgmt', 'filter':
'vdsm-no-mac-spoofing', 'specParams': {},
> 'deviceId': '738c8ebe-b014-4d65-8c78-942aaf12bfb5',
'address': {'bus':
> '0x00', ' slot': '0x03', ' domain': '0x0000',
' type': 'pci', ' function':
> '0x0'}, 'device': 'bridge', 'type':
'interface'}, {'index': '2', 'iface':
> 'ide', 'address': {' controller': '0', '
target': '0', 'unit': '0', ' bus':
> '1', ' type': 'drive'}, 'specParams':
{'path': ''}, 'readonly': 'true',
> 'deviceId': '5611019a-948e-41b3-8ffd-75790ebcdf84', 'path':
'', 'device':
> 'cdrom', 'shared': 'false', 'type': 'disk'},
{'volumeInfo': {'domainID':
> 'f14f471e-0cce-414d-af57-779eeb88c97a', 'volType': 'path',
'leaseOffset': 0,
> 'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a',
> 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a',
'leasePath':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease',
> 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}, 'index':
0, 'iface':
> 'scsi', 'apparentsize': '162135015424', 'imageID':
> '94a298cb-87a5-48cd-b78f-770582b50639', 'readonly': 'false',
'shared':
> 'false', 'truesize': '107119386624', 'type':
'disk', 'domainID':
> 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0',
'format': 'raw',
> 'deviceId': '94a298cb-87a5-48cd-b78f-770582b50639',
'poolID':
> '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk',
'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a',
> 'propagateErrors': 'off', 'optional': 'false',
'bootOrder': '1', 'volumeID':
> 'c4864816-7384-4d0e-927e-29ea36d5fa4a', 'specParams': {},
'volumeChain':
> [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a',
'volType': 'path',
> 'leaseOffset': 0, 'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a',
> 'volumeID': 'c4864816-7384-4d0e-927e-29ea36d5fa4a',
'leasePath':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/94a298cb-87a5-48cd-b78f-770582b50639/c4864816-7384-4d0e-927e-29ea36d5fa4a.lease',
> 'imageID': '94a298cb-87a5-48cd-b78f-770582b50639'}]},
{'address': {'bus':
> '0x00', ' slot': '0x07', ' domain': '0x0000',
' type': 'pci', ' function':
> '0x0'}, 'volumeInfo': {'domainID':
'f14f471e-0cce-414d-af57-779eeb88c97a',
> 'volType': 'path', 'leaseOffset': 0, 'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d',
> 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d',
'leasePath':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease',
> 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}, 'index':
'0', 'iface':
> 'virtio', 'apparentsize': '1073741824', 'imageID':
> '8df43d38-c4c7-4711-bc87-55f35d1550e5', 'readonly': 'false',
'shared':
> 'false', 'truesize': '0', 'type': 'disk',
'domainID':
> 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0',
'format': 'raw',
> 'deviceId': '8df43d38-c4c7-4711-bc87-55f35d1550e5',
'poolID':
> '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk',
'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d',
> 'propagateErrors': 'off', 'optional': 'false',
'volumeID':
> '42786233-04c0-4a9f-b731-6eaa2ca4d40d', 'specParams': {},
'volumeChain':
> [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a',
'volType': 'path',
> 'leaseOffset': 0, 'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d',
> 'volumeID': '42786233-04c0-4a9f-b731-6eaa2ca4d40d',
'leasePath':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/8df43d38-c4c7-4711-bc87-55f35d1550e5/42786233-04c0-4a9f-b731-6eaa2ca4d40d.lease',
> 'imageID': '8df43d38-c4c7-4711-bc87-55f35d1550e5'}]},
{'address': {'
> controller': '0', ' target': '0', 'unit':
'0', ' bus': '0', ' type':
> 'drive'}, 'volumeInfo': {'domainID':
'f14f471e-0cce-414d-af57-779eeb88c97a',
> 'volType': 'path', 'leaseOffset': 0, 'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
> 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
'leasePath':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease',
> 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}, 'index':
'1', 'iface':
> 'scsi', 'apparentsize': '1073741824', 'imageID':
> 'e1886286-3d81-47d0-ae8d-77454e464078', 'readonly': 'false',
'shared':
> 'false', 'truesize': '0', 'type': 'disk',
'domainID':
> 'f14f471e-0cce-414d-af57-779eeb88c97a', 'reqsize': '0',
'format': 'raw',
> 'deviceId': 'e1886286-3d81-47d0-ae8d-77454e464078',
'poolID':
> '18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9', 'device': 'disk',
'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
> 'propagateErrors': 'off', 'optional': 'false',
'volumeID':
> '1ef64ebd-6415-43d1-a34f-28f05cf6dea2', 'specParams': {},
'volumeChain':
> [{'domainID': 'f14f471e-0cce-414d-af57-779eeb88c97a',
'volType': 'path',
> 'leaseOffset': 0, 'path':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
> 'volumeID': '1ef64ebd-6415-43d1-a34f-28f05cf6dea2',
'leasePath':
>
'/rhev/data-center/mnt/_disk01_vm/f14f471e-0cce-414d-af57-779eeb88c97a/images/e1886286-3d81-47d0-ae8d-77454e464078/1ef64ebd-6415-43d1-a34f-28f05cf6dea2.lease',
> 'imageID': 'e1886286-3d81-47d0-ae8d-77454e464078'}]},
{'device': 'ich6',
> 'specParams': {}, 'type': 'sound', 'deviceId':
> 'a1e596e9-218f-46ba-9f32-b9c966e11d73', 'address': {'bus':
'0x00', ' slot':
> '0x04', ' domain': '0x0000', ' type': 'pci',
' function': '0x0'}},
> {'device': 'memballoon', 'specParams': {'model':
'virtio'}, 'type':
> 'balloon', 'deviceId':
'5c04fd7e-7249-4e3a-b8eb-91cce72d5b60', 'target':
> 4194304}], 'smp': '3', 'vmType': 'kvm',
'memSize': 4096, 'displayIp': '0',
> 'spiceSecureChannels':
> 'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard',
> 'smpCoresPerSocket': '1', 'vmName': 'cobra',
'display': 'qxl', 'nice': '0'}
> createXmlElem:<bound method Drive.createXmlElem of <vm.Drive object at
> 0x7f1150113a90>> device:cdrom deviceId:5611019a-948e-41b3-8ffd-75790ebcdf84
> extSharedState:none getLeasesXML:<bound method Drive.getLeasesXML of
> <vm.Drive object at 0x7f1150113a90>> getNextVolumeSize:<bound method
> Drive.getNextVolumeSize of <vm.Drive object at 0x7f1150113a90>>
> getXML:<bound method Drive.getXML of <vm.Drive object at
0x7f1150113a90>>
> hasVolumeLeases:False iface:ide index:2 isDiskReplicationInProgress:<bound
> method Drive.isDiskReplicationInProgress of <vm.Drive object at
> 0x7f1150113a90>> isVdsmImage:<bound method Drive.isVdsmImage of
<vm.Drive
> object at 0x7f1150113a90>> log:<logUtils.SimpleLogAdapter object at
> 0x7f111838af90> name:hdc networkDev:False path: readonly:true reqsize:0
> serial: shared:false specParams:{'path': ''} truesize:0 type:disk
> volExtensionChunk:1024 watermarkLimit:536870912
> Traceback (most recent call last):
> File "/usr/share/vdsm/clientIF.py", line 356, in teardownVolumePath
> res = self.irs.teardownImage(drive['domainID'],
> File "/usr/share/vdsm/vm.py", line 1389, in __getitem__
> raise KeyError(key)
> KeyError: 'domainID'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,186::task::579::TaskManager.Task::(_updateState)
> Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state init -> state
> preparing
> Thread-32158::INFO::2014-01-08
> 11:54:42,187::logUtils::44::dispatcher::(wrapper) Run and protect:
> teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a',
> spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9',
> imgUUID='94a298cb-87a5-48cd-b78f-770582b50639', volUUID=None)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,187::resourceManager::197::ResourceManager.Request::(__init__)
>
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Request
> was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at
'teardownImage'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,187::resourceManager::541::ResourceManager::(registerResource)
> Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
> for lock type 'shared'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,187::resourceManager::600::ResourceManager::(registerResource)
> Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking
> as 'shared' (1 active user)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,187::resourceManager::237::ResourceManager.Request::(grant)
>
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`533f2699-0684-4247-9d5f-a858ffe96fe9`::Granted
> request
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,188::task::811::TaskManager.Task::(resourceAcquired)
> Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::_resourcesAcquired:
> Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,188::task::974::TaskManager.Task::(_decref)
> Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 1 aborting False
> Thread-32158::INFO::2014-01-08
> 11:54:42,188::logUtils::47::dispatcher::(wrapper) Run and protect:
> teardownImage, Return response: None
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,188::task::1168::TaskManager.Task::(prepare)
> Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::finished: None
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,188::task::579::TaskManager.Task::(_updateState)
> Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::moving from state preparing ->
> state finished
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,188::resourceManager::939::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources
> {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef
> 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj:
> 'None'>}
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,188::resourceManager::976::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,189::resourceManager::615::ResourceManager::(releaseResource)
> Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,189::resourceManager::634::ResourceManager::(releaseResource)
> Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active
> users)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,189::resourceManager::640::ResourceManager::(releaseResource)
> Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out
> if anyone is waiting for it.
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,189::resourceManager::648::ResourceManager::(releaseResource) No
> one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a',
> Clearing records.
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,189::task::974::TaskManager.Task::(_decref)
> Task=`30d76cca-4645-4893-8d68-5cc68ba42dc3`::ref 0 aborting False
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,189::task::579::TaskManager.Task::(_updateState)
> Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state init -> state
> preparing
> Thread-32158::INFO::2014-01-08
> 11:54:42,190::logUtils::44::dispatcher::(wrapper) Run and protect:
> teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a',
> spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9',
> imgUUID='8df43d38-c4c7-4711-bc87-55f35d1550e5', volUUID=None)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,190::resourceManager::197::ResourceManager.Request::(__init__)
>
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Request
> was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at
'teardownImage'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,190::resourceManager::541::ResourceManager::(registerResource)
> Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
> for lock type 'shared'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,190::resourceManager::600::ResourceManager::(registerResource)
> Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking
> as 'shared' (1 active user)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,190::resourceManager::237::ResourceManager.Request::(grant)
>
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`0907e117-ba60-4c9f-a190-ebb7e027b4c2`::Granted
> request
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,190::task::811::TaskManager.Task::(resourceAcquired)
> Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::_resourcesAcquired:
> Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,191::task::974::TaskManager.Task::(_decref)
> Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 1 aborting False
> Thread-32158::INFO::2014-01-08
> 11:54:42,191::logUtils::47::dispatcher::(wrapper) Run and protect:
> teardownImage, Return response: None
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,191::task::1168::TaskManager.Task::(prepare)
> Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::finished: None
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,191::task::579::TaskManager.Task::(_updateState)
> Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::moving from state preparing ->
> state finished
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,191::resourceManager::939::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources
> {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef
> 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj:
> 'None'>}
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,191::resourceManager::976::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,191::resourceManager::615::ResourceManager::(releaseResource)
> Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,192::resourceManager::634::ResourceManager::(releaseResource)
> Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active
> users)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,192::resourceManager::640::ResourceManager::(releaseResource)
> Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out
> if anyone is waiting for it.
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,192::resourceManager::648::ResourceManager::(releaseResource) No
> one is waiting for resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a',
> Clearing records.
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,192::task::974::TaskManager.Task::(_decref)
> Task=`4b2bde05-78f2-42b2-a5be-171377c6905e`::ref 0 aborting False
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,192::task::579::TaskManager.Task::(_updateState)
> Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state init -> state
> preparing
> Thread-32158::INFO::2014-01-08
> 11:54:42,192::logUtils::44::dispatcher::(wrapper) Run and protect:
> teardownImage(sdUUID='f14f471e-0cce-414d-af57-779eeb88c97a',
> spUUID='18f6234c-a9de-4fdf-bd9a-2bd90b9f33f9',
> imgUUID='e1886286-3d81-47d0-ae8d-77454e464078', volUUID=None)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,193::resourceManager::197::ResourceManager.Request::(__init__)
>
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Request
> was made in '/usr/share/vdsm/storage/hsm.py' line '3283' at
'teardownImage'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,193::resourceManager::541::ResourceManager::(registerResource)
> Trying to register resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
> for lock type 'shared'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,196::resourceManager::600::ResourceManager::(registerResource)
> Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free. Now locking
> as 'shared' (1 active user)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,196::resourceManager::237::ResourceManager.Request::(grant)
>
ResName=`Storage.f14f471e-0cce-414d-af57-779eeb88c97a`ReqID=`8908ad39-2555-4ef5-a6ae-49e8504de015`::Granted
> request
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,196::task::811::TaskManager.Task::(resourceAcquired)
> Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::_resourcesAcquired:
> Storage.f14f471e-0cce-414d-af57-779eeb88c97a (shared)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,196::task::974::TaskManager.Task::(_decref)
> Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::ref 1 aborting False
> Thread-32158::INFO::2014-01-08
> 11:54:42,196::logUtils::47::dispatcher::(wrapper) Run and protect:
> teardownImage, Return response: None
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,196::task::1168::TaskManager.Task::(prepare)
> Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::finished: None
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,197::task::579::TaskManager.Task::(_updateState)
> Task=`75d1051f-b118-4af8-b9f1-504fcd1802c2`::moving from state preparing ->
> state finished
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,197::resourceManager::939::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources
> {'Storage.f14f471e-0cce-414d-af57-779eeb88c97a': < ResourceRef
> 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a', isValid: 'True' obj:
> 'None'>}
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,197::resourceManager::976::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,197::resourceManager::615::ResourceManager::(releaseResource)
> Trying to release resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a'
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,197::resourceManager::634::ResourceManager::(releaseResource)
> Released resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' (0 active
> users)
> Thread-32158::DEBUG::2014-01-08
> 11:54:42,197::resourceManager::640::ResourceManager::(releaseResource)
> Resource 'Storage.f14f471e-0cce-414d-af57-779eeb88c97a' is free, finding out
> if anyone is waiting for it.
>
>
>
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>