Hi everyone
I've set up a gluster storage with two replicated bricks
DC is up and I created a VM to test gluster storage
If I start the VM WITHOUT any disk attached (only one virtual DVD)
it starts fine.
If I attach a gluster domain disk thin provisioning 30 Gb the Vm
stucks in "waiting for launch" state
I see no special activity on the gluster servers (they serve
several other shares with no troubles at all and even the ISO
domain is a NFS on locally mounted gluster and works fine)
I've double checked all the pre requisites and they look fine (F
19 - gluster setup insecure in both glusterd.vol and volume
options - uid/gid/insecure )
Am I doing something wrong?
I'm even unable to stop the VM from the engine GUI
Any advise?
Thank you
Best regards
CentOS_30.log
2013-10-30 10:17:48.094+0000: starting up
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name CentOS_30 -S -machine
pc-1.0,accel=kvm,usb=off -cpu Westmere -m 2048 -smp
1,sockets=1,cores=1,threads=1 -uuid
200dfb05-461e-49d9-95a2-c0a7c7ced669 -smbios
type=1,manufacturer=oVirt,product=oVirt
Node,version=19-4,serial=316140A0-D7DA-11DD-AF75-08606ED667F7,uuid=200dfb05-461e-49d9-95a2-c0a7c7ced669
-no-user-config -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/CentOS_30.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2013-10-30T10:17:47,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x5 -drive
file=/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/images/11111111-1111-1111-1111-111111111111/CentOS-6.4-x86_64-bin-DVD1.iso,if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial=
-device
ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=1
-drive
file=gluster://172.16.0.100/vms/21e3d67f-fdec-4f87-b09e-53af994a6f87/images/b73a79fe-401b-434c-b023-6de4c3eb100d/202adc86-7857-40eb-8ca6-7cec3ef0d2d6,if=none,id=drive-virtio-disk0,format=raw,serial=b73a79fe-401b-434c-b023-6de4c3eb100d,cache=none,werror=stop,rerror=stop,aio=threads
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=2
-netdev tap,fd=27,id=hostnet0,vhost=on,vhostfd=28 -device
virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:00:00:33,bus=pci.0,addr=0x3,bootindex=3
-chardev
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/200dfb05-461e-49d9-95a2-c0a7c7ced669.com.redhat.rhevm.vdsm,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev
socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/200dfb05-461e-49d9-95a2-c0a7c7ced669.org.qemu.guest_agent.0,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
-chardev spicevmc,id=charchannel2,name=vdagent -device
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel2,name=com.redhat.spice.0
-spice
tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbredir,seamless-migration=on
-k en-us -vga qxl -global qxl-vga.ram_size=67108864 -global
qxl-vga.vram_size=67108864 -device
virtio-balloon-pci,id=balloon0,bus=pci
vdsm.log
Thread-1239::DEBUG::2013-10-30
11:25:03,420::task::974::TaskManager.Task::(_decref)
Task=`d891f3d3-070d-403f-9050-f10651a7fc47`::ref 0 aborting False
Thread-242::DEBUG::2013-10-30
11:25:04,982::domainMonitor::178::Storage.DomainMonitorThread::(_monitorDomain)
Refreshing domain 21e3d67f-fdec-4f87-b09e-53af994a6f87
Thread-242::DEBUG::2013-10-30
11:25:04,992::fileSD::153::Storage.StorageDomain::(__init__)
Reading domain in path
/rhev/data-center/mnt/glusterSD/172.16.0.100:vms/21e3d67f-fdec-4f87-b09e-53af994a6f87
Thread-242::DEBUG::2013-10-30
11:25:04,994::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with FileMetadataRW backend
Thread-242::DEBUG::2013-10-30
11:25:05,005::persistentDict::234::Storage.PersistentDict::(refresh)
read lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=vms',
'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60',
'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'MASTER_VERSION=1',
'POOL_DESCRIPTION=GlusterSD',
'POOL_DOMAINS=21e3d67f-fdec-4f87-b09e-53af994a6f87:Active,8e56a914-3395-48ad-ab4c-c2ca7f9956b4:Active',
'POOL_SPM_ID=1', 'POOL_SPM_LVER=3',
'POOL_UUID=e3cb7694-47f7-4eac-9201-922e381a25dc',
'REMOTE_PATH=172.16.0.100:vms', 'ROLE=Master',
'SDUUID=21e3d67f-fdec-4f87-b09e-53af994a6f87', 'TYPE=GLUSTERFS',
'VERSION=3',
'_SHA_CKSUM=2d0dc3836063d45dd1d8046aaf00bce05c2455ea']
Thread-242::DEBUG::2013-10-30
11:25:05,008::fileSD::535::Storage.StorageDomain::(imageGarbageCollector)
Removing remnants of deleted images []
Thread-242::
INFO::2013-10-30
11:25:05,008::sd::374::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 21e3d67f-fdec-4f87-b09e-53af994a6f87_imageNS
already registered
Thread-242::
INFO::2013-10-30
11:25:05,008::sd::382::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 21e3d67f-fdec-4f87-b09e-53af994a6f87_volumeNS
already registered
Thread-242::DEBUG::2013-10-30
11:25:05,020::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/glusterSD/172.16.0.100:vms/21e3d67f-fdec-4f87-b09e-53af994a6f87/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-242::DEBUG::2013-10-30
11:25:05,026::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n497 bytes
(497 B) copied, 0.000280701 s, 1.8 MB/s\n'; <rc> = 0
Thread-243::DEBUG::2013-10-30
11:25:07,721::domainMonitor::178::Storage.DomainMonitorThread::(_monitorDomain)
Refreshing domain 8e56a914-3395-48ad-ab4c-c2ca7f9956b4
Thread-243::DEBUG::2013-10-30
11:25:07,730::fileSD::153::Storage.StorageDomain::(__init__)
Reading domain in path
/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4
Thread-243::DEBUG::2013-10-30
11:25:07,731::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with FileMetadataRW backend
Thread-243::DEBUG::2013-10-30
11:25:07,740::persistentDict::234::Storage.PersistentDict::(refresh)
read lines (FileMetadataRW)=['CLASS=Iso', 'DESCRIPTION=ISO',
'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5',
'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'MASTER_VERSION=0',
'POOL_UUID=01b79582-75e3-436d-bade-db653207620b,e3cb7694-47f7-4eac-9201-922e381a25dc',
'REMOTE_PATH=172.16.0.5:/home/iso2', 'ROLE=Regular',
'SDUUID=8e56a914-3395-48ad-ab4c-c2ca7f9956b4', 'TYPE=NFS',
'VERSION=0',
'_SHA_CKSUM=b0d102325f279bb8efa68c594cbace0b11c66783']
Thread-243::DEBUG::2013-10-30
11:25:07,741::fileSD::535::Storage.StorageDomain::(imageGarbageCollector)
Removing remnants of deleted images []
Thread-243::
INFO::2013-10-30
11:25:07,741::sd::374::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 8e56a914-3395-48ad-ab4c-c2ca7f9956b4_imageNS
already registered
Thread-243::
INFO::2013-10-30
11:25:07,741::sd::382::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 8e56a914-3395-48ad-ab4c-c2ca7f9956b4_volumeNS
already registered
Thread-243::DEBUG::2013-10-30
11:25:07,757::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-243::DEBUG::2013-10-30
11:25:07,760::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n373 bytes
(373 B) copied, 0.000128829 s, 2.9 MB/s\n'; <rc> = 0
Thread-1246::DEBUG::2013-10-30
11:25:09,944::task::579::TaskManager.Task::(_updateState)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::moving from state
init -> state preparing
Thread-1246::
INFO::2013-10-30
11:25:09,944::logUtils::44::dispatcher::(wrapper) Run and protect:
repoStats(options=None)
Thread-1246::
INFO::2013-10-30
11:25:09,945::logUtils::47::dispatcher::(wrapper) Run and protect:
repoStats, Return response:
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'delay': '0.000280701',
'lastCheck': '4.9', 'code': 0, 'valid': True, 'version': 3},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'delay': '0.000128829',
'lastCheck': '2.2', 'code': 0, 'valid': True, 'version': 0}}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::task::1168::TaskManager.Task::(prepare)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::finished:
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'delay': '0.000280701',
'lastCheck': '4.9', 'code': 0, 'valid': True, 'version': 3},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'delay': '0.000128829',
'lastCheck': '2.2', 'code': 0, 'valid': True, 'version': 0}}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::task::579::TaskManager.Task::(_updateState)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::moving from state
preparing -> state finished
Thread-1246::DEBUG::2013-10-30
11:25:09,945::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::task::974::TaskManager.Task::(_decref)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::ref 0 aborting False
Thread-1249::DEBUG::2013-10-30
11:25:13,507::BindingXMLRPC::177::vds::(wrapper) client
[172.16.0.5]
Thread-1249::DEBUG::2013-10-30
11:25:13,507::task::579::TaskManager.Task::(_updateState)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::moving from state
init -> state preparing
Thread-1249::
INFO::2013-10-30
11:25:13,507::logUtils::44::dispatcher::(wrapper) Run and protect:
getSpmStatus(spUUID='e3cb7694-47f7-4eac-9201-922e381a25dc',
options=None)
Thread-1249::
INFO::2013-10-30
11:25:13,508::logUtils::47::dispatcher::(wrapper) Run and protect:
getSpmStatus, Return response: {'spm_st': {'spmId': 1,
'spmStatus': 'SPM', 'spmLver': 3}}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::task::1168::TaskManager.Task::(prepare)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::finished: {'spm_st':
{'spmId': 1, 'spmStatus': 'SPM', 'spmLver': 3}}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::task::579::TaskManager.Task::(_updateState)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::moving from state
preparing -> state finished
Thread-1249::DEBUG::2013-10-30
11:25:13,508::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::task::974::TaskManager.Task::(_decref)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::ref 0 aborting False
Thread-1250::DEBUG::2013-10-30
11:25:13,510::BindingXMLRPC::177::vds::(wrapper) client
[172.16.0.5]
Thread-1250::DEBUG::2013-10-30
11:25:13,510::task::579::TaskManager.Task::(_updateState)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::moving from state
init -> state preparing
Thread-1250::
INFO::2013-10-30
11:25:13,511::logUtils::44::dispatcher::(wrapper) Run and protect:
getStoragePoolInfo(spUUID='e3cb7694-47f7-4eac-9201-922e381a25dc',
options=None)
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::197::ResourceManager.Request::(__init__)
ResName=`Storage.e3cb7694-47f7-4eac-9201-922e381a25dc`ReqID=`42f323ef-aa03-49e6-893f-0e4ff0764b6b`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '2560' at
'getStoragePoolInfo'
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::541::ResourceManager::(registerResource)
Trying to register resource
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc' for lock type
'shared'
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::600::ResourceManager::(registerResource)
Resource 'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc' is free.
Now locking as 'shared' (1 active user)
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::237::ResourceManager.Request::(grant)
ResName=`Storage.e3cb7694-47f7-4eac-9201-922e381a25dc`ReqID=`42f323ef-aa03-49e6-893f-0e4ff0764b6b`::Granted
request
Thread-1250::DEBUG::2013-10-30
11:25:13,512::task::811::TaskManager.Task::(resourceAcquired)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::_resourcesAcquired:
Storage.e3cb7694-47f7-4eac-9201-922e381a25dc (shared)
Thread-1250::DEBUG::2013-10-30
11:25:13,512::task::974::TaskManager.Task::(_decref)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::ref 1 aborting False
Thread-1250::
INFO::2013-10-30
11:25:13,517::logUtils::47::dispatcher::(wrapper) Run and protect:
getStoragePoolInfo, Return response: {'info': {'spm_id': 1,
'master_uuid': '21e3d67f-fdec-4f87-b09e-53af994a6f87', 'name':
'GlusterSD', 'version': '3', 'domains':
'21e3d67f-fdec-4f87-b09e-53af994a6f87:Active,8e56a914-3395-48ad-ab4c-c2ca7f9956b4:Active',
'pool_status': 'connected', 'isoprefix':
'/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/images/11111111-1111-1111-1111-111111111111',
'type': 'GLUSTERFS', 'master_ver': 1, 'lver': 3}, 'dominfo':
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'status': 'Active',
'diskfree': '239969632256', 'alerts': [], 'version': 3,
'disktotal': '492120440832'},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'status': 'Active',
'diskfree': '201653747712', 'alerts': [], 'version': 0,
'disktotal': '246783410176'}}}
Thread-1250::DEBUG::2013-10-30
11:25:13,517::task::1168::TaskManager.Task::(prepare)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::finished: {'info':
{'spm_id': 1, 'master_uuid':
'21e3d67f-fdec-4f87-b09e-53af994a6f87', 'name': 'GlusterSD',
'version': '3', 'domains':
'21e3d67f-fdec-4f87-b09e-53af994a6f87:Active,8e56a914-3395-48ad-ab4c-c2ca7f9956b4:Active',
'pool_status': 'connected', 'isoprefix':
'/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/images/11111111-1111-1111-1111-111111111111',
'type': 'GLUSTERFS', 'master_ver': 1, 'lver': 3}, 'dominfo':
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'status': 'Active',
'diskfree': '239969632256', 'alerts': [], 'version': 3,
'disktotal': '492120440832'},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'status': 'Active',
'diskfree': '201653747712', 'alerts': [], 'version': 0,
'disktotal': '246783410176'}}}
Thread-1250::DEBUG::2013-10-30
11:25:13,517::task::579::TaskManager.Task::(_updateState)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::moving from state
preparing -> state finished
Thread-1250::DEBUG::2013-10-30
11:25:13,517::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources
{'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc': < ResourceRef
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc', isValid: 'True'
obj: 'None'>}
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc'
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc'
(0 active users)
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc' is free,
finding out if anyone is waiting for it.
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::648::ResourceManager::(releaseResource)
No one is waiting for resource
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc', Clearing records.
Thread-1250::DEBUG::2013-10-30
11:25:13,518::task::974::TaskManager.Task::(_decref)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::ref 0 aborting False
Thread-242::DEBUG::2013-10-30
11:25:15,046::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/glusterSD/172.16.0.100:vms/21e3d67f-fdec-4f87-b09e-53af994a6f87/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-242::DEBUG::2013-10-30
11:25:15,051::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n497 bytes
(497 B) copied, 0.00032077 s, 1.5 MB/s\n'; <rc> = 0
Thread-243::DEBUG::2013-10-30
11:25:17,777::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-243::DEBUG::2013-10-30
11:25:17,780::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n373 bytes
(373 B) copied, 9.9232e-05 s, 3.8 Thread-1239::DEBUG::2013-10-30
11:25:03,420::task::974::TaskManager.Task::(_decref)
Task=`d891f3d3-070d-403f-9050-f10651a7fc47`::ref 0 aborting False
Thread-242::DEBUG::2013-10-30
11:25:04,982::domainMonitor::178::Storage.DomainMonitorThread::(_monitorDomain)
Refreshing domain 21e3d67f-fdec-4f87-b09e-53af994a6f87
Thread-242::DEBUG::2013-10-30
11:25:04,992::fileSD::153::Storage.StorageDomain::(__init__)
Reading domain in path
/rhev/data-center/mnt/glusterSD/172.16.0.100:vms/21e3d67f-fdec-4f87-b09e-53af994a6f87
Thread-242::DEBUG::2013-10-30
11:25:04,994::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with FileMetadataRW backend
Thread-242::DEBUG::2013-10-30
11:25:05,005::persistentDict::234::Storage.PersistentDict::(refresh)
read lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=vms',
'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60',
'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'MASTER_VERSION=1',
'POOL_DESCRIPTION=GlusterSD',
'POOL_DOMAINS=21e3d67f-fdec-4f87-b09e-53af994a6f87:Active,8e56a914-3395-48ad-ab4c-c2ca7f9956b4:Active',
'POOL_SPM_ID=1', 'POOL_SPM_LVER=3',
'POOL_UUID=e3cb7694-47f7-4eac-9201-922e381a25dc',
'REMOTE_PATH=172.16.0.100:vms', 'ROLE=Master',
'SDUUID=21e3d67f-fdec-4f87-b09e-53af994a6f87', 'TYPE=GLUSTERFS',
'VERSION=3',
'_SHA_CKSUM=2d0dc3836063d45dd1d8046aaf00bce05c2455ea']
Thread-242::DEBUG::2013-10-30
11:25:05,008::fileSD::535::Storage.StorageDomain::(imageGarbageCollector)
Removing remnants of deleted images []
Thread-242::
INFO::2013-10-30
11:25:05,008::sd::374::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 21e3d67f-fdec-4f87-b09e-53af994a6f87_imageNS
already registered
Thread-242::
INFO::2013-10-30
11:25:05,008::sd::382::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 21e3d67f-fdec-4f87-b09e-53af994a6f87_volumeNS
already registered
Thread-242::DEBUG::2013-10-30
11:25:05,020::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/glusterSD/172.16.0.100:vms/21e3d67f-fdec-4f87-b09e-53af994a6f87/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-242::DEBUG::2013-10-30
11:25:05,026::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n497 bytes
(497 B) copied, 0.000280701 s, 1.8 MB/s\n'; <rc> = 0
Thread-243::DEBUG::2013-10-30
11:25:07,721::domainMonitor::178::Storage.DomainMonitorThread::(_monitorDomain)
Refreshing domain 8e56a914-3395-48ad-ab4c-c2ca7f9956b4
Thread-243::DEBUG::2013-10-30
11:25:07,730::fileSD::153::Storage.StorageDomain::(__init__)
Reading domain in path
/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4
Thread-243::DEBUG::2013-10-30
11:25:07,731::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with FileMetadataRW backend
Thread-243::DEBUG::2013-10-30
11:25:07,740::persistentDict::234::Storage.PersistentDict::(refresh)
read lines (FileMetadataRW)=['CLASS=Iso', 'DESCRIPTION=ISO',
'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5',
'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'MASTER_VERSION=0',
'POOL_UUID=01b79582-75e3-436d-bade-db653207620b,e3cb7694-47f7-4eac-9201-922e381a25dc',
'REMOTE_PATH=172.16.0.5:/home/iso2', 'ROLE=Regular',
'SDUUID=8e56a914-3395-48ad-ab4c-c2ca7f9956b4', 'TYPE=NFS',
'VERSION=0',
'_SHA_CKSUM=b0d102325f279bb8efa68c594cbace0b11c66783']
Thread-243::DEBUG::2013-10-30
11:25:07,741::fileSD::535::Storage.StorageDomain::(imageGarbageCollector)
Removing remnants of deleted images []
Thread-243::
INFO::2013-10-30
11:25:07,741::sd::374::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 8e56a914-3395-48ad-ab4c-c2ca7f9956b4_imageNS
already registered
Thread-243::
INFO::2013-10-30
11:25:07,741::sd::382::Storage.StorageDomain::(_registerResourceNamespaces)
Resource namespace 8e56a914-3395-48ad-ab4c-c2ca7f9956b4_volumeNS
already registered
Thread-243::DEBUG::2013-10-30
11:25:07,757::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-243::DEBUG::2013-10-30
11:25:07,760::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n373 bytes
(373 B) copied, 0.000128829 s, 2.9 MB/s\n'; <rc> = 0
Thread-1246::DEBUG::2013-10-30
11:25:09,944::task::579::TaskManager.Task::(_updateState)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::moving from state
init -> state preparing
Thread-1246::
INFO::2013-10-30
11:25:09,944::logUtils::44::dispatcher::(wrapper) Run and protect:
repoStats(options=None)
Thread-1246::
INFO::2013-10-30
11:25:09,945::logUtils::47::dispatcher::(wrapper) Run and protect:
repoStats, Return response:
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'delay': '0.000280701',
'lastCheck': '4.9', 'code': 0, 'valid': True, 'version': 3},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'delay': '0.000128829',
'lastCheck': '2.2', 'code': 0, 'valid': True, 'version': 0}}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::task::1168::TaskManager.Task::(prepare)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::finished:
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'delay': '0.000280701',
'lastCheck': '4.9', 'code': 0, 'valid': True, 'version': 3},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'delay': '0.000128829',
'lastCheck': '2.2', 'code': 0, 'valid': True, 'version': 0}}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::task::579::TaskManager.Task::(_updateState)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::moving from state
preparing -> state finished
Thread-1246::DEBUG::2013-10-30
11:25:09,945::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1246::DEBUG::2013-10-30
11:25:09,945::task::974::TaskManager.Task::(_decref)
Task=`c064522b-6fe7-4cb2-b85d-ad1f8185919a`::ref 0 aborting False
Thread-1249::DEBUG::2013-10-30
11:25:13,507::BindingXMLRPC::177::vds::(wrapper) client
[172.16.0.5]
Thread-1249::DEBUG::2013-10-30
11:25:13,507::task::579::TaskManager.Task::(_updateState)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::moving from state
init -> state preparing
Thread-1249::
INFO::2013-10-30
11:25:13,507::logUtils::44::dispatcher::(wrapper) Run and protect:
getSpmStatus(spUUID='e3cb7694-47f7-4eac-9201-922e381a25dc',
options=None)
Thread-1249::
INFO::2013-10-30
11:25:13,508::logUtils::47::dispatcher::(wrapper) Run and protect:
getSpmStatus, Return response: {'spm_st': {'spmId': 1,
'spmStatus': 'SPM', 'spmLver': 3}}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::task::1168::TaskManager.Task::(prepare)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::finished: {'spm_st':
{'spmId': 1, 'spmStatus': 'SPM', 'spmLver': 3}}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::task::579::TaskManager.Task::(_updateState)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::moving from state
preparing -> state finished
Thread-1249::DEBUG::2013-10-30
11:25:13,508::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1249::DEBUG::2013-10-30
11:25:13,508::task::974::TaskManager.Task::(_decref)
Task=`8518162d-eb91-44aa-819a-354e05807cb9`::ref 0 aborting False
Thread-1250::DEBUG::2013-10-30
11:25:13,510::BindingXMLRPC::177::vds::(wrapper) client
[172.16.0.5]
Thread-1250::DEBUG::2013-10-30
11:25:13,510::task::579::TaskManager.Task::(_updateState)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::moving from state
init -> state preparing
Thread-1250::
INFO::2013-10-30
11:25:13,511::logUtils::44::dispatcher::(wrapper) Run and protect:
getStoragePoolInfo(spUUID='e3cb7694-47f7-4eac-9201-922e381a25dc',
options=None)
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::197::ResourceManager.Request::(__init__)
ResName=`Storage.e3cb7694-47f7-4eac-9201-922e381a25dc`ReqID=`42f323ef-aa03-49e6-893f-0e4ff0764b6b`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '2560' at
'getStoragePoolInfo'
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::541::ResourceManager::(registerResource)
Trying to register resource
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc' for lock type
'shared'
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::600::ResourceManager::(registerResource)
Resource 'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc' is free.
Now locking as 'shared' (1 active user)
Thread-1250::DEBUG::2013-10-30
11:25:13,511::resourceManager::237::ResourceManager.Request::(grant)
ResName=`Storage.e3cb7694-47f7-4eac-9201-922e381a25dc`ReqID=`42f323ef-aa03-49e6-893f-0e4ff0764b6b`::Granted
request
Thread-1250::DEBUG::2013-10-30
11:25:13,512::task::811::TaskManager.Task::(resourceAcquired)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::_resourcesAcquired:
Storage.e3cb7694-47f7-4eac-9201-922e381a25dc (shared)
Thread-1250::DEBUG::2013-10-30
11:25:13,512::task::974::TaskManager.Task::(_decref)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::ref 1 aborting False
Thread-1250::
INFO::2013-10-30
11:25:13,517::logUtils::47::dispatcher::(wrapper) Run and protect:
getStoragePoolInfo, Return response: {'info': {'spm_id': 1,
'master_uuid': '21e3d67f-fdec-4f87-b09e-53af994a6f87', 'name':
'GlusterSD', 'version': '3', 'domains':
'21e3d67f-fdec-4f87-b09e-53af994a6f87:Active,8e56a914-3395-48ad-ab4c-c2ca7f9956b4:Active',
'pool_status': 'connected', 'isoprefix':
'/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/images/11111111-1111-1111-1111-111111111111',
'type': 'GLUSTERFS', 'master_ver': 1, 'lver': 3}, 'dominfo':
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'status': 'Active',
'diskfree': '239969632256', 'alerts': [], 'version': 3,
'disktotal': '492120440832'},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'status': 'Active',
'diskfree': '201653747712', 'alerts': [], 'version': 0,
'disktotal': '246783410176'}}}
Thread-1250::DEBUG::2013-10-30
11:25:13,517::task::1168::TaskManager.Task::(prepare)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::finished: {'info':
{'spm_id': 1, 'master_uuid':
'21e3d67f-fdec-4f87-b09e-53af994a6f87', 'name': 'GlusterSD',
'version': '3', 'domains':
'21e3d67f-fdec-4f87-b09e-53af994a6f87:Active,8e56a914-3395-48ad-ab4c-c2ca7f9956b4:Active',
'pool_status': 'connected', 'isoprefix':
'/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/images/11111111-1111-1111-1111-111111111111',
'type': 'GLUSTERFS', 'master_ver': 1, 'lver': 3}, 'dominfo':
{'21e3d67f-fdec-4f87-b09e-53af994a6f87': {'status': 'Active',
'diskfree': '239969632256', 'alerts': [], 'version': 3,
'disktotal': '492120440832'},
'8e56a914-3395-48ad-ab4c-c2ca7f9956b4': {'status': 'Active',
'diskfree': '201653747712', 'alerts': [], 'version': 0,
'disktotal': '246783410176'}}}
Thread-1250::DEBUG::2013-10-30
11:25:13,517::task::579::TaskManager.Task::(_updateState)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::moving from state
preparing -> state finished
Thread-1250::DEBUG::2013-10-30
11:25:13,517::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources
{'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc': < ResourceRef
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc', isValid: 'True'
obj: 'None'>}
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc'
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc'
(0 active users)
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc' is free,
finding out if anyone is waiting for it.
Thread-1250::DEBUG::2013-10-30
11:25:13,518::resourceManager::648::ResourceManager::(releaseResource)
No one is waiting for resource
'Storage.e3cb7694-47f7-4eac-9201-922e381a25dc', Clearing records.
Thread-1250::DEBUG::2013-10-30
11:25:13,518::task::974::TaskManager.Task::(_decref)
Task=`221b537a-2c1b-418f-93ff-bf77d4029097`::ref 0 aborting False
Thread-242::DEBUG::2013-10-30
11:25:15,046::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/glusterSD/172.16.0.100:vms/21e3d67f-fdec-4f87-b09e-53af994a6f87/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-242::DEBUG::2013-10-30
11:25:15,051::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n497 bytes
(497 B) copied, 0.00032077 s, 1.5 MB/s\n'; <rc> = 0
Thread-243::DEBUG::2013-10-30
11:25:17,777::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
'/usr/bin/dd iflag=direct
if=/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/dom_md/metadata
bs=4096 count=1' (cwd None)
Thread-243::DEBUG::2013-10-30
11:25:17,780::fileSD::238::Storage.Misc.excCmd::(getReadDelay)
SUCCESS: <err> = '0+1 records in\n0+1 records out\n373 bytes
(373 B) copied, 9.9232e-05 s, 3.8 MB/s\n'; <rc> = 0
engine.log
2013-10-30 11:17:47,729 INFO
[org.ovirt.engine.core.bll.RunVmOnceCommand]
(ajp--127.0.0.1-8702-6) Running command: RunVmOnceCommand
internal: false. Entities affected : ID:
200dfb05-461e-49d9-95a2-c0a7c7ced669 Type: VM
2013-10-30 11:17:47,743 INFO
[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
(ajp--127.0.0.1-8702-6) START, CreateVmVDSCommand(HostName =
Hypervisor, HostId = da764821-0650-40f0-b7cf-038766fbd2ca,
vmId=200dfb05-461e-49d9-95a2-c0a7c7ced669, vm=VM [CentOS_30]), log
id: 638a3a0e
2013-10-30 11:17:47,765 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
(ajp--127.0.0.1-8702-6) START, CreateVDSCommand(HostName =
Hypervisor, HostId = da764821-0650-40f0-b7cf-038766fbd2ca,
vmId=200dfb05-461e-49d9-95a2-c0a7c7ced669, vm=VM [CentOS_30]), log
id: 297dbca2
2013-10-30 11:17:47,784 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
(ajp--127.0.0.1-8702-6)
org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand
spiceSslCipherSuite=DEFAULT,memSize=2048,kvmEnable=true,smp=1,vmType=kvm,emulatedMachine=pc-1.0,keyboardLayout=en-us,memGuaranteedSize=1365,pitReinjection=false,nice=0,display=qxl,smartcardEnable=false,smpCoresPerSocket=1,spiceSecureChannels=smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard,timeOffset=0,transparentHugePages=true,vmId=200dfb05-461e-49d9-95a2-c0a7c7ced669,devices=[Ljava.util.HashMap;@f3eee3e,acpiEnable=true,vmName=CentOS_30,cpuType=Westmere,custom={}
2013-10-30 11:17:47,785 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
(ajp--127.0.0.1-8702-6) FINISH, CreateVDSCommand, log id: 297dbca2
2013-10-30 11:17:47,786 INFO
[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
(ajp--127.0.0.1-8702-6) FINISH, CreateVmVDSCommand, return:
WaitForLaunch, log id: 638a3a0e
2013-10-30 11:17:47,812 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-6) Correlation ID: 4f8bc4f9, Job ID:
74a77fc0-7e83-44c0-b3d7-334eb38612b4, Call Stack: null, Custom
Event ID: -1, Message: VM CentOS_30 was started by admin@internal
(Host: Hypervisor).
2013-10-30 11:17:51,115 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVdsCommand]
(DefaultQuartzScheduler_Worker-37) START,
FullListVdsCommand(HostName = Hypervisor, HostId =
da764821-0650-40f0-b7cf-038766fbd2ca, vds=Host[Hypervisor],
vmIds=[200dfb05-461e-49d9-95a2-c0a7c7ced669]), log id: 546bd3e2
2013-10-30 11:17:51,119 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVdsCommand]
(DefaultQuartzScheduler_Worker-37) FINISH, FullListVdsCommand,
return: [Ljava.util.HashMap;@318a7164, log id: 546bd3e2
2013-10-30 11:17:51,121 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-37) Received a scsi Device without
an address when processing VM 200dfb05-461e-49d9-95a2-c0a7c7ced669
devices, skipping device: {model=virtio-scsi, device=scsi,
type=controller}
2013-10-30 11:17:51,121 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-37) Received a cdrom Device without
an address when processing VM 200dfb05-461e-49d9-95a2-c0a7c7ced669
devices, skipping device: {shared=false, bootOrder=1, iface=ide,
index=2, specParams={}, device=cdrom,
path=/rhev/data-center/mnt/172.16.0.5:_home_iso2/8e56a914-3395-48ad-ab4c-c2ca7f9956b4/images/11111111-1111-1111-1111-111111111111/CentOS-6.4-x86_64-bin-DVD1.iso,
type=disk, readonly=true,
deviceId=e29c7ac7-f872-43e3-90e2-cb2205c0203f}
2013-10-30 11:17:51,122 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-37) Received a disk Device without
an address when processing VM 200dfb05-461e-49d9-95a2-c0a7c7ced669
devices, skipping device: {shared=false,
volumeID=202adc86-7857-40eb-8ca6-7cec3ef0d2d6, index=0,
propagateErrors=off, format=raw, type=disk, truesize=4096,
reqsize=0, bootOrder=2, iface=virtio,
volumeChain=[Ljava.lang.Object;@8741843,
imageID=b73a79fe-401b-434c-b023-6de4c3eb100d,
domainID=21e3d67f-fdec-4f87-b09e-53af994a6f87, specParams={},
optional=false, volumeInfo={volfileServer=172.16.0.100, volPort=0,
protocol=gluster, volTransport=tcp,
path=vms/21e3d67f-fdec-4f87-b09e-53af994a6f87/images/b73a79fe-401b-434c-b023-6de4c3eb100d/202adc86-7857-40eb-8ca6-7cec3ef0d2d6,
volType=network},
path=/rhev/data-center/e3cb7694-47f7-4eac-9201-922e381a25dc/21e3d67f-fdec-4f87-b09e-53af994a6f87/images/b73a79fe-401b-434c-b023-6de4c3eb100d/202adc86-7857-40eb-8ca6-7cec3ef0d2d6,
device=disk, poolID=e3cb7694-47f7-4eac-9201-922e381a25dc,
readonly=false, deviceId=b73a79fe-401b-434c-b023-6de4c3eb100d,
apparentsize=32212254720}
2013-10-30 11:17:51,123 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-37) Received a memballoon Device
without an address when processing VM
200dfb05-461e-49d9-95a2-c0a7c7ced669 devices, skipping device:
{specParams={model=virtio}, target=2097152, device=memballoon,
type=balloon, deviceId=b61b058d-cdfe-4d7c-8e6c-1bfd631cb26c}