[Users] creating and running a vm fails

Rene Rosenberger r.rosenberger at netbiscuits.com
Thu Mar 29 07:32:03 UTC 2012


Hi,

i have just retried to start the vm what fails again. Looking in /var/log/vsdm/vdsm.log shows last entry on 27.03.2012.

MainThread::INFO::2012-03-27 10:58:13,525::vdsm::70::vds::(run) I am the actual vdsm 4.9-0
MainThread::DEBUG::2012-03-27 10:58:15,120::resourceManager::379::ResourceManager::(registerNamespace) Registering namespace 'Storage'
MainThread::DEBUG::2012-03-27 10:58:15,122::threadPool::45::Misc.ThreadPool::(__init__) Enter - numThreads: 10.0, waitTimeout: 3, maxTasks: 500.0
MainThread::DEBUG::2012-03-27 10:58:15,393::multipath::109::Storage.Multipath::(isEnabled) multipath Defaulting to False
MainThread::DEBUG::2012-03-27 10:58:15,398::multipath::123::Storage.Misc.excCmd::(setupMultipath) '/usr/bin/sudo -n /bin/cp /tmp/tmpkart5V /etc/multipath.conf' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:15,505::multipath::123::Storage.Misc.excCmd::(setupMultipath) SUCCESS: <err> = ''; <rc> = 0
MainThread::DEBUG::2012-03-27 10:58:15,508::multipath::131::Storage.Misc.excCmd::(setupMultipath) '/usr/bin/sudo -n /sbin/multipath -F' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:15,591::multipath::131::Storage.Misc.excCmd::(setupMultipath) FAILED: <err> = ''; <rc> = 1
MainThread::DEBUG::2012-03-27 10:58:15,600::multipath::134::Storage.Misc.excCmd::(setupMultipath) '/usr/bin/sudo -n /sbin/service multipathd restart' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:16,123::multipath::134::Storage.Misc.excCmd::(setupMultipath) SUCCESS: <err> = ''; <rc> = 0
MainThread::DEBUG::2012-03-27 10:58:16,130::hsm::334::Storage.Misc.excCmd::(__validateLvmLockingType) '/usr/bin/sudo -n /sbin/lvm dumpconfig global/locking_type' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:16,424::hsm::334::Storage.Misc.excCmd::(__validateLvmLockingType) SUCCESS: <err> = ''; <rc> = 0
MainThread::DEBUG::2012-03-27 10:58:16,428::lvm::316::OperationMutex::(_reloadpvs) Operation 'lvm reload operation' got the operation mutex
MainThread::DEBUG::2012-03-27 10:58:16,434::lvm::284::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n /sbin/lvm pvs --config " devices { preferred_names = [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter = [ \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1  wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } " --noheadings --units b --nosuffix --separator | -o uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_count,dev_size' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:16,708::lvm::284::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = ''; <rc> = 0
MainThread::DEBUG::2012-03-27 10:58:16,711::lvm::339::OperationMutex::(_reloadpvs) Operation 'lvm reload operation' released the operation mutex
MainThread::DEBUG::2012-03-27 10:58:16,712::lvm::349::OperationMutex::(_reloadvgs) Operation 'lvm reload operation' got the operation mutex
MainThread::DEBUG::2012-03-27 10:58:16,713::lvm::284::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n /sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter = [ \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1  wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } " --noheadings --units b --nosuffix --separator | -o uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:16,906::lvm::284::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '  No volume groups found\n'; <rc> = 0
MainThread::DEBUG::2012-03-27 10:58:16,909::lvm::376::OperationMutex::(_reloadvgs) Operation 'lvm reload operation' released the operation mutex
MainThread::DEBUG::2012-03-27 10:58:16,910::lvm::284::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n /sbin/lvm lvs --config " devices { preferred_names = [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter = [ \\"r%.*%\\" ] }  global {  locking_type=1  prioritise_write_locks=1  wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } " --noheadings --units b --nosuffix --separator | -o uuid,name,vg_name,attr,size,seg_start_pe,devices,tags' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:17,119::lvm::284::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '  No volume groups found\n'; <rc> = 0
Thread-12::DEBUG::2012-03-27 10:58:17,123::misc::1022::SamplingMethod::(__call__) Trying to enter sampling method (storage.sdc.refreshStorage)
MainThread::INFO::2012-03-27 10:58:17,130::dispatcher::118::Storage.Dispatcher::(__init__) Starting StorageDispatcher...
Thread-12::DEBUG::2012-03-27 10:58:17,132::misc::1024::SamplingMethod::(__call__) Got in to sampling method
Thread-12::DEBUG::2012-03-27 10:58:17,141::misc::1022::SamplingMethod::(__call__) Trying to enter sampling method (storage.iscsi.rescan)
Thread-12::DEBUG::2012-03-27 10:58:17,142::misc::1024::SamplingMethod::(__call__) Got in to sampling method
Thread-12::DEBUG::2012-03-27 10:58:17,147::iscsiadm::48::Storage.Misc.excCmd::(_runCmd) '/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None)
MainThread::ERROR::2012-03-27 10:58:17,164::netinfo::126::root::(speed) cannot read eth0 speed
Traceback (most recent call last):
  File "/usr/lib/python2.6/site-packages/vdsm/netinfo.py", line 113, in speed
    s = int(file('/sys/class/net/%s/speed' % dev).read())
IOError: [Errno 22] Invalid argument
MainThread::DEBUG::2012-03-27 10:58:17,198::utils::602::Storage.Misc.excCmd::(execCmd) '/usr/bin/pgrep -xf ksmd' (cwd None)
Thread-12::DEBUG::2012-03-27 10:58:17,297::iscsiadm::48::Storage.Misc.excCmd::(_runCmd) FAILED: <err> = 'iscsiadm: No session found.\n'; <rc> = 21
Thread-12::DEBUG::2012-03-27 10:58:17,304::misc::1032::SamplingMethod::(__call__) Returning last result
Thread-12::DEBUG::2012-03-27 10:58:17,306::supervdsm::83::SuperVdsmProxy::(_killSupervdsm) Could not kill old Super Vdsm [Errno 2] No such file or directory: '/var/run/vdsm/svdsm.pid'
Thread-12::DEBUG::2012-03-27 10:58:17,308::supervdsm::71::SuperVdsmProxy::(_launchSupervdsm) Launching Super Vdsm
Thread-12::DEBUG::2012-03-27 10:58:17,309::supervdsm::74::Storage.Misc.excCmd::(_launchSupervdsm) '/usr/bin/sudo -n /usr/bin/python /usr/share/vdsm/supervdsmServer.pyc f90ef831-7bff-4eb4-a9b5-890538959b22 2858' (cwd None)
MainThread::DEBUG::2012-03-27 10:58:17,390::utils::602::Storage.Misc.excCmd::(execCmd) SUCCESS: <err> = ''; <rc> = 0
MainThread::INFO::2012-03-27 10:58:17,438::vmChannels::139::vds::(settimeout) Setting channels' timeout to 30 seconds.
VM Channels Listener::INFO::2012-03-27 10:58:17,524::vmChannels::127::vds::(run) Starting VM channels listener thread.
MainThread::DEBUG::2012-03-27 10:58:17,864::supervdsmServer::201::SuperVdsm.Server::(main) Making sure I'm root
MainThread::DEBUG::2012-03-27 10:58:17,866::supervdsmServer::205::SuperVdsm.Server::(main) Parsing cmd args
MainThread::DEBUG::2012-03-27 10:58:17,867::supervdsmServer::208::SuperVdsm.Server::(main) Creating PID file
MainThread::DEBUG::2012-03-27 10:58:17,867::supervdsmServer::212::SuperVdsm.Server::(main) Cleaning old socket
MainThread::DEBUG::2012-03-27 10:58:17,868::supervdsmServer::216::SuperVdsm.Server::(main) Setting up keep alive thread
MainThread::DEBUG::2012-03-27 10:58:17,887::supervdsmServer::221::SuperVdsm.Server::(main) Creating remote object manager
MainThread::DEBUG::2012-03-27 10:58:17,894::supervdsmServer::232::SuperVdsm.Server::(main) Started serving super vdsm object
Thread-12::DEBUG::2012-03-27 10:58:19,378::supervdsm::92::SuperVdsmProxy::(_connect) Trying to connect to Super Vdsm
Thread-12::DEBUG::2012-03-27 10:58:19,431::supervdsm::64::SuperVdsmProxy::(__init__) Connected to Super Vdsm
Thread-12::DEBUG::2012-03-27 10:58:19,737::multipath::71::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo -n /sbin/multipath' (cwd None)
Thread-12::DEBUG::2012-03-27 10:58:19,838::multipath::71::Storage.Misc.excCmd::(rescan) FAILED: <err> = ''; <rc> = 1
Thread-12::DEBUG::2012-03-27 10:58:19,841::lvm::457::OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' got the operation mutex
Thread-12::DEBUG::2012-03-27 10:58:19,842::lvm::459::OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' released the operation mutex
Thread-12::DEBUG::2012-03-27 10:58:19,843::lvm::469::OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' got the operation mutex
Thread-12::DEBUG::2012-03-27 10:58:19,844::lvm::471::OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' released the operation mutex
Thread-12::DEBUG::2012-03-27 10:58:19,844::lvm::490::OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' got the operation mutex
Thread-12::DEBUG::2012-03-27 10:58:19,845::lvm::492::OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' released the operation mutex
Thread-12::DEBUG::2012-03-27 10:58:19,846::misc::1032::SamplingMethod::(__call__) Returning last result
Thread-12::DEBUG::2012-03-27 10:58:19,847::hsm::358::Storage.HSM::(__cleanStorageRepository) Started cleaning storage repository at '/rhev/data-center'
Thread-12::DEBUG::2012-03-27 10:58:19,853::hsm::390::Storage.HSM::(__cleanStorageRepository) White list: ['/rhev/data-center/hsm-tasks', '/rhev/data-center/hsm-tasks/*', '/rhev/data-center/mnt']
Thread-12::DEBUG::2012-03-27 10:58:19,857::hsm::391::Storage.HSM::(__cleanStorageRepository) Mount list: []
Thread-12::DEBUG::2012-03-27 10:58:19,857::hsm::393::Storage.HSM::(__cleanStorageRepository) Cleaning leftovers
Thread-12::DEBUG::2012-03-27 10:58:19,858::hsm::436::Storage.HSM::(__cleanStorageRepository) Finished cleaning storage repository at '/rhev/data-center'

Regards, rene


-----Ursprüngliche Nachricht-----
Von: Moran Goldboim [mailto:mgoldboi at redhat.com]
Gesendet: Donnerstag, 29. März 2012 09:06
An: Rene Rosenberger
Cc: Itamar Heim; users at oVirt.org
Betreff: Re: [Users] creating and running a vm fails

I suspect something happened to vdsm - can you attach vdsm logs
(/var/log/vdsm/vdsm.log) from that time and maybe sneak a pick at /var/log/core if we have something there.
Moran.

On 03/29/2012 08:07 AM, Rene Rosenberger wrote:
> Hi,
>
> the host on which the vm should run is up. It has a green arrow.
>
> Regards, rene
>
>
> -----Ursprüngliche Nachricht-----
> Von: Itamar Heim [mailto:iheim at redhat.com]
> Gesendet: Mittwoch, 28. März 2012 22:07
> An: Rene Rosenberger
> Cc: users at oVirt.org
> Betreff: Re: [Users] creating and running a vm fails
>
> On 03/28/2012 02:33 PM, Rene Rosenberger wrote:
>> Hi,
>>
>> when i create a VM with it virtual disk and i want to run it once to
>> install from imported iso file I get the following error message:
>>
>> 2012-03-28 14:31:35,488
>> INFO[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
>> (http--0.0.0.0-8443-5) [cf5d84e] START, CreateVmVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a,
>> vmId=93220e0a-610a-4b20-986e-3c8b0d39e35f,
>> vm=org.ovirt.engine.core.common.businessentities.VM at 6e69e491), log id:
>> 3fc82aad
>>
>> 2012-03-28 14:31:35,498
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
>> (http--0.0.0.0-8443-5) [cf5d84e] START, CreateVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a,
>> vmId=93220e0a-610a-4b20-986e-3c8b0d39e35f,
>> vm=org.ovirt.engine.core.common.businessentities.VM at 6e69e491), log id:
>> 7e63d56b
>>
>> 2012-03-28 14:31:35,620
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
>> (http--0.0.0.0-8443-5) [cf5d84e]
>> org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand
>> spiceSslCipherSuite=DEFAULT,memSize=512,kvmEnable=true,boot=dc,smp=1,
>> e
>> mulatedMachine=pc-0.14,vmType=kvm,keyboardLayout=en-us,pitReinjection
>> =
>> false,nice=0,display=vnc,tabletEnable=true,smpCoresPerSocket=1,spiceS
>> e
>> cureChannels=,spiceMonitors=1,cdrom=/rhev/data-center/13080edc-77ea-1
>> 1
>> e1-b6a4-525400c49d2a/93e2079e-ef3b-452f-af71-5b3e7eb32ba0/images/1111
>> 1
>> 111-1111-1111-1111-111111111111/CentOS-6.2-x86_64-bin-DVD1.iso,timeOf
>> f
>> set=0,transparentHugePages=true,drives=[Ljava.util.Map;@7f0ba1c1,vmId
>> =
>> 93220e0a-610a-4b20-986e-3c8b0d39e35f,acpiEnable=true,vmName=test,cpuT
>> y
>> pe=Opteron_G3,custom={}
>>
>> 2012-03-28 14:31:35,620
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
>> (http--0.0.0.0-8443-5) [cf5d84e] FINISH, CreateVDSCommand, log id:
>> 7e63d56b
>>
>> 2012-03-28 14:31:35,625
>> INFO[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
>> (http--0.0.0.0-8443-5) [cf5d84e]
>> IncreasePendingVms::CreateVmIncreasing
>> vds KVM-DMZ-04 pending vcpu count, now 1. Vm: test
>>
>> 2012-03-28 14:31:35,631
>> INFO[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
>> (http--0.0.0.0-8443-5) [cf5d84e] FINISH, CreateVmVDSCommand, return:
>> WaitForLaunch, log id: 3fc82aad
>>
>> 2012-03-28 14:31:39,316
>> WARN[org.ovirt.engine.core.vdsbroker.VdsManager]
>> (QuartzScheduler_Worker-35)
>> ResourceManager::refreshVdsRunTimeInfo::Failed to refresh VDS , vds =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a : KVM-DMZ-04, VDS Network Error,
>> continuing.
>>
>> VDSNetworkException:
>>
>> 2012-03-28 14:31:41,328
>> WARN[org.ovirt.engine.core.vdsbroker.VdsManager]
>> (QuartzScheduler_Worker-29)
>> ResourceManager::refreshVdsRunTimeInfo::Failed to refresh VDS , vds =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a : KVM-DMZ-04, VDS Network Error,
>> continuing.
>>
>> VDSNetworkException:
>>
>> 2012-03-28 14:31:43,340
>> WARN[org.ovirt.engine.core.vdsbroker.VdsManager]
>> (QuartzScheduler_Worker-45)
>> ResourceManager::refreshVdsRunTimeInfo::Failed to refresh VDS , vds =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a : KVM-DMZ-04, VDS Network Error,
>> continuing.
>>
>> VDSNetworkException:
>>
>> 2012-03-28 14:31:43,519
>> INFO[org.ovirt.engine.core.bll.storage.SetStoragePoolStatusCommand]
>> (QuartzScheduler_Worker-38) [7ed2f412] Running command:
>> SetStoragePoolStatusCommand internal: true. Entities affected :ID:
>> 13080edc-77ea-11e1-b6a4-525400c49d2a Type: StoragePool
>>
>> 2012-03-28 14:31:43,543 ERROR
>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (QuartzScheduler_Worker-38) [7ed2f412]
>> IrsBroker::Failed::GetStoragePoolInfoVDS due to: ConnectException:
>> Connection refused
>>
>> 2012-03-28 14:31:45,351
>> WARN[org.ovirt.engine.core.vdsbroker.VdsManager]
>> (QuartzScheduler_Worker-47)
>> ResourceManager::refreshVdsRunTimeInfo::Failed to refresh VDS , vds =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a : KVM-DMZ-04, VDS Network Error,
>> continuing.
>>
>> VDSNetworkException:
>>
>> 2012-03-28 14:31:47,363
>> WARN[org.ovirt.engine.core.vdsbroker.VdsManager]
>> (QuartzScheduler_Worker-71)
>> ResourceManager::refreshVdsRunTimeInfo::Failed to refresh VDS , vds =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a : KVM-DMZ-04, VDS Network Error,
>> continuing.
>>
>> VDSNetworkException:
>>
>> 2012-03-28 14:31:51,091
>> INFO[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>> (QuartzScheduler_Worker-55) vm test running in db and not running in
>> vds
>> - add to rerun treatment. vds KVM-DMZ-04
>>
>> 2012-03-28 14:31:51,104
>> INFO[org.ovirt.engine.core.bll.InitVdsOnUpCommand]
>> (QuartzScheduler_Worker-55) [6df61eca] Running command:
>> InitVdsOnUpCommand internal: true.
>>
>> 2012-03-28 14:31:51,182
>> INFO[org.ovirt.engine.core.bll.storage.ConnectHostToStoragePoolServer
>> s
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] Running command:
>> ConnectHostToStoragePoolServersCommand internal: true. Entities
>> affected
>> :ID: 13080edc-77ea-11e1-b6a4-525400c49d2a Type: StoragePool
>>
>> 2012-03-28 14:31:51,185
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVD
>> S
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] START,
>> ConnectStorageServerVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a, storagePoolId =
>> 13080edc-77ea-11e1-b6a4-525400c49d2a, storageType = ISCSI,
>> connectionList = [{ id: e25e5093-9240-42ac-a21a-f0c216294944,
>> connection: 192.168.200.32 };]), log id: 11a3b069
>>
>> 2012-03-28 14:31:51,488
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVD
>> S
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] FINISH,
>> ConnectStorageServerVDSCommand, return:
>> {e25e5093-9240-42ac-a21a-f0c216294944=0}, log id: 11a3b069
>>
>> 2012-03-28 14:31:51,489
>> INFO[org.ovirt.engine.core.bll.storage.ConnectHostToStoragePoolServer
>> s
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] Host KVM-DMZ-04 storage
>> connection was succeeded
>>
>> 2012-03-28 14:31:51,492
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVD
>> S
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] START,
>> ConnectStorageServerVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a, storagePoolId =
>> 13080edc-77ea-11e1-b6a4-525400c49d2a, storageType = NFS,
>> connectionList = [{ id: 3dd798da-77ea-11e1-969c-525400c49d2a, connection:
>> oVirt.dynetic.de:/mnt/iso };]), log id: 1d73fc4e
>>
>> 2012-03-28 14:31:51,535
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVD
>> S
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] FINISH,
>> ConnectStorageServerVDSCommand, return:
>> {3dd798da-77ea-11e1-969c-525400c49d2a=0}, log id: 1d73fc4e
>>
>> 2012-03-28 14:31:51,535
>> INFO[org.ovirt.engine.core.bll.storage.ConnectHostToStoragePoolServer
>> s
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] Host KVM-DMZ-04 storage
>> connection was succeeded
>>
>> 2012-03-28 14:31:51,538
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVD
>> S
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] START,
>> ConnectStorageServerVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a, storagePoolId =
>> 13080edc-77ea-11e1-b6a4-525400c49d2a, storageType = NFS,
>> connectionList = [{ id: 57d6595f-1109-49f9-a7f3-f8fe255c34bd, connection:
>> 192.168.200.32:/nfsexport };]), log id: 618343ee
>>
>> 2012-03-28 14:31:51,579
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVD
>> S
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] FINISH,
>> ConnectStorageServerVDSCommand, return:
>> {57d6595f-1109-49f9-a7f3-f8fe255c34bd=0}, log id: 618343ee
>>
>> 2012-03-28 14:31:51,579
>> INFO[org.ovirt.engine.core.bll.storage.ConnectHostToStoragePoolServer
>> s
>> Command]
>> (QuartzScheduler_Worker-55) [48a331d9] Host KVM-DMZ-04 storage
>> connection was succeeded
>>
>> 2012-03-28 14:31:51,594
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStoragePoolVDSC
>> o
>> mmand]
>> (QuartzScheduler_Worker-55) [48a331d9] START,
>> ConnectStoragePoolVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a, storagePoolId =
>> 13080edc-77ea-11e1-b6a4-525400c49d2a, vds_spm_id = 1, masterDomainId
>> = 8ed25a57-f53a-4cf0-bb92-781f3ce36a48, masterVersion = 1), log id:
>> 190d7dea
>>
>> 2012-03-28 14:32:02,207
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStoragePoolVDSC
>> o
>> mmand]
>> (QuartzScheduler_Worker-55) [48a331d9] FINISH,
>> ConnectStoragePoolVDSCommand, log id: 190d7dea
>>
>> 2012-03-28 14:32:02,231
>> INFO[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDire
>> c
>> tor]
>> (QuartzScheduler_Worker-55) [48a331d9] No string for UNASSIGNED type.
>> Use default Log
>>
>> 2012-03-28 14:32:02,234
>> INFO[org.ovirt.engine.core.bll.MultipleActionsRunner]
>> (pool-5-thread-47) [48a331d9] MultipleActionsRunner of type
>> MigrateVmToServer invoked with no actions
>>
>> 2012-03-28 14:32:02,264
>> INFO[org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedComma
>> n
>> d]
>> (QuartzScheduler_Worker-55) [75839ca5] Running command:
>> HandleVdsCpuFlagsOrClusterChangedCommand internal: true. Entities
>> affected :ID: 0e0403a4-78ae-11e1-9c19-525400c49d2a Type: VDS
>>
>> 2012-03-28 14:32:02,272
>> INFO[org.ovirt.engine.core.bll.HandleVdsVersionCommand]
>> (QuartzScheduler_Worker-55) [2dd7e7a] Running command:
>> HandleVdsVersionCommand internal: true. Entities affected :ID:
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a Type: VDS
>>
>> 2012-03-28 14:32:02,276 ERROR
>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>> (QuartzScheduler_Worker-55) [2dd7e7a] Rerun vm
>> 93220e0a-610a-4b20-986e-3c8b0d39e35f. Called from vds KVM-DMZ-04
>>
>> 2012-03-28 14:32:02,280
>> INFO[org.ovirt.engine.core.vdsbroker.UpdateVdsDynamicDataVDSCommand]
>> (pool-5-thread-47) [2dd7e7a] START,
>> UpdateVdsDynamicDataVDSCommand(vdsId
>> = 0e0403a4-78ae-11e1-9c19-525400c49d2a,
>> vdsDynamic=org.ovirt.engine.core.common.businessentities.VdsDynamic at 2
>> f
>> 1f5a9f),
>> log id: 2901421c
>>
>> 2012-03-28 14:32:02,282
>> INFO[org.ovirt.engine.core.vdsbroker.UpdateVdsDynamicDataVDSCommand]
>> (pool-5-thread-47) [2dd7e7a] FINISH, UpdateVdsDynamicDataVDSCommand,
>> log
>> id: 2901421c
>>
>> 2012-03-28 14:32:02,299
>> INFO[org.ovirt.engine.core.vdsbroker.irsbroker.IsValidVDSCommand]
>> (pool-5-thread-47) [2dd7e7a] START, IsValidVDSCommand(storagePoolId =
>> 13080edc-77ea-11e1-b6a4-525400c49d2a, ignoreFailoverLimit = false,
>> compatabilityVersion = null), log id: 2d73522e
>>
>> 2012-03-28 14:32:02,300
>> INFO[org.ovirt.engine.core.vdsbroker.irsbroker.IsValidVDSCommand]
>> (pool-5-thread-47) [2dd7e7a] FINISH, IsValidVDSCommand, return:
>> false, log id: 2d73522e
>>
>> 2012-03-28 14:32:02,301
>> WARN[org.ovirt.engine.core.bll.RunVmOnceCommand]
>> (pool-5-thread-47) [2dd7e7a] CanDoAction of action RunVmOnce failed.
>> Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM
>> , ACTION_TYPE_FAILED_IMAGE_REPOSITORY_NOT_FOUND
>>
>> 2012-03-28 14:32:04,591
>> INFO[org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (QuartzScheduler_Worker-62) hostFromVds::selectedVds - KVM-DMZ-04,
>> spmStatus Free, storage pool Default
>>
>> 2012-03-28 14:32:05,430
>> INFO[org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (QuartzScheduler_Worker-62) SpmStatus on vds
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a: Free
>>
>> 2012-03-28 14:32:05,436
>> INFO[org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (QuartzScheduler_Worker-62) starting spm on vds KVM-DMZ-04, storage
>> pool Default, prevId 1, LVER 0
>>
>> 2012-03-28 14:32:05,439
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.SpmStartVDSCommand]
>> (QuartzScheduler_Worker-62) START, SpmStartVDSCommand(vdsId =
>> 0e0403a4-78ae-11e1-9c19-525400c49d2a, storagePoolId =
>> 13080edc-77ea-11e1-b6a4-525400c49d2a, prevId=1, prevLVER=0,
>> storagePoolFormatType=V2, recoveryMode=Manual, SCSIFencing=false),
>> log
>> id: 25d38cf4
>>
>> 2012-03-28 14:32:05,490
>> INFO[org.ovirt.engine.core.vdsbroker.vdsbroker.SpmStartVDSCommand]
>> (QuartzScheduler_Worker-62) spmStart polling started: taskId =
>> c2a0d9a2-e356-4aa9-bb29-f66a0899d8cc
>>
>> Any idea what to do?
>>
> looks like the host is no up?
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users




More information about the Users mailing list