Dafna,
Libvirtd.log shows no errors, but VM log shows the following:
2014-01-03 13:52:11.296+0000: starting up
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin QEMU_AUDIO_DRV=spice
/usr/bin/qemu-kvm -name OATEST2 -S -machine pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m
1024 -realtime mlock=off -smp 1,sockets=1,cores=1,threads=1 -uuid
d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -smbios type=1,manufacturer=oVirt,product=oVirt
Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C384C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6
-no-user-config -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,server,nowait -mon
chardev=charmonitor,id=monitor,mode=control -rtc base=2014-01-03T13:52:11,driftfix=slew
-no-shutdown -device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive
if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device
ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive
file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5c4b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-8387-ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca992-b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=native
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
-chardev
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev
socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
-chardev spicevmc,id=charchannel2,name=vdagent -device
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel2,name=com.redhat.spice.0
-spice
tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbredir,seamless-migration=on
-k en-us -device qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,addr=0x2
-device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
libvirt: Lock Driver error : unsupported configuration: Read/write, exclusive access,
disks were present, but no leases specified
2014-01-03 13:52:11.306+0000: shutting down
Not sure what you mean with this
http://forums.opensuse.org/english/get-technical-help-here/virtualization....
Do you want me to update libvirt with these repos on the oVirt-Node based installation?
Thanks,
Oliver
-----Ursprüngliche Nachricht-----
Von: Dafna Ron [mailto:dron@redhat.com]
Gesendet: Freitag, 03. Jänner 2014 15:10
An: Albl, Oliver
Cc: users(a)ovirt.org
Betreff: Re: AW: AW: [Users] Host cannot access storage domains
actually, looking at this again, it's a libvirt error and it can be related to selinux
or sasl.
can you also, look at libvirt log and the vm log under /var/log/libvirt?
On 01/03/2014 02:00 PM, Albl, Oliver wrote:
Dafna,
please find the logs below:
ERRORs in vdsm.log on host02:
Thread-61::ERROR::2014-01-03
13:51:48,956::sdc::137::Storage.StorageDomainCache::(_findDomain)
looking for unfetched domain f404398a-97f9-474c-af2c-e8887f53f688
Thread-61::ERROR::2014-01-03
13:51:48,959::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDom
ain) looking for domain f404398a-97f9-474c-af2c-e8887f53f688
Thread-323::ERROR::2014-01-03
13:52:11,527::vm::2132::vm.Vm::(_startUnderlyingVm)
vmId=`d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6`::The vm start process failed Traceback (most
recent call last):
File "/usr/share/vdsm/vm.py", line 2092, in _startUnderlyingVm
self._run()
File "/usr/share/vdsm/vm.py", line 2959, in _run
self._connection.createXML(domxml, flags),
File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", line
76, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 2920, in
createXML
libvirtError: Child quit during startup handshake: Input/output error
Thread-60::ERROR::2014-01-03
13:52:23,111::sdc::137::Storage.StorageDomainCache::(_findDomain)
looking for unfetched domain 52cf84ce-6eda-4337-8c94-491d94f5a18d
Thread-60::ERROR::2014-01-03
13:52:23,111::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDom
ain) looking for domain 52cf84ce-6eda-4337-8c94-491d94f5a18d
Thread-62::ERROR::2014-01-03
13:52:26,353::sdc::137::Storage.StorageDomainCache::(_findDomain)
looking for unfetched domain 7841a1c0-181a-4d43-9a25-b707accb5c4b
Thread-62::ERROR::2014-01-03
13:52:26,355::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDom
ain) looking for domain 7841a1c0-181a-4d43-9a25-b707accb5c4b
engine.log:
2014-01-03 14:52:06,976 INFO
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] START,
IsVmDuringInitiatingVDSCommand( vmId =
d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 5940cf72
2014-01-03 14:52:06,976 INFO
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH,
IsVmDuringInitiatingVDSCommand, return: false, log id: 5940cf72
2014-01-03 14:52:07,057 INFO
[org.ovirt.engine.core.bll.RunVmOnceCommand] (ajp--127.0.0.1-8702-3)
[2ab5cd2] Running command: RunVmOnceCommand internal: false. Entities
affected : ID: d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 Type: VM, ID:
d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 Type: VM
2014-01-03 14:52:07,151 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.IsoPrefixVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] START, IsoPrefixVDSCommand(HostName
= host02, HostId = 6dc7fac6-149e-4445-ace1-3c334a24d52a,
storagePoolId=b33d1793-252b-44ac-9685-3fe56b83c4c9), log id: 1705b611
2014-01-03 14:52:07,152 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.IsoPrefixVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, IsoPrefixVDSCommand, return:
/rhev/data-center/mnt/vmmgmt:_var_lib_exports_iso/f74f052e-0dc6-456d-a
f95-248c2227c2e5/images/11111111-1111-1111-1111-111111111111, log id:
1705b611
2014-01-03 14:52:07,170 INFO
[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] START, CreateVmVDSCommand(HostName =
host02, HostId = 6dc7fac6-149e-4445-ace1-3c334a24d52a,
vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, vm=VM [TEST2]), log id:
27b504de
2014-01-03 14:52:07,190 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] START, CreateVDSCommand(HostName =
host02, HostId = 6dc7fac6-149e-4445-ace1-3c334a24d52a,
vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, vm=VM [TEST2]), log id:
6ad0220
2014-01-03 14:52:08,472 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2]
org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand
spiceSslCipherSuite=DEFAULT,memSize=1024,kvmEnable=true,smp=1,vmType=k
vm,emulatedMachine=pc-1.0,keyboardLayout=en-us,memGuaranteedSize=1024,
pitReinjection=false,nice=0,display=qxl,smartcardEnable=false,smpCores
PerSocket=1,spiceSecureChannels=smain,sinputs,scursor,splayback,srecor
d,sdisplay,susbredir,ssmartcard,timeOffset=0,transparentHugePages=true
,vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6,devices=[Ljava.util.HashMap
;@3692311a,acpiEnable=true,vmName=TEST2,cpuType=SandyBridge,custom={}
2014-01-03 14:52:08,476 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, CreateVDSCommand, log id:
6ad0220
2014-01-03 14:52:08,484 INFO
[org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand]
(ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, CreateVmVDSCommand, return:
WaitForLaunch, log id: 27b504de
2014-01-03 14:52:08,497 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-3) [2ab5cd2] Correlation ID: 2ab5cd2, Job ID:
2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event ID: -1, Message: VM
TEST2 was started by oliver.albl (Host: host02).
2014-01-03 14:52:14,728 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
(DefaultQuartzScheduler_Worker-7) [24696b3e] START,
DestroyVDSCommand(HostName = host02, HostId =
6dc7fac6-149e-4445-ace1-3c334a24d52a,
vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, force=false,
secondsToWait=0, gracefully=false), log id: 6a95ffd5
2014-01-03 14:52:15,783 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
(DefaultQuartzScheduler_Worker-7) [24696b3e] FINISH,
DestroyVDSCommand, log id: 6a95ffd5
2014-01-03 14:52:15,804 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(DefaultQuartzScheduler_Worker-7) [24696b3e] Correlation ID: null, Call Stack: null,
Custom Event ID: -1, Message: VM TEST2 is down. Exit message: Child quit during startup
handshake: Input/output error.
2014-01-03 14:52:15,805 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-7) [24696b3e] Running on vds during
rerun failed vm: null
2014-01-03 14:52:15,805 INFO
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-7) [24696b3e] vm TEST2 running in db
and not running in vds - add to rerun treatment. vds host02
2014-01-03 14:52:15,808 ERROR
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-7) [24696b3e] Rerun vm
d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6. Called from vds host02
2014-01-03 14:52:15,810 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (pool-6-thread-40)
[24696b3e] Correlation ID: 2ab5cd2, Job ID: 2913133b-1301-484e-9887-b110841c8078, Call
Stack: null, Custom Event ID: -1, Message: Failed to run VM TEST2 on Host host02.
2014-01-03 14:52:15,823 INFO
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(pool-6-thread-40) [24696b3e] START, IsVmDuringInitiatingVDSCommand(
vmId = d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 35e1eec
2014-01-03 14:52:15,824 INFO
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(pool-6-thread-40) [24696b3e] FINISH, IsVmDuringInitiatingVDSCommand,
return: false, log id: 35e1eec
2014-01-03 14:52:15,858 WARN
[org.ovirt.engine.core.bll.RunVmOnceCommand] (pool-6-thread-40)
[24696b3e] CanDoAction of action RunVmOnce failed.
Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,
SCHEDULING_ALL_HOSTS_FILTERED_OUT
2014-01-03 14:52:15,862 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (pool-6-thread-40)
[24696b3e] Correlation ID: 2ab5cd2, Job ID: 2913133b-1301-484e-9887-b110841c8078, Call
Stack: null, Custom Event ID: -1, Message: Failed to run VM TEST2 (User: oliver.albl).
Thanks,
Oliver
-----Ursprüngliche Nachricht-----
Von: Dafna Ron [mailto:dron@redhat.com]
Gesendet: Freitag, 03. Jänner 2014 14:51
An: Albl, Oliver
Cc: users(a)ovirt.org
Betreff: Re: AW: [Users] Host cannot access storage domains
Thanks for reporting the issue :)
As for the vm, can you please find the error in vdsm.log and in engine and paste it?
Thanks,
Dafna
On 01/03/2014 01:49 PM, Albl, Oliver wrote:
> Dafna,
>
> you were right, it seems to be a caching issue. Rebooting the host did the job:
>
> Before Reboot:
>
> [root@host01 log]# vdsClient -s 0 getStorageDomainsList
> 52cf84ce-6eda-4337-8c94-491d94f5a18d
> f404398a-97f9-474c-af2c-e8887f53f688
> 7841a1c0-181a-4d43-9a25-b707accb5c4b
>
> [root@host02 log]# vdsClient -s 0 getStorageDomainsList
> 52cf84ce-6eda-4337-8c94-491d94f5a18d
> f404398a-97f9-474c-af2c-e8887f53f688
> 7841a1c0-181a-4d43-9a25-b707accb5c4b
> 925ee53a-69b5-440f-b145-138ada5b452e
>
> After Reboot:
>
> [root@host02 admin]# vdsClient -s 0 getStorageDomainsList
> 52cf84ce-6eda-4337-8c94-491d94f5a18d
> f404398a-97f9-474c-af2c-e8887f53f688
> 7841a1c0-181a-4d43-9a25-b707accb5c4b
>
> So now I have both hosts up and running but when I try to start a VM on the second
host, I receive the following messages in the events pane:
>
> VM TEST2 was started by oliver.albl (Host: host02) VM TEST2 is down.
> Exit message: Child quit during startup handshake: Input/output error.
>
> Thanks again for your help!
> Oliver
>
> -----Ursprüngliche Nachricht-----
> Von: Dafna Ron [mailto:dron@redhat.com]
> Gesendet: Freitag, 03. Jänner 2014 14:22
> An: Albl, Oliver
> Cc: users(a)ovirt.org
> Betreff: Re: [Users] Host cannot access storage domains
>
> yes, please attach the vdsm log
> also, can you run vdsClient 0 getStorageDomainsList and vdsClient 0 getDeviceList on
both hosts?
>
> It might be a cache issue, so can you please restart the host and if it helps attach
output before and after the reboot?
>
> Thanks,
>
> Dafna
>
>
> On 01/03/2014 01:12 PM, Albl, Oliver wrote:
>> Hi,
>>
>> I am starting with oVirt 3.3.2 and I have an issue adding a host to
>> a cluster.
>>
>> I am using oVirt Engine Version 3.3.2-1.el6
>>
>> There is a cluster with one host (installed with oVirt Node - 3.0.3
>> -
>> 1.1.fc19 ISO image) up and running.
>>
>> I installed a second host using the same ISO image.
>>
>> I approved the host in the cluster.
>>
>> When I try to activate the second host, I receive the following
>> messages in the events pane:
>>
>> State was set to Up for host host02.
>>
>> Host host02 reports about one of the Active Storage Domains as
>> Problematic.
>>
>> Host host02 cannot access one of the Storage Domains attached to the
>> Data Center Test303. Stetting Host state to Non-Operational.
>>
>> Failed to connect Host host02 to Storage Pool Test303
>>
>> There are 3 FC Storage Domains configured and visible to both hosts.
>>
>> multipath -ll shows all LUNs on both hosts.
>>
>> The engine.log reports the following about every five minutes:
>>
>> 2014-01-03 13:50:15,408 ERROR
>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (pool-6-thread-44) Domain 7841a1c0-181a-4d43-9a25-b707accb5c4b:
>> LUN_105 check timeot 69.7 is too big
>>
>> 2014-01-03 13:50:15,409 ERROR
>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (pool-6-thread-44) Domain 52cf84ce-6eda-4337-8c94-491d94f5a18d:
>> LUN_103 check timeot 59.6 is too big
>>
>> 2014-01-03 13:50:15,410 ERROR
>> [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44)
>> Storage Domain LUN_105 of pool Test303 is in problem in host host02
>>
>> 2014-01-03 13:50:15,411 ERROR
>> [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44)
>> Storage Domain LUN_103 of pool Test030 is in problem in host host02
>>
>> Please let me know if there are any log files I should attach.
>>
>> Thank you for your help!
>>
>> All the best,
>>
>> Oliver Albl
>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>>
http://lists.ovirt.org/mailman/listinfo/users
> --
> Dafna Ron
--
Dafna Ron
--
Dafna Ron