* Haim Ateya <hateya(a)redhat.com> [2012-01-18 10:15]:
On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote:
>* Haim Ateya<hateya(a)redhat.com> [2012-01-18 08:02]:
>>On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote:
>>>* Haim Ateya<hateya(a)redhat.com> [2012-01-18 07:13]:
>>>>On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper wrote:
>>>>>I've created some f16 VMs that contain both ovirt-engine and a
few
>>>>>to run vdsm as nodes. When I add in the VM host into the engine and
it
>>>>>attempts to install vdsm (even though I've already installed
vdsm) the
>>>>>install fails because the vdsm install script is checking to see if
the
>>>>>host has virt capabilities; since I'm not running nested KVM,
this
>>>>>fails. Is there a way to work around this can enable a VM to be a
host
>>>>>in oVirt? I had heard in the past there was a way to create fake
VMs
>>>>>when attempting to do ovirt-engine stress testing, wondering if that
>>>>>might be of help here.
>>>>>
>>>>>Also, are their vdsm rpms built for RHEL6.x available?
>>>>>
>>>>>Thanks!
>>>>>
>>>>
>>>>Hi Rayn,
>>>>
>>>>- login to your ovirt-engine machine
>>>>- edit
>>>>/usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py
>>>>- comment out the following:
>>>>
>>>> 836 if not oDeploy.virtExplorer(random_num):
>>>> 837 logging.error('virtExplorer test failed')
>>>> 838 return False
>>>>- reinstall host
>>>
>>>So I'm getting further, but now the bootstrap.log has more errors below.
>>>If I follow the test day instructions, it indicates to install vdsm and
>>>includes instructions, but it's clear that ovirt-engine is configured by
>>>default to push out vdsm and install it. If I've already configured and
>>>installed vdsm on the node is there any way to not attempting to
>>>bootstrap
>>>vdsm
>>>at all and just attempt to have it connect?
>>>
>>>
>>>
>>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='VDS PACKAGES'
>>>status='OK' result='qemu-kvm-tools'
>>>message='qemu-kvm-tools-0.15.1-3.fc16.x86_64 '/>
>>>Wed, 18 Jan 2012 08:35:37 DEBUG Basic configuration found, skipping
>>>this step
>>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='CreateConf'
status='OK'
>>>message='Basic configuration found, skipping this step'/>
>>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='CheckLocalHostname'
>>>status='OK' message='Local hostname is correct.'/>
>>>Wed, 18 Jan 2012 08:35:37 DEBUG Bridge ovirtmgmt not found, need to
>>>create it.
>>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry.
>>>url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return.
>>>address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge begin.
>>>Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: read host name:
>>>ichigo-dom223.phx.austin.ibm.com
>>>Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: using host name
>>>ichigo-dom223.phx.austin.ibm.com strIP= 192.168.68.223
>>>Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface IP=192.168.68.223
>>>strIface=engine
>>>Wed, 18 Jan 2012 08:35:37 ERROR makeBridge found existing bridge
>>>named:
>>>engine
>>>Wed, 18 Jan 2012 08:35:37 ERROR makeBridge errored: out=
>>>err=None
>>>ret=None
>>>Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge return.
>>>Wed, 18 Jan 2012 08:35:37 ERROR addNetwork error trying to add
>>>management bridge
>>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetNetworking'
>>>status='FAIL' message='addNetwork error trying to add management
>>>bridge'/>
>>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry.
>>>url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return.
>>>address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile start. IP =
>>>ichigo-dom223.phx.austin.ibm.com port = 8080 fileName =
>>>"/engine.ssh.key.txt"
>>>Wed, 18 Jan 2012 08:35:37 DEBUG /engine.ssh.key.txt failed in HTTPS.
>>>Retrying using HTTP.
>>>Traceback (most recent call last):
>>> File "/tmp/deployUtil.py", line 1334, in getRemoteFile
>>> conn.sock = getSSLSocket(sock, certPath)
>>> File "/tmp/deployUtil.py", line 1178, in getSSLSocket
>>> cert_reqs=ssl.CERT_REQUIRED)
>>> File "/usr/lib64/python2.7/ssl.py", line 372, in wrap_socket
>>> ciphers=ciphers)
>>> File "/usr/lib64/python2.7/ssl.py", line 132, in __init__
>>> ciphers)
>>>SSLError: [Errno 185090050] _ssl.c:340: error:0B084002:x509 certificate
>>>routines:X509_load_cert_crl_file:system lib
>>>Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile end.
>>>Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey start
>>>Wed, 18 Jan 2012 08:35:37 DEBUG Failed to read
>>>/root/.ssh/authorized_keys
>>>Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey: failed to chmod
>>>authorized_keys
>>>Traceback (most recent call last):
>>> File "/tmp/deployUtil.py", line 608, in handleSSHKey
>>> silentRestoreCon(P_ROOT_AUTH_KEYS)
>>> File "/tmp/deployUtil.py", line 576, in silentRestoreCon
>>> import selinux
>>> File "/usr/lib64/python2.7/site-packages/selinux/__init__.py",
line
>>> 26,
>>> in<module>
>>> _selinux = swig_import_helper()
>>> File "/usr/lib64/python2.7/site-packages/selinux/__init__.py",
line
>>> 22,
>>> in swig_import_helper
>>> _mod = imp.load_module('_selinux', fp, pathname, description)
>>>ImportError: /usr/lib64/python2.7/site-packages/selinux/_selinux.so:
>>>undefined symbol: selinux_check_access
>>>Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey end
>>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetSSHAccess'
>>>status='FAIL' message='Failed to write server~s SSH
key.'/>
>>>Wed, 18 Jan 2012 08:35:37 ERROR setSSHAccess test failed
>>>Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='RHEV_INSTALL'
>>>status='FAIL'/>
>>>Wed, 18 Jan 2012 08:35:37 DEBUG **** End VDS Validation ****
>>>
>>>
>>>
>>>
>>>>- add fake_kvm_support = True to your vdsm.conf under
>>>>/etc/vdsm/vdsm.conf
>>>>- restart vdsmd service
>>>
>>
>>please make sure selinux is set at least on permissive mode;
>>
>>sed -i 's/SELINUX=disabled/SELINUX=permissive/g'
>>/etc/sysconfig/selinux
>>
>>reboot and reinstall.
>>
>>anyhow, if this is the case, its a known issue and patch is pending
>>upstream.
>
>I did this, but I was also able to just re-run the installer and
>bootstrap completed. However, now I have another issue.
>
>THe host is marked unresponsive in engine, engine.log shows a
>connectivity issue, but both hosts can ping and share data.
>
>. Stage completed. (Stage: Running second installation script on Host)
>2012-01-18 09:58:08,550 INFO
>[org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper]
>(pool-5-thread-49) RunSSHCommand returns true
>2012-01-18 09:58:08,550 INFO [org.ovirt.engine.core.bll.VdsInstaller]
>(pool-5-thread-49) FinishCommand ended:true
>2012-01-18 09:58:08,554 INFO
>[org.ovirt.engine.core.bll.InstallVdsCommand] (pool-5-thread-49) After
>Installation pool-5-thread-49
>2012-01-18 09:58:08,555 INFO
>[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>(pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId =
>8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot,
>nonOperationalReason=NONE), log id: 703c3cbd
>2012-01-18 09:58:08,560 INFO
>[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>(pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id: 703c3cbd
>2012-01-18 09:58:08,560 INFO [org.ovirt.engine.core.bll.VdsCommand]
>(pool-5-thread-50) Waiting 300 seconds, for server to finish reboot
>process.
>2012-01-18 10:03:08,561 INFO
>[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>(pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId =
>8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive,
>nonOperationalReason=NONE), log id: 3e57bdd2
>2012-01-18 10:03:08,570 INFO
>[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>(pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id: 3e57bdd2
>2012-01-18 10:03:10,201 ERROR
>[org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
>(QuartzScheduler_Worker-35) XML RPC error in command GetCapabilitiesVDS (
>Vds: ichigo-dom224 ), the error was:
>java.util.concurrent.ExecutionException:
>java.lang.reflect.InvocationTargetException
>
>
>I can query vdsm on the on the node:
>
>[root@f16-node1 ~]# vdsClient -s 0 getVdsCaps
> HBAInventory = {'iSCSI': [{'InitiatorName':
> 'iqn.1994-05.com.redhat:2abcda43e16d'}], 'FC': []}
> ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d
> bondings = {'bond4': {'hwaddr': '00:00:00:00:00:00',
'cfg': {},
> 'netmask': '', 'addr': '', 'slaves':
[]}, 'bond0': {'hwaddr':
> '00:00:00:00:00:00', 'cfg': {}, 'netmask': '',
'addr': '', 'slaves':
> []}, 'bond1': {'hwaddr': '00:00:00:00:00:00',
'cfg': {}, 'netmask':
> '', 'addr': '', 'slaves': []}, 'bond2':
{'hwaddr':
> '00:00:00:00:00:00', 'cfg': {}, 'netmask': '',
'addr': '', 'slaves':
> []}, 'bond3': {'hwaddr': '00:00:00:00:00:00',
'cfg': {}, 'netmask':
> '', 'addr': '', 'slaves': []}}
> clusterLevels = ['3.0']
> cpuCores = 1
> cpuFlags =
>
pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1
> cpuModel = Intel(Fake) CPU
> cpuSockets = 1
> cpuSpeed = 2800.482
> emulatedMachines = ['pc-0.14', 'pc', 'fedora-13',
'pc-0.13',
> 'pc-0.12', 'pc-0.11', 'pc-0.10', 'isapc']
> guestOverhead = 65
> hooks = {}
> kvmEnabled = true
> management_ip =
> memSize = 7988
> networks = {'ovirtmgmt': {'addr': '192.168.68.224',
'cfg': {'DEVICE':
> 'ovirtmgmt', 'DELAY': '0', 'BOOTPROTO':
'dhcp', 'TYPE': 'Bridge',
> 'ONBOOT': 'yes'}, 'ports': ['eth0'],
'netmask': '255.255.192.0',
> 'stp': 'off', 'gateway': '192.168.68.1'}}
> nics = {'eth0': {'hwaddr': '00:FF:FE:00:00:E0',
'netmask': '',
> 'speed': 0, 'addr': ''}}
> operatingSystem = {'release': '1', 'version':
'16', 'name': 'Fedora'}
> packages2 = {'kernel': {'release': '7.fc16.x86_64',
'buildtime':
> 1320196248.0, 'version': '3.1.0'}, 'spice-server':
{'release':
> '1.fc16', 'buildtime': '1321276111', 'version':
'0.10.0'}, 'vdsm':
> {'release': '0.fc16', 'buildtime': '1326734129',
'version':
> '4.9.3.1'}, 'qemu-kvm': {'release': '3.fc16',
'buildtime':
> '1321651456', 'version': '0.15.1'}, 'libvirt':
{'release': '4.fc16',
> 'buildtime': '1324326688', 'version': '0.9.6'},
'qemu-img':
> {'release': '3.fc16', 'buildtime': '1321651456',
'version': '0.15.1'}}
> reservedMem = 321
> software_revision = 0
> software_version = 4.9
> supportedProtocols = ['2.2', '2.3']
> supportedRHEVMs = ['3.0']
> uuid = 922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0
> version_name = Snow Man
> vlans = {}
> vmTypes = ['kvm']
>
>
can you check if problem is solved if you run iptables -F ?
It doesn't.
Here's the engine log:
2012-01-18 10:26:50,394 INFO [org.ovirt.engine.core.vdsbroker.ActivateVdsVDSCommand]
(pool-5-thread-50) START, ActivateVdsVDSCommand(vdsId =
8c627fa8-41d8-11e1-8d2f-00fffe0000df), log id: 29fb3bf1
2012-01-18 10:26:50,402 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
(pool-5-thread-50) XML RPC error in command GetCapabilitiesVDS ( Vds: ichigo-dom224 ), the
error was: java.util.concurrent.ExecutionException:
java.lang.reflect.InvocationTargetException
2012-01-18 10:26:50,403 ERROR [org.ovirt.engine.core.vdsbroker.VdsManager]
(pool-5-thread-50) VDS::handleNetworkException Server failed to respond, vds_id =
8c627fa8-41d8-11e1-8d2f-00fffe0000df, vds_name = ichigo-dom224, error =
VDSNetworkException:
2012-01-18 10:26:50,406 INFO [org.ovirt.engine.core.bll.VdsEventListener]
(pool-5-thread-49) ResourceManager::vdsNotResponding entered for Host
8c627fa8-41d8-11e1-8d2f-00fffe0000df, ichigo-dom224
2012-01-18 10:26:50,409 ERROR [org.ovirt.engine.core.bll.VdsNotRespondingTreatmentCommand]
(pool-5-thread-49) Failed to run Fence script on vds:ichigo-dom224, VMs moved to UnKnown
instead.
2012-01-18 10:26:50,413 INFO [org.ovirt.engine.core.vdsbroker.VdsManager]
(pool-5-thread-50) ResourceManager::activateVds - failed to get VDS =
8c627fa8-41d8-11e1-8d2f-00fffe0000df capabilities with error: VDSNetworkException: .
2012-01-18 10:26:50,413 INFO [org.ovirt.engine.core.vdsbroker.VdsManager]
(pool-5-thread-50) ResourceManager::activateVds - failed to activate VDS =
8c627fa8-41d8-11e1-8d2f-00fffe0000df
2012-01-18 10:26:50,416 INFO
[org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand] (pool-5-thread-50)
Running command: HandleVdsCpuFlagsOrClusterChangedCommand internal: true. Entities
affected : ID: 8c627fa8-41d8-11e1-8d2f-00fffe0000df Type: VDS
2012-01-18 10:26:50,416 ERROR
[org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand] (pool-5-thread-50)
Could not find server cpu for server 8c627fa8-41d8-11e1-8d2f-00fffe0000df:ichigo-dom224,
flags: null
2012-01-18 10:26:50,418 WARN [org.ovirt.engine.core.bll.HandleVdsVersionCommand]
(pool-5-thread-50) CanDoAction of action HandleVdsVersion failed.
Reasons:VDS_CANNOT_CHECK_VERSION_HOST_NON_RESPONSIVE
2012-01-18 10:26:50,418 INFO [org.ovirt.engine.core.vdsbroker.ActivateVdsVDSCommand]
(pool-5-thread-50) FINISH, ActivateVdsVDSCommand, log id: 29fb3bf1
2012-01-18 10:26:52,040 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
(QuartzScheduler_Worker-44) XML RPC error in command GetCapabilitiesVDS ( Vds:
ichigo-dom224 ), the error was: java.util.concurrent.ExecutionException:
java.lang.reflect.InvocationTargetException
and from the node:
[root@f16-node1 ~]# iptables -L -nv
Chain INPUT (policy ACCEPT 9147 packets, 1907K bytes)
pkts bytes target prot opt in out source destination
Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
pkts bytes target prot opt in out source destination
Chain OUTPUT (policy ACCEPT 8980 packets, 2465K bytes)
pkts bytes target prot opt in out source destination
[root@f16-node1 ~]# iptables -F
[root@f16-node1 ~]# iptables -L -nv
Chain INPUT (policy ACCEPT 20 packets, 3202 bytes)
pkts bytes target prot opt in out source destination
Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
pkts bytes target prot opt in out source destination
Chain OUTPUT (policy ACCEPT 17 packets, 3924 bytes)
pkts bytes target prot opt in out source destination
--
Ryan Harper
Software Engineer; Linux Technology Center
IBM Corp., Austin, Tx
ryanh(a)us.ibm.com