
----- Original Message -----
* Haim Ateya <hateya@redhat.com> [2012-01-18 10:15]:
On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote:
* Haim Ateya<hateya@redhat.com> [2012-01-18 08:02]:
On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote:
* Haim Ateya<hateya@redhat.com> [2012-01-18 07:13]:
On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper wrote: >I've created some f16 VMs that contain both ovirt-engine and a >few >to run vdsm as nodes. When I add in the VM host into the >engine and it >attempts to install vdsm (even though I've already installed >vdsm) the >install fails because the vdsm install script is checking to >see if the >host has virt capabilities; since I'm not running nested KVM, >this >fails. Is there a way to work around this can enable a VM to >be a host >in oVirt? I had heard in the past there was a way to create >fake VMs >when attempting to do ovirt-engine stress testing, wondering >if that >might be of help here. > >Also, are their vdsm rpms built for RHEL6.x available? > >Thanks! >
Hi Rayn,
- login to your ovirt-engine machine - edit /usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py - comment out the following:
836 if not oDeploy.virtExplorer(random_num): 837 logging.error('virtExplorer test failed') 838 return False - reinstall host
So I'm getting further, but now the bootstrap.log has more errors below. If I follow the test day instructions, it indicates to install vdsm and includes instructions, but it's clear that ovirt-engine is configured by default to push out vdsm and install it. If I've already configured and installed vdsm on the node is there any way to not attempting to bootstrap vdsm at all and just attempt to have it connect?
Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='VDS PACKAGES' status='OK' result='qemu-kvm-tools' message='qemu-kvm-tools-0.15.1-3.fc16.x86_64 '/> Wed, 18 Jan 2012 08:35:37 DEBUG Basic configuration found, skipping this step Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='CreateConf' status='OK' message='Basic configuration found, skipping this step'/> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='CheckLocalHostname' status='OK' message='Local hostname is correct.'/> Wed, 18 Jan 2012 08:35:37 DEBUG Bridge ovirtmgmt not found, need to create it. Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry. url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/ Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return. address=ichigo-dom223.phx.austin.ibm.com port=8080 Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge begin. Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: read host name: ichigo-dom223.phx.austin.ibm.com Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: using host name ichigo-dom223.phx.austin.ibm.com strIP= 192.168.68.223 Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface IP=192.168.68.223 strIface=engine Wed, 18 Jan 2012 08:35:37 ERROR makeBridge found existing bridge named: engine Wed, 18 Jan 2012 08:35:37 ERROR makeBridge errored: out= err=None ret=None Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge return. Wed, 18 Jan 2012 08:35:37 ERROR addNetwork error trying to add management bridge Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetNetworking' status='FAIL' message='addNetwork error trying to add management bridge'/> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry. url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/ Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return. address=ichigo-dom223.phx.austin.ibm.com port=8080 Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile start. IP = ichigo-dom223.phx.austin.ibm.com port = 8080 fileName = "/engine.ssh.key.txt" Wed, 18 Jan 2012 08:35:37 DEBUG /engine.ssh.key.txt failed in HTTPS. Retrying using HTTP. Traceback (most recent call last): File "/tmp/deployUtil.py", line 1334, in getRemoteFile conn.sock = getSSLSocket(sock, certPath) File "/tmp/deployUtil.py", line 1178, in getSSLSocket cert_reqs=ssl.CERT_REQUIRED) File "/usr/lib64/python2.7/ssl.py", line 372, in wrap_socket ciphers=ciphers) File "/usr/lib64/python2.7/ssl.py", line 132, in __init__ ciphers) SSLError: [Errno 185090050] _ssl.c:340: error:0B084002:x509 certificate routines:X509_load_cert_crl_file:system lib Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile end. Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey start Wed, 18 Jan 2012 08:35:37 DEBUG Failed to read /root/.ssh/authorized_keys Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey: failed to chmod authorized_keys Traceback (most recent call last): File "/tmp/deployUtil.py", line 608, in handleSSHKey silentRestoreCon(P_ROOT_AUTH_KEYS) File "/tmp/deployUtil.py", line 576, in silentRestoreCon import selinux File "/usr/lib64/python2.7/site-packages/selinux/__init__.py", line 26, in<module> _selinux = swig_import_helper() File "/usr/lib64/python2.7/site-packages/selinux/__init__.py", line 22, in swig_import_helper _mod = imp.load_module('_selinux', fp, pathname, description) ImportError: /usr/lib64/python2.7/site-packages/selinux/_selinux.so: undefined symbol: selinux_check_access Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey end Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetSSHAccess' status='FAIL' message='Failed to write server~s SSH key.'/> Wed, 18 Jan 2012 08:35:37 ERROR setSSHAccess test failed Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='RHEV_INSTALL' status='FAIL'/> Wed, 18 Jan 2012 08:35:37 DEBUG **** End VDS Validation ****
- add fake_kvm_support = True to your vdsm.conf under /etc/vdsm/vdsm.conf - restart vdsmd service
please make sure selinux is set at least on permissive mode;
sed -i 's/SELINUX=disabled/SELINUX=permissive/g' /etc/sysconfig/selinux
reboot and reinstall.
anyhow, if this is the case, its a known issue and patch is pending upstream.
I did this, but I was also able to just re-run the installer and bootstrap completed. However, now I have another issue.
THe host is marked unresponsive in engine, engine.log shows a connectivity issue, but both hosts can ping and share data.
. Stage completed. (Stage: Running second installation script on Host) 2012-01-18 09:58:08,550 INFO [org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper] (pool-5-thread-49) RunSSHCommand returns true 2012-01-18 09:58:08,550 INFO [org.ovirt.engine.core.bll.VdsInstaller] (pool-5-thread-49) FinishCommand ended:true 2012-01-18 09:58:08,554 INFO [org.ovirt.engine.core.bll.InstallVdsCommand] (pool-5-thread-49) After Installation pool-5-thread-49 2012-01-18 09:58:08,555 INFO [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] (pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId = 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot, nonOperationalReason=NONE), log id: 703c3cbd 2012-01-18 09:58:08,560 INFO [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] (pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id: 703c3cbd 2012-01-18 09:58:08,560 INFO [org.ovirt.engine.core.bll.VdsCommand] (pool-5-thread-50) Waiting 300 seconds, for server to finish reboot process. 2012-01-18 10:03:08,561 INFO [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] (pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId = 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive, nonOperationalReason=NONE), log id: 3e57bdd2 2012-01-18 10:03:08,570 INFO [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] (pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id: 3e57bdd2 2012-01-18 10:03:10,201 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand] (QuartzScheduler_Worker-35) XML RPC error in command GetCapabilitiesVDS ( Vds: ichigo-dom224 ), the error was: java.util.concurrent.ExecutionException: java.lang.reflect.InvocationTargetException
I can query vdsm on the on the node:
[root@f16-node1 ~]# vdsClient -s 0 getVdsCaps HBAInventory = {'iSCSI': [{'InitiatorName': 'iqn.1994-05.com.redhat:2abcda43e16d'}], 'FC': []} ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d bondings = {'bond4': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond0': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond1': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond2': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond3': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}} clusterLevels = ['3.0'] cpuCores = 1 cpuFlags = pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1 cpuModel = Intel(Fake) CPU cpuSockets = 1 cpuSpeed = 2800.482 emulatedMachines = ['pc-0.14', 'pc', 'fedora-13', 'pc-0.13', 'pc-0.12', 'pc-0.11', 'pc-0.10', 'isapc'] guestOverhead = 65 hooks = {} kvmEnabled = true management_ip = memSize = 7988 networks = {'ovirtmgmt': {'addr': '192.168.68.224', 'cfg': {'DEVICE': 'ovirtmgmt', 'DELAY': '0', 'BOOTPROTO': 'dhcp', 'TYPE': 'Bridge', 'ONBOOT': 'yes'}, 'ports': ['eth0'], 'netmask': '255.255.192.0', 'stp': 'off', 'gateway': '192.168.68.1'}} nics = {'eth0': {'hwaddr': '00:FF:FE:00:00:E0', 'netmask': '', 'speed': 0, 'addr': ''}} operatingSystem = {'release': '1', 'version': '16', 'name': 'Fedora'} packages2 = {'kernel': {'release': '7.fc16.x86_64', 'buildtime': 1320196248.0, 'version': '3.1.0'}, 'spice-server': {'release': '1.fc16', 'buildtime': '1321276111', 'version': '0.10.0'}, 'vdsm': {'release': '0.fc16', 'buildtime': '1326734129', 'version': '4.9.3.1'}, 'qemu-kvm': {'release': '3.fc16', 'buildtime': '1321651456', 'version': '0.15.1'}, 'libvirt': {'release': '4.fc16', 'buildtime': '1324326688', 'version': '0.9.6'}, 'qemu-img': {'release': '3.fc16', 'buildtime': '1321651456', 'version': '0.15.1'}} reservedMem = 321 software_revision = 0 software_version = 4.9 supportedProtocols = ['2.2', '2.3'] supportedRHEVMs = ['3.0'] uuid = 922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0 version_name = Snow Man vlans = {} vmTypes = ['kvm']
can you check if problem is solved if you run iptables -F ?
It doesn't.
can you also post the vdsm.log to see if the request made it and was rejected for some reason?
Here's the engine log:
2012-01-18 10:26:50,394 INFO [org.ovirt.engine.core.vdsbroker.ActivateVdsVDSCommand] (pool-5-thread-50) START, ActivateVdsVDSCommand(vdsId = 8c627fa8-41d8-11e1-8d2f-00fffe0000df), log id: 29fb3bf1 2012-01-18 10:26:50,402 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand] (pool-5-thread-50) XML RPC error in command GetCapabilitiesVDS ( Vds: ichigo-dom224 ), the error was: java.util.concurrent.ExecutionException: java.lang.reflect.InvocationTargetException 2012-01-18 10:26:50,403 ERROR [org.ovirt.engine.core.vdsbroker.VdsManager] (pool-5-thread-50) VDS::handleNetworkException Server failed to respond, vds_id = 8c627fa8-41d8-11e1-8d2f-00fffe0000df, vds_name = ichigo-dom224, error = VDSNetworkException: 2012-01-18 10:26:50,406 INFO [org.ovirt.engine.core.bll.VdsEventListener] (pool-5-thread-49) ResourceManager::vdsNotResponding entered for Host 8c627fa8-41d8-11e1-8d2f-00fffe0000df, ichigo-dom224 2012-01-18 10:26:50,409 ERROR [org.ovirt.engine.core.bll.VdsNotRespondingTreatmentCommand] (pool-5-thread-49) Failed to run Fence script on vds:ichigo-dom224, VMs moved to UnKnown instead. 2012-01-18 10:26:50,413 INFO [org.ovirt.engine.core.vdsbroker.VdsManager] (pool-5-thread-50) ResourceManager::activateVds - failed to get VDS = 8c627fa8-41d8-11e1-8d2f-00fffe0000df capabilities with error: VDSNetworkException: . 2012-01-18 10:26:50,413 INFO [org.ovirt.engine.core.vdsbroker.VdsManager] (pool-5-thread-50) ResourceManager::activateVds - failed to activate VDS = 8c627fa8-41d8-11e1-8d2f-00fffe0000df 2012-01-18 10:26:50,416 INFO [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand] (pool-5-thread-50) Running command: HandleVdsCpuFlagsOrClusterChangedCommand internal: true. Entities affected : ID: 8c627fa8-41d8-11e1-8d2f-00fffe0000df Type: VDS 2012-01-18 10:26:50,416 ERROR [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand] (pool-5-thread-50) Could not find server cpu for server 8c627fa8-41d8-11e1-8d2f-00fffe0000df:ichigo-dom224, flags: null 2012-01-18 10:26:50,418 WARN [org.ovirt.engine.core.bll.HandleVdsVersionCommand] (pool-5-thread-50) CanDoAction of action HandleVdsVersion failed. Reasons:VDS_CANNOT_CHECK_VERSION_HOST_NON_RESPONSIVE 2012-01-18 10:26:50,418 INFO [org.ovirt.engine.core.vdsbroker.ActivateVdsVDSCommand] (pool-5-thread-50) FINISH, ActivateVdsVDSCommand, log id: 29fb3bf1 2012-01-18 10:26:52,040 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand] (QuartzScheduler_Worker-44) XML RPC error in command GetCapabilitiesVDS ( Vds: ichigo-dom224 ), the error was: java.util.concurrent.ExecutionException: java.lang.reflect.InvocationTargetException
and from the node: [root@f16-node1 ~]# iptables -L -nv Chain INPUT (policy ACCEPT 9147 packets, 1907K bytes) pkts bytes target prot opt in out source destination
Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) pkts bytes target prot opt in out source destination
Chain OUTPUT (policy ACCEPT 8980 packets, 2465K bytes) pkts bytes target prot opt in out source destination [root@f16-node1 ~]# iptables -F [root@f16-node1 ~]# iptables -L -nv Chain INPUT (policy ACCEPT 20 packets, 3202 bytes) pkts bytes target prot opt in out source destination
Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) pkts bytes target prot opt in out source destination
Chain OUTPUT (policy ACCEPT 17 packets, 3924 bytes) pkts bytes target prot opt in out source destination
-- Ryan Harper Software Engineer; Linux Technology Center IBM Corp., Austin, Tx ryanh@us.ibm.com
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users