* Ayal Baron<abaron(a)redhat.com> [2012-01-18 10:35]:
>
> ----- Original Message -----
>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 10:15]:
>>> On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote:
>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 08:02]:
>>>>> On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote:
>>>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 07:13]:
>>>>>>> On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper wrote:
>>>>>>>> I've created some f16 VMs that contain both
ovirt-engine and a
>>>>>>>> few
>>>>>>>> to run vdsm as nodes. When I add in the VM host into
the
>>>>>>>> engine and it
>>>>>>>> attempts to install vdsm (even though I've already
installed
>>>>>>>> vdsm) the
>>>>>>>> install fails because the vdsm install script is checking
to
>>>>>>>> see if the
>>>>>>>> host has virt capabilities; since I'm not running
nested KVM,
>>>>>>>> this
>>>>>>>> fails. Is there a way to work around this can enable a
VM to
>>>>>>>> be a host
>>>>>>>> in oVirt? I had heard in the past there was a way to
create
>>>>>>>> fake VMs
>>>>>>>> when attempting to do ovirt-engine stress testing,
wondering
>>>>>>>> if that
>>>>>>>> might be of help here.
>>>>>>>>
>>>>>>>> Also, are their vdsm rpms built for RHEL6.x available?
>>>>>>>>
>>>>>>>> Thanks!
>>>>>>>>
>>>>>>> Hi Rayn,
>>>>>>>
>>>>>>> - login to your ovirt-engine machine
>>>>>>> - edit
>>>>>>>
/usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py
>>>>>>> - comment out the following:
>>>>>>>
>>>>>>> 836 if not oDeploy.virtExplorer(random_num):
>>>>>>> 837 logging.error('virtExplorer test
failed')
>>>>>>> 838 return False
>>>>>>> - reinstall host
>>>>>> So I'm getting further, but now the bootstrap.log has more
>>>>>> errors below.
>>>>>> If I follow the test day instructions, it indicates to install
>>>>>> vdsm and
>>>>>> includes instructions, but it's clear that ovirt-engine is
>>>>>> configured by
>>>>>> default to push out vdsm and install it. If I've already
>>>>>> configured and
>>>>>> installed vdsm on the node is there any way to not attempting to
>>>>>> bootstrap
>>>>>> vdsm
>>>>>> at all and just attempt to have it connect?
>>>>>>
>>>>>>
>>>>>>
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='VDS
PACKAGES'
>>>>>> status='OK' result='qemu-kvm-tools'
>>>>>> message='qemu-kvm-tools-0.15.1-3.fc16.x86_64 '/>
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Basic configuration found,
>>>>>> skipping
>>>>>> this step
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='CreateConf'
>>>>>> status='OK'
>>>>>> message='Basic configuration found, skipping this
step'/>
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
>>>>>> component='CheckLocalHostname'
>>>>>> status='OK' message='Local hostname is
correct.'/>
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Bridge ovirtmgmt not found,
>>>>>> need to
>>>>>> create it.
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry.
>>>>>> url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return.
>>>>>>
address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge begin.
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: read host name:
>>>>>>
ichigo-dom223.phx.austin.ibm.com
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: using host name
>>>>>>
ichigo-dom223.phx.austin.ibm.com strIP= 192.168.68.223
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface
>>>>>> IP=192.168.68.223
>>>>>> strIface=engine
>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR makeBridge found existing
>>>>>> bridge
>>>>>> named:
>>>>>> engine
>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR makeBridge errored: out=
>>>>>> err=None
>>>>>> ret=None
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge return.
>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR addNetwork error trying to
>>>>>> add
>>>>>> management bridge
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='SetNetworking'
>>>>>> status='FAIL' message='addNetwork error trying to add
management
>>>>>> bridge'/>
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry.
>>>>>> url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return.
>>>>>>
address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile start. IP =
>>>>>>
ichigo-dom223.phx.austin.ibm.com port = 8080 fileName =
>>>>>> "/engine.ssh.key.txt"
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG /engine.ssh.key.txt failed in
>>>>>> HTTPS.
>>>>>> Retrying using HTTP.
>>>>>> Traceback (most recent call last):
>>>>>> File "/tmp/deployUtil.py", line 1334, in
getRemoteFile
>>>>>> conn.sock = getSSLSocket(sock, certPath)
>>>>>> File "/tmp/deployUtil.py", line 1178, in
getSSLSocket
>>>>>> cert_reqs=ssl.CERT_REQUIRED)
>>>>>> File "/usr/lib64/python2.7/ssl.py", line 372, in
wrap_socket
>>>>>> ciphers=ciphers)
>>>>>> File "/usr/lib64/python2.7/ssl.py", line 132, in
__init__
>>>>>> ciphers)
>>>>>> SSLError: [Errno 185090050] _ssl.c:340: error:0B084002:x509
>>>>>> certificate
>>>>>> routines:X509_load_cert_crl_file:system lib
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile end.
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey start
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Failed to read
>>>>>> /root/.ssh/authorized_keys
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey: failed to chmod
>>>>>> authorized_keys
>>>>>> Traceback (most recent call last):
>>>>>> File "/tmp/deployUtil.py", line 608, in
handleSSHKey
>>>>>> silentRestoreCon(P_ROOT_AUTH_KEYS)
>>>>>> File "/tmp/deployUtil.py", line 576, in
silentRestoreCon
>>>>>> import selinux
>>>>>> File
>>>>>>
"/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>> line
>>>>>> 26,
>>>>>> in<module>
>>>>>> _selinux = swig_import_helper()
>>>>>> File
>>>>>>
"/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>> line
>>>>>> 22,
>>>>>> in swig_import_helper
>>>>>> _mod = imp.load_module('_selinux', fp, pathname,
>>>>>> description)
>>>>>> ImportError:
>>>>>> /usr/lib64/python2.7/site-packages/selinux/_selinux.so:
>>>>>> undefined symbol: selinux_check_access
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey end
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='SetSSHAccess'
>>>>>> status='FAIL' message='Failed to write server~s SSH
key.'/>
>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR setSSHAccess test failed
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='RHEV_INSTALL'
>>>>>> status='FAIL'/>
>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG **** End VDS Validation ****
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>> - add fake_kvm_support = True to your vdsm.conf under
>>>>>>> /etc/vdsm/vdsm.conf
>>>>>>> - restart vdsmd service
>>>>> please make sure selinux is set at least on permissive mode;
>>>>>
>>>>> sed -i 's/SELINUX=disabled/SELINUX=permissive/g'
>>>>> /etc/sysconfig/selinux
>>>>>
>>>>> reboot and reinstall.
>>>>>
>>>>> anyhow, if this is the case, its a known issue and patch is
>>>>> pending
>>>>> upstream.
>>>> I did this, but I was also able to just re-run the installer and
>>>> bootstrap completed. However, now I have another issue.
>>>>
>>>> THe host is marked unresponsive in engine, engine.log shows a
>>>> connectivity issue, but both hosts can ping and share data.
>>>>
>>>> . Stage completed. (Stage: Running second installation script on
>>>> Host)
>>>> 2012-01-18 09:58:08,550 INFO
>>>> [org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper]
>>>> (pool-5-thread-49) RunSSHCommand returns true
>>>> 2012-01-18 09:58:08,550 INFO
>>>> [org.ovirt.engine.core.bll.VdsInstaller]
>>>> (pool-5-thread-49) FinishCommand ended:true
>>>> 2012-01-18 09:58:08,554 INFO
>>>> [org.ovirt.engine.core.bll.InstallVdsCommand] (pool-5-thread-49)
>>>> After
>>>> Installation pool-5-thread-49
>>>> 2012-01-18 09:58:08,555 INFO
>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>> (pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId =
>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot,
>>>> nonOperationalReason=NONE), log id: 703c3cbd
>>>> 2012-01-18 09:58:08,560 INFO
>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>> (pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id:
>>>> 703c3cbd
>>>> 2012-01-18 09:58:08,560 INFO
>>>> [org.ovirt.engine.core.bll.VdsCommand]
>>>> (pool-5-thread-50) Waiting 300 seconds, for server to finish
>>>> reboot
>>>> process.
>>>> 2012-01-18 10:03:08,561 INFO
>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>> (pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId =
>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive,
>>>> nonOperationalReason=NONE), log id: 3e57bdd2
>>>> 2012-01-18 10:03:08,570 INFO
>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>> (pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id:
>>>> 3e57bdd2
>>>> 2012-01-18 10:03:10,201 ERROR
>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
>>>> (QuartzScheduler_Worker-35) XML RPC error in command
>>>> GetCapabilitiesVDS (
>>>> Vds: ichigo-dom224 ), the error was:
>>>> java.util.concurrent.ExecutionException:
>>>> java.lang.reflect.InvocationTargetException
>>>>
>>>>
>>>> I can query vdsm on the on the node:
>>>>
>>>> [root@f16-node1 ~]# vdsClient -s 0 getVdsCaps
>>>> HBAInventory = {'iSCSI': [{'InitiatorName':
>>>> 'iqn.1994-05.com.redhat:2abcda43e16d'}], 'FC': []}
>>>> ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d
>>>> bondings = {'bond4': {'hwaddr':
'00:00:00:00:00:00', 'cfg':
>>>> {},
>>>> 'netmask': '', 'addr': '',
'slaves': []}, 'bond0': {'hwaddr':
>>>> '00:00:00:00:00:00', 'cfg': {}, 'netmask':
'', 'addr': '',
>>>> 'slaves':
>>>> []}, 'bond1': {'hwaddr':
'00:00:00:00:00:00', 'cfg': {},
>>>> 'netmask':
>>>> '', 'addr': '', 'slaves': []},
'bond2': {'hwaddr':
>>>> '00:00:00:00:00:00', 'cfg': {}, 'netmask':
'', 'addr': '',
>>>> 'slaves':
>>>> []}, 'bond3': {'hwaddr':
'00:00:00:00:00:00', 'cfg': {},
>>>> 'netmask':
>>>> '', 'addr': '', 'slaves': []}}
>>>> clusterLevels = ['3.0']
>>>> cpuCores = 1
>>>> cpuFlags =
>>>>
pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1
>>>> cpuModel = Intel(Fake) CPU
>>>> cpuSockets = 1
>>>> cpuSpeed = 2800.482
>>>> emulatedMachines = ['pc-0.14', 'pc',
'fedora-13', 'pc-0.13',
>>>> 'pc-0.12', 'pc-0.11', 'pc-0.10',
'isapc']
>>>> guestOverhead = 65
>>>> hooks = {}
>>>> kvmEnabled = true
>>>> management_ip =
>>>> memSize = 7988
>>>> networks = {'ovirtmgmt': {'addr':
'192.168.68.224', 'cfg':
>>>> {'DEVICE':
>>>> 'ovirtmgmt', 'DELAY': '0',
'BOOTPROTO': 'dhcp', 'TYPE':
>>>> 'Bridge',
>>>> 'ONBOOT': 'yes'}, 'ports': ['eth0'],
'netmask':
>>>> '255.255.192.0',
>>>> 'stp': 'off', 'gateway':
'192.168.68.1'}}
>>>> nics = {'eth0': {'hwaddr':
'00:FF:FE:00:00:E0', 'netmask':
>>>> '',
>>>> 'speed': 0, 'addr': ''}}
>>>> operatingSystem = {'release': '1',
'version': '16', 'name':
>>>> 'Fedora'}
>>>> packages2 = {'kernel': {'release':
'7.fc16.x86_64',
>>>> 'buildtime':
>>>> 1320196248.0, 'version': '3.1.0'},
'spice-server':
>>>> {'release':
>>>> '1.fc16', 'buildtime': '1321276111',
'version': '0.10.0'},
>>>> 'vdsm':
>>>> {'release': '0.fc16', 'buildtime':
'1326734129', 'version':
>>>> '4.9.3.1'}, 'qemu-kvm': {'release':
'3.fc16', 'buildtime':
>>>> '1321651456', 'version': '0.15.1'},
'libvirt': {'release':
>>>> '4.fc16',
>>>> 'buildtime': '1324326688', 'version':
'0.9.6'}, 'qemu-img':
>>>> {'release': '3.fc16', 'buildtime':
'1321651456', 'version':
>>>> '0.15.1'}}
>>>> reservedMem = 321
>>>> software_revision = 0
>>>> software_version = 4.9
>>>> supportedProtocols = ['2.2', '2.3']
>>>> supportedRHEVMs = ['3.0']
>>>> uuid = 922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0
>>>> version_name = Snow Man
>>>> vlans = {}
>>>> vmTypes = ['kvm']
>>>>
>>>>
>>> can you check if problem is solved if you run iptables -F ?
>> It doesn't.
> can you also post the vdsm.log to see if the request made it and was rejected for
some reason?
yes, I think this is the issue:
Thread-1016::ERROR::2012-01-18
11:36:50,986::SecureXMLRPCServer::73::root::(handle_error) client
('192.168.68.223', 58819)
Traceback (most recent call last):
File "/usr/lib64/python2.7/SocketServer.py", line 582, in
process_request_thread
self.finish_request(request, client_address)
File "/usr/share/vdsm/SecureXMLRPCServer.py", line 66, in finish_request
request.do_handshake()
File "/usr/lib64/python2.7/ssl.py", line 296, in do_handshake
self._sslobj.do_handshake()
I have vdsm.conf with ssl=true, however, if I set ssl=false, then I
cannot query vdsm from the localhost client:
[root@f16-node1 vdsm]# vdsClient -s 0 getVdsCaps
with ssl=false, that returns connection refused.
Indeed. If you want to work without ssl, you need to also change the
"UseSecureConnectionWithServers" option_name to "false" in vdc_options
table in ovirt-engine database and restart jboss-as service , so it will
query vdsm without SSL.
--
Thanks,
Rami Vaknin, QE @ Red Hat, TLV, IL.