On Jan 18, 2012, at 19:10, Ryan Harper <ryanh(a)us.ibm.com> wrote:
> * Rami Vaknin <rvaknin(a)redhat.com> [2012-01-18 10:49]:
>> On 01/18/2012 06:39 PM, Ryan Harper wrote:
>>> * Ayal Baron<abaron(a)redhat.com> [2012-01-18 10:35]:
>>>>
>>>> ----- Original Message -----
>>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 10:15]:
>>>>>> On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote:
>>>>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 08:02]:
>>>>>>>> On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote:
>>>>>>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18
07:13]:
>>>>>>>>>> On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper
wrote:
>>>>>>>>>>> I've created some f16 VMs that contain
both ovirt-engine and a
>>>>>>>>>>> few
>>>>>>>>>>> to run vdsm as nodes. When I add in the VM
host into the
>>>>>>>>>>> engine and it
>>>>>>>>>>> attempts to install vdsm (even though
I've already installed
>>>>>>>>>>> vdsm) the
>>>>>>>>>>> install fails because the vdsm install
script is checking to
>>>>>>>>>>> see if the
>>>>>>>>>>> host has virt capabilities; since I'm
not running nested KVM,
>>>>>>>>>>> this
>>>>>>>>>>> fails. Is there a way to work around this
can enable a VM to
>>>>>>>>>>> be a host
>>>>>>>>>>> in oVirt? I had heard in the past there was
a way to create
>>>>>>>>>>> fake VMs
>>>>>>>>>>> when attempting to do ovirt-engine stress
testing, wondering
>>>>>>>>>>> if that
>>>>>>>>>>> might be of help here.
>>>>>>>>>>>
>>>>>>>>>>> Also, are their vdsm rpms built for RHEL6.x
available?
>>>>>>>>>>>
>>>>>>>>>>> Thanks!
>>>>>>>>>>>
>>>>>>>>>> Hi Rayn,
>>>>>>>>>>
>>>>>>>>>> - login to your ovirt-engine machine
>>>>>>>>>> - edit
>>>>>>>>>>
/usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py
>>>>>>>>>> - comment out the following:
>>>>>>>>>>
>>>>>>>>>> 836 if not
oDeploy.virtExplorer(random_num):
>>>>>>>>>> 837 logging.error('virtExplorer
test failed')
>>>>>>>>>> 838 return False
>>>>>>>>>> - reinstall host
>>>>>>>>> So I'm getting further, but now the
bootstrap.log has more
>>>>>>>>> errors below.
>>>>>>>>> If I follow the test day instructions, it indicates
to install
>>>>>>>>> vdsm and
>>>>>>>>> includes instructions, but it's clear that
ovirt-engine is
>>>>>>>>> configured by
>>>>>>>>> default to push out vdsm and install it. If
I've already
>>>>>>>>> configured and
>>>>>>>>> installed vdsm on the node is there any way to not
attempting to
>>>>>>>>> bootstrap
>>>>>>>>> vdsm
>>>>>>>>> at all and just attempt to have it connect?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='VDS PACKAGES'
>>>>>>>>> status='OK' result='qemu-kvm-tools'
>>>>>>>>> message='qemu-kvm-tools-0.15.1-3.fc16.x86_64
'/>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Basic
configuration found,
>>>>>>>>> skipping
>>>>>>>>> this step
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='CreateConf'
>>>>>>>>> status='OK'
>>>>>>>>> message='Basic configuration found, skipping
this step'/>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
>>>>>>>>> component='CheckLocalHostname'
>>>>>>>>> status='OK' message='Local hostname is
correct.'/>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Bridge ovirtmgmt
not found,
>>>>>>>>> need to
>>>>>>>>> create it.
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress
Entry.
>>>>>>>>>
url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress
return.
>>>>>>>>>
address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge
begin.
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface:
read host name:
>>>>>>>>>
ichigo-dom223.phx.austin.ibm.com
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface:
using host name
>>>>>>>>>
ichigo-dom223.phx.austin.ibm.com strIP=
192.168.68.223
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface
>>>>>>>>> IP=192.168.68.223
>>>>>>>>> strIface=engine
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR makeBridge found
existing
>>>>>>>>> bridge
>>>>>>>>> named:
>>>>>>>>> engine
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR makeBridge
errored: out=
>>>>>>>>> err=None
>>>>>>>>> ret=None
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge
return.
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR addNetwork error
trying to
>>>>>>>>> add
>>>>>>>>> management bridge
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='SetNetworking'
>>>>>>>>> status='FAIL' message='addNetwork error
trying to add management
>>>>>>>>> bridge'/>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress
Entry.
>>>>>>>>>
url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress
return.
>>>>>>>>>
address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile
start. IP =
>>>>>>>>>
ichigo-dom223.phx.austin.ibm.com port = 8080
fileName =
>>>>>>>>> "/engine.ssh.key.txt"
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG
/engine.ssh.key.txt failed in
>>>>>>>>> HTTPS.
>>>>>>>>> Retrying using HTTP.
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>> File "/tmp/deployUtil.py", line 1334, in
getRemoteFile
>>>>>>>>> conn.sock = getSSLSocket(sock, certPath)
>>>>>>>>> File "/tmp/deployUtil.py", line 1178, in
getSSLSocket
>>>>>>>>> cert_reqs=ssl.CERT_REQUIRED)
>>>>>>>>> File "/usr/lib64/python2.7/ssl.py", line
372, in wrap_socket
>>>>>>>>> ciphers=ciphers)
>>>>>>>>> File "/usr/lib64/python2.7/ssl.py", line
132, in __init__
>>>>>>>>> ciphers)
>>>>>>>>> SSLError: [Errno 185090050] _ssl.c:340:
error:0B084002:x509
>>>>>>>>> certificate
>>>>>>>>> routines:X509_load_cert_crl_file:system lib
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile
end.
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey
start
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Failed to read
>>>>>>>>> /root/.ssh/authorized_keys
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey:
failed to chmod
>>>>>>>>> authorized_keys
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>> File "/tmp/deployUtil.py", line 608, in
handleSSHKey
>>>>>>>>> silentRestoreCon(P_ROOT_AUTH_KEYS)
>>>>>>>>> File "/tmp/deployUtil.py", line 576, in
silentRestoreCon
>>>>>>>>> import selinux
>>>>>>>>> File
>>>>>>>>>
"/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>>>>> line
>>>>>>>>> 26,
>>>>>>>>> in<module>
>>>>>>>>> _selinux = swig_import_helper()
>>>>>>>>> File
>>>>>>>>>
"/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>>>>> line
>>>>>>>>> 22,
>>>>>>>>> in swig_import_helper
>>>>>>>>> _mod = imp.load_module('_selinux', fp,
pathname,
>>>>>>>>> description)
>>>>>>>>> ImportError:
>>>>>>>>>
/usr/lib64/python2.7/site-packages/selinux/_selinux.so:
>>>>>>>>> undefined symbol: selinux_check_access
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey end
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='SetSSHAccess'
>>>>>>>>> status='FAIL' message='Failed to write
server~s SSH key.'/>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR setSSHAccess test
failed
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='RHEV_INSTALL'
>>>>>>>>> status='FAIL'/>
>>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG **** End VDS
Validation ****
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> - add fake_kvm_support = True to your vdsm.conf
under
>>>>>>>>>> /etc/vdsm/vdsm.conf
>>>>>>>>>> - restart vdsmd service
>>>>>>>> please make sure selinux is set at least on permissive
mode;
>>>>>>>>
>>>>>>>> sed -i
's/SELINUX=disabled/SELINUX=permissive/g'
>>>>>>>> /etc/sysconfig/selinux
>>>>>>>>
>>>>>>>> reboot and reinstall.
>>>>>>>>
>>>>>>>> anyhow, if this is the case, its a known issue and patch
is
>>>>>>>> pending
>>>>>>>> upstream.
>>>>>>> I did this, but I was also able to just re-run the installer
and
>>>>>>> bootstrap completed. However, now I have another issue.
>>>>>>>
>>>>>>> THe host is marked unresponsive in engine, engine.log shows
a
>>>>>>> connectivity issue, but both hosts can ping and share data.
>>>>>>>
>>>>>>> . Stage completed. (Stage: Running second installation
script on
>>>>>>> Host)
>>>>>>> 2012-01-18 09:58:08,550 INFO
>>>>>>>
[org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper]
>>>>>>> (pool-5-thread-49) RunSSHCommand returns true
>>>>>>> 2012-01-18 09:58:08,550 INFO
>>>>>>> [org.ovirt.engine.core.bll.VdsInstaller]
>>>>>>> (pool-5-thread-49) FinishCommand ended:true
>>>>>>> 2012-01-18 09:58:08,554 INFO
>>>>>>> [org.ovirt.engine.core.bll.InstallVdsCommand]
(pool-5-thread-49)
>>>>>>> After
>>>>>>> Installation pool-5-thread-49
>>>>>>> 2012-01-18 09:58:08,555 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>>> (pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId =
>>>>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot,
>>>>>>> nonOperationalReason=NONE), log id: 703c3cbd
>>>>>>> 2012-01-18 09:58:08,560 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>>> (pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id:
>>>>>>> 703c3cbd
>>>>>>> 2012-01-18 09:58:08,560 INFO
>>>>>>> [org.ovirt.engine.core.bll.VdsCommand]
>>>>>>> (pool-5-thread-50) Waiting 300 seconds, for server to
finish
>>>>>>> reboot
>>>>>>> process.
>>>>>>> 2012-01-18 10:03:08,561 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>>> (pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId =
>>>>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive,
>>>>>>> nonOperationalReason=NONE), log id: 3e57bdd2
>>>>>>> 2012-01-18 10:03:08,570 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>>> (pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id:
>>>>>>> 3e57bdd2
>>>>>>> 2012-01-18 10:03:10,201 ERROR
>>>>>>>
[org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
>>>>>>> (QuartzScheduler_Worker-35) XML RPC error in command
>>>>>>> GetCapabilitiesVDS (
>>>>>>> Vds: ichigo-dom224 ), the error was:
>>>>>>> java.util.concurrent.ExecutionException:
>>>>>>> java.lang.reflect.InvocationTargetException
>>>>>>>
>>>>>>>
>>>>>>> I can query vdsm on the on the node:
>>>>>>>
>>>>>>> [root@f16-node1 ~]# vdsClient -s 0 getVdsCaps
>>>>>>> HBAInventory = {'iSCSI':
[{'InitiatorName':
>>>>>>> 'iqn.1994-05.com.redhat:2abcda43e16d'}],
'FC': []}
>>>>>>> ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d
>>>>>>> bondings = {'bond4': {'hwaddr':
'00:00:00:00:00:00', 'cfg':
>>>>>>> {},
>>>>>>> 'netmask': '', 'addr':
'', 'slaves': []}, 'bond0': {'hwaddr':
>>>>>>> '00:00:00:00:00:00', 'cfg': {},
'netmask': '', 'addr': '',
>>>>>>> 'slaves':
>>>>>>> []}, 'bond1': {'hwaddr':
'00:00:00:00:00:00', 'cfg': {},
>>>>>>> 'netmask':
>>>>>>> '', 'addr': '', 'slaves':
[]}, 'bond2': {'hwaddr':
>>>>>>> '00:00:00:00:00:00', 'cfg': {},
'netmask': '', 'addr': '',
>>>>>>> 'slaves':
>>>>>>> []}, 'bond3': {'hwaddr':
'00:00:00:00:00:00', 'cfg': {},
>>>>>>> 'netmask':
>>>>>>> '', 'addr': '', 'slaves':
[]}}
>>>>>>> clusterLevels = ['3.0']
>>>>>>> cpuCores = 1
>>>>>>> cpuFlags =
>>>>>>>
pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1
>>>>>>> cpuModel = Intel(Fake) CPU
>>>>>>> cpuSockets = 1
>>>>>>> cpuSpeed = 2800.482
>>>>>>> emulatedMachines = ['pc-0.14', 'pc',
'fedora-13', 'pc-0.13',
>>>>>>> 'pc-0.12', 'pc-0.11', 'pc-0.10',
'isapc']
>>>>>>> guestOverhead = 65
>>>>>>> hooks = {}
>>>>>>> kvmEnabled = true
>>>>>>> management_ip =
>>>>>>> memSize = 7988
>>>>>>> networks = {'ovirtmgmt': {'addr':
'192.168.68.224', 'cfg':
>>>>>>> {'DEVICE':
>>>>>>> 'ovirtmgmt', 'DELAY': '0',
'BOOTPROTO': 'dhcp', 'TYPE':
>>>>>>> 'Bridge',
>>>>>>> 'ONBOOT': 'yes'}, 'ports':
['eth0'], 'netmask':
>>>>>>> '255.255.192.0',
>>>>>>> 'stp': 'off', 'gateway':
'192.168.68.1'}}
>>>>>>> nics = {'eth0': {'hwaddr':
'00:FF:FE:00:00:E0', 'netmask':
>>>>>>> '',
>>>>>>> 'speed': 0, 'addr': ''}}
>>>>>>> operatingSystem = {'release': '1',
'version': '16', 'name':
>>>>>>> 'Fedora'}
>>>>>>> packages2 = {'kernel': {'release':
'7.fc16.x86_64',
>>>>>>> 'buildtime':
>>>>>>> 1320196248.0, 'version': '3.1.0'},
'spice-server':
>>>>>>> {'release':
>>>>>>> '1.fc16', 'buildtime':
'1321276111', 'version': '0.10.0'},
>>>>>>> 'vdsm':
>>>>>>> {'release': '0.fc16',
'buildtime': '1326734129', 'version':
>>>>>>> '4.9.3.1'}, 'qemu-kvm':
{'release': '3.fc16', 'buildtime':
>>>>>>> '1321651456', 'version':
'0.15.1'}, 'libvirt': {'release':
>>>>>>> '4.fc16',
>>>>>>> 'buildtime': '1324326688',
'version': '0.9.6'}, 'qemu-img':
>>>>>>> {'release': '3.fc16',
'buildtime': '1321651456', 'version':
>>>>>>> '0.15.1'}}
>>>>>>> reservedMem = 321
>>>>>>> software_revision = 0
>>>>>>> software_version = 4.9
>>>>>>> supportedProtocols = ['2.2', '2.3']
>>>>>>> supportedRHEVMs = ['3.0']
>>>>>>> uuid =
922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0
>>>>>>> version_name = Snow Man
>>>>>>> vlans = {}
>>>>>>> vmTypes = ['kvm']
>>>>>>>
>>>>>>>
>>>>>> can you check if problem is solved if you run iptables -F ?
>>>>> It doesn't.
>>>> can you also post the vdsm.log to see if the request made it and was
>>>> rejected for some reason?
>>> yes, I think this is the issue:
>>>
>>> Thread-1016::ERROR::2012-01-18
>>> 11:36:50,986::SecureXMLRPCServer::73::root::(handle_error) client
>>> ('192.168.68.223', 58819)
>>> Traceback (most recent call last):
>>> File "/usr/lib64/python2.7/SocketServer.py", line 582, in
>>> process_request_thread
>>> self.finish_request(request, client_address)
>>> File "/usr/share/vdsm/SecureXMLRPCServer.py", line 66, in
>>> finish_request
>>> request.do_handshake()
>>> File "/usr/lib64/python2.7/ssl.py", line 296, in do_handshake
>>> self._sslobj.do_handshake()
>>>
>>> I have vdsm.conf with ssl=true, however, if I set ssl=false, then I
>>> cannot query vdsm from the localhost client:
>>>
>>> [root@f16-node1 vdsm]# vdsClient -s 0 getVdsCaps
>>>
>>> with ssl=false, that returns connection refused.
>>>
>>>
>> Indeed. If you want to work without ssl, you need to also change the
>> "UseSecureConnectionWithServers" option_name to "false" in
vdc_options
>> table in ovirt-engine database and restart jboss-as service , so it will
>> query vdsm without SSL.
>
> ok, now with ssl=false, and the database updated and jboss-as restart,
> engine sees the host as up, but the local query of the capabilities
> fails.
You should run the command without the "-s" (which stands for secure).
I learn something new every day. vdsClient 0 getVdsCaps works fine now.
--
Ryan Harper
Software Engineer; Linux Technology Center
IBM Corp., Austin, Tx
ryanh(a)us.ibm.com