* Rami Vaknin<rvaknin(a)redhat.com> [2012-01-18 10:49]:
> On 01/18/2012 06:39 PM, Ryan Harper wrote:
>> * Ayal Baron<abaron(a)redhat.com> [2012-01-18 10:35]:
>>> ----- Original Message -----
>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 10:15]:
>>>>> On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote:
>>>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18 08:02]:
>>>>>>> On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote:
>>>>>>>> * Haim Ateya<hateya(a)redhat.com> [2012-01-18
07:13]:
>>>>>>>>> On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper
wrote:
>>>>>>>>>> I've created some f16 VMs that contain both
ovirt-engine and a
>>>>>>>>>> few
>>>>>>>>>> to run vdsm as nodes. When I add in the VM host
into the
>>>>>>>>>> engine and it
>>>>>>>>>> attempts to install vdsm (even though I've
already installed
>>>>>>>>>> vdsm) the
>>>>>>>>>> install fails because the vdsm install script is
checking to
>>>>>>>>>> see if the
>>>>>>>>>> host has virt capabilities; since I'm not
running nested KVM,
>>>>>>>>>> this
>>>>>>>>>> fails. Is there a way to work around this can
enable a VM to
>>>>>>>>>> be a host
>>>>>>>>>> in oVirt? I had heard in the past there was a
way to create
>>>>>>>>>> fake VMs
>>>>>>>>>> when attempting to do ovirt-engine stress
testing, wondering
>>>>>>>>>> if that
>>>>>>>>>> might be of help here.
>>>>>>>>>>
>>>>>>>>>> Also, are their vdsm rpms built for RHEL6.x
available?
>>>>>>>>>>
>>>>>>>>>> Thanks!
>>>>>>>>>>
>>>>>>>>> Hi Rayn,
>>>>>>>>>
>>>>>>>>> - login to your ovirt-engine machine
>>>>>>>>> - edit
>>>>>>>>>
/usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py
>>>>>>>>> - comment out the following:
>>>>>>>>>
>>>>>>>>> 836 if not oDeploy.virtExplorer(random_num):
>>>>>>>>> 837 logging.error('virtExplorer test
failed')
>>>>>>>>> 838 return False
>>>>>>>>> - reinstall host
>>>>>>>> So I'm getting further, but now the bootstrap.log has
more
>>>>>>>> errors below.
>>>>>>>> If I follow the test day instructions, it indicates to
install
>>>>>>>> vdsm and
>>>>>>>> includes instructions, but it's clear that
ovirt-engine is
>>>>>>>> configured by
>>>>>>>> default to push out vdsm and install it. If I've
already
>>>>>>>> configured and
>>>>>>>> installed vdsm on the node is there any way to not
attempting to
>>>>>>>> bootstrap
>>>>>>>> vdsm
>>>>>>>> at all and just attempt to have it connect?
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='VDS PACKAGES'
>>>>>>>> status='OK' result='qemu-kvm-tools'
>>>>>>>> message='qemu-kvm-tools-0.15.1-3.fc16.x86_64
'/>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Basic configuration
found,
>>>>>>>> skipping
>>>>>>>> this step
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='CreateConf'
>>>>>>>> status='OK'
>>>>>>>> message='Basic configuration found, skipping this
step'/>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
>>>>>>>> component='CheckLocalHostname'
>>>>>>>> status='OK' message='Local hostname is
correct.'/>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Bridge ovirtmgmt not
found,
>>>>>>>> need to
>>>>>>>> create it.
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry.
>>>>>>>>
url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return.
>>>>>>>>
address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge begin.
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: read
host name:
>>>>>>>>
ichigo-dom223.phx.austin.ibm.com
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface: using
host name
>>>>>>>>
ichigo-dom223.phx.austin.ibm.com strIP= 192.168.68.223
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG _getMGTIface
>>>>>>>> IP=192.168.68.223
>>>>>>>> strIface=engine
>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR makeBridge found
existing
>>>>>>>> bridge
>>>>>>>> named:
>>>>>>>> engine
>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR makeBridge errored:
out=
>>>>>>>> err=None
>>>>>>>> ret=None
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG makeBridge return.
>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR addNetwork error
trying to
>>>>>>>> add
>>>>>>>> management bridge
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='SetNetworking'
>>>>>>>> status='FAIL' message='addNetwork error
trying to add management
>>>>>>>> bridge'/>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress Entry.
>>>>>>>>
url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getAddress return.
>>>>>>>>
address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile start.
IP =
>>>>>>>>
ichigo-dom223.phx.austin.ibm.com port = 8080 fileName =
>>>>>>>> "/engine.ssh.key.txt"
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG /engine.ssh.key.txt
failed in
>>>>>>>> HTTPS.
>>>>>>>> Retrying using HTTP.
>>>>>>>> Traceback (most recent call last):
>>>>>>>> File "/tmp/deployUtil.py", line 1334, in
getRemoteFile
>>>>>>>> conn.sock = getSSLSocket(sock, certPath)
>>>>>>>> File "/tmp/deployUtil.py", line 1178, in
getSSLSocket
>>>>>>>> cert_reqs=ssl.CERT_REQUIRED)
>>>>>>>> File "/usr/lib64/python2.7/ssl.py", line
372, in wrap_socket
>>>>>>>> ciphers=ciphers)
>>>>>>>> File "/usr/lib64/python2.7/ssl.py", line
132, in __init__
>>>>>>>> ciphers)
>>>>>>>> SSLError: [Errno 185090050] _ssl.c:340:
error:0B084002:x509
>>>>>>>> certificate
>>>>>>>> routines:X509_load_cert_crl_file:system lib
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG getRemoteFile end.
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey start
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG Failed to read
>>>>>>>> /root/.ssh/authorized_keys
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey: failed
to chmod
>>>>>>>> authorized_keys
>>>>>>>> Traceback (most recent call last):
>>>>>>>> File "/tmp/deployUtil.py", line 608, in
handleSSHKey
>>>>>>>> silentRestoreCon(P_ROOT_AUTH_KEYS)
>>>>>>>> File "/tmp/deployUtil.py", line 576, in
silentRestoreCon
>>>>>>>> import selinux
>>>>>>>> File
>>>>>>>>
"/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>>>> line
>>>>>>>> 26,
>>>>>>>> in<module>
>>>>>>>> _selinux = swig_import_helper()
>>>>>>>> File
>>>>>>>>
"/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>>>> line
>>>>>>>> 22,
>>>>>>>> in swig_import_helper
>>>>>>>> _mod = imp.load_module('_selinux', fp,
pathname,
>>>>>>>> description)
>>>>>>>> ImportError:
>>>>>>>> /usr/lib64/python2.7/site-packages/selinux/_selinux.so:
>>>>>>>> undefined symbol: selinux_check_access
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG handleSSHKey end
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='SetSSHAccess'
>>>>>>>> status='FAIL' message='Failed to write
server~s SSH key.'/>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR setSSHAccess test
failed
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
component='RHEV_INSTALL'
>>>>>>>> status='FAIL'/>
>>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG **** End VDS
Validation ****
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>> - add fake_kvm_support = True to your vdsm.conf
under
>>>>>>>>> /etc/vdsm/vdsm.conf
>>>>>>>>> - restart vdsmd service
>>>>>>> please make sure selinux is set at least on permissive mode;
>>>>>>>
>>>>>>> sed -i 's/SELINUX=disabled/SELINUX=permissive/g'
>>>>>>> /etc/sysconfig/selinux
>>>>>>>
>>>>>>> reboot and reinstall.
>>>>>>>
>>>>>>> anyhow, if this is the case, its a known issue and patch is
>>>>>>> pending
>>>>>>> upstream.
>>>>>> I did this, but I was also able to just re-run the installer and
>>>>>> bootstrap completed. However, now I have another issue.
>>>>>>
>>>>>> THe host is marked unresponsive in engine, engine.log shows a
>>>>>> connectivity issue, but both hosts can ping and share data.
>>>>>>
>>>>>> . Stage completed. (Stage: Running second installation script on
>>>>>> Host)
>>>>>> 2012-01-18 09:58:08,550 INFO
>>>>>> [org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper]
>>>>>> (pool-5-thread-49) RunSSHCommand returns true
>>>>>> 2012-01-18 09:58:08,550 INFO
>>>>>> [org.ovirt.engine.core.bll.VdsInstaller]
>>>>>> (pool-5-thread-49) FinishCommand ended:true
>>>>>> 2012-01-18 09:58:08,554 INFO
>>>>>> [org.ovirt.engine.core.bll.InstallVdsCommand] (pool-5-thread-49)
>>>>>> After
>>>>>> Installation pool-5-thread-49
>>>>>> 2012-01-18 09:58:08,555 INFO
>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>> (pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId =
>>>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot,
>>>>>> nonOperationalReason=NONE), log id: 703c3cbd
>>>>>> 2012-01-18 09:58:08,560 INFO
>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>> (pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id:
>>>>>> 703c3cbd
>>>>>> 2012-01-18 09:58:08,560 INFO
>>>>>> [org.ovirt.engine.core.bll.VdsCommand]
>>>>>> (pool-5-thread-50) Waiting 300 seconds, for server to finish
>>>>>> reboot
>>>>>> process.
>>>>>> 2012-01-18 10:03:08,561 INFO
>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>> (pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId =
>>>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive,
>>>>>> nonOperationalReason=NONE), log id: 3e57bdd2
>>>>>> 2012-01-18 10:03:08,570 INFO
>>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>>> (pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id:
>>>>>> 3e57bdd2
>>>>>> 2012-01-18 10:03:10,201 ERROR
>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
>>>>>> (QuartzScheduler_Worker-35) XML RPC error in command
>>>>>> GetCapabilitiesVDS (
>>>>>> Vds: ichigo-dom224 ), the error was:
>>>>>> java.util.concurrent.ExecutionException:
>>>>>> java.lang.reflect.InvocationTargetException
>>>>>>
>>>>>>
>>>>>> I can query vdsm on the on the node:
>>>>>>
>>>>>> [root@f16-node1 ~]# vdsClient -s 0 getVdsCaps
>>>>>> HBAInventory = {'iSCSI': [{'InitiatorName':
>>>>>> 'iqn.1994-05.com.redhat:2abcda43e16d'}],
'FC': []}
>>>>>> ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d
>>>>>> bondings = {'bond4': {'hwaddr':
'00:00:00:00:00:00', 'cfg':
>>>>>> {},
>>>>>> 'netmask': '', 'addr': '',
'slaves': []}, 'bond0': {'hwaddr':
>>>>>> '00:00:00:00:00:00', 'cfg': {},
'netmask': '', 'addr': '',
>>>>>> 'slaves':
>>>>>> []}, 'bond1': {'hwaddr':
'00:00:00:00:00:00', 'cfg': {},
>>>>>> 'netmask':
>>>>>> '', 'addr': '', 'slaves':
[]}, 'bond2': {'hwaddr':
>>>>>> '00:00:00:00:00:00', 'cfg': {},
'netmask': '', 'addr': '',
>>>>>> 'slaves':
>>>>>> []}, 'bond3': {'hwaddr':
'00:00:00:00:00:00', 'cfg': {},
>>>>>> 'netmask':
>>>>>> '', 'addr': '', 'slaves':
[]}}
>>>>>> clusterLevels = ['3.0']
>>>>>> cpuCores = 1
>>>>>> cpuFlags =
>>>>>>
pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1
>>>>>> cpuModel = Intel(Fake) CPU
>>>>>> cpuSockets = 1
>>>>>> cpuSpeed = 2800.482
>>>>>> emulatedMachines = ['pc-0.14', 'pc',
'fedora-13', 'pc-0.13',
>>>>>> 'pc-0.12', 'pc-0.11', 'pc-0.10',
'isapc']
>>>>>> guestOverhead = 65
>>>>>> hooks = {}
>>>>>> kvmEnabled = true
>>>>>> management_ip =
>>>>>> memSize = 7988
>>>>>> networks = {'ovirtmgmt': {'addr':
'192.168.68.224', 'cfg':
>>>>>> {'DEVICE':
>>>>>> 'ovirtmgmt', 'DELAY': '0',
'BOOTPROTO': 'dhcp', 'TYPE':
>>>>>> 'Bridge',
>>>>>> 'ONBOOT': 'yes'}, 'ports':
['eth0'], 'netmask':
>>>>>> '255.255.192.0',
>>>>>> 'stp': 'off', 'gateway':
'192.168.68.1'}}
>>>>>> nics = {'eth0': {'hwaddr':
'00:FF:FE:00:00:E0', 'netmask':
>>>>>> '',
>>>>>> 'speed': 0, 'addr': ''}}
>>>>>> operatingSystem = {'release': '1',
'version': '16', 'name':
>>>>>> 'Fedora'}
>>>>>> packages2 = {'kernel': {'release':
'7.fc16.x86_64',
>>>>>> 'buildtime':
>>>>>> 1320196248.0, 'version': '3.1.0'},
'spice-server':
>>>>>> {'release':
>>>>>> '1.fc16', 'buildtime': '1321276111',
'version': '0.10.0'},
>>>>>> 'vdsm':
>>>>>> {'release': '0.fc16', 'buildtime':
'1326734129', 'version':
>>>>>> '4.9.3.1'}, 'qemu-kvm': {'release':
'3.fc16', 'buildtime':
>>>>>> '1321651456', 'version': '0.15.1'},
'libvirt': {'release':
>>>>>> '4.fc16',
>>>>>> 'buildtime': '1324326688',
'version': '0.9.6'}, 'qemu-img':
>>>>>> {'release': '3.fc16', 'buildtime':
'1321651456', 'version':
>>>>>> '0.15.1'}}
>>>>>> reservedMem = 321
>>>>>> software_revision = 0
>>>>>> software_version = 4.9
>>>>>> supportedProtocols = ['2.2', '2.3']
>>>>>> supportedRHEVMs = ['3.0']
>>>>>> uuid =
922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0
>>>>>> version_name = Snow Man
>>>>>> vlans = {}
>>>>>> vmTypes = ['kvm']
>>>>>>
>>>>>>
>>>>> can you check if problem is solved if you run iptables -F ?
>>>> It doesn't.
>>> can you also post the vdsm.log to see if the request made it and was
>>> rejected for some reason?
>> yes, I think this is the issue:
>>
>> Thread-1016::ERROR::2012-01-18
>> 11:36:50,986::SecureXMLRPCServer::73::root::(handle_error) client
>> ('192.168.68.223', 58819)
>> Traceback (most recent call last):
>> File "/usr/lib64/python2.7/SocketServer.py", line 582, in
>> process_request_thread
>> self.finish_request(request, client_address)
>> File "/usr/share/vdsm/SecureXMLRPCServer.py", line 66, in
>> finish_request
>> request.do_handshake()
>> File "/usr/lib64/python2.7/ssl.py", line 296, in do_handshake
>> self._sslobj.do_handshake()
>>
>> I have vdsm.conf with ssl=true, however, if I set ssl=false, then I
>> cannot query vdsm from the localhost client:
>>
>> [root@f16-node1 vdsm]# vdsClient -s 0 getVdsCaps
>>
>> with ssl=false, that returns connection refused.
>>
>>
> Indeed. If you want to work without ssl, you need to also change the
> "UseSecureConnectionWithServers" option_name to "false" in
vdc_options
> table in ovirt-engine database and restart jboss-as service , so it will
> query vdsm without SSL.
ok, now with ssl=false, and the database updated and jboss-as restart,
engine sees the host as up, but the local query of the capabilities
fails.
[root@f16-node1 vdsm]# vdsClient -s 0 getVdsCaps
If ssl is not enabled, do not use
-s option.
Thanks
Douglas
--
Cheers
Douglas