[Users] test day: using VM has host for vdsm

Rami Vaknin rvaknin at redhat.com
Wed Jan 18 16:46:42 UTC 2012


On 01/18/2012 06:39 PM, Ryan Harper wrote:
> * Ayal Baron<abaron at redhat.com>  [2012-01-18 10:35]:
>>
>> ----- Original Message -----
>>> * Haim Ateya<hateya at redhat.com>  [2012-01-18 10:15]:
>>>> On Wed 18 Jan 2012 06:09:46 PM IST, Ryan Harper wrote:
>>>>> * Haim Ateya<hateya at redhat.com>   [2012-01-18 08:02]:
>>>>>> On Wed 18 Jan 2012 03:48:08 PM IST, Ryan Harper wrote:
>>>>>>> * Haim Ateya<hateya at redhat.com>    [2012-01-18 07:13]:
>>>>>>>> On Wed 18 Jan 2012 02:59:01 PM IST, Ryan Harper wrote:
>>>>>>>>> I've created some f16 VMs that contain both ovirt-engine and a
>>>>>>>>> few
>>>>>>>>> to run vdsm as nodes.  When I add in the VM host into the
>>>>>>>>> engine and it
>>>>>>>>> attempts to install vdsm (even though I've already installed
>>>>>>>>> vdsm) the
>>>>>>>>> install fails because the vdsm install script is checking to
>>>>>>>>> see if the
>>>>>>>>> host has virt capabilities; since I'm not running nested KVM,
>>>>>>>>> this
>>>>>>>>> fails.  Is there a way to work around this can enable a VM to
>>>>>>>>> be a host
>>>>>>>>> in oVirt?  I had heard in the past there was a way to create
>>>>>>>>> fake VMs
>>>>>>>>> when attempting to do ovirt-engine stress testing, wondering
>>>>>>>>> if that
>>>>>>>>> might be of help here.
>>>>>>>>>
>>>>>>>>> Also, are their vdsm rpms built for RHEL6.x available?
>>>>>>>>>
>>>>>>>>> Thanks!
>>>>>>>>>
>>>>>>>> Hi Rayn,
>>>>>>>>
>>>>>>>> - login to your ovirt-engine machine
>>>>>>>> - edit
>>>>>>>> /usr/share/ovirt-engine/engine.ear/components.war/vds/vds_bootstrap.py
>>>>>>>> - comment out the following:
>>>>>>>>
>>>>>>>>     836     if not oDeploy.virtExplorer(random_num):
>>>>>>>>     837         logging.error('virtExplorer test failed')
>>>>>>>>     838         return False
>>>>>>>> - reinstall host
>>>>>>> So I'm getting further, but now the bootstrap.log has more
>>>>>>> errors below.
>>>>>>> If I follow the test day instructions, it indicates to install
>>>>>>> vdsm and
>>>>>>> includes instructions, but it's clear that ovirt-engine is
>>>>>>> configured by
>>>>>>> default to push out vdsm and install it.  If I've already
>>>>>>> configured and
>>>>>>> installed vdsm on the node is there any way to not attempting to
>>>>>>> bootstrap
>>>>>>> vdsm
>>>>>>> at all and just attempt to have it connect?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='VDS PACKAGES'
>>>>>>> status='OK' result='qemu-kvm-tools'
>>>>>>> message='qemu-kvm-tools-0.15.1-3.fc16.x86_64 '/>
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    Basic configuration found,
>>>>>>> skipping
>>>>>>> this step
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='CreateConf'
>>>>>>> status='OK'
>>>>>>> message='Basic configuration found, skipping this step'/>
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP
>>>>>>> component='CheckLocalHostname'
>>>>>>> status='OK' message='Local hostname is correct.'/>
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    Bridge ovirtmgmt not found,
>>>>>>> need to
>>>>>>> create it.
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    getAddress Entry.
>>>>>>> url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    getAddress return.
>>>>>>> address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    makeBridge begin.
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    _getMGTIface: read host name:
>>>>>>> ichigo-dom223.phx.austin.ibm.com
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    _getMGTIface: using host name
>>>>>>> ichigo-dom223.phx.austin.ibm.com strIP= 192.168.68.223
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    _getMGTIface
>>>>>>> IP=192.168.68.223
>>>>>>> strIface=engine
>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR    makeBridge found existing
>>>>>>> bridge
>>>>>>> named:
>>>>>>> engine
>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR    makeBridge errored:  out=
>>>>>>> err=None
>>>>>>> ret=None
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    makeBridge return.
>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR    addNetwork error trying to
>>>>>>> add
>>>>>>> management bridge
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetNetworking'
>>>>>>> status='FAIL' message='addNetwork error trying to add management
>>>>>>> bridge'/>
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    getAddress Entry.
>>>>>>> url=http://ichigo-dom223.phx.austin.ibm.com:8080/Components/vds/
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    getAddress return.
>>>>>>> address=ichigo-dom223.phx.austin.ibm.com port=8080
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    getRemoteFile start. IP =
>>>>>>> ichigo-dom223.phx.austin.ibm.com port = 8080 fileName =
>>>>>>> "/engine.ssh.key.txt"
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    /engine.ssh.key.txt failed in
>>>>>>> HTTPS.
>>>>>>> Retrying using HTTP.
>>>>>>> Traceback (most recent call last):
>>>>>>>    File "/tmp/deployUtil.py", line 1334, in getRemoteFile
>>>>>>>      conn.sock = getSSLSocket(sock, certPath)
>>>>>>>    File "/tmp/deployUtil.py", line 1178, in getSSLSocket
>>>>>>>      cert_reqs=ssl.CERT_REQUIRED)
>>>>>>>    File "/usr/lib64/python2.7/ssl.py", line 372, in wrap_socket
>>>>>>>      ciphers=ciphers)
>>>>>>>    File "/usr/lib64/python2.7/ssl.py", line 132, in __init__
>>>>>>>      ciphers)
>>>>>>> SSLError: [Errno 185090050] _ssl.c:340: error:0B084002:x509
>>>>>>> certificate
>>>>>>> routines:X509_load_cert_crl_file:system lib
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    getRemoteFile end.
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    handleSSHKey start
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    Failed to read
>>>>>>> /root/.ssh/authorized_keys
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    handleSSHKey: failed to chmod
>>>>>>> authorized_keys
>>>>>>> Traceback (most recent call last):
>>>>>>>    File "/tmp/deployUtil.py", line 608, in handleSSHKey
>>>>>>>      silentRestoreCon(P_ROOT_AUTH_KEYS)
>>>>>>>    File "/tmp/deployUtil.py", line 576, in silentRestoreCon
>>>>>>>      import selinux
>>>>>>>    File
>>>>>>>    "/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>>>    line
>>>>>>>    26,
>>>>>>>    in<module>
>>>>>>>      _selinux = swig_import_helper()
>>>>>>>    File
>>>>>>>    "/usr/lib64/python2.7/site-packages/selinux/__init__.py",
>>>>>>>    line
>>>>>>>    22,
>>>>>>>    in swig_import_helper
>>>>>>>      _mod = imp.load_module('_selinux', fp, pathname,
>>>>>>>      description)
>>>>>>> ImportError:
>>>>>>> /usr/lib64/python2.7/site-packages/selinux/_selinux.so:
>>>>>>> undefined symbol: selinux_check_access
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    handleSSHKey end
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='SetSSHAccess'
>>>>>>> status='FAIL' message='Failed to write server~s SSH key.'/>
>>>>>>> Wed, 18 Jan 2012 08:35:37 ERROR    setSSHAccess test failed
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG<BSTRAP component='RHEV_INSTALL'
>>>>>>> status='FAIL'/>
>>>>>>> Wed, 18 Jan 2012 08:35:37 DEBUG    **** End VDS Validation ****
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>> - add fake_kvm_support = True to your vdsm.conf under
>>>>>>>> /etc/vdsm/vdsm.conf
>>>>>>>> - restart vdsmd service
>>>>>> please make sure selinux is set at least on permissive mode;
>>>>>>
>>>>>> sed -i   's/SELINUX=disabled/SELINUX=permissive/g'
>>>>>> /etc/sysconfig/selinux
>>>>>>
>>>>>> reboot and reinstall.
>>>>>>
>>>>>> anyhow, if this is the case, its a known issue and patch is
>>>>>> pending
>>>>>> upstream.
>>>>> I did this, but I was also able to just re-run the installer and
>>>>> bootstrap completed.  However, now I have another issue.
>>>>>
>>>>> THe host is marked unresponsive in engine, engine.log shows a
>>>>> connectivity issue, but both hosts can ping and share data.
>>>>>
>>>>> . Stage completed. (Stage: Running second installation script on
>>>>> Host)
>>>>> 2012-01-18 09:58:08,550 INFO
>>>>> [org.ovirt.engine.core.utils.hostinstall.MinaInstallWrapper]
>>>>> (pool-5-thread-49) RunSSHCommand returns true
>>>>> 2012-01-18 09:58:08,550 INFO
>>>>> [org.ovirt.engine.core.bll.VdsInstaller]
>>>>> (pool-5-thread-49)  FinishCommand ended:true
>>>>> 2012-01-18 09:58:08,554 INFO
>>>>> [org.ovirt.engine.core.bll.InstallVdsCommand] (pool-5-thread-49)
>>>>> After
>>>>> Installation pool-5-thread-49
>>>>> 2012-01-18 09:58:08,555 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> (pool-5-thread-49) START, SetVdsStatusVDSCommand(vdsId =
>>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=Reboot,
>>>>> nonOperationalReason=NONE), log id: 703c3cbd
>>>>> 2012-01-18 09:58:08,560 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> (pool-5-thread-49) FINISH, SetVdsStatusVDSCommand, log id:
>>>>> 703c3cbd
>>>>> 2012-01-18 09:58:08,560 INFO
>>>>> [org.ovirt.engine.core.bll.VdsCommand]
>>>>> (pool-5-thread-50) Waiting 300 seconds, for server to finish
>>>>> reboot
>>>>> process.
>>>>> 2012-01-18 10:03:08,561 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> (pool-5-thread-50) START, SetVdsStatusVDSCommand(vdsId =
>>>>> 8c627fa8-41d8-11e1-8d2f-00fffe0000df, status=NonResponsive,
>>>>> nonOperationalReason=NONE), log id: 3e57bdd2
>>>>> 2012-01-18 10:03:08,570 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> (pool-5-thread-50) FINISH, SetVdsStatusVDSCommand, log id:
>>>>> 3e57bdd2
>>>>> 2012-01-18 10:03:10,201 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand]
>>>>> (QuartzScheduler_Worker-35) XML RPC error in command
>>>>> GetCapabilitiesVDS (
>>>>> Vds: ichigo-dom224 ), the error was:
>>>>> java.util.concurrent.ExecutionException:
>>>>> java.lang.reflect.InvocationTargetException
>>>>>
>>>>>
>>>>> I can query vdsm on the on the node:
>>>>>
>>>>> [root at f16-node1 ~]# vdsClient -s 0 getVdsCaps
>>>>>      HBAInventory = {'iSCSI': [{'InitiatorName':
>>>>>      'iqn.1994-05.com.redhat:2abcda43e16d'}], 'FC': []}
>>>>>      ISCSIInitiatorName = iqn.1994-05.com.redhat:2abcda43e16d
>>>>>      bondings = {'bond4': {'hwaddr': '00:00:00:00:00:00', 'cfg':
>>>>>      {},
>>>>>      'netmask': '', 'addr': '', 'slaves': []}, 'bond0': {'hwaddr':
>>>>>      '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '',
>>>>>      'slaves':
>>>>>      []}, 'bond1': {'hwaddr': '00:00:00:00:00:00', 'cfg': {},
>>>>>      'netmask':
>>>>>      '', 'addr': '', 'slaves': []}, 'bond2': {'hwaddr':
>>>>>      '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '',
>>>>>      'slaves':
>>>>>      []}, 'bond3': {'hwaddr': '00:00:00:00:00:00', 'cfg': {},
>>>>>      'netmask':
>>>>>      '', 'addr': '', 'slaves': []}}
>>>>>      clusterLevels = ['3.0']
>>>>>      cpuCores = 1
>>>>>      cpuFlags =
>>>>>      pge,clflush,sep,syscall,tsc,vmx,cmov,nx,constant_tsc,pat,sse4_1,lm,msr,fpu,fxsr,pae,nopl,mmx,cx8,mce,de,mca,pse,pni,popcnt,apic,sse,sse4_2,lahf_lm,sse2,hypervisor,up,ssse3,cx16,pse36,mtrr,x2apicmodel_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Opteron_G1
>>>>>      cpuModel = Intel(Fake) CPU
>>>>>      cpuSockets = 1
>>>>>      cpuSpeed = 2800.482
>>>>>      emulatedMachines = ['pc-0.14', 'pc', 'fedora-13', 'pc-0.13',
>>>>>      'pc-0.12', 'pc-0.11', 'pc-0.10', 'isapc']
>>>>>      guestOverhead = 65
>>>>>      hooks = {}
>>>>>      kvmEnabled = true
>>>>>      management_ip =
>>>>>      memSize = 7988
>>>>>      networks = {'ovirtmgmt': {'addr': '192.168.68.224', 'cfg':
>>>>>      {'DEVICE':
>>>>>      'ovirtmgmt', 'DELAY': '0', 'BOOTPROTO': 'dhcp', 'TYPE':
>>>>>      'Bridge',
>>>>>      'ONBOOT': 'yes'}, 'ports': ['eth0'], 'netmask':
>>>>>      '255.255.192.0',
>>>>>      'stp': 'off', 'gateway': '192.168.68.1'}}
>>>>>      nics = {'eth0': {'hwaddr': '00:FF:FE:00:00:E0', 'netmask':
>>>>>      '',
>>>>>      'speed': 0, 'addr': ''}}
>>>>>      operatingSystem = {'release': '1', 'version': '16', 'name':
>>>>>      'Fedora'}
>>>>>      packages2 = {'kernel': {'release': '7.fc16.x86_64',
>>>>>      'buildtime':
>>>>>      1320196248.0, 'version': '3.1.0'}, 'spice-server':
>>>>>      {'release':
>>>>>      '1.fc16', 'buildtime': '1321276111', 'version': '0.10.0'},
>>>>>      'vdsm':
>>>>>      {'release': '0.fc16', 'buildtime': '1326734129', 'version':
>>>>>      '4.9.3.1'}, 'qemu-kvm': {'release': '3.fc16', 'buildtime':
>>>>>      '1321651456', 'version': '0.15.1'}, 'libvirt': {'release':
>>>>>      '4.fc16',
>>>>>      'buildtime': '1324326688', 'version': '0.9.6'}, 'qemu-img':
>>>>>      {'release': '3.fc16', 'buildtime': '1321651456', 'version':
>>>>>      '0.15.1'}}
>>>>>      reservedMem = 321
>>>>>      software_revision = 0
>>>>>      software_version = 4.9
>>>>>      supportedProtocols = ['2.2', '2.3']
>>>>>      supportedRHEVMs = ['3.0']
>>>>>      uuid = 922F4AE6-8EEA-4B11-44C4-EA1E1D665AC2_00:FF:FE:00:00:E0
>>>>>      version_name = Snow Man
>>>>>      vlans = {}
>>>>>      vmTypes = ['kvm']
>>>>>
>>>>>
>>>> can you check if problem is solved if you run iptables -F ?
>>> It doesn't.
>> can you also post the vdsm.log to see if the request made it and was rejected for some reason?
> yes, I think this is the issue:
>
> Thread-1016::ERROR::2012-01-18
> 11:36:50,986::SecureXMLRPCServer::73::root::(handle_error) client ('192.168.68.223', 58819)
> Traceback (most recent call last):
>       File "/usr/lib64/python2.7/SocketServer.py", line 582, in process_request_thread
>           self.finish_request(request, client_address)
>       File "/usr/share/vdsm/SecureXMLRPCServer.py", line 66, in finish_request
>           request.do_handshake()
>       File "/usr/lib64/python2.7/ssl.py", line 296, in do_handshake
>           self._sslobj.do_handshake()
>
> I have vdsm.conf with ssl=true, however, if I set ssl=false, then I
> cannot query vdsm from the localhost client:
>
> [root at f16-node1 vdsm]# vdsClient -s 0 getVdsCaps
>
> with ssl=false, that returns connection refused.
>
>
Indeed. If you want to work without ssl, you need to also change the 
"UseSecureConnectionWithServers" option_name to "false" in vdc_options 
table in ovirt-engine database and restart jboss-as service , so it will 
query vdsm without SSL.


-- 

Thanks,

Rami Vaknin, QE @ Red Hat, TLV, IL.




More information about the Users mailing list