Hello,

Thanks for having a look at this issue.

On ovirt-engine - either of the command did not work. 

Here is the output from the hypervisor.

[root@h1 ~]# vdsClient -s 0 getVdsCaps
HBAInventory = {'iSCSI': [{'InitiatorName': 'iqn.1994-05.com.redhat:5e25411e9039'}], 'FC': [{'wwpn': '10000000c9cb2826', 'wwnn': '20000000c9cb2826', 'model': 'LPe1205-CIOv - Emulex LPe1205-CIOv 8Gb 2-port PCIe Fibre Channel Adapter'}, {'wwpn': '10000000c9cb2827', 'wwnn': '20000000c9cb2827', 'model': 'LPe1205-CIOv - Emulex LPe1205-CIOv 8Gb 2-port PCIe Fibre Channel Adapter'}]}
ISCSIInitiatorName = iqn.1994-05.com.redhat:5e25411e9039
bondings = {'bond4': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond0': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond1': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond2': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}, 'bond3': {'hwaddr': '00:00:00:00:00:00', 'cfg': {}, 'netmask': '', 'addr': '', 'slaves': []}}
clusterLevels = ['3.0']
cpuCores = 12
cpuFlags = fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,aperfmperf,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,arat,epb,dts,tpr_shadow,vnmi,flexpriority,ept,vpid,model_486,model_pentium,model_pentium2,model_pentium3,model_pentiumpro,model_qemu32,model_coreduo,model_core2duo,model_n270,model_Conroe,model_Penryn,model_Nehalem,model_Westmere,model_Opteron_G1
cpuModel = Intel(R) Xeon(R) CPU           E5645  @ 2.40GHz
cpuSockets = 2
cpuSpeed = 2400.276
emulatedMachines = ['rhel6.2.0', 'pc', 'rhel6.1.0', 'rhel6.0.0', 'rhel5.5.0', 'rhel5.4.4', 'rhel5.4.0']
guestOverhead = 65
hooks = {'before_vm_start': {'50_vhostmd': {'md5': '1635e45584995d9ad2629939cb0bcdbf'}}, 'after_vm_destroy': {'50_vhostmd': {'md5': '7e7509430304099f38e4a9de966af090'}}, 'before_vm_dehibernate': {'50_vhostmd': {'md5': '1635e45584995d9ad2629939cb0bcdbf'}}, 'before_vm_migrate_destination': {'50_vhostmd': {'md5': '1635e45584995d9ad2629939cb0bcdbf'}}}
kvmEnabled = true
lastClient = 172.29.5.200
lastClientIface = rhevm
management_ip = 172.29.81.141
memSize = 64420
networks = {'rhevm': {'addr': '172.29.81.141', 'cfg': {'IPV6FORWARDING': 'no', 'IPV6INIT': 'no', 'SKIPLIBVIRT': 'True', 'IPADDR': '172.29.81.141', 'ONBOOT': 'yes', 'DELAY': '0', 'DEVICE': 'rhevm', 'NETMASK': '255.255.255.0', 'BOOTPROTO': 'static', 'IPV6_AUTOCONF': 'no', 'PEERNTP': 'yes', 'TYPE': 'Bridge', 'GATEWAY': '172.29.81.1'}, 'ports': ['eth8'], 'netmask': '255.255.255.0', 'stp': 'off', 'gateway': '172.29.81.1'}}
nics = {'eth9': {'hwaddr': '5C:F3:FC:98:19:82', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth8': {'hwaddr': '5C:F3:FC:98:19:80', 'netmask': '', 'speed': 1000, 'addr': ''}, 'eth7': {'hwaddr': '00:00:C9:D2:16:73', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth6': {'hwaddr': '00:00:C9:D2:16:6F', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth5': {'hwaddr': '00:00:C9:D2:16:72', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth4': {'hwaddr': '00:00:C9:D2:16:6E', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth3': {'hwaddr': '00:00:C9:D2:16:71', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth2': {'hwaddr': '00:00:C9:D2:16:6D', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth1': {'hwaddr': '00:00:C9:D2:16:70', 'netmask': '', 'speed': 0, 'addr': ''}, 'eth0': {'hwaddr': '00:00:C9:D2:16:6C', 'netmask': '', 'speed': 0, 'addr': ''}}
operatingSystem = {'release': '20120510.0.el6_2', 'version': '6.2', 'name': 'RHEV Hypervisor'}
packages2 = {'kernel': {'release': '220.17.1.el6.x86_64', 'buildtime': 1335447433.0, 'version': '2.6.32'}, 'spice-server': {'release': '5.el6', 'buildtime': '1319663438', 'version': '0.8.2'}, 'vdsm': {'release': '112.12.el6_2', 'buildtime': '1334495504', 'version': '4.9'}, 'qemu-kvm': {'release': '2.209.el6_2.5', 'buildtime': '1334776518', 'version': '0.12.1.2'}, 'libvirt': {'release': '23.el6_2.8', 'buildtime': '1333958926', 'version': '0.9.4'}, 'qemu-img': {'release': '2.209.el6_2.5', 'buildtime': '1334776518', 'version': '0.12.1.2'}}
reservedMem = 256
software_revision = 112.12
software_version = 4.9
supportedProtocols = ['2.2', '2.3']
supportedRHEVMs = ['3.0']
uuid = C12E325E-3D69-11E1-91C1-5CF3FC981980_00:00:C9:D2:16:6C
version_name = Snow Man
vlans = {}
vmTypes = ['kvm']
[root@h1 ~]#

Regards,
Keyur



On Fri, Nov 9, 2012 at 7:36 PM, Yair Zaslavsky <yzaslavs@redhat.com> wrote:
For troubleshooting -
Can you run on your host machine

vdsClient 0 getVdsCaps (might be vdsClient -s 0 getVdsCaps)

and send us the results?



From: "Keyur Makwana" <kurmakwana@gmail.com>
To: users@ovirt.org
Sent: Friday, November 9, 2012 12:50:22 PM
Subject: [Users] Ovirt Host (3.0) compatability issue with Cluster (3.1)


Hello All,

I am new to Ovirt and trying to Implement Ovirt in my organisation with an view to implement complete open source virtualization solution instead of using VMware or other Paid once.

I have installed ovirt-engine on Fedora 17 and have installed Ovirt node (Hypervisor) - Red Hat Enterprise Virtualization Hypervisor release 6.2 (20120510.0.el6_2)


Post installation engine started fine but while trying to add host its giving me error as below.

Host host1 is compatible with versions (3.0) and cannot join Cluster Default which is set to version 3.1.

I don't know what exact details I would need to share for troubleshooting. Hence I have attached a file with details of rpms installed on ovirt-engine and node. ( One may not like to scroll the entire file)

I might be missing something very silly or I may sound too stupid. 

Ovirt masters : Please guide.

Thanks in advance,
Keyur


_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users