[Users] OVirt-Engine 3.3 RC - Add Fedora 19 host fails
Einav Cohen
ecohen at redhat.com
Wed Sep 4 15:46:05 EDT 2013
> ----- Original Message -----
> From: "Markus Stockhausen" <stockhausen at collogia.de>
> Sent: Wednesday, September 4, 2013 3:19:52 PM
>
> Hello Einav,
>
> thanks for the tip. In ovirt-engine I mapped the ovirtmgmt network to
> interface p49p1 checked the save settings radiobutton and clicked ok.
> The connection was lost again. Now I have:
>
> [root at colovn1 network-scripts]# cat ifcfg-ovirtmgmt
> # Generated by VDSM version 4.12.1-1.fc19
> DEVICE=ovirtmgmt
> ONBOOT=yes
> TYPE=Bridge
> DELAY=0
> DEFROUTE=yes
> NM_CONTROLLED=no
> STP=no
> BOOTPROTO=static
>
> [root at colovn1 network-scripts]# cat ifcfg-p49p1
> # Generated by VDSM version 4.12.1-1.fc19
> DEVICE=p49p1
> ONBOOT=yes
> HWADDR=00:30:48:d7:9e:c4
> BRIDGE=ovirtmgmt
> MTU=1500
> NM_CONTROLLED=no
> STP=no
>
> and what seems to be the old config file:
>
> [root at colovn1 network-scripts]# cat ifcfg-enp1s0
> PEERROUTES="yes"
> IPV6INIT="yes"
> UUID="0bd1b6df-9f88-4dcf-82af-225b14b99d67"
> IPV6_PEERDNS="yes"
> DEFROUTE="yes"
> PEERDNS="yes"
> IPV4_FAILURE_FATAL="no"
> HWADDR="00:30:48:D7:9E:C4"
> BOOTPROTO="none"
> IPV6_DEFROUTE="yes"
> IPV6_AUTOCONF="yes"
> IPV6_FAILURE_FATAL="no"
> IPV6_PEERROUTES="yes"
> TYPE="Ethernet"
> ONBOOT="yes"
> NAME="enp1s0"
> IPADDR="192.168.10.51"
> NETMASK="255.255.255.0"
> GATEWAY="192.168.10.7"
> DNS1="192.168.10.15"
>
> Afterwards I disabled Networkmanager:
>
> [root at colovn1 network-scripts]# systemctl disable NetworkManager
> [root at colovn1 network-scripts]# systemctl stop NetworkManager.service
>
> Now I have no configuration at all. At least I can rech the
> machine through the storage network. Any idea what to
> try next?
make sure that your network service is on ("service network start")
and that it will be automatically started after reboot ("chkconfig
network on").
>
> Markus
>
> ________________________________________
> Von: Einav Cohen [ecohen at redhat.com]
> Gesendet: Mittwoch, 4. September 2013 20:58
> An: Markus Stockhausen
> Cc: users
> Betreff: Re: [Users] OVirt-Engine 3.3 RC - Add Fedora 19 host fails
>
> hi Markus,
>
> during the ovirt 3.3 test day, there was a known issue with regards to the
> ovirtmgmt network - see the 'host installation' note in the Known Issues
> section in [1] - I am not sure if this issue has already been resolved or
> not.
> can you please follow the recommended steps in [1] and see if it helps?
>
> ----
> Thanks,
> Einav
>
> [1] http://www.ovirt.org/OVirt_3.3_TestDay#Known_issues
>
>
> ----- Original Message -----
> > From: "Markus Stockhausen" <stockhausen at collogia.de>
> > To: "users" <users at ovirt.org>
> > Sent: Wednesday, September 4, 2013 2:44:28 PM
> > Subject: [Users] OVirt-Engine 3.3 RC - Add Fedora 19 host fails
> >
> > Hello,
> >
> > I'm trying to add a Fedora 19 host to my newly installed ovirt engine 3.3
> > The ost has vdsm.x86_64 4.12.1-1.fc19 installed from the beta repos.
> > During the installation two interfaces are active:
> >
> > Storage/NFS network:
> > ib0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2044
> > inet 10.10.30.1 netmask 255.255.255.0 broadcast 10.10.30.255
> > Infiniband hardware address can be incorrect! Please read BUGS section in
> > ifconfig(8).
> > infiniband
> > 80:00:00:48:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00
> > txqueuelen 256 (InfiniBand)
> > RX packets 38899 bytes 2213697 (2.1 MiB)
> > RX errors 0 dropped 0 overruns 0 frame 0
> > TX packets 67314 bytes 15746759 (15.0 MiB)
> > TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
> >
> > Frontend/Management network:
> > p49p1: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
> > inet 192.168.10.51 netmask 255.255.255.0 broadcast 192.168.10.255
> > ether 00:30:48:d7:9e:c4 txqueuelen 1000 (Ethernet)
> > RX packets 14631 bytes 9069189 (8.6 MiB)
> > RX errors 0 dropped 354 overruns 0 frame 0
> > TX packets 6552 bytes 2348866 (2.2 MiB)
> > TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
> > device interrupt 16 memory 0xface0000-fad00000
> >
> > The installation (with IP 192.168.10.51) fails and afterwards the frontend
> > interface
> > is down. Attached the VDSM logs.
> >
> > Thanks for your help in advance.
> >
> > Markus
> >
> > supervdsm.log:
> >
> > MainProcess|Thread-16::ERROR::2013-09-04
> > 20:37:32,695::netinfo::239::root::(speed) cannot read ib0 speed
> > Traceback (most recent call last):
> > File "/usr/lib64/python2.7/site-packages/vdsm/netinfo.py", line 235, in
> > speed
> > s = int(file('/sys/class/net/%s/speed' % dev).read())
> > IOError: [Errno 22] Invalid argument
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,698::configNetwork::176::root::(addNetwork) validating network...
> > MainProcess|Thread-16::INFO::2013-09-04
> > 20:37:32,698::configNetwork::191::root::(addNetwork) Adding network
> > ovirtmgmt with vlan=None, bonding=None, nics=['p49p1'],
> > bondingOptions=None,
> > mtu=None, bridged=True, options={'STP': 'no', 'implicitBonding': True}
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,699::ifcfg::366::root::(_persistentBackup) backing up
> > ifcfg-ovirtmgmt: # original file did not exist
> >
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,702::ifcfg::665::Storage.Misc.excCmd::(ifdown) '/usr/sbin/ifdown
> > ovirtmgmt' (cwd None)
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,917::ifcfg::665::Storage.Misc.excCmd::(ifdown) SUCCESS: <err> =
> > '';
> > <rc> = 0
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,917::utils::486::root::(execCmd) '/sbin/ip route show to 0.0.0.0/0
> > table all' (cwd None)
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,919::utils::505::root::(execCmd) SUCCESS: <err> = ''; <rc> = 0
> > MainProcess|Thread-16::ERROR::2013-09-04
> > 20:37:32,923::netinfo::239::root::(speed) cannot read ib0 speed
> > Traceback (most recent call last):
> > File "/usr/lib64/python2.7/site-packages/vdsm/netinfo.py", line 235, in
> > speed
> > s = int(file('/sys/class/net/%s/speed' % dev).read())
> > IOError: [Errno 22] Invalid argument
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,926::ifcfg::366::root::(_persistentBackup) backing up ifcfg-p49p1:
> > # original file did not exist
> >
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:32,926::ifcfg::665::Storage.Misc.excCmd::(ifdown) '/usr/sbin/ifdown
> > p49p1' (cwd None)
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,243::ifcfg::665::Storage.Misc.excCmd::(ifdown) SUCCESS: <err> =
> > 'bridge ovirtmgmt does not exist!\n'; <rc> = 0
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,243::ifcfg::672::Storage.Misc.excCmd::(_ifup) '/usr/sbin/ifup
> > p49p1' (cwd None)
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,354::ifcfg::672::Storage.Misc.excCmd::(_ifup) SUCCESS: <err> = '';
> > <rc> = 0
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,354::__init__::78::root::(_addSourceRoute) Adding source route
> > ovirtmgmt, None, None, None
> > MainProcess|Thread-16::ERROR::2013-09-04
> > 20:37:33,354::sourceRoute::68::root::(configure) ipaddr, mask or gateway
> > not
> > received
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,355::ifcfg::672::Storage.Misc.excCmd::(_ifup) '/usr/sbin/ifup
> > ovirtmgmt' (cwd None)
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,731::ifcfg::672::Storage.Misc.excCmd::(_ifup) SUCCESS: <err> = '';
> > <rc> = 0
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,733::libvirtconnection::101::libvirtconnection::(wrapper) Unknown
> > libvirterror: ecode: 43 edom: 19 level: 2 message: Netzwerk nicht gefunden:
> > Kein Netzwerk mit übereinstimmendem Namen »vdsm-ovirtmgmt«
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,733::ifcfg::245::root::(_atomicNetworkBackup) Backed up ovirtmgmt
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,734::libvirtconnection::101::libvirtconnection::(wrapper) Unknown
> > libvirterror: ecode: 43 edom: 19 level: 2 message: Netzwerk nicht gefunden:
> > Kein Netzwerk mit übereinstimmendem Namen »vdsm-ovirtmgmt«
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,734::ifcfg::254::root::(_persistentNetworkBackup) backing up
> > network ovirtmgmt: # original file did not exist
> >
> > MainProcess|Thread-16::DEBUG::2013-09-04
> > 20:37:33,741::configNetwork::541::setupNetworks::(setupNetworks) Checking
> > connectivity...
> > MainProcess|storageRefresh::DEBUG::2013-09-04
> > 20:37:34,278::misc::817::SamplingMethod::(__call__) Returning last result
> >
> > vdsm.log:
> >
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:32,255::misc::817::SamplingMethod::(__call__) Returning last result
> > Thread-13::DEBUG::2013-09-04
> > 20:37:32,365::BindingXMLRPC::979::vds::(wrapper)
> > client [192.168.10.110]::call ping with () {}
> > Thread-13::DEBUG::2013-09-04
> > 20:37:32,366::BindingXMLRPC::986::vds::(wrapper)
> > return ping with {'status': {'message': 'Done', 'code': 0}}
> > Thread-14::DEBUG::2013-09-04
> > 20:37:32,374::BindingXMLRPC::979::vds::(wrapper)
> > client [192.168.10.110]::call getCapabilities with () {}
> > Thread-14::DEBUG::2013-09-04 20:37:32,562::utils::486::root::(execCmd)
> > '/sbin/ip route show to 0.0.0.0/0 table all' (cwd None)
> > Thread-14::DEBUG::2013-09-04 20:37:32,573::utils::505::root::(execCmd)
> > SUCCESS: <err> = ''; <rc> = 0
> > Thread-14::ERROR::2013-09-04 20:37:32,577::netinfo::239::root::(speed)
> > cannot
> > read ib0 speed
> > Traceback (most recent call last):
> > File "/usr/lib64/python2.7/site-packages/vdsm/netinfo.py", line 235, in
> > speed
> > s = int(file('/sys/class/net/%s/speed' % dev).read())
> > IOError: [Errno 22] Invalid argument
> > Thread-14::DEBUG::2013-09-04
> > 20:37:32,611::BindingXMLRPC::986::vds::(wrapper)
> > return getCapabilities with {'status': {'message': 'Done', 'code': 0},
> > 'info': {'HBAInventory': {'iSCSI': [{'InitiatorName':
> > 'iqn.1994-05.com.redhat:df62bd8ee7d8'}], 'FC': []}, 'packages2': {'kernel':
> > {'release': '200.fc19.x86_64', 'buildtime': 1377795945.0, 'version':
> > '3.10.10'}, 'spice-server': {'release': '1.fc19', 'buildtime': 1375454091L,
> > 'version': '0.12.4'}, 'vdsm': {'release': '1.fc19', 'buildtime':
> > 1377639886L, 'version': '4.12.1'}, 'qemu-kvm': {'release': '7.fc19',
> > 'buildtime': 1376836471L, 'version': '1.4.2'}, 'libvirt': {'release':
> > '1.fc19', 'buildtime': 1375400611L, 'version': '1.0.5.5'}, 'qemu-img':
> > {'release': '7.fc19', 'buildtime': 1376836471L, 'version': '1.4.2'}, 'mom':
> > {'release': '3.fc19', 'buildtime': 1375215820L, 'version': '0.3.2'}},
> > 'cpuModel': 'Intel(R) Core(TM) i7 CPU 920 @ 2.67GHz', 'hooks': {},
> > 'cpuSockets': '1', 'vmTypes': ['kvm'], 'supportedProtocols': ['2.2',
> > '2.3'],
> > 'networks': {}, 'bridges': {}, 'uuid':
> > '49434D53-0200-48D7-3000-D7483000C49E', 'lastClientIface': 'p49p1', 'nics':
> > {'ib0': {'netmask': '255.255.255.0', 'addr': '10.10.30.1', 'hwaddr':
> > '80:00:00:48:fe:80:00:00:00:00:00:00:00:23:7d:ff:ff:94:d3:fd', 'cfg':
> > {'NETWORK': '10.10.30.0', 'IPADDR': '10.10.30.1', 'BROADCAST':
> > '10.10.30.255', 'NETMASK': '255.255.255.0', 'BOOTPROTO': 'static',
> > 'DEVICE':
> > 'ib0', 'TYPE': 'Infiniband', 'ONBOOT': 'yes'}, 'ipv6addrs': [], 'speed': 0,
> > 'mtu': '2044'}, 'ib1': {'netmask': '', 'addr': '', 'hwaddr':
> > '80:00:00:49:fe:80:00:00:00:00:00:00:00:23:7d:ff:ff:94:d3:fe', 'cfg': {},
> > 'ipv6addrs': [], 'speed': 0, 'mtu': '4092'}, 'p50p1': {'netmask': '',
> > 'addr': '', 'hwaddr': '00:30:48:d7:9e:c5', 'cfg': {}, 'ipv6addrs': [],
> > 'speed': 0, 'mtu': '1500'}, 'p49p1': {'netmask': '255.255.255.0', 'addr':
> > '192.168.10.51', 'hwaddr': '00:30:48:d7:9e:c4', 'cfg': {}, 'ipv6addrs': [],
> > 'speed': 1000, 'mtu': '1500'}}, 'software_revision': '1', 'clusterLevels':
> > ['3.0', '3.1', '3.2', '3.3'], 'cpuFlags':
> > u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,nopl,xtopology,nonstop_tsc,aperfmperf,pni,dtes64,monitor,ds_cpl,vmx,est,tm2,ssse3,cx16,xtpr,pdcm,sse4_1,sse4_2,popcnt,lahf_lm,ida,dtherm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_n270',
> > 'ISCSIInitiatorName': 'iqn.1994-05.com.redhat:df62bd8ee7d8',
> > 'netConfigDirty': 'False', 'supportedENGINEs': ['3.0', '3.1', '3.2',
> > '3.3'],
> > 'reservedMem': '321', 'bondings': {'bond0': {'netmask': '', 'addr': '',
> > 'slaves': [], 'hwaddr': 'e6:a8:3e:20:57:b0', 'cfg': {}, 'ipv6addrs': [],
> > 'mtu': '1500'}}, 'software_version': '4.12', 'memSize': '24104',
> > 'cpuSpeed':
> > '2668.000', 'version_name': 'Snow Man', 'vlans': {}, 'cpuCores': '4',
> > 'kvmEnabled': 'true', 'guestOverhead': '65', 'management_ip': '0.0.0.0',
> > 'cpuThreads': '8', 'emulatedMachines': [u'pc', u'q35', u'isapc',
> > u'pc-0.10',
> > u'pc-0.11', u'pc-0.12', u'pc-0.13', u'pc-0.14', u'pc-0.15', u'pc-1.0',
> > u'pc-1.1', u'pc-1.2', u'pc-1.3', u'none'], 'operatingSystem': {'release':
> > '3', 'version': '19', 'name': 'Fedora'}, 'lastClient': '192.168.10.110'}}
> > Thread-15::DEBUG::2013-09-04
> > 20:37:32,640::BindingXMLRPC::979::vds::(wrapper)
> > client [192.168.10.110]::call ping with () {}
> > Thread-15::DEBUG::2013-09-04
> > 20:37:32,640::BindingXMLRPC::986::vds::(wrapper)
> > return ping with {'status': {'message': 'Done', 'code': 0}}
> > Thread-16::DEBUG::2013-09-04
> > 20:37:32,642::BindingXMLRPC::979::vds::(wrapper)
> > client [192.168.10.110]::call setupNetworks with ({'ovirtmgmt': {'nic':
> > 'p49p1', 'STP': 'no', 'bridged': 'true'}}, {}, {'connectivityCheck':
> > 'true',
> > 'connectivityTimeout': 120}) {}
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,278::multipath::111::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo
> > -n /sbin/multipath' (cwd None)
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,315::multipath::111::Storage.Misc.excCmd::(rescan) SUCCESS: <err>
> > =
> > ''; <rc> = 0
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,315::lvm::483::OperationMutex::(_invalidateAllPvs) Operation 'lvm
> > invalidate operation' got the operation mutex
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,315::lvm::485::OperationMutex::(_invalidateAllPvs) Operation 'lvm
> > invalidate operation' released the operation mutex
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,315::lvm::494::OperationMutex::(_invalidateAllVgs) Operation 'lvm
> > invalidate operation' got the operation mutex
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,316::lvm::496::OperationMutex::(_invalidateAllVgs) Operation 'lvm
> > invalidate operation' released the operation mutex
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,316::lvm::514::OperationMutex::(_invalidateAllLvs) Operation 'lvm
> > invalidate operation' got the operation mutex
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,316::lvm::516::OperationMutex::(_invalidateAllLvs) Operation 'lvm
> > invalidate operation' released the operation mutex
> > storageRefresh::DEBUG::2013-09-04
> > 20:37:34,316::misc::817::SamplingMethod::(__call__) Returning last result
> > storageRefresh::WARNING::2013-09-04
> > 20:37:34,316::fileUtils::167::Storage.fileUtils::(createdir) Dir
> > /rhev/data-center/hsm-tasks already exists
> > Thread-16::ERROR::2013-09-04 20:39:36,278::API::1261::vds::(setupNetworks)
> > connectivity check failed
> > Traceback (most recent call last):
> > File "/usr/share/vdsm/API.py", line 1259, in setupNetworks
> > supervdsm.getProxy().setupNetworks(networks, bondings, options)
> > File "/usr/share/vdsm/supervdsm.py", line 50, in __call__
> > return callMethod()
> > File "/usr/share/vdsm/supervdsm.py", line 48, in <lambda>
> > **kwargs)
> > File "<string>", line 2, in setupNetworks
> > File "/usr/lib64/python2.7/multiprocessing/managers.py", line 773, in
> > _callmethod
> > raise convert_to_error(kind, result)
> > ConfigNetworkError: (10, 'connectivity check failed')
> > Thread-16::DEBUG::2013-09-04
> > 20:39:36,280::BindingXMLRPC::986::vds::(wrapper)
> > return setupNetworks with {'status': {'message': 'connectivity check
> > failed', 'code': 10}}
> >
> >
> > _______________________________________________
> > Users mailing list
> > Users at ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
More information about the Users
mailing list