[ovirt-devel] latest python-lxml python-ethtool break VDSM on f19
Roy Golan
rgolan at redhat.com
Mon May 12 06:13:01 UTC 2014
FYI and please guide Gianluca against what to open a BZ
-------- Original Message --------
Subject: Re: [ovirt-users] getVdsCapabilites unexpected exception [was:
Re: AIO 3.4 on fedora 19 initial errors before coming up]
Date: Sun, 11 May 2014 23:49:06 +0200
From: Gianluca Cecchi <gianluca.cecchi at gmail.com>
To: Roy Golan <rgolan at redhat.com>
CC: users <users at ovirt.org>
On Sun, May 11, 2014 at 8:41 PM, Gianluca Cecchi
<gianluca.cecchi at gmail.com <mailto:gianluca.cecchi at gmail.com>> wrote:
On Sun, May 11, 2014 at 8:18 PM, Gianluca Cecchi
<gianluca.cecchi at gmail.com <mailto:gianluca.cecchi at gmail.com>> wrote:
On Sun, May 11, 2014 at 4:10 PM, Roy Golan <rgolan at redhat.com
<mailto:rgolan at redhat.com>> wrote:
The vm will stay in "Waiting.." as the getVdsCaps is failing
and the monitoring of Vms will not take place.
need to fix this "Unexpected error" first. is it a matter of
ssl enabled configuration for host communication? i.e. can
you try vdsClient -s 0 getVdsCaps ?
_______________________________________________
Users mailing list
Users at ovirt.org <mailto:Users at ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users
I didn't change anything on this server.
It is an all-in-one config so both engine and vdsm host are on it.
Yesterday and every previous day I was able to start the system
and start the VM; I only applied the yum update command yeseterday
(with --exclude=sos due to the opened bug) and then I made
shutdown of the system.
Today after startup I got this problem.
[root at tekkaman vdsm]# vdsClient -s 0 getVdsCaps
Unexpected exception
it seems the error in vdsm.log when I run the command above is
of this type:
Thread-25::ERROR::2014-05-11
20:18:02,202::BindingXMLRPC::1086::vds::(wrapper) unexpected error
Traceback (most recent call last):
File "/usr/share/vdsm/BindingXMLRPC.py", line 1070, in wrapper
res = f(*args, **kwargs)
File "/usr/share/vdsm/BindingXMLRPC.py", line 393, in
getCapabilities
ret = api.getCapabilities()
File "/usr/share/vdsm/API.py", line 1185, in getCapabilities
c = caps.get()
File "/usr/share/vdsm/caps.py", line 369, in get
caps.update(netinfo.get())
File "/usr/lib64/python2.7/site-packages/vdsm/netinfo.py",
line 557, in get
netAttr.get('qosOutbound'))
File "/usr/lib64/python2.7/site-packages/vdsm/netinfo.py",
line 487, in _getNetInfo
ipv4addr, ipv4netmask, ipv6addrs = getIpInfo(iface)
File "/usr/lib64/python2.7/site-packages/vdsm/netinfo.py",
line 317, in getIpInfo
ipv6addrs = devInfo.get_ipv6_addresses()
SystemError: error return without exception set
Based on above errors, I think that for some reason these two python
related packages that were updated yesterday are causing some
problems with vdsm.
Can you confirm that you can run ok the 3.4 vdsm with those?
vdsm-4.14.6-0.fc19.x86_64
May 10 21:24:23 Updated: python-ethtool-0.9-2.fc19.x86_64
May 10 21:24:23 Updated: python-lxml-3.3.5-1.fc19.x86_64
I can also try to rollback and see...
I was right.
Against what to bugzilla?
This is a show stopper for fedora 19 ovirt users...
[root at tekkaman log]# vdsClient -s 0 getVdsCaps
Unexpected exception
[root at tekkaman log]# yum downgrade python-lxml python-ethtool
Loaded plugins: fastestmirror, langpacks, refresh-packagekit, versionlock
Dropbox | 951 B 00:00:00
adobe-linux-x86_64 | 951 B 00:00:00
fedora-virt-preview | 2.9 kB 00:00:00
google-chrome | 951 B 00:00:00
livna | 1.3 kB 00:00:00
ovirt-3.3.3 | 2.9 kB 00:00:00
ovirt-stable | 2.9 kB 00:00:00
rpmfusion-free-updates | 3.3 kB 00:00:00
rpmfusion-nonfree-updates | 3.3 kB 00:00:00
updates/19/x86_64/metalink | 28 kB 00:00:00
Loading mirror speeds from cached hostfile
* fedora: mirror.netcologne.de <http://mirror.netcologne.de>
* livna: rpm.livna.org <http://rpm.livna.org>
* rpmfusion-free: mirror.switch.ch <http://mirror.switch.ch>
* rpmfusion-free-updates: mirror.switch.ch <http://mirror.switch.ch>
* rpmfusion-nonfree: mirror.switch.ch <http://mirror.switch.ch>
* rpmfusion-nonfree-updates: mirror.switch.ch <http://mirror.switch.ch>
* updates: mirror.netcologne.de <http://mirror.netcologne.de>
Resolving Dependencies
--> Running transaction check
---> Package python-ethtool.x86_64 0:0.8-1.fc19 will be a downgrade
---> Package python-ethtool.x86_64 0:0.9-2.fc19 will be erased
---> Package python-lxml.x86_64 0:3.2.1-1.fc19 will be a downgrade
---> Package python-lxml.x86_64 0:3.3.5-1.fc19 will be erased
--> Finished Dependency Resolution
Dependencies Resolved
======================================================================================================================================================
Package Arch Version Repository Size
======================================================================================================================================================
Downgrading:
python-ethtool x86_64 0.8-1.fc19 fedora 32 k
python-lxml x86_64 3.2.1-1.fc19 fedora 752 k
Transaction Summary
======================================================================================================================================================
Downgrade 2 Packages
Total download size: 785 k
Is this ok [y/d/N]:
Downloading packages:
(1/2): python-ethtool-0.8-1.fc19.x86_64.rpm | 32 kB 00:00:00
(2/2): python-lxml-3.2.1-1.fc19.x86_64.rpm | 752 kB 00:00:02
------------------------------------------------------------------------------------------------------------------------------------------------------
Total 317 kB/s | 785 kB 00:00:02
Running transaction check
Running transaction test
Transaction test succeeded
Running transaction
Installing : python-lxml-3.2.1-1.fc19.x86_64 1/4
Installing : python-ethtool-0.8-1.fc19.x86_64 2/4
Cleanup : python-lxml-3.3.5-1.fc19.x86_64 3/4
Cleanup : python-ethtool-0.9-2.fc19.x86_64 4/4
Verifying : python-ethtool-0.8-1.fc19.x86_64 1/4
Verifying : python-lxml-3.2.1-1.fc19.x86_64 2/4
Verifying : python-ethtool-0.9-2.fc19.x86_64 3/4
Verifying : python-lxml-3.3.5-1.fc19.x86_64 4/4
Removed:
python-ethtool.x86_64 0:0.9-2.fc19 python-lxml.x86_64 0:3.3.5-1.fc19
Installed:
python-ethtool.x86_64 0:0.8-1.fc19 python-lxml.x86_64 0:3.2.1-1.fc19
Complete!
[root at tekkaman log]# vdsClient -s 0 getVdsCaps
Unexpected exception
[root at tekkaman log]# systemctl restart vdsmd
[root at tekkaman log]# systemctl status vdsmd
vdsmd.service - Virtual Desktop Server Manager
Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled)
Active: active (running) since Sun 2014-05-11 23:44:52 CEST; 16s ago
Process: 13935 ExecStopPost=/usr/libexec/vdsm/vdsmd_init_common.sh
--post-stop (code=exited, status=0/SUCCESS)
Process: 13939 ExecStartPre=/usr/libexec/vdsm/vdsmd_init_common.sh
--pre-start (code=exited, status=0/SUCCESS)
Main PID: 14003 (vdsm)
CGroup: name=systemd:/system/vdsmd.service
├─14003 /usr/bin/python /usr/share/vdsm/vdsm
├─14074 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 32 30
├─14081 rpc.statd --no-notify
├─14103 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 34 32
├─14105 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 41 40
├─14106 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 48 46
├─14107 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 54 48
├─14108 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 41 40
├─14121 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 50 41
├─14123 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 66 65
├─14125 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 71 50
├─14129 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 75 71
├─14131 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 81 75
├─14135 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 92 89
├─14137 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 75 65
└─14138 /usr/bin/python
/usr/share/vdsm/storage/remoteFileHandler.pyc 89 75
May 11 23:44:53 tekkaman.localdomain.local vdsm[14003]: vdsm vds WARNING
Unable to load the json rpc server module. Please make sure it is installed.
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
client step 2
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
parse_server_challenge()
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
ask_user_info()
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
client step 2
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
ask_user_info()
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
make_client_response()
May 11 23:44:53 tekkaman.localdomain.local python[14003]: DIGEST-MD5
client step 3
May 11 23:44:57 tekkaman.localdomain.local rpc.statd[14081]: Version
1.2.7 starting
May 11 23:44:57 tekkaman.localdomain.local rpc.statd[14081]: Flags: TI-RPC
[root at tekkaman log]# vdsClient -s 0 getVdsCaps
HBAInventory = {'FC': [], 'iSCSI': [{'InitiatorName':
'iqn.1994-05.com.redhat:e6aa759a959'}]}
ISCSIInitiatorName = 'iqn.1994-05.com.redhat:e6aa759a959'
bondings = {'bond0': {'addr': '',
'cfg': {},
'hwaddr': '9a:0f:68:19:0f:87',
'ipv6addrs': [],
'mtu': '1500',
'netmask': '',
'slaves': []}}
bridges = {';vdsmdummy;': {'addr': '',
'cfg': {},
'gateway': '',
'ipv6addrs': [],
'ipv6gateway': '::',
'mtu': '1500',
'netmask': '',
'ports': [],
'stp': 'off'},
'ovirtmgmt': {'addr': '192.168.1.101',
'cfg': {'BOOTPROTO': 'none',
'DELAY': '0',
'DEVICE': 'ovirtmgmt',
'GATEWAY': '192.168.1.1',
'IPADDR': '192.168.1.101',
'NETMASK': '255.255.255.0',
'NM_CONTROLLED': 'no',
'ONBOOT': 'yes',
'TYPE': 'Bridge'},
'gateway': '192.168.1.1',
'ipv6addrs': ['fe80::92e6:baff:fec9:f1e1/64'],
'ipv6gateway': '::',
'mtu': '1500',
'netmask': '255.255.255.0',
'ports': ['p10p1'],
'stp': 'off'}}
clusterLevels = ['3.0', '3.1', '3.2', '3.3', '3.4']
cpuCores = '4'
cpuFlags =
'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,mmx,fxsr,sse,sse2,ht,syscall,nx,mmxext,fxsr_opt,pdpe1gb,rdtscp,lm,3dnowext,3dnow,constant_tsc,rep_good,nopl,nonstop_tsc,extd_apicid,pni,monitor,cx16,popcnt,lahf_lm,cmp_legacy,svm,extapic,cr8_legacy,abm,sse4a,misalignsse,3dnowprefetch,osvw,ibs,skinit,wdt,hw_pstate,npt,lbrv,svm_lock,nrip_save,model_athlon,model_Opteron_G3,model_Opteron_G1,model_phenom,model_Opteron_G2'
cpuModel = 'AMD Athlon(tm) II X4 630 Processor'
cpuSockets = '1'
cpuSpeed = '2800.000'
cpuThreads = '4'
emulatedMachines = ['pc',
'pc-q35-1.4',
'pc-q35-1.5',
'q35',
'isapc',
'pc-0.10',
'pc-0.11',
'pc-0.12',
'pc-0.13',
'pc-0.14',
'pc-0.15',
'pc-1.0',
'pc-1.1',
'pc-1.2',
'pc-1.3',
'pc-i440fx-1.4',
'pc-i440fx-1.5',
'none']
guestOverhead = '65'
hooks = {}
kvmEnabled = 'true'
lastClient = '192.168.1.101'
lastClientIface = 'ovirtmgmt'
management_ip = '0.0.0.0'
memSize = '16048'
netConfigDirty = 'False'
networks = {'ovirtmgmt': {'addr': '192.168.1.101',
'bridged': True,
'cfg': {'BOOTPROTO': 'none',
'DELAY': '0',
'DEVICE': 'ovirtmgmt',
'GATEWAY': '192.168.1.1',
'IPADDR': '192.168.1.101',
'NETMASK': '255.255.255.0',
'NM_CONTROLLED': 'no',
'ONBOOT': 'yes',
'TYPE': 'Bridge'},
'gateway': '192.168.1.1',
'iface': 'ovirtmgmt',
'ipv6addrs':
['fe80::92e6:baff:fec9:f1e1/64'],
'ipv6gateway': '::',
'mtu': '1500',
'netmask': '255.255.255.0',
'ports': ['p10p1'],
'stp': 'off'}}
nics = {'p10p1': {'addr': '',
'cfg': {'BRIDGE': 'ovirtmgmt',
'DEVICE': 'p10p1',
'HWADDR': '90:e6:ba:c9:f1:e1',
'NM_CONTROLLED': 'no',
'ONBOOT': 'yes'},
'hwaddr': '90:e6:ba:c9:f1:e1',
'ipv6addrs': ['fe80::92e6:baff:fec9:f1e1/64'],
'mtu': '1500',
'netmask': '',
'speed': 100}}
operatingSystem = {'name': 'Fedora', 'release': '8', 'version': '19'}
packages2 = {'kernel': {'buildtime': 1398276657.0,
'release': '100.fc19.x86_64',
'version': '3.13.11'},
'libvirt': {'buildtime': 1387094943,
'release': '1.fc19',
'version': '1.1.3.2'},
'mom': {'buildtime': 1391183561, 'release': '1.fc19',
'version': '0.4.0'},
'qemu-img': {'buildtime': 1384762225,
'release': '2.fc19',
'version': '1.6.1'},
'qemu-kvm': {'buildtime': 1384762225,
'release': '2.fc19',
'version': '1.6.1'},
'spice-server': {'buildtime': 1383130020,
'release': '3.fc19',
'version': '0.12.4'},
'vdsm': {'buildtime': 1395804204, 'release': '0.fc19',
'version': '4.14.6'}}
reservedMem = '321'
rngSources = ['random']
software_revision = '0'
software_version = '4.14'
supportedENGINEs = ['3.0', '3.1', '3.2', '3.3', '3.4']
supportedProtocols = ['2.2', '2.3']
uuid = 'E0E1001E-8C00-002A-6F9A-90E6BAC9F1E1'
version_name = 'Snow Man'
vlans = {}
vmTypes = ['kvm']
And I'm now able to start and connect to my VM again.
HIH,
Gianluca
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/devel/attachments/20140512/6ad3f44c/attachment-0001.html>
More information about the Devel
mailing list