[ovirt-users] 3. vdsm losing connection to libvirt (Chris Adams)

Nikolai Sednev nsednev at redhat.com
Tue Dec 16 17:10:54 UTC 2014


Hi, 
Can I get engine, libvirt, vdsm, mom, logs from host8 and connectivity log? 
Have you tried installing clean OSs on hosts, especially on problematic host? 
I'd also try to disable JSONRPC on hosts, by putting them to maintenance and then removing JSONRPC from the check box on all hosts, just to compare if it resolves the issue. 



Thanks in advance. 

Best regards, 
Nikolai 
____________________ 
Nikolai Sednev 
Senior Quality Engineer at Compute team 
Red Hat Israel 
34 Jerusalem Road, 
Ra'anana, Israel 43501 

Tel: +972 9 7692043 
Mobile: +972 52 7342734 
Email: nsednev at redhat.com 
IRC: nsednev 

----- Original Message -----

From: users-request at ovirt.org 
To: users at ovirt.org 
Sent: Tuesday, December 16, 2014 5:50:28 PM 
Subject: Users Digest, Vol 39, Issue 98 

Send Users mailing list submissions to 
users at ovirt.org 

To subscribe or unsubscribe via the World Wide Web, visit 
http://lists.ovirt.org/mailman/listinfo/users 
or, via email, send a message with subject or body 'help' to 
users-request at ovirt.org 

You can reach the person managing the list at 
users-owner at ovirt.org 

When replying, please edit your Subject line so it is more specific 
than "Re: Contents of Users digest..." 


Today's Topics: 

1. Re: Free Ovirt Powered Cloud (Lior Vernia) 
2. gluster rpms not found (Pat Pierson) 
3. vdsm losing connection to libvirt (Chris Adams) 
4. Re: Creating new users on oVirt 3.5 (Donny Davis) 
5. Re: gfapi, 3.5.1 (Alex Crow) 


---------------------------------------------------------------------- 

Message: 1 
Date: Tue, 16 Dec 2014 15:55:02 +0200 
From: Lior Vernia <lvernia at redhat.com> 
To: Donny Davis <donny at cloudspin.me> 
Cc: users at ovirt.org 
Subject: Re: [ovirt-users] Free Ovirt Powered Cloud 
Message-ID: <549039B6.2010804 at redhat.com> 
Content-Type: text/plain; charset=ISO-8859-1 

Hi Donny, 

On 15/12/14 18:24, Donny Davis wrote: 
> Hi guys, I'm providing a free public cloud solution entirely based on 
> vanilla oVirt called cloudspin.me <http://cloudspin.me> 
> 

This looks great! :) 

> It runs on IPv6, and I am looking for people to use the system, host 
> services and report back to me with their results. 
> 

Do you also use IPv6 internally in your deployment? e.g. assign IPv6 
addresses to your hosts, storage domain, power management etc.? We'd be 
very interested to hear what works and what doesn't. And perhaps help 
push forward what doesn't, if you need it :) 

> Data I am looking for 
> 
> Connection Speed - Is it comparable to other services 
> 
> User experience - Are there any changes recommended 
> 
> Does it work for you - What does, and does not work for you. 
> 
> 
> 
> I am trying to get funding to keep this a free resource for everyone to 
> use. (not from here:) 
> 
> I am completely open to any and all suggestions, and or help with 
> things. I am a one man show at the moment. 
> 
> If anyone has any questions please email me back 
> 
> Donny D 
> 
> 
> 
> 
> 
> _______________________________________________ 
> Users mailing list 
> Users at ovirt.org 
> http://lists.ovirt.org/mailman/listinfo/users 
> 


------------------------------ 

Message: 2 
Date: Tue, 16 Dec 2014 09:08:57 -0500 
From: Pat Pierson <ihasn2004 at gmail.com> 
To: nathan at robotics.net 
Cc: "users at ovirt.org" <users at ovirt.org> 
Subject: [ovirt-users] gluster rpms not found 
Message-ID: 
<CAMRYiEiKL1MEGoHWjKtnhW3DXjouU0w3hs5zFx75sfBL8M4JaQ at mail.gmail.com> 
Content-Type: text/plain; charset="utf-8" 

Nathan, 
Did you find a work around for this? I am running into the same issue. 

Is there a way to force vdsm to see gluster? Or a way to manually run the 
search so I can see why it fails? 


>*<> 
*nathan stratton | vp technology | broadsoft, inc | +1-240-404-6580 
|www.broadsoft.com 


On Fri, Jun 20, 2014 at 11:01 AM, Nathan Stratton <nathan at 
robotics.net <http://lists.ovirt.org/mailman/listinfo/users>> 
wrote: 

>* Actually I have vdsm-gluster, that is why vdsm tries to find the gluster 
*>* packages. Is there a way I can run the vdsm gluster rpm search manually to 
*>* see what is going wrong? 
*>>* [root at virt01a <http://lists.ovirt.org/mailman/listinfo/users> 
~]# yum list installed |grep vdsm 
*>* vdsm.x86_64 4.14.9-0.el6 @ovirt-3.4-stable 
*>>* vdsm-cli.noarch 4.14.9-0.el6 @ovirt-3.4-stable 
*>>* vdsm-gluster.noarch 4.14.9-0.el6 @ovirt-3.4-stable 
*>>* vdsm-python.x86_64 4.14.9-0.el6 @ovirt-3.4-stable 
*>>* vdsm-python-zombiereaper.noarch 
*>* vdsm-xmlrpc.noarch 4.14.9-0.el6 @ovirt-3.4-stable 
*>>>>* ><> 
*>* nathan stratton | vp technology | broadsoft, inc | +1-240-404-6580 
<%2B1-240-404-6580> | 
*>* www.broadsoft.com <http://www.broadsoft.com/> 
*>>>* On Thu, Jun 19, 2014 at 8:39 PM, Andrew Lau <andrew at 
andrewklau.com <http://lists.ovirt.org/mailman/listinfo/users>> wrote: 
*>>>* You're missing vdsm-gluster 
*>>>>* yum install vdsm-gluster 
*>>>>* On Fri, Jun 20, 2014 at 6:24 AM, Nathan Stratton <nathan at 
robotics.net <http://lists.ovirt.org/mailman/listinfo/users>> 
*>>* wrote: 
*>>* > I am running ovirt 3.4 and have gluster installed: 
*>>* > 
*>>* > [root at virt01a 
<http://lists.ovirt.org/mailman/listinfo/users>]# yum list installed 
|grep gluster 
*>>* > glusterfs.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > glusterfs-api.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > glusterfs-cli.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > glusterfs-fuse.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > glusterfs-libs.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > glusterfs-rdma.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > glusterfs-server.x86_64 3.5.0-2.el6 @ovirt-glusterfs-epel 
*>>* > 
*>>* > However vdsm can't seem to find them: 
*>>* > 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,250::caps::458::root::(_getKeyPackages) rpm package 
*>>* glusterfs-rdma 
*>>* > not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,250::caps::458::root::(_getKeyPackages) rpm package 
*>>* glusterfs-fuse 
*>>* > not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,251::caps::458::root::(_getKeyPackages) rpm package 
*>>* gluster-swift 
*>>* > not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,252::caps::458::root::(_getKeyPackages) rpm package 
*>>* > gluster-swift-object not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,252::caps::458::root::(_getKeyPackages) rpm package glusterfs 
*>>* not 
*>>* > found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,252::caps::458::root::(_getKeyPackages) rpm package 
*>>* > gluster-swift-plugin not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,254::caps::458::root::(_getKeyPackages) rpm package 
*>>* > gluster-swift-account not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,254::caps::458::root::(_getKeyPackages) rpm package 
*>>* > gluster-swift-proxy not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,254::caps::458::root::(_getKeyPackages) rpm package 
*>>* > gluster-swift-doc not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,255::caps::458::root::(_getKeyPackages) rpm package 
*>>* > glusterfs-server not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,255::caps::458::root::(_getKeyPackages) rpm package 
*>>* > gluster-swift-container not found 
*>>* > Thread-13::DEBUG::2014-06-19 
*>>* > 16:15:57,255::caps::458::root::(_getKeyPackages) rpm package 
*>>* > glusterfs-geo-replication not found 
*>>* > 
*>>* > Any ideas? 
*>>* > 
*>>* >><> 
*>>* > nathan stratton | vp technology | broadsoft, inc | 
+1-240-404-6580 <%2B1-240-404-6580> | 
*>>* > www.broadsoft.com <http://www.broadsoft.com/> 
*>>* > 
*>>* > _______________________________________________ 
*>>* > Users mailing list 
*>>* > Users at ovirt.org <http://lists.ovirt.org/mailman/listinfo/users> 
*>>* > http://lists.ovirt.org/mailman/listinfo/users 
<http://lists.ovirt.org/mailman/listinfo/users> 
*>>* > 
*>>>>-------------- next part -------------- 
An HTML attachment was scrubbed... 
URL: <http://lists.ovirt.org/pipermail/users/attachments/20140621/9b14c8fe/attachment.html> 


-- 
Patrick Pierson 
-------------- next part -------------- 
An HTML attachment was scrubbed... 
URL: <http://lists.ovirt.org/pipermail/users/attachments/20141216/58d14872/attachment-0001.html> 

------------------------------ 

Message: 3 
Date: Tue, 16 Dec 2014 08:48:48 -0600 
From: Chris Adams <cma at cmadams.net> 
To: users at ovirt.org 
Subject: [ovirt-users] vdsm losing connection to libvirt 
Message-ID: <20141216144848.GA1708 at cmadams.net> 
Content-Type: text/plain; charset=us-ascii 

I have a oVirt setup that has three nodes, all running CentOS 7, with a 
hosted engine running CentOS 6. Two of the nodes (node8 and node9) are 
configured for hosted engine, and the third (node2) is just a "regular" 
node (as you might guess from the names, more nodes are coming as I 
migrate VMs to oVirt). 

On one node, node8, vdsm periodically loses its connection to libvirt, 
which causes vdsm to restart. There doesn't appear to be any trigger 
that I can see (not time of day, load, etc. related). The engine VM is 
up and running on node8 (don't know if that has anything to do with it). 

I get some entries in /var/log/messages repeated continuously; the 
"ovirt-ha-broker: sending ioctl 5401 to a partition" I mentioned before, 
and the following: 

Dec 15 20:56:23 node8 journal: User record for user '107' was not found: No such file or directory 
Dec 15 20:56:23 node8 journal: Group record for user '107' was not found: No such file or directory 

I don't think those have any relevance (don't know where they come 
from); filtering those out, I see: 

Dec 15 20:56:33 node8 journal: End of file while reading data: Input/output error 
Dec 15 20:56:33 node8 journal: Tried to close invalid fd 0 
Dec 15 20:56:38 node8 journal: vdsm root WARNING connection to libvirt broken. ecode: 1 edom: 7 
Dec 15 20:56:38 node8 journal: vdsm root CRITICAL taking calling process down. 
Dec 15 20:56:38 node8 journal: vdsm vds ERROR libvirt error 
Dec 15 20:56:38 node8 journal: ovirt-ha-broker mgmt_bridge.MgmtBridge ERROR Failed to getVdsCapabilities: Error 16 from getVdsCapabilities: Unexpected exception 
Dec 15 20:56:45 node8 journal: End of file while reading data: Input/output error 
Dec 15 20:56:45 node8 vdsmd_init_common.sh: vdsm: Running run_final_hooks 
Dec 15 20:56:45 node8 systemd: Starting Virtual Desktop Server Manager... 
<and then all the normal-looking vdsm startup> 

It is happening about once a day, but not at any regular interval or 
time (was 02:23 Sunday, then 20:56 Monday). 

vdsm.log has this at that time: 

Thread-601576::DEBUG::2014-12-15 20:56:38,715::BindingXMLRPC::1132::vds::(wrapper) client [127.0.0.1]::call getCapabilities with () {} 
Thread-601576::DEBUG::2014-12-15 20:56:38,718::utils::738::root::(execCmd) /sbin/ip route show to 0.0.0.0/0 table all (cwd None) 
Thread-601576::DEBUG::2014-12-15 20:56:38,746::utils::758::root::(execCmd) SUCCESS: <err> = ''; <rc> = 0 
Thread-601576::WARNING::2014-12-15 20:56:38,754::libvirtconnection::135::root::(wrapper) connection to libvirt broken. ecode: 1 edom: 7 
Thread-601576::CRITICAL::2014-12-15 20:56:38,754::libvirtconnection::137::root::(wrapper) taking calling process down. 
MainThread::DEBUG::2014-12-15 20:56:38,754::vdsm::58::vds::(sigtermHandler) Received signal 15 
Thread-601576::DEBUG::2014-12-15 20:56:38,755::libvirtconnection::143::root::(wrapper) Unknown libvirterror: ecode: 1 edom: 7 level: 2 message: internal error: client socket is closed 
MainThread::DEBUG::2014-12-15 20:56:38,755::protocoldetector::135::vds.MultiProtocolAcceptor::(stop) Stopping Acceptor 
MainThread::INFO::2014-12-15 20:56:38,755::__init__::563::jsonrpc.JsonRpcServer::(stop) Stopping JsonRPC Server 
Detector thread::DEBUG::2014-12-15 20:56:38,756::protocoldetector::106::vds.MultiProtocolAcceptor::(_cleanup) Cleaning Acceptor 
MainThread::INFO::2014-12-15 20:56:38,757::vmchannels::188::vds::(stop) VM channels listener was stopped. 
MainThread::INFO::2014-12-15 20:56:38,758::momIF::91::MOM::(stop) Shutting down MOM 
MainThread::DEBUG::2014-12-15 20:56:38,759::task::595::Storage.TaskManager.Task::(_updateState) Task=`26c7680c-23e2-42bb-964c-272e778a168a`::moving from state init -> state preparing 
MainThread::INFO::2014-12-15 20:56:38,759::logUtils::44::dispatcher::(wrapper) Run and protect: prepareForShutdown(options=None) 
Thread-601576::ERROR::2014-12-15 20:56:38,755::BindingXMLRPC::1142::vds::(wrapper) libvirt error 
Traceback (most recent call last): 
File "/usr/share/vdsm/rpc/BindingXMLRPC.py", line 1135, in wrapper 
res = f(*args, **kwargs) 
File "/usr/share/vdsm/rpc/BindingXMLRPC.py", line 463, in getCapabilities 
ret = api.getCapabilities() 
File "/usr/share/vdsm/API.py", line 1245, in getCapabilities 
c = caps.get() 
File "/usr/share/vdsm/caps.py", line 615, in get 
caps.update(netinfo.get()) 
File "/usr/lib/python2.7/site-packages/vdsm/netinfo.py", line 812, in get 
nets = networks() 
File "/usr/lib/python2.7/site-packages/vdsm/netinfo.py", line 119, in networks 
allNets = ((net, net.name()) for net in conn.listAllNetworks(0)) 
File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line 129, in wrapper 
__connections.get(id(target)).pingLibvirt() 
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 3642, in getLibVersion 
if ret == -1: raise libvirtError ('virConnectGetLibVersion() failed', conn=self) 
libvirtError: internal error: client socket is closed 


-- 
Chris Adams <cma at cmadams.net> 


------------------------------ 

Message: 4 
Date: Tue, 16 Dec 2014 07:57:16 -0700 
From: "Donny Davis" <donny at cloudspin.me> 
To: "'Alon Bar-Lev'" <alonbl at redhat.com>, "'Fedele Stabile'" 
<fedele.stabile at fis.unical.it> 
Cc: users at ovirt.org 
Subject: Re: [ovirt-users] Creating new users on oVirt 3.5 
Message-ID: <008801d01940$9682f2f0$c388d8d0$@cloudspin.me> 
Content-Type: text/plain; charset="us-ascii" 

Check out my write-up on AAA, 
I tried my best to break it down, and make it simple 

https://cloudspin.me/ovirt-simple-ldap-aaa/ 

-----Original Message----- 
From: users-bounces at ovirt.org [mailto:users-bounces at ovirt.org] On Behalf Of 
Alon Bar-Lev 
Sent: Tuesday, December 16, 2014 1:49 AM 
To: Fedele Stabile 
Cc: users at ovirt.org 
Subject: Re: [ovirt-users] Creating new users on oVirt 3.5 



----- Original Message ----- 
> From: "Fedele Stabile" <fedele.stabile at fis.unical.it> 
> To: users at ovirt.org 
> Sent: Monday, December 15, 2014 8:05:28 PM 
> Subject: [ovirt-users] Creating new users on oVirt 3.5 
> 
> Hello, 
> I have to create some users on my oVirt 3.5 infrastructure. 
> On FridayI was following istructions on 
> http://www.ovirt.org/LDAP_Quick_Start 
> LDAP Quick Start 
> so I correctly created a OpenLDAP server and a Kerberos service, but 
> this morning I read that the instructions are obsolete... 
> Now I'm trying to understand how to implement the new mechanism... but 
> I'm in troubles: 
> 1) run yum install ovirt-engine-extension-aaa-ldap 
> 2) copied files in /etc/ovirt-engine/extensions.d and modified the 
> name in fis.unical.it-auth(n/z).properties 
> 3) copied files in /etc/ovirt-engine/aaa but now I can't do anything 
> 
> Can you help me with newbye instructions to install the aaa-extensions? 
> Thank you very much 
> Fedele Stabile 

Hello, 

Have you read[1]? 
We of course need help in improving documentation :) Can you please send 
engine.log when starting up engine so I can see if there are any issues? 
Please make sure that at /etc/ovirt-engine/extensions.d you set the 
config.profile.file.1 to absolute file, /etc/ovirt-enigne/aaa/ as we wait 
for 3.5.1 to support relative names. 

The simplest sequence is: 

1. copy recursive /usr/share/ovirt-engine-extension-aaa-ldap/examples/simple 
to /etc/ovirt-engine 2. edit /etc/ovirt-engine/extension.d/* replace ../aaa 
to /etc/ovirt-engine/aaa this is pending 3.5.1. 
3. edit /etc/ovirt-engine/aaa/ldap1.properties and set vars.server, 
vars.user, vars.password to meet your setup. 
4. restart engine. 
5. send me engine.log 

Regards, 
Alon 

[1] 
http://gerrit.ovirt.org/gitweb?p=ovirt-engine-extension-aaa-ldap.git;a=blob; 
f=README;hb=HEAD 
_______________________________________________ 
Users mailing list 
Users at ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users 



------------------------------ 

Message: 5 
Date: Tue, 16 Dec 2014 15:50:23 +0000 
From: Alex Crow <acrow at integrafin.co.uk> 
To: users at ovirt.org 
Subject: Re: [ovirt-users] gfapi, 3.5.1 
Message-ID: <549054BF.2090105 at integrafin.co.uk> 
Content-Type: text/plain; charset=utf-8; format=flowed 

Hi, 

Anyone know if this is due to work correctly in the next iteration of 3.5? 

Thanks 

Alex 

On 09/12/14 10:33, Alex Crow wrote: 
> Hi, 
> 
> Will the vdsm patches to properly enable libgfapi storage for VMs (and 
> matching refactored code in the hosted-engine setup scripts) for VMs 
> make it into 3.5.1? It's not in the snapshots yet it seems. 
> 
> I notice it's in master/3.6 snapshot but something stops the HA stuff 
> in self-hosted setups from connecting storage: 
> 
> from Master test setup: 
> /var/log/ovirt-hosted-engine-ha/broker.log 
> 
> MainThread::INFO::2014-12-08 
> 19:22:56,287::hosted_engine::222::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_hostname) 
> Found certificate common name: 172.17.10.50 
> MainThread::WARNING::2014-12-08 
> 19:22:56,395::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:23:11,501::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:23:26,610::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:23:41,717::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:23:56,824::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::ERROR::2014-12-08 
> 19:24:11,840::hosted_engine::500::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed trying to connect storage: 
> MainThread::ERROR::2014-12-08 
> 19:24:11,840::agent::173::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) 
> Error: 'Failed trying to connect storage' - trying to restart agent 
> MainThread::WARNING::2014-12-08 
> 19:24:16,845::agent::176::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) 
> Restarting agent, attempt '8' 
> MainThread::INFO::2014-12-08 
> 19:24:16,855::hosted_engine::222::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_hostname) 
> Found certificate common name: 172.17.10.50 
> MainThread::WARNING::2014-12-08 
> 19:24:16,962::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:24:32,069::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:24:47,181::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:25:02,288::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::WARNING::2014-12-08 
> 19:25:17,389::hosted_engine::497::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed to connect storage, waiting '15' seconds before the next attempt 
> MainThread::ERROR::2014-12-08 
> 19:25:32,404::hosted_engine::500::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm) 
> Failed trying to connect storage: 
> MainThread::ERROR::2014-12-08 
> 19:25:32,404::agent::173::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) 
> Error: 'Failed trying to connect storage' - trying to restart agent 
> MainThread::WARNING::2014-12-08 
> 19:25:37,409::agent::176::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) 
> Restarting agent, attempt '9' 
> MainThread::ERROR::2014-12-08 
> 19:25:37,409::agent::178::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) 
> Too many errors occurred, giving up. Please review the log and 
> consider filing a bug. 
> MainThread::INFO::2014-12-08 
> 19:25:37,409::agent::118::ovirt_hosted_engine_ha.agent.agent.Agent::(run) 
> Agent shutting down 
> (END) - Next: /var/log/ovirt-hosted-engine-ha/broker.log 
> 
> vdsm.log: 
> 
> Detector thread::DEBUG::2014-12-08 
> 19:20:45,458::protocoldetector::214::vds.MultiProtocolAcceptor::(_remove_connection) 
> Removing connection 127.0.0.1:53083 
> Detector thread::DEBUG::2014-12-08 
> 19:20:45,458::BindingXMLRPC::1193::XmlDetector::(handleSocket) xml 
> over http detected from ('127.0.0.1', 53083) 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,459::BindingXMLRPC::318::vds::(wrapper) client [127.0.0.1] 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,460::task::592::Storage.TaskManager.Task::(_updateState) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::moving from state init -> 
> state preparing 
> Thread-44::INFO::2014-12-08 
> 19:20:45,460::logUtils::48::dispatcher::(wrapper) Run and protect: 
> connectStorageServer(domType=1, 
> spUUID='ab2b5ee7-9aa7-426f-9d58-5e7d3840ad81', conList=[{'connection': 
> 'zebulon.ifa.net:/engine', 'iqn': ',', 'protocol_version': '3' 
> , 'kvm': 'password', '=': 'user', ',': '='}], options=None) 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,461::hsm::2384::Storage.HSM::(__prefetchDomains) nfs local 
> path: /rhev/data-center/mnt/zebulon.ifa.net:_engine 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,462::hsm::2408::Storage.HSM::(__prefetchDomains) Found SD 
> uuids: (u'd3240928-dae9-4ed0-8a28-7ab552455063',) 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::hsm::2464::Storage.HSM::(connectStorageServer) knownSDs: 
> {d3240928-dae9-4ed0-8a28-7ab552455063: storage.nfsSD.findDomain} 
> Thread-44::ERROR::2014-12-08 
> 19:20:45,463::task::863::Storage.TaskManager.Task::(_setError) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::Unexpected error 
> Traceback (most recent call last): 
> File "/usr/share/vdsm/storage/task.py", line 870, in _run 
> return fn(*args, **kargs) 
> File "/usr/share/vdsm/logUtils.py", line 49, in wrapper 
> res = f(*args, **kwargs) 
> File "/usr/share/vdsm/storage/hsm.py", line 2466, in 
> connectStorageServer 
> res.append({'id': conDef["id"], 'status': status}) 
> KeyError: 'id' 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::882::Storage.TaskManager.Task::(_run) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::Task._run: 
> b5accf8f-014a-412d-9fb8-9e9447d49b72 (1, 
> 'ab2b5ee7-9aa7-426f-9d58-5e7d3840ad81', [{'kvm': 'password', ',': '=', 
> 'conn 
> ection': 'zebulon.ifa.net:/engine', 'iqn': ',', 'protocol_version': 
> '3', '=': 'user'}]) {} failed - stopping task 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::1214::Storage.TaskManager.Task::(stop) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::stopping in state 
> preparing (force False) 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::990::Storage.TaskManager.Task::(_decref) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::ref 1 aborting True 
> Thread-44::INFO::2014-12-08 
> 19:20:45,463::task::1168::Storage.TaskManager.Task::(prepare) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::aborting: Task is 
> aborted: u"'id'" - code 100 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::1173::Storage.TaskManager.Task::(prepare) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::Prepare: aborted: 'id' 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::990::Storage.TaskManager.Task::(_decref) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::ref 0 aborting True 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::925::Storage.TaskManager.Task::(_doAbort) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::Task._doAbort: force False 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {} 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,463::task::592::Storage.TaskManager.Task::(_updateState) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::moving from state 
> preparing -> state aborting 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,464::task::547::Storage.TaskManager.Task::(__state_aborting) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::_aborting: recover policy 
> none 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,464::task::592::Storage.TaskManager.Task::(_updateState) 
> Task=`b5accf8f-014a-412d-9fb8-9e9447d49b72`::moving from state 
> aborting -> state failed 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,464::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) 
> Owner.releaseAll requests {} resources {} 
> Thread-44::DEBUG::2014-12-08 
> 19:20:45,464::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {} 
> Thread-44::ERROR::2014-12-08 
> 19:20:45,464::dispatcher::79::Storage.Dispatcher::(wrapper) 'id' 
> Traceback (most recent call last): 
> File "/usr/share/vdsm/storage/dispatcher.py", line 71, in wrapper 
> result = ctask.prepare(func, *args, **kwargs) 
> File "/usr/share/vdsm/storage/task.py", line 103, in wrapper 
> return m(self, *a, **kw) 
> File "/usr/share/vdsm/storage/task.py", line 1176, in prepare 
> raise self.error 
> KeyError: 'id' 
> clientIFinit::ERROR::2014-12-08 
> 19:20:48,190::clientIF::460::vds::(_recoverExistingVms) Vm's recovery 
> failed 
> Traceback (most recent call last): 
> File "/usr/share/vdsm/clientIF.py", line 404, in _recoverExistingVms 
> caps.CpuTopology().cores()) 
> File "/usr/share/vdsm/caps.py", line 200, in __init__ 
> self._topology = _getCpuTopology(capabilities) 
> File "/usr/share/vdsm/caps.py", line 232, in _getCpuTopology 
> capabilities = _getFreshCapsXMLStr() 
> File "/usr/share/vdsm/caps.py", line 222, in _getFreshCapsXMLStr 
> return libvirtconnection.get().getCapabilities() 
> File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", 
> line 157, in get 
> passwd) 
> File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", 
> line 102, in open_connection 
> return utils.retry(libvirtOpen, timeout=10, sleep=0.2) 
> File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 935, in 
> retry 
> return func() 
> File "/usr/lib64/python2.7/site-packages/libvirt.py", line 102, in 
> openAuth 
> if ret is None:raise libvirtError('virConnectOpenAuth() failed') 
> libvirtError: authentication failed: polkit: 
> polkit\56retains_authorization_after_challenge=1 
> Authorization requires authentication but no agent is available. 
> 
> 

-- 
This message is intended only for the addressee and may contain 
confidential information. Unless you are that person, you may not 
disclose its contents or use it in any way and are requested to delete 
the message along with any attachments and notify us immediately. 
"Transact" is operated by Integrated Financial Arrangements plc. 29 
Clement's Lane, London EC4N 7AE. Tel: (020) 7608 4900 Fax: (020) 7608 
5300. (Registered office: as above; Registered in England and Wales 
under number: 3727592). Authorised and regulated by the Financial 
Conduct Authority (entered on the Financial Services Register; no. 190856). 



------------------------------ 

_______________________________________________ 
Users mailing list 
Users at ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users 


End of Users Digest, Vol 39, Issue 98 
************************************* 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20141216/9759da91/attachment-0001.html>


More information about the Users mailing list