[ovirt-users] Options not being passed fence_ipmilan, Ovirt3.5 on Centos 7.1 hosts

Mike Lindsay mike.lindsay at cbc.ca
Wed Apr 22 13:45:27 UTC 2015


Hi Eli,

All nodes are the same in the cluster (fresh install):

[mpc-ovirt-node03 ~]# yum list installed *OVIRT* *VDSM* *FENCE*
Loaded plugins: fastestmirror, langpacks, priorities
Loading mirror speeds from cached hostfile
 * base: centos.mirror.rafal.ca
 * epel: fedora-epel.mirror.iweb.com
 * extras: less.cogeco.net
 * ovirt-3.5: resources.ovirt.org
 * ovirt-3.5-epel: fedora-epel.mirror.iweb.com
 * rpmforge: repoforge.mirror.constant.com
 * updates: centos.mirror.rafal.ca
183 packages excluded due to repository priority protections
Installed Packages
fence-agents-all.x86_64
4.0.11-11.el7_1
@updates
fence-agents-apc.x86_64
4.0.11-11.el7_1
@updates
fence-agents-apc-snmp.x86_64
4.0.11-11.el7_1
@updates
fence-agents-bladecenter.x86_64
4.0.11-11.el7_1
@updates
fence-agents-brocade.x86_64
4.0.11-11.el7_1
@updates
fence-agents-cisco-mds.x86_64
4.0.11-11.el7_1
@updates
fence-agents-cisco-ucs.x86_64
4.0.11-11.el7_1
@updates
fence-agents-common.x86_64
4.0.11-11.el7_1
@updates
fence-agents-drac5.x86_64
4.0.11-11.el7_1
@updates
fence-agents-eaton-snmp.x86_64
4.0.11-11.el7_1
@updates
fence-agents-eps.x86_64
4.0.11-11.el7_1
@updates
fence-agents-hpblade.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ibmblade.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ifmib.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ilo-mp.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ilo-ssh.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ilo2.x86_64
4.0.11-11.el7_1
@updates
fence-agents-intelmodular.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ipdu.x86_64
4.0.11-11.el7_1
@updates
fence-agents-ipmilan.x86_64
4.0.11-11.el7_1
@updates
fence-agents-kdump.x86_64
4.0.11-11.el7_1
@updates
fence-agents-rhevm.x86_64
4.0.11-11.el7_1
@updates
fence-agents-rsb.x86_64
4.0.11-11.el7_1
@updates
fence-agents-scsi.x86_64
4.0.11-11.el7_1
@updates
fence-agents-vmware-soap.x86_64
4.0.11-11.el7_1
@updates
fence-agents-wti.x86_64
4.0.11-11.el7_1
@updates
fence-virt.x86_64
0.3.2-1.el7
@base
libgovirt.x86_64
0.3.1-3.el7
@base
ovirt-engine-sdk-python.noarch
3.5.1.0-1.el7.centos
@ovirt-3.5
ovirt-host-deploy.noarch
1.3.1-1.el7
@ovirt-3.5
ovirt-hosted-engine-ha.noarch
1.2.5-1.el7.centos
@ovirt-3.5
ovirt-hosted-engine-setup.noarch
1.2.2-1.el7.centos
@ovirt-3.5
ovirt-release35.noarch
002-1
@/ovirt-release35
vdsm.x86_64
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-cli.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-gluster.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-jsonrpc.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-python.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-python-zombiereaper.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-xmlrpc.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5
vdsm-yajsonrpc.noarch
4.16.10-8.gitc937927.el7
@ovirt-3.5

Cheers,
Mike


On 22 April 2015 at 03:00, Eli Mesika <emesika at redhat.com> wrote:

>
>
> ----- Original Message -----
> > From: "Mike Lindsay" <mike.lindsay at cbc.ca>
> > To: users at ovirt.org
> > Sent: Tuesday, April 21, 2015 8:16:18 PM
> > Subject: [ovirt-users] Options not being passed fence_ipmilan,
> Ovirt3.5 on Centos 7.1 hosts
> >
> > Hi All,
> >
> > I have a bit of an issue with a new install of Ovirt 3.5 (our 3.4
> cluster is
> > working fine) in a 4 node cluster.
> >
> > When I test fencing (or cause a kernal panic triggering a fence) the
> fencing
> > fails. On investigation it appears that the fencing options are not being
> > passed to the fencing script (fence_ipmilan in this case):
> >
> > Fence options under GUI(as entered in the gui): lanplus, ipport=623,
> > power_wait=4, privlvl=operator
> >
> > from vdsm.log on the fence proxy node:
> >
> > Thread-818296::DEBUG::2015-04-21
> 12:39:39,136::API::1209::vds::(fenceNode)
> >
> fenceNode(addr=x.x.x.x,port=,agent=ipmilan,user=stonith,passwd=XXXX,action=status,secure=False,options=
> > power_wait=4
> > Thread-818296::DEBUG::2015-04-21
> 12:39:39,137::utils::739::root::(execCmd)
> > /usr/sbin/fence_ipmilan (cwd None)
> > Thread-818296::DEBUG::2015-04-21
> 12:39:39,295::utils::759::root::(execCmd)
> > FAILED: <err> = 'Failed: Unable to obtain correct plug status or plug is
> not
> > available\n\n\n'; <rc> = 1
> > Thread-818296::DEBUG::2015-04-21 12:39:39,296::API::1164::vds::(fence)
> rc 1
> > inp agent=fence_ipmilan
> > Thread-818296::DEBUG::2015-04-21
> 12:39:39,296::API::1235::vds::(fenceNode) rc
> > 1 in agent=fence_ipmilan
> > Thread-818296::DEBUG::2015-04-21
> > 12:39:39,297::stompReactor::163::yajsonrpc.StompServer::(send) Sending
> > response
> >
> >
> > from engine.log on the engine:
> > 2015-04-21 12:39:38,843 INFO
> > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > (ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
> Event
> > ID: -1, Message: Host mpc-ovirt-node03 from cluster Default was chosen
> as a
> > proxy to execute Status command on Host mpc-ovirt-node04.
> > 2015-04-21 12:39:38,845 INFO [org.ovirt.engine.core.bll.FenceExecutor]
> > (ajp--127.0.0.1-8702-4) Using Host mpc-ovirt-node03 from cluster Default
> as
> > proxy to execute Status command on Host
> > 2015-04-21 12:39:38,885 INFO [org.ovirt.engine.core.bll.FenceExecutor]
> > (ajp--127.0.0.1-8702-4) Executing <Status> Power Management command,
> Proxy
> > Host:mpc-ovirt-node03, Agent:ipmilan, Target Host:, Management
> IP:x.x.x.x,
> > User:stonith, Options: power_wait=4, ipport=623,
> privlvl=operator,lanplus,
> > Fencing policy:null
> > 2015-04-21 12:39:38,921 INFO
> > [org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
> > (ajp--127.0.0.1-8702-4) START, FenceVdsVDSCommand(HostName =
> > mpc-ovirt-node03, HostId = 5613a489-589d-4e89-ab01-3642795eedb8,
> targetVdsId
> > = dbfa4e85-3e97-4324-b222-bf40a491db08, action = Status, ip = x.x.x.x,
> port
> > = , type = ipmilan, user = stonith, password = ******, options = '
> > power_wait=4, ipport=623, privlvl=operator,lanplus', policy = 'null'),
> log
> > id: 774f328
> > 2015-04-21 12:39:39,338 WARN
> > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > (ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
> Event
> > ID: -1, Message: Power Management test failed for Host
> mpc-ovirt-node04.Done
> > 2015-04-21 12:39:39,339 INFO
> > [org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
> > (ajp--127.0.0.1-8702-4) FINISH, FenceVdsVDSCommand, return: Test
> Succeeded,
> > unknown, log id: 774f328
> > 2015-04-21 12:39:39,340 WARN [org.ovirt.engine.core.bll.FenceExecutor]
> > (ajp--127.0.0.1-8702-4) Fencing operation failed with proxy host
> > 5613a489-589d-4e89-ab01-3642795eedb8, trying another proxy...
> > 2015-04-21 12:39:39,594 INFO
> > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > (ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
> Event
> > ID: -1, Message: Host mpc-ovirt-node01 from cluster Default was chosen
> as a
> > proxy to execute Status command on Host mpc-ovirt-node04.
> > 2015-04-21 12:39:39,595 INFO [org.ovirt.engine.core.bll.FenceExecutor]
> > (ajp--127.0.0.1-8702-4) Using Host mpc-ovirt-node01 from cluster Default
> as
> > proxy to execute Status command on Host
> > 2015-04-21 12:39:39,598 INFO [org.ovirt.engine.core.bll.FenceExecutor]
> > (ajp--127.0.0.1-8702-4) Executing <Status> Power Management command,
> Proxy
> > Host:mpc-ovirt-node01, Agent:ipmilan, Target Host:, Management
> IP:x.x.x.x,
> > User:stonith, Options: power_wait=4, ipport=623,
> privlvl=operator,lanplus,
> > Fencing policy:null
> > 2015-04-21 12:39:39,634 INFO
> > [org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
> > (ajp--127.0.0.1-8702-4) START, FenceVdsVDSCommand(HostName =
> > mpc-ovirt-node01, HostId = c3e8be6e-ac54-4861-b774-17ba5cc66dc6,
> targetVdsId
> > = dbfa4e85-3e97-4324-b222-bf40a491db08, action = Status, ip = x.x.x.x,
> port
> > = , type = ipmilan, user = stonith, password = ******, options = '
> > power_wait=4, ipport=623, privlvl=operator,lanplus', policy = 'null'),
> log
> > id: 6369eb1
> > 2015-04-21 12:39:40,056 WARN
> > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > (ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
> Event
> > ID: -1, Message: Power Management test failed for Host
> mpc-ovirt-node04.Done
> > 2015-04-21 12:39:40,057 INFO
> > [org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
> > (ajp--127.0.0.1-8702-4) FINISH, FenceVdsVDSCommand, return: Test
> Succeeded,
> > unknown, log id: 6369eb1
> >
> >
> > For verification I temporarily replaced /usr/sbin/fence_ipmilan with a
> shell
> > script that dumps the env plus any cli args passed into a log file:
> >
> > -------------------------- Tue Apr 21 12:39:39 EDT 2015
> > ----------------------------
> > ENV DUMP:
> > LC_ALL=C
> > USER=vdsm
> > PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
> > PWD=/
> > LANG=en_CA.UTF-8
> > LIBVIRT_LOG_FILTERS=
> > SHLVL=1
> > HOME=/var/lib/vdsm
> > LOGNAME=vdsm
> > LIBVIRT_LOG_OUTPUTS=
> > _=/usr/bin/env
> >
> > ------------------------------
> > CLI DUMP:
> >
> > <this is where the cli args should be listed>
> >
> >
> > Version info:
> > libvirt version: 1.2.8, package: 16.el7_1.2 (CentOS BuildSystem <
> > http://bugs.centos.org >, 2015-03-26-23:17:42, worker1.bsys.centos.org )
> > fence_ipmilan: 4.0.11 (built Mon Apr 13 13:22:18 UTC 2015)
> > vdsm.x86_64: 4.16.10-8.gitc937927.el7
> > ovirt-engine.noarch: 3.5.1.1-1.el6
> >
> > Engine os: Centos 6.6
> > Host os: Centos 7.1.1503
> >
> > I've found some old post from 2012 that describe the same problem. Has
> anyone
> > else run into this?
> >
> > Any thought or suggestions would be appreciated.
>
> Hi Mike
>
> Seems from the logs above that the oVirt engine got the options correctly
> but somehow it is ignored by VDSM
> Can you please share installed VDSM RPM version on the host that served as
> a proxy for this operation
> Please also provide the cluster version of the cluster that contains the
> proxy host as well
>
> Thanks
>
> Eli
>
>
>
> >
> > Cheers,
> > Mike
> >
> > _______________________________________________
> > Users mailing list
> > Users at ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>



-- 
 *Mike Lindsay*
Senior Systems Administrator
*Technological Maintenance and Support*
*CBC/SRC*
mike.lindsay at cbc.ca
(o) 416-205-8992
(c) 416-819-2841
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150422/e44e7bc8/attachment-0001.html>


More information about the Users mailing list