soft negative affinity
by Kapetanakis Giannis
Hi,
Can someone comment on this?
https://bugzilla.redhat.com/show_bug.cgi?id=1207255
I've defined a soft negative affinity group for two VMs.
To my understanding if the there are at least 2 nodes available in the
cluster
then the VMs SHOULD start on different nodes.
This does not happen. They start on the same node.
If I make it hard then it works.
However I don't want to make it hard since if there is only one node
available in the cluster then one vm will stay down.
regards,
Giannis
9 years, 7 months
Minutes: oVirt Weekly Sync (Open Format)
by Brian Proffitt
Minutes: http://ovirt.org/meetings/ovirt/2015/ovirt.2015-04-22-14.02.html
Minutes (text): http://ovirt.org/meetings/ovirt/2015/ovirt.2015-04-22-14.02.txt
Log: http://ovirt.org/meetings/ovirt/2015/ovirt.2015-04-22-14.02.log.html
=======================================
#ovirt: oVirt Weekly Sync (Open Format)
=======================================
Meeting started by bkp at 14:02:19 UTC. The full logs are available at
http://ovirt.org/meetings/ovirt/2015/ovirt.2015-04-22-14.02.log.html .
Meeting summary
---------------
* Agenda (bkp, 14:04:08)
* 3.5 news (bkp, 14:04:33)
* 3.6 news/discussion (bkp, 14:04:45)
* Open Discussion (bkp, 14:04:55)
* 3.5 news/discussion (bkp, 14:06:35)
* 3.5 general status email:
http://lists.ovirt.org/pipermail/devel/2015-April/010315.html (bkp,
14:12:58)
* 3.5.2 looks good for April 28 (bkp, 14:13:01)
* 3.5.2, only one bug in QA (bkp, 14:13:03)
* sbonazzo needs maintainers to review the package list he sent on RC4
release, to be sure that listed packages correspond to the ones
maintainers want to release (bkp, 14:13:07)
* 3.6 news/discussion (bkp, 14:13:29)
* Brief general status for 3.6 here:
http://lists.ovirt.org/pipermail/devel/2015-April/010310.html (bkp,
14:28:29)
* el6 repo broken by the removal of vdsm, infra and node (fabiand)
must address in the next couple of days (bkp, 14:28:32)
* Node needs to be dropped on el6, which, according to sbonazzo some
related job is still building RPMs for el6 (bkp, 14:28:35)
* Today (22.4.15) is feature submission deadline for 3.6 (bkp,
14:28:38)
* Submitted features need to be reviewed. (bkp, 14:28:40)
* An alpha release of 3.6 is scheduled for May 6 (bkp, 14:28:43)
* Open Discussion (bkp, 14:28:44)
* Inquiry about jdk 1.8 / wildfly support for dev infra, no initial
response. Using jdk 1.7 as a fallback is not optimal, sbonazzo
reports (bkp, 14:28:46)
* Node is building nicely and the team is cleaning up and preparing
for 3.6, fabiand reports (bkp, 14:28:49)
* On the issue of server hardening, dcaro reports that we are waiting
for the sec team to finish the checks and give more feedback on best
practices (bkp, 14:53:45)
* bkp suggests that office hours be held for set half-hour periods,
moderated by a rotation of community members. Format should be open,
answering all questions and discussing any cross-team/community-wide
issues as needed. Twice a week, Tuesdays and Thursdays. (bkp,
14:53:48)
* ACTION: bkp will ask on community mailing lists to finalize
discussion on office hour format (bkp, 14:53:52)
Meeting ended at 14:54:10 UTC.
Action Items
------------
* bkp will ask on community mailing lists to finalize discussion on
office hour format
Action Items, by person
-----------------------
* bkp
* bkp will ask on community mailing lists to finalize discussion on
office hour format
* **UNASSIGNED**
* (none)
People Present (lines said)
---------------------------
* bkp (73)
* sbonazzo (56)
* dcaro (14)
* fabiand (7)
* mr_chris (4)
* ovirtbot (2)
* awels (1)
* misc (1)
Generated by `MeetBot`_ 0.1.4
.. _`MeetBot`: http://wiki.debian.org/MeetBot
--
Brian Proffitt
Community Liaison
oVirt
Open Source and Standards, Red Hat - http://community.redhat.com
Phone: +1 574 383 9BKP
IRC: bkp @ OFTC
9 years, 7 months
Options not being passed fence_ipmilan, Ovirt3.5 on Centos 7.1 hosts
by Mike Lindsay
Hi All,
I have a bit of an issue with a new install of Ovirt 3.5 (our 3.4 cluster
is working fine) in a 4 node cluster.
When I test fencing (or cause a kernal panic triggering a fence) the
fencing fails. On investigation it appears that the fencing options are not
being passed to the fencing script (fence_ipmilan in this case):
Fence options under GUI(as entered in the gui): lanplus, ipport=623,
power_wait=4, privlvl=operator
from vdsm.log on the fence proxy node:
Thread-818296::DEBUG::2015-04-21 12:39:39,136::API::1209::vds::(fenceNode)
fenceNode(addr=x.x.x.x,port=,agent=ipmilan,user=stonith,passwd=XXXX,action=status,secure=False,options=
power_wait=4
Thread-818296::DEBUG::2015-04-21 12:39:39,137::utils::739::root::(execCmd)
/usr/sbin/fence_ipmilan (cwd None)
Thread-818296::DEBUG::2015-04-21 12:39:39,295::utils::759::root::(execCmd)
FAILED: <err> = 'Failed: Unable to obtain correct plug status or plug is
not available\n\n\n'; <rc> = 1
Thread-818296::DEBUG::2015-04-21 12:39:39,296::API::1164::vds::(fence) rc 1
inp agent=fence_ipmilan
Thread-818296::DEBUG::2015-04-21 12:39:39,296::API::1235::vds::(fenceNode)
rc 1 in agent=fence_ipmilan
Thread-818296::DEBUG::2015-04-21
12:39:39,297::stompReactor::163::yajsonrpc.StompServer::(send) Sending
response
from engine.log on the engine:
2015-04-21 12:39:38,843 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
Event ID: -1, Message: Host mpc-ovirt-node03 from cluster Default was
chosen as a proxy to execute Status command on Host mpc-ovirt-node04.
2015-04-21 12:39:38,845 INFO [org.ovirt.engine.core.bll.FenceExecutor]
(ajp--127.0.0.1-8702-4) Using Host mpc-ovirt-node03 from cluster Default as
proxy to execute Status command on Host
2015-04-21 12:39:38,885 INFO [org.ovirt.engine.core.bll.FenceExecutor]
(ajp--127.0.0.1-8702-4) Executing <Status> Power Management command, Proxy
Host:mpc-ovirt-node03, Agent:ipmilan, Target Host:, Management IP:x.x.x.x,
User:stonith, Options: power_wait=4, ipport=623, privlvl=operator,lanplus,
Fencing policy:null
2015-04-21 12:39:38,921 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
(ajp--127.0.0.1-8702-4) START, FenceVdsVDSCommand(HostName =
mpc-ovirt-node03, HostId = 5613a489-589d-4e89-ab01-3642795eedb8,
targetVdsId = dbfa4e85-3e97-4324-b222-bf40a491db08, action = Status, ip =
x.x.x.x, port = , type = ipmilan, user = stonith, password = ******,
options = ' power_wait=4, ipport=623, privlvl=operator,lanplus', policy =
'null'), log id: 774f328
2015-04-21 12:39:39,338 WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
Event ID: -1, Message: Power Management test failed for Host
mpc-ovirt-node04.Done
2015-04-21 12:39:39,339 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
(ajp--127.0.0.1-8702-4) FINISH, FenceVdsVDSCommand, return: Test Succeeded,
unknown, log id: 774f328
2015-04-21 12:39:39,340 WARN [org.ovirt.engine.core.bll.FenceExecutor]
(ajp--127.0.0.1-8702-4) Fencing operation failed with proxy host
5613a489-589d-4e89-ab01-3642795eedb8, trying another proxy...
2015-04-21 12:39:39,594 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
Event ID: -1, Message: Host mpc-ovirt-node01 from cluster Default was
chosen as a proxy to execute Status command on Host mpc-ovirt-node04.
2015-04-21 12:39:39,595 INFO [org.ovirt.engine.core.bll.FenceExecutor]
(ajp--127.0.0.1-8702-4) Using Host mpc-ovirt-node01 from cluster Default as
proxy to execute Status command on Host
2015-04-21 12:39:39,598 INFO [org.ovirt.engine.core.bll.FenceExecutor]
(ajp--127.0.0.1-8702-4) Executing <Status> Power Management command, Proxy
Host:mpc-ovirt-node01, Agent:ipmilan, Target Host:, Management IP:x.x.x.x,
User:stonith, Options: power_wait=4, ipport=623, privlvl=operator,lanplus,
Fencing policy:null
2015-04-21 12:39:39,634 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
(ajp--127.0.0.1-8702-4) START, FenceVdsVDSCommand(HostName =
mpc-ovirt-node01, HostId = c3e8be6e-ac54-4861-b774-17ba5cc66dc6,
targetVdsId = dbfa4e85-3e97-4324-b222-bf40a491db08, action = Status, ip =
x.x.x.x, port = , type = ipmilan, user = stonith, password = ******,
options = ' power_wait=4, ipport=623, privlvl=operator,lanplus', policy =
'null'), log id: 6369eb1
2015-04-21 12:39:40,056 WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ajp--127.0.0.1-8702-4) Correlation ID: null, Call Stack: null, Custom
Event ID: -1, Message: Power Management test failed for Host
mpc-ovirt-node04.Done
2015-04-21 12:39:40,057 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.FenceVdsVDSCommand]
(ajp--127.0.0.1-8702-4) FINISH, FenceVdsVDSCommand, return: Test Succeeded,
unknown, log id: 6369eb1
For verification I temporarily replaced /usr/sbin/fence_ipmilan with a
shell script that dumps the env plus any cli args passed into a log file:
-------------------------- Tue Apr 21 12:39:39 EDT 2015
----------------------------
ENV DUMP:
LC_ALL=C
USER=vdsm
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
PWD=/
LANG=en_CA.UTF-8
LIBVIRT_LOG_FILTERS=
SHLVL=1
HOME=/var/lib/vdsm
LOGNAME=vdsm
LIBVIRT_LOG_OUTPUTS=
_=/usr/bin/env
------------------------------
CLI DUMP:
<this is where the cli args should be listed>
Version info:
libvirt version: 1.2.8, package: 16.el7_1.2 (CentOS BuildSystem <
http://bugs.centos.org>, 2015-03-26-23:17:42, worker1.bsys.centos.org)
fence_ipmilan: 4.0.11 (built Mon Apr 13 13:22:18 UTC 2015)
vdsm.x86_64: 4.16.10-8.gitc937927.el7
ovirt-engine.noarch: 3.5.1.1-1.el6
Engine os: Centos 6.6
Host os: Centos 7.1.1503
I've found some old post from 2012 that describe the same problem. Has
anyone else run into this?
Any thought or suggestions would be appreciated.
Cheers,
Mike
9 years, 7 months
Is it possible to limit the number and speed of paralel STORAGE migrations?
by Ernest Beinrohr
This is a multi-part message in MIME format.
--------------000203080307020305010009
Content-Type: text/plain; charset=utf-8; format=flowed
Content-Transfer-Encoding: 7bit
Ovirt uses dd and qemu-img for live migration. Is it possible to limit
the number of concurrent live storage moves or limit the bandwidth used?
I'd like to move about 30 disks to another storage during the night, but
each takes about 30 minutes each and if more than one runs, it chokes my
storage.
--
Ernest Beinrohr, AXON PRO
Ing <http://www.beinrohr.sk/ing.php>, RHCE
<http://www.beinrohr.sk/rhce.php>, RHCVA
<http://www.beinrohr.sk/rhce.php>, LPIC
<http://www.beinrohr.sk/lpic.php>, VCA <http://www.beinrohr.sk/vca.php>,
+421-2-62410360 +421-903-482603
--------------000203080307020305010009
Content-Type: text/html; charset=utf-8
Content-Transfer-Encoding: 7bit
<html>
<head>
<meta http-equiv="content-type" content="text/html; charset=utf-8">
</head>
<body bgcolor="#FFFFFF" text="#000000">
Ovirt uses dd and qemu-img for live migration. Is it possible to
limit the number of concurrent live storage moves or limit the
bandwidth used?<br>
<br>
<br>
I'd like to move about 30 disks to another storage during the night,
but each takes about 30 minutes each and if more than one runs, it
chokes my storage. <br>
<br>
<div class="moz-signature">-- <br>
<div id="oernii_footer" style="color: gray;">
<span style="font-family: Lucida Console, Luxi Mono, Courier,
monospace; font-size: 90%;">
Ernest Beinrohr, AXON PRO<br>
<a style="text-decoration: none; color: gray;"
href="http://www.beinrohr.sk/ing.php">Ing</a>, <a
style="text-decoration: none; color: gray;"
href="http://www.beinrohr.sk/rhce.php">RHCE</a>, <a
style="text-decoration: none; color: gray;"
href="http://www.beinrohr.sk/rhce.php">RHCVA</a>, <a
style="text-decoration: none; color: gray;"
href="http://www.beinrohr.sk/lpic.php">LPIC</a>, <a
style="text-decoration: none; color: gray;"
href="http://www.beinrohr.sk/vca.php">VCA</a>, <br>
+421-2-62410360 +421-903-482603
<br>
</span> </div>
<img
src="http://nojsstats.appspot.com/UA-44497096-1/email.beinrohr.sk"
moz-do-not-send="true" border="0" width="1" height="1">
</div>
</body>
</html>
--------------000203080307020305010009--
9 years, 7 months
[ACTION NEEDED][QE] oVirt 3.6.0 status
by Sandro Bonazzola
Hi, here's an update on 3.6 status on integration / rel-eng side
The tracker bug for 3.6.0 [1] currently shows no blockers.
Repository closure is currently broken on Fedora 20 due to a missing required dependency on recent libvirt, a patch fixing this issue is currently
under review[4].
VDSM builds for EL6 are no more available on master snapshot.
ACTION: please review jenkins jobs relying on VDSM to be available on EL6.
There are 521 bugs [2] targeted to 3.6.0.
NEW ASSIGNED POST Total
<unassigned> 1 0 0 1
docs 10 0 0 10
external 1 0 0 1
gluster 32 26 16 74
i18n 2 0 0 2
infra 53 5 7 65
integration 35 3 9 47
network 42 1 9 52
node 26 3 2 31
sla 45 3 2 50
spice 1 0 0 1
storage 67 7 6 80
ux 27 0 2 29
virt 60 6 9 75
Total 402 54 62 518
Features submission is now CLOSED as per current release schedule.
ACTION: begin reviewing features tracked in the google doc[3]
On Integration side:
* Progress on Glance docker integration
* Progress on backup / restore RFEs
On Release engineering side:
* CentOS Virt SIG: CBS tag hierarchy created
* Fedora: pushed some dependencies for Fedora inclusion
* Worked with infra team fixing jenkins issues
ACTION: community members are welcome to join QE effort[5] by testing nightly master snapshot[6] on test systems
[1] https://bugzilla.redhat.com/1155425
[2] https://bugzilla.redhat.com/buglist.cgi?quicksearch=target_release%3A3.6....
[3] http://goo.gl/9X3G49
[4] https://gerrit.ovirt.org/39249
[5] http://www.ovirt.org/OVirt_Quality_Assurance
[6] http://www.ovirt.org/Install_nightly_snapshot
--
Sandro Bonazzola
Better technology. Faster innovation. Powered by community collaboration.
See how it works at redhat.com
9 years, 7 months
storage issue's with oVirt 3.5.1 + Nexenta NFS
by Maikel vd Mosselaar
Hi,
We are running ovirt 3.5.1 with 3 nodes and seperate engine.
All on CentOS 6.6:
3 x nodes
1 x engine
1 x storage nexenta with NFS
For multiple weeks we are experiencing issues of our nodes that cannot
access the storage at random moments (atleast thats what the nodes think).
When the nodes are complaining about a unavailable storage then the load
rises up to +200 on all three nodes, this causes that all running VMs
are unaccessible. During this process oVirt event viewer shows some i/o
storage error messages, when this happens random VMs get paused and will
not be resumed anymore (this almost happens every time but not all the
VMs get paused).
During the event we tested the accessibility from the nodes to the
storage and it looks like it is working normal, at least we can do a normal
"ls" on the storage without any delay of showing the contents.
We tried multiple things that we thought it causes this issue but
nothing worked so far.
* rebooting storage / nodes / engine.
* disabling offsite rsync backups.
* moved the biggest VMs with highest load to different platform outside
of oVirt.
* checked the wsize and rsize on the nfs mounts, storage and nodes are
correct according to the "NFS troubleshooting page" on ovirt.org.
The environment is running in production so we are not free to test
everything.
I can provide log files if needed.
Kind Regards,
Maikel
9 years, 7 months
Move/Migrate Storage Domain to new devices
by Dael Maselli
Hi,
I have a data storage domain that use one FC LUN. I need to move all
data to a new storage server.
I tried by move single disks to a new storage domain but some cannot be
moved, I think because they are thin-cloned by template.
When I worked with LVM I use to do a simple pvmove leaving the VG
intact, is there something similar (online or in maintenance) in oVirt?
Can I just do a pvmove from the SPM host o it's going to destroy everything?
Thank you very much.
Regards,
Dael Maselli.
--
___________________________________________________________________
Dael Maselli --- INFN-LNF Computing Service -- +39.06.9403.2214
___________________________________________________________________
* http://www.lnf.infn.it/~dmaselli/ *
___________________________________________________________________
Democracy is two wolves and a lamb voting on what to have for lunch
___________________________________________________________________
9 years, 7 months