[HEADS UP] CentOS 7.3 is rolling out, need qemu-kvm-ev 2.6
by Sandro Bonazzola
Hi,
as you probably noticed, CentOS 7.3 is rolling out right now and will be
announced pretty soon.
Please note that due to new libvirt shipped within CentOS 7.3, it is
required to update qemu-kvm-ev to 2.6 at the same time.
If you're using CentOS Virt SIG rpms / repos for oVirt this should happen
automatically.
If you're using only ovirt repositories be sure to have qemu-kvm-ev 2.6 if
you update to CentOS 7.3.
In terms of ovirt repositories, qemu-kvm-ev 2.6 is available right now in
ovirt-master-snapshot-static, ovirt-4.0-snapshot-static, and ovirt-4.0-pre
(contains 4.0.6 RC4 rpms going to be announced in a few minutes.)
Thanks,
--
Sandro Bonazzola
Better technology. Faster innovation. Powered by community collaboration.
See how it works at redhat.com
7 years, 11 months
Using zRam with oVirt Nodes
by Fernando Frediani
Hello folks.
On simple libvirt/KVM hosts in order to improve RAM usage and avoid swap
to disk I use zRam with swap to RAM technique. So create half of amount
of host memory in zRam disk divided by the number of CPU cores. Works
pretty well.
Has anyone tried it with oVirt Nodes or even has it been considered as a
feature to simply turn on/off ?
Thanks
Fernando
7 years, 11 months
Re: [ovirt-users] HA Score
by knarra
On 01/01/2017 06:23 PM, Doron Fediuck wrote:
> Care to share the HE agents log files?
Hi Doron,
I have collected logs from all the machines and placed it in the
link below.
http://rhsqe-repo.lab.eng.blr.redhat.com/sosreports/HC/ovirt_logs/
Thanks
kasturi
> On Fri, Dec 30, 2016 at 9:17 AM, knarra <knarra(a)redhat.com> wrote:
>> Hi,
>>
>> I have latest 4.1 installed and i see that HA score on hosts keeps going
>> to 0 and comes back to 3400. This behavior is something which i am
>> observing with 4.1 and i see that it takes considerable amount of time to
>> get back to normal state. Any reason why it takes such a long time?
>>
>> Thanks
>>
>> kasturi.
>>
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
7 years, 11 months
New oVirt user
by Jim Kusznir
Hello:
I've been involved in virtualization from its very early days, and been
running linux virtualization solutions off and on for a decade.
Previously, I was always frustrated with the long feature list offered by
many linux virtualization systems but with no reasonable way to manage
that. It seemed that I had to spend an inordinate amount of time doing
everything by hand. Thus, when I found oVirt, I was ecstatic!
Unfortunately, at that time I changed employment (or rather left employment
and became self-employed), and didn't have any reason to build my own virt
cluster..until now!
So I'm back with oVirt, and actually deploying a small 3-node cluster. I
intend to run on it:
VoIP Server
Web Server
Business backend server
UniFi management server
Monitoring server (zabbix)
Not a heavy load, and 3 servers is probably overkill, but I need this to
work, and it sounds like 3 is the magic entry level for all the
cluster/failover stuff to work. For now, my intent is to use a single SSD
on each node with gluster for the storage backend. I figure if all the
failover stuff actually working, if I loose a node due to disk failure, its
not the end of the world. I can rebuild it, reconnect gluster, and restart
everything. As this is for a startup business, funds are thin at the
moment, so I'm trying to cut a couple corners that don't affect overall
reliability. If this side of the business grows more, I would likely
invest in some dedicated servers.
So far, I've based my efforts around this guide on oVirt's website:
http://www.ovirt.org/blog/2016/08/up-and-running-with-ovirt-4-0-and-glust...
My cluster is currently functioning, but not entirely correctly. Some of
it is gut feel, some of it is specific test cases (more to follow). First,
some areas that lacked clarity and the choices I made in them:
Early on, Jason talks about using a dedicated gluster network for the
gluster storage sync'ing. I liked that idea, and as I had 4 nics on each
machine, I thought dedicating one or two to gluster would be fine. So, on
my clean, bare machines, I setup another network with private NiCs and put
it on a standalone switch. I added hostnames with a designator (-g on the
end) for the IPs for all three nodes into /etc/hosts on all three nodes so
now each node can resolve itself and the other nodes on the -g name (and
private IP) as well as their main host name and "more public" (but not
public) IP.
Then, for gdeploy, I put the hostnames in as the -g hostnames, as I didn't
see anywhere to tell gluster to use the private network. I think this is a
place I went wrong, but didn't realize it until the end....
I set up the gdeploy script (it took a few times, and a few OS rebuilds to
get it just right...), and ran it, and it was successful! When complete, I
had a working gluster cluster and the right software installed on each node!
I set up the engine on node1, and that worked, and I was able to log in to
the web gui. I mistakenly skipped the web gui enable gluster service
before doing the engine vm reboot to complete the engine setup process, but
I did go back in after the reboot and do that. After doing that, I was
notified in the gui that there were additional nodes, did I want to add
them. Initially, I skipped that and went back to the command line as Jason
suggests. Unfortunately, it could not find any other nodes through his
method, and it didn't work. Combine that with the warnings that I should
not be using the command line method, and it would be removed in the next
release, I went back to the gui and attempted to add the nodes that way.
Here's where things appeared to go wrong...It showed me two additional
nodes, but ONLY by their -g (private gluster) hostname. And the ssh
fingerprints were not populated, so it would not let me proceed. After
messing with this for a bit, I realized that the engine cannot get to the
nodes via the gluster interface (and as far as I knew, it shouldn't).
Working late at night, I let myself "hack it up" a bit, and on the engine
VM, I added /etc/hosts entries for the -g hostnames pointing to the main
IPs. It then populated the ssh host keys and let me add them in. Ok, so
things appear to be working..kinda. I noticed at this point that ALL
aspects of the gui became VERY slow. Clicking in and typing in any field
felt like I was on ssh over a satellite link. Everything felt a bit worse
than the early days of vSphere....Painfully slow. but it was still
working, so I pressed on.
I configured gluster storage. Eventually I was successful, but initially
it would only let me add a "Data" storage domain, the drop-down menu did
NOT contain iso, export, or anything else... Somehow, on its own, after
leaving and re-entering that tab a few times, iso and export materialized
on their own in the menu, so I was able to finish that setup.
Ok, all looks good. I wanted to try out his little tip on adding a VM,
too. I saw "ovirt-imiage-repository" in the "external providers" section,
but he mentioned it in the storage section. It wasn't there on mine, and
in external providers, I couldn't find anyway to do anything useful. I
tried and fumbled with this, and still, I have not figured out how to use
this feature. It would be nice....
Anyway, I moved on for now. As I was skeptical that things were set up
correctly, i tried putting node 1 (which was running my engine, and was NOT
set up with the -g hostname) into maintence mode, to see if it really did
smoothly failover. It failed to go into maintence mode (left it for 12
hours, too!). I suspect its because of the hostnames/networks in use.
Oh, I forgot to mention...I did follow the instructions in Jason's guide to
set up the gluster network in ovirt and map that to the right physical
interface on all 3 nodes. I also moved migration from the main network to
the gluster network as Jason had suggested.
So...How badly did I do? How do I fix the issues? (I'm not opposed to
starting from scratch again, either...I've already done that 3-4 times in
the early phases of getting the gdeploy script down, and I already have
kickstart files setup with a network environment...I was rebuilding that
often! I just need to know how to fix my setup this time....)
I do greatly appreciate others' help and insight. I am in the IRC channel
under kusznir currently, too.
--Jim
7 years, 11 months
Current status of 4.0.6 | EL7.3?
by Derek Atkins
Hi,
What is the current status of Ovirt 4.0.6 and EL7.3? From previous
threads it seemed to me that there was a potential compatibility issue
with the 7.3 kernel and an updated version of vdsm or qemu? I just want
to ensure any potential issues have been cleared up before I upgrade my
systems.
Thanks,
-derek
--
Derek Atkins 617-623-3745
derek(a)ihtfp.com www.ihtfp.com
Computer and Internet Security Consultant
7 years, 11 months
Re: [ovirt-users] Watchdog device
by Michal Skrivanek
> On 01 Jan 2017, at 15:18, Doron Fediuck <dfediuck(a)redhat.com> wrote:
>
> Hi Gary,
> this is a known issue we're working on. if the VM is down it should
> work as expected (edit the VM and close the dialog).
Engine.log might help. Can you attach it? Supposing you mean the VM is
created and not running, it should work..
> Please try and let me know. Another option is to do this until the bug
> is resolved is using the REST API.
>
> Note that this is currently working on on Linux guests (there are no
> updated Windows drivers). You will need to configure
> the device in the guest as explained by the docs.
>
> Doron
>
>
>> On Sun, Jan 1, 2017 at 9:28 AM, Gary Pedretty <gary(a)ravnalaska.net> wrote:
>> How do you add the supported watchdog device to a VM that has already been
>> created? I have tried adding it via the High Availability tab in the Edit
>> VM Dialog, but it does not retain the setting when I close the dialog.
>>
>> Latest version os Overt running as self-hosted engine
>>
>> Gary
>>
>>
>> ------------------------------------------------------------------------
>> Gary Pedretty gary(a)ravnalaska.net
>> Systems Manager www.flyravn.com
>> Ravn Alaska /\ 907-450-7251
>> 5245 Airport Industrial Road / \/\ 907-450-7238 fax
>> Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment
>> Serving Alaska's Interior / \/ /\ \ \/\ "Love your neighbor as
>> Having a heatwave, its summer yourself” Matt 22:39
>> ------------------------------------------------------------------------
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
7 years, 11 months
Re: [ovirt-users] hosted_storage locked
by Evgenia Tokar
It seems that the import of the storage domain failed since it was
inaccessible:
2016-12-28 07:06:25,406 ERROR
[org.ovirt.engine.core.bll.storage.domain.GetExistingStorageDomainListQuery]
(org.ovirt.thread.pool-6-thread-46) [] Exception:
org.ovirt.engine.core.common.errors.EngineException: EngineException:
org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException:
VDSGenericException: VDSErrorException: Failed to
HSMGetStorageDomainInfoVDS, error = Domain is either partially accessible
or entirely inaccessible: (u'247b250c-ccd3-4056-853f-f03a05369da9',), code
= 379 (Failed with error StorageDomainAccessError and code 379)
....
2016-12-28 07:06:25,413 ERROR
[org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-46) [] Failed query for all Storage
Domains. The import command can not proceed without this info
Once that happened the command was logged in the db as failed, and every
engine restart the engine tries to run it again (compensate mechanism).
However, since the import storage domain command is not "compensatable"
this fails and the domain remains locked.
2016-12-28 07:09:02,643 ERROR [org.ovirt.engine.core.bll.CommandsFactory]
(ServerService Thread Pool -- 56) [] CommandsFactory : Failed to get type
information using reflection for Class
'org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand',
Command Id '2ab30734-bbe4-4f50-897f-f1b393a31852':
org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand.<init>(org.ovirt.engine.core.compat.Guid)
2016-12-28 07:09:02,643 ERROR [org.ovirt.engine.core.bll.CommandsFactory]
(ServerService Thread Pool -- 56) [] Exception:
java.lang.NoSuchMethodException:
org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand.<init>(org.ovirt.engine.core.compat.Guid)
....
2016-12-28 07:09:02,646 ERROR [org.ovirt.engine.core.bll.Backend]
(ServerService Thread Pool -- 56) [] Failed to run compensation on startup
for Command
'org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand',
Command Id '2ab30734-bbe4-4f50-897f-f1b393a31852'
I would suggest:
1. Verify the you can access the storage domain.
2. If you can access the domain, you can try removing it from the ui (I
think the destroy option should be available).
After the removal the engine will try to reimport it, and hopefully
will succeed.
Removing the storage domain may cause the engine and the ha-agent to
restart.
Jenny
On Wed, Dec 28, 2016 at 6:29 PM, Gary Pedretty <gary(a)ravnalaska.net> wrote:
> Here is the engine log from the engine vm.
>
> Gary
>
>
>
>
> ------------------------------------------------------------------------
> Gary Pedretty gary(a)ravnalaska.net
> <gary(a)eraalaska.net>
> Systems Manager www.flyravn.com
> Ravn Alaska /\ 907-450-7251
> <(907)%20450-7251>
> 5245 Airport Industrial Road / \/\ 907-450-7238
> <(907)%20450-7238> fax
> Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment
> Serving All of Alaska / \/ /\ \ \/\ “Love your neighbor as
> Really loving the record green up date! Summmer!! yourself” Matt 22:39
> ------------------------------------------------------------------------
>
>
>
>
>
>
>
>
>
>
>
>
> On Dec 28, 2016, at 1:12 AM, Evgenia Tokar <etokar(a)redhat.com> wrote:
>
> Hi!
>
> Can you attach the engine log from the hosted engine vm?
>
> Thanks,
> Jenny
>
>
> On Wed, Dec 28, 2016 at 10:56 AM, Gary Pedretty <gary(a)ravnalaska.net>
> wrote:
>
>> Just setup a new hosted-engine Data Center. With 3 hosts and gluster
>> based storage domains. Used the latest installers. Everything went well,
>> until I got the engine deployed and each host setup and went to the web
>> interface for the first time.
>>
>> After adding the master storage domain, the hosted_storage added itself
>> like usual, but remained locked. Everything is working just fine with the
>> exception of no listing of the engine as a VM under that tab or the icon on
>> the current host. The Volumes tab does not show anything on the
>> hosted_storage, but all the other volumes including the master data and a
>> second data all show normal status for bricks. Error log is not reporting
>> anything unusual either. All the gluster volumes including the engine
>> (hosted_storage) show good status via command line in the hosts.
>>
>> Going to let it run for a day or so before I add any VMs.
>>
>> Any ideas?
>>
>> Gary
>>
>> ------------------------------------------------------------------------
>> Gary Pedretty gary(a)ravnalaska.net
>> <gary(a)eraalaska.net>
>> Systems Manager www.flyravn.com
>> Ravn Alaska /\ 907-450-7251
>> <(907)%20450-7251>
>> 5245 Airport Industrial Road / \/\ 907-450-7238
>> <(907)%20450-7238> fax
>> Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment
>> Serving All of Alaska / \/ /\ \ \/\ “Love your neighbor as
>> Really loving the record green up date! Summmer!! yourself” Matt 22:39
>> ------------------------------------------------------------------------
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
>
7 years, 11 months