The best HCI config with 8Nodes and 2 sites?
by Arman Khalatyan
Hello everybody,
I would like to deploy HCI with our 2 buildings each with 8 compute nodes.
Each host has a mirrored OS disks and 1 slot for the SSD. So I will use SSD
for the glusterfs.
my question is what is the best type of the glusterfs volume?
I can leave with 8way mirror but what happened if the connection between
buildings will go down?
where will my ovirt-engine start?
3 years, 9 months
One host 4.4.5 upgade failed
by ozmen62@hotmail.com
Hi,
We have 2 hosts, upgarded to 4.4.5.
Engine and one of the hosts was successfull
But, of of the hosts stucked and doesnt upgrade.
successful one has these kernels
# vmlinuz-4.18.0-240.1.1.el8_3.x86_64
# vmlinuz-4.18.0-240.15.1.el8_3.x86_64
Unsuccessful one has;
# vmlinuz-4.18.0-193.28.1.el8_2.x86_64
# vmlinuz-4.18.0-240.1.1.el8_3.x86_64
Unsuccessful one has "/usr/share/ovirt-node-ng-image-update/" directory and image but we dont make it upgrade
This is the some part of the log file from engine;
FAILED! => {"changed": true, "cmd": ["vdsm-tool", "configure", "--force"], "delta": "0:00:10.773484", "end": "2021-03-23 09:00:08.285785", "msg": "non-zero return code", "rc": 1, "start": "2021-03-23 08:59:57.512301", "stderr": "Error: ServiceOperationError: _systemctlStop failed\nb'Job for vdsmd.service canceled.\\n' ", "stderr_lines": ["Error: ServiceOperationError: _systemctlStop failed", "b'Job for vdsmd.service canceled.\\n' "], "stdout": "\nChecking configuration status...\n\nlibvirt is already configured for vdsm\nSUCCESS: ssl configured to true. No conflicts\nsanlock is configured for vdsm\nlvm is configured for vdsm\nCurrent revision of multipath.conf detected, preserving\nabrt is already configured for vdsm\nManaged volume database is already configured", "stdout_lines": ["", "Checking configuration status...", "", "libvirt is already configured for vdsm", "SUCCESS: ssl configured to true. No conflicts", "sanlock is configured for vdsm", "lvm is configured for vdsm", "Cur
rent revision of multipath.conf detected, preserving", "abrt is already configured for vdsm", "Managed volume database is already configured"]}
when we triggered the dnf update/upgrade it says "there is no upgrade" both engine web page and cli
Is there any suggestion for upgrading the host, it's stucked on "NonOperational" mode
3 years, 9 months
user portal
by Enrico Becchetti
Dear all,
I'm using ovirt 4.3.2 with its engine on a virtual machine. The nodes
are all Centos 7.7.
Both engine and hypervisor systems work on a 10.0.0.0 private network.
Now I would like to let users access the ovirt web page (user portal)
and for this
I must necessarily add a second network interface to the engine by
inserting a public ip. I can't use NAT.
Can you give me any advice for this operation ?
Can I add the network interface and then run engine-setup ?
Will oVirt be accessible from both ip addresses at the end of this
operation ?
Lots of thanks.
Enrico
--
_______________________________________________________________________
Enrico Becchetti Servizio di Calcolo e Reti
Istituto Nazionale di Fisica Nucleare - Sezione di Perugia
Via Pascoli,c/o Dipartimento di Fisica 06123 Perugia (ITALY)
Phone:+39 075 5852777 Skype:enrico_becchetti
Mail: Enrico.Becchetti<at>pg.infn.it
______________________________________________________________________
3 years, 9 months
What do some of the terms mean in the documentation?
by jenia.ivlev@gmail.com
Hello
I want to install oVirt virtualization software on my ArchLinux as a lab for learning and experimenting.
I'm having trouble going through the documentation because I don't understand how the different pieces fit together.
Specifically, what are the roles of the oVirt Node, oVirt Host and oVirt Engine?
Thanks
3 years, 9 months
No option to remove OpenStack Glance storage domain.
by Gary Taylor
Hi,
I was playing around a few months ago and added an ovirt-image-repository storage domain. There wasn't a real purpose for doing it, just trying to learn and play around. I tried to remove it earlier but couldn't figure out how because the Remove button is greyed out. I got busy and forgot about it. I am trying to clean it up now but the Remove button is still greyed out for that domain. How do I get rid of it now? It isn't being used. It's unattached. I'm the admin.
https://imgur.com/KBKUu16.png
oVirt Open Virtualization Manager
Software Version:4.4.4.7-1.el8
Thank-you,
Gary
3 years, 9 months
libvirtd.service is down
by jenia.ivlev@gmail.com
Hello.
I have an error in my oVirt node 4.4: internal error: "Network is already in use by interface ens3"
This is the output from journalctl:
[root@localhost ~]# sudo journalctl --unit libvirtd.service
-- Logs begin at Sun 2021-03-21 16:13:11 UTC, end at Sun 2021-03-21 16:20:37 UTC. --
Mar 21 16:13:24 localhost.localdomain systemd[1]: Starting Virtualization daemon...
Mar 21 16:13:25 localhost.localdomain systemd[1]: Started Virtualization daemon.
Mar 21 16:13:25 localhost.localdomain libvirtd[1626]: libvirt version: 6.6.0, package: 13.el8 (CBS <cbs(a)centos.org>, 2021-03-01-13:29:02, )
Mar 21 16:13:25 localhost.localdomain libvirtd[1626]: hostname: localhost.localdomain
Mar 21 16:13:25 localhost.localdomain libvirtd[1626]: internal error: Network is already in use by interface ens3
Mar 21 16:15:25 localhost.localdomain systemd[1]: libvirtd.service: Succeeded.
Mar 21 16:19:30 localhost.localdomain systemd[1]: Starting Virtualization daemon...
Mar 21 16:19:30 localhost.localdomain systemd[1]: Started Virtualization daemon.
And this is the output from systemctl:
[root@localhost ~]# sudo systemctl status libvirtd.service
â libvirtd.service - Virtualization daemon
Loaded: loaded (/usr/lib/systemd/system/libvirtd.service; enabled; vendor preset: enabled)
Drop-In: /etc/systemd/system/libvirtd.service.d
ââunlimited-core.conf
Active: inactive (dead) since Sun 2021-03-21 16:15:25 UTC; 2min 36s ago
Docs: man:libvirtd(8)
https://libvirt.org
Process: 1626 ExecStart=/usr/sbin/libvirtd $LIBVIRTD_ARGS (code=exited, status=0/SUCCESS)
Main PID: 1626 (code=exited, status=0/SUCCESS)
Mar 21 16:13:24 localhost.localdomain systemd[1]: Starting Virtualization daemon...
Mar 21 16:13:25 localhost.localdomain systemd[1]: Started Virtualization daemon.
Mar 21 16:13:25 localhost.localdomain libvirtd[1626]: libvirt version: 6.6.0, package: 13.el8 (CBS <cbs(a)centos.org>, 2021-03-01-13:29:02, )
Mar 21 16:13:25 localhost.localdomain libvirtd[1626]: hostname: localhost.localdomain
Mar 21 16:13:25 localhost.localdomain libvirtd[1626]: internal error: Network is already in use by interface ens3
Mar 21 16:15:25 localhost.localdomain systemd[1]: libvirtd.service: Succeeded.
I managed to successfully start libvirtd.service by doing systemctl start libvirtd.service but I want to correct this bug somehow.
Anyone encountered this before?
Thanks
3 years, 9 months
Re: oVirt process that writes to Gluster mount
by Ben
Well, I'll be -- you're absolutely right, and I'm a bit embarrassed I
didn't consider that before. The node that's not healing shows connections
from 2 FUSE clients, which I expect, and 3 glustershd, which I also expect.
[root@gluster1 ~]# gluster volume status ssd-san client-list
Client connections for volume ssd-san
Name count
----- ------
glustershd 3
fuse 2
total clients for volume ssd-san : 5
-----------------------------------------------------------------
But the secondary node, which is constantly healing, *shows that it's
missing a FUSE connection*:
[root@gluster2 ~]# gluster volume status ssd-san client-list
Client connections for volume ssd-san
Name count
----- ------
glustershd 3
fuse 1
total clients for volume ssd-san : 4
-----------------------------------------------------------------
I had to restart the glusterd service on node 2 twice before the FUSE
client reconnected and stayed connected.
Thanks a ton, I really appreciate your help!
On Mon, Mar 22, 2021 at 12:37 AM Strahil Nikolov <hunter86_bg(a)yahoo.com>
wrote:
> Healing must happen only after a maintenance (for example patch + reboot)
> on any of the nodes .
> Once the node is up , the FUSE client (any host) should reconnect to all
> gluster bricks and write to all bricks simultaneously.
>
> If you got constant healing, this indicates that a client is not writing
> to all bricks.
>
> Check with the following command if there is such client:
> 'gluster volume status *VOLNAME* clients'
>
>
> Best Regards,
> Strahil Nikolov
>
> On Mon, Mar 22, 2021 at 3:24, Ben
> <gravyfish(a)gmail.com> wrote:
> Sorry, just saw this -- I'm not sure I understand what you mean, but in
> any case, the healing process does complete when I stop all of my VMs,
> which I believe indicates that something about the oVirt writes to Gluster
> is causing the problem in the first place.
>
> On Sun, Mar 14, 2021 at 8:06 AM Strahil Nikolov <hunter86_bg(a)yahoo.com>
> wrote:
>
> Are you sure that gluster volume client's count is the same on all nodes ?
>
> Best Regards,
> Strahil Nikolov
>
> On Sat, Mar 13, 2021 at 23:58, Ben
> <gravyfish(a)gmail.com> wrote:
> Hi, I could use some help with a problem I'm having with the Gluster
> storage servers I use in my oVirt data center. I first noticed the problem
> when files would constantly heal after rebooting one of the Gluster nodes
> -- in the replica 2/arbiter, the node that remained online and the arbiter
> would begin healing files and never finish.
>
> I raised the issue with the helpful folks over at Gluster:
> https://github.com/gluster/glusterfs/issues/2226
>
> The short version is this: after running a tcpdump and noticing malformed
> RPC calls to Gluster from one of my oVirt nodes, they're looking for a
> stack trace of whatever process is running I/O on the Gluster cluster from
> oVirt in order to figure out what it's doing and if the write problems
> could cause the indefinite healing I'm seeing. After checking the qemu
> PIDs, it doesn't look like they are actually performing the writes -- is
> there a particular part of the oVirt stack I can look at to find the write
> operations to Gluster? I don't see anything else doing read/write on the VM
> image files on the Gluster mount, but I could be missing something.
>
> NB: I'm using a traditional Gluster setup with the FUSE client, not
> hyperconverged.
>
> Thanks in advance for any assistance.
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/76H5BY6IH55...
>
>
3 years, 9 months
oVirt process that writes to Gluster mount
by Ben
Hi, I could use some help with a problem I'm having with the Gluster
storage servers I use in my oVirt data center. I first noticed the problem
when files would constantly heal after rebooting one of the Gluster nodes
-- in the replica 2/arbiter, the node that remained online and the arbiter
would begin healing files and never finish.
I raised the issue with the helpful folks over at Gluster:
https://github.com/gluster/glusterfs/issues/2226
The short version is this: after running a tcpdump and noticing malformed
RPC calls to Gluster from one of my oVirt nodes, they're looking for a
stack trace of whatever process is running I/O on the Gluster cluster from
oVirt in order to figure out what it's doing and if the write problems
could cause the indefinite healing I'm seeing. After checking the qemu
PIDs, it doesn't look like they are actually performing the writes -- is
there a particular part of the oVirt stack I can look at to find the write
operations to Gluster? I don't see anything else doing read/write on the VM
image files on the Gluster mount, but I could be missing something.
NB: I'm using a traditional Gluster setup with the FUSE client, not
hyperconverged.
Thanks in advance for any assistance.
3 years, 9 months
oVirt4.4.5 and gluster op-version
by Jiří Sléžka
Hello,
I have just upgraded my 2hosts+1arbiter hci cluster to 4.4.5. Gluster is
not managed by oVirt and currently is in op.version 70200
gluster volume get all cluster.op-version
cluster.op-version 70200
after gluster upgrade max-op-version is
gluster volume get all cluster.max-op-version
cluster.max-op-version 80000
can I (or should I) switch to 80000 op-version?
Thanks,
Jiri
3 years, 9 months