Engine across Clusters
by Abe E
Has anyone setup hype converged gluster (3Nodes) and then added more after while maintaining access to the engine?
An oversight on my end was 2 fold, Engine gluster being on engine nodes and new nodes requiring their own cluster due to different CPU type.
So basically I am trying to see if I can setup a new cluster for my other nodes that require it while trying to give them ability to run the engine and ofcourse because they arent part of the engine cluster, we all know how that goes. Has anyone dealt with this or worked around it, any advices?
2 years, 1 month
Gluster issue with brick going down
by Chris Adams
I have a hyper-converged cluster running oVirt 4.4.10 and Gluster 8.6.
Periodically, one brick of one volume will drop out, but it's seemingly
random as to which volume and brick is affected. All I see in the brick
log is:
[2022-03-19 13:27:36.360727] W [MSGID: 113075] [posix-helpers.c:2135:posix_fs_health_check] 0-vmstore-posix: aio_read_cmp_buf() on /gluster_bricks/vmstore/vmstore/.glusterfs/health_check returned ret is -1 error is Structure needs cleaning
[2022-03-19 13:27:36.361160] M [MSGID: 113075] [posix-helpers.c:2214:posix_health_check_thread_proc] 0-vmstore-posix: health-check failed, going down
[2022-03-19 13:27:36.361395] M [MSGID: 113075] [posix-helpers.c:2232:posix_health_check_thread_proc] 0-vmstore-posix: still alive! -> SIGTERM
Searching around, I see references to similar issues, but no real
solutions. I see a suggestion that changing the health-check-interval
from 10 to 30 seconds helps, but it looks like 30 seconds is the default
with this version of Gluster (and I don't see it explicitly set for any
of my volumes).
While "Structure needs cleaning" appears to be an XFS filesystem error,
I don't see any XFS errors from the kernel.
This is a low I/O cluster - the storage network is on two 10 gig
switches with a two-port LAG to each server, but typically is only
seeing a few tens of megabits per second.
--
Chris Adams <cma(a)cmadams.net>
2 years, 1 month
Unable to deploy ovirt 4.4 on alma 8.5
by Richa Gupta
Hi Team,
While installing ovirt 4.4 on alma Linux 8.5 we are facing following issue:
[ INFO ] TASK [ovirt.ovirt.engine_setup : Install oVirt Engine package]
[ ERROR ] fatal: [localhost -> 192.168.222.56]: FAILED! => {"changed": false, "msg": "Failed to download metadata for repo 'ovirt-4.4': Cannot download repomd.xml: Cannot download repodata/repomd.xml: All mirrors were tried", "rc": 1, "results": []}
Using repo as following:
[ovirt-4.4]
name=Latest oVirt 4.4 Release
#baseurl=https://resources.ovirt.org/pub/ovirt-4.4/rpm/el$releasever/
mirrorlist=https://mirrorlist.ovirt.org/mirrorlist-ovirt-4.4-el$releasever
enabled=1
countme=1
fastestmirror=1
gpgcheck=1
gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-ovirt-4.4
Can someone please help in resolving this issue?
2 years, 1 month
all hosts lost scenario - restore engine from backup file - LOCKED HOST 4.3 version
by goosesk blabla
Hi,
I have problem with LOCKED host after restore of engine.
i am trying many disaster scenarios and one from them is loosing of all hosts with self hosted engine included with backup file of engine only.
When all hosts and engine VM were destroyed at the same time, i installed new self hosted engine on hardware from one lost host. Then tried restore engine
engine-backup --mode=restore --scope=all --file=backup/file/ovirt-engine-backup --log=backup/log/ovirt-engine-backup.log --no-restore-permissions --provision-db --provision-dwh-db –provision-reports-db
After this, lost engine was restored successfully. I had offline Datacentrer, dead hosts and VM. I added new hosts, which were able to connect to domain storage automatically and I was able to start VM.
New host cannot has the same IP as already dead host or hardware UID. This can be solved by setup old dead host to maintenance mode and then delete host. Then the same hardware and IP can be reused.
But, host where old engine was running is LOCKED. You cannot migrate engine VM, cannot start hosted engine from new engine GUI, cannot setup host to maintenance mode and delete.. This is problem when your hardware for hosts is limited.
I would like to ask how to solve this situation, Is there any way how to “reinstall” old hosts? I see that SSL certificates were changed with new installation of new hosts, but I don’t know if there is way how to enable old dead hosts.
Is there any way how to destroy old engine VM to add host back to work ?
Thank you
2 years, 1 month
Hosted engine deployment fails with storage domaine error
by Eugène Ngontang
Hi,
Hope you are well on your end.
I'm still trying to setup a brand new hosted engine but I'm having "Hosted
engine deployment fails with storage domaine error" issue, like shown in
the screenshot.
I can't figure out what's going on, I have checked the NFS server and
everything seems to be working fine, as well as the nfs mount point on the
client side (RHV host).
Please can one here help moving fast with this troubleshooting? Any idea?
Regards,
Eugène NG
--
LesCDN <http://lescdn.com>
engontang(a)lescdn.com
------------------------------------------------------------
*Aux hommes il faut un chef, et au*
* chef il faut des hommes!L'habit ne fait pas le moine, mais lorsqu'on te
voit on te juge!*
2 years, 1 month
Re: mdadm vs. JBOD
by jonas@rabe.ch
Thanks to Nikolov and Strahil for the valuable input! I was off for a few weeks, so I would like to apologize if I'm potentially reviving a zombie thread.
I am a bit confused about where to go with this environment after the discontinuation of the hyperconverged setup. What alternative options are there for us? Or do you think going the Gluster way would still be advisable, even though it seems as it is being discontinued over time?
Thanks for any input on this!
Best regards,
Jonas
On 1/22/22 14:31, Strahil Nikolov via Users wrote:
> Using the wizzard is utilizing the Gluster Andible roles.
> I would highly recommend using it, unless you know what you are doing (for example storage alignment when using Hardware raid).
>
> Keep in mind that the DHT xlator (the logic in distributed volumes) is shard aware, so your shards are spread between subvolumes and additional performance can be gained.So using replicated-distributed volumes have their benefits.
>
> If you decide to avoid the software raid, use only replica3 volumes as with SSDs/NVMEs usually the failures are not physical, but logical (maximum writes reached -> predictive failure -> total failure).
>
> Also, consider mounting via noatime/relatime and context="system_u:object_r:glusterd_brick_t:s0" for your gluster bricks.
>
> Best Regards,
> Strahil Nikolov
>
>
> > On Fri, Jan 21, 2022 at 11:00, Gilboa Davara
> > <gilboad(a)gmail.com> mailto:gilboad@gmail.com wrote:
> >
> >
> > _______________________________________________
> > Users mailing list -- users(a)ovirt.org mailto:users@ovirt.org
> > To unsubscribe send an email to users-leave(a)ovirt.org mailto:users-leave@ovirt.org
> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> > https://lists.ovirt.org/archives/list/users@ovirt.org/message/U2ZEWLRF5D6...
> >
> >
>
2 years, 1 month