Using NFS on Gluster
by Indivar Nair
Hi All,
We would like to set up a 2 node mirrored Gluster Storage using the
ovirt installation tools, but use it as an NFS storage.
We want to do this for 2 reasons -
1. We have a high-speed link between both the Gluster Nodes which we would
like to use for brick replication rather than ovirt node's 1G link
2. We have fast SSD disks on the ovirt nodes which we would like to use for
local NFS Caching (FS Caching) on the VM Images.
In short, we would like to manage Gluster stroage using the ovirt web
interface, but access the Gluster storage as an NFS Server.
How can we do this?
Thanks.
Regards,
Indivar Nair
5 years, 8 months
Re: OVirt Gluster Fail
by Strahil
Hi Andrea,
My guess is that while node2 was in maintenance , node3 brick(s) have died, or there were some pending heals.
For backup, you can use anything that works for KVM, but the hard part is to get the configuration of each VM. If the VM is running, you can use 'virsh dumpxml domain' to get the configuration of the running VM, but this won't work for VM that are off.
Why firewalld was not stopped - my guess is a rare bug that is hard to reproduce.
Best Regards,
Strahil Nikolov
On Mar 26, 2019 17:10, Andrea Milan <commramius(a)tiscali.it> wrote:
>
> Hi Sahina, Strahil
>
>
>
> thank you for the information, I managed to start the heal and restore both the hosted engine and the Vms.
>
>
>
> This is the log’s on all nodes
>
>
>
> [2019-03-26 08:30:58.462329] I [MSGID: 104045] [glfs-master.c:91:notify] 0-gfapi: New graph 676c6e6f-6465-3032-2e61-736370642e6c (0) coming up
>
> [2019-03-26 08:30:58.462364] I [MSGID: 114020] [client.c:2356:notify] 0-asc-client-0: parent translators are ready, attempting connect on transport
>
> [2019-03-26 08:30:58.464374] I [MSGID: 114020] [client.c:2356:notify] 0-asc-client-1: parent translators are ready, attempting connect on transport
>
> [2019-03-26 08:30:58.464898] I [rpc-clnt.c:1965:rpc_clnt_reconfig] 0-asc-client-0: changing port to 49438 (from 0)
>
> [2019-03-26 08:30:58.466148] I [MSGID: 114020] [client.c:2356:notify] 0-asc-client-3: parent translators are ready, attempting connect on transport
>
> [2019-03-26 08:30:58.468028] E [socket.c:2309:socket_connect_finish] 0-asc-client-0: connection to 192.170.254.3:49438 failed (Nessun instradamento per l'host)
>
> [2019-03-26 08:30:58.468054] I [rpc-clnt.c:1965:rpc_clnt_reconfig] 0-asc-client-1: changing port to 49441 (from 0)
>
> [2019-03-26 08:30:58.470040] I [rpc-clnt.c:1965:rpc_clnt_reconfig] 0-asc-client-3: changing port to 49421 (from 0)
>
> [2019-03-26 08:30:58.471345] I [MSGID: 114057] [client-handshake.c:1440:select_server_supported_programs] 0-asc-client-1: Using Program GlusterFS 3.3, Num (1298437), Version (330)
>
> [2019-03-26 08:30:58.472642] I [MSGID: 114046] [client-handshake.c:1216:client_setvolume_cbk] 0-asc-client-1: Connected to asc-client-1, attached to remote volume '/bricks/asc/brick'.
>
> [2019-03-26 08:30:58.472659] I [MSGID: 114047] [client-handshake.c:1227:client_setvolume_cbk] 0-asc-client-1: Server and Client lk-version numbers are not same, reopening the fds
>
> [2019-03-26 08:30:58.472714] I [MSGID: 108005] [afr-common.c:4387:afr_notify] 0-asc-replicate-0: Subvolume 'asc-client-1' came back up; going online.
>
> [2019-03-26 08:30:58.472731] I [MSGID: 114035] [client-handshake.c:202:client_set_lk_version_cbk] 0-asc-client-1: Server lk version = 1
>
> [2019-03-26 08:30:58.473112] E [socket.c:2309:socket_connect_finish] 0-asc-client-3: connection to 192.170.254.6:49421 failed (Nessun instradamento per l'host)
>
> [2019-03-26 08:30:58.473152] W [MSGID: 108001] [afr-common.c:4467:afr_notify] 0-asc-replicate-0: Client-quorum is not met
>
> [2019-03-26 08:30:58.477699] I [MSGID: 108031] [afr-common.c:2157:afr_local_discovery_cbk] 0-asc-replicate-0: selecting local read_child asc-client-1
>
> [2019-03-26 08:30:58.477804] I [MSGID: 104041] [glfs-resolve.c:885:__glfs_active_subvol] 0-asc: switched to graph 676c6e6f-6465-3032-2e61-736370642e6c (0)
>
>
>
>
>
> I analyzed the single nodes and I realized that the firewalld service has been stopped on all nodes.
>
> Firewalld re-enabled the heal started automatically, and the “gluster heal volume VOLNAME info” immediately gave correct connections.
>
>
>
> the recovery of the single bricks started immediatly.
>
> When finished that I have correctly detected and start the host-engine.
>
>
>
> I wanted to tell you about the sequence that led me to the block:
>
>
>
> 1) Node03 in maintenance by hosted-engine.
>
> 2) Maintenance performed and restarted.
>
> 3) Repositioned active node03
>
> 4) Heal automatic controlled with Ovirt Manager.
>
> 5) Heal completed correctly.
>
> 6) Node02 put into maintenance.
>
> 7) During the shutdown of the Node02 some VMs have gone to Pause, Ovirt Manager has signaled the block of the Node01 and immediately the host-engine has stopped.
>
> 8) Restarted the Node02, I saw that the gluster had the peer but there was no healing between the nodes.
>
> I had to close everything, and the situation that presented itself was that of previous emails.
>
>
>
> Questions:
>
> - Why did the Node02 in maintenance block Node01?
>
> - Why was restarting the system not restarting the firewalld service? Is it also managed by vdsm?
>
> - What is the correct way to backup virtual machines on an external machine? We use Ovirt4.1
>
> - can backup be used outside of Ovirt? Es qemu-kvm standard ...
>
> Thanks for all.
> Best regards
> Andrea Milan
>
> Il 25.03.2019 11:53 Sahina Bose ha scritto:
>>
>> You will first need to restore connectivity between the gluster peers
>>
>> for heal to work. So restart glusterd on all hosts as Strahil
>>
>> mentioned, and check if "gluster peer status" returns the other nodes
>>
>> as connected. If not, please check the glusterd log to see what's
>>
>> causing the issue. Share the logs if we need to look at it, along with
>>
>> the version info
>>
>>
>>
>>
>>
>> On Sun, Mar 24, 2019 at 1:08 AM Strahil <hunter86_bg(a)yahoo.com> wrote:
>>>
>>> Hi Andrea, The cluster volumes might have sharding enabled and thus files larger than shard size can be recovered only via cluster. You can try to restart gluster on all nodes and force heal: 1. Kill gluster processes: systemctl stop glusterd /usr/share/glusterfs/scripts/stop-all-gluster-processes.sh 2. Start gluster: systemctl start glusterd 3. Force heal: for i in $(gluster volume list); do gluster volume heal $i full ; done sleep 300 for i in $(gluster volume list); do gluster volume heal $i info summary ; done Best Regards, Strahil Nikolov On Mar 23, 2019 13:51, commramius(a)tiscali.it wrote: > > During maintenance of a machine the hosted engine crashed. > At that point there was no more chance of managing anything. > > The VMs have paused, and were no longer manageable. > I restarted the machine, but one point all the bricks were no longer reachable. > > Now I am in a situation where the engine support is no longer loaded. > > The gluster sees the peers connected and the services turned on for the various bricks, but fails to heal the messages that I find for each machine are the following > > # gluster volume heal engine info > Brick 192.170.254.3:/bricks/engine/brick > > . > . > . > > Status: Connected Number of entries: 190 > > Brick 192.170.254.4:/bricks/engine/brick > Status: Il socket di destinazione non è connesso > Number of entries: - > > Brick 192.170.254.6:/bricks/engine/brick > Status: Il socket di destinazione non è connesso > Number of entries: - > > this for all the bricks (some have no heal to do because the machines inside were turned off). > > In practice all the bricks see only localhost as connected. > > How can I restore the machines? > Is there a way to read data from the physical machine and export it so that it can be reused? > Unfortunately we need to access that data. > > Someone can help me. > > Thanks Andrea > _______________________________________________ > Users mailing list -- users(a)ovirt.org > To unsubscribe send an email to users-leave(a)ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ > List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/EOIY7ZU4GOE... _______________________________________________ Users mailing list -- users(a)ovirt.org To unsubscribe send an email to users-leave(a)ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/NMBDYBOY4TZ...
>
>
>
>
> Con OpenStar hai Giga, SMS e i minuti che vuoi da 4,99€ al mese, per sempre. Cambi gratis quando e come vuoi e in più hai 6 mesi di INFINTY! http://tisca.li/myopen
>
5 years, 8 months
Changing Ovirt hosted engine memory
by Dionysis K
Hello! i am new user to ovirt and i was trying to increase the memory of the hosted engine from 5120 MB to 10240 MB and the reports on the events are those bellow, the environment the hosted engine runs is on 3 ovirt nodes with version 4.2.8 and the hosted engine is updates with the latest version, any ideas how to set up things
Failed to hot set memory to VM HostedEngine. Underlying error message: unsupported configuration: Attaching memory device with size '5242880' would exceed domain's maxMemory config
VDSM command SetAmountOfMemoryVDS failed: unsupported configuration: Attaching memory device with size '5242880' would exceed domain's maxMemory config
5 years, 8 months
oVirt 4.3 - create windows2012 vm failed
by jingjie.jiang@oracle.com
Hi,
The oVirt 4.3 is installed on CentOS 7.6.
Windows2012 vm failed with following error messages:
2019-03-18 09:19:12,941-04 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-15) [] EVENT_ID: VM_DOWN_ERROR(119), VM win12_nfs is down with error. Exit message: internal error: qemu unexpectedly closed the monitor: 2019-03-18T13:19:11.837850Z qemu-kvm: warning: All CPU(s) up to maxcpus should be described in NUMA config, ability to start up with partial NUMA mappings is obsoleted and will be removed in future
Hyper-V SynIC is not supported by kernel
2019-03-18T13:19:11.861071Z qemu-kvm: kvm_init_vcpu failed: Function not implemented.
Anyone suggestion?
Thanks,
Jingjie
5 years, 8 months
Re: OVirt Gluster Fail
by Strahil
Hi Andrea,
The cluster volumes might have sharding enabled and thus files larger than shard size can be recovered only via cluster.
You can try to restart gluster on all nodes and force heal:
1. Kill gluster processes:
systemctl stop glusterd
/usr/share/glusterfs/scripts/stop-all-gluster-processes.sh
2. Start gluster:
systemctl start glusterd
3. Force heal:
for i in $(gluster volume list); do gluster volume heal $i full ; done
sleep 300
for i in $(gluster volume list); do gluster volume heal $i info summary ; done
Best Regards,
Strahil NikolovOn Mar 23, 2019 13:51, commramius(a)tiscali.it wrote: > > During maintenance of a machine the hosted engine crashed. > At that point there was no more chance of managing anything. > > The VMs have paused, and were no longer manageable. > I restarted the machine, but one point all the bricks were no longer reachable. > > Now I am in a situation where the engine support is no longer loaded. > > The gluster sees the peers connected and the services turned on for the various bricks, but fails to heal the messages that I find for each machine are the following > > # gluster volume heal engine info > Brick 192.170.254.3:/bricks/engine/brick > > . > . > . > > Status: Connected Number of entries: 190 > > Brick 192.170.254.4:/bricks/engine/brick > Status: Il socket di destinazione non è connesso > Number of entries: - > > Brick 192.170.254.6:/bricks/engine/brick > Status: Il socket di destinazione non è connesso > Number of entries: - > > this for all the bricks (some have no heal to do because the machines inside were turned off). > > In practice all the bricks see only localhost as connected. > > How can I restore the machines? > Is there a way to read data from the physical machine and export it so that it can be reused? > Unfortunately we need to access that data. > > Someone can help me. > > Thanks Andrea > _______________________________________________ > Users mailing list -- users(a)ovirt.org > To unsubscribe send an email to users-leave(a)ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ > List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/EOIY7ZU4GOE...
5 years, 8 months
oVirt 2019 Survey
by Sandro Bonazzola
As we continue to develop oVirt 4.3 and future releases, the Development
and Integration teams at Red Hat would value insights on how you are
deploying the oVirt environment.
Please help us to hit the mark by completing this short survey. Survey will
close on March 31th.
If you're managing multiple oVirt deployments with very different use cases
or very different deployments you can consider answering this survey
multiple times.
Please note the answers to this survey will be publicly accessible.
This survey is under oVirt Privacy Policy available at
https://www.ovirt.org/site/privacy-policy.html
The survey is available at https://goo.gl/forms/QOl5gDJdR83dzVPT2
Thanks,
--
SANDRO BONAZZOLA
MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
Red Hat EMEA <https://www.redhat.com/>
sbonazzo(a)redhat.com
<https://red.ht/sig>
5 years, 8 months
How to create a template from VM Portal
by Wood Peter
Hi,
How do I create a template from a VM in "VM Portal"?
I'm using version 4.2.8.2 and I don't see the option to convert a VM to a
template anywhere. The documentation refers to the Administration Portal
interface.
Is it possible to create a template from the "VM Portal" or one must login
to "Administration Portal"?
Any help would be appreciated.
Thank you,
-- Peter
5 years, 8 months
API POST ovf file instead of libvirt KVM regular xml
by Nathanaël Blanchet
Hi all,
I've got a ovf file where I need to reuse nic defintions. I want to post them to the REST API, but in API documentation, I'm only able to find examples with libvirt definition.
So my question is : does ovf file are supported to be directly posted or is there a way to convert ovf file to libvirtxml so as to post them to the API?
Thank you.
5 years, 8 months
Gluster event logging
by Ron Jerome
I'm running oVirt 4.3.2, and just upgraded gluster to 5.5. I see that the gluster event daemon now works, however the events are not being processed by the ovirt engine. On the engine side I'm seeing:
"engine.log:2019-03-25 17:14:04,707-04 ERROR [org.ovirt.engine.core.services.GlusterEventsWebHookServlet] (default task-198) [] Error processing event data" in the engine log.
On the gluster servers in event.log, I'm seeing:
"[2019-03-25 17:14:04,704] WARNING [utils - 272:publish_to_webhook] - Event push failed to URL: http://ovirt-engine:80/ovirt-engine/services/glusterevents, Event: {"event": "CLIENT_CONNECT", "message": {"brick_path": "/gluster_bricks/ISO/ISO", "client_identifier": "10.10.15.12:49120", "client_uid": "CTX_ID:d988b64d-2996-44c4-88ed-ee768c9e0c21-GRAPH_ID:0-PID:43711-HOST:ovirt2-PC_NAME:ISO-client-2-RECON_NO:-0", "server_identifier": "10.10.15.12:49155", "subdir_mount": "(null)"}, "nodeid": "9a6ea54e-c692-45bd-889b-ed20c6fb1145", "ts": 1553548444}, Status Code: 500"
How can I determine why these events are not getting processed?
Thanks,
5 years, 8 months
Hosted Engine Storage migration
by joel+ovirt@robisons.us
Hello listmembers!
I'm in search of some steps to use to migrate my Hosted-Engine from one NFS storage domain to another different NFS storage domain. One of the possible ways to accomplish this task may already be a feature with oVirt 4.1, I'm using 4.2
https://lists.ovirt.org/pipermail/users/2017-January/078739.html (New feature in 4.1?)
Will someone elaborate the steps to use this new feature and how I could use it to accomplish this task? the bug link in the email linked above has a comment (https://bugzilla.redhat.com/show_bug.cgi?id=1409112#c12) - does this mean I need to redeploy the hosted-engine? I'm not able to follow as I'm new to oVirt and still learning the ins and outs and design if I'm on the wrong path here please correct me.
Thank you,
-Joel
5 years, 8 months