deprecating export domain?
by Charles Kozler
Hello,
I recently read on this list from a redhat member that export domain is
either being deprecated or looking at being deprecated
To that end, can you share details? Can you share any notes/postings/bz's
that document this? I would imagine something like this would be discussed
in larger audience
This seems like a somewhat significant change to make and I am curious
where this is scheduled? Currently, a lot of my backups rely explicitly on
an export domain for online snapshots, so I'd like to plan accordingly
Thanks!
2 years, 7 months
Support for Shared SAS storage
by Vinícius Ferrão
Hello,
I’ve two compute nodes with SAS Direct Attached sharing the same disks.
Looking at the supported types I can’t see this on the documentation: https://www.ovirt.org/documentation/admin-guide/chap-Storage.html
There’s is local storage on this documentation, but my case is two machines, both using SAS, connected to the same machines. It’s the VRTX hardware from Dell.
Is there any support for this? It should be just like Fibre Channel and iSCSI, but with SAS instead.
Thanks,
2 years, 7 months
Shutdown procedure for single host HCI Gluster
by Gianluca Cecchi
Hello,
I'm testing the single node HCI with ovirt-node-ng 4.3.9 iso.
Very nice and many improvements over the last time I tried it. Good!
I have a doubt related to shutdown procedure of the server.
Here below my steps:
- Shutdown all VMs (except engine)
- Put into maintenance data and vmstore domains
- Enable Global HA Maintenance
- Shutdown engine
- Shutdown hypervisor
It seems that the last step doesn't end and I had to brutally power off the
hypervisor.
Here the screenshot regarding infinite failure in unmounting
/gluster_bricks/engine
https://drive.google.com/file/d/1ee0HG21XmYVA0t7LYo5hcFx1iLxZdZ-E/view?us...
What would be the right step to do before the final shutdown of hypervisor?
Thanks,
Gianluca
2 years, 8 months
Upgrade ovirt from 3.4 to 4.3
by lu.alfonsi@almaviva.it
Good morning,
i have a difficult enviroment with 20 Hypervisors based on ovirt 3.4.3-1 and i would like to reach the 4.3 version. Which are the best steps to achieve these objective?
Thanks in advance
Luigi
2 years, 9 months
Re: Single instance scaleup.
by Strahil
Hi Leo,
As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes.
You can use the following for adding the bricks:
https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Ad...
Best Regards,
Strahil NikolivOn May 26, 2019 10:54, Leo David <leoalex(a)gmail.com> wrote:
>
> Hi Stahil,
> Thank you so much for yout input !
>
> gluster volume info
>
>
> Volume Name: engine
> Type: Distribute
> Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.80.191:/gluster_bricks/engine/engine
> Options Reconfigured:
> nfs.disable: on
> transport.address-family: inet
> storage.owner-uid: 36
> storage.owner-gid: 36
> features.shard: on
> performance.low-prio-threads: 32
> performance.strict-o-direct: off
> network.remote-dio: off
> network.ping-timeout: 30
> user.cifs: off
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> cluster.eager-lock: enable
> Volume Name: ssd-samsung
> Type: Distribute
> Volume ID: 76576cc6-220b-4651-952d-99846178a19e
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.80.191:/gluster_bricks/sdc/data
> Options Reconfigured:
> cluster.eager-lock: enable
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> user.cifs: off
> network.ping-timeout: 30
> network.remote-dio: off
> performance.strict-o-direct: on
> performance.low-prio-threads: 32
> features.shard: on
> storage.owner-gid: 36
> storage.owner-uid: 36
> transport.address-family: inet
> nfs.disable: on
>
> The other two hosts will be 192.168.80.192/193 - this is gluster dedicated network over 10GB sfp+ switch.
> - host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array )
> - host 3 has:
> - 1 ssd for OS
> - 1 ssd - for adding to engine volume in a full replica 3
> - 2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung )
> So the plan is to have "engine" scaled in a full replica 3, and "ssd-samsung" scalled in a replica 3 arbitrated.
>
>
>
>
> On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg(a)yahoo.com> wrote:
>>
>> Hi Leo,
>>
>> Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'.
>> If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary.
>>
>> What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s interfase(s).
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> On May 26, 2019 07:52, Leo David <leoalex(a)gmail.com> wrote:
>>>
>>> Hello Everyone,
>>> Can someone help me to clarify this ?
>>> I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle.
>>> Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ?
>>> If so, could you share with me what would it be the procedure ?
>>> Thank you very much !
>>>
>>> Leo
>
>
>
> --
> Best regards, Leo David
2 years, 9 months
Failed to add storage domain
by thunderlight1@gmail.com
Hi!
I have installed oVirt using the iso ovirt-node-ng-installer-4.3.2-2019031908.el7. I the did run the Host-engine deployment through Cockpit.
I got an error when it tries to create the domain storage. It sucessfully mounted the NFS-share on the host. Bellow is the error I got:
2019-04-14 10:40:38,967+0200 INFO ansible skipped {'status': 'SKIPPED', 'ansible_task': u'Check storage domain free space', 'ansible_host': u'localhost', 'ansible_playbook': u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_type': 'task'}
2019-04-14 10:40:38,967+0200 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7fb6918ad9d0> kwargs
2019-04-14 10:40:39,516+0200 INFO ansible task start {'status': 'OK', 'ansible_task': u'ovirt.hosted_engine_setup : Activate storage domain', 'ansible_playbook': u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_type': 'task'}
2019-04-14 10:40:39,516+0200 DEBUG ansible on_any args TASK: ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False
2019-04-14 10:40:41,923+0200 DEBUG var changed: host "localhost" var "otopi_storage_domain_details" type "<type 'dict'>" value: "{
"changed": false,
"exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_xSFxOp/__main__.py\", line 664, in main\n storage_domains_module.post_create_check(sd_id)\n File \"/tmp/ansible_ovirt_storage_domain_payload_xSFxOp/__main__.py\", line 526, in post_create_check\n id=storage_domain.id,\n File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(response)\n File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response
, body)\n File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\nError: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n",
"failed": true,
"msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."
}"
2019-04-14 10:40:41,924+0200 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<type 'list'>" value: "[]"
2019-04-14 10:40:41,924+0200 DEBUG var changed: host "localhost" var "play_hosts" type "<type 'list'>" value: "[]"
2019-04-14 10:40:41,924+0200 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<type 'list'>" value: "[]"
2019-04-14 10:40:41,924+0200 ERROR ansible failed {'status': 'FAILED', 'ansible_type': 'task', 'ansible_task': u'Activate storage domain', 'ansible_result': u'type: <type \'dict\'>\nstr: {\'_ansible_parsed\': True, u\'exception\': u\'Traceback (most recent call last):\\n File "/tmp/ansible_ovirt_storage_domain_payload_xSFxOp/__main__.py", line 664, in main\\n storage_domains_module.post_create_check(sd_id)\\n File "/tmp/ansible_ovirt_storage_domain_payload_xSFxOp/__main__.py", line 526', 'task_duration': 2, 'ansible_host': u'localhost', 'ansible_playbook': u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml'}
2019-04-14 10:40:41,924+0200 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7fb691843190> kwargs ignore_errors:None
2019-04-14 10:40:41,928+0200 INFO ansible stats {
"ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
"ansible_playbook_duration": "00:37 Minutes",
"ansible_result": "type: <type 'dict'>\nstr: {u'localhost': {'unreachable': 0, 'skipped': 6, 'ok': 23, 'changed': 1, 'failures': 1}}",
"ansible_type": "finish",
"status": "FAILED"
}
2019-04-14 10:40:41,928+0200 INFO SUMMARY:
Duration Task Name
-------- --------
[ < 1 sec ] Execute just a specific set of steps
[ 00:01 ] Force facts gathering
[ 00:01 ] Check local VM dir stat
[ 00:01 ] Obtain SSO token using username/password credentials
[ 00:01 ] Fetch host facts
[ < 1 sec ] Fetch cluster ID
[ 00:01 ] Fetch cluster facts
[ 00:01 ] Fetch Datacenter facts
[ < 1 sec ] Fetch Datacenter ID
[ < 1 sec ] Fetch Datacenter name
[ 00:02 ] Add NFS storage domain
[ 00:01 ] Get storage domain details
[ 00:01 ] Find the appliance OVF
[ 00:01 ] Parse OVF
[ < 1 sec ] Get required size
[ FAILED ] Activate storage domain
2019-04-14 10:40:41,928+0200 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7fb69404eb90> kwargs
Any suggestions on how fix this?
2 years, 10 months
How to connect to a guest with vGPU ?
by Josep Manel Andrés Moscardó
Hi,
I got vGPU through mdev working but I am wondering how I would connect
to the client and make use of the GPU. So far I try to access the
console through SPICE and at some point in the boot process it switches
to GPU and I cannot see anything else.
Thanks.
--
Josep Manel Andrés Moscardó
Systems Engineer, IT Operations
EMBL Heidelberg
T +49 6221 387-8394
2 years, 10 months
Issues encountered performing HE install on 4.3
by Alan G
Hi,
I hit a few issues while performing a recent HE install of 4.3. While I managed to find solutions/workarounds to all the problems I thought I might share them here
* As defined in the Ansible defaults the temp dir for building the local HE VM is /var/tmp. I was 80M short of the required space and there did not appear to be a (supported) way to specify a different location. I ended up having to do a bind mount of /var/tmp to get me through the install. Would be nice to be able to specify a custom location.
* Permissive umask required. Our CIS CentOS 7 build requires that default umask is 027. This breaks the installer as it creates the VM image under /var/tmp as root and cannot then access it as qemu user. As the temp files are cleaned up on failure it took me a while to track this one down. My solution was to temporarily set the umask to 022 for the session while running the installer. It would be nice if the installer either handled this by doing a chmod/chown as required, or at least doing a umask pre-check and failing with a meaningful error.
* SSH root login required on host. Again for CIS we have "PermitRoologin no" configured in sshd. This means the add host task fails on the Engine, but instead of a hard failure we get a timeout on the installer. Which left me chasing some imagined routing/bridging/DNS issue. Eventually I realised I could get to the engine logs and found the issue but took several hours. Would be nice if the installer could either support a sudo option or at least perform a root login pre-check and fail with a meaningful error.
Thanks,
Alan
2 years, 10 months
Ubuntu 18.04 and 16.04 cloud images hang at boot up
by suaro@live.com
I'm using oVirt 4.3 (latest ) and able to successfully provision Centos VMs without any problems.
When I attempt to provision Ubuntu VMs, they hang at startup.
The console shows :
...
...
[ 4.010016] Btrfs loaded
[ 101.268594] random: nonblocking pool is initialized
It stays like this indefinitely.
Again, I have no problems with Centos images, but need Ubuntu
Any tips greatly appreciated.
2 years, 10 months
Re: vm console problem
by David David
tested on four different workstations with: fedora20, fedora31 and
windows10(remote-manager last vers)
вс, 29 мар. 2020 г. в 12:39, Strahil Nikolov <hunter86_bg(a)yahoo.com>:
> On March 29, 2020 9:47:02 AM GMT+03:00, David David <dd432690(a)gmail.com>
> wrote:
> >I did as you said:
> >copied from engine /etc/ovirt-engine/ca.pem onto my desktop into
> >/etc/pki/ca-trust/source/anchors and then run update-ca-trust
> >it didn’t help, still the same errors
> >
> >
> >пт, 27 мар. 2020 г. в 21:56, Strahil Nikolov <hunter86_bg(a)yahoo.com>:
> >
> >> On March 27, 2020 12:23:10 PM GMT+02:00, David David
> ><dd432690(a)gmail.com>
> >> wrote:
> >> >here is debug from opening console.vv by remote-viewer
> >> >
> >> >2020-03-27 14:09 GMT+04:00, Milan Zamazal <mzamazal(a)redhat.com>:
> >> >> David David <dd432690(a)gmail.com> writes:
> >> >>
> >> >>> yes i have
> >> >>> console.vv attached
> >> >>
> >> >> It looks the same as mine.
> >> >>
> >> >> There is a difference in our logs, you have
> >> >>
> >> >> Possible auth 19
> >> >>
> >> >> while I have
> >> >>
> >> >> Possible auth 2
> >> >>
> >> >> So I still suspect a wrong authentication method is used, but I
> >don't
> >> >> have any idea why.
> >> >>
> >> >> Regards,
> >> >> Milan
> >> >>
> >> >>> 2020-03-26 21:38 GMT+04:00, Milan Zamazal <mzamazal(a)redhat.com>:
> >> >>>> David David <dd432690(a)gmail.com> writes:
> >> >>>>
> >> >>>>> copied from qemu server all certs except "cacrl" to my
> >> >desktop-station
> >> >>>>> into /etc/pki/
> >> >>>>
> >> >>>> This is not needed, the CA certificate is included in console.vv
> >> >and no
> >> >>>> other certificate should be needed.
> >> >>>>
> >> >>>>> but remote-viewer is still didn't work
> >> >>>>
> >> >>>> The log looks like remote-viewer is attempting certificate
> >> >>>> authentication rather than password authentication. Do you have
> >> >>>> password in console.vv? It should look like:
> >> >>>>
> >> >>>> [virt-viewer]
> >> >>>> type=vnc
> >> >>>> host=192.168.122.2
> >> >>>> port=5900
> >> >>>> password=fxLazJu6BUmL
> >> >>>> # Password is valid for 120 seconds.
> >> >>>> ...
> >> >>>>
> >> >>>> Regards,
> >> >>>> Milan
> >> >>>>
> >> >>>>> 2020-03-26 2:22 GMT+04:00, Nir Soffer <nsoffer(a)redhat.com>:
> >> >>>>>> On Wed, Mar 25, 2020 at 12:45 PM David David
> ><dd432690(a)gmail.com>
> >> >>>>>> wrote:
> >> >>>>>>>
> >> >>>>>>> ovirt 4.3.8.2-1.el7
> >> >>>>>>> gtk-vnc2-1.0.0-1.fc31.x86_64
> >> >>>>>>> remote-viewer version 8.0-3.fc31
> >> >>>>>>>
> >> >>>>>>> can't open vm console by remote-viewer
> >> >>>>>>> vm has vnc console protocol
> >> >>>>>>> when click on console button to connect to a vm, the
> >> >remote-viewer
> >> >>>>>>> console disappear immediately
> >> >>>>>>>
> >> >>>>>>> remote-viewer debug in attachment
> >> >>>>>>
> >> >>>>>> You an issue with the certificates:
> >> >>>>>>
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.238:
> >> >>>>>> ../src/vncconnection.c Set credential 2 libvirt
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Searching for certs in /etc/pki
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Searching for certs in /root/.pki
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Failed to find certificate
> >CA/cacert.pem
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c No CA certificate provided, using
> >GNUTLS
> >> >global
> >> >>>>>> trust
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Failed to find certificate CA/cacrl.pem
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Failed to find certificate
> >> >>>>>> libvirt/private/clientkey.pem
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Failed to find certificate
> >> >>>>>> libvirt/clientcert.pem
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Waiting for missing credentials
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c Got all credentials
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.239:
> >> >>>>>> ../src/vncconnection.c No CA certificate provided; trying the
> >> >system
> >> >>>>>> trust store instead
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.240:
> >> >>>>>> ../src/vncconnection.c Using the system trust store and CRL
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.240:
> >> >>>>>> ../src/vncconnection.c No client cert or key provided
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.240:
> >> >>>>>> ../src/vncconnection.c No CA revocation list provided
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.241:
> >> >>>>>> ../src/vncconnection.c Handshake was blocking
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.243:
> >> >>>>>> ../src/vncconnection.c Handshake was blocking
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.251:
> >> >>>>>> ../src/vncconnection.c Handshake was blocking
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.298:
> >> >>>>>> ../src/vncconnection.c Handshake done
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.298:
> >> >>>>>> ../src/vncconnection.c Validating
> >> >>>>>> (remote-viewer:2721): gtk-vnc-DEBUG: 11:56:25.301:
> >> >>>>>> ../src/vncconnection.c Error: The certificate is not trusted
> >> >>>>>>
> >> >>>>>> Adding people that may know more about this.
> >> >>>>>>
> >> >>>>>> Nir
> >> >>>>>>
> >> >>>>>>
> >> >>>>
> >> >>>>
> >> >>
> >> >>
> >>
> >> Hello,
> >>
> >> You can try to take the engine's CA (maybe it's useless) and put it
> >on
> >> your system in:
> >> /etc/pki/ca-trust/source/anchors (if it's EL7 or a Fedora) and then
> >run
> >> update-ca-trust
> >>
> >> Best Regards,
> >> Strahil Nikolov
> >>
>
> Hey David,
>
> What is you workstation's OS ?
> Also, have you tried from another workstation ?
>
> Best Regards,
> Strahil Nikolov
>
2 years, 11 months