On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell <mrrex4(a)gmail.com> wrote:
The output of ovirt-hosted-engine-cleanup
[root@node-1 ~]# ovirt-hosted-engine-cleanup
This will de-configure the host to run ovirt-hosted-engine-setup from
scratch.
Caution, this operation should be used with care.
Are you sure you want to proceed? [y/n]
y
-=== Destroy hosted-engine VM ===-
You must run deploy first
-=== Stop HA services ===-
-=== Shutdown sanlock ===-
shutdown force 1 wait 0
shutdown done 0
-=== Disconnecting the hosted-engine storage domain ===-
You must run deploy first
-=== De-configure VDSM networks ===-
-=== Stop other services ===-
-=== De-configure external daemons ===-
-=== Removing configuration files ===-
? /etc/init/libvirtd.conf already missing
- removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml
? /etc/ovirt-hosted-engine/answers.conf already missing
? /etc/ovirt-hosted-engine/hosted-engine.conf already missing
- removing /etc/vdsm/vdsm.conf
- removing /etc/pki/vdsm/certs/cacert.pem
- removing /etc/pki/vdsm/certs/vdsmcert.pem
- removing /etc/pki/vdsm/keys/vdsmkey.pem
- removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem
- removing /etc/pki/vdsm/libvirt-spice/ca-key.pem
- removing /etc/pki/vdsm/libvirt-spice/server-cert.pem
- removing /etc/pki/vdsm/libvirt-spice/server-key.pem
? /etc/pki/CA/cacert.pem already missing
? /etc/pki/libvirt/*.pem already missing
? /etc/pki/libvirt/private/*.pem already missing
? /etc/pki/ovirt-vmconsole/*.pem already missing
- removing /var/cache/libvirt/qemu
? /var/run/ovirt-hosted-engine-ha/* already missing
[root@node-1 ~]#
Output of sanlock client status:
[root@node-1 ~]# sanlock client status
[root@node-1 ~]#
Thank you for your help!
On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi <stirabos(a)redhat.com>
wrote:
>
>
> On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell <mrrex4(a)gmail.com>
> wrote:
>
>> After letting this sit for a few days, does anyone have any ideas as to
>> how to deal with my situation? Would anyone like me to send the SOS report
>> directly to them? It's a 9MB file.
>>
>> If nothing comes up, I'm going to try and sift through the SOS report
>> tonight, but I won't know what I'm trying to find.
>>
>> Thank you for any and all help.
>>
>> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola <sbonazzo(a)redhat.com>
>> wrote:
>>
>>>
>>>
>>> On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell <mrrex4(a)gmail.com>
>>> wrote:
>>>
>>>> Ran the 4 commands listed above, no errors on the screen.
>>>>
>>>> Started the hosted-engine standard setup from the web-UI.
>>>>
>>>> Using iSCSI for the storage.
>>>>
>>>> Using mostly default options, I got these errors in the web-UI.
>>>>
>>>> Error creating Volume Group: Failed to initialize physical device:
>>>>
("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",)
>>>> Failed to execute stage 'Misc configuration': Failed to
initialize
>>>> physical device: ("[u'/dev/mapper/36589cfc00000
>>>> 0de7482638fcfcebbbb4']",)
>>>> Hosted Engine deployment failed: this system is not reliable, please
>>>> check the issue,fix and redeploy
>>>>
>>>> I rebuilt my iSCSI (I don't think I cleaned it up from a previous
>>>> install).
>>>> Re-ran the above 4 commands.
>>>> Restarted hosted engine standard setup from web-UI.
>>>> Install moved past "Connecting Storage Pool" so I believe the
above
>>>> was my fault.
>>>>
>>>> These are the last messages displayed on the web-UI.
>>>> Creating Storage Pool
>>>> Connecting Storage Pool
>>>> Verifying sanlock lockspace initialization
>>>> Creating Image for 'hosted-engine.lockspace' ...
>>>> Image for 'hosted-engine.lockspace' created successfully
>>>> Creating Image for 'hosted-engine.metadata' ...
>>>> Image for 'hosted-engine.metadata' created successfully
>>>> Creating VM Image
>>>> Extracting disk image from OVF archive (could take a few minutes
>>>> depending on archive size)
>>>> Validating pre-allocated volume size
>>>> Uploading volume to data domain (could take a few minutes depending on
>>>> archive size)
>>>>
>>>> At the host terminal, I got the error "watchdog watchdog0: watchdog
>>>> did not stop!"
>>>> Then the host restarted.
>>>>
>>>
>>> Simone, can you help here?
>>>
>>>
> Ok, sorry for the delay.
> The second installation attempt seams fine but it seams that
> ovirt-hosted-engine-cleanup failed stopping sanlock and so the
> watchdog kick in rebooting your system in the middle of deployment attempt.
>
> could you please post the output of
> ovirt-hosted-engine-cleanup
> sanlock client status
> ?
>
>
>
>>
>>>
>>>
>>>>
>>>> This is as far as I've gotten in previous attempts.
>>>>
>>>> Attaching the hosted-engine-setup log.
>>>>
>>>> The SOS report is 9MB and the ovirt users group will drop the email.
>>>>
>>>> On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola
<sbonazzo(a)redhat.com
>>>> > wrote:
>>>>
>>>>>
>>>>>
>>>>> On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell
<mrrex4(a)gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Now that you have identified the problem, should I run the
following
>>>>>> commands and send you another SOS?
>>>>>>
>>>>>> ovirt-hosted-engine-cleanup
>>>>>> vdsm-tool configure --force
>>>>>> systemctl restart libvirtd
>>>>>> systemctl restart vdsm
>>>>>>
>>>>>> Or is there a different plan in mind?
>>>>>>
>>>>>
>>>>> I would have expected someone from virt team to follow up for
further
>>>>> investigations :-)
>>>>> above commands should work.
>>>>>
>>>>>
>>>>>
>>>>>>
>>>>>> Thank you,
>>>>>>
>>>>>> Brendan
>>>>>>
>>>>>> On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola <
>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell
<mrrex4(a)gmail.com
>>>>>>> > wrote:
>>>>>>>
>>>>>>>> Can you please elaborate about the failure you see here
and how
>>>>>>>> are you trying to manually partition the host?
>>>>>>>>
>>>>>>>> Sure, I will start from the beginning.
>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604
>>>>>>>> <(201)%20705-2604>.iso
>>>>>>>> - During installation I setup one of the two interfaces
and check
>>>>>>>> the box to automatically use the connection.
>>>>>>>> - I'm currently providing a host name of
node-1.test.net until I
>>>>>>>> have a successful process.
>>>>>>>> - I configure date and time for my timezone and to use an
internal
>>>>>>>> NTP server.
>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0
SanDisk
>>>>>>>> flash drive, check the box that I would like to make
additional space, and
>>>>>>>> click done. In the reclaim disk space window, I click
delete all, and then
>>>>>>>> reclaim space. I go back into the Installation
Destination, select that I
>>>>>>>> will configure partitioning, and click done. The Manual
Partitioning
>>>>>>>> window opens, I use the option to automatically create
mount points.
>>>>>>>>
>>>>>>>
>>>>>>> In this screen, please change partitioning scheme from LVM to
LVM
>>>>>>> Thin Provisioning: it should solve your following error.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>> At this point, /boot is 1024MB, /var is 15GB, / is
88.11 GB, and
>>>>>>>> swap is 11.57GB. I then change / to 23.11 GB, update
settings, change /var
>>>>>>>> to 80GB, update settings again, and click done. I accept
the changes and
>>>>>>>> begin installation.
>>>>>>>>
>>>>>>>> I tried these changes based on this article:
http://www.ovirt.org/
>>>>>>>>
documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/
>>>>>>>>
>>>>>>>> The article does say that you can specify a different
directory
>>>>>>>> than /var/tmp, but I don't recall seeing that
option.
>>>>>>>>
>>>>>>>
>>>>>>> If the setup detects not enough space in /var/tmp for
extracting
>>>>>>> the appliance it will ask about a different directory.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> After some time, I get the following error:
>>>>>>>> There was an error running the kickstart script at line
7. This
>>>>>>>> is a fatal error and installation will be aborted. The
details of this
>>>>>>>> error are:
>>>>>>>>
>>>>>>>> [INFO] Trying to create a manageable base from
'/'
>>>>>>>> [ERROR] LVM Thin Provisioning partitioning scheme is
required.
>>>>>>>> For autoinstall via Kickstart with LVM Thin Provisioning
check options
>>>>>>>> --thinpool and --grow. Please consult documentation for
details.
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> ^^ this one should be solved by the LVM Thin Provisioning
scheme
>>>>>>> mentioned above..
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> Traceback (most recent call last):
>>>>>>>> File "/usr/lib64/python2.7/runpy.py", line 162,
in
>>>>>>>> _run_module_as_main "__main__", fname, loader,
pkg_name)
>>>>>>>> File "/usr/lib64/python2.7runpy.py", line 72,
in _run_code exec
>>>>>>>> code in run_globals
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/__main__.py",
>>>>>>>> line 51, in <module> CliApplication()
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/__init__.py",
>>>>>>>> line 82, in CliApplication()
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/hooks.py", line
>>>>>>>> 120, in emit cb(self.context, *args)
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/plugins/core.py",
>>>>>>>> line 169, in post_argparse layout.initialize(args.source,
args.init_nvr)
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/plugins/core.py",
>>>>>>>> line 216, in initialize
self.app.imgbase.init_layout_from(source,
>>>>>>>> init_nvr)
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line
>>>>>>>> 271, in init_layout_from self.init_tags_on(existing_lv)
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line
>>>>>>>> 243, in init_tags_on pool = lv.thinpool()
>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/lvm.py", line
>>>>>>>> 250, in thinpool raise MissingLvmThinPool()
imgbased.lvm.MissingLvmThinPoo
>>>>>>>> l
>>>>>>>>
>>>>>>>> At this point, the only option is to exit the installer.
>>>>>>>>
>>>>>>>> ****************************
>>>>>>>>
>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is
not
>>>>>>>> supported anymore.
>>>>>>>>
>>>>>>>> Not a problem.
>>>>>>>>
>>>>>>>> ****************************
>>>>>>>>
>>>>>>>> Can you please provide hosted engine setup logs or better
a full
>>>>>>>> sos report? (sosreport -a)
>>>>>>>>
>>>>>>>> Again, the process I'm following:
>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604
>>>>>>>> <(201)%20705-2604>.iso
>>>>>>>> - During installation I setup one of the two interfaces
and check
>>>>>>>> the box to automatically use the connection.
>>>>>>>> - I'm currently providing a host name of
node-1.test.net until I
>>>>>>>> have a successful process.
>>>>>>>> - I configure date and time for my timezone and to use an
internal
>>>>>>>> NTP server.
>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0
SanDisk
>>>>>>>> flash drive, check the box that I would like to make
additional space, and
>>>>>>>> click done. In the reclaim disk space window, I click
delete all, and then
>>>>>>>> reclaim space.
>>>>>>>> - Begin Installation and set a root password.
>>>>>>>> - Perform a yum update - no packages marked for update
(as
>>>>>>>> expected)
>>>>>>>> - Use vi to update /etc/hosts with a reference for
node-1.test.net
>>>>>>>> and
engine.test.net
>>>>>>>> - First attempt at hosted-engine from web-UI
>>>>>>>> - Setup downloads and installs ovirt-engine-applianc
>>>>>>>> e-4.1-20170523.1.el7.centos.noarch.rpm
>>>>>>>> *Failed to execute stage 'Environment setup':
Failed to
>>>>>>>> reconfigure libvirt for VDSM
>>>>>>>> *Hosted Engine deployment failed
>>>>>>>> - Attached SOS report
>>>>>>>> The checksum is: aa56097edc0b63c49caaf1a1fde021bc
>>>>>>>>
>>>>>>>> At this point, I would run ovirt-hosted-engine-cleanup
and I
>>>>>>>> would get further along in the install process. However,
because this is a
>>>>>>>> fresh install, I'm going to leave things here for now
so you can review the
>>>>>>>> SOS.
>>>>>>>>
>>>>>>>
>>>>>>> Thanks for the SOS report!
>>>>>>> Hosted Engine setup fails on:
>>>>>>>
>>>>>>> 2017-05-30 19:24:39 DEBUG
otopi.plugins.gr_he_setup.system.vdsmenv
>>>>>>> plugin.execute:921 execute-output: ('/bin/vdsm-tool',
'configure',
>>>>>>> '--force') stdout:
>>>>>>>
>>>>>>> Checking configuration status...
>>>>>>>
>>>>>>> Current revision of multipath.conf detected, preserving
>>>>>>> lvm is configured for vdsm
>>>>>>> libvirt is already configured for vdsm
>>>>>>> SUCCESS: ssl configured to true. No conflicts
>>>>>>>
>>>>>>> Running configure...
>>>>>>> Reconfiguration of libvirt is done.
>>>>>>>
>>>>>>> 2017-05-30 19:24:39 DEBUG
otopi.plugins.gr_he_setup.system.vdsmenv
>>>>>>> plugin.execute:926 execute-output: ('/bin/vdsm-tool',
'configure',
>>>>>>> '--force') stderr:
>>>>>>> Error: ServiceOperationError: _systemctlStart failed
>>>>>>> Job for libvirtd.service failed because the control process
exited
>>>>>>> with error code. See "systemctl status
libvirtd.service" and "journalctl
>>>>>>> -xe" for details.
>>>>>>>
>>>>>>> At the same time journalctl shows:
>>>>>>>
>>>>>>> May 30 19:24:39
node-1.test.net libvirtd[20954]: libvirt
version:
>>>>>>> 2.0.0, package: 10.el7_3.5 (CentOS BuildSystem <
>>>>>>>
http://bugs.centos.org>, 2017-03-03-02:09:45,
c1bm.rdu2.centos.org)
>>>>>>> May 30 19:24:39
node-1.test.net libvirtd[20954]: hostname:
>>>>>>>
node-1.test.net
>>>>>>> May 30 19:24:39
node-1.test.net libvirtd[20954]: The server
>>>>>>> certificate /etc/pki/vdsm/certs/vdsmcert.pem is not yet
active
>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]: libvirtd.service:
main
>>>>>>> process exited, code=exited, status=6/NOTCONFIGURED
>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]: Failed to start
>>>>>>> Virtualization daemon.
>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]: Unit
libvirtd.service
>>>>>>> entered failed state.
>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]: libvirtd.service
>>>>>>> failed.
>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]: libvirtd.service
>>>>>>> holdoff time over, scheduling restart.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> ****************************
>>>>>>>>
>>>>>>>> I'd like to understand the issues you faced before
suggesting to
>>>>>>>> restart from scratch.
>>>>>>>>
>>>>>>>> Too late... I did two re-installs to get a more accurate
account
>>>>>>>> of my install process for above.
>>>>>>>>
>>>>>>>> ****************************
>>>>>>>>
>>>>>>>> Thank you for your help!
>>>>>>>>
>>>>>>>> Brendan
>>>>>>>>
>>>>>>>> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola <
>>>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell
<
>>>>>>>>> mrrex4(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> oVirt users list,
>>>>>>>>>>
>>>>>>>>>> Long story short, I've been spending weeks on
this project for
>>>>>>>>>> my home lab with no success.
>>>>>>>>>>
>>>>>>>>>> I would like to successfully install two nodes
that host a
>>>>>>>>>> highly available engine with an iSCSI storage
back-end.
>>>>>>>>>>
>>>>>>>>>> I have read through most, if not all, of the
guides on
ovirt.org
>>>>>>>>>> with no substantial help.
>>>>>>>>>>
>>>>>>>>>> Successfully, I have done the following:
>>>>>>>>>> Install oVirt Engine on a bare metal system,
added a node, and
>>>>>>>>>> started exploring - not desired.
>>>>>>>>>> Install oVirt Node 4.0.6 on a bare metal system -
fails if
>>>>>>>>>> partitions are not done automatically.
>>>>>>>>>> Install oVirt Node 4.1.2 on a bare metal system -
fails if
>>>>>>>>>> partitions are not done automatically.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Can you please elaborate about the failure you see
here and how
>>>>>>>>> are you trying to manually partition the host?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> My process after installing a Node:
>>>>>>>>>> Run a yum update - just to be sure, but I am
using latest iso
>>>>>>>>>> images from downloads section.
>>>>>>>>>> Edit /etc/hosts for local name resolution - the
goal is to host
>>>>>>>>>> DNS as a virtual machine, eventually.
>>>>>>>>>> On 4.1 if I install ovirt-engine-appliance from
yum, it does
>>>>>>>>>> simplify one step in the hosted engine setup. If
I do this on 4.0 it
>>>>>>>>>> discards the image and uses the default.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0
is not
>>>>>>>>> supported anymore.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> On 4.1 the hosted engine setup fails immediately
unless I run
>>>>>>>>>> the hosted engine cleanup from the shell.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Can you please provide hosted engine setup logs or
better a full
>>>>>>>>> sos report? (sosreport -a)
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> If I do this, I can typically get to the point
of installing.
>>>>>>>>>>
>>>>>>>>>> When I do get to the installation phase, I get to
a point just
>>>>>>>>>> after extracting the OVA that I get a message on
the shell saying something
>>>>>>>>>> about the watchdog running the whole time and
then the node reboots.
>>>>>>>>>>
>>>>>>>>>> I found one email thread that sounded like my
issue and
>>>>>>>>>> suggested the following commands:
>>>>>>>>>> vdsm-tool configure --force
>>>>>>>>>> systemctl restart libvirtd
>>>>>>>>>> systemctl restart vdsmd
>>>>>>>>>>
>>>>>>>>>> Unfortunately, these commands did not help my
situation like the
>>>>>>>>>> other individual.
>>>>>>>>>>
>>>>>>>>>> What log file would everyone like to see first?
Given that I
>>>>>>>>>> still consider myself relatively new to Linux,
please identify the path for
>>>>>>>>>> the log file requested.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> See above
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Also, because I plan on performing a clean
install for this
>>>>>>>>>> thread using my process from above (I'm not
expecting my outcome to be any
>>>>>>>>>> different), are there any tips and tricks that
might result in a success?
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> I'd like to understand the issues you faced
before suggesting to
>>>>>>>>> restart from scratch.
>>>>>>>>> Adding some people who may help as well.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Thank you for any and all help,
>>>>>>>>>> Brendan
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Users mailing list
>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> SANDRO BONAZZOLA
>>>>>>>>>
>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG
VIRTUALIZATION
>>>>>>>>> R&D
>>>>>>>>>
>>>>>>>>> Red Hat EMEA <
https://www.redhat.com/>
>>>>>>>>> <
https://red.ht/sig>
>>>>>>>>> TRIED. TESTED. TRUSTED.
<
https://redhat.com/trusted>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> SANDRO BONAZZOLA
>>>>>>>
>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG
VIRTUALIZATION R&D
>>>>>>>
>>>>>>> Red Hat EMEA <
https://www.redhat.com/>
>>>>>>> <
https://red.ht/sig>
>>>>>>> TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> SANDRO BONAZZOLA
>>>>>
>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION
R&D
>>>>>
>>>>> Red Hat EMEA <
https://www.redhat.com/>
>>>>> <
https://red.ht/sig>
>>>>> TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>>
>>> SANDRO BONAZZOLA
>>>
>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D
>>>
>>> Red Hat EMEA <
https://www.redhat.com/>
>>> <
https://red.ht/sig>
>>> TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>
>>>
>>
>>
>