Now that you have identified the problem, should I run the following
commands and send you another SOS?
ovirt-hosted-engine-cleanup
vdsm-tool configure --force
systemctl restart libvirtd
systemctl restart vdsm
Or is there a different plan in mind?
Thank you,
Brendan
On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola <sbonazzo(a)redhat.com>
wrote:
On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell <mrrex4(a)gmail.com>
wrote:
> Can you please elaborate about the failure you see here and how are you
> trying to manually partition the host?
>
> Sure, I will start from the beginning.
> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 <(201)%20705-2604>
> .iso
> - During installation I setup one of the two interfaces and check the box
> to automatically use the connection.
> - I'm currently providing a host name of
node-1.test.net until I have a
> successful process.
> - I configure date and time for my timezone and to use an internal NTP
> server.
> - On Installation Destination, I pick my 128GB USB3.0 SanDisk flash
> drive, check the box that I would like to make additional space, and click
> done. In the reclaim disk space window, I click delete all, and then
> reclaim space. I go back into the Installation Destination, select that I
> will configure partitioning, and click done. The Manual Partitioning
> window opens, I use the option to automatically create mount points.
>
In this screen, please change partitioning scheme from LVM to LVM Thin
Provisioning: it should solve your following error.
> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, and swap
> is 11.57GB. I then change / to 23.11 GB, update settings, change /var to
> 80GB, update settings again, and click done. I accept the changes and
> begin installation.
>
> I tried these changes based on this article:
http://www.ovirt.org/
> documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/
>
> The article does say that you can specify a different directory than
> /var/tmp, but I don't recall seeing that option.
>
If the setup detects not enough space in /var/tmp for extracting the
appliance it will ask about a different directory.
>
> After some time, I get the following error:
> There was an error running the kickstart script at line 7. This is a
> fatal error and installation will be aborted. The details of this error
> are:
>
> [INFO] Trying to create a manageable base from '/'
> [ERROR] LVM Thin Provisioning partitioning scheme is required. For
> autoinstall via Kickstart with LVM Thin Provisioning check options
> --thinpool and --grow. Please consult documentation for details.
>
^^ this one should be solved by the LVM Thin Provisioning scheme mentioned
above..
>
> Traceback (most recent call last):
> File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main
> "__main__", fname, loader, pkg_name)
> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code exec code in
> run_globals
> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", line 51,
> in <module> CliApplication()
> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", line 82,
> in CliApplication()
> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", line 120, in
> emit cb(self.context, *args)
> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", line
> 169, in post_argparse layout.initialize(args.source, args.init_nvr)
> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", line
> 216, in initialize self.app.imgbase.init_layout_from(source, init_nvr)
> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line 271,
> in init_layout_from self.init_tags_on(existing_lv)
> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line 243,
> in init_tags_on pool = lv.thinpool()
> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line 250, in
> thinpool raise MissingLvmThinPool() imgbased.lvm.MissingLvmThinPool
>
> At this point, the only option is to exit the installer.
>
> ****************************
>
> Being this a new install, please use 4.1. oVirt 4.0 is not supported
> anymore.
>
> Not a problem.
>
> ****************************
>
> Can you please provide hosted engine setup logs or better a full sos
> report? (sosreport -a)
>
> Again, the process I'm following:
> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 <(201)%20705-2604>
> .iso
> - During installation I setup one of the two interfaces and check the box
> to automatically use the connection.
> - I'm currently providing a host name of
node-1.test.net until I have a
> successful process.
> - I configure date and time for my timezone and to use an internal NTP
> server.
> - On Installation Destination, I pick my 128GB USB3.0 SanDisk flash
> drive, check the box that I would like to make additional space, and click
> done. In the reclaim disk space window, I click delete all, and then
> reclaim space.
> - Begin Installation and set a root password.
> - Perform a yum update - no packages marked for update (as expected)
> - Use vi to update /etc/hosts with a reference for
node-1.test.net and
>
engine.test.net
> - First attempt at hosted-engine from web-UI
> - Setup downloads and installs ovirt-engine-applianc
> e-4.1-20170523.1.el7.centos.noarch.rpm
> *Failed to execute stage 'Environment setup': Failed to reconfigure
> libvirt for VDSM
> *Hosted Engine deployment failed
> - Attached SOS report
> The checksum is: aa56097edc0b63c49caaf1a1fde021bc
>
> At this point, I would run ovirt-hosted-engine-cleanup and I would get
> further along in the install process. However, because this is a fresh
> install, I'm going to leave things here for now so you can review the SOS.
>
Thanks for the SOS report!
Hosted Engine setup fails on:
2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv
plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure',
'--force') stdout:
Checking configuration status...
Current revision of multipath.conf detected, preserving
lvm is configured for vdsm
libvirt is already configured for vdsm
SUCCESS: ssl configured to true. No conflicts
Running configure...
Reconfiguration of libvirt is done.
2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv
plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure',
'--force') stderr:
Error: ServiceOperationError: _systemctlStart failed
Job for libvirtd.service failed because the control process exited with
error code. See "systemctl status libvirtd.service" and "journalctl
-xe"
for details.
At the same time journalctl shows:
May 30 19:24:39
node-1.test.net libvirtd[20954]: libvirt version: 2.0.0,
package: 10.el7_3.5 (CentOS BuildSystem <
http://bugs.centos.org>,
2017-03-03-02:09:45,
c1bm.rdu2.centos.org)
May 30 19:24:39
node-1.test.net libvirtd[20954]: hostname:
node-1.test.net
May 30 19:24:39
node-1.test.net libvirtd[20954]: The server certificate
/etc/pki/vdsm/certs/vdsmcert.pem is not yet active
May 30 19:24:39
node-1.test.net systemd[1]: libvirtd.service: main
process exited, code=exited, status=6/NOTCONFIGURED
May 30 19:24:39
node-1.test.net systemd[1]: Failed to start
Virtualization daemon.
May 30 19:24:39
node-1.test.net systemd[1]: Unit libvirtd.service entered
failed state.
May 30 19:24:39
node-1.test.net systemd[1]: libvirtd.service failed.
May 30 19:24:39
node-1.test.net systemd[1]: libvirtd.service holdoff time
over, scheduling restart.
>
> ****************************
>
> I'd like to understand the issues you faced before suggesting to restart
> from scratch.
>
> Too late... I did two re-installs to get a more accurate account of my
> install process for above.
>
> ****************************
>
> Thank you for your help!
>
> Brendan
>
> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola <sbonazzo(a)redhat.com>
> wrote:
>
>>
>>
>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell <mrrex4(a)gmail.com>
>> wrote:
>>
>>> oVirt users list,
>>>
>>> Long story short, I've been spending weeks on this project for my home
>>> lab with no success.
>>>
>>> I would like to successfully install two nodes that host a highly
>>> available engine with an iSCSI storage back-end.
>>>
>>> I have read through most, if not all, of the guides on
ovirt.org with
>>> no substantial help.
>>>
>>> Successfully, I have done the following:
>>> Install oVirt Engine on a bare metal system, added a node, and started
>>> exploring - not desired.
>>> Install oVirt Node 4.0.6 on a bare metal system - fails if partitions
>>> are not done automatically.
>>> Install oVirt Node 4.1.2 on a bare metal system - fails if partitions
>>> are not done automatically.
>>>
>>
>> Can you please elaborate about the failure you see here and how are you
>> trying to manually partition the host?
>>
>>
>>
>>
>>
>>>
>>> My process after installing a Node:
>>> Run a yum update - just to be sure, but I am using latest iso images
>>> from downloads section.
>>> Edit /etc/hosts for local name resolution - the goal is to host DNS as
>>> a virtual machine, eventually.
>>> On 4.1 if I install ovirt-engine-appliance from yum, it does simplify
>>> one step in the hosted engine setup. If I do this on 4.0 it discards the
>>> image and uses the default.
>>>
>>
>> Being this a new install, please use 4.1. oVirt 4.0 is not supported
>> anymore.
>>
>>
>>
>>> On 4.1 the hosted engine setup fails immediately unless I run the
>>> hosted engine cleanup from the shell.
>>>
>>
>> Can you please provide hosted engine setup logs or better a full sos
>> report? (sosreport -a)
>>
>>
>>
>>
>>> If I do this, I can typically get to the point of installing.
>>>
>>> When I do get to the installation phase, I get to a point just after
>>> extracting the OVA that I get a message on the shell saying something about
>>> the watchdog running the whole time and then the node reboots.
>>>
>>> I found one email thread that sounded like my issue and suggested the
>>> following commands:
>>> vdsm-tool configure --force
>>> systemctl restart libvirtd
>>> systemctl restart vdsmd
>>>
>>> Unfortunately, these commands did not help my situation like the other
>>> individual.
>>>
>>> What log file would everyone like to see first? Given that I still
>>> consider myself relatively new to Linux, please identify the path for the
>>> log file requested.
>>>
>>
>> See above
>>
>>
>>>
>>> Also, because I plan on performing a clean install for this thread
>>> using my process from above (I'm not expecting my outcome to be any
>>> different), are there any tips and tricks that might result in a success?
>>>
>>
>> I'd like to understand the issues you faced before suggesting to restart
>> from scratch.
>> Adding some people who may help as well.
>>
>>
>>
>>>
>>> Thank you for any and all help,
>>> Brendan
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users(a)ovirt.org
>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>>
>> --
>>
>> SANDRO BONAZZOLA
>>
>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D
>>
>> Red Hat EMEA <
https://www.redhat.com/>
>> <
https://red.ht/sig>
>> TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>
>>
>
>
--
SANDRO BONAZZOLA
ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D
Red Hat EMEA <
https://www.redhat.com/>
<
https://red.ht/sig>
TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>