HI Brendan,
Can you please send the output for systemctl status vdsmd and journalctl -u
vdsmd.service ?
Thanks,
On Wed, Jun 7, 2017 at 9:32 AM, Sandro Bonazzola <sbonazzo(a)redhat.com>
wrote:
On Tue, Jun 6, 2017 at 2:56 PM, Brendan Hartzell <mrrex4(a)gmail.com> wrote:
> Upon login to the server, to watch terminal output, I noticed that the
> node status is degraded.
>
> [root@node-1 ~]# nodectl check
> Status: WARN
> Bootloader ... OK
> Layer boot entries ... OK
> Valid boot entries ... OK
> Mount points ... OK
> Separate /var ... OK
> Discard is used ... OK
> Basic storage ... OK
> Initialized VG ... OK
> Initialized Thin Pool ... OK
> Initialized LVs ... OK
> Thin storage ... OK
> Checking available space in thinpool ... OK
> Checking thinpool auto-extend ... OK
> vdsmd ... BAD
>
Yuval, can you help here?
>
>
> Pressing forward with the retry using the web-UI.
>
> After resetting my iSCSI storage (on the storage server side), Install
> started.
>
> Status in the web-UI:
> Creating Storage Domain
> Creating Storage Pool
> Connecting Storage Pool
> Verifying sanlock lockspace initialization
> Creating Image for 'hosted-engine.lockspace' ...
> Image for 'hosted-engine.lockspace' created successfully
> Creating Image for 'hosted-engine.metadata' ...
> Image for 'hosted-engine.metadata' created successfully
> Creating VM Image
> Extracting disk image from OVF archive (could take a few minutes
> depending on archive size)
> Validating pre-allocated volume size
>
> Output from the terminal:
> [45863.076979]watchdog watchdog0: watchdog did not stop!
>
> System restarted.
>
> Attaching ovirt-hosted-engine-setup log.
>
> I'm running an SOS report, but it's too big for the users list. I can
> email it directly to you upon request.
>
> On Tue, Jun 6, 2017 at 12:12 AM, Simone Tiraboschi <stirabos(a)redhat.com>
> wrote:
>
>>
>>
>> On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell <mrrex4(a)gmail.com>
>> wrote:
>>
>>> As requested,
>>>
>>
>> It seams fine, there are no pending locks now.
>> Could you please retry?
>>
>>
>>>
>>> The output of ovirt-hosted-engine-cleanup
>>>
>>> [root@node-1 ~]# ovirt-hosted-engine-cleanup
>>> This will de-configure the host to run ovirt-hosted-engine-setup from
>>> scratch.
>>> Caution, this operation should be used with care.
>>>
>>> Are you sure you want to proceed? [y/n]
>>> y
>>> -=== Destroy hosted-engine VM ===-
>>> You must run deploy first
>>> -=== Stop HA services ===-
>>> -=== Shutdown sanlock ===-
>>> shutdown force 1 wait 0
>>> shutdown done 0
>>> -=== Disconnecting the hosted-engine storage domain ===-
>>> You must run deploy first
>>> -=== De-configure VDSM networks ===-
>>> -=== Stop other services ===-
>>> -=== De-configure external daemons ===-
>>> -=== Removing configuration files ===-
>>> ? /etc/init/libvirtd.conf already missing
>>> - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml
>>> ? /etc/ovirt-hosted-engine/answers.conf already missing
>>> ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing
>>> - removing /etc/vdsm/vdsm.conf
>>> - removing /etc/pki/vdsm/certs/cacert.pem
>>> - removing /etc/pki/vdsm/certs/vdsmcert.pem
>>> - removing /etc/pki/vdsm/keys/vdsmkey.pem
>>> - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem
>>> - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem
>>> - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem
>>> - removing /etc/pki/vdsm/libvirt-spice/server-key.pem
>>> ? /etc/pki/CA/cacert.pem already missing
>>> ? /etc/pki/libvirt/*.pem already missing
>>> ? /etc/pki/libvirt/private/*.pem already missing
>>> ? /etc/pki/ovirt-vmconsole/*.pem already missing
>>> - removing /var/cache/libvirt/qemu
>>> ? /var/run/ovirt-hosted-engine-ha/* already missing
>>> [root@node-1 ~]#
>>>
>>> Output of sanlock client status:
>>> [root@node-1 ~]# sanlock client status
>>> [root@node-1 ~]#
>>>
>>> Thank you for your help!
>>>
>>> On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi
<stirabos(a)redhat.com>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell
<mrrex4(a)gmail.com>
>>>> wrote:
>>>>
>>>>> After letting this sit for a few days, does anyone have any ideas as
>>>>> to how to deal with my situation? Would anyone like me to send the
SOS
>>>>> report directly to them? It's a 9MB file.
>>>>>
>>>>> If nothing comes up, I'm going to try and sift through the SOS
report
>>>>> tonight, but I won't know what I'm trying to find.
>>>>>
>>>>> Thank you for any and all help.
>>>>>
>>>>> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola
<sbonazzo(a)redhat.com
>>>>> > wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell
<mrrex4(a)gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Ran the 4 commands listed above, no errors on the screen.
>>>>>>>
>>>>>>> Started the hosted-engine standard setup from the web-UI.
>>>>>>>
>>>>>>> Using iSCSI for the storage.
>>>>>>>
>>>>>>> Using mostly default options, I got these errors in the
web-UI.
>>>>>>>
>>>>>>> Error creating Volume Group: Failed to initialize physical
device:
>>>>>>>
("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",)
>>>>>>> Failed to execute stage 'Misc configuration': Failed
to initialize
>>>>>>> physical device: ("[u'/dev/mapper/36589cfc00000
>>>>>>> 0de7482638fcfcebbbb4']",)
>>>>>>> Hosted Engine deployment failed: this system is not
reliable,
>>>>>>> please check the issue,fix and redeploy
>>>>>>>
>>>>>>> I rebuilt my iSCSI (I don't think I cleaned it up from a
previous
>>>>>>> install).
>>>>>>> Re-ran the above 4 commands.
>>>>>>> Restarted hosted engine standard setup from web-UI.
>>>>>>> Install moved past "Connecting Storage Pool" so I
believe the above
>>>>>>> was my fault.
>>>>>>>
>>>>>>> These are the last messages displayed on the web-UI.
>>>>>>> Creating Storage Pool
>>>>>>> Connecting Storage Pool
>>>>>>> Verifying sanlock lockspace initialization
>>>>>>> Creating Image for 'hosted-engine.lockspace' ...
>>>>>>> Image for 'hosted-engine.lockspace' created
successfully
>>>>>>> Creating Image for 'hosted-engine.metadata' ...
>>>>>>> Image for 'hosted-engine.metadata' created
successfully
>>>>>>> Creating VM Image
>>>>>>> Extracting disk image from OVF archive (could take a few
minutes
>>>>>>> depending on archive size)
>>>>>>> Validating pre-allocated volume size
>>>>>>> Uploading volume to data domain (could take a few minutes
depending
>>>>>>> on archive size)
>>>>>>>
>>>>>>> At the host terminal, I got the error "watchdog
watchdog0: watchdog
>>>>>>> did not stop!"
>>>>>>> Then the host restarted.
>>>>>>>
>>>>>>
>>>>>> Simone, can you help here?
>>>>>>
>>>>>>
>>>> Ok, sorry for the delay.
>>>> The second installation attempt seams fine but it seams that
>>>> ovirt-hosted-engine-cleanup failed stopping sanlock and so the
>>>> watchdog kick in rebooting your system in the middle of deployment
attempt.
>>>>
>>>> could you please post the output of
>>>> ovirt-hosted-engine-cleanup
>>>> sanlock client status
>>>> ?
>>>>
>>>>
>>>>
>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> This is as far as I've gotten in previous attempts.
>>>>>>>
>>>>>>> Attaching the hosted-engine-setup log.
>>>>>>>
>>>>>>> The SOS report is 9MB and the ovirt users group will drop the
email.
>>>>>>>
>>>>>>> On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola <
>>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell <
>>>>>>>> mrrex4(a)gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Now that you have identified the problem, should I
run the
>>>>>>>>> following commands and send you another SOS?
>>>>>>>>>
>>>>>>>>> ovirt-hosted-engine-cleanup
>>>>>>>>> vdsm-tool configure --force
>>>>>>>>> systemctl restart libvirtd
>>>>>>>>> systemctl restart vdsm
>>>>>>>>>
>>>>>>>>> Or is there a different plan in mind?
>>>>>>>>>
>>>>>>>>
>>>>>>>> I would have expected someone from virt team to follow up
for
>>>>>>>> further investigations :-)
>>>>>>>> above commands should work.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thank you,
>>>>>>>>>
>>>>>>>>> Brendan
>>>>>>>>>
>>>>>>>>> On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola
<
>>>>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell
<
>>>>>>>>>> mrrex4(a)gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Can you please elaborate about the failure
you see here and how
>>>>>>>>>>> are you trying to manually partition the
host?
>>>>>>>>>>>
>>>>>>>>>>> Sure, I will start from the beginning.
>>>>>>>>>>> - Using:
ovirt-node-ng-installer-ovirt-4.1-2017052604
>>>>>>>>>>> <(201)%20705-2604>.iso
>>>>>>>>>>> - During installation I setup one of the two
interfaces and
>>>>>>>>>>> check the box to automatically use the
connection.
>>>>>>>>>>> - I'm currently providing a host name of
node-1.test.net until
>>>>>>>>>>> I have a successful process.
>>>>>>>>>>> - I configure date and time for my timezone
and to use an
>>>>>>>>>>> internal NTP server.
>>>>>>>>>>> - On Installation Destination, I pick my
128GB USB3.0 SanDisk
>>>>>>>>>>> flash drive, check the box that I would like
to make additional space, and
>>>>>>>>>>> click done. In the reclaim disk space
window, I click delete all, and then
>>>>>>>>>>> reclaim space. I go back into the
Installation Destination, select that I
>>>>>>>>>>> will configure partitioning, and click done.
The Manual Partitioning
>>>>>>>>>>> window opens, I use the option to
automatically create mount points.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> In this screen, please change partitioning scheme
from LVM to
>>>>>>>>>> LVM Thin Provisioning: it should solve your
following error.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> At this point, /boot is 1024MB, /var is
15GB, / is 88.11 GB,
>>>>>>>>>>> and swap is 11.57GB. I then change / to
23.11 GB, update settings, change
>>>>>>>>>>> /var to 80GB, update settings again, and
click done. I accept the changes
>>>>>>>>>>> and begin installation.
>>>>>>>>>>>
>>>>>>>>>>> I tried these changes based on this article:
>>>>>>>>>>>
http://www.ovirt.org/documentation/self-hosted/chap
>>>>>>>>>>> -Deploying_Self-Hosted_Engine/
>>>>>>>>>>>
>>>>>>>>>>> The article does say that you can specify a
different directory
>>>>>>>>>>> than /var/tmp, but I don't recall seeing
that option.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> If the setup detects not enough space in /var/tmp
for extracting
>>>>>>>>>> the appliance it will ask about a different
directory.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> After some time, I get the following error:
>>>>>>>>>>> There was an error running the kickstart
script at line 7.
>>>>>>>>>>> This is a fatal error and installation will
be aborted. The details of
>>>>>>>>>>> this error are:
>>>>>>>>>>>
>>>>>>>>>>> [INFO] Trying to create a manageable base
from '/'
>>>>>>>>>>> [ERROR] LVM Thin Provisioning partitioning
scheme is required.
>>>>>>>>>>> For autoinstall via Kickstart with LVM Thin
Provisioning check options
>>>>>>>>>>> --thinpool and --grow. Please consult
documentation for details.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> ^^ this one should be solved by the LVM Thin
Provisioning scheme
>>>>>>>>>> mentioned above..
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>> File
"/usr/lib64/python2.7/runpy.py", line 162, in
>>>>>>>>>>> _run_module_as_main "__main__",
fname, loader, pkg_name)
>>>>>>>>>>> File
"/usr/lib64/python2.7runpy.py", line 72, in _run_code
>>>>>>>>>>> exec code in run_globals
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/__main__.py",
>>>>>>>>>>> line 51, in <module> CliApplication()
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/__init__.py",
>>>>>>>>>>> line 82, in CliApplication()
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/hooks.py",
>>>>>>>>>>> line 120, in emit cb(self.context, *args)
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/plugins/core.py",
>>>>>>>>>>> line 169, in post_argparse
layout.initialize(args.source, args.init_nvr)
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/plugins/core.py",
>>>>>>>>>>> line 216, in initialize
self.app.imgbase.init_layout_from(source,
>>>>>>>>>>> init_nvr)
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/imgbase.py",
>>>>>>>>>>> line 271, in init_layout_from
self.init_tags_on(existing_lv)
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/imgbase.py",
>>>>>>>>>>> line 243, in init_tags_on pool =
lv.thinpool()
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/imgbased/lvm.py", line
>>>>>>>>>>> 250, in thinpool raise MissingLvmThinPool()
imgbased.lvm.MissingLvmThinPoo
>>>>>>>>>>> l
>>>>>>>>>>>
>>>>>>>>>>> At this point, the only option is to exit the
installer.
>>>>>>>>>>>
>>>>>>>>>>> ****************************
>>>>>>>>>>>
>>>>>>>>>>> Being this a new install, please use 4.1.
oVirt 4.0 is not
>>>>>>>>>>> supported anymore.
>>>>>>>>>>>
>>>>>>>>>>> Not a problem.
>>>>>>>>>>>
>>>>>>>>>>> ****************************
>>>>>>>>>>>
>>>>>>>>>>> Can you please provide hosted engine setup
logs or better a
>>>>>>>>>>> full sos report? (sosreport -a)
>>>>>>>>>>>
>>>>>>>>>>> Again, the process I'm following:
>>>>>>>>>>> - Using:
ovirt-node-ng-installer-ovirt-4.1-2017052604
>>>>>>>>>>> <(201)%20705-2604>.iso
>>>>>>>>>>> - During installation I setup one of the two
interfaces and
>>>>>>>>>>> check the box to automatically use the
connection.
>>>>>>>>>>> - I'm currently providing a host name of
node-1.test.net until
>>>>>>>>>>> I have a successful process.
>>>>>>>>>>> - I configure date and time for my timezone
and to use an
>>>>>>>>>>> internal NTP server.
>>>>>>>>>>> - On Installation Destination, I pick my
128GB USB3.0 SanDisk
>>>>>>>>>>> flash drive, check the box that I would like
to make additional space, and
>>>>>>>>>>> click done. In the reclaim disk space window,
I click delete all, and then
>>>>>>>>>>> reclaim space.
>>>>>>>>>>> - Begin Installation and set a root
password.
>>>>>>>>>>> - Perform a yum update - no packages marked
for update (as
>>>>>>>>>>> expected)
>>>>>>>>>>> - Use vi to update /etc/hosts with a
reference for
>>>>>>>>>>>
node-1.test.net and
engine.test.net
>>>>>>>>>>> - First attempt at hosted-engine from web-UI
>>>>>>>>>>> - Setup downloads and installs
ovirt-engine-applianc
>>>>>>>>>>> e-4.1-20170523.1.el7.centos.noarch.rpm
>>>>>>>>>>> *Failed to execute stage 'Environment
setup': Failed to
>>>>>>>>>>> reconfigure libvirt for VDSM
>>>>>>>>>>> *Hosted Engine deployment failed
>>>>>>>>>>> - Attached SOS report
>>>>>>>>>>> The checksum is:
aa56097edc0b63c49caaf1a1fde021bc
>>>>>>>>>>>
>>>>>>>>>>> At this point, I would run
ovirt-hosted-engine-cleanup and I
>>>>>>>>>>> would get further along in the install
process. However, because this is a
>>>>>>>>>>> fresh install, I'm going to leave things
here for now so you can review the
>>>>>>>>>>> SOS.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Thanks for the SOS report!
>>>>>>>>>> Hosted Engine setup fails on:
>>>>>>>>>>
>>>>>>>>>> 2017-05-30 19:24:39 DEBUG
otopi.plugins.gr_he_setup.system.vdsmenv
>>>>>>>>>> plugin.execute:921 execute-output:
('/bin/vdsm-tool', 'configure',
>>>>>>>>>> '--force') stdout:
>>>>>>>>>>
>>>>>>>>>> Checking configuration status...
>>>>>>>>>>
>>>>>>>>>> Current revision of multipath.conf detected,
preserving
>>>>>>>>>> lvm is configured for vdsm
>>>>>>>>>> libvirt is already configured for vdsm
>>>>>>>>>> SUCCESS: ssl configured to true. No conflicts
>>>>>>>>>>
>>>>>>>>>> Running configure...
>>>>>>>>>> Reconfiguration of libvirt is done.
>>>>>>>>>>
>>>>>>>>>> 2017-05-30 19:24:39 DEBUG
otopi.plugins.gr_he_setup.system.vdsmenv
>>>>>>>>>> plugin.execute:926 execute-output:
('/bin/vdsm-tool', 'configure',
>>>>>>>>>> '--force') stderr:
>>>>>>>>>> Error: ServiceOperationError: _systemctlStart
failed
>>>>>>>>>> Job for libvirtd.service failed because the
control process
>>>>>>>>>> exited with error code. See "systemctl
status libvirtd.service" and
>>>>>>>>>> "journalctl -xe" for details.
>>>>>>>>>>
>>>>>>>>>> At the same time journalctl shows:
>>>>>>>>>>
>>>>>>>>>> May 30 19:24:39
node-1.test.net libvirtd[20954]:
libvirt
>>>>>>>>>> version: 2.0.0, package: 10.el7_3.5 (CentOS
BuildSystem <
>>>>>>>>>>
http://bugs.centos.org>, 2017-03-03-02:09:45,
>>>>>>>>>>
c1bm.rdu2.centos.org)
>>>>>>>>>> May 30 19:24:39
node-1.test.net libvirtd[20954]:
hostname:
>>>>>>>>>>
node-1.test.net
>>>>>>>>>> May 30 19:24:39
node-1.test.net libvirtd[20954]:
The server
>>>>>>>>>> certificate /etc/pki/vdsm/certs/vdsmcert.pem is
not yet active
>>>>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]:
libvirtd.service:
>>>>>>>>>> main process exited, code=exited,
status=6/NOTCONFIGURED
>>>>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]:
Failed to start
>>>>>>>>>> Virtualization daemon.
>>>>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]: Unit
>>>>>>>>>> libvirtd.service entered failed state.
>>>>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]:
libvirtd.service
>>>>>>>>>> failed.
>>>>>>>>>> May 30 19:24:39
node-1.test.net systemd[1]:
libvirtd.service
>>>>>>>>>> holdoff time over, scheduling restart.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> ****************************
>>>>>>>>>>>
>>>>>>>>>>> I'd like to understand the issues you
faced before suggesting
>>>>>>>>>>> to restart from scratch.
>>>>>>>>>>>
>>>>>>>>>>> Too late... I did two re-installs to get a
more accurate
>>>>>>>>>>> account of my install process for above.
>>>>>>>>>>>
>>>>>>>>>>> ****************************
>>>>>>>>>>>
>>>>>>>>>>> Thank you for your help!
>>>>>>>>>>>
>>>>>>>>>>> Brendan
>>>>>>>>>>>
>>>>>>>>>>> On Tue, May 30, 2017 at 12:17 AM, Sandro
Bonazzola <
>>>>>>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, May 30, 2017 at 6:49 AM, Brendan
Hartzell <
>>>>>>>>>>>> mrrex4(a)gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> oVirt users list,
>>>>>>>>>>>>>
>>>>>>>>>>>>> Long story short, I've been
spending weeks on this project
>>>>>>>>>>>>> for my home lab with no success.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I would like to successfully install
two nodes that host a
>>>>>>>>>>>>> highly available engine with an iSCSI
storage back-end.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have read through most, if not all,
of the guides on
>>>>>>>>>>>>>
ovirt.org with no substantial help.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Successfully, I have done the
following:
>>>>>>>>>>>>> Install oVirt Engine on a bare metal
system, added a node,
>>>>>>>>>>>>> and started exploring - not desired.
>>>>>>>>>>>>> Install oVirt Node 4.0.6 on a bare
metal system - fails if
>>>>>>>>>>>>> partitions are not done
automatically.
>>>>>>>>>>>>> Install oVirt Node 4.1.2 on a bare
metal system - fails if
>>>>>>>>>>>>> partitions are not done
automatically.
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Can you please elaborate about the
failure you see here and
>>>>>>>>>>>> how are you trying to manually partition
the host?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> My process after installing a Node:
>>>>>>>>>>>>> Run a yum update - just to be sure,
but I am using latest iso
>>>>>>>>>>>>> images from downloads section.
>>>>>>>>>>>>> Edit /etc/hosts for local name
resolution - the goal is to
>>>>>>>>>>>>> host DNS as a virtual machine,
eventually.
>>>>>>>>>>>>> On 4.1 if I install
ovirt-engine-appliance from yum, it does
>>>>>>>>>>>>> simplify one step in the hosted
engine setup. If I do this on 4.0 it
>>>>>>>>>>>>> discards the image and uses the
default.
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Being this a new install, please use 4.1.
oVirt 4.0 is not
>>>>>>>>>>>> supported anymore.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>> On 4.1 the hosted engine setup fails
immediately unless I run
>>>>>>>>>>>>> the hosted engine cleanup from the
shell.
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Can you please provide hosted engine
setup logs or better a
>>>>>>>>>>>> full sos report? (sosreport -a)
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>> If I do this, I can typically get
to the point of
>>>>>>>>>>>>> installing.
>>>>>>>>>>>>>
>>>>>>>>>>>>> When I do get to the installation
phase, I get to a point
>>>>>>>>>>>>> just after extracting the OVA that I
get a message on the shell saying
>>>>>>>>>>>>> something about the watchdog running
the whole time and then the node
>>>>>>>>>>>>> reboots.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I found one email thread that sounded
like my issue and
>>>>>>>>>>>>> suggested the following commands:
>>>>>>>>>>>>> vdsm-tool configure --force
>>>>>>>>>>>>> systemctl restart libvirtd
>>>>>>>>>>>>> systemctl restart vdsmd
>>>>>>>>>>>>>
>>>>>>>>>>>>> Unfortunately, these commands did not
help my situation like
>>>>>>>>>>>>> the other individual.
>>>>>>>>>>>>>
>>>>>>>>>>>>> What log file would everyone like to
see first? Given that I
>>>>>>>>>>>>> still consider myself relatively new
to Linux, please identify the path for
>>>>>>>>>>>>> the log file requested.
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> See above
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Also, because I plan on performing a
clean install for this
>>>>>>>>>>>>> thread using my process from above
(I'm not expecting my outcome to be any
>>>>>>>>>>>>> different), are there any tips and
tricks that might result in a success?
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> I'd like to understand the issues you
faced before suggesting
>>>>>>>>>>>> to restart from scratch.
>>>>>>>>>>>> Adding some people who may help as well.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thank you for any and all help,
>>>>>>>>>>>>> Brendan
>>>>>>>>>>>>>
>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>> Users mailing list
>>>>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>>
>>>>>>>>>>>> SANDRO BONAZZOLA
>>>>>>>>>>>>
>>>>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING,
EMEA ENG
>>>>>>>>>>>> VIRTUALIZATION R&D
>>>>>>>>>>>>
>>>>>>>>>>>> Red Hat EMEA
<
https://www.redhat.com/>
>>>>>>>>>>>> <
https://red.ht/sig>
>>>>>>>>>>>> TRIED. TESTED. TRUSTED.
<
https://redhat.com/trusted>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>>
>>>>>>>>>> SANDRO BONAZZOLA
>>>>>>>>>>
>>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG
VIRTUALIZATION
>>>>>>>>>> R&D
>>>>>>>>>>
>>>>>>>>>> Red Hat EMEA <
https://www.redhat.com/>
>>>>>>>>>> <
https://red.ht/sig>
>>>>>>>>>> TRIED. TESTED. TRUSTED.
<
https://redhat.com/trusted>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>>
>>>>>>>> SANDRO BONAZZOLA
>>>>>>>>
>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG
VIRTUALIZATION
>>>>>>>> R&D
>>>>>>>>
>>>>>>>> Red Hat EMEA <
https://www.redhat.com/>
>>>>>>>> <
https://red.ht/sig>
>>>>>>>> TRIED. TESTED. TRUSTED.
<
https://redhat.com/trusted>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>>
>>>>>> SANDRO BONAZZOLA
>>>>>>
>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION
R&D
>>>>>>
>>>>>> Red Hat EMEA <
https://www.redhat.com/>
>>>>>> <
https://red.ht/sig>
>>>>>> TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>
--
SANDRO BONAZZOLA
ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D
Red Hat EMEA <
https://www.redhat.com/>
<
https://red.ht/sig>
TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>