
On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell <mrrex4@gmail.com> wrote:
As requested,
It seams fine, there are no pending locks now. Could you please retry?
The output of ovirt-hosted-engine-cleanup
[root@node-1 ~]# ovirt-hosted-engine-cleanup This will de-configure the host to run ovirt-hosted-engine-setup from scratch. Caution, this operation should be used with care.
Are you sure you want to proceed? [y/n] y -=== Destroy hosted-engine VM ===- You must run deploy first -=== Stop HA services ===- -=== Shutdown sanlock ===- shutdown force 1 wait 0 shutdown done 0 -=== Disconnecting the hosted-engine storage domain ===- You must run deploy first -=== De-configure VDSM networks ===- -=== Stop other services ===- -=== De-configure external daemons ===- -=== Removing configuration files ===- ? /etc/init/libvirtd.conf already missing - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml ? /etc/ovirt-hosted-engine/answers.conf already missing ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing - removing /etc/vdsm/vdsm.conf - removing /etc/pki/vdsm/certs/cacert.pem - removing /etc/pki/vdsm/certs/vdsmcert.pem - removing /etc/pki/vdsm/keys/vdsmkey.pem - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem - removing /etc/pki/vdsm/libvirt-spice/server-key.pem ? /etc/pki/CA/cacert.pem already missing ? /etc/pki/libvirt/*.pem already missing ? /etc/pki/libvirt/private/*.pem already missing ? /etc/pki/ovirt-vmconsole/*.pem already missing - removing /var/cache/libvirt/qemu ? /var/run/ovirt-hosted-engine-ha/* already missing [root@node-1 ~]#
Output of sanlock client status: [root@node-1 ~]# sanlock client status [root@node-1 ~]#
Thank you for your help!
On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi <stirabos@redhat.com> wrote:
On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell <mrrex4@gmail.com> wrote:
After letting this sit for a few days, does anyone have any ideas as to how to deal with my situation? Would anyone like me to send the SOS report directly to them? It's a 9MB file.
If nothing comes up, I'm going to try and sift through the SOS report tonight, but I won't know what I'm trying to find.
Thank you for any and all help.
On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola <sbonazzo@redhat.com> wrote:
On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell <mrrex4@gmail.com> wrote:
Ran the 4 commands listed above, no errors on the screen.
Started the hosted-engine standard setup from the web-UI.
Using iSCSI for the storage.
Using mostly default options, I got these errors in the web-UI.
Error creating Volume Group: Failed to initialize physical device: ("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",) Failed to execute stage 'Misc configuration': Failed to initialize physical device: ("[u'/dev/mapper/36589cfc00000 0de7482638fcfcebbbb4']",) Hosted Engine deployment failed: this system is not reliable, please check the issue,fix and redeploy
I rebuilt my iSCSI (I don't think I cleaned it up from a previous install). Re-ran the above 4 commands. Restarted hosted engine standard setup from web-UI. Install moved past "Connecting Storage Pool" so I believe the above was my fault.
These are the last messages displayed on the web-UI. Creating Storage Pool Connecting Storage Pool Verifying sanlock lockspace initialization Creating Image for 'hosted-engine.lockspace' ... Image for 'hosted-engine.lockspace' created successfully Creating Image for 'hosted-engine.metadata' ... Image for 'hosted-engine.metadata' created successfully Creating VM Image Extracting disk image from OVF archive (could take a few minutes depending on archive size) Validating pre-allocated volume size Uploading volume to data domain (could take a few minutes depending on archive size)
At the host terminal, I got the error "watchdog watchdog0: watchdog did not stop!" Then the host restarted.
Simone, can you help here?
Ok, sorry for the delay. The second installation attempt seams fine but it seams that ovirt-hosted-engine-cleanup failed stopping sanlock and so the watchdog kick in rebooting your system in the middle of deployment attempt.
could you please post the output of ovirt-hosted-engine-cleanup sanlock client status ?
This is as far as I've gotten in previous attempts.
Attaching the hosted-engine-setup log.
The SOS report is 9MB and the ovirt users group will drop the email.
On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola <sbonazzo@redhat.com
wrote:
On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell <mrrex4@gmail.com> wrote:
> Now that you have identified the problem, should I run the following > commands and send you another SOS? > > ovirt-hosted-engine-cleanup > vdsm-tool configure --force > systemctl restart libvirtd > systemctl restart vdsm > > Or is there a different plan in mind? >
I would have expected someone from virt team to follow up for further investigations :-) above commands should work.
> > Thank you, > > Brendan > > On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola < > sbonazzo@redhat.com> wrote: > >> >> >> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell <mrrex4@gmail.com >> > wrote: >> >>> Can you please elaborate about the failure you see here and how >>> are you trying to manually partition the host? >>> >>> Sure, I will start from the beginning. >>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>> <(201)%20705-2604>.iso >>> - During installation I setup one of the two interfaces and check >>> the box to automatically use the connection. >>> - I'm currently providing a host name of node-1.test.net until I >>> have a successful process. >>> - I configure date and time for my timezone and to use an internal >>> NTP server. >>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>> flash drive, check the box that I would like to make additional space, and >>> click done. In the reclaim disk space window, I click delete all, and then >>> reclaim space. I go back into the Installation Destination, select that I >>> will configure partitioning, and click done. The Manual Partitioning >>> window opens, I use the option to automatically create mount points. >>> >> >> In this screen, please change partitioning scheme from LVM to LVM >> Thin Provisioning: it should solve your following error. >> >> >> >> >>> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, and >>> swap is 11.57GB. I then change / to 23.11 GB, update settings, change /var >>> to 80GB, update settings again, and click done. I accept the changes and >>> begin installation. >>> >>> I tried these changes based on this article: http://www.ovirt.org/ >>> documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/ >>> >>> The article does say that you can specify a different directory >>> than /var/tmp, but I don't recall seeing that option. >>> >> >> If the setup detects not enough space in /var/tmp for extracting >> the appliance it will ask about a different directory. >> >> >> >>> >>> After some time, I get the following error: >>> There was an error running the kickstart script at line 7. This >>> is a fatal error and installation will be aborted. The details of this >>> error are: >>> >>> [INFO] Trying to create a manageable base from '/' >>> [ERROR] LVM Thin Provisioning partitioning scheme is required. >>> For autoinstall via Kickstart with LVM Thin Provisioning check options >>> --thinpool and --grow. Please consult documentation for details. >>> >> >> >> ^^ this one should be solved by the LVM Thin Provisioning scheme >> mentioned above.. >> >> >> >>> >>> Traceback (most recent call last): >>> File "/usr/lib64/python2.7/runpy.py", line 162, in >>> _run_module_as_main "__main__", fname, loader, pkg_name) >>> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code exec >>> code in run_globals >>> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", >>> line 51, in <module> CliApplication() >>> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", >>> line 82, in CliApplication() >>> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", line >>> 120, in emit cb(self.context, *args) >>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>> line 169, in post_argparse layout.initialize(args.source, args.init_nvr) >>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>> line 216, in initialize self.app.imgbase.init_layout_from(source, >>> init_nvr) >>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line >>> 271, in init_layout_from self.init_tags_on(existing_lv) >>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", line >>> 243, in init_tags_on pool = lv.thinpool() >>> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line >>> 250, in thinpool raise MissingLvmThinPool() imgbased.lvm.MissingLvmThinPoo >>> l >>> >>> At this point, the only option is to exit the installer. >>> >>> **************************** >>> >>> Being this a new install, please use 4.1. oVirt 4.0 is not >>> supported anymore. >>> >>> Not a problem. >>> >>> **************************** >>> >>> Can you please provide hosted engine setup logs or better a full >>> sos report? (sosreport -a) >>> >>> Again, the process I'm following: >>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>> <(201)%20705-2604>.iso >>> - During installation I setup one of the two interfaces and check >>> the box to automatically use the connection. >>> - I'm currently providing a host name of node-1.test.net until I >>> have a successful process. >>> - I configure date and time for my timezone and to use an internal >>> NTP server. >>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>> flash drive, check the box that I would like to make additional space, and >>> click done. In the reclaim disk space window, I click delete all, and then >>> reclaim space. >>> - Begin Installation and set a root password. >>> - Perform a yum update - no packages marked for update (as >>> expected) >>> - Use vi to update /etc/hosts with a reference for node-1.test.net >>> and engine.test.net >>> - First attempt at hosted-engine from web-UI >>> - Setup downloads and installs ovirt-engine-applianc >>> e-4.1-20170523.1.el7.centos.noarch.rpm >>> *Failed to execute stage 'Environment setup': Failed to >>> reconfigure libvirt for VDSM >>> *Hosted Engine deployment failed >>> - Attached SOS report >>> The checksum is: aa56097edc0b63c49caaf1a1fde021bc >>> >>> At this point, I would run ovirt-hosted-engine-cleanup and I >>> would get further along in the install process. However, because this is a >>> fresh install, I'm going to leave things here for now so you can review the >>> SOS. >>> >> >> Thanks for the SOS report! >> Hosted Engine setup fails on: >> >> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >> plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure', >> '--force') stdout: >> >> Checking configuration status... >> >> Current revision of multipath.conf detected, preserving >> lvm is configured for vdsm >> libvirt is already configured for vdsm >> SUCCESS: ssl configured to true. No conflicts >> >> Running configure... >> Reconfiguration of libvirt is done. >> >> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >> plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure', >> '--force') stderr: >> Error: ServiceOperationError: _systemctlStart failed >> Job for libvirtd.service failed because the control process exited >> with error code. See "systemctl status libvirtd.service" and "journalctl >> -xe" for details. >> >> At the same time journalctl shows: >> >> May 30 19:24:39 node-1.test.net libvirtd[20954]: libvirt version: >> 2.0.0, package: 10.el7_3.5 (CentOS BuildSystem < >> http://bugs.centos.org>, 2017-03-03-02:09:45, c1bm.rdu2.centos.org) >> May 30 19:24:39 node-1.test.net libvirtd[20954]: hostname: >> node-1.test.net >> May 30 19:24:39 node-1.test.net libvirtd[20954]: The server >> certificate /etc/pki/vdsm/certs/vdsmcert.pem is not yet active >> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service: main >> process exited, code=exited, status=6/NOTCONFIGURED >> May 30 19:24:39 node-1.test.net systemd[1]: Failed to start >> Virtualization daemon. >> May 30 19:24:39 node-1.test.net systemd[1]: Unit libvirtd.service >> entered failed state. >> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >> failed. >> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >> holdoff time over, scheduling restart. >> >> >> >> >> >> >> >> >> >>> >>> **************************** >>> >>> I'd like to understand the issues you faced before suggesting to >>> restart from scratch. >>> >>> Too late... I did two re-installs to get a more accurate account >>> of my install process for above. >>> >>> **************************** >>> >>> Thank you for your help! >>> >>> Brendan >>> >>> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola < >>> sbonazzo@redhat.com> wrote: >>> >>>> >>>> >>>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell < >>>> mrrex4@gmail.com> wrote: >>>> >>>>> oVirt users list, >>>>> >>>>> Long story short, I've been spending weeks on this project for >>>>> my home lab with no success. >>>>> >>>>> I would like to successfully install two nodes that host a >>>>> highly available engine with an iSCSI storage back-end. >>>>> >>>>> I have read through most, if not all, of the guides on ovirt.org >>>>> with no substantial help. >>>>> >>>>> Successfully, I have done the following: >>>>> Install oVirt Engine on a bare metal system, added a node, and >>>>> started exploring - not desired. >>>>> Install oVirt Node 4.0.6 on a bare metal system - fails if >>>>> partitions are not done automatically. >>>>> Install oVirt Node 4.1.2 on a bare metal system - fails if >>>>> partitions are not done automatically. >>>>> >>>> >>>> Can you please elaborate about the failure you see here and how >>>> are you trying to manually partition the host? >>>> >>>> >>>> >>>> >>>> >>>>> >>>>> My process after installing a Node: >>>>> Run a yum update - just to be sure, but I am using latest iso >>>>> images from downloads section. >>>>> Edit /etc/hosts for local name resolution - the goal is to host >>>>> DNS as a virtual machine, eventually. >>>>> On 4.1 if I install ovirt-engine-appliance from yum, it does >>>>> simplify one step in the hosted engine setup. If I do this on 4.0 it >>>>> discards the image and uses the default. >>>>> >>>> >>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>> supported anymore. >>>> >>>> >>>> >>>>> On 4.1 the hosted engine setup fails immediately unless I run >>>>> the hosted engine cleanup from the shell. >>>>> >>>> >>>> Can you please provide hosted engine setup logs or better a full >>>> sos report? (sosreport -a) >>>> >>>> >>>> >>>> >>>>> If I do this, I can typically get to the point of installing. >>>>> >>>>> When I do get to the installation phase, I get to a point just >>>>> after extracting the OVA that I get a message on the shell saying something >>>>> about the watchdog running the whole time and then the node reboots. >>>>> >>>>> I found one email thread that sounded like my issue and >>>>> suggested the following commands: >>>>> vdsm-tool configure --force >>>>> systemctl restart libvirtd >>>>> systemctl restart vdsmd >>>>> >>>>> Unfortunately, these commands did not help my situation like the >>>>> other individual. >>>>> >>>>> What log file would everyone like to see first? Given that I >>>>> still consider myself relatively new to Linux, please identify the path for >>>>> the log file requested. >>>>> >>>> >>>> See above >>>> >>>> >>>>> >>>>> Also, because I plan on performing a clean install for this >>>>> thread using my process from above (I'm not expecting my outcome to be any >>>>> different), are there any tips and tricks that might result in a success? >>>>> >>>> >>>> I'd like to understand the issues you faced before suggesting to >>>> restart from scratch. >>>> Adding some people who may help as well. >>>> >>>> >>>> >>>>> >>>>> Thank you for any and all help, >>>>> Brendan >>>>> >>>>> _______________________________________________ >>>>> Users mailing list >>>>> Users@ovirt.org >>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>> >>>>> >>>> >>>> >>>> -- >>>> >>>> SANDRO BONAZZOLA >>>> >>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION >>>> R&D >>>> >>>> Red Hat EMEA <https://www.redhat.com/> >>>> <https://red.ht/sig> >>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>> >>> >>> >> >> >> -- >> >> SANDRO BONAZZOLA >> >> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >> >> Red Hat EMEA <https://www.redhat.com/> >> <https://red.ht/sig> >> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >> > >
--
SANDRO BONAZZOLA
ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D
Red Hat EMEA <https://www.redhat.com/> <https://red.ht/sig> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
--
SANDRO BONAZZOLA
ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D
Red Hat EMEA <https://www.redhat.com/> <https://red.ht/sig> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>