On the host that has the Hosted Engine VM, the sanlock.log reports:
2017-06-27 17:30:20+0100 1043742 [7307]: add_lockspace
207221b2-959b-426b-b945-18e1adfed62f:3:/dev/207221b2-959b-426b-b945-18e1adfed62f/ids:0
conflicts with name of list1 s5
207221b2-959b-426b-b945-18e1adfed62f:1:/dev/207221b2-959b-426b-b945-18e1adfed62f/ids:0
Again, I'm not sure what has happened here.
On Tue, Jun 27, 2017 at 5:26 PM, cmc <iucounu(a)gmail.com> wrote:
I see this on the host it is trying to migrate in /var/log/sanlock:
2017-06-27 17:10:40+0100 527703 [2407]: s3528 lockspace
207221b2-959b-426b-b945-18e1adfed62f:1:/dev/207221b2-959b-426b-b945-18e1adfed62f/ids:0
2017-06-27 17:13:00+0100 527843 [27446]: s3528 delta_acquire host_id 1
busy1 1 2 1042692 3d4ec963-8486-43a2-a7d9-afa82508f89f.kvm-ldn-03
2017-06-27 17:13:01+0100 527844 [2407]: s3528 add_lockspace fail result -262
The sanlock service is running. Why would this occur?
Thanks,
C
On Tue, Jun 27, 2017 at 5:21 PM, cmc <iucounu(a)gmail.com> wrote:
> Hi Martin,
>
> Thanks for the reply. I have done this, and the deployment completed
> without error. However, it still will not allow the Hosted Engine
> migrate to another host. The
> /etc/ovirt-hosted-engine/hosted-engine.conf got created ok on the host
> I re-installed, but the ovirt-ha-broker.service, though it starts,
> reports:
>
> --------------------8<-------------------
>
> Jun 27 14:58:26 kvm-ldn-01 systemd[1]: Starting oVirt Hosted Engine
> High Availability Communications Broker...
> Jun 27 14:58:27 kvm-ldn-01 ovirt-ha-broker[6101]: ovirt-ha-broker
> ovirt_hosted_engine_ha.broker.storage_broker.StorageBroker ERROR
> Failed to read metadata from
>
/rhev/data-center/mnt/blockSD/207221b2-959b-426b-b945-18e1adfed62f/ha_agent/hosted-engine.metadata
> Traceback (most
> recent call last):
> File
>
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py",
> line 129, in get_raw_stats_for_service_type
> f =
> os.open(path, direct_flag | os.O_RDONLY | os.O_SYNC)
> OSError: [Errno 2]
> No such file or directory:
>
'/rhev/data-center/mnt/blockSD/207221b2-959b-426b-b945-18e1adfed62f/ha_agent/hosted-engine.metadata'
>
> --------------------8<-------------------
>
> I checked the path, and it exists. I can run 'less -f' on it fine. The
> perms are slightly different on the host that is running the VM vs the
> one that is reporting errors (600 vs 660), ownership is vdsm:qemu. Is
> this a san locking issue?
>
> Thanks for any help,
>
> Cam
>
> On Tue, Jun 27, 2017 at 1:41 PM, Martin Sivak <msivak(a)redhat.com> wrote:
>>> Should it be? It was not in the instructions for the migration from
>>> bare-metal to Hosted VM
>>
>> The hosted engine will only migrate to hosts that have the services
>> running. Please put one other host to maintenance and select Hosted
>> engine action: DEPLOY in the reinstall dialog.
>>
>> Best regards
>>
>> Martin Sivak
>>
>> On Tue, Jun 27, 2017 at 1:23 PM, cmc <iucounu(a)gmail.com> wrote:
>>> I changed the 'os.other.devices.display.protocols.value.3.6 =
>>> spice/qxl,vnc/cirrus,vnc/qxl' line to have the same display protocols
>>> as 4 and the hosted engine now appears in the list of VMs. I am
>>> guessing the compatibility version was causing it to use the 3.6
>>> version. However, I am still unable to migrate the engine VM to
>>> another host. When I try putting the host it is currently on into
>>> maintenance, it reports:
>>>
>>> Error while executing action: Cannot switch the Host(s) to Maintenance mode.
>>> There are no available hosts capable of running the engine VM.
>>>
>>> Running 'hosted-engine --vm-status' still shows 'Engine status:
>>> unknown stale-data'.
>>>
>>> The ovirt-ha-broker service is only running on one host. It was set to
>>> 'disabled' in systemd. It won't start as there is no
>>> /etc/ovirt-hosted-engine/hosted-engine.conf on the other two hosts.
>>> Should it be? It was not in the instructions for the migration from
>>> bare-metal to Hosted VM
>>>
>>> Thanks,
>>>
>>> Cam
>>>
>>> On Thu, Jun 22, 2017 at 1:07 PM, cmc <iucounu(a)gmail.com> wrote:
>>>> Hi Tomas,
>>>>
>>>> So in my /usr/share/ovirt-engine/conf/osinfo-defaults.properties on my
>>>> engine VM, I have:
>>>>
>>>> os.other.devices.display.protocols.value =
spice/qxl,vnc/vga,vnc/qxl,vnc/cirrus
>>>> os.other.devices.display.protocols.value.3.6 =
spice/qxl,vnc/cirrus,vnc/qxl
>>>>
>>>> That seems to match - I assume since this is 4.1, the 3.6 should not
apply
>>>>
>>>> Is there somewhere else I should be looking?
>>>>
>>>> Thanks,
>>>>
>>>> Cam
>>>>
>>>> On Thu, Jun 22, 2017 at 11:40 AM, Tomas Jelinek
<tjelinek(a)redhat.com> wrote:
>>>>>
>>>>>
>>>>> On Thu, Jun 22, 2017 at 12:38 PM, Michal Skrivanek
>>>>> <michal.skrivanek(a)redhat.com> wrote:
>>>>>>
>>>>>>
>>>>>> > On 22 Jun 2017, at 12:31, Martin Sivak
<msivak(a)redhat.com> wrote:
>>>>>> >
>>>>>> > Tomas, what fields are needed in a VM to pass the check that
causes
>>>>>> > the following error?
>>>>>> >
>>>>>> >>>>> WARN
[org.ovirt.engine.core.bll.exportimport.ImportVmCommand]
>>>>>> >>>>> (org.ovirt.thread.pool-6-thread-23) []
Validation of action
>>>>>> >>>>> 'ImportVm'
>>>>>> >>>>> failed for user SYSTEM. Reasons:
VAR__ACTION__IMPORT
>>>>>> >>>>>
>>>>>> >>>>>
,VAR__TYPE__VM,ACTION_TYPE_FAILED_ILLEGAL_VM_DISPLAY_TYPE_IS_NOT_SUPPORTED_BY_OS
>>>>>>
>>>>>> to match the OS and VM Display type;-)
>>>>>> Configuration is in osinfo….e.g. if that is import from older
releases on
>>>>>> Linux this is typically caused by the cahgen of cirrus to vga for
non-SPICE
>>>>>> VMs
>>>>>
>>>>>
>>>>> yep, the default supported combinations for 4.0+ is this:
>>>>> os.other.devices.display.protocols.value =
>>>>> spice/qxl,vnc/vga,vnc/qxl,vnc/cirrus
>>>>>
>>>>>>
>>>>>>
>>>>>> >
>>>>>> > Thanks.
>>>>>> >
>>>>>> > On Thu, Jun 22, 2017 at 12:19 PM, cmc
<iucounu(a)gmail.com> wrote:
>>>>>> >> Hi Martin,
>>>>>> >>
>>>>>> >>>
>>>>>> >>> just as a random comment, do you still have the
database backup from
>>>>>> >>> the bare metal -> VM attempt? It might be
possible to just try again
>>>>>> >>> using it. Or in the worst case.. update the
offending value there
>>>>>> >>> before restoring it to the new engine instance.
>>>>>> >>
>>>>>> >> I still have the backup. I'd rather do the latter,
as re-running the
>>>>>> >> HE deployment is quite lengthy and involved (I have to
re-initialise
>>>>>> >> the FC storage each time). Do you know what the
offending value(s)
>>>>>> >> would be? Would it be in the Postgres DB or in a config
file
>>>>>> >> somewhere?
>>>>>> >>
>>>>>> >> Cheers,
>>>>>> >>
>>>>>> >> Cam
>>>>>> >>
>>>>>> >>> Regards
>>>>>> >>>
>>>>>> >>> Martin Sivak
>>>>>> >>>
>>>>>> >>> On Thu, Jun 22, 2017 at 11:39 AM, cmc
<iucounu(a)gmail.com> wrote:
>>>>>> >>>> Hi Yanir,
>>>>>> >>>>
>>>>>> >>>> Thanks for the reply.
>>>>>> >>>>
>>>>>> >>>>> First of all, maybe a chain reaction of :
>>>>>> >>>>> WARN
[org.ovirt.engine.core.bll.exportimport.ImportVmCommand]
>>>>>> >>>>> (org.ovirt.thread.pool-6-thread-23) []
Validation of action
>>>>>> >>>>> 'ImportVm'
>>>>>> >>>>> failed for user SYSTEM. Reasons:
VAR__ACTION__IMPORT
>>>>>> >>>>>
>>>>>> >>>>>
,VAR__TYPE__VM,ACTION_TYPE_FAILED_ILLEGAL_VM_DISPLAY_TYPE_IS_NOT_SUPPORTED_BY_OS
>>>>>> >>>>> is causing the hosted engine vm not to be
set up correctly and
>>>>>> >>>>> further
>>>>>> >>>>> actions were made when the hosted engine vm
wasnt in a stable state.
>>>>>> >>>>>
>>>>>> >>>>> As for now, are you trying to revert back to
a previous/initial
>>>>>> >>>>> state ?
>>>>>> >>>>
>>>>>> >>>> I'm not trying to revert it to a previous
state for now. This was a
>>>>>> >>>> migration from a bare metal engine, and it
didn't report any error
>>>>>> >>>> during the migration. I'd had some problems
on my first attempts at
>>>>>> >>>> this migration, whereby it never completed (due
to a proxy issue) but
>>>>>> >>>> I managed to resolve this. Do you know of a way
to get the Hosted
>>>>>> >>>> Engine VM into a stable state, without
rebuilding the entire cluster
>>>>>> >>>> from scratch (since I have a lot of VMs on it)?
>>>>>> >>>>
>>>>>> >>>> Thanks for any help.
>>>>>> >>>>
>>>>>> >>>> Regards,
>>>>>> >>>>
>>>>>> >>>> Cam
>>>>>> >>>>
>>>>>> >>>>> Regards,
>>>>>> >>>>> Yanir
>>>>>> >>>>>
>>>>>> >>>>> On Wed, Jun 21, 2017 at 4:32 PM, cmc
<iucounu(a)gmail.com> wrote:
>>>>>> >>>>>>
>>>>>> >>>>>> Hi Jenny/Martin,
>>>>>> >>>>>>
>>>>>> >>>>>> Any idea what I can do here? The hosted
engine VM has no log on any
>>>>>> >>>>>> host in /var/log/libvirt/qemu, and I
fear that if I need to put the
>>>>>> >>>>>> host into maintenance, e.g., to upgrade
it that I created it on
>>>>>> >>>>>> (which
>>>>>> >>>>>> I think is hosting it), or if it fails
for any reason, it won't get
>>>>>> >>>>>> migrated to another host, and I will not
be able to manage the
>>>>>> >>>>>> cluster. It seems to be a very dangerous
position to be in.
>>>>>> >>>>>>
>>>>>> >>>>>> Thanks,
>>>>>> >>>>>>
>>>>>> >>>>>> Cam
>>>>>> >>>>>>
>>>>>> >>>>>> On Wed, Jun 21, 2017 at 11:48 AM, cmc
<iucounu(a)gmail.com> wrote:
>>>>>> >>>>>>> Thanks Martin. The hosts are all
part of the same cluster.
>>>>>> >>>>>>>
>>>>>> >>>>>>> I get these errors in the engine.log
on the engine:
>>>>>> >>>>>>>
>>>>>> >>>>>>> 2017-06-19 03:28:05,030Z WARN
>>>>>> >>>>>>>
[org.ovirt.engine.core.bll.exportimport.ImportVmCommand]
>>>>>> >>>>>>> (org.ovirt.thread.pool-6-thread-23)
[] Validation of action
>>>>>> >>>>>>> 'ImportVm'
>>>>>> >>>>>>> failed for user SYST
>>>>>> >>>>>>> EM. Reasons:
>>>>>> >>>>>>>
>>>>>> >>>>>>>
VAR__ACTION__IMPORT,VAR__TYPE__VM,ACTION_TYPE_FAILED_ILLEGAL_VM_DISPLAY_TYPE_IS_NOT_SUPPORTED_BY_OS
>>>>>> >>>>>>> 2017-06-19 03:28:05,030Z INFO
>>>>>> >>>>>>>
[org.ovirt.engine.core.bll.exportimport.ImportVmCommand]
>>>>>> >>>>>>> (org.ovirt.thread.pool-6-thread-23)
[] Lock freed to object
>>>>>> >>>>>>>
'EngineLock:{exclusiveLocks='[a
>>>>>> >>>>>>>
79e6b0e-fff4-4cba-a02c-4c00be151300=<VM,
>>>>>> >>>>>>>
ACTION_TYPE_FAILED_VM_IS_BEING_IMPORTED$VmName HostedEngine>,
>>>>>> >>>>>>> HostedEngine=<VM_NAME,
ACTION_TYPE_FAILED_NAME_ALREADY_USED>]',
>>>>>> >>>>>>> sharedLocks=
>>>>>> >>>>>>>
'[a79e6b0e-fff4-4cba-a02c-4c00be151300=<REMOTE_VM,
>>>>>> >>>>>>>
ACTION_TYPE_FAILED_VM_IS_BEING_IMPORTED$VmName HostedEngine>]'}'
>>>>>> >>>>>>> 2017-06-19 03:28:05,030Z ERROR
>>>>>> >>>>>>>
[org.ovirt.engine.core.bll.HostedEngineImporter]
>>>>>> >>>>>>> (org.ovirt.thread.pool-6-thread-23)
[] Failed importing the Hosted
>>>>>> >>>>>>> Engine VM
>>>>>> >>>>>>>
>>>>>> >>>>>>> The sanlock.log reports conflicts on
that same host, and a
>>>>>> >>>>>>> different
>>>>>> >>>>>>> error on the other hosts, not sure
if they are related.
>>>>>> >>>>>>>
>>>>>> >>>>>>> And this in the
/var/log/ovirt-hosted-engine-ha/agent log on the
>>>>>> >>>>>>> host
>>>>>> >>>>>>> which I deployed the hosted engine
VM on:
>>>>>> >>>>>>>
>>>>>> >>>>>>> MainThread::ERROR::2017-06-19
>>>>>> >>>>>>>
>>>>>> >>>>>>>
>>>>>> >>>>>>>
13:09:49,743::ovf_store::124::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(getEngineVMOVF)
>>>>>> >>>>>>> Unable to extract HEVM OVF
>>>>>> >>>>>>> MainThread::ERROR::2017-06-19
>>>>>> >>>>>>>
>>>>>> >>>>>>>
>>>>>> >>>>>>>
13:09:49,743::config::445::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(_get_vm_conf_content_from_ovf_store)
>>>>>> >>>>>>> Failed extracting VM OVF from the
OVF_STORE volume, falling back
>>>>>> >>>>>>> to
>>>>>> >>>>>>> initial vm.conf
>>>>>> >>>>>>>
>>>>>> >>>>>>> I've seen some of these issues
reported in bugzilla, but they were
>>>>>> >>>>>>> for
>>>>>> >>>>>>> older versions of oVirt (and appear
to be resolved).
>>>>>> >>>>>>>
>>>>>> >>>>>>> I will install that package on the
other two hosts, for which I
>>>>>> >>>>>>> will
>>>>>> >>>>>>> put them in maintenance as vdsm is
installed as an upgrade. I
>>>>>> >>>>>>> guess
>>>>>> >>>>>>> restarting vdsm is a good idea after
that?
>>>>>> >>>>>>>
>>>>>> >>>>>>> Thanks,
>>>>>> >>>>>>>
>>>>>> >>>>>>> Campbell
>>>>>> >>>>>>>
>>>>>> >>>>>>> On Wed, Jun 21, 2017 at 10:51 AM,
Martin Sivak <msivak(a)redhat.com>
>>>>>> >>>>>>> wrote:
>>>>>> >>>>>>>> Hi,
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> you do not have to install it on
all hosts. But you should have
>>>>>> >>>>>>>> more
>>>>>> >>>>>>>> than one and ideally all hosted
engine enabled nodes should
>>>>>> >>>>>>>> belong to
>>>>>> >>>>>>>> the same engine cluster.
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> Best regards
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> Martin Sivak
>>>>>> >>>>>>>>
>>>>>> >>>>>>>> On Wed, Jun 21, 2017 at 11:29
AM, cmc <iucounu(a)gmail.com> wrote:
>>>>>> >>>>>>>>> Hi Jenny,
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> Does ovirt-hosted-engine-ha
need to be installed across all
>>>>>> >>>>>>>>> hosts?
>>>>>> >>>>>>>>> Could that be the reason it
is failing to see it properly?
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> Thanks,
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> Cam
>>>>>> >>>>>>>>>
>>>>>> >>>>>>>>> On Mon, Jun 19, 2017 at 1:27
PM, cmc <iucounu(a)gmail.com> wrote:
>>>>>> >>>>>>>>>> Hi Jenny,
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> Logs are attached. I can
see errors in there, but am unsure how
>>>>>> >>>>>>>>>> they
>>>>>> >>>>>>>>>> arose.
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> Thanks,
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> Campbell
>>>>>> >>>>>>>>>>
>>>>>> >>>>>>>>>> On Mon, Jun 19, 2017 at
12:29 PM, Evgenia Tokar
>>>>>> >>>>>>>>>>
<etokar(a)redhat.com>
>>>>>> >>>>>>>>>> wrote:
>>>>>> >>>>>>>>>>> From the output it
looks like the agent is down, try starting
>>>>>> >>>>>>>>>>> it by
>>>>>> >>>>>>>>>>> running:
>>>>>> >>>>>>>>>>> systemctl start
ovirt-ha-agent.
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> The engine is
supposed to see the hosted engine storage domain
>>>>>> >>>>>>>>>>> and
>>>>>> >>>>>>>>>>> import it
>>>>>> >>>>>>>>>>> to the system, then
it should import the hosted engine vm.
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> Can you attach the
agent log from the host
>>>>>> >>>>>>>>>>>
(/var/log/ovirt-hosted-engine-ha/agent.log)
>>>>>> >>>>>>>>>>> and the engine log
from the engine vm
>>>>>> >>>>>>>>>>>
(/var/log/ovirt-engine/engine.log)?
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> Thanks,
>>>>>> >>>>>>>>>>> Jenny
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>> On Mon, Jun 19, 2017
at 12:41 PM, cmc <iucounu(a)gmail.com>
>>>>>> >>>>>>>>>>> wrote:
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> Hi Jenny,
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> What version
are you running?
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
4.1.2.2-1.el7.centos
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> For the
hosted engine vm to be imported and displayed in the
>>>>>> >>>>>>>>>>>>> engine, you
>>>>>> >>>>>>>>>>>>> must first
create a master storage domain.
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> To provide a bit
more detail: this was a migration of a
>>>>>> >>>>>>>>>>>> bare-metal
>>>>>> >>>>>>>>>>>> engine in an
existing cluster to a hosted engine VM for that
>>>>>> >>>>>>>>>>>> cluster.
>>>>>> >>>>>>>>>>>> As part of this
migration, I built an entirely new host and
>>>>>> >>>>>>>>>>>> ran
>>>>>> >>>>>>>>>>>>
'hosted-engine --deploy' (followed these instructions:
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
http://www.ovirt.org/documentation/self-hosted/chap-Migrating_from_Bare_M...).
>>>>>> >>>>>>>>>>>> I restored the
backup from the engine and it completed
>>>>>> >>>>>>>>>>>> without any
>>>>>> >>>>>>>>>>>> errors. I
didn't see any instructions regarding a master
>>>>>> >>>>>>>>>>>> storage
>>>>>> >>>>>>>>>>>> domain in the
page above. The cluster has two existing master
>>>>>> >>>>>>>>>>>> storage
>>>>>> >>>>>>>>>>>> domains, one is
fibre channel, which is up, and one ISO
>>>>>> >>>>>>>>>>>> domain,
>>>>>> >>>>>>>>>>>> which
>>>>>> >>>>>>>>>>>> is currently
offline.
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> What do you
mean the hosted engine commands are failing?
>>>>>> >>>>>>>>>>>>> What
>>>>>> >>>>>>>>>>>>> happens
>>>>>> >>>>>>>>>>>>> when
>>>>>> >>>>>>>>>>>>> you run
hosted-engine --vm-status now?
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> Interestingly,
whereas when I ran it before, it exited with
>>>>>> >>>>>>>>>>>> no
>>>>>> >>>>>>>>>>>> output
>>>>>> >>>>>>>>>>>> and a return
code of '1', it now reports:
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> --== Host 1
status ==--
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>
conf_on_shared_storage : True
>>>>>> >>>>>>>>>>>> Status
up-to-date : False
>>>>>> >>>>>>>>>>>> Hostname
:
>>>>>> >>>>>>>>>>>>
kvm-ldn-03.ldn.fscfc.co.uk
>>>>>> >>>>>>>>>>>> Host ID
: 1
>>>>>> >>>>>>>>>>>> Engine status
: unknown stale-data
>>>>>> >>>>>>>>>>>> Score
: 0
>>>>>> >>>>>>>>>>>> stopped
: True
>>>>>> >>>>>>>>>>>> Local
maintenance : False
>>>>>> >>>>>>>>>>>> crc32
: 0217f07b
>>>>>> >>>>>>>>>>>>
local_conf_timestamp : 2911
>>>>>> >>>>>>>>>>>> Host timestamp
: 2897
>>>>>> >>>>>>>>>>>> Extra metadata
(valid at timestamp):
>>>>>> >>>>>>>>>>>>
metadata_parse_version=1
>>>>>> >>>>>>>>>>>>
metadata_feature_version=1
>>>>>> >>>>>>>>>>>>
timestamp=2897 (Thu Jun 15 16:22:54 2017)
>>>>>> >>>>>>>>>>>>
host-id=1
>>>>>> >>>>>>>>>>>> score=0
>>>>>> >>>>>>>>>>>>
vm_conf_refresh_time=2911 (Thu Jun 15 16:23:08 2017)
>>>>>> >>>>>>>>>>>>
conf_on_shared_storage=True
>>>>>> >>>>>>>>>>>>
maintenance=False
>>>>>> >>>>>>>>>>>>
state=AgentStopped
>>>>>> >>>>>>>>>>>>
stopped=True
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> Yet I can login
to the web GUI fine. I guess it is not HA due
>>>>>> >>>>>>>>>>>> to
>>>>>> >>>>>>>>>>>> being
>>>>>> >>>>>>>>>>>> in an unknown
state currently? Does the hosted-engine-ha rpm
>>>>>> >>>>>>>>>>>> need
>>>>>> >>>>>>>>>>>> to
>>>>>> >>>>>>>>>>>> be installed
across all nodes in the cluster, btw?
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> Thanks for the
help,
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>> Cam
>>>>>> >>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> Jenny Tokar
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>> On Thu, Jun
15, 2017 at 6:32 PM, cmc <iucounu(a)gmail.com>
>>>>>> >>>>>>>>>>>>> wrote:
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> Hi,
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> I've
migrated from a bare-metal engine to a hosted engine.
>>>>>> >>>>>>>>>>>>>> There
>>>>>> >>>>>>>>>>>>>> were
>>>>>> >>>>>>>>>>>>>> no
errors during the install, however, the hosted engine
>>>>>> >>>>>>>>>>>>>> did not
>>>>>> >>>>>>>>>>>>>> get
>>>>>> >>>>>>>>>>>>>> started.
I tried running:
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>>
hosted-engine --status
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> on the
host I deployed it on, and it returns nothing (exit
>>>>>> >>>>>>>>>>>>>> code
>>>>>> >>>>>>>>>>>>>> is 1
>>>>>> >>>>>>>>>>>>>>
however). I could not ping it either. So I tried starting
>>>>>> >>>>>>>>>>>>>> it via
>>>>>> >>>>>>>>>>>>>>
'hosted-engine --vm-start' and it returned:
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> Virtual
machine does not exist
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> But it
then became available. I logged into it
>>>>>> >>>>>>>>>>>>>>
successfully. It
>>>>>> >>>>>>>>>>>>>> is not
>>>>>> >>>>>>>>>>>>>> in the
list of VMs however.
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> Any
ideas why the hosted-engine commands fail, and why it
>>>>>> >>>>>>>>>>>>>> is not
>>>>>> >>>>>>>>>>>>>> in
>>>>>> >>>>>>>>>>>>>> the list
of virtual machines?
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> Thanks
for any help,
>>>>>> >>>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>> Cam
>>>>>> >>>>>>>>>>>>>>
_______________________________________________
>>>>>> >>>>>>>>>>>>>> Users
mailing list
>>>>>> >>>>>>>>>>>>>>
Users(a)ovirt.org
>>>>>> >>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>>>
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>>>
>>>>>> >>>>>>>>>
_______________________________________________
>>>>>> >>>>>>>>> Users mailing list
>>>>>> >>>>>>>>> Users(a)ovirt.org
>>>>>> >>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >>>>>>
_______________________________________________
>>>>>> >>>>>> Users mailing list
>>>>>> >>>>>> Users(a)ovirt.org
>>>>>> >>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >>>>>
>>>>>> >>>>>
>>>>>> > _______________________________________________
>>>>>> > Users mailing list
>>>>>> > Users(a)ovirt.org
>>>>>> >
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >
>>>>>> >
>>>>>>
>>>>>