On Wed, Dec 2, 2015 at 12:19 PM, Budur Nagaraju <nbudoor(a)gmail.com> wrote:
I have installed KVM in the nested environment in ESXi6.x version is
that
recommended ?
I often use KVM over KVM in nested environment but honestly I never tried
to run KVM over ESXi but I suspect that all of your issues comes from there.
apart from Hosted engine is there any other alternate way to
configure
Engine HA cluster ?
Nothing else from the project. You can use two external VMs in cluster with
pacemaker but it's completely up to you.
-Nagaraju
On Wed, Dec 2, 2015 at 4:11 PM, Simone Tiraboschi <stirabos(a)redhat.com>
wrote:
>
>
> On Wed, Dec 2, 2015 at 11:25 AM, Budur Nagaraju <nbudoor(a)gmail.com>
> wrote:
>
>> pls fine the logs from the below mentioned URL,
>>
>>
http://pastebin.com/ZeKyyFbN
>>
>
> OK, the issue is here:
>
> Thread-88::ERROR::2015-12-02
> 15:06:27,735::vm::2358::vm.Vm::(_startUnderlyingVm)
> vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::The vm start process failed
> Traceback (most recent call last):
> File "/usr/share/vdsm/virt/vm.py", line 2298, in _startUnderlyingVm
> self._run()
> File "/usr/share/vdsm/virt/vm.py", line 3363, in _run
> self._connection.createXML(domxml, flags),
> File "/usr/lib/python2.6/site-packages/vdsm/libvirtconnection.py", line
> 119, in wrapper
> ret = f(*args, **kwargs)
> File "/usr/lib64/python2.6/site-packages/libvirt.py", line 2709, in
> createXML
> if ret is None:raise libvirtError('virDomainCreateXML() failed',
> conn=self)
> libvirtError: unsupported configuration: Domain requires KVM, but it is
> not available. Check that virtualization is enabled in the host BIOS, and
> host configuration is setup to load the kvm modules.
> Thread-88::DEBUG::2015-12-02
> 15:06:27,751::vm::2813::vm.Vm::(setDownStatus)
> vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::Changed state to Down:
> unsupported configuration: Domain requires KVM, but it is not available.
> Check that virtualization is enabled in the host BIOS, and host
> configuration is setup to load the kvm modules. (code=1)
>
> but it's pretty strange cause hosted-engine-setup already explicitly
> check for visualization support and just exits with a clear error if not.
> Did you played with the kvm module while hosted-engine-setup was running?
>
> Can you please hosted-engine-setup logs?
>
>
>>
>> On Fri, Nov 27, 2015 at 6:39 PM, Simone Tiraboschi <stirabos(a)redhat.com>
>> wrote:
>>
>>>
>>>
>>> On Fri, Nov 27, 2015 at 12:42 PM, Maxim Kovgan <kovganm(a)gmail.com>
>>> wrote:
>>>
>>>> Maybe even makes sense to open a bugzilla ticket already. Better safe
>>>> than sorry.
>>>>
>>>
>>> We still need at least one log file to understand what happened.
>>>
>>>
>>>> On Nov 27, 2015 11:35 AM, "Simone Tiraboschi"
<stirabos(a)redhat.com>
>>>> wrote:
>>>>
>>>>>
>>>>> On Fri, Nov 27, 2015 at 10:10 AM, Budur Nagaraju
<nbudoor(a)gmail.com>
>>>>> wrote:
>>>>>
>>>>>> I do not know what logs you are expecting ? the logs which I got
is
>>>>>> pasted in the mail if you require in pastebin let me know I will
upload
>>>>>> there .
>>>>>>
>>>>>
>>>>>
>>>>> Please run sosreport utility and share the resulting archive where
>>>>> you prefer.
>>>>> You can follow this guide:
>>>>>
http://www.linuxtechi.com/how-to-create-sosreport-in-linux/
>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Nov 27, 2015 at 1:58 PM, Sandro Bonazzola <
>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Nov 27, 2015 at 8:34 AM, Budur Nagaraju
<nbudoor(a)gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I got only 10lines to in the vdsm logs and are below ,
>>>>>>>>
>>>>>>>>
>>>>>>> Can you please provide full sos report?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> [root@he /]# tail -f /var/log/vdsm/vdsm.log
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,360::resourceManager::616::Storage.ResourceManager::(releaseResource)
>>>>>>>> Trying to release resource
'Storage.HsmDomainMonitorLock'
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,360::resourceManager::635::Storage.ResourceManager::(releaseResource)
>>>>>>>> Released resource 'Storage.HsmDomainMonitorLock'
(0 active users)
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,360::resourceManager::641::Storage.ResourceManager::(releaseResource)
>>>>>>>> Resource 'Storage.HsmDomainMonitorLock' is free,
finding out if anyone is
>>>>>>>> waiting for it.
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,360::resourceManager::649::Storage.ResourceManager::(releaseResource)
>>>>>>>> No one is waiting for resource
'Storage.HsmDomainMonitorLock', Clearing
>>>>>>>> records.
>>>>>>>> Thread-100::INFO::2015-11-27
>>>>>>>> 12:58:57,360::logUtils::47::dispatcher::(wrapper) Run and
protect:
>>>>>>>> stopMonitoringDomain, Return response: None
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,361::task::1191::Storage.TaskManager.Task::(prepare)
>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::finished:
None
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,361::task::595::Storage.TaskManager.Task::(_updateState)
>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::moving from
state preparing ->
>>>>>>>> state finished
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,361::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>>>>>>> Owner.releaseAll requests {} resources {}
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,361::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>>>>>>> Owner.cancelAll requests {}
>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>
12:58:57,361::task::993::Storage.TaskManager.Task::(_decref)
>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::ref 0
aborting False
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Thu, Nov 26, 2015 at 4:20 PM, Simone Tiraboschi <
>>>>>>>> stirabos(a)redhat.com> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Thu, Nov 26, 2015 at 11:05 AM, Budur Nagaraju
<
>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *Below are the entire logs*
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>> Sorry, with the entire log I mean if you can attach
or share
>>>>>>>>> somewhere the whole /var/log/vdsm/vdsm.log cause the
latest ten lines are
>>>>>>>>> not enough to point out the issue.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/vdsm.log *
>>>>>>>>>>
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>> Detected protocol xml from 127.0.0.1:50944
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>> http detected from ('127.0.0.1', 50944)
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>> Adding connection from 127.0.0.1:50945
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>> Connection removed from 127.0.0.1:50945
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>> Detected protocol xml from 127.0.0.1:50945
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>> http detected from ('127.0.0.1', 50945)
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>> Adding connection from 127.0.0.1:50946
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>> Connection removed from 127.0.0.1:50946
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>> Detected protocol xml from 127.0.0.1:50946
>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>
15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>> http detected from ('127.0.0.1', 50946)
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/supervdsm.log
*
>>>>>>>>>>
>>>>>>>>>> MainProcess::DEBUG::2015-11-26
>>>>>>>>>>
15:13:30,234::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> call readMultipathConf with () {}
>>>>>>>>>> MainProcess::DEBUG::2015-11-26
>>>>>>>>>>
15:13:30,234::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> return readMultipathConf with ['# RHEV
REVISION 1.1', '', 'defaults {',
>>>>>>>>>> ' polling_interval 5', '
getuid_callout
>>>>>>>>>> "/lib/udev/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n"',
>>>>>>>>>> ' no_path_retry fail', '
user_friendly_names no', '
>>>>>>>>>> flush_on_last_del yes', '
fast_io_fail_tmo 5', '
>>>>>>>>>> dev_loss_tmo 30', ' max_fds
4096', '}', '',
>>>>>>>>>> 'devices {', 'device {', '
vendor "HITACHI"', '
>>>>>>>>>> product "DF.*"',
' getuid_callout
>>>>>>>>>> "/lib/udev/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n"',
>>>>>>>>>> '}', 'device {', ' vendor
"COMPELNT"', '
>>>>>>>>>> product "Compellent
Vol"', ' no_path_retry
>>>>>>>>>> fail', '}', 'device {', '
# multipath.conf.default', '
>>>>>>>>>> vendor "DGC"',
' product ".*"', '
>>>>>>>>>> product_blacklist "LUNZ"',
' path_grouping_policy
>>>>>>>>>> "group_by_prio"', '
path_checker "emc_clariion"', '
>>>>>>>>>> hardware_handler "1 emc"',
' prio "emc"', '
>>>>>>>>>> failback immediate', '
rr_weight
>>>>>>>>>> "uniform"', ' # vdsm
required configuration', '
>>>>>>>>>> getuid_callout "/lib/udev/scsi_id
--whitelisted
>>>>>>>>>> --replace-whitespace --device=/dev/%n"',
' features "0"',
>>>>>>>>>> ' no_path_retry fail',
'}', '}']
>>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26
>>>>>>>>>>
15:13:31,365::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> call getHardwareInfo with () {}
>>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26
>>>>>>>>>>
15:13:31,397::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> return getHardwareInfo with
{'systemProductName': 'KVM', 'systemUUID':
>>>>>>>>>> 'f91632f2-7a17-4ddb-9631-742f82a77480',
'systemFamily': 'Red Hat Enterprise
>>>>>>>>>> Linux', 'systemVersion': 'RHEL
7.0.0 PC (i440FX + PIIX, 1996)',
>>>>>>>>>> 'systemManufacturer': 'Red Hat'}
>>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26
>>>>>>>>>>
15:13:35,393::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> call validateAccess with ('qemu',
('qemu', 'kvm'),
>>>>>>>>>>
'/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {}
>>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26
>>>>>>>>>>
15:13:35,395::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> return validateAccess with None
>>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26
>>>>>>>>>>
15:13:36,067::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> call validateAccess with ('qemu',
('qemu', 'kvm'),
>>>>>>>>>>
'/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {}
>>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26
>>>>>>>>>>
15:13:36,069::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> return validateAccess with None
>>>>>>>>>> MainProcess|PolicyEngine::DEBUG::2015-11-26
>>>>>>>>>>
15:13:40,619::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> call ksmTune with ({'run': 0},) {}
>>>>>>>>>> MainProcess|PolicyEngine::DEBUG::2015-11-26
>>>>>>>>>>
15:13:40,619::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>> return ksmTune with None
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *[root@he ~]# tail -f
/var/log/vdsm/connectivity.log *
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2015-11-26
15:02:02,632:DEBUG:recent_client:False
>>>>>>>>>> 2015-11-26 15:04:44,975:DEBUG:recent_client:True
>>>>>>>>>> 2015-11-26
15:05:15,039:DEBUG:recent_client:False
>>>>>>>>>> 2015-11-26 15:07:23,311:DEBUG:recent_client:True
>>>>>>>>>> 2015-11-26
15:08:25,774:DEBUG:recent_client:True,
>>>>>>>>>> ovirtmgmt:(operstate:up speed:0 duplex:unknown),
lo:(operstate:up speed:0
>>>>>>>>>> duplex:unknown), ;vdsmdummy;:(operstate:down
speed:0 duplex:unknown),
>>>>>>>>>> bond0:(operstate:down speed:0 duplex:unknown),
eth0:(operstate:up speed:100
>>>>>>>>>> duplex:full)
>>>>>>>>>> 2015-11-26
15:08:55,845:DEBUG:recent_client:False
>>>>>>>>>> 2015-11-26 15:08:59,859:DEBUG:recent_client:True
>>>>>>>>>> 2015-11-26
15:09:29,929:DEBUG:recent_client:False
>>>>>>>>>> 2015-11-26
15:13:32,292:DEBUG:recent_client:True,
>>>>>>>>>> ovirtmgmt:(operstate:up speed:0 duplex:unknown),
lo:(operstate:up speed:0
>>>>>>>>>> duplex:unknown), ;vdsmdummy;:(operstate:down
speed:0 duplex:unknown),
>>>>>>>>>> bond0:(operstate:down speed:0 duplex:unknown),
eth0:(operstate:up speed:100
>>>>>>>>>> duplex:full)
>>>>>>>>>> 2015-11-26
15:14:02,363:DEBUG:recent_client:False
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/mom.log *
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2015-11-26 15:13:30,581 - mom.Policy - INFO -
Loaded policy
>>>>>>>>>> '04-cputune'
>>>>>>>>>> 2015-11-26 15:13:30,581 - mom.PolicyEngine - INFO
- Policy
>>>>>>>>>> Engine starting
>>>>>>>>>> 2015-11-26 15:13:30,582 - mom.RPCServer - INFO -
RPC Server is
>>>>>>>>>> disabled
>>>>>>>>>> 2015-11-26 15:13:40,618 - mom.Controllers.KSM -
INFO - Updating
>>>>>>>>>> KSM configuration: pages_to_scan:0
merge_across_nodes:8 run:0
>>>>>>>>>> sleep_millisecs:0
>>>>>>>>>> 2015-11-26 15:14:51,492 - mom.RPCServer - INFO -
getStatistics()
>>>>>>>>>> 2015-11-26 15:14:56,962 - mom.RPCServer - INFO -
getStatistics()
>>>>>>>>>> 2015-11-26 15:15:02,451 - mom.RPCServer - INFO -
getStatistics()
>>>>>>>>>> 2015-11-26 15:15:07,777 - mom.RPCServer - INFO -
getStatistics()
>>>>>>>>>> 2015-11-26 15:15:13,267 - mom.RPCServer - INFO -
getStatistics()
>>>>>>>>>> 2015-11-26 15:15:18,765 - mom.RPCServer - INFO -
getStatistics()
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Thu, Nov 26, 2015 at 3:28 PM, Budur Nagaraju
<
>>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Below are the logs,
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> [root@he ~]# tail -f /var/log/vdsm/vdsm.log
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>> Detected protocol xml from 127.0.0.1:50944
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>> http detected from ('127.0.0.1',
50944)
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>> Adding connection from 127.0.0.1:50945
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>> Connection removed from 127.0.0.1:50945
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>> Detected protocol xml from 127.0.0.1:50945
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>> http detected from ('127.0.0.1',
50945)
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>> Adding connection from 127.0.0.1:50946
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>> Connection removed from 127.0.0.1:50946
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>> Detected protocol xml from 127.0.0.1:50946
>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>
15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>> http detected from ('127.0.0.1',
50946)
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Nov 26, 2015 at 3:06 PM, Simone
Tiraboschi <
>>>>>>>>>>> stirabos(a)redhat.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Nov 26, 2015 at 10:33 AM, Budur
Nagaraju <
>>>>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> I have done a fresh installation and
now am getting the below
>>>>>>>>>>>>> error,
>>>>>>>>>>>>>
>>>>>>>>>>>>> [ INFO ] Updating hosted-engine
configuration
>>>>>>>>>>>>> [ INFO ] Stage: Transaction commit
>>>>>>>>>>>>> [ INFO ] Stage: Closing up
>>>>>>>>>>>>> The following network ports
should be opened:
>>>>>>>>>>>>> tcp:5900
>>>>>>>>>>>>> tcp:5901
>>>>>>>>>>>>> udp:5900
>>>>>>>>>>>>> udp:5901
>>>>>>>>>>>>> An example of the required
configuration for
>>>>>>>>>>>>> iptables can be found at:
>>>>>>>>>>>>>
/etc/ovirt-hosted-engine/iptables.example
>>>>>>>>>>>>> In order to configure
firewalld, copy the files from
>>>>>>>>>>>>>
/etc/ovirt-hosted-engine/firewalld to
>>>>>>>>>>>>> /etc/firewalld/services
>>>>>>>>>>>>> and execute the following
commands:
>>>>>>>>>>>>> firewall-cmd -service
hosted-console
>>>>>>>>>>>>> [ INFO ] Creating VM
>>>>>>>>>>>>> [ ERROR ] Failed to execute stage
'Closing up': Cannot set
>>>>>>>>>>>>> temporary password for console
connection. The VM may not have been
>>>>>>>>>>>>> created: please check VDSM logs
>>>>>>>>>>>>> [ INFO ] Stage: Clean up
>>>>>>>>>>>>> [ INFO ] Generating answer file
>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126145701.conf'
>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination
>>>>>>>>>>>>> [ INFO ] Stage: Termination
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> [root@he ovirt]# tail -f
/var/log/vdsm/
>>>>>>>>>>>>> backup/ connectivity.log
mom.log
>>>>>>>>>>>>> supervdsm.log vdsm.log
>>>>>>>>>>>>> [root@he ovirt]# tail -f
/var/log/vdsm/vdsm.log
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:07,564::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:42741
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:07,564::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>> http detected from
('127.0.0.1', 42741)
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:07,644::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>>> Adding connection from
127.0.0.1:42742
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,088::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>>> Connection removed from
127.0.0.1:42742
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,088::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:42742
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,088::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>> http detected from
('127.0.0.1', 42742)
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,171::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>>> Adding connection from
127.0.0.1:42743
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,572::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>>> Connection removed from
127.0.0.1:42743
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,573::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:42743
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
14:57:08,573::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>> http detected from
('127.0.0.1', 42743)
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> It failed before, can you please attach
the whole VDSM logs?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 2:01 PM,
Simone Tiraboschi <
>>>>>>>>>>>>> stirabos(a)redhat.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 7:30 AM,
Budur Nagaraju <
>>>>>>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Its a fresh setup ,I have
deleted all the vms ,still am
>>>>>>>>>>>>>>> facing same issues .
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Can you please paste the output
of
>>>>>>>>>>>>>> vdsClient -s 0 list
>>>>>>>>>>>>>> ?
>>>>>>>>>>>>>> thanks
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 11:56
AM, Oved Ourfali <
>>>>>>>>>>>>>>> oourfali(a)redhat.com>
wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Seems like you have
existing VMs running on the host (you
>>>>>>>>>>>>>>>> can check that by looking
for qemu processes on your host).
>>>>>>>>>>>>>>>> Is that a clean
deployment, or was the host used before
>>>>>>>>>>>>>>>> for running VMs?
>>>>>>>>>>>>>>>> Perhaps you already ran
the hosted engine setup, and the
>>>>>>>>>>>>>>>> VM was left there?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> CC-ing Sandro who is more
familiar in that than me.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>> Oved
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at
7:07 AM, Budur Nagaraju <
>>>>>>>>>>>>>>>> nbudoor(a)gmail.com>
wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> HI
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Getting below error
while configuring Hosted engine,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> root@he ~]#
hosted-engine --deploy
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Initializing
>>>>>>>>>>>>>>>>> [ INFO ] Generating
a temporary VNC password.
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Environment setup
>>>>>>>>>>>>>>>>> Continuing
will configure this host for serving
>>>>>>>>>>>>>>>>> as hypervisor and
create a VM where you have to install oVirt Engine
>>>>>>>>>>>>>>>>> afterwards.
>>>>>>>>>>>>>>>>> Are you
sure you want to continue? (Yes,
>>>>>>>>>>>>>>>>> No)[Yes]: yes
>>>>>>>>>>>>>>>>>
Configuration files: []
>>>>>>>>>>>>>>>>> Log file:
>>>>>>>>>>>>>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20151126102302-bkozgk.log
>>>>>>>>>>>>>>>>> Version:
otopi-1.3.2 (otopi-1.3.2-1.el6)
>>>>>>>>>>>>>>>>> It has been
detected that this program is
>>>>>>>>>>>>>>>>> executed through an
SSH connection without using screen.
>>>>>>>>>>>>>>>>> Continuing
with the installation may lead to
>>>>>>>>>>>>>>>>> broken installation
if the network connection fails.
>>>>>>>>>>>>>>>>> It is
highly recommended to abort the
>>>>>>>>>>>>>>>>> installation and run
it inside a screen session using command "screen".
>>>>>>>>>>>>>>>>> Do you want
to continue anyway? (Yes, No)[No]:
>>>>>>>>>>>>>>>>> yes
>>>>>>>>>>>>>>>>> [WARNING] Cannot
detect if hardware supports
>>>>>>>>>>>>>>>>> virtualization
>>>>>>>>>>>>>>>>> [ INFO ] Bridge
ovirtmgmt already created
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Environment packages setup
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Programs detection
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Environment setup
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *[ ERROR ] The
following VMs has been found:
>>>>>>>>>>>>>>>>>
2b8d6d91-d838-44f6-ae3b-c92cda014280[ ERROR ] Failed to execute stage
>>>>>>>>>>>>>>>>> 'Environment
setup': Cannot setup Hosted Engine with other VMs running*
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Clean up
>>>>>>>>>>>>>>>>> [ INFO ] Generating
answer file
>>>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126102310.conf'
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Pre-termination
>>>>>>>>>>>>>>>>> [ INFO ] Stage:
Termination
>>>>>>>>>>>>>>>>> [root@he ~]#
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>> Users mailing list
>>>>>>>>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>> Users mailing list
>>>>>>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list
>>>>>>>> Users(a)ovirt.org
>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Sandro Bonazzola
>>>>>>> Better technology. Faster innovation. Powered by community
>>>>>>> collaboration.
>>>>>>> See how it works at
redhat.com
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users(a)ovirt.org
>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>>
>>>
>>
>