On Mon, Jan 4, 2016 at 3:06 PM, Budur Nagaraju <nbudoor(a)gmail.com> wrote:
The issue is here:
Thread-84::ERROR::2016-01-04
19:31:42,304::vm::2358::vm.Vm::(_startUnderlyingVm)
vmId=`3d3edc54-ceae-43e5-84a4-50a21c31d9cd`::The vm start process failed
Traceback (most recent call last):
File "/usr/share/vdsm/virt/vm.py", line 2298, in _startUnderlyingVm
self._run()
File "/usr/share/vdsm/virt/vm.py", line 3363, in _run
self._connection.createXML(domxml, flags),
File "/usr/lib/python2.6/site-packages/vdsm/libvirtconnection.py", line
119, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib64/python2.6/site-packages/libvirt.py", line 2709, in
createXML
if ret is None:raise libvirtError('virDomainCreateXML() failed',
conn=self)
libvirtError: unsupported configuration: Domain requires KVM, but it is not
available. Check that virtualization is enabled in the host BIOS, and host
configuration is setup to load the kvm modules.
libvirt refuses to start the engine VM cause KVM is not available.
Can you please check it?
Thanks,
Nagaraju
On Wed, Dec 2, 2015 at 5:35 PM, Simone Tiraboschi <stirabos(a)redhat.com>
wrote:
>
>
> On Wed, Dec 2, 2015 at 12:19 PM, Budur Nagaraju <nbudoor(a)gmail.com>
> wrote:
>
>> I have installed KVM in the nested environment in ESXi6.x version is
>> that recommended ?
>>
>
> I often use KVM over KVM in nested environment but honestly I never tried
> to run KVM over ESXi but I suspect that all of your issues comes from there.
>
>
>> apart from Hosted engine is there any other alternate way to configure
>> Engine HA cluster ?
>>
>
> Nothing else from the project. You can use two external VMs in cluster
> with pacemaker but it's completely up to you.
>
>
>>
>>
>> -Nagaraju
>>
>>
>> On Wed, Dec 2, 2015 at 4:11 PM, Simone Tiraboschi <stirabos(a)redhat.com>
>> wrote:
>>
>>>
>>>
>>> On Wed, Dec 2, 2015 at 11:25 AM, Budur Nagaraju <nbudoor(a)gmail.com>
>>> wrote:
>>>
>>>> pls fine the logs from the below mentioned URL,
>>>>
>>>>
http://pastebin.com/ZeKyyFbN
>>>>
>>>
>>> OK, the issue is here:
>>>
>>> Thread-88::ERROR::2015-12-02
>>> 15:06:27,735::vm::2358::vm.Vm::(_startUnderlyingVm)
>>> vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::The vm start process failed
>>> Traceback (most recent call last):
>>> File "/usr/share/vdsm/virt/vm.py", line 2298, in
_startUnderlyingVm
>>> self._run()
>>> File "/usr/share/vdsm/virt/vm.py", line 3363, in _run
>>> self._connection.createXML(domxml, flags),
>>> File
"/usr/lib/python2.6/site-packages/vdsm/libvirtconnection.py",
>>> line 119, in wrapper
>>> ret = f(*args, **kwargs)
>>> File "/usr/lib64/python2.6/site-packages/libvirt.py", line 2709,
in
>>> createXML
>>> if ret is None:raise libvirtError('virDomainCreateXML() failed',
>>> conn=self)
>>> libvirtError: unsupported configuration: Domain requires KVM, but it is
>>> not available. Check that virtualization is enabled in the host BIOS, and
>>> host configuration is setup to load the kvm modules.
>>> Thread-88::DEBUG::2015-12-02
>>> 15:06:27,751::vm::2813::vm.Vm::(setDownStatus)
>>> vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::Changed state to Down:
>>> unsupported configuration: Domain requires KVM, but it is not available.
>>> Check that virtualization is enabled in the host BIOS, and host
>>> configuration is setup to load the kvm modules. (code=1)
>>>
>>> but it's pretty strange cause hosted-engine-setup already explicitly
>>> check for visualization support and just exits with a clear error if not.
>>> Did you played with the kvm module while hosted-engine-setup was
>>> running?
>>>
>>> Can you please hosted-engine-setup logs?
>>>
>>>
>>>>
>>>> On Fri, Nov 27, 2015 at 6:39 PM, Simone Tiraboschi <
>>>> stirabos(a)redhat.com> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Fri, Nov 27, 2015 at 12:42 PM, Maxim Kovgan
<kovganm(a)gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Maybe even makes sense to open a bugzilla ticket already. Better
>>>>>> safe than sorry.
>>>>>>
>>>>>
>>>>> We still need at least one log file to understand what happened.
>>>>>
>>>>>
>>>>>> On Nov 27, 2015 11:35 AM, "Simone Tiraboschi"
<stirabos(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>> On Fri, Nov 27, 2015 at 10:10 AM, Budur Nagaraju
<nbudoor(a)gmail.com
>>>>>>> > wrote:
>>>>>>>
>>>>>>>> I do not know what logs you are expecting ? the logs
which I got
>>>>>>>> is pasted in the mail if you require in pastebin let me
know I will upload
>>>>>>>> there .
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Please run sosreport utility and share the resulting archive
where
>>>>>>> you prefer.
>>>>>>> You can follow this guide:
>>>>>>>
http://www.linuxtechi.com/how-to-create-sosreport-in-linux/
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Nov 27, 2015 at 1:58 PM, Sandro Bonazzola <
>>>>>>>> sbonazzo(a)redhat.com> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Nov 27, 2015 at 8:34 AM, Budur Nagaraju <
>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> I got only 10lines to in the vdsm logs and are
below ,
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>> Can you please provide full sos report?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> [root@he /]# tail -f /var/log/vdsm/vdsm.log
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,360::resourceManager::616::Storage.ResourceManager::(releaseResource)
>>>>>>>>>> Trying to release resource
'Storage.HsmDomainMonitorLock'
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,360::resourceManager::635::Storage.ResourceManager::(releaseResource)
>>>>>>>>>> Released resource
'Storage.HsmDomainMonitorLock' (0 active users)
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,360::resourceManager::641::Storage.ResourceManager::(releaseResource)
>>>>>>>>>> Resource 'Storage.HsmDomainMonitorLock'
is free, finding out if anyone is
>>>>>>>>>> waiting for it.
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,360::resourceManager::649::Storage.ResourceManager::(releaseResource)
>>>>>>>>>> No one is waiting for resource
'Storage.HsmDomainMonitorLock', Clearing
>>>>>>>>>> records.
>>>>>>>>>> Thread-100::INFO::2015-11-27
>>>>>>>>>> 12:58:57,360::logUtils::47::dispatcher::(wrapper)
Run and protect:
>>>>>>>>>> stopMonitoringDomain, Return response: None
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,361::task::1191::Storage.TaskManager.Task::(prepare)
>>>>>>>>>>
Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::finished: None
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,361::task::595::Storage.TaskManager.Task::(_updateState)
>>>>>>>>>>
Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::moving from state preparing ->
>>>>>>>>>> state finished
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,361::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>>>>>>>>> Owner.releaseAll requests {} resources {}
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,361::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>>>>>>>>> Owner.cancelAll requests {}
>>>>>>>>>> Thread-100::DEBUG::2015-11-27
>>>>>>>>>>
12:58:57,361::task::993::Storage.TaskManager.Task::(_decref)
>>>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::ref
0 aborting False
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Thu, Nov 26, 2015 at 4:20 PM, Simone
Tiraboschi <
>>>>>>>>>> stirabos(a)redhat.com> wrote:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Nov 26, 2015 at 11:05 AM, Budur
Nagaraju <
>>>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *Below are the entire logs*
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>> Sorry, with the entire log I mean if you can
attach or share
>>>>>>>>>>> somewhere the whole /var/log/vdsm/vdsm.log
cause the latest ten lines are
>>>>>>>>>>> not enough to point out the issue.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *[root@he ~]# tail -f
/var/log/vdsm/vdsm.log *
>>>>>>>>>>>>
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:50944
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>> http detected from ('127.0.0.1',
50944)
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>> Adding connection from 127.0.0.1:50945
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>> Connection removed from 127.0.0.1:50945
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:50945
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>> http detected from ('127.0.0.1',
50945)
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>> Adding connection from 127.0.0.1:50946
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>> Connection removed from 127.0.0.1:50946
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:50946
>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>
15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>> http detected from ('127.0.0.1',
50946)
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *[root@he ~]# tail -f
/var/log/vdsm/supervdsm.log *
>>>>>>>>>>>>
>>>>>>>>>>>> MainProcess::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:30,234::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> call readMultipathConf with () {}
>>>>>>>>>>>> MainProcess::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:30,234::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> return readMultipathConf with ['#
RHEV REVISION 1.1', '', 'defaults {',
>>>>>>>>>>>> ' polling_interval 5',
' getuid_callout
>>>>>>>>>>>> "/lib/udev/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n"',
>>>>>>>>>>>> ' no_path_retry
fail', ' user_friendly_names no', '
>>>>>>>>>>>> flush_on_last_del yes', '
fast_io_fail_tmo 5', '
>>>>>>>>>>>> dev_loss_tmo 30', '
max_fds 4096', '}', '',
>>>>>>>>>>>> 'devices {', 'device {',
' vendor "HITACHI"', '
>>>>>>>>>>>> product
"DF.*"', ' getuid_callout
>>>>>>>>>>>> "/lib/udev/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n"',
>>>>>>>>>>>> '}', 'device {', '
vendor "COMPELNT"', '
>>>>>>>>>>>> product "Compellent
Vol"', ' no_path_retry
>>>>>>>>>>>> fail', '}', 'device
{', ' # multipath.conf.default', '
>>>>>>>>>>>> vendor
"DGC"', ' product ".*"', '
>>>>>>>>>>>> product_blacklist
"LUNZ"', ' path_grouping_policy
>>>>>>>>>>>> "group_by_prio"', '
path_checker "emc_clariion"', '
>>>>>>>>>>>> hardware_handler "1
emc"', ' prio "emc"', '
>>>>>>>>>>>> failback immediate',
' rr_weight
>>>>>>>>>>>> "uniform"', ' # vdsm
required configuration', '
>>>>>>>>>>>> getuid_callout
"/lib/udev/scsi_id --whitelisted
>>>>>>>>>>>> --replace-whitespace
--device=/dev/%n"', ' features "0"',
>>>>>>>>>>>> ' no_path_retry
fail', '}', '}']
>>>>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:31,365::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> call getHardwareInfo with () {}
>>>>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:31,397::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> return getHardwareInfo with
{'systemProductName': 'KVM', 'systemUUID':
>>>>>>>>>>>>
'f91632f2-7a17-4ddb-9631-742f82a77480', 'systemFamily': 'Red Hat
Enterprise
>>>>>>>>>>>> Linux', 'systemVersion':
'RHEL 7.0.0 PC (i440FX + PIIX, 1996)',
>>>>>>>>>>>> 'systemManufacturer': 'Red
Hat'}
>>>>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:35,393::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> call validateAccess with ('qemu',
('qemu', 'kvm'),
>>>>>>>>>>>>
'/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {}
>>>>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:35,395::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> return validateAccess with None
>>>>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:36,067::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> call validateAccess with ('qemu',
('qemu', 'kvm'),
>>>>>>>>>>>>
'/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {}
>>>>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:36,069::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> return validateAccess with None
>>>>>>>>>>>>
MainProcess|PolicyEngine::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:40,619::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> call ksmTune with ({'run': 0},)
{}
>>>>>>>>>>>>
MainProcess|PolicyEngine::DEBUG::2015-11-26
>>>>>>>>>>>>
15:13:40,619::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>>>>>>>>>>>> return ksmTune with None
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *[root@he ~]# tail -f
/var/log/vdsm/connectivity.log *
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2015-11-26
15:02:02,632:DEBUG:recent_client:False
>>>>>>>>>>>> 2015-11-26
15:04:44,975:DEBUG:recent_client:True
>>>>>>>>>>>> 2015-11-26
15:05:15,039:DEBUG:recent_client:False
>>>>>>>>>>>> 2015-11-26
15:07:23,311:DEBUG:recent_client:True
>>>>>>>>>>>> 2015-11-26
15:08:25,774:DEBUG:recent_client:True,
>>>>>>>>>>>> ovirtmgmt:(operstate:up speed:0
duplex:unknown), lo:(operstate:up speed:0
>>>>>>>>>>>> duplex:unknown),
;vdsmdummy;:(operstate:down speed:0 duplex:unknown),
>>>>>>>>>>>> bond0:(operstate:down speed:0
duplex:unknown), eth0:(operstate:up speed:100
>>>>>>>>>>>> duplex:full)
>>>>>>>>>>>> 2015-11-26
15:08:55,845:DEBUG:recent_client:False
>>>>>>>>>>>> 2015-11-26
15:08:59,859:DEBUG:recent_client:True
>>>>>>>>>>>> 2015-11-26
15:09:29,929:DEBUG:recent_client:False
>>>>>>>>>>>> 2015-11-26
15:13:32,292:DEBUG:recent_client:True,
>>>>>>>>>>>> ovirtmgmt:(operstate:up speed:0
duplex:unknown), lo:(operstate:up speed:0
>>>>>>>>>>>> duplex:unknown),
;vdsmdummy;:(operstate:down speed:0 duplex:unknown),
>>>>>>>>>>>> bond0:(operstate:down speed:0
duplex:unknown), eth0:(operstate:up speed:100
>>>>>>>>>>>> duplex:full)
>>>>>>>>>>>> 2015-11-26
15:14:02,363:DEBUG:recent_client:False
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *[root@he ~]# tail -f
/var/log/vdsm/mom.log *
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2015-11-26 15:13:30,581 - mom.Policy -
INFO - Loaded policy
>>>>>>>>>>>> '04-cputune'
>>>>>>>>>>>> 2015-11-26 15:13:30,581 -
mom.PolicyEngine - INFO - Policy
>>>>>>>>>>>> Engine starting
>>>>>>>>>>>> 2015-11-26 15:13:30,582 - mom.RPCServer -
INFO - RPC Server is
>>>>>>>>>>>> disabled
>>>>>>>>>>>> 2015-11-26 15:13:40,618 -
mom.Controllers.KSM - INFO -
>>>>>>>>>>>> Updating KSM configuration:
pages_to_scan:0 merge_across_nodes:8 run:0
>>>>>>>>>>>> sleep_millisecs:0
>>>>>>>>>>>> 2015-11-26 15:14:51,492 - mom.RPCServer -
INFO -
>>>>>>>>>>>> getStatistics()
>>>>>>>>>>>> 2015-11-26 15:14:56,962 - mom.RPCServer -
INFO -
>>>>>>>>>>>> getStatistics()
>>>>>>>>>>>> 2015-11-26 15:15:02,451 - mom.RPCServer -
INFO -
>>>>>>>>>>>> getStatistics()
>>>>>>>>>>>> 2015-11-26 15:15:07,777 - mom.RPCServer -
INFO -
>>>>>>>>>>>> getStatistics()
>>>>>>>>>>>> 2015-11-26 15:15:13,267 - mom.RPCServer -
INFO -
>>>>>>>>>>>> getStatistics()
>>>>>>>>>>>> 2015-11-26 15:15:18,765 - mom.RPCServer -
INFO -
>>>>>>>>>>>> getStatistics()
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Nov 26, 2015 at 3:28 PM, Budur
Nagaraju <
>>>>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Below are the logs,
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> [root@he ~]# tail -f
/var/log/vdsm/vdsm.log
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:50944
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>> http detected from
('127.0.0.1', 50944)
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>>> Adding connection from
127.0.0.1:50945
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>>> Connection removed from
127.0.0.1:50945
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:50945
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>> http detected from
('127.0.0.1', 50945)
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>>> Adding connection from
127.0.0.1:50946
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>>> Connection removed from
127.0.0.1:50946
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:50946
>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26
>>>>>>>>>>>>>
15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>> http detected from
('127.0.0.1', 50946)
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 3:06 PM,
Simone Tiraboschi <
>>>>>>>>>>>>> stirabos(a)redhat.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 10:33 AM,
Budur Nagaraju <
>>>>>>>>>>>>>> nbudoor(a)gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I have done a fresh
installation and now am getting the
>>>>>>>>>>>>>>> below error,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [ INFO ] Updating
hosted-engine configuration
>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction
commit
>>>>>>>>>>>>>>> [ INFO ] Stage: Closing up
>>>>>>>>>>>>>>> The following
network ports should be opened:
>>>>>>>>>>>>>>> tcp:5900
>>>>>>>>>>>>>>> tcp:5901
>>>>>>>>>>>>>>> udp:5900
>>>>>>>>>>>>>>> udp:5901
>>>>>>>>>>>>>>> An example of the
required configuration for
>>>>>>>>>>>>>>> iptables can be found at:
>>>>>>>>>>>>>>>
/etc/ovirt-hosted-engine/iptables.example
>>>>>>>>>>>>>>> In order to
configure firewalld, copy the files
>>>>>>>>>>>>>>> from
>>>>>>>>>>>>>>>
/etc/ovirt-hosted-engine/firewalld to
>>>>>>>>>>>>>>> /etc/firewalld/services
>>>>>>>>>>>>>>> and execute the
following commands:
>>>>>>>>>>>>>>> firewall-cmd
-service hosted-console
>>>>>>>>>>>>>>> [ INFO ] Creating VM
>>>>>>>>>>>>>>> [ ERROR ] Failed to execute
stage 'Closing up': Cannot set
>>>>>>>>>>>>>>> temporary password for
console connection. The VM may not have been
>>>>>>>>>>>>>>> created: please check VDSM
logs
>>>>>>>>>>>>>>> [ INFO ] Stage: Clean up
>>>>>>>>>>>>>>> [ INFO ] Generating answer
file
>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126145701.conf'
>>>>>>>>>>>>>>> [ INFO ] Stage:
Pre-termination
>>>>>>>>>>>>>>> [ INFO ] Stage: Termination
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [root@he ovirt]# tail -f
/var/log/vdsm/
>>>>>>>>>>>>>>> backup/
connectivity.log mom.log
>>>>>>>>>>>>>>> supervdsm.log vdsm.log
>>>>>>>>>>>>>>> [root@he ovirt]# tail -f
/var/log/vdsm/vdsm.log
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:07,564::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:42741
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:07,564::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>>>> http detected from
('127.0.0.1', 42741)
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:07,644::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>>>>> Adding connection from
127.0.0.1:42742
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,088::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>>>>> Connection removed from
127.0.0.1:42742
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,088::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:42742
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,088::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>>>> http detected from
('127.0.0.1', 42742)
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,171::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>>>>>>>>>>>> Adding connection from
127.0.0.1:42743
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,572::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>>>>>>>>>>>> Connection removed from
127.0.0.1:42743
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,573::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>>>>>>>>>>>> Detected protocol xml from
127.0.0.1:42743
>>>>>>>>>>>>>>> Detector
thread::DEBUG::2015-11-26
>>>>>>>>>>>>>>>
14:57:08,573::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>>>>>>>>>>>> http detected from
('127.0.0.1', 42743)
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> It failed before, can you please
attach the whole VDSM logs?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 2:01
PM, Simone Tiraboschi <
>>>>>>>>>>>>>>> stirabos(a)redhat.com>
wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at
7:30 AM, Budur Nagaraju <
>>>>>>>>>>>>>>>> nbudoor(a)gmail.com>
wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Its a fresh setup ,I
have deleted all the vms ,still am
>>>>>>>>>>>>>>>>> facing same issues .
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Can you please paste the
output of
>>>>>>>>>>>>>>>> vdsClient -s 0 list
>>>>>>>>>>>>>>>> ?
>>>>>>>>>>>>>>>> thanks
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Thu, Nov 26, 2015
at 11:56 AM, Oved Ourfali <
>>>>>>>>>>>>>>>>>
oourfali(a)redhat.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Seems like you
have existing VMs running on the host
>>>>>>>>>>>>>>>>>> (you can check
that by looking for qemu processes on your host).
>>>>>>>>>>>>>>>>>> Is that a clean
deployment, or was the host used before
>>>>>>>>>>>>>>>>>> for running VMs?
>>>>>>>>>>>>>>>>>> Perhaps you
already ran the hosted engine setup, and the
>>>>>>>>>>>>>>>>>> VM was left
there?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> CC-ing Sandro who
is more familiar in that than me.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>> Oved
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Thu, Nov 26,
2015 at 7:07 AM, Budur Nagaraju <
>>>>>>>>>>>>>>>>>>
nbudoor(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> HI
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Getting below
error while configuring Hosted engine,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> root@he ~]#
hosted-engine --deploy
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Initializing
>>>>>>>>>>>>>>>>>>> [ INFO ]
Generating a temporary VNC password.
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Environment setup
>>>>>>>>>>>>>>>>>>>
Continuing will configure this host for
>>>>>>>>>>>>>>>>>>> serving as
hypervisor and create a VM where you have to install oVirt
>>>>>>>>>>>>>>>>>>> Engine
afterwards.
>>>>>>>>>>>>>>>>>>> Are
you sure you want to continue? (Yes,
>>>>>>>>>>>>>>>>>>> No)[Yes]:
yes
>>>>>>>>>>>>>>>>>>>
Configuration files: []
>>>>>>>>>>>>>>>>>>> Log
file:
>>>>>>>>>>>>>>>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20151126102302-bkozgk.log
>>>>>>>>>>>>>>>>>>>
Version: otopi-1.3.2 (otopi-1.3.2-1.el6)
>>>>>>>>>>>>>>>>>>> It
has been detected that this program is
>>>>>>>>>>>>>>>>>>> executed
through an SSH connection without using screen.
>>>>>>>>>>>>>>>>>>>
Continuing with the installation may lead to
>>>>>>>>>>>>>>>>>>> broken
installation if the network connection fails.
>>>>>>>>>>>>>>>>>>> It
is highly recommended to abort the
>>>>>>>>>>>>>>>>>>> installation
and run it inside a screen session using command "screen".
>>>>>>>>>>>>>>>>>>> Do
you want to continue anyway? (Yes,
>>>>>>>>>>>>>>>>>>> No)[No]: yes
>>>>>>>>>>>>>>>>>>> [WARNING]
Cannot detect if hardware supports
>>>>>>>>>>>>>>>>>>>
virtualization
>>>>>>>>>>>>>>>>>>> [ INFO ]
Bridge ovirtmgmt already created
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Environment packages setup
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Programs detection
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Environment setup
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> *[ ERROR ]
The following VMs has been found:
>>>>>>>>>>>>>>>>>>>
2b8d6d91-d838-44f6-ae3b-c92cda014280[ ERROR ] Failed to execute stage
>>>>>>>>>>>>>>>>>>>
'Environment setup': Cannot setup Hosted Engine with other VMs running*
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Clean up
>>>>>>>>>>>>>>>>>>> [ INFO ]
Generating answer file
>>>>>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126102310.conf'
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Pre-termination
>>>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Termination
>>>>>>>>>>>>>>>>>>> [root@he ~]#
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>>>> Users mailing
list
>>>>>>>>>>>>>>>>>>>
Users(a)ovirt.org
>>>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>> Users mailing list
>>>>>>>>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Users mailing list
>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Sandro Bonazzola
>>>>>>>>> Better technology. Faster innovation. Powered by
community
>>>>>>>>> collaboration.
>>>>>>>>> See how it works at
redhat.com
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users(a)ovirt.org
>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>
>>
>