[ovirt-users] HA cluster
Budur Nagaraju
nbudoor at gmail.com
Fri Nov 27 07:34:27 UTC 2015
I got only 10lines to in the vdsm logs and are below ,
[root at he /]# tail -f /var/log/vdsm/vdsm.log
Thread-100::DEBUG::2015-11-27
12:58:57,360::resourceManager::616::Storage.ResourceManager::(releaseResource)
Trying to release resource 'Storage.HsmDomainMonitorLock'
Thread-100::DEBUG::2015-11-27
12:58:57,360::resourceManager::635::Storage.ResourceManager::(releaseResource)
Released resource 'Storage.HsmDomainMonitorLock' (0 active users)
Thread-100::DEBUG::2015-11-27
12:58:57,360::resourceManager::641::Storage.ResourceManager::(releaseResource)
Resource 'Storage.HsmDomainMonitorLock' is free, finding out if anyone is
waiting for it.
Thread-100::DEBUG::2015-11-27
12:58:57,360::resourceManager::649::Storage.ResourceManager::(releaseResource)
No one is waiting for resource 'Storage.HsmDomainMonitorLock', Clearing
records.
Thread-100::INFO::2015-11-27
12:58:57,360::logUtils::47::dispatcher::(wrapper) Run and protect:
stopMonitoringDomain, Return response: None
Thread-100::DEBUG::2015-11-27
12:58:57,361::task::1191::Storage.TaskManager.Task::(prepare)
Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::finished: None
Thread-100::DEBUG::2015-11-27
12:58:57,361::task::595::Storage.TaskManager.Task::(_updateState)
Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::moving from state preparing ->
state finished
Thread-100::DEBUG::2015-11-27
12:58:57,361::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-100::DEBUG::2015-11-27
12:58:57,361::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-100::DEBUG::2015-11-27
12:58:57,361::task::993::Storage.TaskManager.Task::(_decref)
Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::ref 0 aborting False
On Thu, Nov 26, 2015 at 4:20 PM, Simone Tiraboschi <stirabos at redhat.com>
wrote:
>
>
> On Thu, Nov 26, 2015 at 11:05 AM, Budur Nagaraju <nbudoor at gmail.com>
> wrote:
>
>>
>>
>>
>> *Below are the entire logs*
>>
>>
> Sorry, with the entire log I mean if you can attach or share somewhere the
> whole /var/log/vdsm/vdsm.log cause the latest ten lines are not enough to
> point out the issue.
>
>
>>
>>
>>
>>
>> *[root at he ~]# tail -f /var/log/vdsm/vdsm.log *
>>
>> Detector thread::DEBUG::2015-11-26
>> 15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>> Detected protocol xml from 127.0.0.1:50944
>> Detector thread::DEBUG::2015-11-26
>> 15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>> http detected from ('127.0.0.1', 50944)
>> Detector thread::DEBUG::2015-11-26
>> 15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>> Adding connection from 127.0.0.1:50945
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>> Connection removed from 127.0.0.1:50945
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>> Detected protocol xml from 127.0.0.1:50945
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>> http detected from ('127.0.0.1', 50945)
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>> Adding connection from 127.0.0.1:50946
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>> Connection removed from 127.0.0.1:50946
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>> Detected protocol xml from 127.0.0.1:50946
>> Detector thread::DEBUG::2015-11-26
>> 15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>> http detected from ('127.0.0.1', 50946)
>>
>>
>>
>>
>> *[root at he ~]# tail -f /var/log/vdsm/supervdsm.log *
>>
>> MainProcess::DEBUG::2015-11-26
>> 15:13:30,234::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>> call readMultipathConf with () {}
>> MainProcess::DEBUG::2015-11-26
>> 15:13:30,234::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>> return readMultipathConf with ['# RHEV REVISION 1.1', '', 'defaults {',
>> ' polling_interval 5', ' getuid_callout
>> "/lib/udev/scsi_id --whitelisted --replace-whitespace --device=/dev/%n"',
>> ' no_path_retry fail', ' user_friendly_names no', '
>> flush_on_last_del yes', ' fast_io_fail_tmo 5', '
>> dev_loss_tmo 30', ' max_fds 4096', '}', '',
>> 'devices {', 'device {', ' vendor "HITACHI"', '
>> product "DF.*"', ' getuid_callout
>> "/lib/udev/scsi_id --whitelisted --replace-whitespace --device=/dev/%n"',
>> '}', 'device {', ' vendor "COMPELNT"', '
>> product "Compellent Vol"', ' no_path_retry
>> fail', '}', 'device {', ' # multipath.conf.default', '
>> vendor "DGC"', ' product ".*"', '
>> product_blacklist "LUNZ"', ' path_grouping_policy
>> "group_by_prio"', ' path_checker "emc_clariion"', '
>> hardware_handler "1 emc"', ' prio "emc"', '
>> failback immediate', ' rr_weight
>> "uniform"', ' # vdsm required configuration', '
>> getuid_callout "/lib/udev/scsi_id --whitelisted
>> --replace-whitespace --device=/dev/%n"', ' features "0"',
>> ' no_path_retry fail', '}', '}']
>> MainProcess|Thread-13::DEBUG::2015-11-26
>> 15:13:31,365::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>> call getHardwareInfo with () {}
>> MainProcess|Thread-13::DEBUG::2015-11-26
>> 15:13:31,397::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>> return getHardwareInfo with {'systemProductName': 'KVM', 'systemUUID':
>> 'f91632f2-7a17-4ddb-9631-742f82a77480', 'systemFamily': 'Red Hat Enterprise
>> Linux', 'systemVersion': 'RHEL 7.0.0 PC (i440FX + PIIX, 1996)',
>> 'systemManufacturer': 'Red Hat'}
>> MainProcess|Thread-21::DEBUG::2015-11-26
>> 15:13:35,393::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>> call validateAccess with ('qemu', ('qemu', 'kvm'),
>> '/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {}
>> MainProcess|Thread-21::DEBUG::2015-11-26
>> 15:13:35,395::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>> return validateAccess with None
>> MainProcess|Thread-22::DEBUG::2015-11-26
>> 15:13:36,067::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>> call validateAccess with ('qemu', ('qemu', 'kvm'),
>> '/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {}
>> MainProcess|Thread-22::DEBUG::2015-11-26
>> 15:13:36,069::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>> return validateAccess with None
>> MainProcess|PolicyEngine::DEBUG::2015-11-26
>> 15:13:40,619::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
>> call ksmTune with ({'run': 0},) {}
>> MainProcess|PolicyEngine::DEBUG::2015-11-26
>> 15:13:40,619::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
>> return ksmTune with None
>>
>>
>>
>> *[root at he ~]# tail -f /var/log/vdsm/connectivity.log *
>>
>>
>> 2015-11-26 15:02:02,632:DEBUG:recent_client:False
>> 2015-11-26 15:04:44,975:DEBUG:recent_client:True
>> 2015-11-26 15:05:15,039:DEBUG:recent_client:False
>> 2015-11-26 15:07:23,311:DEBUG:recent_client:True
>> 2015-11-26 15:08:25,774:DEBUG:recent_client:True, ovirtmgmt:(operstate:up
>> speed:0 duplex:unknown), lo:(operstate:up speed:0 duplex:unknown),
>> ;vdsmdummy;:(operstate:down speed:0 duplex:unknown), bond0:(operstate:down
>> speed:0 duplex:unknown), eth0:(operstate:up speed:100 duplex:full)
>> 2015-11-26 15:08:55,845:DEBUG:recent_client:False
>> 2015-11-26 15:08:59,859:DEBUG:recent_client:True
>> 2015-11-26 15:09:29,929:DEBUG:recent_client:False
>> 2015-11-26 15:13:32,292:DEBUG:recent_client:True, ovirtmgmt:(operstate:up
>> speed:0 duplex:unknown), lo:(operstate:up speed:0 duplex:unknown),
>> ;vdsmdummy;:(operstate:down speed:0 duplex:unknown), bond0:(operstate:down
>> speed:0 duplex:unknown), eth0:(operstate:up speed:100 duplex:full)
>> 2015-11-26 15:14:02,363:DEBUG:recent_client:False
>>
>>
>>
>>
>> *[root at he ~]# tail -f /var/log/vdsm/mom.log *
>>
>>
>> 2015-11-26 15:13:30,581 - mom.Policy - INFO - Loaded policy '04-cputune'
>> 2015-11-26 15:13:30,581 - mom.PolicyEngine - INFO - Policy Engine starting
>> 2015-11-26 15:13:30,582 - mom.RPCServer - INFO - RPC Server is disabled
>> 2015-11-26 15:13:40,618 - mom.Controllers.KSM - INFO - Updating KSM
>> configuration: pages_to_scan:0 merge_across_nodes:8 run:0 sleep_millisecs:0
>> 2015-11-26 15:14:51,492 - mom.RPCServer - INFO - getStatistics()
>> 2015-11-26 15:14:56,962 - mom.RPCServer - INFO - getStatistics()
>> 2015-11-26 15:15:02,451 - mom.RPCServer - INFO - getStatistics()
>> 2015-11-26 15:15:07,777 - mom.RPCServer - INFO - getStatistics()
>> 2015-11-26 15:15:13,267 - mom.RPCServer - INFO - getStatistics()
>> 2015-11-26 15:15:18,765 - mom.RPCServer - INFO - getStatistics()
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Thu, Nov 26, 2015 at 3:28 PM, Budur Nagaraju <nbudoor at gmail.com>
>> wrote:
>>
>>> Below are the logs,
>>>
>>>
>>> [root at he ~]# tail -f /var/log/vdsm/vdsm.log
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>> Detected protocol xml from 127.0.0.1:50944
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>> http detected from ('127.0.0.1', 50944)
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>> Adding connection from 127.0.0.1:50945
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>> Connection removed from 127.0.0.1:50945
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>> Detected protocol xml from 127.0.0.1:50945
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>> http detected from ('127.0.0.1', 50945)
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>> Adding connection from 127.0.0.1:50946
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>> Connection removed from 127.0.0.1:50946
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>> Detected protocol xml from 127.0.0.1:50946
>>> Detector thread::DEBUG::2015-11-26
>>> 15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>> http detected from ('127.0.0.1', 50946)
>>>
>>>
>>>
>>> On Thu, Nov 26, 2015 at 3:06 PM, Simone Tiraboschi <stirabos at redhat.com>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Thu, Nov 26, 2015 at 10:33 AM, Budur Nagaraju <nbudoor at gmail.com>
>>>> wrote:
>>>>
>>>>> I have done a fresh installation and now am getting the below error,
>>>>>
>>>>> [ INFO ] Updating hosted-engine configuration
>>>>> [ INFO ] Stage: Transaction commit
>>>>> [ INFO ] Stage: Closing up
>>>>> The following network ports should be opened:
>>>>> tcp:5900
>>>>> tcp:5901
>>>>> udp:5900
>>>>> udp:5901
>>>>> An example of the required configuration for iptables can be
>>>>> found at:
>>>>> /etc/ovirt-hosted-engine/iptables.example
>>>>> In order to configure firewalld, copy the files from
>>>>> /etc/ovirt-hosted-engine/firewalld to /etc/firewalld/services
>>>>> and execute the following commands:
>>>>> firewall-cmd -service hosted-console
>>>>> [ INFO ] Creating VM
>>>>> [ ERROR ] Failed to execute stage 'Closing up': Cannot set temporary
>>>>> password for console connection. The VM may not have been created: please
>>>>> check VDSM logs
>>>>> [ INFO ] Stage: Clean up
>>>>> [ INFO ] Generating answer file
>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126145701.conf'
>>>>> [ INFO ] Stage: Pre-termination
>>>>> [ INFO ] Stage: Termination
>>>>>
>>>>>
>>>>>
>>>>> [root at he ovirt]# tail -f /var/log/vdsm/
>>>>> backup/ connectivity.log mom.log
>>>>> supervdsm.log vdsm.log
>>>>> [root at he ovirt]# tail -f /var/log/vdsm/vdsm.log
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:07,564::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>> Detected protocol xml from 127.0.0.1:42741
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:07,564::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>> http detected from ('127.0.0.1', 42741)
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:07,644::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>> Adding connection from 127.0.0.1:42742
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,088::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>> Connection removed from 127.0.0.1:42742
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,088::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>> Detected protocol xml from 127.0.0.1:42742
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,088::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>> http detected from ('127.0.0.1', 42742)
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,171::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection)
>>>>> Adding connection from 127.0.0.1:42743
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,572::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection)
>>>>> Connection removed from 127.0.0.1:42743
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,573::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read)
>>>>> Detected protocol xml from 127.0.0.1:42743
>>>>> Detector thread::DEBUG::2015-11-26
>>>>> 14:57:08,573::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml over
>>>>> http detected from ('127.0.0.1', 42743)
>>>>>
>>>>>
>>>>
>>>> It failed before, can you please attach the whole VDSM logs?
>>>>
>>>>
>>>>>
>>>>> On Thu, Nov 26, 2015 at 2:01 PM, Simone Tiraboschi <
>>>>> stirabos at redhat.com> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Nov 26, 2015 at 7:30 AM, Budur Nagaraju <nbudoor at gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Its a fresh setup ,I have deleted all the vms ,still am facing same
>>>>>>> issues .
>>>>>>>
>>>>>>>
>>>>>> Can you please paste the output of
>>>>>> vdsClient -s 0 list
>>>>>> ?
>>>>>> thanks
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> On Thu, Nov 26, 2015 at 11:56 AM, Oved Ourfali <oourfali at redhat.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi
>>>>>>>>
>>>>>>>> Seems like you have existing VMs running on the host (you can check
>>>>>>>> that by looking for qemu processes on your host).
>>>>>>>> Is that a clean deployment, or was the host used before for running
>>>>>>>> VMs?
>>>>>>>> Perhaps you already ran the hosted engine setup, and the VM was
>>>>>>>> left there?
>>>>>>>>
>>>>>>>> CC-ing Sandro who is more familiar in that than me.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Oved
>>>>>>>>
>>>>>>>> On Thu, Nov 26, 2015 at 7:07 AM, Budur Nagaraju <nbudoor at gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> HI
>>>>>>>>>
>>>>>>>>> Getting below error while configuring Hosted engine,
>>>>>>>>>
>>>>>>>>> root at he ~]# hosted-engine --deploy
>>>>>>>>> [ INFO ] Stage: Initializing
>>>>>>>>> [ INFO ] Generating a temporary VNC password.
>>>>>>>>> [ INFO ] Stage: Environment setup
>>>>>>>>> Continuing will configure this host for serving as
>>>>>>>>> hypervisor and create a VM where you have to install oVirt Engine
>>>>>>>>> afterwards.
>>>>>>>>> Are you sure you want to continue? (Yes, No)[Yes]: yes
>>>>>>>>> Configuration files: []
>>>>>>>>> Log file:
>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20151126102302-bkozgk.log
>>>>>>>>> Version: otopi-1.3.2 (otopi-1.3.2-1.el6)
>>>>>>>>> It has been detected that this program is executed
>>>>>>>>> through an SSH connection without using screen.
>>>>>>>>> Continuing with the installation may lead to broken
>>>>>>>>> installation if the network connection fails.
>>>>>>>>> It is highly recommended to abort the installation and
>>>>>>>>> run it inside a screen session using command "screen".
>>>>>>>>> Do you want to continue anyway? (Yes, No)[No]: yes
>>>>>>>>> [WARNING] Cannot detect if hardware supports virtualization
>>>>>>>>> [ INFO ] Bridge ovirtmgmt already created
>>>>>>>>> [ INFO ] Stage: Environment packages setup
>>>>>>>>> [ INFO ] Stage: Programs detection
>>>>>>>>> [ INFO ] Stage: Environment setup
>>>>>>>>>
>>>>>>>>> *[ ERROR ] The following VMs has been found:
>>>>>>>>> 2b8d6d91-d838-44f6-ae3b-c92cda014280[ ERROR ] Failed to execute stage
>>>>>>>>> 'Environment setup': Cannot setup Hosted Engine with other VMs running*
>>>>>>>>> [ INFO ] Stage: Clean up
>>>>>>>>> [ INFO ] Generating answer file
>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126102310.conf'
>>>>>>>>> [ INFO ] Stage: Pre-termination
>>>>>>>>> [ INFO ] Stage: Termination
>>>>>>>>> [root at he ~]#
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list
>>>>>>>>> Users at ovirt.org
>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users at ovirt.org
>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20151127/0a086086/attachment-0001.html>
More information about the Users
mailing list