On Wed, Nov 15, 2017 at 4:00 PM, Yedidyah Bar David <didi@redhat.com> wrote:
On Wed, Nov 15, 2017 at 1:35 PM, Dafna Ron <dron@redhat.com> wrote:
> Didi,
>
> Thank you for your detailed explanation and for taking the time to debug
> this issue.
>
> I opened the following Jira's:
>
> 1. for increasing entropy in the hosts:
> https://ovirt-jira.atlassian.net/browse/OVIRT-1763

I do not think anymore this is the main reason for slowness, although
it might still be useful to verify/improve.

yuvalt pointed out in a private discussion that openssl lib does nothing
related to random numbers in its pre/post install scripts.

It does call ldconfig, as do several other packages, which can take
quite a lot of time. Some took less. 3 minutes is definitely not
reasonable.

Also see this, from engine log:

2017-11-13 11:07:17,026-05 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
Installing Host lago-upgrade-from-release-suite-master-host0. Yum
update: 398/570: 1:NetworkManager-team-1.8.0-11.el7_4.x86_64.
2017-11-13 11:07:30,573-05 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
Installing Host lago-upgrade-from-release-suite-master-host0. Yum
obsoleting: 399/570: 1:NetworkManager-ppp-1.8.0-11.el7_4.x86_64.
2017-11-13 11:07:45,137-05 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
Installing Host lago-upgrade-from-release-suite-master-host0. Yum
update: 400/570: 1:NetworkManager-tui-1.8.0-11.el7_4.x86_64.
2017-11-13 11:07:57,842-05 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
Installing Host lago-upgrade-from-release-suite-master-host0. Yum
update: 401/570: audit-2.7.6-3.el7.x86_64.

That's ~ 15 seconds per package, and the first 3 have no scripts at all.

I'd say there is some serious storage issue there - bad disk, loaded
storage network/hardware, something like this.

That was my guess as well. Sometimes selinux relabeling takes time (I believe ovirt console does that).
Y.
 

>
> 2. I added a comment to the not all logs are downloaded Jira regarded a
> workaround which would save the logs on a different location:
> https://ovirt-jira.atlassian.net/browse/OVIRT-1583
>
> 3. adding /tmp logs to the job logs:
> https://ovirt-jira.atlassian.net/browse/OVIRT-1764
>
> Again, thank you for your help Didi.
>
> Dafna
>
>
>
> On 11/15/2017 09:01 AM, Yedidyah Bar David wrote:
>> On Tue, Nov 14, 2017 at 5:48 PM, Dafna Ron <dron@redhat.com> wrote:
>>> Hi,
>>>
>>> We had a failure in upgrade suite for 002_bootstrap.add_hosts. I am not
>>> seeing any error that can suggest on an issue in engine.
>>>
>>> I can see in the host messages host that we have stopped writing to the log
>>> for 15 minutes and it may suggest that there is something that is keeping
>>> the host from starting which causes us to fail the test on timeout.
>>>
>>> However,  i can use some help in determining the cause for this failure and
>>> weather its connected to the bootstrap_add_host test in upgrade.
>>>
>>> Link to suspected patches: As I said, I do not think its related, but this
>>> is the patch that was reported.
>>>
>>> https://gerrit.ovirt.org/#/c/83854/
>>>
>>>
>>> Link to Job:
>>>
>>> http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/3795/
>>>
>>>
>>> Link to all logs:
>>>
>>> http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/3795/artifact/
>>>
>>>
>>> (Relevant) error snippet from the log:
>>>
>>> <error>
>>>
>>> Test error:
>>>
>>> Error Message
>>>
>>> False != True after 900 seconds
>>>
>>> Stacktrace
>>>
>>> Traceback (most recent call last):
>>>   File "/usr/lib64/python2.7/unittest/case.py", line 369, in run
>>>     testMethod()
>>>   File "/usr/lib/python2.7/site-packages/nose/case.py", line 197, in runTest
>>>     self.test(*self.arg)
>>>   File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 129, in
>>> wrapped_test
>>>     test()
>>>   File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 59, in
>>> wrapper
>>>     return func(get_test_prefix(), *args, **kwargs)
>>>   File
>>> "/home/jenkins/workspace/ovirt-master_change-queue-tester/ovirt-system-tests/upgrade-from-release-suite-master/test-scenarios-after-upgrade/002_bootstrap.py",
>>> line 187, in add_hosts
>>>     testlib.assert_true_within(_host_is_up_4, timeout=15*60)
>>>   File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 263, in
>>> assert_true_within
>>>     assert_equals_within(func, True, timeout, allowed_exceptions)
>>>   File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 237, in
>>> assert_equals_within
>>>     '%s != %s after %s seconds' % (res, value, timeout)
>>> AssertionError: False != True after 900 seconds
>>>
>> Hi,
>>
>> It took me way too long to find the log file that includes the above
>> stack trace. There is a known bug that pressing '(all files in zip)' does
>> not get all of them, only some. If we can't fix it, please add something
>> that will create such a zip/tar/whatever as part of the job, so that we
>> do not rely on jenkins.
>>
>> Something causes things to be very slow, no idea what in particular,
>> but could not find a real problem.
>>
>> Since we timed out and killed the engine in the middle of host-deploy,
>> we do not have full logs of it. These are copied to the engine from the
>> host only after it finishes. Before that, the log is written on the host
>> in /tmp. We might want to make ost collect that as well, to help debug
>> similar cases.
>>
>> What we can see in engine log is that we start installing the host [1]:
>>
>> 2017-11-13 11:00:12,331-05 INFO
>> [org.ovirt.engine.core.bll.hostdeploy.InstallVdsInternalCommand]
>> (EE-ManagedThreadFactory-engine-Thread-1) [45082420] Before
>> Installation host 83090668-908c-4f49-8690-5348ec12f931,
>> lago-upgrade-from-release-suite-master-host0
>> 2017-11-13 11:00:12,354-05 INFO
>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> (EE-ManagedThreadFactory-engine-Thread-1) [45082420] START,
>> SetVdsStatusVDSCommand(HostName =
>> lago-upgrade-from-release-suite-master-host0,
>> SetVdsStatusVDSCommandParameters:{hostId='83090668-908c-4f49-8690-5348ec12f931',
>> status='Installing', nonOperationalReason='NONE',
>> stopSpmFailureLogged='false', maintenanceReason='null'}), log id:
>> 22a831f7
>>
>> First line from host-deploy is:
>>
>> 2017-11-13 11:00:13,335-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Stage:
>> Initializing.
>>
>> which is one second later, ok.
>>
>> ...
>>
>> 2017-11-13 11:00:13,459-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Stage:
>> Environment packages setup.
>>
>> It goes install (immediately) packages that it needs for itself,
>> in this case it updates the package 'dmidecode', from 2.12-9
>> to 3.0-5. I guess our image is a bit old.
>>
>> ...
>>
>> 2017-11-13 11:00:19,516-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> Verify: 2/2: dmidecode.x86_64 1:2.12-9.el7 - ud.
>>
>> So it took (the equivalent of 'yum update dmidecode') 6 seconds.
>> Reasonable.
>>
>> ...
>>
>> Later on, it starts installing/updating the actual packages it
>> should install/update. Starts with:
>>
>> 2017-11-13 11:00:24,982-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> Status: Downloading Packages.
>>
>> Then there are many lines about each package download. First is:
>>
>> 2017-11-13 11:00:25,049-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> Download/Verify: GeoIP-1.5.0-11.el7.x86_64.
>>
>> Last is:
>>
>> 2017-11-13 11:00:43,790-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> Download/Verify: zlib-1.2.7-17.el7.x86_64.
>>
>> Later we see it updates 570 packages. So downloading 570 packages
>> took ~ 20 seconds, good too. Then it installs them. First is:
>>
>> 2017-11-13 11:00:46,114-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> update: 1/570: libgcc-4.8.5-16.el7.x86_64.
>>
>> And then many similar ones. I only skimmed through them, trying to
>> find large gaps - didn't write a script - and a large one I find is:
>>
>> 2017-11-13 11:10:04,522-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> updated: 506/570: audit.
>>
>> 2017-11-13 11:13:06,483-05 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> (VdsDeploy) [45082420] EVENT_ID: VDS_INSTALL_IN_PROGRESS(509),
>> Installing Host lago-upgrade-from-release-suite-master-host0. Yum
>> updated: 507/570: openssl-libs.
>>
>> In between these two lines, there are a few unrelated ones, from some
>> other engine thread. So something caused it to need 3 minutes to install
>> openssl-libs.
>>
>> A wild guess: Perhaps we have low entropy, and openssl does something
>> that needs entropy?
>>
>> Checking engine-setup log there [2], I see that the following command
>> took almost 1.5 minutes to run:
>>
>> 2017-11-13 10:58:19,459-0500 DEBUG
>> otopi.plugins.ovirt_engine_setup.ovirt_engine.config.sso
>> plugin.executeRaw:813 execute:
>> ('/usr/share/ovirt-engine/bin/ovirt-engine-crypto-tool.sh',
>> 'pbe-encode', '--password=env:pass'), executable='None', cwd='None',
>> env={'pass': '**FILTERED**', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s',
>> 'SSH_CLIENT': '192.168.200.1 52372 22', 'SELINUX_USE_CURRENT_RANGE':
>> '', 'LOGNAME': 'root', 'USER': 'root', 'OVIRT_ENGINE_JAVA_HOME':
>> u'/usr/lib/jvm/jre', 'PATH':
>> '/opt/rh/rh-postgresql95/root/usr/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin',
>> 'HOME': '/root', 'OVIRT_JBOSS_HOME':
>> '/usr/share/ovirt-engine-wildfly', 'LD_LIBRARY_PATH':
>> '/opt/rh/rh-postgresql95/root/usr/lib64', 'LANG': 'en_US.UTF-8',
>> 'SHELL': '/bin/bash', 'LIBRARY_PATH':
>> '/opt/rh/rh-postgresql95/root/usr/lib64', 'SHLVL': '4',
>> 'POSTGRESQLENV':
>> 'COMMAND/pg_dump=str:/opt/rh/rh-postgresql95/root/usr/bin/pg_dump
>>     COMMAND/psql=str:/opt/rh/rh-postgresql95/root/usr/bin/psql
>>  COMMAND/pg_restore=str:/opt/rh/rh-postgresql95/root/usr/bin/pg_restore
>>          COMMAND/postgresql-setup=str:/opt/rh/rh-postgresql95/root/usr/bin/postgresql-setup
>>          OVESETUP_PROVISIONING/postgresService=str:rh-postgresql95-postgresql
>>          OVESETUP_PROVISIONING/postgresConf=str:/var/opt/rh/rh-postgresql95/lib/pgsql/data/postgresql.conf
>>          OVESETUP_PROVISIONING/postgresPgHba=str:/var/opt/rh/rh-postgresql95/lib/pgsql/data/pg_hba.conf
>>          OVESETUP_PROVISIONING/postgresPgVersion=str:/var/opt/rh/rh-postgresql95/lib/pgsql/data/PG_VERSION',
>> 'MANPATH': '/opt/rh/rh-postgresql95/root/usr/share/man:', 'X_SCLS':
>> 'rh-postgresql95 ', 'XDG_RUNTIME_DIR': '/run/user/0',
>> 'OVIRT_ENGINE_JAVA_HOME_FORCE': '1', 'PYTHONPATH':
>> '/usr/share/ovirt-engine/setup/bin/..::', 'SELINUX_ROLE_REQUESTED':
>> '', 'MAIL': '/var/mail/root', 'PKG_CONFIG_PATH':
>> '/opt/rh/rh-postgresql95/root/usr/lib64/pkgconfig', 'XDG_SESSION_ID':
>> '14', 'sclenv': 'rh-postgresql95', 'XDG_CONFIG_DIRS':
>> '/etc/opt/rh/rh-postgresql95/xdg:/etc/xdg', 'JAVACONFDIRS':
>> '/etc/opt/rh/rh-postgresql95/java:/etc/java',
>> 'SELINUX_LEVEL_REQUESTED': '', 'XDG_DATA_DIRS':
>> '/opt/rh/rh-postgresql95/root/usr/share', 'PWD': '/root', 'CPATH':
>> '/opt/rh/rh-postgresql95/root/usr/include', 'OTOPI_LOGFILE':
>> '/var/log/ovirt-engine/setup/ovirt-engine-setup-20171113105548-qrc7zo.log',
>> 'SSH_CONNECTION': '192.168.200.1 52372 192.168.200.3 22',
>> 'OTOPI_EXECDIR': '/root'}
>> 2017-11-13 10:59:43,545-0500 DEBUG
>> otopi.plugins.ovirt_engine_setup.ovirt_engine.config.sso
>> plugin.executeRaw:863 execute-result:
>> ('/usr/share/ovirt-engine/bin/ovirt-engine-crypto-tool.sh',
>> 'pbe-encode', '--password=env:pass'), rc=0
>> 2017-11-13 10:59:43,546-0500 DEBUG
>> otopi.plugins.ovirt_engine_setup.ovirt_engine.config.sso
>> plugin.execute:921 execute-output:
>> ('/usr/share/ovirt-engine/bin/ovirt-engine-crypto-tool.sh',
>> 'pbe-encode', '--password=env:pass') stdout:
>> eyJhcnRpZmFjdCI6IkVudmVsb3BlUEJFIiwic2FsdCI6ImNjSmhFbHRnUEJxeUlNTUJSaU1OdFRYL3M0RGRJT1hOSWJjV2F1NFZGT0U9Iiwic2VjcmV0IjoiK2ZlTzVyZm9kNGlsVmZLRENaRjdseVVQZHZnWnBTWUF0cnBYUWVpQnJaTT0iLCJ2ZXJzaW9uIjoiMSIsIml0ZXJhdGlvbnMiOiI0MDAwIiwiYWxnb3JpdGhtIjoiUEJLREYyV2l0aEhtYWNTSEExIn0=
>>
>> 2017-11-13 10:59:43,546-0500 DEBUG
>> otopi.plugins.ovirt_engine_setup.ovirt_engine.config.sso
>> plugin.execute:926 execute-output:
>> ('/usr/share/ovirt-engine/bin/ovirt-engine-crypto-tool.sh',
>> 'pbe-encode', '--password=env:pass') stderr:
>>
>> This is almost always due to not enough entropy.
>>
>> So please check if everything - from physical machines to
>> lago/ost/libvirt/etc., makes sure to supply all VMs enough
>> entropy. We usually do this using virtio-rng.
>>
>> [1] http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/3795/artifact/exported-artifacts/upgrade-from-release-suit-master-el7/test_logs/upgrade-from-release-suite-master/post-002_bootstrap.py/lago-upgrade-from-release-suite-master-engine/_var_log/ovirt-engine/engine.log
>> [2] http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/3795/artifact/exported-artifacts/upgrade-from-release-suit-master-el7/test_logs/upgrade-from-release-suite-master/post-002_bootstrap.py/lago-upgrade-from-release-suite-master-engine/_var_log/ovirt-engine/setup/ovirt-engine-setup-20171113105548-qrc7zo.log
>>
>>> lago log:
>>>
>>> 2017-11-13
>>> 15:31:23,212::log_utils.py::__enter__::600::lago.prefix::INFO::ESC[0mESC[0m
>>> 2017-11-13
>>> 15:31:23,213::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:0468ed2f-b174-4d94-bc66-2b6e08087a86:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13
>>> 15:31:23,213::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:91372413-f7fd-4b72-85b9-9f5216ca7ae9:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13 15:31:23,213::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13 15:31:23,213::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13
>>> 15:31:26,220::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-engine: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.3
>>> 2017-11-13
>>> 15:31:26,221::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-host0: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.2
>>> 2017-11-13 15:31:27,222::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:91372413-f7fd-4b72-85b9-9f5216ca7ae9:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13 15:31:27,222::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:0468ed2f-b174-4d94-bc66-2b6e08087a86:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:27,222::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-engine:
>>> Timed out (in 4 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13 15:31:27,222::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-host0:
>>> Timed out (in 4 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13
>>> 15:31:28,224::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:fc88f641-b012-4636-a471-9ccaaf361a53:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13
>>> 15:31:28,224::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:afb01a46-6338-407f-b7c9-9d5c6b91404d:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:28,224::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13 15:31:28,225::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13
>>> 15:31:28,226::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-host0: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.2
>>> 2017-11-13 15:31:29,228::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:afb01a46-6338-407f-b7c9-9d5c6b91404d:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:29,228::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-host0:
>>> Timed out (in 1 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13
>>> 15:31:29,229::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-engine: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.3
>>> 2017-11-13 15:31:30,229::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:fc88f641-b012-4636-a471-9ccaaf361a53:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13
>>> 15:31:30,230::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:9df5ddca-bbc4-485a-9f3a-b0b9a5bb5990:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:30,230::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-engine:
>>> Timed out (in 2 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13 15:31:30,230::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13
>>> 15:31:30,231::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-host0: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.2
>>> 2017-11-13
>>> 15:31:31,231::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:53fe67da-a632-49fe-b697-e8a2c4cb7d23:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13 15:31:31,232::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13
>>> 15:31:31,232::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-engine: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.3
>>> 2017-11-13 15:31:31,232::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:9df5ddca-bbc4-485a-9f3a-b0b9a5bb5990:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:31,233::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-host0:
>>> Timed out (in 1 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13 15:31:32,234::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:53fe67da-a632-49fe-b697-e8a2c4cb7d23:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13 15:31:32,234::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-engine:
>>> Timed out (in 1 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13
>>> 15:31:32,234::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:1969bf5b-4e66-43d7-91c6-a28949e98fe8:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:32,234::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13
>>> 15:31:32,235::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-host0: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.2
>>> 2017-11-13
>>> 15:31:33,235::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:5fd0e3c3-d83c-46f9-9c88-bc739aa9c430:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13 15:31:33,235::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13 15:31:33,236::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:1969bf5b-4e66-43d7-91c6-a28949e98fe8:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13
>>> 15:31:33,236::ssh.py::get_ssh_client::354::lago.ssh::DEBUG::Socket error
>>> connecting to lago-upgrade-from-release-suite-master-engine: [Errno None]
>>> Unable to connect to port 22 on 192.168.200.3
>>> 2017-11-13 15:31:33,237::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-host0:
>>> Timed out (in 1 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13 15:31:34,238::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:5fd0e3c3-d83c-46f9-9c88-bc739aa9c430:Get ssh client for
>>> lago-upgrade-from-release-suite-master-engine:
>>> 2017-11-13 15:31:34,238::ssh.py::wait_for_ssh::129::lago.ssh::DEBUG::Got
>>> exception while sshing to lago-upgrade-from-release-suite-master-engine:
>>> Timed out (in 1 s) trying to ssh to
>>> lago-upgrade-from-release-suite-master-engine
>>> 2017-11-13
>>> 15:31:34,238::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:22550d6e-c397-43dc-9691-c8d9debfe416:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:34,239::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 1 tries for lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13 15:31:34,407::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:22550d6e-c397-43dc-9691-c8d9debfe416:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:34,639::ssh.py::ssh::58::lago.ssh::DEBUG::Running bb49adfa
>>> on lago-upgrade-from-release-suite-master-host0: true
>>> 2017-11-13 15:31:34,665::ssh.py::ssh::81::lago.ssh::DEBUG::Command bb49adfa
>>> on lago-upgrade-from-release-suite-master-host0 returned with 0
>>> 2017-11-13 15:31:34,665::ssh.py::wait_for_ssh::153::lago.ssh::DEBUG::Wait
>>> succeeded for ssh to lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13
>>> 15:31:34,665::log_utils.py::__exit__::611::lago.prefix::INFO::ESC[32mSuccessESC[0m
>>> (in 0:00:11)
>>> 2017-11-13
>>> 15:31:34,666::log_utils.py::__enter__::600::lago.prefix::INFO::ESC[0mESC[0m
>>> 2017-11-13
>>> 15:31:34,666::log_utils.py::__enter__::600::lago.ssh::DEBUG::start
>>> task:4e2a7421-28f7-4679-955f-7261fe377939:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:34,667::ssh.py::get_ssh_client::339::lago.ssh::DEBUG::Still
>>> got 100 tries for lago-upgrade-from-release-suite-master-host0
>>> 2017-11-13 15:31:34,805::log_utils.py::__exit__::611::lago.ssh::DEBUG::end
>>> task:4e2a7421-28f7-4679-955f-7261fe377939:Get ssh client for
>>> lago-upgrade-from-release-suite-master-host0:
>>> 2017-11-13 15:31:34,906::ssh.py::ssh::58::lago.ssh::DEBUG::Running bb724bf2
>>> on lago-upgrade-from-release-suite-master-host0: bash -s < "set -xe
>>> DIST=$(uname -r | sed -r  's/^.*\.([^\.]+)\.[^\.]+$/\1/')
>>>
>>>
>>> from host:
>>>
>>> ions. Proceeding anyway.
>>> Nov 13 10:31:34 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Configuration file /usr/lib/systemd/system/ebtables.service is marked
>>> executable. Please remove executable permission bits. Proceeding anyway.
>>> Nov 13 10:31:34 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Configuration file /usr/lib/systemd/system/wpa_supplicant.service is marked
>>> executable. Please remove executable permission bits. Proceeding anyway.
>>> Nov 13 10:31:55 lago-upgrade-from-release-suite-master-host0 yum[1294]:
>>> Updated: iptables-1.4.21-18.2.el7_4.x86_64
>>> Nov 13 10:31:56 lago-upgrade-from-release-suite-master-host0 systemd-logind:
>>> Removed session 3.
>>> Nov 13 10:31:56 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Removed slice user-0.slice.
>>> Nov 13 10:31:56 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Stopping user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Created slice user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Starting user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Started Session 4 of user root.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Starting Session 4 of user root.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd-logind:
>>> New session 4 of user root.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd-logind:
>>> Removed session 4.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Removed slice user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Stopping user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Created slice user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Starting user-0.slice.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Started Session 5 of user root.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Starting Session 5 of user root.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd-logind:
>>> New session 5 of user root.
>>> Nov 13 10:44:00 lago-upgrade-from-release-suite-master-host0 systemd-logind:
>>> Removed session 5.
>>> Nov 13 10:44:01 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Removed slice user-0.slice.
>>> Nov 13 10:44:01 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Stopping user-0.slice.
>>> Nov 13 10:46:27 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Starting Cleanup of Temporary Directories...
>>> Nov 13 10:46:27 lago-upgrade-from-release-suite-master-host0 systemd:
>>> Started Cleanup of Temporary Directories.
>>> Nov 13 10:54:36 lago-upgrade-from-release-suite-master-host0 dhclient[906]:
>>> DHCPREQUEST on eth0 to 192.168.200.1 port 67 (xid=0x564629cb)
>>> Nov 13 10:54:36 lago-upgrade-from-release-suite-master-host0 dhclient[906]:
>>> DHCPACK from 192.168.200.1 (xid=0x564629cb)
>>>
>>> </error>
>>>
>>>
>>
>>
>



--
Didi