Hi Jayme,
It getting hard to debug your issue over the mailing list.
Can you please open a bug in bugzilla and attach all the information you
know?
The versions you are using, config files from engine machine
and ansible log, vars.yaml , integ.ini from the master0 vm in /root
directory, .
Sorry for the inconvenience.
Best,
--
Shirly Radco
BI Senior Software Engineer
Red Hat <
I actually see the pods running on master0 if I do this:
@master0 master]# oc project kube-system
Now using project "kube-system" on server "
https://openshift-master.cloud.xxxxxxx.com:8443".
[root@master0 master]# oc get pods
NAME READY STATUS
RESTARTS AGE
master-api-master0.cloud.xxxxxxxx.com 1/1 Running 0
22m
master-controllers-master0.cloud.xxxxxxxx.com 1/1 Running 0
22m
master-etcd-master0.cloud.xxxxxxxxxx 1/1 Running 0
22m
So I wonder why the ansible "Wait for control plane pods to appear" task
is looping
- name: Wait for control plane pods to appear
oc_obj:
state: list
kind: pod
name: "master-{{ item }}-{{ l_kubelet_node_name | lower }}"
namespace: kube-system
register: control_plane_pods
until:
- "'results' in control_plane_pods"
- "'results' in control_plane_pods.results"
- control_plane_pods.results.results | length > 0
retries: 60
delay: 5
with_items:
- "{{ 'etcd' if inventory_hostname in groups['oo_etcd_to_config']
else
omit }}"
- api
- controllers
ignore_errors: true
On Tue, May 28, 2019 at 4:23 PM Jayme <jaymef(a)gmail.com> wrote:
> I just tried again from scratch this time making sure a proper wildcard
> DNS entry existed and without using the set /etc/hosts option and am still
> running in to the pods issue. Can anyone confirm if this requires a public
> external IP to work? I am working on an internal DNS zone here and natted
> ips.
>
> On Tue, May 28, 2019 at 3:28 PM Edward Berger <edwberger(a)gmail.com>
> wrote:
>
>> In my case it was a single bare metal host, so that would be equivalent
>> to disabling iptables on the master0 VM you're installing to, in your ovirt
>> scenario.
>>
>> On Tue, May 28, 2019 at 1:25 PM Jayme <jaymef(a)gmail.com> wrote:
>>
>>> Do you mean the iptables firewall on the server being installed to i.e.
>>> master0 or the actual oVirt host that the master0 VM is running on? I did
>>> try flushing iptables rules on master0 VM then ran plays again from
>>> installer VM but fail at the same point.
>>>
>>> Does this log message have anything to do with the issue, /etc/cni
>>> directory does not even exist on master0 VM.
>>>
>>> May 28 17:23:35 master0 origin-node: W0528 17:23:35.012902 10434
>>> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d
>>> May 28 17:23:35 master0 origin-node: E0528 17:23:35.013398 10434
>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready:
>>> cni config uninitialized
>>>
>>>
>>>
>>> On Tue, May 28, 2019 at 1:19 PM Edward Berger <edwberger(a)gmail.com>
>>> wrote:
>>>
>>>> > TASK [openshift_control_plane : Wait for control plane pods to
>>>> appear] *********
>>>> > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857
>>>> ************
>>>> > FAILED - RETRYING: Wait for control plane pods to appear (60
retries
>>>> left).
>>>> > FAILED - RETRYING: Wait for control plane pods to appear (59
retries
>>>> left).
>>>> > It eventually counts all the way down to zero and fails.
>>>>
>>>> This looks a lot like the issues I saw when the host firewall
>>>> (iptables) was blocking another OKD all-in-one-host install script [1].
>>>> Disabling iptables allowed the installation to continue for my proof
>>>> of concept "cluster".
>>>>
>>>> [
1]https://github.com/gshipley/installcentos
>>>>
>>>> The other error I had with [1] was it was trying to install a couple
>>>> of packages (zile and python2-pip) from EPEL with the repo disabled.
>>>>
>>>>
>>>>
>>>> On Tue, May 28, 2019 at 10:41 AM Jayme <jaymef(a)gmail.com> wrote:
>>>>
>>>>> Shirly,
>>>>>
>>>>> Oh and I should mention that I did verify that NetworkManager was
>>>>> installed on the master0 VM and enabled/started the second go around.
So
>>>>> that service is there and running.
>>>>>
>>>>> # systemctl list-unit-files | grep Network
>>>>> dbus-org.freedesktop.NetworkManager.service
>>>>> enabled
>>>>> NetworkManager-dispatcher.service
>>>>> enabled
>>>>> NetworkManager-wait-online.service
>>>>> enabled
>>>>> NetworkManager.service
>>>>> enabled
>>>>>
>>>>> On Tue, May 28, 2019 at 11:13 AM Jayme <jaymef(a)gmail.com>
wrote:
>>>>>
>>>>>> Shirly,
>>>>>>
>>>>>> I appreciate the help with this. Unfortunately I am still
running
>>>>>> in to the same problem. So far I've tried to
install/enable/start
>>>>>> NetworkManager on the existing "master0" server and
re-ran the plans from
>>>>>> the installer VM. I ran in to the same problem waiting for
control plane
>>>>>> pods and same errors in syslog.
>>>>>>
>>>>>> So I wiped everything out, killed the template along with the
>>>>>> installer and master VMs. On oVirt engine (I am running
4.3.3.7-1 stable)
>>>>>> I did have ovirt-engine-metrics-1.3.0x rpm installed, no yum
updates
>>>>>> available on an update check. So I installed
>>>>>>
http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm
>>>>>> then proceeded to install the latest version of
ovirt-engine-metrics which
>>>>>> gave me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted
engine.
>>>>>>
>>>>>> After that package was installed I proceeded to follow steps
from
>>>>>> beginning outlined at:
>>>>>>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_... --
>>>>>> I ran in to the docker check issue again (same as my initial
email) so I
>>>>>> disabled that and again got as far as starting control plane pods
before
>>>>>> failure.
>>>>>>
>>>>>> Not sure where to go from here at this point. The only thing I
can
>>>>>> think of that I did differently vs the instructions outlined
above is that
>>>>>> I have not crated the wildcard DNS record, however I did set
configs to
>>>>>> create /etc/hosts entries and they /etc/hosts on the machines
have the
>>>>>> proper IPs assigned for all hostnames (automatically added by the
ansible
>>>>>> plays).
>>>>>>
>>>>>> Any ideas how I can get past the plane pods issue?
>>>>>>
>>>>>> Thanks!
>>>>>>
>>>>>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco
<sradco(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> The latest release of 4.3.z should already include a fix for
this
>>>>>>> issue, ovirt-engine-metrics-1.3.1 rpm.
>>>>>>>
>>>>>>> The issue is that it requires the NetworkManagar to be
installed,
>>>>>>> running and enabled for it to work.
>>>>>>>
>>>>>>> You can install it manually on the master0 vm , start and
enable it
>>>>>>> or you can also install the updated rpm from the nightly
builds if your
>>>>>>> environment is oVirt 4.2.z:
>>>>>>>
>>>>>>>
https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-e...
>>>>>>>
>>>>>>> Relevant patches are:
>>>>>>>
https://gerrit.ovirt.org/#/c/99959/
>>>>>>>
https://gerrit.ovirt.org/#/c/99718/
>>>>>>>
>>>>>>> Best regards,
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> Shirly Radco
>>>>>>>
>>>>>>> BI Senior Software Engineer
>>>>>>>
>>>>>>> Red Hat <
https://www.redhat.com/>
>>>>>>>
>>>>>>> <
https://www.redhat.com/>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, May 27, 2019 at 4:41 PM Jayme
<jaymef(a)gmail.com> wrote:
>>>>>>>
>>>>>>>> I managed to get past that but am running in to another
problem
>>>>>>>> later in the process on the control plane pods to appear
task. I thought
>>>>>>>> perhaps a glitch in the process from the failed docker
step previously so
>>>>>>>> after a few more runs I tried killing everything and
restarting the metrics
>>>>>>>> process again from the very beginning and end up hitting
the same issue
>>>>>>>> with control plane pods even though all other steps/tasks
seem to be
>>>>>>>> working.
>>>>>>>>
>>>>>>>> I'm just getting this:
>>>>>>>>
>>>>>>>> TASK [openshift_control_plane : Wait for control plane
pods to
>>>>>>>> appear] *********
>>>>>>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180)
0:14:33.857
>>>>>>>> ************
>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear
(60
>>>>>>>> retries left).
>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear
(59
>>>>>>>> retries left).
>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear
(58
>>>>>>>> retries left).
>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear
(57
>>>>>>>> retries left).
>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear
(56
>>>>>>>> retries left).
>>>>>>>>
>>>>>>>> It eventually counts all the way down to zero and fails.
>>>>>>>>
>>>>>>>> In syslog of the master0 server I'm seeing some
errors related to
>>>>>>>> cni config
>>>>>>>>
>>>>>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with
files=None
>>>>>>>> kind=pod force=False all_namespaces=None
field_selector=None
>>>>>>>> namespace=kube-system delete_after=False
>>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig
content=None state=list
>>>>>>>> debug=False selector=None
name=master-api-master0.xxxxxx.com
>>>>>>>> May 27 13:39:09 master0 origin-node: W0527
13:39:09.064230 20150
>>>>>>>> cni.go:172] Unable to update cni config: No networks
found in /etc/cni/net.d
>>>>>>>> May 27 13:39:09 master0 origin-node: E0527
13:39:09.064670 20150
>>>>>>>> kubelet.go:2101] Container runtime network not ready:
NetworkReady=false
>>>>>>>> reason:NetworkPluginNotReady message:docker: network
plugin is not ready:
>>>>>>>> cni config uninitialized
>>>>>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with
files=None
>>>>>>>> kind=pod force=False all_namespaces=None
field_selector=None
>>>>>>>> namespace=kube-system delete_after=False
>>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig
content=None state=list
>>>>>>>> debug=False selector=None
name=master-api-master0.xxxxxx.com
>>>>>>>> May 27 13:39:14 master0 origin-node: W0527
13:39:14.066911 20150
>>>>>>>> cni.go:172] Unable to update cni config: No networks
found in /etc/cni/net.d
>>>>>>>> May 27 13:39:14 master0 origin-node: E0527
13:39:14.067321 20150
>>>>>>>> kubelet.go:2101] Container runtime network not ready:
NetworkReady=false
>>>>>>>> reason:NetworkPluginNotReady message:docker: network
plugin is not ready:
>>>>>>>> cni config uninitialized
>>>>>>>> May 27 13:39:14 master0 origin-node: E0527
13:39:14.814705 20150
>>>>>>>> summary.go:102] Failed to get system container stats for
>>>>>>>> "/system.slice/origin-node.service": failed to
get cgroup stats for
>>>>>>>> "/system.slice/origin-node.service": failed to
get container info for
>>>>>>>> "/system.slice/origin-node.service": unknown
container
>>>>>>>> "/system.slice/origin-node.service"
>>>>>>>> May 27 13:39:19 master0 origin-node: W0527
13:39:19.069450 20150
>>>>>>>> cni.go:172] Unable to update cni config: No networks
found in /etc/cni/net.d
>>>>>>>> May 27 13:39:19 master0 origin-node: E0527
13:39:19.069850 20150
>>>>>>>> kubelet.go:2101] Container runtime network not ready:
NetworkReady=false
>>>>>>>> reason:NetworkPluginNotReady message:docker: network
plugin is not ready:
>>>>>>>> cni config uninitialized
>>>>>>>>
>>>>>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco
<sradco(a)redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi Jayme,
>>>>>>>>>
>>>>>>>>> Thank you for reaching out.
>>>>>>>>> Please try rerunning the ansible playbook.
>>>>>>>>> If this doesn't work, try adding to the integ.ini
in the metrics
>>>>>>>>> vm
>>>>>>>>> openshift_disable_check=docker_storage
>>>>>>>>> and rerun the ansible playbook again.
>>>>>>>>>
>>>>>>>>> Please update how it goes.
>>>>>>>>>
>>>>>>>>> Best regards,
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> Shirly Radco
>>>>>>>>>
>>>>>>>>> BI Senior Software Engineer
>>>>>>>>>
>>>>>>>>> Red Hat <
https://www.redhat.com/>
>>>>>>>>>
>>>>>>>>> <
https://www.redhat.com/>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme
<jaymef(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> I'm running in to this ansible error during
oVirt metrics
>>>>>>>>>> installation (following procedures at:
>>>>>>>>>>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_...
>>>>>>>>>> )
>>>>>>>>>>
>>>>>>>>>> This is happening late in the process, after
successfully
>>>>>>>>>> deploying the installation VM and then running
second step from the metrics
>>>>>>>>>> VM.
>>>>>>>>>>
>>>>>>>>>> CHECK [memory_availability :
master0.xxxxxx.com]
>>>>>>>>>>
*************************************************************************************************
>>>>>>>>>> fatal: [
master0.xxxxxxx.com]: FAILED! =>
{"changed": true,
>>>>>>>>>> "checks":
{"disk_availability": {}, "docker_image_availability":
>>>>>>>>>> {"changed": true},
"docker_storage": {"failed": true, "failures":
>>>>>>>>>> [["OpenShiftCheckException",
"Could not find imported module support code
>>>>>>>>>> for docker_info. Looked for either
AnsibleDockerClient.py or
>>>>>>>>>> docker_common.py\nTraceback (most recent call
last):\n File
>>>>>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>>>>>> line 225, in run_check\n result =
check.run()\n File
>>>>>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>>>>>> line 53, in run\n docker_info =
self.execute_module(\"docker_info\",
>>>>>>>>>> {})\n File
>>>>>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>>>>>> line 211, in execute_module\n result =
self._execute_module(module_name,
>>>>>>>>>> module_args, self.tmp, self.task_vars)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>>>>>> line 809, in _execute_module\n (module_style,
shebang, module_data,
>>>>>>>>>> module_path) =
self._configure_module(module_name=module_name,
>>>>>>>>>> module_args=module_args, task_vars=task_vars)\n
File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>>>>>> line 203, in _configure_module\n
environment=final_environment)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>>>>>> line 1023, in modify_module\n
environment=environment)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>>>>>> line 859, in _find_module_utils\n
recursive_finder(module_name,
>>>>>>>>>> b_module_data, py_module_names, py_module_cache,
zf)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>>>>>> line 621, in recursive_finder\n raise
AnsibleError('
>>>>>>>>>> '.join(msg))\nAnsibleError: Could not find
imported module support code for
>>>>>>>>>> docker_info. Looked for either
AnsibleDockerClient.py or
>>>>>>>>>> docker_common.py\n"]], "msg":
"Could not find imported module support code
>>>>>>>>>> for docker_info. Looked for either
AnsibleDockerClient.py or
>>>>>>>>>> docker_common.py\nTraceback (most recent call
last):\n File
>>>>>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>>>>>> line 225, in run_check\n result =
check.run()\n File
>>>>>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>>>>>> line 53, in run\n docker_info =
self.execute_module(\"docker_info\",
>>>>>>>>>> {})\n File
>>>>>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>>>>>> line 211, in execute_module\n result =
self._execute_module(module_name,
>>>>>>>>>> module_args, self.tmp, self.task_vars)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>>>>>> line 809, in _execute_module\n (module_style,
shebang, module_data,
>>>>>>>>>> module_path) =
self._configure_module(module_name=module_name,
>>>>>>>>>> module_args=module_args, task_vars=task_vars)\n
File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>>>>>> line 203, in _configure_module\n
environment=final_environment)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>>>>>> line 1023, in modify_module\n
environment=environment)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>>>>>> line 859, in _find_module_utils\n
recursive_finder(module_name,
>>>>>>>>>> b_module_data, py_module_names, py_module_cache,
zf)\n File
>>>>>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>>>>>> line 621, in recursive_finder\n raise
AnsibleError('
>>>>>>>>>> '.join(msg))\nAnsibleError: Could not find
imported module support code for
>>>>>>>>>> docker_info. Looked for either
AnsibleDockerClient.py or
>>>>>>>>>> docker_common.py\n"},
"memory_availability": {}, "package_availability":
>>>>>>>>>> {"changed": false,
"invocation": {"module_args": {"packages":
["PyYAML",
>>>>>>>>>> "bash-completion", "bind",
"ceph-common", "dnsmasq", "docker", "firewalld",
>>>>>>>>>> "flannel", "glusterfs-fuse",
"httpd-tools", "iptables",
>>>>>>>>>> "iptables-services",
"iscsi-initiator-utils", "libselinux-python",
>>>>>>>>>> "nfs-utils", "ntp",
"openssl", "origin", "origin-clients",
>>>>>>>>>> "origin-hyperkube",
"origin-node", "pyparted", "python-httplib2",
>>>>>>>>>> "yum-utils"]}}},
"package_version": {"changed": false, "invocation":
>>>>>>>>>> {"module_args":
{"package_list": [{"check_multi": false, "name":
"origin",
>>>>>>>>>> "version": ""},
{"check_multi": false, "name": "origin-master",
"version":
>>>>>>>>>> ""}, {"check_multi": false,
"name": "origin-node", "version": ""}],
>>>>>>>>>> "package_mgr": "yum"}}}},
"msg": "One or more checks failed",
>>>>>>>>>> "playbook_context":
"install"}
>>>>>>>>>>
>>>>>>>>>> NO MORE HOSTS LEFT
>>>>>>>>>>
*******************************************************************************************************************************************
>>>>>>>>>>
>>>>>>>>>> PLAY RECAP
>>>>>>>>>>
***************************************************************************************************************************************************
>>>>>>>>>> localhost : ok=35 changed=1
unreachable=0
>>>>>>>>>> failed=0 skipped=16 rescued=0
ignored=0
>>>>>>>>>>
master0.xxxxxxx.com : ok=96 changed=6
unreachable=0
>>>>>>>>>> failed=1 skipped=165 rescued=0 ignored=0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> INSTALLER STATUS
>>>>>>>>>>
*********************************************************************************************************************************************
>>>>>>>>>> Initialization : Complete (0:00:16)
>>>>>>>>>> Health Check : In Progress (0:00:36)
>>>>>>>>>> This phase can be restarted by running:
>>>>>>>>>> playbooks/openshift-checks/pre-install.yml
>>>>>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151)
>>>>>>>>>> 0:01:56.339 ************
>>>>>>>>>>
>>>>>>>>>>
===============================================================================
>>>>>>>>>> Run health checks (install) - EL
>>>>>>>>>>
---------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 36.15s
>>>>>>>>>> os_firewall : Ensure iptables services are not
enabled
>>>>>>>>>>
------------------------------------------------------------------------------------------------
>>>>>>>>>> 2.74s
>>>>>>>>>> openshift_repos : Ensure libselinux-python is
installed
>>>>>>>>>>
-----------------------------------------------------------------------------------------------
>>>>>>>>>> 1.77s
>>>>>>>>>> openshift_repos : refresh cache
>>>>>>>>>>
-----------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.60s
>>>>>>>>>> Gather Cluster facts
>>>>>>>>>>
----------------------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.51s
>>>>>>>>>> container_runtime : Fixup SELinux permissions for
docker
>>>>>>>>>>
----------------------------------------------------------------------------------------------
>>>>>>>>>> 1.33s
>>>>>>>>>> container_runtime : Place
additional/blocked/insecure registries
>>>>>>>>>> in /etc/containers/registries.conf
>>>>>>>>>>
--------------------------------------------------- 1.30s
>>>>>>>>>> Ensure openshift-ansible installer package deps
are installed
>>>>>>>>>>
-----------------------------------------------------------------------------------------
>>>>>>>>>> 1.29s
>>>>>>>>>> container_runtime : Install Docker
>>>>>>>>>>
--------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.17s
>>>>>>>>>> Initialize openshift.node.sdn_mtu
>>>>>>>>>>
---------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.13s
>>>>>>>>>> os_firewall : Install firewalld packages
>>>>>>>>>>
--------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.13s
>>>>>>>>>> container_runtime : Set various Docker options
>>>>>>>>>>
--------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.11s
>>>>>>>>>> install NetworkManager
>>>>>>>>>>
--------------------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.10s
>>>>>>>>>> openshift_repos : Configure correct origin
release repository
>>>>>>>>>>
-----------------------------------------------------------------------------------------
>>>>>>>>>> 1.05s
>>>>>>>>>> container_runtime : Get current installed Docker
version
>>>>>>>>>>
----------------------------------------------------------------------------------------------
>>>>>>>>>> 1.04s
>>>>>>>>>> openshift_repos : Configure origin gpg keys
>>>>>>>>>>
-----------------------------------------------------------------------------------------------------------
>>>>>>>>>> 1.04s
>>>>>>>>>> openshift_repos : Remove
openshift_additional.repo file
>>>>>>>>>>
-----------------------------------------------------------------------------------------------
>>>>>>>>>> 0.99s
>>>>>>>>>> container_runtime : Setup the docker-storage for
overlay
>>>>>>>>>>
----------------------------------------------------------------------------------------------
>>>>>>>>>> 0.96s
>>>>>>>>>> Detecting Operating System from ostree_booted
>>>>>>>>>>
---------------------------------------------------------------------------------------------------------
>>>>>>>>>> 0.95s
>>>>>>>>>> Gather Cluster facts
>>>>>>>>>>
----------------------------------------------------------------------------------------------------------------------------------
>>>>>>>>>> 0.92s
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Failure summary:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 1. Hosts:
master0.xxxxxxx.com
>>>>>>>>>> Play: OpenShift Health Checks
>>>>>>>>>> Task: Run health checks (install) - EL
>>>>>>>>>> Message: One or more checks failed
>>>>>>>>>> Details: check "docker_storage":
>>>>>>>>>> Could not find imported module
support code for
>>>>>>>>>> docker_info. Looked for either
AnsibleDockerClient.py or docker_common.py
>>>>>>>>>> Traceback (most recent call
last):
>>>>>>>>>> File
>>>>>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py",
>>>>>>>>>> line 225, in run_check
>>>>>>>>>> result = check.run()
>>>>>>>>>> File
>>>>>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py",
>>>>>>>>>> line 53, in run
>>>>>>>>>> docker_info =
>>>>>>>>>> self.execute_module("docker_info", {})
>>>>>>>>>> File
>>>>>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py",
>>>>>>>>>> line 211, in execute_module
>>>>>>>>>> result =
self._execute_module(module_name,
>>>>>>>>>> module_args, self.tmp, self.task_vars)
>>>>>>>>>> File
>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line
>>>>>>>>>> 809, in _execute_module
>>>>>>>>>> (module_style, shebang,
module_data,
>>>>>>>>>> module_path) =
self._configure_module(module_name=module_name,
>>>>>>>>>> module_args=module_args, task_vars=task_vars)
>>>>>>>>>> File
>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line
>>>>>>>>>> 203, in _configure_module
>>>>>>>>>>
environment=final_environment)
>>>>>>>>>> File
>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>>>>>> 1023, in modify_module
>>>>>>>>>> environment=environment)
>>>>>>>>>> File
>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>>>>>> 859, in _find_module_utils
>>>>>>>>>> recursive_finder(module_name,
b_module_data,
>>>>>>>>>> py_module_names, py_module_cache, zf)
>>>>>>>>>> File
>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>>>>>> 621, in recursive_finder
>>>>>>>>>> raise AnsibleError('
'.join(msg))
>>>>>>>>>> AnsibleError: Could not find
imported module
>>>>>>>>>> support code for docker_info. Looked for either
AnsibleDockerClient.py or
>>>>>>>>>> docker_common.py
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> The execution of "install_okd.yaml"
includes checks designed to
>>>>>>>>>> fail early if the requirements of the playbook
are not met. One or more of
>>>>>>>>>> these checks failed. To disregard these
results,explicitly disable checks
>>>>>>>>>> by setting an Ansible variable:
>>>>>>>>>> openshift_disable_check=docker_storage
>>>>>>>>>> Failing check names are shown in the failure
details above. Some
>>>>>>>>>> checks may be configurable by variables if your
requirements are different
>>>>>>>>>> from the defaults; consult check documentation.
>>>>>>>>>> Variables can be set in the inventory or passed
on the command
>>>>>>>>>> line using the -e flag to ansible-playbook.
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>>>>>> To unsubscribe send an email to
users-leave(a)ovirt.org
>>>>>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>>>>>> oVirt Code of Conduct:
>>>>>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>>> List Archives:
>>>>>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G...
>>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>> Users mailing list -- users(a)ovirt.org
>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct:
>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>> List Archives:
>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CN6G76Z7MVN...
>>>>>
>>>>