I managed to get past that but am running in to another problem later in
the process on the control plane pods to appear task. I thought perhaps a
glitch in the process from the failed docker step previously so after a few
more runs I tried killing everything and restarting the metrics process
again from the very beginning and end up hitting the same issue with
control plane pods even though all other steps/tasks seem to be working.
I'm just getting this:
TASK [openshift_control_plane : Wait for control plane pods to appear]
*********
Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857
************
FAILED - RETRYING: Wait for control plane pods to appear (60 retries left).
FAILED - RETRYING: Wait for control plane pods to appear (59 retries left).
FAILED - RETRYING: Wait for control plane pods to appear (58 retries left).
FAILED - RETRYING: Wait for control plane pods to appear (57 retries left).
FAILED - RETRYING: Wait for control plane pods to appear (56 retries left).
It eventually counts all the way down to zero and fails.
In syslog of the master0 server I'm seeing some errors related to cni config
May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None kind=pod
force=False all_namespaces=None field_selector=None namespace=kube-system
delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig
content=None state=list debug=False selector=None name=
May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150
cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d
May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150
kubelet.go:2101] Container runtime network not ready: NetworkReady=false
reason:NetworkPluginNotReady message:docker: network plugin is not ready:
cni config uninitialized
May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None kind=pod
force=False all_namespaces=None field_selector=None namespace=kube-system
delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig
content=None state=list debug=False selector=None name=
May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150
cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d
May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150
kubelet.go:2101] Container runtime network not ready: NetworkReady=false
reason:NetworkPluginNotReady message:docker: network plugin is not ready:
cni config uninitialized
May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150
summary.go:102] Failed to get system container stats for
"/system.slice/origin-node.service": failed to get cgroup stats for
"/system.slice/origin-node.service": failed to get container info for
"/system.slice/origin-node.service": unknown container
"/system.slice/origin-node.service"
May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150
cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d
May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150
kubelet.go:2101] Container runtime network not ready: NetworkReady=false
reason:NetworkPluginNotReady message:docker: network plugin is not ready:
cni config uninitialized
On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sradco(a)redhat.com> wrote:
Hi Jayme,
Thank you for reaching out.
Please try rerunning the ansible playbook.
If this doesn't work, try adding to the integ.ini in the metrics vm
openshift_disable_check=docker_storage
and rerun the ansible playbook again.
Please update how it goes.
Best regards,
--
Shirly Radco
BI Senior Software Engineer
Red Hat <
https://www.redhat.com/>
<
https://www.redhat.com/>
On Sun, May 26, 2019 at 9:34 PM Jayme <jaymef(a)gmail.com> wrote:
> I'm running in to this ansible error during oVirt metrics installation
> (following procedures at:
>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_...
> )
>
> This is happening late in the process, after successfully deploying the
> installation VM and then running second step from the metrics VM.
>
> CHECK [memory_availability :
master0.xxxxxx.com]
>
*************************************************************************************************
> fatal: [
master0.xxxxxxx.com]: FAILED! => {"changed": true,
"checks":
> {"disk_availability": {}, "docker_image_availability":
{"changed": true},
> "docker_storage": {"failed": true, "failures":
[["OpenShiftCheckException",
> "Could not find imported module support code for docker_info. Looked for
> either AnsibleDockerClient.py or docker_common.py\nTraceback (most recent
> call last):\n File
>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
> line 225, in run_check\n result = check.run()\n File
>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
> line 53, in run\n docker_info = self.execute_module(\"docker_info\",
> {})\n File
>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
> line 211, in execute_module\n result = self._execute_module(module_name,
> module_args, self.tmp, self.task_vars)\n File
> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
> line 809, in _execute_module\n (module_style, shebang, module_data,
> module_path) = self._configure_module(module_name=module_name,
> module_args=module_args, task_vars=task_vars)\n File
> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
> line 203, in _configure_module\n environment=final_environment)\n File
> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
> line 1023, in modify_module\n environment=environment)\n File
> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
> line 859, in _find_module_utils\n recursive_finder(module_name,
> b_module_data, py_module_names, py_module_cache, zf)\n File
> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
> line 621, in recursive_finder\n raise AnsibleError('
> '.join(msg))\nAnsibleError: Could not find imported module support code for
> docker_info. Looked for either AnsibleDockerClient.py or
> docker_common.py\n"]], "msg": "Could not find imported module
support code
> for docker_info. Looked for either AnsibleDockerClient.py or
> docker_common.py\nTraceback (most recent call last):\n File
>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
> line 225, in run_check\n result = check.run()\n File
>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
> line 53, in run\n docker_info = self.execute_module(\"docker_info\",
> {})\n File
>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
> line 211, in execute_module\n result = self._execute_module(module_name,
> module_args, self.tmp, self.task_vars)\n File
> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
> line 809, in _execute_module\n (module_style, shebang, module_data,
> module_path) = self._configure_module(module_name=module_name,
> module_args=module_args, task_vars=task_vars)\n File
> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
> line 203, in _configure_module\n environment=final_environment)\n File
> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
> line 1023, in modify_module\n environment=environment)\n File
> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
> line 859, in _find_module_utils\n recursive_finder(module_name,
> b_module_data, py_module_names, py_module_cache, zf)\n File
> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
> line 621, in recursive_finder\n raise AnsibleError('
> '.join(msg))\nAnsibleError: Could not find imported module support code for
> docker_info. Looked for either AnsibleDockerClient.py or
> docker_common.py\n"}, "memory_availability": {},
"package_availability":
> {"changed": false, "invocation": {"module_args":
{"packages": ["PyYAML",
> "bash-completion", "bind", "ceph-common",
"dnsmasq", "docker", "firewalld",
> "flannel", "glusterfs-fuse", "httpd-tools",
"iptables",
> "iptables-services", "iscsi-initiator-utils",
"libselinux-python",
> "nfs-utils", "ntp", "openssl", "origin",
"origin-clients",
> "origin-hyperkube", "origin-node", "pyparted",
"python-httplib2",
> "yum-utils"]}}}, "package_version": {"changed": false,
"invocation":
> {"module_args": {"package_list": [{"check_multi":
false, "name": "origin",
> "version": ""}, {"check_multi": false,
"name": "origin-master", "version":
> ""}, {"check_multi": false, "name":
"origin-node", "version": ""}],
> "package_mgr": "yum"}}}}, "msg": "One or more
checks failed",
> "playbook_context": "install"}
>
> NO MORE HOSTS LEFT
>
*******************************************************************************************************************************************
>
> PLAY RECAP
>
***************************************************************************************************************************************************
> localhost : ok=35 changed=1 unreachable=0
> failed=0 skipped=16 rescued=0 ignored=0
>
master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 failed=1
> skipped=165 rescued=0 ignored=0
>
>
> INSTALLER STATUS
>
*********************************************************************************************************************************************
> Initialization : Complete (0:00:16)
> Health Check : In Progress (0:00:36)
> This phase can be restarted by running:
> playbooks/openshift-checks/pre-install.yml
> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) 0:01:56.339
> ************
>
> ===============================================================================
> Run health checks (install) - EL
>
---------------------------------------------------------------------------------------------------------------------
> 36.15s
> os_firewall : Ensure iptables services are not enabled
>
------------------------------------------------------------------------------------------------
> 2.74s
> openshift_repos : Ensure libselinux-python is installed
>
-----------------------------------------------------------------------------------------------
> 1.77s
> openshift_repos : refresh cache
>
-----------------------------------------------------------------------------------------------------------------------
> 1.60s
> Gather Cluster facts
>
----------------------------------------------------------------------------------------------------------------------------------
> 1.51s
> container_runtime : Fixup SELinux permissions for docker
>
----------------------------------------------------------------------------------------------
> 1.33s
> container_runtime : Place additional/blocked/insecure registries in
> /etc/containers/registries.conf
> --------------------------------------------------- 1.30s
> Ensure openshift-ansible installer package deps are installed
>
-----------------------------------------------------------------------------------------
> 1.29s
> container_runtime : Install Docker
>
--------------------------------------------------------------------------------------------------------------------
> 1.17s
> Initialize openshift.node.sdn_mtu
>
---------------------------------------------------------------------------------------------------------------------
> 1.13s
> os_firewall : Install firewalld packages
>
--------------------------------------------------------------------------------------------------------------
> 1.13s
> container_runtime : Set various Docker options
>
--------------------------------------------------------------------------------------------------------
> 1.11s
> install NetworkManager
>
--------------------------------------------------------------------------------------------------------------------------------
> 1.10s
> openshift_repos : Configure correct origin release repository
>
-----------------------------------------------------------------------------------------
> 1.05s
> container_runtime : Get current installed Docker version
>
----------------------------------------------------------------------------------------------
> 1.04s
> openshift_repos : Configure origin gpg keys
>
-----------------------------------------------------------------------------------------------------------
> 1.04s
> openshift_repos : Remove openshift_additional.repo file
>
-----------------------------------------------------------------------------------------------
> 0.99s
> container_runtime : Setup the docker-storage for overlay
>
----------------------------------------------------------------------------------------------
> 0.96s
> Detecting Operating System from ostree_booted
>
---------------------------------------------------------------------------------------------------------
> 0.95s
> Gather Cluster facts
>
----------------------------------------------------------------------------------------------------------------------------------
> 0.92s
>
>
> Failure summary:
>
>
> 1. Hosts:
master0.xxxxxxx.com
> Play: OpenShift Health Checks
> Task: Run health checks (install) - EL
> Message: One or more checks failed
> Details: check "docker_storage":
> Could not find imported module support code for
> docker_info. Looked for either AnsibleDockerClient.py or docker_common.py
> Traceback (most recent call last):
> File
>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py",
> line 225, in run_check
> result = check.run()
> File
>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py",
> line 53, in run
> docker_info = self.execute_module("docker_info", {})
> File
>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py",
> line 211, in execute_module
> result = self._execute_module(module_name,
> module_args, self.tmp, self.task_vars)
> File
> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py",
line
> 809, in _execute_module
> (module_style, shebang, module_data, module_path) =
> self._configure_module(module_name=module_name, module_args=module_args,
> task_vars=task_vars)
> File
> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py",
line
> 203, in _configure_module
> environment=final_environment)
> File
> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
> 1023, in modify_module
> environment=environment)
> File
> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
> 859, in _find_module_utils
> recursive_finder(module_name, b_module_data,
> py_module_names, py_module_cache, zf)
> File
> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
> 621, in recursive_finder
> raise AnsibleError(' '.join(msg))
> AnsibleError: Could not find imported module support code
> for docker_info. Looked for either AnsibleDockerClient.py or
> docker_common.py
>
>
> The execution of "install_okd.yaml" includes checks designed to fail
> early if the requirements of the playbook are not met. One or more of these
> checks failed. To disregard these results,explicitly disable checks by
> setting an Ansible variable:
> openshift_disable_check=docker_storage
> Failing check names are shown in the failure details above. Some checks
> may be configurable by variables if your requirements are different from
> the defaults; consult check documentation.
> Variables can be set in the inventory or passed on the command line using
> the -e flag to ansible-playbook.
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G...
>