TASK [openshift_control_plane : Wait for control plane pods to
appear]
*********
Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857
************
FAILED - RETRYING: Wait for control plane pods to appear (60 retries
left).
FAILED - RETRYING: Wait for control plane pods to appear (59 retries
left).
It eventually counts all the way down to zero and fails.
This looks a lot like the issues I saw when the host firewall (iptables)
was blocking another OKD all-in-one-host install script [1].
Disabling iptables allowed the installation to continue for my proof of
concept "cluster".
[
1]https://github.com/gshipley/installcentos
The other error I had with [1] was it was trying to install a couple of
packages (zile and python2-pip) from EPEL with the repo disabled.
On Tue, May 28, 2019 at 10:41 AM Jayme <jaymef(a)gmail.com> wrote:
> Shirly,
>
> Oh and I should mention that I did verify that NetworkManager was
> installed on the master0 VM and enabled/started the second go around. So
> that service is there and running.
>
> # systemctl list-unit-files | grep Network
> dbus-org.freedesktop.NetworkManager.service
> enabled
> NetworkManager-dispatcher.service
> enabled
> NetworkManager-wait-online.service
> enabled
> NetworkManager.service
> enabled
>
> On Tue, May 28, 2019 at 11:13 AM Jayme <jaymef(a)gmail.com> wrote:
>
>> Shirly,
>>
>> I appreciate the help with this. Unfortunately I am still running in to
>> the same problem. So far I've tried to install/enable/start NetworkManager
>> on the existing "master0" server and re-ran the plans from the
installer
>> VM. I ran in to the same problem waiting for control plane pods and same
>> errors in syslog.
>>
>> So I wiped everything out, killed the template along with the installer
>> and master VMs. On oVirt engine (I am running 4.3.3.7-1 stable) I did have
>> ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an
>> update check. So I installed
>>
http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then
>> proceeded to install the latest version of ovirt-engine-metrics which gave
>> me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine.
>>
>> After that package was installed I proceeded to follow steps from
>> beginning outlined at:
>>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_... --
>> I ran in to the docker check issue again (same as my initial email) so I
>> disabled that and again got as far as starting control plane pods before
>> failure.
>>
>> Not sure where to go from here at this point. The only thing I can think
>> of that I did differently vs the instructions outlined above is that I have
>> not crated the wildcard DNS record, however I did set configs to create
>> /etc/hosts entries and they /etc/hosts on the machines have the proper IPs
>> assigned for all hostnames (automatically added by the ansible plays).
>>
>> Any ideas how I can get past the plane pods issue?
>>
>> Thanks!
>>
>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <sradco(a)redhat.com> wrote:
>>
>>> Hi,
>>>
>>> The latest release of 4.3.z should already include a fix for this issue,
>>> ovirt-engine-metrics-1.3.1 rpm.
>>>
>>> The issue is that it requires the NetworkManagar to be installed,
>>> running and enabled for it to work.
>>>
>>> You can install it manually on the master0 vm , start and enable it or
>>> you can also install the updated rpm from the nightly builds if your
>>> environment is oVirt 4.2.z:
>>>
>>>
https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-e...
>>>
>>> Relevant patches are:
>>>
https://gerrit.ovirt.org/#/c/99959/
>>>
https://gerrit.ovirt.org/#/c/99718/
>>>
>>> Best regards,
>>>
>>> --
>>>
>>> Shirly Radco
>>>
>>> BI Senior Software Engineer
>>>
>>> Red Hat <
https://www.redhat.com/>
>>>
>>> <
https://www.redhat.com/>
>>>
>>>
>>> On Mon, May 27, 2019 at 4:41 PM Jayme <jaymef(a)gmail.com> wrote:
>>>
>>>> I managed to get past that but am running in to another problem later
>>>> in the process on the control plane pods to appear task. I thought
>>>> perhaps a glitch in the process from the failed docker step previously
so
>>>> after a few more runs I tried killing everything and restarting the
metrics
>>>> process again from the very beginning and end up hitting the same issue
>>>> with control plane pods even though all other steps/tasks seem to be
>>>> working.
>>>>
>>>> I'm just getting this:
>>>>
>>>
TASK [openshift_control_plane : Wait for
control plane pods to appear]
>>>> *********
>>>
Monday 27 May 2019 13:31:54 +0000
(0:00:00.180) 0:14:33.857
>>>> ************
>>>
FAILED - RETRYING: Wait for control plane
pods to appear (60 retries
>>>> left).
>>>
FAILED - RETRYING: Wait for control plane
pods to appear (59 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 retries
>>>> left).
>>>>
>>>> It eventually counts all the way down to zero and fails.
>>>>
>>>> In syslog of the master0 server I'm seeing some errors related to
cni
>>>> config
>>>>
>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None
>>>> kind=pod force=False all_namespaces=None field_selector=None
>>>> namespace=kube-system delete_after=False
>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list
>>>> debug=False selector=None
name=master-api-master0.xxxxxx.com
>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150
>>>> cni.go:172] Unable to update cni config: No networks found in
/etc/cni/net.d
>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not
ready:
>>>> cni config uninitialized
>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None
>>>> kind=pod force=False all_namespaces=None field_selector=None
>>>> namespace=kube-system delete_after=False
>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list
>>>> debug=False selector=None
name=master-api-master0.xxxxxx.com
>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150
>>>> cni.go:172] Unable to update cni config: No networks found in
/etc/cni/net.d
>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not
ready:
>>>> cni config uninitialized
>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150
>>>> summary.go:102] Failed to get system container stats for
>>>> "/system.slice/origin-node.service": failed to get cgroup stats
for
>>>> "/system.slice/origin-node.service": failed to get container
info for
>>>> "/system.slice/origin-node.service": unknown container
>>>> "/system.slice/origin-node.service"
>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150
>>>> cni.go:172] Unable to update cni config: No networks found in
/etc/cni/net.d
>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not
ready:
>>>> cni config uninitialized
>>>>
>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sradco(a)redhat.com>
wrote:
>>>>
>>>>> Hi Jayme,
>>>>>
>>>>> Thank you for reaching out.
>>>>> Please try rerunning the ansible playbook.
>>>>> If this doesn't work, try adding to the integ.ini in the metrics
vm
>>>>> openshift_disable_check=docker_storage
>>>>> and rerun the ansible playbook again.
>>>>>
>>>>> Please update how it goes.
>>>>>
>>>>> Best regards,
>>>>>
>>>>> --
>>>>>
>>>>> Shirly Radco
>>>>>
>>>>> BI Senior Software Engineer
>>>>>
>>>>> Red Hat <
https://www.redhat.com/>
>>>>>
>>>>> <
https://www.redhat.com/>
>>>>>
>>>>>
>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <jaymef(a)gmail.com>
wrote:
>>>>>
>>>>>> I'm running in to this ansible error during oVirt metrics
>>>>>> installation (following procedures at:
>>>>>>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_...
>>>>>> )
>>>>>>
>>>>>> This is happening late in the process, after successfully
deploying
>>>>>> the installation VM and then running second step from the metrics
VM.
>>>>>>
>>>>>> CHECK [memory_availability :
master0.xxxxxx.com]
>>>>>>
*************************************************************************************************
>>>>>> fatal: [
master0.xxxxxxx.com]: FAILED! => {"changed":
true, "checks":
>>>>>> {"disk_availability": {},
"docker_image_availability": {"changed": true},
>>>>>> "docker_storage": {"failed": true,
"failures": [["OpenShiftCheckException",
>>>>>> "Could not find imported module support code for
docker_info. Looked for
>>>>>> either AnsibleDockerClient.py or docker_common.py\nTraceback
(most recent
>>>>>> call last):\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>> line 225, in run_check\n result = check.run()\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>> line 53, in run\n docker_info =
self.execute_module(\"docker_info\",
>>>>>> {})\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>> line 211, in execute_module\n result =
self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 809, in _execute_module\n (module_style, shebang,
module_data,
>>>>>> module_path) = self._configure_module(module_name=module_name,
>>>>>> module_args=module_args, task_vars=task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 203, in _configure_module\n
environment=final_environment)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 1023, in modify_module\n environment=environment)\n
File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 859, in _find_module_utils\n
recursive_finder(module_name,
>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 621, in recursive_finder\n raise AnsibleError('
>>>>>> '.join(msg))\nAnsibleError: Could not find imported module
support code for
>>>>>> docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\n"]], "msg": "Could not find
imported module support code
>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\nTraceback (most recent call last):\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>> line 225, in run_check\n result = check.run()\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>> line 53, in run\n docker_info =
self.execute_module(\"docker_info\",
>>>>>> {})\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>> line 211, in execute_module\n result =
self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 809, in _execute_module\n (module_style, shebang,
module_data,
>>>>>> module_path) = self._configure_module(module_name=module_name,
>>>>>> module_args=module_args, task_vars=task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 203, in _configure_module\n
environment=final_environment)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 1023, in modify_module\n environment=environment)\n
File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 859, in _find_module_utils\n
recursive_finder(module_name,
>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 621, in recursive_finder\n raise AnsibleError('
>>>>>> '.join(msg))\nAnsibleError: Could not find imported module
support code for
>>>>>> docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\n"}, "memory_availability": {},
"package_availability":
>>>>>> {"changed": false, "invocation":
{"module_args": {"packages": ["PyYAML",
>>>>>> "bash-completion", "bind",
"ceph-common", "dnsmasq", "docker", "firewalld",
>>>>>> "flannel", "glusterfs-fuse",
"httpd-tools", "iptables",
>>>>>> "iptables-services", "iscsi-initiator-utils",
"libselinux-python",
>>>>>> "nfs-utils", "ntp", "openssl",
"origin", "origin-clients",
>>>>>> "origin-hyperkube", "origin-node",
"pyparted", "python-httplib2",
>>>>>> "yum-utils"]}}}, "package_version":
{"changed": false, "invocation":
>>>>>> {"module_args": {"package_list":
[{"check_multi": false, "name": "origin",
>>>>>> "version": ""}, {"check_multi":
false, "name": "origin-master", "version":
>>>>>> ""}, {"check_multi": false, "name":
"origin-node", "version": ""}],
>>>>>> "package_mgr": "yum"}}}}, "msg":
"One or more checks failed",
>>>>>> "playbook_context": "install"}
>>>>>>
>>>>>> NO MORE HOSTS LEFT
>>>>>>
*******************************************************************************************************************************************
>>>>>>
>>>>>> PLAY RECAP
>>>>>>
***************************************************************************************************************************************************
>>>>>> localhost : ok=35 changed=1 unreachable=0
>>>>>> failed=0 skipped=16 rescued=0 ignored=0
>>>>>>
master0.xxxxxxx.com : ok=96 changed=6 unreachable=0
failed=1
>>>>>> skipped=165 rescued=0 ignored=0
>>>>>>
>>>>>>
>>>>>> INSTALLER STATUS
>>>>>>
*********************************************************************************************************************************************
>>>>>> Initialization : Complete (0:00:16)
>>>>>> Health Check : In Progress (0:00:36)
>>>>>> This phase can be restarted by running:
>>>>>> playbooks/openshift-checks/pre-install.yml
>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151)
0:01:56.339
>>>>>> ************
>>>>>>
>>>>>>
===============================================================================
>>>>>> Run health checks (install) - EL
>>>>>>
---------------------------------------------------------------------------------------------------------------------
>>>>>> 36.15s
>>>>>> os_firewall : Ensure iptables services are not enabled
>>>>>>
------------------------------------------------------------------------------------------------
>>>>>> 2.74s
>>>>>> openshift_repos : Ensure libselinux-python is installed
>>>>>>
-----------------------------------------------------------------------------------------------
>>>>>> 1.77s
>>>>>> openshift_repos : refresh cache
>>>>>>
-----------------------------------------------------------------------------------------------------------------------
>>>>>> 1.60s
>>>>>> Gather Cluster facts
>>>>>>
----------------------------------------------------------------------------------------------------------------------------------
>>>>>> 1.51s
>>>>>> container_runtime : Fixup SELinux permissions for docker
>>>>>>
----------------------------------------------------------------------------------------------
>>>>>> 1.33s
>>>>>> container_runtime : Place additional/blocked/insecure registries
in
>>>>>> /etc/containers/registries.conf
>>>>>> --------------------------------------------------- 1.30s
>>>>>> Ensure openshift-ansible installer package deps are installed
>>>>>>
-----------------------------------------------------------------------------------------
>>>>>> 1.29s
>>>>>> container_runtime : Install Docker
>>>>>>
--------------------------------------------------------------------------------------------------------------------
>>>>>> 1.17s
>>>>>> Initialize openshift.node.sdn_mtu
>>>>>>
---------------------------------------------------------------------------------------------------------------------
>>>>>> 1.13s
>>>>>> os_firewall : Install firewalld packages
>>>>>>
--------------------------------------------------------------------------------------------------------------
>>>>>> 1.13s
>>>>>> container_runtime : Set various Docker options
>>>>>>
--------------------------------------------------------------------------------------------------------
>>>>>> 1.11s
>>>>>> install NetworkManager
>>>>>>
--------------------------------------------------------------------------------------------------------------------------------
>>>>>> 1.10s
>>>>>> openshift_repos : Configure correct origin release repository
>>>>>>
-----------------------------------------------------------------------------------------
>>>>>> 1.05s
>>>>>> container_runtime : Get current installed Docker version
>>>>>>
----------------------------------------------------------------------------------------------
>>>>>> 1.04s
>>>>>> openshift_repos : Configure origin gpg keys
>>>>>>
-----------------------------------------------------------------------------------------------------------
>>>>>> 1.04s
>>>>>> openshift_repos : Remove openshift_additional.repo file
>>>>>>
-----------------------------------------------------------------------------------------------
>>>>>> 0.99s
>>>>>> container_runtime : Setup the docker-storage for overlay
>>>>>>
----------------------------------------------------------------------------------------------
>>>>>> 0.96s
>>>>>> Detecting Operating System from ostree_booted
>>>>>>
---------------------------------------------------------------------------------------------------------
>>>>>> 0.95s
>>>>>> Gather Cluster facts
>>>>>>
----------------------------------------------------------------------------------------------------------------------------------
>>>>>> 0.92s
>>>>>>
>>>>>>
>>>>>> Failure summary:
>>>>>>
>>>>>>
>>>>>> 1. Hosts:
master0.xxxxxxx.com
>>>>>> Play: OpenShift Health Checks
>>>>>> Task: Run health checks (install) - EL
>>>>>> Message: One or more checks failed
>>>>>> Details: check "docker_storage":
>>>>>> Could not find imported module support code for
>>>>>> docker_info. Looked for either AnsibleDockerClient.py or
docker_common.py
>>>>>> Traceback (most recent call last):
>>>>>> File
>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py",
>>>>>> line 225, in run_check
>>>>>> result = check.run()
>>>>>> File
>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py",
>>>>>> line 53, in run
>>>>>> docker_info =
self.execute_module("docker_info",
>>>>>> {})
>>>>>> File
>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py",
>>>>>> line 211, in execute_module
>>>>>> result = self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line
>>>>>> 809, in _execute_module
>>>>>> (module_style, shebang, module_data,
module_path)
>>>>>> = self._configure_module(module_name=module_name,
module_args=module_args,
>>>>>> task_vars=task_vars)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line
>>>>>> 203, in _configure_module
>>>>>> environment=final_environment)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>> 1023, in modify_module
>>>>>> environment=environment)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>> 859, in _find_module_utils
>>>>>> recursive_finder(module_name, b_module_data,
>>>>>> py_module_names, py_module_cache, zf)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>> 621, in recursive_finder
>>>>>> raise AnsibleError(' '.join(msg))
>>>>>> AnsibleError: Could not find imported module
support
>>>>>> code for docker_info. Looked for either AnsibleDockerClient.py
or
>>>>>> docker_common.py
>>>>>>
>>>>>>
>>>>>> The execution of "install_okd.yaml" includes checks
designed to fail
>>>>>> early if the requirements of the playbook are not met. One or
more of these
>>>>>> checks failed. To disregard these results,explicitly disable
checks by
>>>>>> setting an Ansible variable:
>>>>>> openshift_disable_check=docker_storage
>>>>>> Failing check names are shown in the failure details above. Some
>>>>>> checks may be configurable by variables if your requirements are
different
>>>>>> from the defaults; consult check documentation.
>>>>>> Variables can be set in the inventory or passed on the command
line
>>>>>> using the -e flag to ansible-playbook.
>>>>>> _______________________________________________
>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G...
>>>>>>
>>>>> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CN6G76Z7MVN...
>