Do you mean the iptables firewall on the server being installed to i.e.
master0 or the actual oVirt host that the master0 VM is running on? I did
try flushing iptables rules on master0 VM then ran plays again from
installer VM but fail at the same point.
Does this log message have anything to do with the issue, /etc/cni
directory does not even exist on master0 VM.
May 28 17:23:35 master0 origin-node: W0528 17:23:35.012902 10434
cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d
May 28 17:23:35 master0 origin-node: E0528 17:23:35.013398 10434
kubelet.go:2101] Container runtime network not ready: NetworkReady=false
reason:NetworkPluginNotReady message:docker: network plugin is not ready:
cni config uninitialized
On Tue, May 28, 2019 at 1:19 PM Edward Berger <edwberger(a)gmail.com> wrote:
> TASK [openshift_control_plane : Wait for control plane pods to
appear]
*********
> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857
************
> FAILED - RETRYING: Wait for control plane pods to appear (60 retries
left).
> FAILED - RETRYING: Wait for control plane pods to appear (59 retries
left).
> It eventually counts all the way down to zero and fails.
This looks a lot like the issues I saw when the host firewall (iptables)
was blocking another OKD all-in-one-host install script [1].
Disabling iptables allowed the installation to continue for my proof of
concept "cluster".
[
1]https://github.com/gshipley/installcentos
The other error I had with [1] was it was trying to install a couple of
packages (zile and python2-pip) from EPEL with the repo disabled.
On Tue, May 28, 2019 at 10:41 AM Jayme <jaymef(a)gmail.com> wrote:
> Shirly,
>
> Oh and I should mention that I did verify that NetworkManager was
> installed on the master0 VM and enabled/started the second go around. So
> that service is there and running.
>
> # systemctl list-unit-files | grep Network
> dbus-org.freedesktop.NetworkManager.service
> enabled
> NetworkManager-dispatcher.service
> enabled
> NetworkManager-wait-online.service
> enabled
> NetworkManager.service
> enabled
>
> On Tue, May 28, 2019 at 11:13 AM Jayme <jaymef(a)gmail.com> wrote:
>
>> Shirly,
>>
>> I appreciate the help with this. Unfortunately I am still running in to
>> the same problem. So far I've tried to install/enable/start NetworkManager
>> on the existing "master0" server and re-ran the plans from the
installer
>> VM. I ran in to the same problem waiting for control plane pods and same
>> errors in syslog.
>>
>> So I wiped everything out, killed the template along with the installer
>> and master VMs. On oVirt engine (I am running 4.3.3.7-1 stable) I did have
>> ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an
>> update check. So I installed
>>
http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then
>> proceeded to install the latest version of ovirt-engine-metrics which gave
>> me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine.
>>
>> After that package was installed I proceeded to follow steps from
>> beginning outlined at:
>>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_... --
>> I ran in to the docker check issue again (same as my initial email) so I
>> disabled that and again got as far as starting control plane pods before
>> failure.
>>
>> Not sure where to go from here at this point. The only thing I can
>> think of that I did differently vs the instructions outlined above is that
>> I have not crated the wildcard DNS record, however I did set configs to
>> create /etc/hosts entries and they /etc/hosts on the machines have the
>> proper IPs assigned for all hostnames (automatically added by the ansible
>> plays).
>>
>> Any ideas how I can get past the plane pods issue?
>>
>> Thanks!
>>
>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <sradco(a)redhat.com> wrote:
>>
>>> Hi,
>>>
>>> The latest release of 4.3.z should already include a fix for this
>>> issue, ovirt-engine-metrics-1.3.1 rpm.
>>>
>>> The issue is that it requires the NetworkManagar to be installed,
>>> running and enabled for it to work.
>>>
>>> You can install it manually on the master0 vm , start and enable it or
>>> you can also install the updated rpm from the nightly builds if your
>>> environment is oVirt 4.2.z:
>>>
>>>
https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-e...
>>>
>>> Relevant patches are:
>>>
https://gerrit.ovirt.org/#/c/99959/
>>>
https://gerrit.ovirt.org/#/c/99718/
>>>
>>> Best regards,
>>>
>>> --
>>>
>>> Shirly Radco
>>>
>>> BI Senior Software Engineer
>>>
>>> Red Hat <
https://www.redhat.com/>
>>>
>>> <
https://www.redhat.com/>
>>>
>>>
>>> On Mon, May 27, 2019 at 4:41 PM Jayme <jaymef(a)gmail.com> wrote:
>>>
>>>> I managed to get past that but am running in to another problem later
>>>> in the process on the control plane pods to appear task. I thought
>>>> perhaps a glitch in the process from the failed docker step previously
so
>>>> after a few more runs I tried killing everything and restarting the
metrics
>>>> process again from the very beginning and end up hitting the same issue
>>>> with control plane pods even though all other steps/tasks seem to be
>>>> working.
>>>>
>>>> I'm just getting this:
>>>>
>>>> TASK [openshift_control_plane : Wait for control plane pods to appear]
>>>> *********
>>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857
>>>> ************
>>>> FAILED - RETRYING: Wait for control plane pods to appear (60 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (59 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 retries
>>>> left).
>>>>
>>>> It eventually counts all the way down to zero and fails.
>>>>
>>>> In syslog of the master0 server I'm seeing some errors related to
cni
>>>> config
>>>>
>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None
>>>> kind=pod force=False all_namespaces=None field_selector=None
>>>> namespace=kube-system delete_after=False
>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list
>>>> debug=False selector=None
name=master-api-master0.xxxxxx.com
>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150
>>>> cni.go:172] Unable to update cni config: No networks found in
/etc/cni/net.d
>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not
ready:
>>>> cni config uninitialized
>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None
>>>> kind=pod force=False all_namespaces=None field_selector=None
>>>> namespace=kube-system delete_after=False
>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list
>>>> debug=False selector=None
name=master-api-master0.xxxxxx.com
>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150
>>>> cni.go:172] Unable to update cni config: No networks found in
/etc/cni/net.d
>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not
ready:
>>>> cni config uninitialized
>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150
>>>> summary.go:102] Failed to get system container stats for
>>>> "/system.slice/origin-node.service": failed to get cgroup stats
for
>>>> "/system.slice/origin-node.service": failed to get container
info for
>>>> "/system.slice/origin-node.service": unknown container
>>>> "/system.slice/origin-node.service"
>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150
>>>> cni.go:172] Unable to update cni config: No networks found in
/etc/cni/net.d
>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not
ready:
>>>> cni config uninitialized
>>>>
>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sradco(a)redhat.com>
>>>> wrote:
>>>>
>>>>> Hi Jayme,
>>>>>
>>>>> Thank you for reaching out.
>>>>> Please try rerunning the ansible playbook.
>>>>> If this doesn't work, try adding to the integ.ini in the metrics
vm
>>>>> openshift_disable_check=docker_storage
>>>>> and rerun the ansible playbook again.
>>>>>
>>>>> Please update how it goes.
>>>>>
>>>>> Best regards,
>>>>>
>>>>> --
>>>>>
>>>>> Shirly Radco
>>>>>
>>>>> BI Senior Software Engineer
>>>>>
>>>>> Red Hat <
https://www.redhat.com/>
>>>>>
>>>>> <
https://www.redhat.com/>
>>>>>
>>>>>
>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <jaymef(a)gmail.com>
wrote:
>>>>>
>>>>>> I'm running in to this ansible error during oVirt metrics
>>>>>> installation (following procedures at:
>>>>>>
https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_...
>>>>>> )
>>>>>>
>>>>>> This is happening late in the process, after successfully
deploying
>>>>>> the installation VM and then running second step from the metrics
VM.
>>>>>>
>>>>>> CHECK [memory_availability :
master0.xxxxxx.com]
>>>>>>
*************************************************************************************************
>>>>>> fatal: [
master0.xxxxxxx.com]: FAILED! => {"changed":
true,
>>>>>> "checks": {"disk_availability": {},
"docker_image_availability":
>>>>>> {"changed": true}, "docker_storage":
{"failed": true, "failures":
>>>>>> [["OpenShiftCheckException", "Could not find
imported module support code
>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\nTraceback (most recent call last):\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>> line 225, in run_check\n result = check.run()\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>> line 53, in run\n docker_info =
self.execute_module(\"docker_info\",
>>>>>> {})\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>> line 211, in execute_module\n result =
self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 809, in _execute_module\n (module_style, shebang,
module_data,
>>>>>> module_path) = self._configure_module(module_name=module_name,
>>>>>> module_args=module_args, task_vars=task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 203, in _configure_module\n
environment=final_environment)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 1023, in modify_module\n environment=environment)\n
File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 859, in _find_module_utils\n
recursive_finder(module_name,
>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 621, in recursive_finder\n raise AnsibleError('
>>>>>> '.join(msg))\nAnsibleError: Could not find imported module
support code for
>>>>>> docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\n"]], "msg": "Could not find
imported module support code
>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\nTraceback (most recent call last):\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>> line 225, in run_check\n result = check.run()\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>> line 53, in run\n docker_info =
self.execute_module(\"docker_info\",
>>>>>> {})\n File
>>>>>>
\"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>> line 211, in execute_module\n result =
self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 809, in _execute_module\n (module_style, shebang,
module_data,
>>>>>> module_path) = self._configure_module(module_name=module_name,
>>>>>> module_args=module_args, task_vars=task_vars)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 203, in _configure_module\n
environment=final_environment)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 1023, in modify_module\n environment=environment)\n
File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 859, in _find_module_utils\n
recursive_finder(module_name,
>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File
>>>>>>
\"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 621, in recursive_finder\n raise AnsibleError('
>>>>>> '.join(msg))\nAnsibleError: Could not find imported module
support code for
>>>>>> docker_info. Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\n"}, "memory_availability": {},
"package_availability":
>>>>>> {"changed": false, "invocation":
{"module_args": {"packages": ["PyYAML",
>>>>>> "bash-completion", "bind",
"ceph-common", "dnsmasq", "docker", "firewalld",
>>>>>> "flannel", "glusterfs-fuse",
"httpd-tools", "iptables",
>>>>>> "iptables-services", "iscsi-initiator-utils",
"libselinux-python",
>>>>>> "nfs-utils", "ntp", "openssl",
"origin", "origin-clients",
>>>>>> "origin-hyperkube", "origin-node",
"pyparted", "python-httplib2",
>>>>>> "yum-utils"]}}}, "package_version":
{"changed": false, "invocation":
>>>>>> {"module_args": {"package_list":
[{"check_multi": false, "name": "origin",
>>>>>> "version": ""}, {"check_multi":
false, "name": "origin-master", "version":
>>>>>> ""}, {"check_multi": false, "name":
"origin-node", "version": ""}],
>>>>>> "package_mgr": "yum"}}}}, "msg":
"One or more checks failed",
>>>>>> "playbook_context": "install"}
>>>>>>
>>>>>> NO MORE HOSTS LEFT
>>>>>>
*******************************************************************************************************************************************
>>>>>>
>>>>>> PLAY RECAP
>>>>>>
***************************************************************************************************************************************************
>>>>>> localhost : ok=35 changed=1 unreachable=0
>>>>>> failed=0 skipped=16 rescued=0 ignored=0
>>>>>>
master0.xxxxxxx.com : ok=96 changed=6 unreachable=0
>>>>>> failed=1 skipped=165 rescued=0 ignored=0
>>>>>>
>>>>>>
>>>>>> INSTALLER STATUS
>>>>>>
*********************************************************************************************************************************************
>>>>>> Initialization : Complete (0:00:16)
>>>>>> Health Check : In Progress (0:00:36)
>>>>>> This phase can be restarted by running:
>>>>>> playbooks/openshift-checks/pre-install.yml
>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151)
0:01:56.339
>>>>>> ************
>>>>>>
>>>>>>
===============================================================================
>>>>>> Run health checks (install) - EL
>>>>>>
---------------------------------------------------------------------------------------------------------------------
>>>>>> 36.15s
>>>>>> os_firewall : Ensure iptables services are not enabled
>>>>>>
------------------------------------------------------------------------------------------------
>>>>>> 2.74s
>>>>>> openshift_repos : Ensure libselinux-python is installed
>>>>>>
-----------------------------------------------------------------------------------------------
>>>>>> 1.77s
>>>>>> openshift_repos : refresh cache
>>>>>>
-----------------------------------------------------------------------------------------------------------------------
>>>>>> 1.60s
>>>>>> Gather Cluster facts
>>>>>>
----------------------------------------------------------------------------------------------------------------------------------
>>>>>> 1.51s
>>>>>> container_runtime : Fixup SELinux permissions for docker
>>>>>>
----------------------------------------------------------------------------------------------
>>>>>> 1.33s
>>>>>> container_runtime : Place additional/blocked/insecure registries
in
>>>>>> /etc/containers/registries.conf
>>>>>> --------------------------------------------------- 1.30s
>>>>>> Ensure openshift-ansible installer package deps are installed
>>>>>>
-----------------------------------------------------------------------------------------
>>>>>> 1.29s
>>>>>> container_runtime : Install Docker
>>>>>>
--------------------------------------------------------------------------------------------------------------------
>>>>>> 1.17s
>>>>>> Initialize openshift.node.sdn_mtu
>>>>>>
---------------------------------------------------------------------------------------------------------------------
>>>>>> 1.13s
>>>>>> os_firewall : Install firewalld packages
>>>>>>
--------------------------------------------------------------------------------------------------------------
>>>>>> 1.13s
>>>>>> container_runtime : Set various Docker options
>>>>>>
--------------------------------------------------------------------------------------------------------
>>>>>> 1.11s
>>>>>> install NetworkManager
>>>>>>
--------------------------------------------------------------------------------------------------------------------------------
>>>>>> 1.10s
>>>>>> openshift_repos : Configure correct origin release repository
>>>>>>
-----------------------------------------------------------------------------------------
>>>>>> 1.05s
>>>>>> container_runtime : Get current installed Docker version
>>>>>>
----------------------------------------------------------------------------------------------
>>>>>> 1.04s
>>>>>> openshift_repos : Configure origin gpg keys
>>>>>>
-----------------------------------------------------------------------------------------------------------
>>>>>> 1.04s
>>>>>> openshift_repos : Remove openshift_additional.repo file
>>>>>>
-----------------------------------------------------------------------------------------------
>>>>>> 0.99s
>>>>>> container_runtime : Setup the docker-storage for overlay
>>>>>>
----------------------------------------------------------------------------------------------
>>>>>> 0.96s
>>>>>> Detecting Operating System from ostree_booted
>>>>>>
---------------------------------------------------------------------------------------------------------
>>>>>> 0.95s
>>>>>> Gather Cluster facts
>>>>>>
----------------------------------------------------------------------------------------------------------------------------------
>>>>>> 0.92s
>>>>>>
>>>>>>
>>>>>> Failure summary:
>>>>>>
>>>>>>
>>>>>> 1. Hosts:
master0.xxxxxxx.com
>>>>>> Play: OpenShift Health Checks
>>>>>> Task: Run health checks (install) - EL
>>>>>> Message: One or more checks failed
>>>>>> Details: check "docker_storage":
>>>>>> Could not find imported module support code for
>>>>>> docker_info. Looked for either AnsibleDockerClient.py or
docker_common.py
>>>>>> Traceback (most recent call last):
>>>>>> File
>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py",
>>>>>> line 225, in run_check
>>>>>> result = check.run()
>>>>>> File
>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py",
>>>>>> line 53, in run
>>>>>> docker_info =
self.execute_module("docker_info",
>>>>>> {})
>>>>>> File
>>>>>>
"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py",
>>>>>> line 211, in execute_module
>>>>>> result = self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line
>>>>>> 809, in _execute_module
>>>>>> (module_style, shebang, module_data,
module_path)
>>>>>> = self._configure_module(module_name=module_name,
module_args=module_args,
>>>>>> task_vars=task_vars)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line
>>>>>> 203, in _configure_module
>>>>>> environment=final_environment)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>> 1023, in modify_module
>>>>>> environment=environment)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>> 859, in _find_module_utils
>>>>>> recursive_finder(module_name, b_module_data,
>>>>>> py_module_names, py_module_cache, zf)
>>>>>> File
>>>>>>
"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line
>>>>>> 621, in recursive_finder
>>>>>> raise AnsibleError(' '.join(msg))
>>>>>> AnsibleError: Could not find imported module
support
>>>>>> code for docker_info. Looked for either AnsibleDockerClient.py
or
>>>>>> docker_common.py
>>>>>>
>>>>>>
>>>>>> The execution of "install_okd.yaml" includes checks
designed to fail
>>>>>> early if the requirements of the playbook are not met. One or
more of these
>>>>>> checks failed. To disregard these results,explicitly disable
checks by
>>>>>> setting an Ansible variable:
>>>>>> openshift_disable_check=docker_storage
>>>>>> Failing check names are shown in the failure details above. Some
>>>>>> checks may be configurable by variables if your requirements are
different
>>>>>> from the defaults; consult check documentation.
>>>>>> Variables can be set in the inventory or passed on the command
line
>>>>>> using the -e flag to ansible-playbook.
>>>>>> _______________________________________________
>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G...
>>>>>>
>>>>> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CN6G76Z7MVN...
>