I actually see the pods running on master0 if I do this: @master0 master]# oc project kube-system Now using project "kube-system" on server " https://openshift-master.cloud.xxxxxxx.com:8443". [root@master0 master]# oc get pods NAME READY STATUS RESTARTS AGE master-api-master0.cloud.xxxxxxxx.com 1/1 Running 0 22m master-controllers-master0.cloud.xxxxxxxx.com 1/1 Running 0 22m master-etcd-master0.cloud.xxxxxxxxxx 1/1 Running 0 22m
So I wonder why the ansible "Wait for control plane pods to appear" task is looping - name: Wait for control plane pods to appear oc_obj: state: list kind: pod name: "master-{{ item }}-{{ l_kubelet_node_name | lower }}" namespace: kube-system register: control_plane_pods until: - "'results' in control_plane_pods" - "'results' in control_plane_pods.results" - control_plane_pods.results.results | length > 0 retries: 60 delay: 5 with_items: - "{{ 'etcd' if inventory_hostname in groups['oo_etcd_to_config'] else omit }}" - api - controllers ignore_errors: true On Tue, May 28, 2019 at 4:23 PM Jayme <jay...@gmail.com> wrote: > I just tried again from scratch this time making sure a proper wildcard > DNS entry existed and without using the set /etc/hosts option and am still > running in to the pods issue. Can anyone confirm if this requires a public > external IP to work? I am working on an internal DNS zone here and natted > ips. > > On Tue, May 28, 2019 at 3:28 PM Edward Berger <edwber...@gmail.com> wrote: > >> In my case it was a single bare metal host, so that would be equivalent >> to disabling iptables on the master0 VM you're installing to, in your ovirt >> scenario. >> >> On Tue, May 28, 2019 at 1:25 PM Jayme <jay...@gmail.com> wrote: >> >>> Do you mean the iptables firewall on the server being installed to i.e. >>> master0 or the actual oVirt host that the master0 VM is running on? I did >>> try flushing iptables rules on master0 VM then ran plays again from >>> installer VM but fail at the same point. >>> >>> Does this log message have anything to do with the issue, /etc/cni >>> directory does not even exist on master0 VM. >>> >>> May 28 17:23:35 master0 origin-node: W0528 17:23:35.012902 10434 >>> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >>> May 28 17:23:35 master0 origin-node: E0528 17:23:35.013398 10434 >>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>> cni config uninitialized >>> >>> >>> >>> On Tue, May 28, 2019 at 1:19 PM Edward Berger <edwber...@gmail.com> >>> wrote: >>> >>>> > TASK [openshift_control_plane : Wait for control plane pods to >>>> appear] ********* >>>> > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>> ************ >>>> > FAILED - RETRYING: Wait for control plane pods to appear (60 retries >>>> left). >>>> > FAILED - RETRYING: Wait for control plane pods to appear (59 retries >>>> left). >>>> > It eventually counts all the way down to zero and fails. >>>> >>>> This looks a lot like the issues I saw when the host firewall >>>> (iptables) was blocking another OKD all-in-one-host install script [1]. >>>> Disabling iptables allowed the installation to continue for my proof of >>>> concept "cluster". >>>> >>>> [1]https://github.com/gshipley/installcentos >>>> >>>> The other error I had with [1] was it was trying to install a couple of >>>> packages (zile and python2-pip) from EPEL with the repo disabled. >>>> >>>> >>>> >>>> On Tue, May 28, 2019 at 10:41 AM Jayme <jay...@gmail.com> wrote: >>>> >>>>> Shirly, >>>>> >>>>> Oh and I should mention that I did verify that NetworkManager was >>>>> installed on the master0 VM and enabled/started the second go around. So >>>>> that service is there and running. >>>>> >>>>> # systemctl list-unit-files | grep Network >>>>> dbus-org.freedesktop.NetworkManager.service >>>>> enabled >>>>> NetworkManager-dispatcher.service >>>>> enabled >>>>> NetworkManager-wait-online.service >>>>> enabled >>>>> NetworkManager.service >>>>> enabled >>>>> >>>>> On Tue, May 28, 2019 at 11:13 AM Jayme <jay...@gmail.com> wrote: >>>>> >>>>>> Shirly, >>>>>> >>>>>> I appreciate the help with this. Unfortunately I am still running in >>>>>> to the same problem. So far I've tried to install/enable/start >>>>>> NetworkManager on the existing "master0" server and re-ran the plans from >>>>>> the installer VM. I ran in to the same problem waiting for control plane >>>>>> pods and same errors in syslog. >>>>>> >>>>>> So I wiped everything out, killed the template along with the >>>>>> installer and master VMs. On oVirt engine (I am running 4.3.3.7-1 >>>>>> stable) >>>>>> I did have ovirt-engine-metrics-1.3.0x rpm installed, no yum updates >>>>>> available on an update check. So I installed >>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then >>>>>> proceeded to install the latest version of ovirt-engine-metrics which >>>>>> gave >>>>>> me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. >>>>>> >>>>>> After that package was installed I proceeded to follow steps from >>>>>> beginning outlined at: >>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>> -- >>>>>> I ran in to the docker check issue again (same as my initial email) so I >>>>>> disabled that and again got as far as starting control plane pods before >>>>>> failure. >>>>>> >>>>>> Not sure where to go from here at this point. The only thing I can >>>>>> think of that I did differently vs the instructions outlined above is >>>>>> that >>>>>> I have not crated the wildcard DNS record, however I did set configs to >>>>>> create /etc/hosts entries and they /etc/hosts on the machines have the >>>>>> proper IPs assigned for all hostnames (automatically added by the ansible >>>>>> plays). >>>>>> >>>>>> Any ideas how I can get past the plane pods issue? >>>>>> >>>>>> Thanks! >>>>>> >>>>>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <sra...@redhat.com> >>>>>> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> The latest release of 4.3.z should already include a fix for this >>>>>>> issue, ovirt-engine-metrics-1.3.1 rpm. >>>>>>> >>>>>>> The issue is that it requires the NetworkManagar to be installed, >>>>>>> running and enabled for it to work. >>>>>>> >>>>>>> You can install it manually on the master0 vm , start and enable it >>>>>>> or you can also install the updated rpm from the nightly builds if your >>>>>>> environment is oVirt 4.2.z: >>>>>>> >>>>>>> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm >>>>>>> >>>>>>> Relevant patches are: >>>>>>> https://gerrit.ovirt.org/#/c/99959/ >>>>>>> https://gerrit.ovirt.org/#/c/99718/ >>>>>>> >>>>>>> Best regards, >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Shirly Radco >>>>>>> >>>>>>> BI Senior Software Engineer >>>>>>> >>>>>>> Red Hat <https://www.redhat.com/> >>>>>>> >>>>>>> <https://www.redhat.com/> >>>>>>> >>>>>>> >>>>>>> On Mon, May 27, 2019 at 4:41 PM Jayme <jay...@gmail.com> wrote: >>>>>>> >>>>>>>> I managed to get past that but am running in to another problem >>>>>>>> later in the process on the control plane pods to appear task. I >>>>>>>> thought >>>>>>>> perhaps a glitch in the process from the failed docker step previously >>>>>>>> so >>>>>>>> after a few more runs I tried killing everything and restarting the >>>>>>>> metrics >>>>>>>> process again from the very beginning and end up hitting the same issue >>>>>>>> with control plane pods even though all other steps/tasks seem to be >>>>>>>> working. >>>>>>>> >>>>>>>> I'm just getting this: >>>>>>>> >>>>>>>> TASK [openshift_control_plane : Wait for control plane pods to >>>>>>>> appear] ********* >>>>>>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>>>>>> ************ >>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (60 >>>>>>>> retries left). >>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (59 >>>>>>>> retries left). >>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 >>>>>>>> retries left). >>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 >>>>>>>> retries left). >>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 >>>>>>>> retries left). >>>>>>>> >>>>>>>> It eventually counts all the way down to zero and fails. >>>>>>>> >>>>>>>> In syslog of the master0 server I'm seeing some errors related to >>>>>>>> cni config >>>>>>>> >>>>>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None >>>>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>>>> namespace=kube-system delete_after=False >>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 >>>>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>>>> /etc/cni/net.d >>>>>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 >>>>>>>> kubelet.go:2101] Container runtime network not ready: >>>>>>>> NetworkReady=false >>>>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not >>>>>>>> ready: >>>>>>>> cni config uninitialized >>>>>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None >>>>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>>>> namespace=kube-system delete_after=False >>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 >>>>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>>>> /etc/cni/net.d >>>>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 >>>>>>>> kubelet.go:2101] Container runtime network not ready: >>>>>>>> NetworkReady=false >>>>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not >>>>>>>> ready: >>>>>>>> cni config uninitialized >>>>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 >>>>>>>> summary.go:102] Failed to get system container stats for >>>>>>>> "/system.slice/origin-node.service": failed to get cgroup stats for >>>>>>>> "/system.slice/origin-node.service": failed to get container info for >>>>>>>> "/system.slice/origin-node.service": unknown container >>>>>>>> "/system.slice/origin-node.service" >>>>>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 >>>>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>>>> /etc/cni/net.d >>>>>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 >>>>>>>> kubelet.go:2101] Container runtime network not ready: >>>>>>>> NetworkReady=false >>>>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not >>>>>>>> ready: >>>>>>>> cni config uninitialized >>>>>>>> >>>>>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sra...@redhat.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Jayme, >>>>>>>>> >>>>>>>>> Thank you for reaching out. >>>>>>>>> Please try rerunning the ansible playbook. >>>>>>>>> If this doesn't work, try adding to the integ.ini in the metrics >>>>>>>>> vm >>>>>>>>> openshift_disable_check=docker_storage >>>>>>>>> and rerun the ansible playbook again. >>>>>>>>> >>>>>>>>> Please update how it goes. >>>>>>>>> >>>>>>>>> Best regards, >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> Shirly Radco >>>>>>>>> >>>>>>>>> BI Senior Software Engineer >>>>>>>>> >>>>>>>>> Red Hat <https://www.redhat.com/> >>>>>>>>> >>>>>>>>> <https://www.redhat.com/> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <jay...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> I'm running in to this ansible error during oVirt metrics >>>>>>>>>> installation (following procedures at: >>>>>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>>>>>> ) >>>>>>>>>> >>>>>>>>>> This is happening late in the process, after successfully >>>>>>>>>> deploying the installation VM and then running second step from the >>>>>>>>>> metrics >>>>>>>>>> VM. >>>>>>>>>> >>>>>>>>>> CHECK [memory_availability : master0.xxxxxx.com] >>>>>>>>>> ************************************************************************************************* >>>>>>>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, >>>>>>>>>> "checks": {"disk_availability": {}, "docker_image_availability": >>>>>>>>>> {"changed": true}, "docker_storage": {"failed": true, "failures": >>>>>>>>>> [["OpenShiftCheckException", "Could not find imported module support >>>>>>>>>> code >>>>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>>>> line 53, in run\n docker_info = >>>>>>>>>> self.execute_module(\"docker_info\", >>>>>>>>>> {})\n File >>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>>>> line 211, in execute_module\n result = >>>>>>>>>> self._execute_module(module_name, >>>>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>> line 809, in _execute_module\n (module_style, shebang, >>>>>>>>>> module_data, >>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>> line 203, in _configure_module\n environment=final_environment)\n >>>>>>>>>> File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>>>> code for >>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>> docker_common.py\n"]], "msg": "Could not find imported module >>>>>>>>>> support code >>>>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>>>> line 53, in run\n docker_info = >>>>>>>>>> self.execute_module(\"docker_info\", >>>>>>>>>> {})\n File >>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>>>> line 211, in execute_module\n result = >>>>>>>>>> self._execute_module(module_name, >>>>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>> line 809, in _execute_module\n (module_style, shebang, >>>>>>>>>> module_data, >>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>> line 203, in _configure_module\n environment=final_environment)\n >>>>>>>>>> File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>>>> code for >>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>> docker_common.py\n"}, "memory_availability": {}, >>>>>>>>>> "package_availability": >>>>>>>>>> {"changed": false, "invocation": {"module_args": {"packages": >>>>>>>>>> ["PyYAML", >>>>>>>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", >>>>>>>>>> "firewalld", >>>>>>>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>>>>>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>>>>>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>>>>>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>>>>>>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>>>>>>>>> {"module_args": {"package_list": [{"check_multi": false, "name": >>>>>>>>>> "origin", >>>>>>>>>> "version": ""}, {"check_multi": false, "name": "origin-master", >>>>>>>>>> "version": >>>>>>>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>>>>>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>>>>>>>> "playbook_context": "install"} >>>>>>>>>> >>>>>>>>>> NO MORE HOSTS LEFT >>>>>>>>>> ******************************************************************************************************************************************* >>>>>>>>>> >>>>>>>>>> PLAY RECAP >>>>>>>>>> *************************************************************************************************************************************************** >>>>>>>>>> localhost : ok=35 changed=1 unreachable=0 >>>>>>>>>> failed=0 skipped=16 rescued=0 ignored=0 >>>>>>>>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 >>>>>>>>>> failed=1 skipped=165 rescued=0 ignored=0 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> INSTALLER STATUS >>>>>>>>>> ********************************************************************************************************************************************* >>>>>>>>>> Initialization : Complete (0:00:16) >>>>>>>>>> Health Check : In Progress (0:00:36) >>>>>>>>>> This phase can be restarted by running: >>>>>>>>>> playbooks/openshift-checks/pre-install.yml >>>>>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) >>>>>>>>>> 0:01:56.339 ************ >>>>>>>>>> >>>>>>>>>> =============================================================================== >>>>>>>>>> Run health checks (install) - EL >>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 36.15s >>>>>>>>>> os_firewall : Ensure iptables services are not enabled >>>>>>>>>> ------------------------------------------------------------------------------------------------ >>>>>>>>>> 2.74s >>>>>>>>>> openshift_repos : Ensure libselinux-python is installed >>>>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>>>> 1.77s >>>>>>>>>> openshift_repos : refresh cache >>>>>>>>>> ----------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.60s >>>>>>>>>> Gather Cluster facts >>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.51s >>>>>>>>>> container_runtime : Fixup SELinux permissions for docker >>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>> 1.33s >>>>>>>>>> container_runtime : Place additional/blocked/insecure registries >>>>>>>>>> in /etc/containers/registries.conf >>>>>>>>>> --------------------------------------------------- 1.30s >>>>>>>>>> Ensure openshift-ansible installer package deps are installed >>>>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>>>> 1.29s >>>>>>>>>> container_runtime : Install Docker >>>>>>>>>> -------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.17s >>>>>>>>>> Initialize openshift.node.sdn_mtu >>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.13s >>>>>>>>>> os_firewall : Install firewalld packages >>>>>>>>>> -------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.13s >>>>>>>>>> container_runtime : Set various Docker options >>>>>>>>>> -------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.11s >>>>>>>>>> install NetworkManager >>>>>>>>>> -------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.10s >>>>>>>>>> openshift_repos : Configure correct origin release repository >>>>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>>>> 1.05s >>>>>>>>>> container_runtime : Get current installed Docker version >>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>> 1.04s >>>>>>>>>> openshift_repos : Configure origin gpg keys >>>>>>>>>> ----------------------------------------------------------------------------------------------------------- >>>>>>>>>> 1.04s >>>>>>>>>> openshift_repos : Remove openshift_additional.repo file >>>>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>>>> 0.99s >>>>>>>>>> container_runtime : Setup the docker-storage for overlay >>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>> 0.96s >>>>>>>>>> Detecting Operating System from ostree_booted >>>>>>>>>> --------------------------------------------------------------------------------------------------------- >>>>>>>>>> 0.95s >>>>>>>>>> Gather Cluster facts >>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>> 0.92s >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Failure summary: >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 1. Hosts: master0.xxxxxxx.com >>>>>>>>>> Play: OpenShift Health Checks >>>>>>>>>> Task: Run health checks (install) - EL >>>>>>>>>> Message: One or more checks failed >>>>>>>>>> Details: check "docker_storage": >>>>>>>>>> Could not find imported module support code for >>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>> docker_common.py >>>>>>>>>> Traceback (most recent call last): >>>>>>>>>> File >>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>>>>>>>> line 225, in run_check >>>>>>>>>> result = check.run() >>>>>>>>>> File >>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>>>>>>>> line 53, in run >>>>>>>>>> docker_info = >>>>>>>>>> self.execute_module("docker_info", {}) >>>>>>>>>> File >>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>>>>>>>> line 211, in execute_module >>>>>>>>>> result = self._execute_module(module_name, >>>>>>>>>> module_args, self.tmp, self.task_vars) >>>>>>>>>> File >>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>>>> line >>>>>>>>>> 809, in _execute_module >>>>>>>>>> (module_style, shebang, module_data, >>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>> module_args=module_args, task_vars=task_vars) >>>>>>>>>> File >>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>>>> line >>>>>>>>>> 203, in _configure_module >>>>>>>>>> environment=final_environment) >>>>>>>>>> File >>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>> line >>>>>>>>>> 1023, in modify_module >>>>>>>>>> environment=environment) >>>>>>>>>> File >>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>> line >>>>>>>>>> 859, in _find_module_utils >>>>>>>>>> recursive_finder(module_name, b_module_data, >>>>>>>>>> py_module_names, py_module_cache, zf) >>>>>>>>>> File >>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>> line >>>>>>>>>> 621, in recursive_finder >>>>>>>>>> raise AnsibleError(' '.join(msg)) >>>>>>>>>> AnsibleError: Could not find imported module >>>>>>>>>> support code for docker_info. Looked for either >>>>>>>>>> AnsibleDockerClient.py or >>>>>>>>>> docker_common.py >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> The execution of "install_okd.yaml" includes checks designed to >>>>>>>>>> fail early if the requirements of the playbook are not met. One or >>>>>>>>>> more of >>>>>>>>>> these checks failed. To disregard these results,explicitly disable >>>>>>>>>> checks >>>>>>>>>> by setting an Ansible variable: >>>>>>>>>> openshift_disable_check=docker_storage >>>>>>>>>> Failing check names are shown in the failure details above. Some >>>>>>>>>> checks may be configurable by variables if your requirements are >>>>>>>>>> different >>>>>>>>>> from the defaults; consult check documentation. >>>>>>>>>> Variables can be set in the inventory or passed on the command >>>>>>>>>> line using the -e flag to ansible-playbook. >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>> List Archives: >>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>>>>>>>> >>>>>>>>> _______________________________________________ >>>>> Users mailing list -- users@ovirt.org >>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/ >>>>> >>>>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/MHYCRVEQVZN7UQ43ZNYJOWAT2ERTYEIH/