Hi Jayme, It getting hard to debug your issue over the mailing list. Can you please open a bug in bugzilla and attach all the information you know? The versions you are using, config files from engine machine and ansible log, vars.yaml , integ.ini from the master0 vm in /root directory, .
Sorry for the inconvenience. Best, -- Shirly Radco BI Senior Software Engineer Red Hat <https://www.redhat.com/> <https://www.redhat.com/> On Tue, May 28, 2019 at 10:47 PM Jayme <jay...@gmail.com> wrote: > I actually see the pods running on master0 if I do this: > > @master0 master]# oc project kube-system > Now using project "kube-system" on server " > https://openshift-master.cloud.xxxxxxx.com:8443". > [root@master0 master]# oc get pods > NAME READY STATUS > RESTARTS AGE > master-api-master0.cloud.xxxxxxxx.com 1/1 Running 0 > 22m > master-controllers-master0.cloud.xxxxxxxx.com 1/1 Running 0 > 22m > master-etcd-master0.cloud.xxxxxxxxxx 1/1 Running 0 > 22m > > So I wonder why the ansible "Wait for control plane pods to appear" task > is looping > > - name: Wait for control plane pods to appear > oc_obj: > state: list > kind: pod > name: "master-{{ item }}-{{ l_kubelet_node_name | lower }}" > namespace: kube-system > register: control_plane_pods > until: > - "'results' in control_plane_pods" > - "'results' in control_plane_pods.results" > - control_plane_pods.results.results | length > 0 > retries: 60 > delay: 5 > with_items: > - "{{ 'etcd' if inventory_hostname in groups['oo_etcd_to_config'] else > omit }}" > - api > - controllers > ignore_errors: true > > On Tue, May 28, 2019 at 4:23 PM Jayme <jay...@gmail.com> wrote: > >> I just tried again from scratch this time making sure a proper wildcard >> DNS entry existed and without using the set /etc/hosts option and am still >> running in to the pods issue. Can anyone confirm if this requires a public >> external IP to work? I am working on an internal DNS zone here and natted >> ips. >> >> On Tue, May 28, 2019 at 3:28 PM Edward Berger <edwber...@gmail.com> >> wrote: >> >>> In my case it was a single bare metal host, so that would be equivalent >>> to disabling iptables on the master0 VM you're installing to, in your ovirt >>> scenario. >>> >>> On Tue, May 28, 2019 at 1:25 PM Jayme <jay...@gmail.com> wrote: >>> >>>> Do you mean the iptables firewall on the server being installed to i.e. >>>> master0 or the actual oVirt host that the master0 VM is running on? I did >>>> try flushing iptables rules on master0 VM then ran plays again from >>>> installer VM but fail at the same point. >>>> >>>> Does this log message have anything to do with the issue, /etc/cni >>>> directory does not even exist on master0 VM. >>>> >>>> May 28 17:23:35 master0 origin-node: W0528 17:23:35.012902 10434 >>>> cni.go:172] Unable to update cni config: No networks found in >>>> /etc/cni/net.d >>>> May 28 17:23:35 master0 origin-node: E0528 17:23:35.013398 10434 >>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>> cni config uninitialized >>>> >>>> >>>> >>>> On Tue, May 28, 2019 at 1:19 PM Edward Berger <edwber...@gmail.com> >>>> wrote: >>>> >>>>> > TASK [openshift_control_plane : Wait for control plane pods to >>>>> appear] ********* >>>>> > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>>> ************ >>>>> > FAILED - RETRYING: Wait for control plane pods to appear (60 retries >>>>> left). >>>>> > FAILED - RETRYING: Wait for control plane pods to appear (59 retries >>>>> left). >>>>> > It eventually counts all the way down to zero and fails. >>>>> >>>>> This looks a lot like the issues I saw when the host firewall >>>>> (iptables) was blocking another OKD all-in-one-host install script [1]. >>>>> Disabling iptables allowed the installation to continue for my proof >>>>> of concept "cluster". >>>>> >>>>> [1]https://github.com/gshipley/installcentos >>>>> >>>>> The other error I had with [1] was it was trying to install a couple >>>>> of packages (zile and python2-pip) from EPEL with the repo disabled. >>>>> >>>>> >>>>> >>>>> On Tue, May 28, 2019 at 10:41 AM Jayme <jay...@gmail.com> wrote: >>>>> >>>>>> Shirly, >>>>>> >>>>>> Oh and I should mention that I did verify that NetworkManager was >>>>>> installed on the master0 VM and enabled/started the second go around. So >>>>>> that service is there and running. >>>>>> >>>>>> # systemctl list-unit-files | grep Network >>>>>> dbus-org.freedesktop.NetworkManager.service >>>>>> enabled >>>>>> NetworkManager-dispatcher.service >>>>>> enabled >>>>>> NetworkManager-wait-online.service >>>>>> enabled >>>>>> NetworkManager.service >>>>>> enabled >>>>>> >>>>>> On Tue, May 28, 2019 at 11:13 AM Jayme <jay...@gmail.com> wrote: >>>>>> >>>>>>> Shirly, >>>>>>> >>>>>>> I appreciate the help with this. Unfortunately I am still running >>>>>>> in to the same problem. So far I've tried to install/enable/start >>>>>>> NetworkManager on the existing "master0" server and re-ran the plans >>>>>>> from >>>>>>> the installer VM. I ran in to the same problem waiting for control >>>>>>> plane >>>>>>> pods and same errors in syslog. >>>>>>> >>>>>>> So I wiped everything out, killed the template along with the >>>>>>> installer and master VMs. On oVirt engine (I am running 4.3.3.7-1 >>>>>>> stable) >>>>>>> I did have ovirt-engine-metrics-1.3.0x rpm installed, no yum updates >>>>>>> available on an update check. So I installed >>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm >>>>>>> then proceeded to install the latest version of ovirt-engine-metrics >>>>>>> which >>>>>>> gave me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. >>>>>>> >>>>>>> After that package was installed I proceeded to follow steps from >>>>>>> beginning outlined at: >>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>>> -- >>>>>>> I ran in to the docker check issue again (same as my initial email) so I >>>>>>> disabled that and again got as far as starting control plane pods before >>>>>>> failure. >>>>>>> >>>>>>> Not sure where to go from here at this point. The only thing I can >>>>>>> think of that I did differently vs the instructions outlined above is >>>>>>> that >>>>>>> I have not crated the wildcard DNS record, however I did set configs to >>>>>>> create /etc/hosts entries and they /etc/hosts on the machines have the >>>>>>> proper IPs assigned for all hostnames (automatically added by the >>>>>>> ansible >>>>>>> plays). >>>>>>> >>>>>>> Any ideas how I can get past the plane pods issue? >>>>>>> >>>>>>> Thanks! >>>>>>> >>>>>>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <sra...@redhat.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> The latest release of 4.3.z should already include a fix for this >>>>>>>> issue, ovirt-engine-metrics-1.3.1 rpm. >>>>>>>> >>>>>>>> The issue is that it requires the NetworkManagar to be installed, >>>>>>>> running and enabled for it to work. >>>>>>>> >>>>>>>> You can install it manually on the master0 vm , start and enable it >>>>>>>> or you can also install the updated rpm from the nightly builds if your >>>>>>>> environment is oVirt 4.2.z: >>>>>>>> >>>>>>>> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm >>>>>>>> >>>>>>>> Relevant patches are: >>>>>>>> https://gerrit.ovirt.org/#/c/99959/ >>>>>>>> https://gerrit.ovirt.org/#/c/99718/ >>>>>>>> >>>>>>>> Best regards, >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> Shirly Radco >>>>>>>> >>>>>>>> BI Senior Software Engineer >>>>>>>> >>>>>>>> Red Hat <https://www.redhat.com/> >>>>>>>> >>>>>>>> <https://www.redhat.com/> >>>>>>>> >>>>>>>> >>>>>>>> On Mon, May 27, 2019 at 4:41 PM Jayme <jay...@gmail.com> wrote: >>>>>>>> >>>>>>>>> I managed to get past that but am running in to another problem >>>>>>>>> later in the process on the control plane pods to appear task. I >>>>>>>>> thought >>>>>>>>> perhaps a glitch in the process from the failed docker step >>>>>>>>> previously so >>>>>>>>> after a few more runs I tried killing everything and restarting the >>>>>>>>> metrics >>>>>>>>> process again from the very beginning and end up hitting the same >>>>>>>>> issue >>>>>>>>> with control plane pods even though all other steps/tasks seem to be >>>>>>>>> working. >>>>>>>>> >>>>>>>>> I'm just getting this: >>>>>>>>> >>>>>>>>> TASK [openshift_control_plane : Wait for control plane pods to >>>>>>>>> appear] ********* >>>>>>>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>>>>>>> ************ >>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (60 >>>>>>>>> retries left). >>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (59 >>>>>>>>> retries left). >>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 >>>>>>>>> retries left). >>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 >>>>>>>>> retries left). >>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 >>>>>>>>> retries left). >>>>>>>>> >>>>>>>>> It eventually counts all the way down to zero and fails. >>>>>>>>> >>>>>>>>> In syslog of the master0 server I'm seeing some errors related to >>>>>>>>> cni config >>>>>>>>> >>>>>>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None >>>>>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>>>>> namespace=kube-system delete_after=False >>>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>>>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 >>>>>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>>>>> /etc/cni/net.d >>>>>>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 >>>>>>>>> kubelet.go:2101] Container runtime network not ready: >>>>>>>>> NetworkReady=false >>>>>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not >>>>>>>>> ready: >>>>>>>>> cni config uninitialized >>>>>>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None >>>>>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>>>>> namespace=kube-system delete_after=False >>>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>>>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 >>>>>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>>>>> /etc/cni/net.d >>>>>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 >>>>>>>>> kubelet.go:2101] Container runtime network not ready: >>>>>>>>> NetworkReady=false >>>>>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not >>>>>>>>> ready: >>>>>>>>> cni config uninitialized >>>>>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 >>>>>>>>> summary.go:102] Failed to get system container stats for >>>>>>>>> "/system.slice/origin-node.service": failed to get cgroup stats for >>>>>>>>> "/system.slice/origin-node.service": failed to get container info for >>>>>>>>> "/system.slice/origin-node.service": unknown container >>>>>>>>> "/system.slice/origin-node.service" >>>>>>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 >>>>>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>>>>> /etc/cni/net.d >>>>>>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 >>>>>>>>> kubelet.go:2101] Container runtime network not ready: >>>>>>>>> NetworkReady=false >>>>>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not >>>>>>>>> ready: >>>>>>>>> cni config uninitialized >>>>>>>>> >>>>>>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sra...@redhat.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi Jayme, >>>>>>>>>> >>>>>>>>>> Thank you for reaching out. >>>>>>>>>> Please try rerunning the ansible playbook. >>>>>>>>>> If this doesn't work, try adding to the integ.ini in the metrics >>>>>>>>>> vm >>>>>>>>>> openshift_disable_check=docker_storage >>>>>>>>>> and rerun the ansible playbook again. >>>>>>>>>> >>>>>>>>>> Please update how it goes. >>>>>>>>>> >>>>>>>>>> Best regards, >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> Shirly Radco >>>>>>>>>> >>>>>>>>>> BI Senior Software Engineer >>>>>>>>>> >>>>>>>>>> Red Hat <https://www.redhat.com/> >>>>>>>>>> >>>>>>>>>> <https://www.redhat.com/> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <jay...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> I'm running in to this ansible error during oVirt metrics >>>>>>>>>>> installation (following procedures at: >>>>>>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>>>>>>> ) >>>>>>>>>>> >>>>>>>>>>> This is happening late in the process, after successfully >>>>>>>>>>> deploying the installation VM and then running second step from the >>>>>>>>>>> metrics >>>>>>>>>>> VM. >>>>>>>>>>> >>>>>>>>>>> CHECK [memory_availability : master0.xxxxxx.com] >>>>>>>>>>> ************************************************************************************************* >>>>>>>>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, >>>>>>>>>>> "checks": {"disk_availability": {}, "docker_image_availability": >>>>>>>>>>> {"changed": true}, "docker_storage": {"failed": true, "failures": >>>>>>>>>>> [["OpenShiftCheckException", "Could not find imported module >>>>>>>>>>> support code >>>>>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>>>>> line 53, in run\n docker_info = >>>>>>>>>>> self.execute_module(\"docker_info\", >>>>>>>>>>> {})\n File >>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>>>>> line 211, in execute_module\n result = >>>>>>>>>>> self._execute_module(module_name, >>>>>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>> line 809, in _execute_module\n (module_style, shebang, >>>>>>>>>>> module_data, >>>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>> line 203, in _configure_module\n >>>>>>>>>>> environment=final_environment)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>>>>> code for >>>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>> docker_common.py\n"]], "msg": "Could not find imported module >>>>>>>>>>> support code >>>>>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>>>>> line 53, in run\n docker_info = >>>>>>>>>>> self.execute_module(\"docker_info\", >>>>>>>>>>> {})\n File >>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>>>>> line 211, in execute_module\n result = >>>>>>>>>>> self._execute_module(module_name, >>>>>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>> line 809, in _execute_module\n (module_style, shebang, >>>>>>>>>>> module_data, >>>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>> line 203, in _configure_module\n >>>>>>>>>>> environment=final_environment)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>>>>> code for >>>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>> docker_common.py\n"}, "memory_availability": {}, >>>>>>>>>>> "package_availability": >>>>>>>>>>> {"changed": false, "invocation": {"module_args": {"packages": >>>>>>>>>>> ["PyYAML", >>>>>>>>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", >>>>>>>>>>> "firewalld", >>>>>>>>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>>>>>>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>>>>>>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>>>>>>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>>>>>>>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>>>>>>>>>> {"module_args": {"package_list": [{"check_multi": false, "name": >>>>>>>>>>> "origin", >>>>>>>>>>> "version": ""}, {"check_multi": false, "name": "origin-master", >>>>>>>>>>> "version": >>>>>>>>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>>>>>>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>>>>>>>>> "playbook_context": "install"} >>>>>>>>>>> >>>>>>>>>>> NO MORE HOSTS LEFT >>>>>>>>>>> ******************************************************************************************************************************************* >>>>>>>>>>> >>>>>>>>>>> PLAY RECAP >>>>>>>>>>> *************************************************************************************************************************************************** >>>>>>>>>>> localhost : ok=35 changed=1 unreachable=0 >>>>>>>>>>> failed=0 skipped=16 rescued=0 ignored=0 >>>>>>>>>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 >>>>>>>>>>> failed=1 skipped=165 rescued=0 ignored=0 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> INSTALLER STATUS >>>>>>>>>>> ********************************************************************************************************************************************* >>>>>>>>>>> Initialization : Complete (0:00:16) >>>>>>>>>>> Health Check : In Progress (0:00:36) >>>>>>>>>>> This phase can be restarted by running: >>>>>>>>>>> playbooks/openshift-checks/pre-install.yml >>>>>>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) >>>>>>>>>>> 0:01:56.339 ************ >>>>>>>>>>> >>>>>>>>>>> =============================================================================== >>>>>>>>>>> Run health checks (install) - EL >>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 36.15s >>>>>>>>>>> os_firewall : Ensure iptables services are not enabled >>>>>>>>>>> ------------------------------------------------------------------------------------------------ >>>>>>>>>>> 2.74s >>>>>>>>>>> openshift_repos : Ensure libselinux-python is installed >>>>>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.77s >>>>>>>>>>> openshift_repos : refresh cache >>>>>>>>>>> ----------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.60s >>>>>>>>>>> Gather Cluster facts >>>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.51s >>>>>>>>>>> container_runtime : Fixup SELinux permissions for docker >>>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.33s >>>>>>>>>>> container_runtime : Place additional/blocked/insecure registries >>>>>>>>>>> in /etc/containers/registries.conf >>>>>>>>>>> --------------------------------------------------- 1.30s >>>>>>>>>>> Ensure openshift-ansible installer package deps are installed >>>>>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>>>>> 1.29s >>>>>>>>>>> container_runtime : Install Docker >>>>>>>>>>> -------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.17s >>>>>>>>>>> Initialize openshift.node.sdn_mtu >>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.13s >>>>>>>>>>> os_firewall : Install firewalld packages >>>>>>>>>>> -------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.13s >>>>>>>>>>> container_runtime : Set various Docker options >>>>>>>>>>> -------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.11s >>>>>>>>>>> install NetworkManager >>>>>>>>>>> -------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.10s >>>>>>>>>>> openshift_repos : Configure correct origin release repository >>>>>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>>>>> 1.05s >>>>>>>>>>> container_runtime : Get current installed Docker version >>>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.04s >>>>>>>>>>> openshift_repos : Configure origin gpg keys >>>>>>>>>>> ----------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 1.04s >>>>>>>>>>> openshift_repos : Remove openshift_additional.repo file >>>>>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>>>>> 0.99s >>>>>>>>>>> container_runtime : Setup the docker-storage for overlay >>>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>>> 0.96s >>>>>>>>>>> Detecting Operating System from ostree_booted >>>>>>>>>>> --------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 0.95s >>>>>>>>>>> Gather Cluster facts >>>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>> 0.92s >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Failure summary: >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 1. Hosts: master0.xxxxxxx.com >>>>>>>>>>> Play: OpenShift Health Checks >>>>>>>>>>> Task: Run health checks (install) - EL >>>>>>>>>>> Message: One or more checks failed >>>>>>>>>>> Details: check "docker_storage": >>>>>>>>>>> Could not find imported module support code for >>>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>> docker_common.py >>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>> File >>>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>>>>>>>>> line 225, in run_check >>>>>>>>>>> result = check.run() >>>>>>>>>>> File >>>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>>>>>>>>> line 53, in run >>>>>>>>>>> docker_info = >>>>>>>>>>> self.execute_module("docker_info", {}) >>>>>>>>>>> File >>>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>>>>>>>>> line 211, in execute_module >>>>>>>>>>> result = self._execute_module(module_name, >>>>>>>>>>> module_args, self.tmp, self.task_vars) >>>>>>>>>>> File >>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>>>>> line >>>>>>>>>>> 809, in _execute_module >>>>>>>>>>> (module_style, shebang, module_data, >>>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>>> module_args=module_args, task_vars=task_vars) >>>>>>>>>>> File >>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>>>>> line >>>>>>>>>>> 203, in _configure_module >>>>>>>>>>> environment=final_environment) >>>>>>>>>>> File >>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>>> line >>>>>>>>>>> 1023, in modify_module >>>>>>>>>>> environment=environment) >>>>>>>>>>> File >>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>>> line >>>>>>>>>>> 859, in _find_module_utils >>>>>>>>>>> recursive_finder(module_name, b_module_data, >>>>>>>>>>> py_module_names, py_module_cache, zf) >>>>>>>>>>> File >>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>>> line >>>>>>>>>>> 621, in recursive_finder >>>>>>>>>>> raise AnsibleError(' '.join(msg)) >>>>>>>>>>> AnsibleError: Could not find imported module >>>>>>>>>>> support code for docker_info. Looked for either >>>>>>>>>>> AnsibleDockerClient.py or >>>>>>>>>>> docker_common.py >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> The execution of "install_okd.yaml" includes checks designed to >>>>>>>>>>> fail early if the requirements of the playbook are not met. One or >>>>>>>>>>> more of >>>>>>>>>>> these checks failed. To disregard these results,explicitly disable >>>>>>>>>>> checks >>>>>>>>>>> by setting an Ansible variable: >>>>>>>>>>> openshift_disable_check=docker_storage >>>>>>>>>>> Failing check names are shown in the failure details above. Some >>>>>>>>>>> checks may be configurable by variables if your requirements are >>>>>>>>>>> different >>>>>>>>>>> from the defaults; consult check documentation. >>>>>>>>>>> Variables can be set in the inventory or passed on the command >>>>>>>>>>> line using the -e flag to ansible-playbook. >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>>> List Archives: >>>>>>>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>> Users mailing list -- users@ovirt.org >>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>> oVirt Code of Conduct: >>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>> List Archives: >>>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/ >>>>>> >>>>>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/MKTNTLQD5VDK3XBO3XVHJIAKJCX3EIC5/