Hi,
Not sure about the error, but I've noticed a task that I haven't seen
during my installation attempts (also Origin 3.9 on a cluster).
From what I see, "origin_control_plane" is an ansible role that's
present on the master branch of the openshift-ansible repo, but not on
the release-3.9 branch.
https://docs.openshift.org/latest/install_config/install/host_preparation.html#preparing-for-advanced-installations-origin
ststes that we should use the release-3.9 branch and that the master is
intended for the currently developed version of OShift Origin.
Hope it helps!
On 24.05.2018 21:52, Jason Marshall wrote:
Good afternoon,
I am attempting to do an advanced installation of Origin 3.9 in a
cluster with a single master and 2 nodes, with a node role on the
master server.
I am able to run the prerequisites.yml playbook with no issue. The
deploy_cluster.yml fails at the point where the origin.node service
attempts to start on the master server. The error that comes up is:
TASK [openshift_control_plane : Start and enable self-hosting node]
********************************************************************************************************************
fatal: [openshift-master.expdev.local]: FAILED! => {"changed": false,
"msg": "Unable to restart service origin-node: Job for
origin-node.service failed because the control process exited with
error code. See \"systemctl status origin-node.service\" and
\"journalctl -xe\" for details.\n"}
...ignoring
<snip>
"May 24 14:40:53 cmhldshftlab01.expdev.local
origin-node[2657]: /usr/local/bin/openshift-node: line 17:
/usr/bin/openshift-node-config: No such file or directory",
"May 24 14:40:53 cmhldshftlab01.expdev.local systemd[1]:
origin-node.service: main process exited, code=exited, status=1/FAILURE",
"May 24 14:40:53 cmhldshftlab01.expdev.local systemd[1]:
Failed to start OpenShift Node.",
"May 24 14:40:53 cmhldshftlab01.expdev.local systemd[1]: Unit
origin-node.service entered failed state.",
"May 24 14:40:53 cmhldshftlab01.expdev.local systemd[1]:
origin-node.service failed."
<snip>
INSTALLER STATUS
***********************************************************************************************************************************************************************
Initialization : Complete (0:00:33)
Health Check : Complete (0:00:24)
Node Preparation : Complete (0:00:01)
etcd Install : Complete (0:00:41)
Load Balancer Install : Complete (0:00:18)
Master Install : In Progress (0:01:47)
This phase can be restarted by running:
playbooks/openshift-master/config.yml
Failure summary:
1. Hosts: openshift-master.expdev.local
Play: Configure masters
Task: openshift_control_plane : fail
Message: Node start failed.
I go looking for openshift-node-config, and can't find it anywhere.
Nor can I find where this file comes from, even when using "yum
whatprovides" or a find command in the openshift-ansible directory I
am installing from.
Am I running into a potential configuration issue, or a bug with the
version of origin I am running? My openshift-ansible folder was pulled
down at around 2PM Eastern today, as I refreshed it to see if there
was any difference in behavior.
Any suggestions or troubleshooting tips would be most appreciated.
Thank you,
Jason
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users