Are you hitting https://github.com/openshift/openshift-ansible/pull/9598 ?

On 10/9/18 11:25 AM, Dan Pungă wrote:
Thanks for the reply Scott!

I've used the release branches for both 3.9 and 3.10 of the openshift-ansible 
project, yes.
I've initially checked the openshift_facts.py script flow in the 3.9 branch; 
now looking at the 3.10 version, I do see the change that you're pointing.

On 09.10.2018 05:40, Scott Dodson wrote:
Dan, are you using the latest from release-3.10 branch? I believe we've disabled the IaaS interrogation when you've not configured a cloud provider via openshift-ansible in the latest on that branch.

On Mon, Oct 8, 2018, 7:38 PM Dan Pungă <dan.pu...@gmail.com 
<mailto:dan.pu...@gmail.com>> wrote:

    I've done a bit of digging and apparently my problem is precisely connected 
to the fact that I'm running the cluster on the OpenStack provider.

    Briefly put, the openshift_facts playbook relies on the 
openshift-ansible/roles/openshift_facts/library/openshift_facts.py script. This 
script uses the ansible.module_utils tools to
    discover the underlying system, including any existing IaaS provider with 
its detailis. In my case it discovers the OpenStack provider and when setting 
the hostnames, the provider
    configuration takes precedence over whatever I've configured at the VM 
level.

    In my case, I haven't properly set up the FQDNs/hostnames at the OpenStack 
level. Instead, after I've created and launched the instances, I disabled at 
the VM level the ability of the
    cloud provider to reset my hostname definition/configuration and I thought 
this would be enough.

    I guess I'll try a reinstall on a lab environment with the 
openshift_facts.py script modified so that it passes over the Openstack check 
and hope it does what I'd expect, which is to be
    agnostic to the type of hosts on which I install.
    I actually thought that the only way the OpenShift/OKD installer would try 
to integrate with a provider was if I'd specifically set the 
openshift_cloudprovider_kind variable in the
    inventory file along with the rest of the specific variables.

    Regards,
    Dan Pungă

    On 08.10.2018 18:44, Dan Pungă wrote:
    Hello all,

    I'm trying to upgrade a working cluster from Openshift Origin 3.9 to OKD 
3.10 and the control plane update fails at one point with host not found.
    I've looked abit over the problem and found this issue on github: 
https://github.com/openshift/openshift-ansible/issues/9935 where michaelgugino 
points out that "when upgrading from
    3.9, your hostnames match the node names in 'oc get nodes' otherwise, we won't 
be able to find the CSRs for your nodes."

    In fact my issue is precisely this: the node names are in fact their IPs 
and not the hostnames of the specific machines. It was something that I saw 
upon installation, but as the 3.9
    cluster was functioning all right, I let it be.
    The idea is that I (think) I have the DNS resolution set up properly, with 
all machines being able to resolve each-other by FQDNs, however the 3.9 
installer configured the node names
    with their respective IP addresses and I don't know how to address this.
    I should mention that the cluster is deployed inside an Openstack project, 
but the install config doesn't use OpenShift-Openstack configuration. However 
when running the
    ~/openshift-ansible/playbooks/byo/openshift_facts.yml I get references to the 
underlying openstack(somehow the installer "figures out" the undelying 
Openstack and treats it as a
    provider, the way I see it). I've pasted the output for one of the nodes 
below.

    Has any of you come across this node name config problem and were you able 
to solve it?
    Is there any procedure to change node names of a working cluster? I should 
say that the masters are also nodes(infrasructure), so I'm guessing the 
procedure, if there is one, would
    have to do with deprecating one master at a time, while for the nodes with 
a delete/change config/re-add procedure.

    Thank you!

    Output from openshift_facts playbook:

    ok: [node1.oshift-pinfold.intra] => {
        "result": {
            "ansible_facts": {
                "openshift": {
                    "common": {
                        "all_hostnames": [
    "node1.oshift-pinfold.intra",
                            "192.168.150.22"
                        ],
                        "config_base": "/etc/origin",
                        "deployment_subtype": "basic",
                        "deployment_type": "origin",
                        "dns_domain": "cluster.local",
                        "examples_content_version": "v3.9",
                        "generate_no_proxy_hosts": true,
                        "hostname": "192.168.150.22",
                        "internal_hostnames": [
                            "192.168.150.22"
                        ],
                        "ip": "192.168.150.22",
                        "kube_svc_ip": "172.30.0.1",
                        "portal_net": "172.30.0.0/16 <http://172.30.0.0/16>",
                        "public_hostname": "node1.oshift-pinfold.intra",
                        "public_ip": "192.168.150.22",
                        "short_version": "3.9",
                        "version": "3.9.0",
                        "version_gte_3_10": false,
                        "version_gte_3_6": true,
                        "version_gte_3_7": true,
                        "version_gte_3_8": true,
                        "version_gte_3_9": true
                    },
                    "current_config": {
                        "roles": [
                            "node"
                        ]
                    },
                    "node": {
                        "bootstrapped": false,
                        "nodename": "192.168.150.22",
                        "sdn_mtu": "1408"
                    },
                    "provider": {
                        "metadata": {
                            "availability_zone": "nova",
                            "ec2_compat": {
                                "ami-id": "None",
                                "ami-launch-index": "0",
                                "ami-manifest-path": "FIXME",
                                "block-device-mapping": {
                                    "ami": "vda",
                                    "ebs0": "/dev/vda",
                                    "ebs1": "/dev/vdb",
                                    "root": "/dev/vda"
                                },
                                "hostname": "node1.novalocal",
                                "instance-action": "none",
                                "instance-id": "i-00000583",
                                "instance-type": "1cpu-2ram-20disk",
                                "local-hostname": "node1.novalocal",
                                "local-ipv4": "192.168.150.22",
                                "placement": {
                                    "availability-zone": "nova"
                                },
                                "public-hostname": "node1.novalocal",
                                "public-ipv4": [],
                                "public-keys/": "0=xxxxxxxxxxx",
                                "reservation-id": "r-la13azpq",
                                "security-groups": [
                                    "DefaultInternal",
                                    "oshift-node"
                                ]
                            },
                            "hostname": "node1.novalocal",
                            "keys": [
                                {
                                    "data": "ssh-rsa 
AAAA........................................................... Generated-by-Nova",
                                    "name": "xxxxxxxxxx",
                                    "type": "ssh"
                                }
                            ],
                            "launch_index": 0,
                            "name": "node1",
                            "project_id": "2a2.........................",
                            "uuid": "80bb................................."
                        },
                        "name": "openstack",
                        "network": {
                            "hostname": "192.168.150.22",
                            "interfaces": [],
                            "ip": "192.168.150.22",
                            "ipv6_enabled": false,
                            "public_hostname": [],
                            "public_ip": []
                        },
                        "zone": "nova"
                    }
                }
            },
            "changed": false,
            "failed": false
        }
    }

    _______________________________________________
    users mailing list
    users@lists.openshift.redhat.com <mailto:users@lists.openshift.redhat.com>
    http://lists.openshift.redhat.com/openshiftmm/listinfo/users



_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users


_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to