Trevor,

    Responded to your email with the logs, but that is still pending
moderator approval to be posted to the list (since the size of the message
was above the size cap). While a moderator makes up his / her mind, I
thought that I'd repeat some of my observations:

E1006 11:39:34.041543       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:39:34.041641       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []
W1006 11:40:12.831667       1 reflector.go:289]
k8s.io/client-go/informers/factory.go:133: watch of *v1.ConfigMap ended
with: too old resource version: 17351 (19676)
W1006 11:40:30.881153       1 reflector.go:289]
k8s.io/client-go/informers/factory.go:133: watch of *v1.ConfigMap ended
with: too old resource version: 18133 (19758)
E1006 11:45:13.003476       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:45:13.003613       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []
W1006 11:46:25.850256       1 reflector.go:289]
k8s.io/client-go/informers/factory.go:133: watch of *v1.ConfigMap ended
with: too old resource version: 19478 (21246)
E1006 11:46:26.883137       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:46:26.883217       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []
W1006 11:47:00.837769       1 reflector.go:289]
k8s.io/client-go/informers/factory.go:133: watch of *v1.ConfigMap ended
with: too old resource version: 19826 (21404)
W1006 11:47:03.833901       1 reflector.go:289]
k8s.io/client-go/informers/factory.go:133: watch of *v1.Deployment ended
with: too old resource version: 13636 (14162)
E1006 11:47:28.977942       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:47:28.978052       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []
E1006 11:47:29.001738       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:47:29.001853       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []
E1006 11:47:29.031826       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:47:29.031924       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []
E1006 11:47:29.455882       1 status.go:71] RouteSyncProgressing FailedHost
route is not available at canonical host []
E1006 11:47:29.456081       1 controller.go:129] {Console Console} failed
with: route is not available at canonical host []

     This is a default install with three masters. Previously, I had done
an install with one master, and even though that was hanging, the console
route had been defined (and so, I imagine, there would have been an A
record set up in DNS). Currently, I see an A record for the master, but not
for the console. I'm wondering if there is perhaps a screwup on my DNS
setup that is contributing to the problem.

    As of right now (probably another 30 mins past the install timing out),
I still see the following messages in the cluster version operator log:

I1006 12:31:16.462865       1 sync_worker.go:745] Update error 294 of 432:
ClusterOperatorNotAvailable Cluster operator console has not yet reported
success (*errors.errorString: cluster operator console is not done; it is
available=false, progressing=true, degraded=false)
I1006 12:31:16.462876       1 sync_worker.go:745] Update error 135 of 432:
ClusterOperatorNotAvailable Cluster operator authentication is still
updating (*errors.errorString: cluster operator authentication is still
updating)
I1006 12:31:16.462884       1 sync_worker.go:745] Update error 260 of 432:
ClusterOperatorNotAvailable Cluster operator monitoring is still updating
(*errors.errorString: cluster operator monitoring is still updating)
I1006 12:31:16.462890       1 sync_worker.go:745] Update error 183 of 432:
ClusterOperatorNotAvailable Cluster operator ingress is still updating
(*errors.errorString: cluster operator ingress is still updating)
I1006 12:31:16.462897       1 sync_worker.go:745] Update error 170 of 432:
ClusterOperatorNotAvailable Cluster operator image-registry is still
updating (*errors.errorString: cluster operator image-registry is still
updating)
E1006 12:31:16.462944       1 sync_worker.go:311] unable to synchronize
image (waiting 2m52.525702462s): Some cluster operators are still updating:
authentication, console, image-registry, ingress, monitoring

Regards,
Marvin

On Sat, Oct 5, 2019 at 6:22 PM W. Trevor King <wk...@redhat.com> wrote:

> On Sat, Oct 5, 2019 at 11:22 AM Just Marvin wrote:
> > INFO Destroying the bootstrap resources...
> > INFO Waiting up to 30m0s for the cluster at
> https://api.one.discworld.a.random.domain:6443 to initialize...
> > FATAL failed to initialize the cluster: Working towards
> 4.2.0-0.nightly-2019-10-01-210901: 99% complete
> > ...
> >     How do I track down what went wrong. And at this point, is it just a
> matter of waiting for a while? Suppose I let it go for a few hours, will
> there be a way to see if the initialization did complete?
>
> Might be.  You can launch additional waiters with 'openshift-install
> wait-for install-complete'.  You can also see exactly what the
> cluster-version operator is stuck on by looking in the cluster-version
> operator pod's logs.  Or you can inspect the ClusterOperator resources
> and see if any of the core operators has more-specific complaints.  Or
> you can 'oc adm must-gather' to get a tarball of OpenStack components
> to send to us if you'd rather have us poke around.
>
> Cheers,
> Trevor
>
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to