Hmm, in my case I don't need to need to restart to cause the problem to
happen. Is there some way to run nightlies of openshift:release-3.7 using
the openshift-ansible? So that I can verify it's fixed for me?

On Mon, Jan 8, 2018 at 12:23 PM Jordan Liggitt <jligg...@redhat.com> wrote:

> Garbage collection in particular could be related to
> https://bugzilla.redhat.com/show_bug.cgi?id=1525699 (fixed in
> https://github.com/openshift/origin/pull/17818 but not included in a
> point release yet)
>
>
> On Jan 7, 2018, at 8:17 PM, Joel Pearson <japear...@agiledigital.com.au>
> wrote:
>
> Hi,
>
> Has anyone else noticed that the new OpenShift Origin 3.7 Template Broker
> seems super flaky?
>
> For example, if I deploy a Jenkins (Persistent or Ephemeral), and then I
> modify the route, by adding an annotation for example:
>
> kubernetes.io/tls-acme: 'true'
>
> I have https://github.com/tnozicka/openshift-acme Installed in the
> cluster which then grabs an SSL cert for me, adds it to the route, then
> moments later all resources from the template are garbage collected for no
> apparent reason.
>
> I also got the same behaviour when I modified the service account the
> Jenkins template uses, I added an additional route so I added a new "
> serviceaccounts.openshift.io/oauth-redirectreference.jenkins:" entry. It
> took a bit longer (like 12 hours), but it all disappeared again.  I have a
> suspicion that if you modify any object that a template created, then
> eventually the template broker will remove all objects it created.
>
> Is there any way to disable the new template broker and use the old
> template system?
>
> In Origin 3.6 it was flawless and worked with openshift-acme without any
> problems at all.
>
> I should mention that if I create things manually then it works fine, I
> can use openshift-acme, and all my resources don't vanish at whim.
>
> Here is a snippet of the logs, you can see the acme points are removed
> after successfully getting a cert, and then moments later, the deleting
> starts:
>
> Jan 08 00:26:47 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:47.648255       1 leaderelection.go:199] successfully renewed
> lease kube-service-catalog/service-catalog-controller-manager
> Jan 08 00:26:47 master-0.openshift.staging.local origin-node[26684]: I0108
> 00:26:47.744777   26749 roundrobin.go:338] LoadBalancerRR: Removing
> endpoints for jenkins-test/acme-9cv97q5dn8:
> Jan 08 00:26:47 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:47.744777   26749 roundrobin.go:338] LoadBalancerRR: Removing
> endpoints for jenkins-test/acme-9cv97q5dn8:
> Jan 08 00:26:47 master-0.openshift.staging.local origin-node[26684]: I0108
> 00:26:47.762005   26749 ovs.go:143] Error executing ovs-ofctl: ovs-ofctl:
> None: invalid IP address
> Jan 08 00:26:47 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:47.762005   26749 ovs.go:143] Error executing ovs-ofctl:
> ovs-ofctl: None: invalid IP address
> Jan 08 00:26:47 master-0.openshift.staging.local dockerd-current[23329]:
> E0108 00:26:47.765091   26749 sdn_controller.go:284] Error deleting OVS
> flows for service &{{ } {acme-9cv97q5dn8  jenkins-test
> /api/v1/namespaces/jenkins-test/services/acme-9cv97q5dn8
> 94c6b3b3-f40a-11e7-88e5-fa163eb8ca3a 622382 0 2018-01-08 00:26:34 +0000 UTC
> <nil> <nil> map[] map[] [] nil [] } {ClusterIP [{http TCP 80 {0 80 } 0}]
> map[] None  []  None []  0} {{[]}}}: exit status 1
> Jan 08 00:26:47 master-0.openshift.staging.local origin-node[26684]: E0108
> 00:26:47.765091   26749 sdn_controller.go:284] Error deleting OVS flows for
> service &{{ } {acme-9cv97q5dn8  jenkins-test
> /api/v1/namespaces/jenkins-test/services/acme-9cv97q5dn8
> 94c6b3b3-f40a-11e7-88e5-fa163eb8ca3a 622382 0 2018-01-08 00:26:34 +0000 UTC
> <nil> <nil> map[] map[] [] nil [] } {ClusterIP [{http TCP 80 {0 80 } 0}]
> map[] None  []  None []  0} {{[]}}}: exit status 1
> Jan 08 00:26:48 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:48.139090       1 rest.go:362] Starting watch for
> /api/v1/namespaces, rv=622418 labels= fields= timeout=8m38s
> Jan 08 00:26:48 master-0.openshift.staging.local origin-master-api[23448]:
> I0108 00:26:48.139090       1 rest.go:362] Starting watch for
> /api/v1/namespaces, rv=622418 labels= fields= timeout=8m38s
> Jan 08 00:26:49 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:49.668205       1 leaderelection.go:199] successfully renewed
> lease kube-service-catalog/service-catalog-controller-manager
> Jan 08 00:26:49 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:49.885207       1 garbagecollector.go:291] processing item [
> template.openshift.io/v1/TemplateInstance, namespace: jenkins-test, name:
> e3639aec-bbbc-4170-b0e4-3b63735af348, uid:
> 915d585d-f408-11e7-88e5-fa163eb8ca3a]
> Jan 08 00:26:49 master-0.openshift.staging.local
> origin-master-controllers[73353]: I0108 00:26:49.885207       1
> garbagecollector.go:291] processing item [
> template.openshift.io/v1/TemplateInstance, namespace: jenkins-test, name:
> e3639aec-bbbc-4170-b0e4-3b63735af348, uid:
> 915d585d-f408-11e7-88e5-fa163eb8ca3a]
> Jan 08 00:26:49 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:49.904249       1 garbagecollector.go:394] delete object [
> template.openshift.io/v1/TemplateInstance, namespace: jenkins-test, name:
> e3639aec-bbbc-4170-b0e4-3b63735af348, uid:
> 915d585d-f408-11e7-88e5-fa163eb8ca3a] with propagation policy Background
> Jan 08 00:26:49 master-0.openshift.staging.local
> origin-master-controllers[73353]: I0108 00:26:49.904249       1
> garbagecollector.go:394] delete object [
> template.openshift.io/v1/TemplateInstance, namespace: jenkins-test, name:
> e3639aec-bbbc-4170-b0e4-3b63735af348, uid:
> 915d585d-f408-11e7-88e5-fa163eb8ca3a] with propagation policy Background
> Jan 08 00:26:49 master-0.openshift.staging.local dockerd-current[23329]:
> I0108 00:26:49.910964       1 garbagecollector.go:291] processing item [
> apps.openshift.io/v1/DeploymentConfig, namespace: jenkins-test, name:
> jenkins, uid: 91759f72-f408-11e7-88e5-fa163eb8ca3a]
>
> Any ideas? Has anyone else seen this?  Considering
> "openshift-ansible-service-broker" is deployed in a broken state by
> openshift-ansible on the release-3.7 branch (for origin, I think enterprise
> would work as the tags exist), it makes me think that not many people are
> using the new service brokers that are talked about here:
> https://blog.openshift.com/whats-new-in-openshift-3-7-service-catalog-and-brokers/
>
> Thanks,
>
> Joel
>
> _______________________________________________
> users mailing list
> users@lists.openshift.redhat.com
> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>
>
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to