Would you mind posting the issue link here so I can keep up on it? I'm seeing some errors like those too.
-- Mateus Caruccio / Master of Puppets GetupCloud.com We make the infrastructure invisible Gartner Cool Vendor 2017 2017-09-05 18:28 GMT-03:00 Clayton Coleman <ccole...@redhat.com>: > Please open a bug in openshift/origin and we'll triage it there. > > On Tue, Sep 5, 2017 at 5:14 PM, Patrick Tescher <patr...@outtherelabs.com> > wrote: > >> The pods are still “terminating” and have been stuck in that state. New >> pods have come and gone since then but the stuck ones are still stuck. >> >> >> On Sep 5, 2017, at 2:13 PM, Clayton Coleman <ccole...@redhat.com> wrote: >> >> So the errors recur continuously for a given pod once they start >> happening? >> >> On Tue, Sep 5, 2017 at 5:07 PM, Patrick Tescher <patr...@outtherelabs.com >> > wrote: >> >>> No patches have been applied since we upgraded to 3.6.0 over a week ago. >>> The errors just popped up for a few different pods in different namespaces. >>> The only thing we did today was launch a stateful set in a new namespace. >>> Those pods were not the ones throwing this error. >>> >>> >>> On Sep 5, 2017, at 1:19 PM, Clayton Coleman <ccole...@redhat.com> wrote: >>> >>> Were any patches applied to the system? Some of these are normal if >>> they happen for a brief period of time. Are you seeing these errors >>> continuously for the same pod over and over? >>> >>> On Tue, Sep 5, 2017 at 3:23 PM, Patrick Tescher < >>> patr...@outtherelabs.com> wrote: >>> >>>> This morning our cluster started experiencing an odd error on multiple >>>> nodes. Pods are stuck in the terminating phase. In our node log I see the >>>> following: >>>> >>>> Sep 5 19:17:22 ip-10-0-1-184 origin-node: E0905 >>>> 19:17:22.043257 112306 nestedpendingoperations.go:262] Operation for >>>> "\"kubernetes.io/secret/182285ee-9267-11e7-b7be-06415eb17bbf >>>> -default-token-f18hx\ >>>> <http://kubernetes.io/secret/182285ee-9267-11e7-b7be-06415eb17bbf-default-token-f18hx%5C>" >>>> (\"182285ee-9267-11e7-b7be-06415eb17bbf\")" failed. No retries >>>> permitted until 2017-09-05 19:17:22.543230782 +0000 UTC >>>> (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for >>>> volume "kubernetes.io/secret/182285ee-9267-11e7-b7be-06415eb17bbf-d >>>> efault-token-f18hx" (volume.spec.Name <http://volume.spec.name/>: >>>> "default-token-f18hx") pod "182285ee-9267-11e7-b7be-06415eb17bbf" >>>> (UID: "182285ee-9267-11e7-b7be-06415eb17bbf") with: remove >>>> /var/lib/origin/openshift.local.volumes/pods/182285ee-9267-1 >>>> 1e7-b7be-06415eb17bbf/volumes/kubernetes.io~secret/default-token-f18hx: >>>> device or resource busy >>>> >>>> That path is not mounted (running mount does not list it) and running >>>> fuser -v on that directory does not show anything. Trying to rmdir results >>>> in a similar error: >>>> >>>> sudo rmdir var/lib/origin/openshift.local.volumes/pods/182285ee-9267-11 >>>> e7-b7be-06415eb17bbf/volumes/kubernetes.io~secret/default-token-f18hx >>>> rmdir: failed to remove ‘var/lib/origin/openshift.loca >>>> l.volumes/pods/182285ee-9267-11e7-b7be-06415eb17bbf/volumes/ >>>> kubernetes.io~secret/default-token-f18hx’: No such file or directory >>>> >>>> Is anyone else getting this error? >>>> >>>> >>>> _______________________________________________ >>>> dev mailing list >>>> dev@lists.openshift.redhat.com >>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/dev >>>> >>>> >>> >>> >> >> > > _______________________________________________ > dev mailing list > dev@lists.openshift.redhat.com > http://lists.openshift.redhat.com/openshiftmm/listinfo/dev > >
_______________________________________________ dev mailing list dev@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/dev