Good find on that bug. Our upgrade guide can help you get started on a fix.
https://docs.openshift.com/container-platform/3.3/install_config/upgrading/index.html Mark On Fri, Nov 18, 2016 at 3:13 AM, Lionel Orellana <lione...@gmail.com> wrote: > > This sounds very very familiar: https://github.com/ > kubernetes/kubernetes/issues/30637 > > Particularly comment: https://github.com/kubernetes/kubernetes/issues/ > 30637#issuecomment-243276076 > > That is a nasty bug. How can I upgrade Kubernetes in my cluster? > > My current versions are > > -bash-4.2$ oc version > oc v1.3.0 > kubernetes v1.3.0+52492b4 > features: Basic-Auth GSSAPI Kerberos SPNEGO > > Server https://poc-docker01.aipo.gov.au:8443 > openshift v1.3.0 > kubernetes v1.3.0+52492b4 > > > On 18 November 2016 at 18:18, Lionel Orellana <lione...@gmail.com> wrote: > >> Files in other dirs in the same NFS server don't get deleted (e.g. >> <server name>/poc_runtime/test/) >> >> There is something in my Openshift node deleting files in <server >> name>/poc_runtime/evs as soon as I put them there! >> >> On 18 November 2016 at 18:04, Lionel Orellana <lione...@gmail.com> wrote: >> >>> >>> In fact, whatever is deleting my files is still doing it: >>> >>> [root@poc-docker03 evs]# touch x >>> [root@poc-docker03 evs]# ls >>> [root@poc-docker03 evs]# >>> >>> evs is a path on an NFS volume that I have added directly to some >>> deployment configs >>> >>> - >>> name: evs >>> nfs: >>> server: <server name> >>> path: /poc_runtime/evs >>> >>> If I stop the origin-service on one particular node the file doesn't >>> disappear. >>> >>> [root@poc-docker03 evs]# touch x >>> [root@poc-docker03 evs]# ls >>> x >>> [root@poc-docker03 evs]# >>> >>> When I restart the origin-node service I see a lot of errors like this >>> >>> Failed cleaning pods: [remove /var/lib/origin/openshift.loca >>> l.volumes/pods/1b7e3a16-ab08-11e6-8618-005056915814/volumes/ >>> kubernetes.io~nfs device or resource bus >>> Failed to remove orphaned pod xxxxx dir; err: remove >>> /var/lib/origin/openshift.local.volumes/pods/xxxx/volumes/kubernetes.io >>> ~nfs/*evs*: device or resource bus >>> >>> Despite the fact that the error says that it couldn't remove it, what >>> exactly is it trying to do here? Is it possible that this process >>> previously deleted the data in the evs folder? >>> >>> >>> >>> >>> On 18 November 2016 at 16:45, Lionel Orellana <lione...@gmail.com> >>> wrote: >>> >>>> What about NFS volumes added directly in build configs. >>>> >>>> volumes: >>>> - >>>> name: jenkins-volume-1 >>>> nfs: >>>> server: <server name> >>>> path: /poc_runtime/jenkins/home >>>> >>>> >>>> We just restarted all the servers hosting my openshift cluster and the >>>> all data in the path above disappeared. Simply by restarting the host VM! >>>> >>>> >>>> >>>> On 18 November 2016 at 16:19, Lionel Orellana <lione...@gmail.com> >>>> wrote: >>>> >>>>> Thanks Mark >>>>> >>>>> On 18 November 2016 at 15:09, Mark Turansky <mtura...@redhat.com> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Thu, Nov 17, 2016 at 10:41 PM, Lionel Orellana <lione...@gmail.com >>>>>> > wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> Couple of questions regarding Persistent Volumes, in particular NFS >>>>>>> ones. >>>>>>> >>>>>>> 1) If I have a PV configured with the Retain policy it is not clear >>>>>>> to me how this PV can be reused after the bound PVC is deleted. Deleting >>>>>>> the PVC makes the PV status "Released". How do I make it "Available" >>>>>>> again >>>>>>> without losing the data? >>>>>>> >>>>>> >>>>>> You can keep the PVC around longer if you intend to reuse it between >>>>>> pods. There is no way for a PV to go from Released to Available again in >>>>>> your scenario. You would have to delete and recreate the PV. It's a >>>>>> pointer >>>>>> to real storage (the NFS share), so you're just recreating the pointer. >>>>>> The >>>>>> data in the NFS volume itself is untouched. >>>>>> >>>>>> >>>>>> >>>>>>> >>>>>>> 2) Is there anything (e.g. all nodes crashing due to some underlying >>>>>>> infrastructure failure) that would cause the data in a "Retain" volume >>>>>>> to >>>>>>> be wiped out? We had a problem with all our vmware servers (where I >>>>>>> host >>>>>>> my openshift POC) and all my NFS mounted volumes were wiped out. The >>>>>>> storage guys assure me that nothing at their end caused that and it must >>>>>>> have been a running process that did it. >>>>>>> >>>>>> >>>>>> "Retain" is just a flag to the recycling process to leave that PV >>>>>> alone when it's Released. The PV's retention policy wouldn't cause >>>>>> everything to be deleted. NFS volumes on the node are no different than >>>>>> if >>>>>> you called "mount" yourself. There is nothing inherent in OpenShift >>>>>> itself >>>>>> that is running in that share that would wipe out data. >>>>>> >>>>>> >>>>>> >>>>>>> >>>>>>> Thanks >>>>>>> >>>>>>> Lionel. >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.openshift.redhat.com >>>>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users