OpenShift is a distribution of Kubernetes, so I don't think you can upgrade Kubernetes without upgrading OpenShift.
On Fri, Nov 18, 2016 at 1:52 PM, Lionel Orellana <lione...@gmail.com> wrote: > So the fix is on Kubernetes 1.3.6. The upgrade guide you mention is for > Openshift as a whole unless I'm missing something. > On Sat., 19 Nov. 2016 at 12:29 am, Mark Turansky <mtura...@redhat.com> > wrote: > >> Good find on that bug. Our upgrade guide can help you get started on a >> fix. >> >> https://docs.openshift.com/container-platform/3.3/ >> install_config/upgrading/index.html >> >> Mark >> >> On Fri, Nov 18, 2016 at 3:13 AM, Lionel Orellana <lione...@gmail.com> >> wrote: >> >> >> This sounds very very familiar: https://github.com/ >> kubernetes/kubernetes/issues/30637 >> >> Particularly comment: https://github.com/kubernetes/kubernetes/issues/ >> 30637#issuecomment-243276076 >> >> That is a nasty bug. How can I upgrade Kubernetes in my cluster? >> >> My current versions are >> >> -bash-4.2$ oc version >> oc v1.3.0 >> kubernetes v1.3.0+52492b4 >> features: Basic-Auth GSSAPI Kerberos SPNEGO >> >> Server https://poc-docker01.aipo.gov.au:8443 >> openshift v1.3.0 >> kubernetes v1.3.0+52492b4 >> >> >> On 18 November 2016 at 18:18, Lionel Orellana <lione...@gmail.com> wrote: >> >> Files in other dirs in the same NFS server don't get deleted (e.g. >> <server name>/poc_runtime/test/) >> >> There is something in my Openshift node deleting files in <server >> name>/poc_runtime/evs as soon as I put them there! >> >> On 18 November 2016 at 18:04, Lionel Orellana <lione...@gmail.com> wrote: >> >> >> In fact, whatever is deleting my files is still doing it: >> >> [root@poc-docker03 evs]# touch x >> [root@poc-docker03 evs]# ls >> [root@poc-docker03 evs]# >> >> evs is a path on an NFS volume that I have added directly to some >> deployment configs >> >> - >> name: evs >> nfs: >> server: <server name> >> path: /poc_runtime/evs >> >> If I stop the origin-service on one particular node the file doesn't >> disappear. >> >> [root@poc-docker03 evs]# touch x >> [root@poc-docker03 evs]# ls >> x >> [root@poc-docker03 evs]# >> >> When I restart the origin-node service I see a lot of errors like this >> >> Failed cleaning pods: [remove /var/lib/origin/openshift. >> local.volumes/pods/1b7e3a16-ab08-11e6-8618-005056915814/volumes/ >> kubernetes.io~nfs device or resource bus >> Failed to remove orphaned pod xxxxx dir; err: remove >> /var/lib/origin/openshift.local.volumes/pods/xxxx/volumes/kubernetes.io >> ~nfs/*evs*: device or resource bus >> >> Despite the fact that the error says that it couldn't remove it, what >> exactly is it trying to do here? Is it possible that this process >> previously deleted the data in the evs folder? >> >> >> >> >> On 18 November 2016 at 16:45, Lionel Orellana <lione...@gmail.com> wrote: >> >> What about NFS volumes added directly in build configs. >> >> volumes: >> - >> name: jenkins-volume-1 >> nfs: >> server: <server name> >> path: /poc_runtime/jenkins/home >> >> >> We just restarted all the servers hosting my openshift cluster and the >> all data in the path above disappeared. Simply by restarting the host VM! >> >> >> >> On 18 November 2016 at 16:19, Lionel Orellana <lione...@gmail.com> wrote: >> >> Thanks Mark >> >> On 18 November 2016 at 15:09, Mark Turansky <mtura...@redhat.com> wrote: >> >> >> >> On Thu, Nov 17, 2016 at 10:41 PM, Lionel Orellana <lione...@gmail.com> >> wrote: >> >> Hi, >> >> Couple of questions regarding Persistent Volumes, in particular NFS ones. >> >> 1) If I have a PV configured with the Retain policy it is not clear to me >> how this PV can be reused after the bound PVC is deleted. Deleting the PVC >> makes the PV status "Released". How do I make it "Available" again without >> losing the data? >> >> >> You can keep the PVC around longer if you intend to reuse it between >> pods. There is no way for a PV to go from Released to Available again in >> your scenario. You would have to delete and recreate the PV. It's a pointer >> to real storage (the NFS share), so you're just recreating the pointer. The >> data in the NFS volume itself is untouched. >> >> >> >> >> 2) Is there anything (e.g. all nodes crashing due to some underlying >> infrastructure failure) that would cause the data in a "Retain" volume to >> be wiped out? We had a problem with all our vmware servers (where I host >> my openshift POC) and all my NFS mounted volumes were wiped out. The >> storage guys assure me that nothing at their end caused that and it must >> have been a running process that did it. >> >> >> "Retain" is just a flag to the recycling process to leave that PV alone >> when it's Released. The PV's retention policy wouldn't cause everything to >> be deleted. NFS volumes on the node are no different than if you called >> "mount" yourself. There is nothing inherent in OpenShift itself that is >> running in that share that would wipe out data. >> >> >> >> >> Thanks >> >> Lionel. >> >> _______________________________________________ >> users mailing list >> users@lists.openshift.redhat.com >> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >> >> >> >> >> >> >> >> >> > _______________________________________________ > users mailing list > users@lists.openshift.redhat.com > http://lists.openshift.redhat.com/openshiftmm/listinfo/users > > -- Alex Wauck // DevOps Engineer *E X O S I T E* *www.exosite.com <http://www.exosite.com/>* Making Machines More Human.
_______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users