It's likely, don't have an eta yet while the scope of the pick is assessed.
On Thu, Nov 24, 2016 at 5:52 PM, Lionel Orellana <lione...@gmail.com> wrote: > This is a pretty bad issue in Kubernetes. We are talking about deleting > data from NFS volumes. Lucky for me I'm just doing a POC. Is this not > considered bad enough to warrant a patch release for Origin 1.3.x? > > Cheers > > Lionel. > > On 19 November 2016 at 07:38, Lionel Orellana <lione...@gmail.com> wrote: > >> The only "released" version of Openshift that includes Kubernetes 1.3.6 >> is v1.4.0.-alpha1. I don't want to upgrade to an alpha1 release. >> >> Can I request a patch of Openshift Origin to include Kubernetes 1.3.6 or >> higher? ( the Kubernetes 1.3 branch is up to 1.3.10). >> >> On 19 November 2016 at 07:26, Alex Wauck <alexwa...@exosite.com> wrote: >> >>> OpenShift is a distribution of Kubernetes, so I don't think you can >>> upgrade Kubernetes without upgrading OpenShift. >>> >>> On Fri, Nov 18, 2016 at 1:52 PM, Lionel Orellana <lione...@gmail.com> >>> wrote: >>> >>>> So the fix is on Kubernetes 1.3.6. The upgrade guide you mention is for >>>> Openshift as a whole unless I'm missing something. >>>> On Sat., 19 Nov. 2016 at 12:29 am, Mark Turansky <mtura...@redhat.com> >>>> wrote: >>>> >>>>> Good find on that bug. Our upgrade guide can help you get started on a >>>>> fix. >>>>> >>>>> https://docs.openshift.com/container-platform/3.3/install_co >>>>> nfig/upgrading/index.html >>>>> >>>>> Mark >>>>> >>>>> On Fri, Nov 18, 2016 at 3:13 AM, Lionel Orellana <lione...@gmail.com> >>>>> wrote: >>>>> >>>>> >>>>> This sounds very very familiar: https://github.com/k >>>>> ubernetes/kubernetes/issues/30637 >>>>> >>>>> Particularly comment: https://github.com/ku >>>>> bernetes/kubernetes/issues/30637#issuecomment-243276076 >>>>> >>>>> That is a nasty bug. How can I upgrade Kubernetes in my cluster? >>>>> >>>>> My current versions are >>>>> >>>>> -bash-4.2$ oc version >>>>> oc v1.3.0 >>>>> kubernetes v1.3.0+52492b4 >>>>> features: Basic-Auth GSSAPI Kerberos SPNEGO >>>>> >>>>> Server https://poc-docker01.aipo.gov.au:8443 >>>>> openshift v1.3.0 >>>>> kubernetes v1.3.0+52492b4 >>>>> >>>>> >>>>> On 18 November 2016 at 18:18, Lionel Orellana <lione...@gmail.com> >>>>> wrote: >>>>> >>>>> Files in other dirs in the same NFS server don't get deleted (e.g. >>>>> <server name>/poc_runtime/test/) >>>>> >>>>> There is something in my Openshift node deleting files in <server >>>>> name>/poc_runtime/evs as soon as I put them there! >>>>> >>>>> On 18 November 2016 at 18:04, Lionel Orellana <lione...@gmail.com> >>>>> wrote: >>>>> >>>>> >>>>> In fact, whatever is deleting my files is still doing it: >>>>> >>>>> [root@poc-docker03 evs]# touch x >>>>> [root@poc-docker03 evs]# ls >>>>> [root@poc-docker03 evs]# >>>>> >>>>> evs is a path on an NFS volume that I have added directly to some >>>>> deployment configs >>>>> >>>>> - >>>>> name: evs >>>>> nfs: >>>>> server: <server name> >>>>> path: /poc_runtime/evs >>>>> >>>>> If I stop the origin-service on one particular node the file doesn't >>>>> disappear. >>>>> >>>>> [root@poc-docker03 evs]# touch x >>>>> [root@poc-docker03 evs]# ls >>>>> x >>>>> [root@poc-docker03 evs]# >>>>> >>>>> When I restart the origin-node service I see a lot of errors like this >>>>> >>>>> Failed cleaning pods: [remove /var/lib/origin/openshift.loca >>>>> l.volumes/pods/1b7e3a16-ab08-11e6-8618-005056915814/volumes/ >>>>> kubernetes.io~nfs device or resource bus >>>>> Failed to remove orphaned pod xxxxx dir; err: remove >>>>> /var/lib/origin/openshift.local.volumes/pods/xxxx/volumes/ku >>>>> bernetes.io~nfs/*evs*: device or resource bus >>>>> >>>>> Despite the fact that the error says that it couldn't remove it, what >>>>> exactly is it trying to do here? Is it possible that this process >>>>> previously deleted the data in the evs folder? >>>>> >>>>> >>>>> >>>>> >>>>> On 18 November 2016 at 16:45, Lionel Orellana <lione...@gmail.com> >>>>> wrote: >>>>> >>>>> What about NFS volumes added directly in build configs. >>>>> >>>>> volumes: >>>>> - >>>>> name: jenkins-volume-1 >>>>> nfs: >>>>> server: <server name> >>>>> path: /poc_runtime/jenkins/home >>>>> >>>>> >>>>> We just restarted all the servers hosting my openshift cluster and the >>>>> all data in the path above disappeared. Simply by restarting the host VM! >>>>> >>>>> >>>>> >>>>> On 18 November 2016 at 16:19, Lionel Orellana <lione...@gmail.com> >>>>> wrote: >>>>> >>>>> Thanks Mark >>>>> >>>>> On 18 November 2016 at 15:09, Mark Turansky <mtura...@redhat.com> >>>>> wrote: >>>>> >>>>> >>>>> >>>>> On Thu, Nov 17, 2016 at 10:41 PM, Lionel Orellana <lione...@gmail.com> >>>>> wrote: >>>>> >>>>> Hi, >>>>> >>>>> Couple of questions regarding Persistent Volumes, in particular NFS >>>>> ones. >>>>> >>>>> 1) If I have a PV configured with the Retain policy it is not clear to >>>>> me how this PV can be reused after the bound PVC is deleted. Deleting the >>>>> PVC makes the PV status "Released". How do I make it "Available" again >>>>> without losing the data? >>>>> >>>>> >>>>> You can keep the PVC around longer if you intend to reuse it between >>>>> pods. There is no way for a PV to go from Released to Available again in >>>>> your scenario. You would have to delete and recreate the PV. It's a >>>>> pointer >>>>> to real storage (the NFS share), so you're just recreating the pointer. >>>>> The >>>>> data in the NFS volume itself is untouched. >>>>> >>>>> >>>>> >>>>> >>>>> 2) Is there anything (e.g. all nodes crashing due to some underlying >>>>> infrastructure failure) that would cause the data in a "Retain" volume to >>>>> be wiped out? We had a problem with all our vmware servers (where I host >>>>> my openshift POC) and all my NFS mounted volumes were wiped out. The >>>>> storage guys assure me that nothing at their end caused that and it must >>>>> have been a running process that did it. >>>>> >>>>> >>>>> "Retain" is just a flag to the recycling process to leave that PV >>>>> alone when it's Released. The PV's retention policy wouldn't cause >>>>> everything to be deleted. NFS volumes on the node are no different than if >>>>> you called "mount" yourself. There is nothing inherent in OpenShift itself >>>>> that is running in that share that would wipe out data. >>>>> >>>>> >>>>> >>>>> >>>>> Thanks >>>>> >>>>> Lionel. >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.openshift.redhat.com >>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.openshift.redhat.com >>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >>>> >>>> >>> >>> >>> -- >>> >>> Alex Wauck // DevOps Engineer >>> >>> *E X O S I T E* >>> *www.exosite.com <http://www.exosite.com/>* >>> >>> Making Machines More Human. >>> >>> >> > > _______________________________________________ > users mailing list > users@lists.openshift.redhat.com > http://lists.openshift.redhat.com/openshiftmm/listinfo/users > >
_______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users