Files in other dirs in the same NFS server don't get deleted (e.g. <server name>/poc_runtime/test/)
There is something in my Openshift node deleting files in <server name>/poc_runtime/evs as soon as I put them there! On 18 November 2016 at 18:04, Lionel Orellana <lione...@gmail.com> wrote: > > In fact, whatever is deleting my files is still doing it: > > [root@poc-docker03 evs]# touch x > [root@poc-docker03 evs]# ls > [root@poc-docker03 evs]# > > evs is a path on an NFS volume that I have added directly to some > deployment configs > > - > name: evs > nfs: > server: <server name> > path: /poc_runtime/evs > > If I stop the origin-service on one particular node the file doesn't > disappear. > > [root@poc-docker03 evs]# touch x > [root@poc-docker03 evs]# ls > x > [root@poc-docker03 evs]# > > When I restart the origin-node service I see a lot of errors like this > > Failed cleaning pods: [remove /var/lib/origin/openshift. > local.volumes/pods/1b7e3a16-ab08-11e6-8618-005056915814/volumes/ > kubernetes.io~nfs device or resource bus > Failed to remove orphaned pod xxxxx dir; err: remove > /var/lib/origin/openshift.local.volumes/pods/xxxx/volumes/kubernetes.io > ~nfs/*evs*: device or resource bus > > Despite the fact that the error says that it couldn't remove it, what > exactly is it trying to do here? Is it possible that this process > previously deleted the data in the evs folder? > > > > > On 18 November 2016 at 16:45, Lionel Orellana <lione...@gmail.com> wrote: > >> What about NFS volumes added directly in build configs. >> >> volumes: >> - >> name: jenkins-volume-1 >> nfs: >> server: <server name> >> path: /poc_runtime/jenkins/home >> >> >> We just restarted all the servers hosting my openshift cluster and the >> all data in the path above disappeared. Simply by restarting the host VM! >> >> >> >> On 18 November 2016 at 16:19, Lionel Orellana <lione...@gmail.com> wrote: >> >>> Thanks Mark >>> >>> On 18 November 2016 at 15:09, Mark Turansky <mtura...@redhat.com> wrote: >>> >>>> >>>> >>>> On Thu, Nov 17, 2016 at 10:41 PM, Lionel Orellana <lione...@gmail.com> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> Couple of questions regarding Persistent Volumes, in particular NFS >>>>> ones. >>>>> >>>>> 1) If I have a PV configured with the Retain policy it is not clear to >>>>> me how this PV can be reused after the bound PVC is deleted. Deleting the >>>>> PVC makes the PV status "Released". How do I make it "Available" again >>>>> without losing the data? >>>>> >>>> >>>> You can keep the PVC around longer if you intend to reuse it between >>>> pods. There is no way for a PV to go from Released to Available again in >>>> your scenario. You would have to delete and recreate the PV. It's a pointer >>>> to real storage (the NFS share), so you're just recreating the pointer. The >>>> data in the NFS volume itself is untouched. >>>> >>>> >>>> >>>>> >>>>> 2) Is there anything (e.g. all nodes crashing due to some underlying >>>>> infrastructure failure) that would cause the data in a "Retain" volume to >>>>> be wiped out? We had a problem with all our vmware servers (where I host >>>>> my openshift POC) and all my NFS mounted volumes were wiped out. The >>>>> storage guys assure me that nothing at their end caused that and it must >>>>> have been a running process that did it. >>>>> >>>> >>>> "Retain" is just a flag to the recycling process to leave that PV alone >>>> when it's Released. The PV's retention policy wouldn't cause everything to >>>> be deleted. NFS volumes on the node are no different than if you called >>>> "mount" yourself. There is nothing inherent in OpenShift itself that is >>>> running in that share that would wipe out data. >>>> >>>> >>>> >>>>> >>>>> Thanks >>>>> >>>>> Lionel. >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.openshift.redhat.com >>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >>>>> >>>>> >>>> >>> >> >
_______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users