Good find on that bug. Our upgrade guide can help you get started on a fix.


https://docs.openshift.com/container-platform/3.3/install_config/upgrading/index.html

Mark

On Fri, Nov 18, 2016 at 3:13 AM, Lionel Orellana <lione...@gmail.com> wrote:

>
> This sounds very very familiar: https://github.com/
> kubernetes/kubernetes/issues/30637
>
> Particularly comment: https://github.com/kubernetes/kubernetes/issues/
> 30637#issuecomment-243276076
>
> That is a nasty bug. How can I upgrade Kubernetes in my cluster?
>
> My current versions are
>
> -bash-4.2$ oc version
> oc v1.3.0
> kubernetes v1.3.0+52492b4
> features: Basic-Auth GSSAPI Kerberos SPNEGO
>
> Server https://poc-docker01.aipo.gov.au:8443
> openshift v1.3.0
> kubernetes v1.3.0+52492b4
>
>
> On 18 November 2016 at 18:18, Lionel Orellana <lione...@gmail.com> wrote:
>
>> Files in other dirs in the same NFS server don't get deleted (e.g.
>> <server name>/poc_runtime/test/)
>>
>> There is something in my Openshift node deleting files in <server
>> name>/poc_runtime/evs as soon as I put them there!
>>
>> On 18 November 2016 at 18:04, Lionel Orellana <lione...@gmail.com> wrote:
>>
>>>
>>> In fact, whatever is deleting my files is still doing it:
>>>
>>> [root@poc-docker03 evs]# touch x
>>> [root@poc-docker03 evs]# ls
>>> [root@poc-docker03 evs]#
>>>
>>> evs is a path on an NFS volume that I have added directly to some
>>> deployment configs
>>>
>>>  -
>>>           name: evs
>>>           nfs:
>>>             server: <server name>
>>>             path: /poc_runtime/evs
>>>
>>> If I stop the origin-service on one particular node the file doesn't
>>> disappear.
>>>
>>> [root@poc-docker03 evs]# touch x
>>> [root@poc-docker03 evs]# ls
>>> x
>>> [root@poc-docker03 evs]#
>>>
>>> When I restart the origin-node service I see a lot of errors like this
>>>
>>>  Failed cleaning pods: [remove /var/lib/origin/openshift.loca
>>> l.volumes/pods/1b7e3a16-ab08-11e6-8618-005056915814/volumes/
>>> kubernetes.io~nfs device or resource bus
>>>  Failed to remove orphaned pod xxxxx dir; err: remove
>>> /var/lib/origin/openshift.local.volumes/pods/xxxx/volumes/kubernetes.io
>>> ~nfs/*evs*: device or resource bus
>>>
>>> Despite the fact that the error says that it couldn't remove it, what
>>> exactly is it trying to do here? Is it possible that this process
>>> previously deleted the data in the evs folder?
>>>
>>>
>>>
>>>
>>> On 18 November 2016 at 16:45, Lionel Orellana <lione...@gmail.com>
>>> wrote:
>>>
>>>> What about NFS volumes added directly in build configs.
>>>>
>>>> volumes:
>>>>         -
>>>>           name: jenkins-volume-1
>>>>           nfs:
>>>>             server: <server name>
>>>>             path: /poc_runtime/jenkins/home
>>>>
>>>>
>>>> We just restarted all the servers hosting my openshift cluster and the
>>>> all data in the path above disappeared. Simply by restarting the host VM!
>>>>
>>>>
>>>>
>>>> On 18 November 2016 at 16:19, Lionel Orellana <lione...@gmail.com>
>>>> wrote:
>>>>
>>>>> Thanks Mark
>>>>>
>>>>> On 18 November 2016 at 15:09, Mark Turansky <mtura...@redhat.com>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Nov 17, 2016 at 10:41 PM, Lionel Orellana <lione...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> Couple of questions regarding Persistent Volumes, in particular NFS
>>>>>>> ones.
>>>>>>>
>>>>>>> 1) If I have a PV configured with the Retain policy it is not clear
>>>>>>> to me how this PV can be reused after the bound PVC is deleted. Deleting
>>>>>>> the PVC makes the PV status "Released". How do I make it "Available" 
>>>>>>> again
>>>>>>> without losing the data?
>>>>>>>
>>>>>>
>>>>>> You can keep the PVC around longer if you intend to reuse it between
>>>>>> pods. There is no way for a PV to go from Released to Available again in
>>>>>> your scenario. You would have to delete and recreate the PV. It's a 
>>>>>> pointer
>>>>>> to real storage (the NFS share), so you're just recreating the pointer. 
>>>>>> The
>>>>>> data in the NFS volume itself is untouched.
>>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> 2) Is there anything (e.g. all nodes crashing due to some underlying
>>>>>>> infrastructure failure) that would cause the data in a "Retain" volume 
>>>>>>> to
>>>>>>> be wiped out? We had a problem with all our vmware servers  (where I 
>>>>>>> host
>>>>>>> my openshift POC)  and all my NFS mounted volumes were wiped out. The
>>>>>>> storage guys assure me that nothing at their end caused that and it must
>>>>>>> have been a running process that did it.
>>>>>>>
>>>>>>
>>>>>> "Retain" is just a flag to the recycling process to leave that PV
>>>>>> alone when it's Released. The PV's retention policy wouldn't cause
>>>>>> everything to be deleted. NFS volumes on the node are no different than 
>>>>>> if
>>>>>> you called "mount" yourself. There is nothing inherent in OpenShift 
>>>>>> itself
>>>>>> that is running in that share that would wipe out data.
>>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> Thanks
>>>>>>>
>>>>>>> Lionel.
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> users@lists.openshift.redhat.com
>>>>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to