I believe I'm seeing a problem with using GlusterFS volumes when you terminate a pod that is using a gluster backed PVC. This is with Origin 3.7.1. I did this:

1. create new project
2. deployed a pod
3. added a volume to the pod usingĀ  a gluster backed PVC.
4. rsh to the pod and check the volume can be written to
5. delete the project

After stage 3 the volume was working OK in the pod and the volume was reported by hekati.

After stage 5 the PVC was no longer present, glusterfs volume was no longer see by hekati (so far so good) but the pod was stuck in the 'Terminating' state and the project did not get deleted. It looks like the container that was running in the pod had been deleted. Even after one hour it was still stuck in the terminating state.

Looking deeper it looks like the mount on the host on which the pod was running was still present. e.g. this was still found in /etc/mtab:

10.0.0.15:vol_a8866bf3769c987aee5c919305b89529 /var/lib/origin/openshift.local.volumes/pods/51a4ef9e-44b4-11e8-b523-fa163ea80da9/volumes/kubernetes.io~glusterfs/pvc-28d4eb2e-44b4-11e8-b523-fa163ea80da9 fuse.glusterfs rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072 0 0

Manually unmounting this mount resulted in the pod finally terminating and (after a short delay) the project being deleted.

Looks like the cleanup processes are not quite correct?

Tim



_______________________________________________
users mailing list
[email protected]
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to