Hello all!

I have setup a cluster with 3 glusterFS nodes for disk persistence just as specified in the docs. I have configured the inventory file to install the containerized version to be used by Openshift's integrated registry. This works fine.

Now I wanted to install the metrics component and I followed the procedure described here: https://docs.openshift.org/latest/install_config/persistent_storage/persistent_storage_glusterfs.html#install-example-infra

I end up with openshift-infra project set up, but with 3 pods failing to start and I think this has to do with the PVC for cassandra that fails to create.

oc get pvc metrics-cassandra-1 -o yaml

apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  annotations:
    control-plane.alpha.kubernetes.io/leader: '{"holderIdentity":"8ef584d1-5923-11e8-8730-0a580a830040","leaseDurationSeconds":15,"acquireTime":"2018-05-17T00:38:34Z","renewTime":"2018-05-17T00:55:33Z","leaderTransitions":0}'
    kubectl.kubernetes.io/last-applied-configuration: |
{"apiVersion":"v1","kind":"PersistentVolumeClaim","metadata":{"annotations":{"volume.beta.kubernetes.io/storage-provisioner":"gluster.org/glusterblock"},"labels":{"metrics-infra":"hawkular-cassandra"},"name":"metrics-cassandra-1","namespace":"openshift-infra"},"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"6Gi"}},"storageClassName":"glusterfs-registry-block"}}
    volume.beta.kubernetes.io/storage-provisioner: gluster.org/glusterblock
  creationTimestamp: 2018-05-17T00:38:34Z
  labels:
    metrics-infra: hawkular-cassandra
  name: metrics-cassandra-1
  namespace: openshift-infra
  resourceVersion: "1204482"
  selfLink: /api/v1/namespaces/openshift-infra/persistentvolumeclaims/metrics-cassandra-1
  uid: a18b8c20-596a-11e8-8a63-fa163ed601cb
spec:
  accessModes:
  - ReadWriteOnce
  resources:
    requests:
      storage: 6Gi
  storageClassName: glusterfs-registry-block
status:
  phase: Pending

oc describe pvc metrics-cassandra-1 shows these warnings:

 36m        23m        13    gluster.org/glusterblock glusterblock-registry-provisioner-dc-1-tljbb 8ef584d1-5923-11e8-8730-0a580a830040            Warning ProvisioningFailed    Failed to provision volume with StorageClass "glusterfs-registry-block": failed to create volume: [heketi] failed to create volume: Failed to allocate new block volume: No space   36m        21m        14    gluster.org/glusterblock glusterblock-registry-provisioner-dc-1-tljbb 8ef584d1-5923-11e8-8730-0a580a830040            Normal Provisioning        External provisioner is provisioning volume for claim "openshift-infra/metrics-cassandra-1"   21m        21m        1    gluster.org/glusterblock glusterblock-registry-provisioner-dc-1-tljbb 8ef584d1-5923-11e8-8730-0a580a830040            Warning ProvisioningFailed    Failed to provision volume with StorageClass "glusterfs-registry-block": failed to create volume: [heketi] failed to create volume: Post http://heketi-registry-default.apps.my.net/blockvolumes: dial tcp: lookup heketi-registry-default.apps.my.net on 192.168.150.16:53: no such host

In the default project, if I check the logs for heketi-registry, I get a lot of

[heketi] ERROR 2018/05/17 00:46:47 /src/github.com/heketi/heketi/apps/glusterfs/operations.go:909: Create Block Volume Build Failed: No space
[negroni] Started POST /blockvolumes
[heketi] INFO 2018/05/17 00:49:02 Loaded simple allocator
[heketi] INFO 2018/05/17 00:49:02 brick_num: 0
[heketi] INFO 2018/05/17 00:49:02 brick_num: 0
[heketi] INFO 2018/05/17 00:49:02 brick_num: 0
[heketi] INFO 2018/05/17 00:49:02 brick_num: 0
[heketi] INFO 2018/05/17 00:49:02 brick_num: 1
[negroni] Completed 500 Internal Server Error in 7.091238ms

For the other glusterFS-related pod, I see the same errors reported by the pvc creation

oc logs -f glusterblock-registry-provisioner-dc-1-tljbb -n default

I0516 22:38:49.136388       1 controller.go:1167] scheduleOperation[lock-provision-openshift-infra/metrics-cassandra-1[1191fb8d-5959-11e8-94c9-fa163e1cba7f]] I0516 22:38:49.166658       1 leaderelection.go:156] attempting to acquire leader lease... I0516 22:38:49.197051       1 leaderelection.go:178] successfully acquired lease to provision for pvc openshift-infra/metrics-cassandra-1 I0516 22:38:49.197122       1 controller.go:1167] scheduleOperation[provision-openshift-infra/metrics-cassandra-1[1191fb8d-5959-11e8-94c9-fa163e1cba7f]] E0516 22:38:49.207257       1 glusterblock-provisioner.go:441] BLOCK VOLUME NAME I RECEIEVED: E0516 22:38:49.207288       1 glusterblock-provisioner.go:449] BLOCK VOLUME CREATE REQUEST: &{Size:6 Clusters:[] Name: Hacount:3 Auth:true} E0516 22:38:49.355122       1 glusterblock-provisioner.go:451] BLOCK VOLUME RESPONSE: <nil> E0516 22:38:49.355204       1 glusterblock-provisioner.go:453] [heketi] failed to create volume: Failed to allocate new block volume: No space E0516 22:38:49.355262       1 controller.go:895] Failed to provision volume for claim "openshift-infra/metrics-cassandra-1" with StorageClass "glusterfs-registry-block": failed to create volume: [heketi] failed to create volume: Failed to allocate new block volume: No space E0516 22:38:49.355365       1 goroutinemap.go:165] Operation for "provision-openshift-infra/metrics-cassandra-1[1191fb8d-5959-11e8-94c9-fa163e1cba7f]" failed. No retries permitted until 2018-05-16 22:40:51.355301022 +0000 UTC m=+23465.283195247 (durationBeforeRetry 2m2s). Error: "failed to create volume: [heketi] failed to create volume: Failed to allocate new block volume: No space" I0516 22:38:51.241605       1 leaderelection.go:198] stopped trying to renew lease to provision for pvc openshift-infra/metrics-cassandra-1, task failed

Regarding the no space message, I am certain that there is space on the device (if there isn't some glusterFS config that's done on the servers which prevents them to extend/create the volumes). All disks have the same 26GB capacity and lvs  on one of the machines shows:

LV VG                                  Attr       LSize Pool                                Origin Data%  Meta%  Move Log Cpy%Sync Convert   docker-pool rootvg                              twi-aot--- <4,16g                                            52,37 2,62
  home rootvg                              -wi-ao---- 1,00g
  root rootvg                              -wi-ao---- 2,00g
  swap rootvg                              -wi-a----- 2,00g
  tmp rootvg                              -wi-ao---- 1,17g
  usr rootvg                              -wi-ao---- 4,00g
  var rootvg                              -wi-ao---- 4,00g
  brick_7aa3a789badd1ae620a2bbefe51b8c73 vg_bd61a1e6f317bb9decade964449c12e8 Vwi-aotz--  2,00g tp_7aa3a789badd1ae620a2bbefe51b8c73 0,71   brick_8818ffee7ab2244ca721b7d15ea1e514 vg_bd61a1e6f317bb9decade964449c12e8 Vwi-aotz--  5,00g tp_8818ffee7ab2244ca721b7d15ea1e514 7,57   tp_7aa3a789badd1ae620a2bbefe51b8c73 vg_bd61a1e6f317bb9decade964449c12e8 twi-aotz-- 2,00g                                            0,71 0,33   tp_8818ffee7ab2244ca721b7d15ea1e514 vg_bd61a1e6f317bb9decade964449c12e8 twi-aotz-- 5,00g                                            7,57   0,29

Any ideas where to look for misconfigurations?

Thank you,
Dan Pungă
_______________________________________________
users mailing list
[email protected]
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to