This is an automated email from the ASF dual-hosted git repository.

zhongxjian pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/dubbo-kubernetes.git


The following commit(s) were added to refs/heads/master by this push:
     new 6fcef42a delete base resource
6fcef42a is described below

commit 6fcef42ad12b01ddead9505ec29a101b20565177
Author: mfordjody <[email protected]>
AuthorDate: Sun Apr 20 10:09:12 2025 +0800

    delete base resource
---
 .../charts/base/templates/ingress-controller.yaml  |  85 -------
 manifests/charts/base/templates/ingress.yaml       |  72 ------
 manifests/charts/base/templates/rook-jobs.yaml     | 259 ---------------------
 manifests/charts/base/values.yaml                  |  64 +----
 4 files changed, 1 insertion(+), 479 deletions(-)

diff --git a/manifests/charts/base/templates/ingress-controller.yaml 
b/manifests/charts/base/templates/ingress-controller.yaml
deleted file mode 100644
index 0d51d866..00000000
--- a/manifests/charts/base/templates/ingress-controller.yaml
+++ /dev/null
@@ -1,85 +0,0 @@
-apiVersion: v1
-kind: ServiceAccount
-metadata:
-  name: dubbo-kubernetes-ingress-controller
-  namespace: {{ .Release.Namespace }}
----
-apiVersion: rbac.authorization.k8s.io/v1
-kind: ClusterRole
-metadata:
-  name: dubbo-kubernetes-ingress-controller-clusterRole
-rules:
-  - apiGroups: [""]
-    resources: ["services", "secrets"]
-    verbs: ["get", "watch", "list"]
-  - apiGroups: ["extensions", "networking.k8s.io"]
-    resources: ["ingresses"]
-    verbs: ["get", "watch", "list", "update"]
-  - apiGroups: ["networking.k8s.io"]
-    resources: ["ingresses/status"]
-    verbs: ["update"]
----
-apiVersion: rbac.authorization.k8s.io/v1
-kind: ClusterRoleBinding
-metadata:
-  name: dubbo-kubernetes-ingress-controller-clusterRoleBinding
-roleRef:
-  apiGroup: rbac.authorization.k8s.io
-  kind: ClusterRole
-  name: dubbo-kubernetes-ingress-controller-clusterRole
-subjects:
-  - kind: ServiceAccount
-    name: dubbo-kubernetes-ingress-controller
-    namespace: {{ .Release.Namespace }}
----
-apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: dubbo-kubernetes-ingress-controller
-  namespace: {{ .Release.Namespace }}
-  labels:
-    app: ingress-controller
-spec:
-  replicas: 1
-  selector:
-    matchLabels:
-      app: ingress-controller
-  template:
-    metadata:
-      labels:
-        app: ingress-controller
-    spec:
-      hostNetwork: true
-      serviceAccountName: dubbo-kubernetes-ingress-controller
-      containers:
-      - name: dubbo-kubernetes-ingress-controller
-        image: docker.io/mfordjody/dubbo-kubernetes-ingress-controller:dev
-        imagePullPolicy: IfNotPresent
-        resources:
-          limits:
-            cpu: "100m"
-            memory: "100Mi"
-          requests:
-            cpu: "100m"
-            memory: "100Mi"
----
-apiVersion: networking.k8s.io/v1
-kind: IngressClass
-metadata:
-  name: dubbo
-spec:
-  controller: "dubbo.io/dubbo-kubernetes-ingress-controller"
----
-{{ $ingress := .Values.ingress }}
-{{- $ca := genCA "dubbo-kubernetes-ca" 365 }}
-{{- $cert := genSignedCert "" nil (list "dubbo-kubernetes-user") 365 $ca }}
-apiVersion: v1
-kind: Secret
-metadata:
-  name: dubbo-kubernetes-ingress-tls
-  namespace: {{ .Release.Namespace }}
-type: kubernetes.io/tls
-data:
-  tls.crt: {{ $cert.Cert | b64enc | quote }}
-  tls.key: {{ $cert.Key | b64enc | quote }}
-  ca.crt: {{ $ca.Cert | b64enc | quote }}
diff --git a/manifests/charts/base/templates/ingress.yaml 
b/manifests/charts/base/templates/ingress.yaml
deleted file mode 100644
index d7971c57..00000000
--- a/manifests/charts/base/templates/ingress.yaml
+++ /dev/null
@@ -1,72 +0,0 @@
-{{ $ingress := .Values.ingress }}
-apiVersion: networking.k8s.io/v1
-kind: Ingress
-metadata:
-  name: admin-dashboard
-  annotations:
-    nginx.ingress.kubernetes.io/ssl-redirect: "true"
-spec:
-  tls:
-  - hosts:
-    - {{ $ingress.hosts.admin }}
-    secretName: dubbo-kubernetes-ingress-tls
-  ingressClassName: {{ $ingress.class |default "dubbo" }}
-  rules:
-  - host: {{ $ingress.hosts.admin | default "dubbo" }}
-    http:
-      paths:
-        - path: /
-          pathType: Prefix
-          backend:
-            service:
-              name: admin-dashboard
-              port:
-                number: 8080
----
-apiVersion: networking.k8s.io/v1
-kind: Ingress
-metadata:
-  name: kube-prometheus-prometheus
-  annotations:
-    nginx.ingress.kubernetes.io/ssl-redirect: "true"
-spec:
-  tls:
-  - hosts:
-    - {{ $ingress.hosts.prometheus }}
-    secretName: dubbo-kubernetes-ingress-tls
-  ingressClassName: {{ $ingress.class | default "dubbo" }}
-  rules:
-  - host: {{ $ingress.hosts.prometheus }}
-    http:
-      paths:
-        - path: /
-          pathType: Prefix
-          backend:
-            service:
-              name: kube-prometheus-prometheus
-              port:
-                number: 9090
----
-apiVersion: networking.k8s.io/v1
-kind: Ingress
-metadata:
-  name: grafana
-  annotations:
-    nginx.ingress.kubernetes.io/ssl-redirect: "true"
-spec:
-  tls:
-  - hosts:
-    - {{ $ingress.hosts.grafana }}
-    secretName: dubbo-kubernetes-ingress-tls
-  ingressClassName: {{ $ingress.class | default "dubbo" }}
-  rules:
-  - host: {{ $ingress.hosts.grafana }}
-    http:
-      paths:
-        - path: /
-          pathType: Prefix
-          backend:
-            service:
-              name: grafana
-              port:
-                number: 80
diff --git a/manifests/charts/base/templates/rook-jobs.yaml 
b/manifests/charts/base/templates/rook-jobs.yaml
deleted file mode 100644
index fb04bc87..00000000
--- a/manifests/charts/base/templates/rook-jobs.yaml
+++ /dev/null
@@ -1,259 +0,0 @@
-{{ $rook := .Values.csi }}
-apiVersion: v1
-kind: ConfigMap
-metadata:
-  name: rook-csi-driver-config
-  namespace: {{ .Release.Namespace }}
-  labels:
-    app: rook-csi-driver
-data:
-  rook-setup.sh: |-
-    #!/usr/bin/env bash
-
-    set -ex
-
-    if [ ! -d "rook" ]; then
-    echo "Pull Rook Project"
-    git clone https://github.com/rook/rook.git
-    else
-    echo "Rook Project Exist,Continue editing..."
-    fi
-
-    cd rook/deploy/examples/
-    sed -i 's/ROOK_DISCOVER_DEVICES_INTERVAL: 
"60m"/ROOK_DISCOVER_DEVICES_INTERVAL: "60s"/' operator.yaml
-    sed -i 's/useAllNodes: true/useAllNodes: false/' cluster.yaml
-    sed -i 's/useAllDevices: true/useAllDevices: false/' cluster.yaml
-    sed -i '/^  mon:/,/^    count:/s/^\(    count:\) [0-9]\+/\1 1/' 
cluster.yaml
-    sed -i '/mon:/, /allowMultiplePerNode:/ s/allowMultiplePerNode: 
false/allowMultiplePerNode: true/' cluster.yaml
-    sed -i '/^  mgr:/,/^    count:/s/^\(    count:\) [0-9]\+/\1 1/' 
cluster.yaml
-    sed -i '/resources:/,/^[^ ]/ {
-      /resources:/c\
-      resources:\
-        mon:\
-          limits:\
-            cpu: {{ $rook.mon.limits.cpu | default "102m" | quote }}\
-            memory: {{ $rook.mon.limits.memory | default "256Mi" | quote }}\
-          requests:\
-            cpu: {{ $rook.mon.requests.cpu | default "102m" | quote }}\
-            memory: {{ $rook.mon.requests.memory | default "256Mi" | quote }}\
-        mgr:\
-          limits:\
-            cpu: {{ $rook.mgr.limits.cpu | default "102m" | quote }}\
-            memory: {{ $rook.mgr.limits.memory | default "512Mi" | quote }}\
-          requests:\
-            cpu: {{ $rook.mgr.requests.cpu | default "102m" | quote }}\
-            memory: {{ $rook.mgr.requests.memory | default "512Mi" | quote }}\
-        mds:\
-          limits:\
-            cpu: {{ $rook.mds.limits.cpu | default "102m" | quote }}\
-            memory: {{ $rook.mds.limits.memory | default "256Mi" | quote }}\
-          requests:\
-            cpu: {{ $rook.mds.requests.cpu | default "102m" | quote }}\
-            memory: {{ $rook.mds.requests.memory | default "256Mi" | quote }}\
-        osd:\
-          limits:\
-            cpu: {{ $rook.osd.limits.cpu | default "512m" | quote }}\
-            memory: {{ $rook.osd.limits.memory | default "2048Mi" | quote }}\
-          requests:\
-            cpu: {{ $rook.osd.requests.cpu | default "512m" | quote }}\
-            memory: {{ $rook.osd.requests.memory | default "2048Mi" | quote }}
-    }' cluster.yaml
-
-    sed -i 's/^  # placement:/  placement:/' cluster.yaml
-    sed -i '/placement:/,/^[^ ]/ {
-      /placement:/c\
-      placement:\
-        mon:\
-          nodeAffinity:\
-            requiredDuringSchedulingIgnoredDuringExecution:\
-              nodeSelectorTerms:\
-              - matchExpressions:\
-                - key: ceph-mon\
-                  operator: In\
-                  values:\
-                  - enabled\
-        mgr:\
-          nodeAffinity:\
-            requiredDuringSchedulingIgnoredDuringExecution:\
-              nodeSelectorTerms:\
-              - matchExpressions:\
-                - key: ceph-mgr\
-                  operator: In\
-                  values:\
-                  - enabled\
-        mds:\
-          nodeAffinity:\
-            requiredDuringSchedulingIgnoredDuringExecution:\
-              nodeSelectorTerms:\
-              - matchExpressions:\
-                - key: ceph-mds\
-                  operator: In\
-                  values:\
-                  - enabled\
-        osd:\
-          nodeAffinity:\
-            requiredDuringSchedulingIgnoredDuringExecution:\
-              nodeSelectorTerms:\
-              - matchExpressions:\
-                - key: ceph-osd\
-                  operator: In\
-                  values:\
-                  - enabled
-    }' cluster.yaml
-
-    sed -i 's/^    # nodes:/    nodes:/' cluster.yaml
-    sed -i '/nodes:/,/^[^ ]/ {
-      /nodes:/c\
-        nodes:\
-        - name: {{ $rook.nodes.name | quote }}\
-          devices: # specific devices to use for storage can be specified for 
each node\
-            - name: {{ $rook.nodes.devices.name | quote }}\
-              config:\
-                storeType: {{ $rook.nodes.devices.storeType  }}\
-                journalSizeMB: {{ $rook.nodes.devices.journalSizeMB | default 
"4096" | quote }}
-    }' cluster.yaml
-
-    kubectl label node {{ $rook.labelSelector.node }} ceph-mon=enabled
-    kubectl label node {{ $rook.labelSelector.node }} ceph-mgr=enabled
-    kubectl label node {{ $rook.labelSelector.node }} ceph-mds=enabled
-    kubectl label node {{ $rook.labelSelector.node }} ceph-osd=enabled
-
-    kubectl apply -f crds.yaml -f common.yaml -f operator.yaml
-    kubectl apply -f cluster.yaml -f toolbox.yaml -f 
dashboard-external-https.yaml
-
-    # 至少需要 1 个 osd
-    sed -i '/resources:/,/^[^ ]/ {
-      /resources:/c\
-        resources:\
-          limits:\
-            cpu: {{ $rook.cephfs.limits.cpu | default "256m" | quote}}\
-            memory: {{ $rook.cephfs.limits.memory | default "512Mi" | quote }}\
-          requests:\
-            cpu: {{ $rook.cephfs.requests.cpu | default "256m" | quote }}\
-            memory: {{ $rook.cephfs.requests.memory | default "512Mi" | quote 
}}
-    }' filesystem-test.yaml
-
-    sed -i '/podAntiAffinity:/,/- rook-ceph-mds/ s/^/      # /' 
filesystem-test.yaml
-    sed -i '/preferredDuringSchedulingIgnoredDuringExecution:/,/- 
rook-ceph-mds/ s/^/      # /' filesystem-test.yaml
-    sed -i '/topologyKey: kubernetes.io\/hostname/ s/^/            #/' 
filesystem-test.yaml
-    sed -i '/topologyKey: topology.kubernetes.io\/zone/ s/^/            #/' 
filesystem-test.yaml
-    sed -i '/podAntiAffinity:/,/- rook-ceph-mds/ s/^/      # /' 
filesystem-test.yaml
-    sed -i '/preferredDuringSchedulingIgnoredDuringExecution:/,/- 
rook-ceph-mds/ s/^/      # /' filesystem-test.yaml
-    sed -i '/topologyKey: kubernetes.io\/hostname/ s/^/            #/' 
filesystem-test.yaml
-    sed -i '/topologyKey: topology.kubernetes.io\/zone/ s/^/            #/' 
filesystem-test.yaml
-    sed -i '/placement:/,/^[^ ]/ {
-      /placement:/c\
-        placement:\
-          nodeAffinity:\
-            requiredDuringSchedulingIgnoredDuringExecution:\
-              nodeSelectorTerms:\
-              - matchExpressions:\
-                - key: ceph-mds\
-                  operator: In\
-                  values:\
-                  - enabled
-    }' filesystem-test.yaml
-
-    # 至少需要 1 个 osd
-    kubectl apply -f filesystem-test.yaml ; kubectl apply -f 
csi/cephfs/storageclass.yaml
-
-    sed -i '/^spec:/a\
-    resources:\
-    limits:\
-    cpu: "256m"\
-    memory: "512Mi"\
-    requests:\
-    cpu: "256m"\
-    memory: "512Mi"
-    ' csi/rbd/storageclass-test.yaml
-
-    # 至少需要 1 个 osd
-    kubectl apply -f csi/rbd/storageclass-test.yaml
----
-apiVersion: batch/v1
-kind: Job
-metadata:
-  name: rook-job
-  namespace: {{ .Release.Namespace }}
-  labels:
-    app: rook-csi-driver
-spec:
-  template:
-    metadata:
-      name: rook-job
-    spec:
-      hostNetwork: true
-      restartPolicy: OnFailure
-      serviceAccountName: rook-admin
-      containers:
-        - name: rook-setup
-          image: mfordjody/rook-setup:dev
-          imagePullPolicy: IfNotPresent
-          command:
-            - /bin/sh
-            - -c
-            - sh /scripts/rook-setup.sh
-          volumeMounts:
-            - name: scripts
-              mountPath: /scripts
-      volumes:
-        - name: scripts
-          configMap:
-            name: rook-csi-driver-config
-            defaultMode: 0755
----
-apiVersion: v1
-kind: ServiceAccount
-metadata:
-  name: rook-admin
-  namespace: {{ .Release.Namespace }}
----
-apiVersion: rbac.authorization.k8s.io/v1
-kind: ClusterRole
-metadata:
-  name: rook-admin-role
-  namespace: {{ .Release.Namespace }}
-rules:
-  - apiGroups: [""]
-    resources: ["nodes"]
-    verbs: ["get", "list", "watch", "patch", "update"]
-  - apiGroups: [""]
-    resources: ["pods", "services", "configmaps", "namespaces"]
-    verbs: ["*"]
-  - apiGroups: ["apps"]
-    resources: ["deployments"]
-    verbs: ["*"]
-  - apiGroups: ["batch"]
-    resources: ["jobs"]
-    verbs: ["*"]
-  - apiGroups: ["apiextensions.k8s.io"]
-    resources: ["customresourcedefinitions"]
-    verbs: ["*"]
-  - apiGroups: ["rbac.authorization.k8s.io"]
-    resources: ["clusterroles", "clusterrolebindings"]
-    verbs: ["*"]
-  - apiGroups: ["rbac.authorization.k8s.io"]
-    resources: ["roles", "rolebindings"]
-    verbs: ["*"]
-  - apiGroups: [""]
-    resources: ["serviceaccounts"]
-    verbs: ["*"]
-  - apiGroups: ["ceph.rook.io"]
-    resources: 
["cephclusters","cephfilesystemsubvolumegroups","cephfilesystems","cephblockpools"]
-    verbs: ["*"]
-  - apiGroups: ["storage.k8s.io"]
-    resources: ["storageclasses"]
-    verbs: ["*"]
----
-apiVersion: rbac.authorization.k8s.io/v1
-kind: ClusterRoleBinding
-metadata:
-  name: rook-admin-binding
-roleRef:
-  apiGroup: rbac.authorization.k8s.io
-  kind: ClusterRole
-  name: rook-admin-role
-subjects:
-  - kind: ServiceAccount
-    name: rook-admin
-    namespace: {{ .Release.Namespace }}
\ No newline at end of file
diff --git a/manifests/charts/base/values.yaml 
b/manifests/charts/base/values.yaml
index 80ccdc37..69ad003d 100644
--- a/manifests/charts/base/values.yaml
+++ b/manifests/charts/base/values.yaml
@@ -15,66 +15,4 @@
 
 _internal_default_values_not_set:
   global: {}
-  base: {}
-
-  ingress:
-    class: ""
-    hosts:
-      admin: admin.example.local
-      prometheus: prom.example.local
-      grafana: grafana.example.local
-
-  csi:
-    labelSelector:
-      node: ""
-    mon:
-      limits:
-        cpu: ""
-        memory: ""
-      requests:
-        cpu: ""
-        memory: ""
-    mgr:
-      limits:
-        cpu: ""
-        memory: ""
-      requests:
-        cpu: ""
-        memory: ""
-    mds:
-      limits:
-        cpu: ""
-        memory: ""
-      requests:
-        cpu: ""
-        memory: ""
-    osd:
-      limits:
-        cpu: ""
-        memory: ""
-      requests:
-        cpu: ""
-        memory: ""
-
-    nodes:
-      name: ""
-      devices:
-        name: ""
-        storeType: bluestore
-        journalSizeMB: ""
-
-    rbd:
-      limits:
-        cpu: ""
-        memory: ""
-      requests:
-        cpu: ""
-        memory: ""
-
-    cephfs:
-      limits:
-        cpu: ""
-        memory: ""
-      requests:
-        cpu: ""
-        memory: ""
\ No newline at end of file
+  base: {}
\ No newline at end of file

Reply via email to