I've Airflow installed in namespace - airflow, and i'm using NFS to store
the dags & pyspark code. I want to run the airflow jobs in different
namespace - airflow-spark-apps .. however, i'm unable to do this since the
PVC associated with the ns - airflow-spark-apps is not accessible to
airflow deployed in ns - airflow.
Here is the helm chart deployment :
```
helm upgrade --install airflow apache-airflow/airflow \
--namespace airflow \
--set dags.persistence.enabled=true \
--set dags.persistence.existingClaim=airflow-dags-pvc \
--set dags.persistence.subPath="airflow-dags" \
--set global.persistence.existingClaimNamespace=airflow-spark-apps \
--set dags.gitSync.enabled=false \
--set
images.airflow.repository=artifacts.versa-networks.com:8443/airflow-image
\
--set images.airflow.tag=0.0.1 \
--set scheduler.resources.requests.memory="1024Mi" \
--set scheduler.resources.requests.cpu="500m" \
--set scheduler.resources.limits.memory="2048Mi" \
--set scheduler.resources.limits.cpu="1000m" \
--set webserver.resources.requests.memory="512Mi" \
--set webserver.resources.requests.cpu="250m" \
--set webserver.resources.limits.memory="1024Mi" \
--set webserver.resources.limits.cpu="500m" \
--set workers.resources.requests.memory="1024Mi" \
--set workers.resources.requests.cpu="500m" \
--set workers.resources.limits.memory="4096Mi" \
--set workers.resources.limits.cpu="2000m" \
--version 1.9.0 \
--set serviceAccount.create=false \
--set serviceAccount.name=airflow \
--set config.kubernetes_executor.multi_namespace_mode=True \
--set
config.kubernetes_executor.multi_namespace_mode_namespace_list="airflow-spark-apps"
\
--debug
```
PVC & PV created in namespace - airflow-spark-apps:
```
apiVersion: v1
kind: PersistentVolume
metadata:
name: airflow-dags-pv
namespace: storage
labels:
app: airflow-dags
spec:
capacity:
storage: 5Gi
accessModes:
- ReadWriteMany
nfs:
server: nfs-service.storage.svc.cluster.local
path: "/exports/airflow-dags"
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: airflow-dags-pvc
namespace: airflow-spark-apps
spec:
accessModes:
- ReadWriteMany
storageClassName: ""
resources:
requests:
storage: 5Gi
selector:
matchLabels:
app: airflow-dags
```
Airflow worker is not starting up, error :
```
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedScheduling 67s (x33 over 161m) default-scheduler
0/4 nodes are available: persistentvolumeclaim "airflow-dags-pvc" not
found. preemption: 0/4 nodes are available: 4 Preemption is not
helpful for scheduling..
```
How do I fix this ? Can I have the airflow jobs running in namespace -
airflow-spark-apps, while the airflow install is in ns - airflow ?
tia!