jedcunningham commented on code in PR #31066:
URL: https://github.com/apache/airflow/pull/31066#discussion_r1278136190


##########
chart/INSTALL:
##########
@@ -12,3 +12,28 @@ kubectl create namespace airflow
 
 ## Install the chart in that namespace
 helm install airflow -n airflow .
+
+# Naming conventions
+
+# For new installations it is highly recommended to start using standard 
naming conventions.
+# It is not enabled by default as this may cause unexpected behaviours on 
existing installations
+helm install airflow -n airflow . --set useStandardNaming=true
+
+# For existing installations, you can use standard naming conventions by 
adding useStandardNaming=true.
+# All your resources will be recreated with a new name and helm will delete 
previous resources.
+
+helm upgrade airflow -n airflow . --set useStandardNaming=true
+# or
+helm upgrade airflow -n airflow . --set 
fullnameOverride=airflow-override,useStandardNaming=true
+
+# Note: This won't delete existing PVCs for logs used by 
statefulset/deployments, but it will recreate them with brand new PVCs.
+# If you do want to preserve logs history you'll need to manually copy the 
data of these volumes into the new volumes after
+# deployment. Depending on what storage backend/class you're using this 
procedure may vary. If you don't mind starting
+# with fresh logs/redis volumes, you can just delete the old pvcs that will be 
names, for example:
+
+kubectl delete pvc -n airflow logs-gta-triggerer-0
+kubectl delete pvc -n airflow logs-gta-worker-0
+kubectl delete pvc -n airflow redis-db-gta-redis-0
+
+# If you do not change useStandardNaming or fullnameOverride after upgrade, 
you can proceed as usual and no unexpected
+# behaviours will be presented

Review Comment:
   This part should probably also be covered in the newsfragment.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to