GitHub user vishnuvs369 created a discussion: Kubernetes Cluster Scale Fails – 
Worker Nodes Not Joining After Scale Operation (CloudStack 4.19.0.1)

### problem

**We attempted to scale the cluster from 8 to 9 worker nodes using the Scale 
Cluster option in the CloudStack UI.**

The new worker VM was created successfully and is in Running state in 
CloudStack. However, Kubernetes does not register the new node.

After the scale operation:

CloudStack shows 9 worker nodes

kubectl get nodes continues to show only 8 worker nodes

UI status stucks on scaling.


**Error observed in CloudStack Management Server logs:**

ERROR ... Unexpected exception while executing ScaleKubernetesClusterCmd
at KubernetesClusterResourceModifierActionWorker.removeSshFirewallRule
at KubernetesClusterScaleWorker.scaleKubernetesClusterIsolatedNetworkRules

### versions

Environment details:

CloudStack version: 4.19.0.1

Cluster type: CloudStack Kubernetes Service (CKS)

Initial cluster size:

1 Control Plane

8 Worker Nodes (working fine)

Scale target: 9 Worker Nodes

Global setting cloud.kubernetes.cluster.max.size was increased from 10 to 50 
prior to scaling.

### The steps to reproduce the bug

**Steps to Reproduce**

Deploy a Kubernetes cluster using CloudStack Kubernetes Service (CKS) on 
CloudStack 4.19.0.1 with the following configuration:

1 Control Plane node

8 Worker nodes

>From the CloudStack UI, navigate to:

Kubernetes → Clusters → <cluster-name> → Scale Cluster


Scale the cluster by increasing the worker node count from 8 to 9 and submit 
the scale operation.

Observe the following behavior:

The new worker VM is created successfully and shows Running state in CloudStack.

The scale task stucks on scaling.

kubectl get nodes still showing node count as 8


### What to do about it?

is there any workaround to resolve this without restarting the cluster

GitHub link: https://github.com/apache/cloudstack/discussions/12577

----
This is an automatically sent email for [email protected].
To unsubscribe, please send an email to: [email protected]

Reply via email to