Hi,

Please run the following commands as root:

mkdir -p /root/.kube
cp -i /etc/kubernetes/admin.conf /root/.kube/config

After then the kubectl commands should work

-Wei

On Thu, 15 Feb 2024 at 13:53, Wally B <wvbauman...@gmail.com> wrote:

> What command do you suggest I run?
>
> kubeconfig returns command not found
>
> on your PR I see
>
> kubeadm join is being called out as well but I wanted to verify what you
> wanted me to test first.
>
> On Thu, Feb 15, 2024 at 2:41 AM Wei ZHOU <ustcweiz...@gmail.com> wrote:
>
> > Hi Wally,
> >
> > I think the cluster is working fine.
> > The kubeconfig is missing in extra nodes. I have just created a PR for
> it:
> > https://github.com/apache/cloudstack/pull/8658
> > You can run the command on the control nodes which should fix the
> problem.
> >
> >
> > -Wei
> >
> > On Thu, 15 Feb 2024 at 09:31, Wally B <wvbauman...@gmail.com> wrote:
> >
> > > 3 Nodes
> > >
> > > Control 1 -- No Errors
> > >
> > > kubectl get nodes
> > > NAME                                        STATUS   ROLES
>  AGE
> > >  VERSION
> > > pz-dev-k8s-ncus-00001-control-18dabdb141b   Ready    control-plane
>  2m6s
> > > v1.28.4
> > > pz-dev-k8s-ncus-00001-control-18dabdb6ad6   Ready    control-plane
>  107s
> > > v1.28.4
> > > pz-dev-k8s-ncus-00001-control-18dabdbc0a8   Ready    control-plane
>  108s
> > > v1.28.4
> > > pz-dev-k8s-ncus-00001-node-18dabdc1644      Ready    <none>
> 115s
> > > v1.28.4
> > > pz-dev-k8s-ncus-00001-node-18dabdc6c16      Ready    <none>
> 115s
> > > v1.28.4
> > >
> > >
> > > kubectl get pods --all-namespaces
> > > NAMESPACE              NAME
> > >                READY   STATUS    RESTARTS        AGE
> > > kube-system            coredns-5dd5756b68-g84vk
> > >                1/1     Running   0               2m46s
> > > kube-system            coredns-5dd5756b68-kf92x
> > >                1/1     Running   0               2m46s
> > > kube-system            etcd-pz-dev-k8s-ncus-00001-control-18dabdb141b
> > >                1/1     Running   0               2m50s
> > > kube-system            etcd-pz-dev-k8s-ncus-00001-control-18dabdb6ad6
> > >                1/1     Running   0               2m16s
> > > kube-system            etcd-pz-dev-k8s-ncus-00001-control-18dabdbc0a8
> > >                1/1     Running   0               2m37s
> > > kube-system
> > >  kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabdb141b
> 1/1
> > >   Running   0               2m52s
> > > kube-system
> > >  kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabdb6ad6
> 1/1
> > >   Running   1 (2m16s ago)   2m15s
> > > kube-system
> > >  kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabdbc0a8
> 1/1
> > >   Running   0               2m37s
> > > kube-system
> > >  kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabdb141b
>  1/1
> > >   Running   1 (2m25s ago)   2m51s
> > > kube-system
> > >  kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabdb6ad6
>  1/1
> > >   Running   0               2m18s
> > > kube-system
> > >  kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabdbc0a8
>  1/1
> > >   Running   0               2m37s
> > > kube-system            kube-proxy-445qx
> > >                1/1     Running   0               2m37s
> > > kube-system            kube-proxy-8swdg
> > >                1/1     Running   0               2m2s
> > > kube-system            kube-proxy-bl9rx
> > >                1/1     Running   0               2m47s
> > > kube-system            kube-proxy-pv8gj
> > >                1/1     Running   0               2m43s
> > > kube-system            kube-proxy-v7cw2
> > >                1/1     Running   0               2m43s
> > > kube-system
> > >  kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabdb141b
> 1/1
> > >   Running   1 (2m22s ago)   2m50s
> > > kube-system
> > >  kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabdb6ad6
> 1/1
> > >   Running   0               2m15s
> > > kube-system
> > >  kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabdbc0a8
> 1/1
> > >   Running   0               2m37s
> > > kube-system            weave-net-8dvl5
> > >                 2/2     Running   0               2m37s
> > > kube-system            weave-net-c54bz
> > >                 2/2     Running   0               2m43s
> > > kube-system            weave-net-lv8l4
> > >                 2/2     Running   1 (2m42s ago)   2m47s
> > > kube-system            weave-net-vg6td
> > >                 2/2     Running   0               2m2s
> > > kube-system            weave-net-vq9s4
> > >                 2/2     Running   0               2m43s
> > > kubernetes-dashboard   dashboard-metrics-scraper-5657497c4c-4k886
> > >                1/1     Running   0               2m46s
> > > kubernetes-dashboard   kubernetes-dashboard-5b749d9495-jpbxl
> > >                 1/1     Running   1 (2m22s ago)   2m46s
> > >
> > >
> > >
> > >
> > > Control 2: Errors at the CLI
> > > Failed to start Execute cloud user/final scripts.
> > >
> > > kubectl get nodes
> > > E0215 08:27:07.797825    2772 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:07.798759    2772 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:07.801039    2772 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:07.801977    2772 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:07.804029    2772 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > The connection to the server localhost:8080 was refused - did you
> specify
> > > the right host or port?
> > >
> > > kubectl get pods --all-namespaces
> > > E0215 08:29:41.818452    2811 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:41.819935    2811 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:41.820883    2811 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:41.822680    2811 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:41.823571    2811 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > The connection to the server localhost:8080 was refused - did you
> specify
> > > the right host or port?
> > >
> > > Ping Google: Success
> > > Ping Control Node 1: Success
> > >
> > >
> > > Control 3: Errors at the CLI
> > > Failed to start Execute cloud user/final scripts.
> > > Failed to start deploy-kube-system.service.
> > >
> > > kubectl get nodes
> > > E0215 08:27:15.057313    2697 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:15.058538    2697 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:15.059260    2697 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:15.061599    2697 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:27:15.062029    2697 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > The connection to the server localhost:8080 was refused - did you
> specify
> > > the right host or port?
> > >
> > >
> > > kubectl get pods --all-namespaces
> > > E0215 08:29:57.108716    2736 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:57.109533    2736 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:57.111372    2736 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:57.112074    2736 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > E0215 08:29:57.113956    2736 memcache.go:265] couldn't get current
> > server
> > > API group list: Get "http://localhost:8080/api?timeout=32s": dial tcp
> > > 127.0.0.1:8080: connect: connection refused
> > > The connection to the server localhost:8080 was refused - did you
> specify
> > > the right host or port?
> > >
> > >
> > > Ping Google: Success
> > > Ping Control Node 1: Success
> > >
> > >
> > > On Thu, Feb 15, 2024 at 2:17 AM Wei ZHOU <ustcweiz...@gmail.com>
> wrote:
> > >
> > > > Can you try with 3 control nodes ?
> > > >
> > > > -Wei
> > > >
> > > > On Thu, 15 Feb 2024 at 09:13, Wally B <wvbauman...@gmail.com> wrote:
> > > >
> > > > > - zone type :
> > > > >         Core
> > > > > - network type:
> > > > >         Advanced
> > > > >         Isolated Network inside a Redundant VPC (same results in
> just
> > > an
> > > > > Isolated network without VPC)
> > > > > - number of control nodes:
> > > > >         2 Control Nodes (HA Cluster)
> > > > >
> > > > > We were able to deploy k8s in the past, not sure what changed.
> > > > >
> > > > > Thanks!
> > > > > -Wally
> > > > >
> > > > > On Thu, Feb 15, 2024 at 2:04 AM Wei ZHOU <ustcweiz...@gmail.com>
> > > wrote:
> > > > >
> > > > > > Hi,
> > > > > >
> > > > > > can you share
> > > > > > - zone type
> > > > > > - network type
> > > > > > - number of control nodes
> > > > > >
> > > > > >
> > > > > > -Wei
> > > > > >
> > > > > > On Thu, 15 Feb 2024 at 08:52, Wally B <wvbauman...@gmail.com>
> > wrote:
> > > > > >
> > > > > > > So
> > > > > > >
> > > > > > > Recreating the Sec Storage VM Fixed the Cert issue and I was
> able
> > > to
> > > > > > > install K8s 1.28.4 Binaries. --- THANKS Wei ZHOU !
> > > > > > >
> > > > > > >
> > > > > > > Im still getting
> > > > > > >
> > > > > > > [FAILED] Failed to start Execute cloud user/final scripts.
> > > > > > >
> > > > > > > on 1 control and 1 worker.
> > > > > > >
> > > > > > > *Control 1 --  pz-dev-k8s-ncus-00001-control-18dabaf66c1  --
> > :*
> > > No
> > > > > > > errors at the CLI
> > > > > > >
> > > > > > > kubectl get nodes
> > > > > > > NAME                                        STATUS   ROLES
> > > > >  AGE
> > > > > > >   VERSION
> > > > > > > pz-dev-k8s-ncus-00001-control-18dabaf0edb   Ready
> > control-plane
> > > > >  5m2s
> > > > > > >  v1.28.4
> > > > > > > pz-dev-k8s-ncus-00001-control-18dabaf66c1   Ready
> > control-plane
> > > > > >  4m44s
> > > > > > >   v1.28.4
> > > > > > > pz-dev-k8s-ncus-00001-node-18dabafb0bd      Ready    <none>
> > > > > > 4m47s
> > > > > > >   v1.28.4
> > > > > > > pz-dev-k8s-ncus-00001-node-18dabb006bc      Ready    <none>
> > > > > > 4m47s
> > > > > > >   v1.28.4
> > > > > > >
> > > > > > >
> > > > > > > kubectl get pods --all-namespaces
> > > > > > > NAMESPACE              NAME
> > > > > > >                READY   STATUS    RESTARTS        AGE
> > > > > > > kube-system            coredns-5dd5756b68-295gb
> > > > > > >                1/1     Running   0               5m32s
> > > > > > > kube-system            coredns-5dd5756b68-cdwvw
> > > > > > >                1/1     Running   0               5m33s
> > > > > > > kube-system
> > > etcd-pz-dev-k8s-ncus-00001-control-18dabaf0edb
> > > > > > >                1/1     Running   0               5m36s
> > > > > > > kube-system
> > > etcd-pz-dev-k8s-ncus-00001-control-18dabaf66c1
> > > > > > >                1/1     Running   0               5m23s
> > > > > > > kube-system
> > > > > > >  kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabaf0edb
> > > > > 1/1
> > > > > > >   Running   0               5m36s
> > > > > > > kube-system
> > > > > > >  kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabaf66c1
> > > > > 1/1
> > > > > > >   Running   0               5m23s
> > > > > > > kube-system
> > > > > > >
> > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabaf0edb
> > > > >  1/1
> > > > > > >   Running   1 (5m13s ago)   5m36s
> > > > > > > kube-system
> > > > > > >
> > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabaf66c1
> > > > >  1/1
> > > > > > >   Running   0               5m23s
> > > > > > > kube-system            kube-proxy-2m8zb
> > > > > > >                1/1     Running   0               5m26s
> > > > > > > kube-system            kube-proxy-cwpjg
> > > > > > >                1/1     Running   0               5m33s
> > > > > > > kube-system            kube-proxy-l2vbf
> > > > > > >                1/1     Running   0               5m26s
> > > > > > > kube-system            kube-proxy-qhlqt
> > > > > > >                1/1     Running   0               5m23s
> > > > > > > kube-system
> > > > > > >  kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabaf0edb
> > > > > 1/1
> > > > > > >   Running   1 (5m8s ago)    5m36s
> > > > > > > kube-system
> > > > > > >  kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabaf66c1
> > > > > 1/1
> > > > > > >   Running   0               5m23s
> > > > > > > kube-system            weave-net-5cs26
> > > > > > >                 2/2     Running   1 (5m9s ago)    5m26s
> > > > > > > kube-system            weave-net-9zqrw
> > > > > > >                 2/2     Running   1 (5m28s ago)   5m33s
> > > > > > > kube-system            weave-net-fcwtr
> > > > > > >                 2/2     Running   0               5m23s
> > > > > > > kube-system            weave-net-lh2dh
> > > > > > >                 2/2     Running   1 (4m41s ago)   5m26s
> > > > > > > kubernetes-dashboard
>  dashboard-metrics-scraper-5657497c4c-r284t
> > > > > > >                1/1     Running   0               5m32s
> > > > > > > kubernetes-dashboard   kubernetes-dashboard-5b749d9495-vtwdd
> > > > > > >                 1/1     Running   0               5m32s
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > *Control 2 ---  pz-dev-k8s-ncus-00001-control-18dabaf66c1   :*
> > > > > [FAILED]
> > > > > > > Failed to start Execute cloud user/final scripts.
> > > > > > >
> > > > > > > kubectl get nodes
> > > > > > > E0215 07:38:33.314561    2643 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:38:33.316751    2643 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:38:33.317754    2643 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:38:33.319181    2643 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:38:33.319975    2643 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > The connection to the server localhost:8080 was refused - did
> you
> > > > > specify
> > > > > > > the right host or port?
> > > > > > >
> > > > > > >
> > > > > > > kubectl get pods --all-namespaces
> > > > > > > E0215 07:42:23.786704    2700 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:42:23.787455    2700 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:42:23.789529    2700 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:42:23.790051    2700 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > E0215 07:42:23.791742    2700 memcache.go:265] couldn't get
> > current
> > > > > > server
> > > > > > > API group list: Get "http://localhost:8080/api?timeout=32s":
> > dial
> > > > tcp
> > > > > > > 127.0.0.1:8080: connect: connection refused
> > > > > > > The connection to the server localhost:8080 was refused - did
> you
> > > > > specify
> > > > > > > the right host or port?
> > > > > > >
> > > > > > >
> > > > > > > */var/log/daemon.log*
> > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://docs.google.com/document/d/1KuIx0jI4TuAXPgACY3rJQz3L2B8AjeqOL0Fm5r4YF5M/edit?usp=sharing
> > > > > > >
> > > > > > > */var/log/messages*
> > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://docs.google.com/document/d/15xet6kxI9rdgi4RkIHqtn-Wywph4h1Coyt_cyrJYkv4/edit?usp=sharing
> > > > > > >
> > > > > > > On Thu, Feb 15, 2024 at 1:21 AM Wei ZHOU <
> ustcweiz...@gmail.com>
> > > > > wrote:
> > > > > > >
> > > > > > > > Destroy ssvm and retry when new ssvm is Up  ?
> > > > > > > >
> > > > > > > > -Wei
> > > > > > > >
> > > > > > > > 在 2024年2月15日星期四,Wally B <wvbauman...@gmail.com> 写道:
> > > > > > > >
> > > > > > > > > Super Weird. I have two other versions added successfully
> but
> > > now
> > > > > > when
> > > > > > > I
> > > > > > > > > try to add an ISO/version I get the following on the
> > management
> > > > > host.
> > > > > > > > This
> > > > > > > > > is the first time I've tried adding a K8s version since
> > 4.18.0
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > tail -f
> /var/log/cloudstack/management/management-server.log
> > |
> > > > grep
> > > > > > > ERROR
> > > > > > > > >
> > > > > > > > > 2024-02-15 06:26:18,900 DEBUG [c.c.a.t.Request]
> > > > > > > > > (AgentManager-Handler-5:null) (logid:) Seq
> > > > 48-6373437897659383816:
> > > > > > > > > Processing:  { Ans: , MgmtId: 15643723020152, via: 48, Ver:
> > v1,
> > > > > > Flags:
> > > > > > > > 10,
> > > > > > > > > [{"com.cloud.agent.api.storage.DownloadAnswer":{"
> > > > > > > > > jobId":"39d72d08-ab48-47dd-b09a-eee3ed816f4d","
> > > > > > > > > downloadPct":"0","errorString":"PKIX
> > > > > > > > > path building failed:
> > > > > > > > > sun.security.provider.certpath.SunCertPathBuilderException:
> > > > unable
> > > > > to
> > > > > > > > find
> > > > > > > > > valid certification path to requested
> > > > > > > > > target","downloadStatus":"DOWNLOAD_ERROR","downloadPath"
> > > > > > > > > :"/mnt/SecStorage/73075a0a-38a1-3631-8170-8887c04f6073/
> > > > > > > > > template/tmpl/1/223/dnld9180711723601784047tmp_","
> > > > > > > > > installPath":"template/tmpl/1/223","templateSize":"(0
> > > > > > > > > bytes) 0","templatePhySicalSize":"(0 bytes)
> > > > > > > > > 0","checkSum":"4dfb9d8be2191bc8bc4b89d78795a5
> > > > > > > > > b","result":"true","details":"PKIX
> > > > > > > > > path building failed:
> > > > > > > > > sun.security.provider.certpath.SunCertPathBuilderException:
> > > > unable
> > > > > to
> > > > > > > > find
> > > > > > > > > valid certification path to requested
> > > > > > > > > target","wait":"0","bypassHostMaintenance":"false"}}] }
> > > > > > > > >
> > > > > > > > > 2024-02-15 06:26:18,937 ERROR
> > > > [o.a.c.s.i.BaseImageStoreDriverImpl]
> > > > > > > > > (RemoteHostEndPoint-5:ctx-55063062) (logid:e21177cb) Failed
> > to
> > > > > > register
> > > > > > > > > template: b6e79c5a-38d4-4cf5-8606-e6f209b6b4c2 with error:
> > PKIX
> > > > > path
> > > > > > > > > building failed:
> > > > > > > > > sun.security.provider.certpath.SunCertPathBuilderException:
> > > > unable
> > > > > to
> > > > > > > > find
> > > > > > > > > valid certification path to requested target
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > On Wed, Feb 14, 2024 at 11:27 PM Wei ZHOU <
> > > ustcweiz...@gmail.com
> > > > >
> > > > > > > wrote:
> > > > > > > > >
> > > > > > > > > > Can you try 1.27.8 or 1.28.4 on
> > > > > > https://download.cloudstack.org/cks/
> > > > > > > ?
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > -Wei
> > > > > > > > > >
> > > > > > > > > > 在 2024年2月15日星期四,Wally B <wvbauman...@gmail.com> 写道:
> > > > > > > > > >
> > > > > > > > > > > Hello Everyone!
> > > > > > > > > > >
> > > > > > > > > > > We are currently attempting to deploy k8s clusters and
> > are
> > > > > > running
> > > > > > > > into
> > > > > > > > > > > issues with the deployment.
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > Current CS Environment:
> > > > > > > > > > >
> > > > > > > > > > > CloudStack Verison: 4.19.0 (Same issue before we
> upgraded
> > > > from
> > > > > > > > 4.18.1).
> > > > > > > > > > > Hypervisor Type: Ubuntu 20.04.03 KVM
> > > > > > > > > > > Attempted K8s Bins: 1.23.3, 1.27.3
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > ======== ISSUE =========
> > > > > > > > > > >
> > > > > > > > > > > For some reason when we attempt the cluster
> provisioning
> > > all
> > > > of
> > > > > > the
> > > > > > > > VMs
> > > > > > > > > > > start up, SSH Keys are installed, but then at least 1,
> > > > > sometimes
> > > > > > 2
> > > > > > > of
> > > > > > > > > the
> > > > > > > > > > > VMs (control and/or worker) we get:
> > > > > > > > > > >
> > > > > > > > > > > [FAILED] Failed to start deploy-kube-system.service.
> > > > > > > > > > > [FAILED] Failed to start Execute cloud user/final
> > scripts.
> > > > > > > > > > >
> > > > > > > > > > > The Cloudstack UI just says:
> > > > > > > > > > > Create Kubernetes cluster test-cluster in progress
> > > > > > > > > > > for about an hour (I assume this is the 3600 second
> > > timeout)
> > > > > and
> > > > > > > then
> > > > > > > > > > > fails.
> > > > > > > > > > >
> > > > > > > > > > > In the users event log it stays on:
> > > > > > > > > > > INFO KUBERNETES.CLUSTER.CREATE
> > > > > > > > > > > Scheduled
> > > > > > > > > > > Creating Kubernetes cluster. Cluster Id: XXX
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > I can ssh into the VMs with their assigned private
> keys.
> > I
> > > > > > > attempted
> > > > > > > > to
> > > > > > > > > > run
> > > > > > > > > > > the deploy-kube-system script but it just says already
> > > > > > provisioned!
> > > > > > > > I'm
> > > > > > > > > > not
> > > > > > > > > > > sure how I would Execute cloud user/final scripts. If I
> > > > attempt
> > > > > > to
> > > > > > > > stop
> > > > > > > > > > the
> > > > > > > > > > > cluster and start it again nothing seems to change.
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > Any help would be appreciated, I can provide any
> details
> > as
> > > > > they
> > > > > > > are
> > > > > > > > > > > needed!
> > > > > > > > > > >
> > > > > > > > > > > Thanks!
> > > > > > > > > > > Wally
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to