Hi, As I understand, 1. After upgrading, you need to patch the system vms or recreate them. Not a bug I think. 2. a minor issue which does not impact the provisioning and operation of CKS cluster. 3. Looks like a network misconfiguration, but did it work before ?
-Wei On Thu, 15 Feb 2024 at 16:39, Wally B <[email protected]> wrote: > As a quick add-on. After running those commands and getting the kubectl > commands working the Error in the management log is > > tail -f /var/log/cloudstack/management/management-server.log | grep ERROR > > 2024-02-15 14:09:41,124 ERROR [c.c.k.c.a.KubernetesClusterActionWorker] > (API-Job-Executor-4:ctx-29ed2b8e job-12348 ctx-3355553d) (logid:ae448a2e) > Failed to setup Kubernetes cluster : pz-dev-k8s-ncus-00001 in usable state > as unable to access control node VMs of the cluster > > 2024-02-15 14:09:41,129 ERROR [c.c.a.ApiAsyncJobDispatcher] > (API-Job-Executor-4:ctx-29ed2b8e job-12348) (logid:ae448a2e) Unexpected > exception while executing > > org.apache.cloudstack.api.command.user.kubernetes.cluster.CreateKubernetesClusterCmd > > 2024-02-15 14:33:01,117 ERROR [c.c.k.c.a.KubernetesClusterActionWorker] > (API-Job-Executor-17:ctx-0685d548 job-12552 ctx-997de847) (logid:fda8fc82) > Failed to setup Kubernetes cluster : pz-dev-k8s-ncus-00001 in usable state > as unable to access control node VMs of the cluster > > > did a quick test-netconnection from my pc to the control node and got > > > > Test-NetConnection 99.xx.xx.xxx -p 6443 > > > > ComputerName : 99.xx.xx.xxx > RemoteAddress : 99.xx.xx.xxx > RemotePort : 6443 > InterfaceAlias : Ethernet > SourceAddress : xxx.xxx.xxx.xxx > TcpTestSucceeded : True > > > So I did a test to see If I could get it from my Management hosts (on the > same public ip range as the Virtual Router Public IP). and I got a TTL > Expired. > > > > > To wrap it up there were 3 issues. > > > 1. Needed to delete and re-provision the Secondary Storage System Virtual > Machine after upgrading from 4.18.1 to 4.19.0 > 2. Needed to fix additional control nodes not getting the kubeadm.conf > copied correctly (Wei PR) > 3. Needed to fix some routing on our end since were were bouncing between > our L3 TOR ->Firewall <- ISP Routers > > Thanks again for all the help, everyone! > Wally > > On Thu, Feb 15, 2024 at 7:24 AM Wally B <[email protected]> wrote: > > > Thanks Wei ZHOU! > > > > That fixed the kubectl command issue but the cluster still just sits at > > > > Create Kubernetes cluster k8s-cluster-1 in progress > > > > Maybe this is just a UI issue? Unfortunately If I stop the k8s cluster > > after it errors out it just stays in the error state. > > > > 1. Click Stop Kubernetes cluster > > 2. UI Says it successfully stopped. > > 3. Try to Start the Cluster but the power button just says Stop > > Kubernetes cluster and the UI Status stays in the error state. > > > > > > On Thu, Feb 15, 2024 at 7:02 AM Wei ZHOU <[email protected]> wrote: > > > >> Hi, > >> > >> Please run the following commands as root: > >> > >> mkdir -p /root/.kube > >> cp -i /etc/kubernetes/admin.conf /root/.kube/config > >> > >> After then the kubectl commands should work > >> > >> -Wei > >> > >> On Thu, 15 Feb 2024 at 13:53, Wally B <[email protected]> wrote: > >> > >> > What command do you suggest I run? > >> > > >> > kubeconfig returns command not found > >> > > >> > on your PR I see > >> > > >> > kubeadm join is being called out as well but I wanted to verify what > you > >> > wanted me to test first. > >> > > >> > On Thu, Feb 15, 2024 at 2:41 AM Wei ZHOU <[email protected]> > wrote: > >> > > >> > > Hi Wally, > >> > > > >> > > I think the cluster is working fine. > >> > > The kubeconfig is missing in extra nodes. I have just created a PR > for > >> > it: > >> > > https://github.com/apache/cloudstack/pull/8658 > >> > > You can run the command on the control nodes which should fix the > >> > problem. > >> > > > >> > > > >> > > -Wei > >> > > > >> > > On Thu, 15 Feb 2024 at 09:31, Wally B <[email protected]> > wrote: > >> > > > >> > > > 3 Nodes > >> > > > > >> > > > Control 1 -- No Errors > >> > > > > >> > > > kubectl get nodes > >> > > > NAME STATUS ROLES > >> > AGE > >> > > > VERSION > >> > > > pz-dev-k8s-ncus-00001-control-18dabdb141b Ready control-plane > >> > 2m6s > >> > > > v1.28.4 > >> > > > pz-dev-k8s-ncus-00001-control-18dabdb6ad6 Ready control-plane > >> > 107s > >> > > > v1.28.4 > >> > > > pz-dev-k8s-ncus-00001-control-18dabdbc0a8 Ready control-plane > >> > 108s > >> > > > v1.28.4 > >> > > > pz-dev-k8s-ncus-00001-node-18dabdc1644 Ready <none> > >> > 115s > >> > > > v1.28.4 > >> > > > pz-dev-k8s-ncus-00001-node-18dabdc6c16 Ready <none> > >> > 115s > >> > > > v1.28.4 > >> > > > > >> > > > > >> > > > kubectl get pods --all-namespaces > >> > > > NAMESPACE NAME > >> > > > READY STATUS RESTARTS AGE > >> > > > kube-system coredns-5dd5756b68-g84vk > >> > > > 1/1 Running 0 2m46s > >> > > > kube-system coredns-5dd5756b68-kf92x > >> > > > 1/1 Running 0 2m46s > >> > > > kube-system > >> etcd-pz-dev-k8s-ncus-00001-control-18dabdb141b > >> > > > 1/1 Running 0 2m50s > >> > > > kube-system > >> etcd-pz-dev-k8s-ncus-00001-control-18dabdb6ad6 > >> > > > 1/1 Running 0 2m16s > >> > > > kube-system > >> etcd-pz-dev-k8s-ncus-00001-control-18dabdbc0a8 > >> > > > 1/1 Running 0 2m37s > >> > > > kube-system > >> > > > kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabdb141b > >> > 1/1 > >> > > > Running 0 2m52s > >> > > > kube-system > >> > > > kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabdb6ad6 > >> > 1/1 > >> > > > Running 1 (2m16s ago) 2m15s > >> > > > kube-system > >> > > > kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabdbc0a8 > >> > 1/1 > >> > > > Running 0 2m37s > >> > > > kube-system > >> > > > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabdb141b > >> > 1/1 > >> > > > Running 1 (2m25s ago) 2m51s > >> > > > kube-system > >> > > > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabdb6ad6 > >> > 1/1 > >> > > > Running 0 2m18s > >> > > > kube-system > >> > > > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabdbc0a8 > >> > 1/1 > >> > > > Running 0 2m37s > >> > > > kube-system kube-proxy-445qx > >> > > > 1/1 Running 0 2m37s > >> > > > kube-system kube-proxy-8swdg > >> > > > 1/1 Running 0 2m2s > >> > > > kube-system kube-proxy-bl9rx > >> > > > 1/1 Running 0 2m47s > >> > > > kube-system kube-proxy-pv8gj > >> > > > 1/1 Running 0 2m43s > >> > > > kube-system kube-proxy-v7cw2 > >> > > > 1/1 Running 0 2m43s > >> > > > kube-system > >> > > > kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabdb141b > >> > 1/1 > >> > > > Running 1 (2m22s ago) 2m50s > >> > > > kube-system > >> > > > kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabdb6ad6 > >> > 1/1 > >> > > > Running 0 2m15s > >> > > > kube-system > >> > > > kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabdbc0a8 > >> > 1/1 > >> > > > Running 0 2m37s > >> > > > kube-system weave-net-8dvl5 > >> > > > 2/2 Running 0 2m37s > >> > > > kube-system weave-net-c54bz > >> > > > 2/2 Running 0 2m43s > >> > > > kube-system weave-net-lv8l4 > >> > > > 2/2 Running 1 (2m42s ago) 2m47s > >> > > > kube-system weave-net-vg6td > >> > > > 2/2 Running 0 2m2s > >> > > > kube-system weave-net-vq9s4 > >> > > > 2/2 Running 0 2m43s > >> > > > kubernetes-dashboard dashboard-metrics-scraper-5657497c4c-4k886 > >> > > > 1/1 Running 0 2m46s > >> > > > kubernetes-dashboard kubernetes-dashboard-5b749d9495-jpbxl > >> > > > 1/1 Running 1 (2m22s ago) 2m46s > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > Control 2: Errors at the CLI > >> > > > Failed to start Execute cloud user/final scripts. > >> > > > > >> > > > kubectl get nodes > >> > > > E0215 08:27:07.797825 2772 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:07.798759 2772 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:07.801039 2772 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:07.801977 2772 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:07.804029 2772 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > The connection to the server localhost:8080 was refused - did you > >> > specify > >> > > > the right host or port? > >> > > > > >> > > > kubectl get pods --all-namespaces > >> > > > E0215 08:29:41.818452 2811 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:41.819935 2811 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:41.820883 2811 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:41.822680 2811 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:41.823571 2811 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > The connection to the server localhost:8080 was refused - did you > >> > specify > >> > > > the right host or port? > >> > > > > >> > > > Ping Google: Success > >> > > > Ping Control Node 1: Success > >> > > > > >> > > > > >> > > > Control 3: Errors at the CLI > >> > > > Failed to start Execute cloud user/final scripts. > >> > > > Failed to start deploy-kube-system.service. > >> > > > > >> > > > kubectl get nodes > >> > > > E0215 08:27:15.057313 2697 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:15.058538 2697 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:15.059260 2697 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:15.061599 2697 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:27:15.062029 2697 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > The connection to the server localhost:8080 was refused - did you > >> > specify > >> > > > the right host or port? > >> > > > > >> > > > > >> > > > kubectl get pods --all-namespaces > >> > > > E0215 08:29:57.108716 2736 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:57.109533 2736 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:57.111372 2736 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:57.112074 2736 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > E0215 08:29:57.113956 2736 memcache.go:265] couldn't get > current > >> > > server > >> > > > API group list: Get "http://localhost:8080/api?timeout=32s": dial > >> tcp > >> > > > 127.0.0.1:8080: connect: connection refused > >> > > > The connection to the server localhost:8080 was refused - did you > >> > specify > >> > > > the right host or port? > >> > > > > >> > > > > >> > > > Ping Google: Success > >> > > > Ping Control Node 1: Success > >> > > > > >> > > > > >> > > > On Thu, Feb 15, 2024 at 2:17 AM Wei ZHOU <[email protected]> > >> > wrote: > >> > > > > >> > > > > Can you try with 3 control nodes ? > >> > > > > > >> > > > > -Wei > >> > > > > > >> > > > > On Thu, 15 Feb 2024 at 09:13, Wally B <[email protected]> > >> wrote: > >> > > > > > >> > > > > > - zone type : > >> > > > > > Core > >> > > > > > - network type: > >> > > > > > Advanced > >> > > > > > Isolated Network inside a Redundant VPC (same results > in > >> > just > >> > > > an > >> > > > > > Isolated network without VPC) > >> > > > > > - number of control nodes: > >> > > > > > 2 Control Nodes (HA Cluster) > >> > > > > > > >> > > > > > We were able to deploy k8s in the past, not sure what changed. > >> > > > > > > >> > > > > > Thanks! > >> > > > > > -Wally > >> > > > > > > >> > > > > > On Thu, Feb 15, 2024 at 2:04 AM Wei ZHOU < > [email protected] > >> > > >> > > > wrote: > >> > > > > > > >> > > > > > > Hi, > >> > > > > > > > >> > > > > > > can you share > >> > > > > > > - zone type > >> > > > > > > - network type > >> > > > > > > - number of control nodes > >> > > > > > > > >> > > > > > > > >> > > > > > > -Wei > >> > > > > > > > >> > > > > > > On Thu, 15 Feb 2024 at 08:52, Wally B < > [email protected]> > >> > > wrote: > >> > > > > > > > >> > > > > > > > So > >> > > > > > > > > >> > > > > > > > Recreating the Sec Storage VM Fixed the Cert issue and I > was > >> > able > >> > > > to > >> > > > > > > > install K8s 1.28.4 Binaries. --- THANKS Wei ZHOU ! > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > Im still getting > >> > > > > > > > > >> > > > > > > > [FAILED] Failed to start Execute cloud user/final scripts. > >> > > > > > > > > >> > > > > > > > on 1 control and 1 worker. > >> > > > > > > > > >> > > > > > > > *Control 1 -- pz-dev-k8s-ncus-00001-control-18dabaf66c1 > -- > >> > > :* > >> > > > No > >> > > > > > > > errors at the CLI > >> > > > > > > > > >> > > > > > > > kubectl get nodes > >> > > > > > > > NAME STATUS ROLES > >> > > > > > AGE > >> > > > > > > > VERSION > >> > > > > > > > pz-dev-k8s-ncus-00001-control-18dabaf0edb Ready > >> > > control-plane > >> > > > > > 5m2s > >> > > > > > > > v1.28.4 > >> > > > > > > > pz-dev-k8s-ncus-00001-control-18dabaf66c1 Ready > >> > > control-plane > >> > > > > > > 4m44s > >> > > > > > > > v1.28.4 > >> > > > > > > > pz-dev-k8s-ncus-00001-node-18dabafb0bd Ready > <none> > >> > > > > > > 4m47s > >> > > > > > > > v1.28.4 > >> > > > > > > > pz-dev-k8s-ncus-00001-node-18dabb006bc Ready > <none> > >> > > > > > > 4m47s > >> > > > > > > > v1.28.4 > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > kubectl get pods --all-namespaces > >> > > > > > > > NAMESPACE NAME > >> > > > > > > > READY STATUS RESTARTS AGE > >> > > > > > > > kube-system coredns-5dd5756b68-295gb > >> > > > > > > > 1/1 Running 0 5m32s > >> > > > > > > > kube-system coredns-5dd5756b68-cdwvw > >> > > > > > > > 1/1 Running 0 5m33s > >> > > > > > > > kube-system > >> > > > etcd-pz-dev-k8s-ncus-00001-control-18dabaf0edb > >> > > > > > > > 1/1 Running 0 5m36s > >> > > > > > > > kube-system > >> > > > etcd-pz-dev-k8s-ncus-00001-control-18dabaf66c1 > >> > > > > > > > 1/1 Running 0 5m23s > >> > > > > > > > kube-system > >> > > > > > > > kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabaf0edb > >> > > > > > 1/1 > >> > > > > > > > Running 0 5m36s > >> > > > > > > > kube-system > >> > > > > > > > kube-apiserver-pz-dev-k8s-ncus-00001-control-18dabaf66c1 > >> > > > > > 1/1 > >> > > > > > > > Running 0 5m23s > >> > > > > > > > kube-system > >> > > > > > > > > >> > > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabaf0edb > >> > > > > > 1/1 > >> > > > > > > > Running 1 (5m13s ago) 5m36s > >> > > > > > > > kube-system > >> > > > > > > > > >> > > kube-controller-manager-pz-dev-k8s-ncus-00001-control-18dabaf66c1 > >> > > > > > 1/1 > >> > > > > > > > Running 0 5m23s > >> > > > > > > > kube-system kube-proxy-2m8zb > >> > > > > > > > 1/1 Running 0 5m26s > >> > > > > > > > kube-system kube-proxy-cwpjg > >> > > > > > > > 1/1 Running 0 5m33s > >> > > > > > > > kube-system kube-proxy-l2vbf > >> > > > > > > > 1/1 Running 0 5m26s > >> > > > > > > > kube-system kube-proxy-qhlqt > >> > > > > > > > 1/1 Running 0 5m23s > >> > > > > > > > kube-system > >> > > > > > > > kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabaf0edb > >> > > > > > 1/1 > >> > > > > > > > Running 1 (5m8s ago) 5m36s > >> > > > > > > > kube-system > >> > > > > > > > kube-scheduler-pz-dev-k8s-ncus-00001-control-18dabaf66c1 > >> > > > > > 1/1 > >> > > > > > > > Running 0 5m23s > >> > > > > > > > kube-system weave-net-5cs26 > >> > > > > > > > 2/2 Running 1 (5m9s ago) 5m26s > >> > > > > > > > kube-system weave-net-9zqrw > >> > > > > > > > 2/2 Running 1 (5m28s ago) 5m33s > >> > > > > > > > kube-system weave-net-fcwtr > >> > > > > > > > 2/2 Running 0 5m23s > >> > > > > > > > kube-system weave-net-lh2dh > >> > > > > > > > 2/2 Running 1 (4m41s ago) 5m26s > >> > > > > > > > kubernetes-dashboard > >> > dashboard-metrics-scraper-5657497c4c-r284t > >> > > > > > > > 1/1 Running 0 5m32s > >> > > > > > > > kubernetes-dashboard > kubernetes-dashboard-5b749d9495-vtwdd > >> > > > > > > > 1/1 Running 0 5m32s > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > *Control 2 --- pz-dev-k8s-ncus-00001-control-18dabaf66c1 > >> :* > >> > > > > > [FAILED] > >> > > > > > > > Failed to start Execute cloud user/final scripts. > >> > > > > > > > > >> > > > > > > > kubectl get nodes > >> > > > > > > > E0215 07:38:33.314561 2643 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:38:33.316751 2643 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:38:33.317754 2643 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:38:33.319181 2643 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:38:33.319975 2643 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > The connection to the server localhost:8080 was refused - > >> did > >> > you > >> > > > > > specify > >> > > > > > > > the right host or port? > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > kubectl get pods --all-namespaces > >> > > > > > > > E0215 07:42:23.786704 2700 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:42:23.787455 2700 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:42:23.789529 2700 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:42:23.790051 2700 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > E0215 07:42:23.791742 2700 memcache.go:265] couldn't > get > >> > > current > >> > > > > > > server > >> > > > > > > > API group list: Get " > http://localhost:8080/api?timeout=32s > >> ": > >> > > dial > >> > > > > tcp > >> > > > > > > > 127.0.0.1:8080: connect: connection refused > >> > > > > > > > The connection to the server localhost:8080 was refused - > >> did > >> > you > >> > > > > > specify > >> > > > > > > > the right host or port? > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > */var/log/daemon.log* > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > >> > > >> > https://docs.google.com/document/d/1KuIx0jI4TuAXPgACY3rJQz3L2B8AjeqOL0Fm5r4YF5M/edit?usp=sharing > >> > > > > > > > > >> > > > > > > > */var/log/messages* > >> > > > > > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > >> > > >> > https://docs.google.com/document/d/15xet6kxI9rdgi4RkIHqtn-Wywph4h1Coyt_cyrJYkv4/edit?usp=sharing > >> > > > > > > > > >> > > > > > > > On Thu, Feb 15, 2024 at 1:21 AM Wei ZHOU < > >> > [email protected]> > >> > > > > > wrote: > >> > > > > > > > > >> > > > > > > > > Destroy ssvm and retry when new ssvm is Up ? > >> > > > > > > > > > >> > > > > > > > > -Wei > >> > > > > > > > > > >> > > > > > > > > 在 2024年2月15日星期四,Wally B <[email protected]> 写道: > >> > > > > > > > > > >> > > > > > > > > > Super Weird. I have two other versions added > >> successfully > >> > but > >> > > > now > >> > > > > > > when > >> > > > > > > > I > >> > > > > > > > > > try to add an ISO/version I get the following on the > >> > > management > >> > > > > > host. > >> > > > > > > > > This > >> > > > > > > > > > is the first time I've tried adding a K8s version > since > >> > > 4.18.0 > >> > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > tail -f > >> > /var/log/cloudstack/management/management-server.log > >> > > | > >> > > > > grep > >> > > > > > > > ERROR > >> > > > > > > > > > > >> > > > > > > > > > 2024-02-15 06:26:18,900 DEBUG [c.c.a.t.Request] > >> > > > > > > > > > (AgentManager-Handler-5:null) (logid:) Seq > >> > > > > 48-6373437897659383816: > >> > > > > > > > > > Processing: { Ans: , MgmtId: 15643723020152, via: 48, > >> Ver: > >> > > v1, > >> > > > > > > Flags: > >> > > > > > > > > 10, > >> > > > > > > > > > [{"com.cloud.agent.api.storage.DownloadAnswer":{" > >> > > > > > > > > > jobId":"39d72d08-ab48-47dd-b09a-eee3ed816f4d"," > >> > > > > > > > > > downloadPct":"0","errorString":"PKIX > >> > > > > > > > > > path building failed: > >> > > > > > > > > > > >> sun.security.provider.certpath.SunCertPathBuilderException: > >> > > > > unable > >> > > > > > to > >> > > > > > > > > find > >> > > > > > > > > > valid certification path to requested > >> > > > > > > > > > > target","downloadStatus":"DOWNLOAD_ERROR","downloadPath" > >> > > > > > > > > > > :"/mnt/SecStorage/73075a0a-38a1-3631-8170-8887c04f6073/ > >> > > > > > > > > > template/tmpl/1/223/dnld9180711723601784047tmp_"," > >> > > > > > > > > > installPath":"template/tmpl/1/223","templateSize":"(0 > >> > > > > > > > > > bytes) 0","templatePhySicalSize":"(0 bytes) > >> > > > > > > > > > 0","checkSum":"4dfb9d8be2191bc8bc4b89d78795a5 > >> > > > > > > > > > b","result":"true","details":"PKIX > >> > > > > > > > > > path building failed: > >> > > > > > > > > > > >> sun.security.provider.certpath.SunCertPathBuilderException: > >> > > > > unable > >> > > > > > to > >> > > > > > > > > find > >> > > > > > > > > > valid certification path to requested > >> > > > > > > > > > target","wait":"0","bypassHostMaintenance":"false"}}] > } > >> > > > > > > > > > > >> > > > > > > > > > 2024-02-15 06:26:18,937 ERROR > >> > > > > [o.a.c.s.i.BaseImageStoreDriverImpl] > >> > > > > > > > > > (RemoteHostEndPoint-5:ctx-55063062) (logid:e21177cb) > >> Failed > >> > > to > >> > > > > > > register > >> > > > > > > > > > template: b6e79c5a-38d4-4cf5-8606-e6f209b6b4c2 with > >> error: > >> > > PKIX > >> > > > > > path > >> > > > > > > > > > building failed: > >> > > > > > > > > > > >> sun.security.provider.certpath.SunCertPathBuilderException: > >> > > > > unable > >> > > > > > to > >> > > > > > > > > find > >> > > > > > > > > > valid certification path to requested target > >> > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > On Wed, Feb 14, 2024 at 11:27 PM Wei ZHOU < > >> > > > [email protected] > >> > > > > > > >> > > > > > > > wrote: > >> > > > > > > > > > > >> > > > > > > > > > > Can you try 1.27.8 or 1.28.4 on > >> > > > > > > https://download.cloudstack.org/cks/ > >> > > > > > > > ? > >> > > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > -Wei > >> > > > > > > > > > > > >> > > > > > > > > > > 在 2024年2月15日星期四,Wally B <[email protected]> 写道: > >> > > > > > > > > > > > >> > > > > > > > > > > > Hello Everyone! > >> > > > > > > > > > > > > >> > > > > > > > > > > > We are currently attempting to deploy k8s clusters > >> and > >> > > are > >> > > > > > > running > >> > > > > > > > > into > >> > > > > > > > > > > > issues with the deployment. > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > Current CS Environment: > >> > > > > > > > > > > > > >> > > > > > > > > > > > CloudStack Verison: 4.19.0 (Same issue before we > >> > upgraded > >> > > > > from > >> > > > > > > > > 4.18.1). > >> > > > > > > > > > > > Hypervisor Type: Ubuntu 20.04.03 KVM > >> > > > > > > > > > > > Attempted K8s Bins: 1.23.3, 1.27.3 > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > ======== ISSUE ========= > >> > > > > > > > > > > > > >> > > > > > > > > > > > For some reason when we attempt the cluster > >> > provisioning > >> > > > all > >> > > > > of > >> > > > > > > the > >> > > > > > > > > VMs > >> > > > > > > > > > > > start up, SSH Keys are installed, but then at > least > >> 1, > >> > > > > > sometimes > >> > > > > > > 2 > >> > > > > > > > of > >> > > > > > > > > > the > >> > > > > > > > > > > > VMs (control and/or worker) we get: > >> > > > > > > > > > > > > >> > > > > > > > > > > > [FAILED] Failed to start > deploy-kube-system.service. > >> > > > > > > > > > > > [FAILED] Failed to start Execute cloud user/final > >> > > scripts. > >> > > > > > > > > > > > > >> > > > > > > > > > > > The Cloudstack UI just says: > >> > > > > > > > > > > > Create Kubernetes cluster test-cluster in progress > >> > > > > > > > > > > > for about an hour (I assume this is the 3600 > second > >> > > > timeout) > >> > > > > > and > >> > > > > > > > then > >> > > > > > > > > > > > fails. > >> > > > > > > > > > > > > >> > > > > > > > > > > > In the users event log it stays on: > >> > > > > > > > > > > > INFO KUBERNETES.CLUSTER.CREATE > >> > > > > > > > > > > > Scheduled > >> > > > > > > > > > > > Creating Kubernetes cluster. Cluster Id: XXX > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > I can ssh into the VMs with their assigned private > >> > keys. > >> > > I > >> > > > > > > > attempted > >> > > > > > > > > to > >> > > > > > > > > > > run > >> > > > > > > > > > > > the deploy-kube-system script but it just says > >> already > >> > > > > > > provisioned! > >> > > > > > > > > I'm > >> > > > > > > > > > > not > >> > > > > > > > > > > > sure how I would Execute cloud user/final scripts. > >> If I > >> > > > > attempt > >> > > > > > > to > >> > > > > > > > > stop > >> > > > > > > > > > > the > >> > > > > > > > > > > > cluster and start it again nothing seems to > change. > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > > >> > > > > > > > > > > > Any help would be appreciated, I can provide any > >> > details > >> > > as > >> > > > > > they > >> > > > > > > > are > >> > > > > > > > > > > > needed! > >> > > > > > > > > > > > > >> > > > > > > > > > > > Thanks! > >> > > > > > > > > > > > Wally > >> > > > > > > > > > > > > >> > > > > > > > > > > > >> > > > > > > > > > > >> > > > > > > > > > >> > > > > > > > > >> > > > > > > > >> > > > > > > >> > > > > > >> > > > > >> > > > >> > > >> > > >
