Advertisement
pup_seba

"kubeadm init" output

Feb 11th, 2020
264
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 5.31 KB | None | 0 0
  1. [root@kube-master-01 ~]# kubeadm init --upload-certs
  2. W0211 12:18:13.719466 2892 validation.go:28] Cannot validate kube-proxy config - no validator is available
  3. W0211 12:18:13.719633 2892 validation.go:28] Cannot validate kubelet config - no validator is available
  4. [init] Using Kubernetes version: v1.17.2
  5. [preflight] Running pre-flight checks
  6. [preflight] Pulling images required for setting up a Kubernetes cluster
  7. [preflight] This might take a minute or two, depending on the speed of your internet connection
  8. [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
  9. [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
  10. [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
  11. [kubelet-start] Starting the kubelet
  12. [certs] Using certificateDir folder "/etc/kubernetes/pki"
  13. [certs] Generating "ca" certificate and key
  14. [certs] Generating "apiserver" certificate and key
  15. [certs] apiserver serving cert is signed for DNS names [kube-master-01 kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 172.16.1.176]
  16. [certs] Generating "apiserver-kubelet-client" certificate and key
  17. [certs] Generating "front-proxy-ca" certificate and key
  18. [certs] Generating "front-proxy-client" certificate and key
  19. [certs] Generating "etcd/ca" certificate and key
  20. [certs] Generating "etcd/server" certificate and key
  21. [certs] etcd/server serving cert is signed for DNS names [kube-master-01 localhost] and IPs [172.16.1.176 127.0.0.1 ::1]
  22. [certs] Generating "etcd/peer" certificate and key
  23. [certs] etcd/peer serving cert is signed for DNS names [kube-master-01 localhost] and IPs [172.16.1.176 127.0.0.1 ::1]
  24. [certs] Generating "etcd/healthcheck-client" certificate and key
  25. [certs] Generating "apiserver-etcd-client" certificate and key
  26. [certs] Generating "sa" key and public key
  27. [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
  28. [kubeconfig] Writing "admin.conf" kubeconfig file
  29. [kubeconfig] Writing "kubelet.conf" kubeconfig file
  30. [kubeconfig] Writing "controller-manager.conf" kubeconfig file
  31. [kubeconfig] Writing "scheduler.conf" kubeconfig file
  32. [control-plane] Using manifest folder "/etc/kubernetes/manifests"
  33. [control-plane] Creating static Pod manifest for "kube-apiserver"
  34. [control-plane] Creating static Pod manifest for "kube-controller-manager"
  35. W0211 12:19:40.571202 2892 manifests.go:214] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
  36. [control-plane] Creating static Pod manifest for "kube-scheduler"
  37. W0211 12:19:40.573160 2892 manifests.go:214] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"
  38. [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
  39. [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
  40. [kubelet-check] Initial timeout of 40s passed.
  41. [kubelet-check] It seems like the kubelet isn't running or healthy.
  42. [kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.
  43. [kubelet-check] It seems like the kubelet isn't running or healthy.
  44. [kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.
  45. [kubelet-check] It seems like the kubelet isn't running or healthy.
  46. [kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.
  47. [kubelet-check] It seems like the kubelet isn't running or healthy.
  48. [kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.
  49. [kubelet-check] It seems like the kubelet isn't running or healthy.
  50. [kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get http://localhost:10248/healthz: dial tcp 127.0.0.1:10248: connect: connection refused.
  51.  
  52. Unfortunately, an error has occurred:
  53. timed out waiting for the condition
  54.  
  55. This error is likely caused by:
  56. - The kubelet is not running
  57. - The kubelet is unhealthy due to a misconfiguration of the node in some way (required cgroups disabled)
  58.  
  59. If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
  60. - 'systemctl status kubelet'
  61. - 'journalctl -xeu kubelet'
  62.  
  63. Additionally, a control plane component may have crashed or exited when started by the container runtime.
  64. To troubleshoot, list all containers using your preferred container runtimes CLI, e.g. docker.
  65. Here is one example how you may list all Kubernetes containers running in docker:
  66. - 'docker ps -a | grep kube | grep -v pause'
  67. Once you have found the failing container, you can inspect its logs with:
  68. - 'docker logs CONTAINERID'
  69. error execution phase wait-control-plane: couldn't initialize a Kubernetes cluster
  70. To see the stack trace of this error execute with --v=5 or higher
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement