-
Notifications
You must be signed in to change notification settings - Fork 6.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failed to scale worker nodes due to the different cgroup driver between docker and kubelet #5262
Comments
Oct 16 15:43:53 cn-hz-wl-test-k8s-02 kubelet[78362]: F1016 15:43:53.891981 78362 server.go:273] failed to run Kubelet: failed to create kubelet: misconfiguration: kubelet cgroup driver: "systemd" is different from docker cgroup driver: "cgroupfs" sample question. -rw-r--r-- 1 root root 603 Oct 16 15:43 /etc/kubernetes/kubelet-config.yaml |
I have also been affected by this issue. I tried to scale the cluster, the playbook finished apparently fine, but the node was not added and all the other nodes passed to NotReady status because of the change of the kubelet cgroup driver change. Using the extra-var It must be also taken into account that the scale.yml playbook restarts the docker daemon of the complete cluster, making all the pods unavailable during a few seconds. |
I am also facing this issue, not sure if i should change the cgroup driver of kubelet or docker. I guess the workaround works, but not sure if that is what we want here. |
According to this issue, maybe the new worker node could be added with |
Same issue here with Kubespray v2.11.0 + CentOS 7
|
from roles/container-engine/containerd/defaults/main.yml file remove kubelet_cgroup_driver:
reference: |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
Stale issues rot after 30d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
Rotten issues close after 30d of inactivity. Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
@fejta-bot: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Environment:
Cloud provider or hardware configuration: Azure VM
OS (
printf "$(uname -srm)\n$(cat /etc/os-release)\n"
):ansible --version
):Kubespray version (commit) (
git rev-parse --short HEAD
):Network plugin used:
Calico
Copy of your inventory file:
inital cluster
scale target
Other place changed
group_vars/all/k8s-cluster/k8s-cluster.yml
Command used to invoke ansible:
Deploy inital cluster
Scale, add worker, fail
Scale, add worker, successful
Output of ansible run:
Deploy log:
https://gist.github.com/ydye/1f1a9cf63583a273942b5df8eae95963
log of scale failure due to cgroup drivers
https://gist.github.com/ydye/6ed258d72cfedf5389d0590ef15f5ac3
scale successfully
https://gist.github.com/ydye/4752fdd33116522e74cec1b3b496f394
Anything else do we need to know:
I think the 3 nodes which I want to add into the cluster is fine. Because when I try to deploy a cluster with the 6 nodes (include the 3 nodes) with the cluster.yml, the deployment will be successful.
The text was updated successfully, but these errors were encountered: