0

I have a one-node cluster deployed with kubespray. I tried 2.14 - the same thing. The cluster successfully starts after deployment, and works for a long time. When I restart the machine all the docker container (apiserver, scheduler, cni, calico, etc.), except etcd, are not working. Systemd kubelet service is up, but its log shows that it is not healthy.

Sep 25 16:07:09 node1 systemd[1]: Started Kubernetes Kubelet Server.
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.378828    4273 flags.go:33] FLAG: --add-dir-header="false"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.378989    4273 flags.go:33] FLAG: --address="0.0.0.0"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379003    4273 flags.go:33] FLAG: --allowed-unsafe-sysctls="[]"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379009    4273 flags.go:33] FLAG: --alsologtostderr="false"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379014    4273 flags.go:33] FLAG: --anonymous-auth="true"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379019    4273 flags.go:33] FLAG: --application-metrics-count-limit="100"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379023    4273 flags.go:33] FLAG: --authentication-token-webhook="false"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379027    4273 flags.go:33] FLAG: --authentication-token-webhook-cache-ttl="2m0s"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379032    4273 flags.go:33] FLAG: --authorization-mode="AlwaysAllow"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379698    4273 flags.go:33] FLAG: --volume-stats-agg-period="1m0s"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.379734    4273 feature_gate.go:243] feature gates: &{map[]}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.381339    4273 feature_gate.go:243] feature gates: &{map[]}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.381391    4273 feature_gate.go:243] feature gates: &{map[]}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391492    4273 mount_linux.go:168] Detected OS with systemd
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391633    4273 server.go:416] Version: v1.17.12
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391728    4273 feature_gate.go:243] feature gates: &{map[]}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391804    4273 feature_gate.go:243] feature gates: &{map[]}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391934    4273 plugins.go:100] No cloud provider specified.
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391956    4273 server.go:532] No cloud provider specified: "" from the config file: ""
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.391968    4273 server.go:822] Client rotation is on, will bootstrap in background
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.393945    4273 bootstrap.go:84] Current kubeconfig file contents are still valid, no bootstrap necessary
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.394040    4273 certificate_store.go:129] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem".
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.394294    4273 server.go:849] Starting client certificate rotation.
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.394311    4273 certificate_manager.go:275] Certificate rotation is enabled.
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.394462    4273 certificate_manager.go:531] Certificate expiration is 2021-09-25 15:58:29 +0000 UTC, rotation deadline is 2021-08-09 11:53:23.825031508 +0000 UTC
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.394497    4273 certificate_manager.go:281] Waiting 7627h46m14.430536992s for next certificate rotation
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.394860    4273 dynamic_cafile_content.go:166] Starting client-ca-bundle::/etc/kubernetes/ssl/ca.crt
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.395055    4273 manager.go:146] cAdvisor running in container: "/sys/fs/cgroup/cpu,cpuacct/system.slice/kubelet.service"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.414349    4273 fs.go:125] Filesystem UUIDs: map[2020-09-25-15-46-36-00:/dev/sr0 55122f25-8946-4857-952e-48dc1647b4e1:/dev/vda1 88DB-6C04:/dev/vda15]
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.414375    4273 fs.go:126] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:25 fsType:tmpfs blockSize:0} /dev/vda1:{mountpoint:/ major:252 minor:1 fsType:ext4 blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/lock:{mountpoint:/run/lock major:0 minor:26 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:61 fsType:tmpfs blockSize:0} /sys/fs/cgroup:{mountpoint:/sys/fs/cgroup major:0 minor:27 fsType:tmpfs blockSize:0}]
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.418995    4273 manager.go:193] Machine: {NumCores:8 CpuFrequency:2494136 MemoryCapacity:50641403904 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:6f481a3404734917921596637175f7df SystemUUID:6F481A34-0473-4917-9215-96637175F7DF BootID:4499a41b-6deb-4cf6-906f-a88d6794e657 Filesystems:[{Device:/sys/fs/cgroup DeviceMajor:0 DeviceMinor:27 Capacity:25320701952 Type:vfs Inodes:6181812 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:61 Capacity:5064138752 Type:vfs Inodes:6181812 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5064142848 Type:vfs Inodes:6181812 HasInodes:true} {Device:/dev/vda1 DeviceMajor:252 DeviceMinor:1 Capacity:103880232960 Type:vfs Inodes:12902400 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:25 Capacity:25320701952 Type:vfs Inodes:6181812 HasInodes:true} {Device:/run/lock DeviceMajor:0 DeviceMinor:26 Capacity:5242880 Type:vfs Inodes:6181812 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:107374182400 Scheduler:none}] NetworkDevices:[{Name:ens3 MacAddress:fa:16:3e:26:40:ce Speed:-1 Mtu:1500}] Topology:[{Id:0 Memory:50641403904 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:1 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[1] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:2 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[2] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:3 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[3] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:4 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[4] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:5 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[5] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:6 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[6] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]} {Id:7 Memory:0 HugePages:[] Cores:[{Id:0 Threads:[7] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:4194304 Type:Unified Level:2}]}] Caches:[{Size:16777216 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420041    4273 manager.go:199] Version: {KernelVersion:4.15.0-118-generic ContainerOsVersion:Ubuntu 18.04.5 LTS DockerVersion:18.09.7 DockerAPIVersion:1.39 CadvisorVersion: CadvisorRevision:}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420193    4273 server.go:642] --cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420530    4273 container_manager_linux.go:265] container manager verified user specified cgroup-root exists: []
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420548    4273 container_manager_linux.go:270] Creating Container Manager object based on Node Config: {RuntimeCgroupsName:/systemd/system.slice SystemCgroupsName: KubeletCgroupsName:/systemd/system.slice ContainerRuntime:docker CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:cgroupfs KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[cpu:{i:{value:200 scale:-3} d:{Dec:<nil>} s:200m Format:DecimalSI} memory:{i:{value:512 scale:6} d:{Dec:<nil>} s:512M Format:DecimalSI}] SystemReserved:map[] HardEvictionThresholds:[{Signal:imagefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.15} GracePeriod:0s MinReclaim:<nil>} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.available Operator:LessThan Value:{Quantity:<nil> Percentage:0.1} GracePeriod:0s MinReclaim:<nil>} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity:<nil> Percentage:0.05} GracePeriod:0s MinReclaim:<nil>}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerReconcilePeriod:10s ExperimentalPodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420695    4273 fake_topology_manager.go:29] [fake topologymanager] NewFakeManager
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420702    4273 container_manager_linux.go:305] Creating device plugin manager: true
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420712    4273 manager.go:126] Creating Device Plugin manager at /var/lib/kubelet/device-plugins/kubelet.sock
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420731    4273 fake_topology_manager.go:39] [fake topologymanager] AddHintProvider HintProvider:  &{kubelet.sock /var/lib/kubelet/device-plugins/ map[] {0 0} <nil> {{} [0 0 0]} 0x1b1eed0 0x6eb9ef8 0x1b1f7a0 map[] map[] map[] map[] map[] 0xc000a6d7d0 [0] 0x6eb9ef8}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420763    4273 state_mem.go:36] [cpumanager] initializing new in-memory state store
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420837    4273 state_mem.go:84] [cpumanager] updated default cpuset: ""
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420843    4273 state_mem.go:92] [cpumanager] updated cpuset assignments: "map[]"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420851    4273 state_checkpoint.go:101] [cpumanager] state checkpoint: restored state from checkpoint
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420855    4273 state_checkpoint.go:102] [cpumanager] state checkpoint: defaultCPUSet:
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420862    4273 fake_topology_manager.go:39] [fake topologymanager] AddHintProvider HintProvider:  &{{0 0} 0x6eb9ef8 10000000000 0xc00049ea20 <nil> <nil> <nil> <nil> map[cpu:{{200 -3} {<nil>}  DecimalSI} memory:{{616857600 0} {<nil>}  DecimalSI}] 0x6eb9ef8}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420925    4273 server.go:1056] Using root directory: /var/lib/kubelet
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420947    4273 kubelet.go:286] Adding pod path: /etc/kubernetes/manifests
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420974    4273 file.go:68] Watching path "/etc/kubernetes/manifests"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.420987    4273 kubelet.go:311] Watching apiserver
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.421729    4273 reflector.go:153] k8s.io/kubernetes/pkg/kubelet/kubelet.go:449: Failed to list *v1.Service: Get https://10.34.12.170:6443/api/v1/services?limit=500&resourceVersion=0: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1&limit=500&resourceVersion=0: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1&limit=500&resourceVersion=0: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.424494    4273 client.go:75] Connecting to docker on unix:///var/run/docker.sock
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.424517    4273 client.go:104] Start docker client with request timeout=2m0s
Sep 25 16:07:09 node1 kubelet[4273]: W0925 16:07:09.425742    4273 docker_service.go:564] Hairpin mode set to "promiscuous-bridge" but kubenet is not enabled, falling back to "hairpin-veth"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.425764    4273 docker_service.go:240] Hairpin mode set to "hairpin-veth"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.459931    4273 plugins.go:166] Loaded network plugin "cni"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.459991    4273 docker_service.go:255] Docker cri networking managed by cni
Sep 25 16:07:09 node1 Labels:[] ExperimentalBuild:false ServerVersion:18.09.7 ClusterStore: ClusterAdvertise: Runtimes:map[runc:{Path:runc Args:[]}] DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:[] Nodes:0 Managers:0 Cluster:<nil> Warnings:[]} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:8fba4e9a7d01810a393d5d25a3621dc101981175 Expected:8fba4e9a7d01810a393d5d25a3621dc101981175} RuncCommit:{ID:dc9208a3303feef5b3839f4323d9beb36df0a9dd Expected:dc9208a3303feef5b3839f4323d9beb36df0a9dd} InitCommit:{ID:fec3683 Expected:fec3683} SecurityOptions:[name=apparmor name=seccomp,profile=default] ProductLicense:Community Engine Warnings:[WARNING: No swap limit support]}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.468858    4273 docker_service.go:273] Setting cgroupDriver to cgroupfs
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.468983    4273 kubelet.go:642] Starting the GRPC server for the docker CRI shim.
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.469073    4273 container_manager_linux.go:118] Configure resource-only container "/systemd/system.slice" with memory limit: 35448982732
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.469091    4273 docker_server.go:59] Start dockershim grpc server
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478669    4273 remote_runtime.go:59] parsed scheme: ""
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478689    4273 remote_runtime.go:59] scheme "" not registered, fallback to default scheme
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478714    4273 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0  <nil>}] <nil>}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478722    4273 clientconn.go:577] ClientConn switching balancer to "pick_first"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478749    4273 remote_image.go:50] parsed scheme: ""
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478755    4273 remote_image.go:50] scheme "" not registered, fallback to default scheme
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478764    4273 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0  <nil>}] <nil>}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478769    4273 clientconn.go:577] ClientConn switching balancer to "pick_first"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478889    4273 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc00026e020, CONNECTING
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.478928    4273 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc000a42020, CONNECTING
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.479399    4273 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc00026e020, READY
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.479426    4273 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc000a42020, READY
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.870533    4273 aws_credentials.go:77] while getting AWS credentials NoCredentialProviders: no valid providers in chain. Deprecated.
Sep 25 16:07:09 node1 kubelet[4273]:         For verbose messaging see aws.Config.CredentialsChainVerboseErrors
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872026    4273 kuberuntime_manager.go:217] Container runtime docker initialized, version: 18.09.7, apiVersion: 1.39.0
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872387    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/aws-ebs"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872485    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/gce-pd"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872550    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/cinder"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872602    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/azure-disk"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872651    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/azure-file"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872701    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/vsphere-volume"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872763    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/empty-dir"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872813    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/git-repo"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872872    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/host-path"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872926    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/nfs"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.872976    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/secret"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873041    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/iscsi"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873091    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/glusterfs"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873151    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/rbd"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873200    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/quobyte"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873251    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/cephfs"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873298    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/downward-api"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873347    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/fc"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873397    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/flocker"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873445    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/configmap"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873496    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/projected"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873556    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/portworx-volume"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873623    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/scaleio"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873681    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/local-volume"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873729    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/storageos"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.873816    4273 plugins.go:631] Loaded volume plugin "kubernetes.io/csi"
Sep 25 16:07:09 node1: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.875105    4273 server.go:1114] Started kubelet
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.875145    4273 kubelet.go:1305] Image garbage collection failed once. Stats initialization may not have completed yet: failed to get imageFs info: unable to find data in memory cache
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.875186    4273 server.go:144] Starting to listen on 10.34.12.170:10250
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.875457    4273 event.go:272] Unable to write event: 'Post https://10.34.12.170:6443/api/v1/namespaces/default/events: dial tcp 10.34.12.170:6443: connect: connection refused' (may retry after sleeping)
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.876063    4273 fs_resource_analyzer.go:64] Starting FS ResourceAnalyzer
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.876153    4273 volume_manager.go:263] The desired_state_of_world populator starts
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.876163    4273 volume_manager.go:265] Starting Kubelet Volume Manager
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.876550    4273 desired_state_of_world_populator.go:138] Desired state populator starts to run
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.876729    4273 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.CSIDriver: Get https://10.34.12.170:6443/apis/storage.k8s.io/v1beta1/csidrivers?limit=500&resourceVersion=0: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1?timeout=10s: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.877999    4273 server.go:384] Adding debug handlers to kubelet server.
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891555    4273 factory.go:356] Registering Docker factory
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891578    4273 factory.go:54] Registering systemd factory
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891705    4273 clientconn.go:104] parsed scheme: "unix"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891714    4273 clientconn.go:104] scheme "unix" not registered, fallback to default scheme
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891735    4273 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{unix:///run/containerd/containerd.sock 0  <nil>}] <nil>}
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891743    4273 clientconn.go:577] ClientConn switching balancer to "pick_first"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.891797    4273 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc00037cd00, CONNECTING
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.892584    4273 balancer_conn_wrappers.go:127] pickfirstBalancer: HandleSubConnStateChange: 0xc00037cd00, READY
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.893173    4273 factory.go:137] Registering containerd factory
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.893533    4273 factory.go:101] Registering Raw factory
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.893662    4273 manager.go:1158] Started watching for new ooms in manager
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.894666    4273 manager.go:272] Starting recovery of all containers
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.897183    4273 kubelet_network_linux.go:150] Not using `--random-fully` in the MASQUERADE rule for iptables because the local version of iptables does not support it
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.898224    4273 status_manager.go:157] Starting to sync pod status with apiserver
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.898331    4273 kubelet.go:1823] Starting kubelet main sync loop.
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.898413    4273 kubelet.go:1847] skipping pod synchronization - [container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]
Sep 25 16:07:09 node1 kubelet[4273]: E0925 16:07:09.899415    4273 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.RuntimeClass: Get https://10.34.12.170:6443/apis/node.k8s.io/v1beta1/runtimeclasses?limit=500&resourceVersion=0: dial tcp 10.34.12.170:6443: connect: connection refused
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.904467    4273 manager.go:277] Recovery completed
Sep 25 16:07:09 node1 kubelet[4273]: W0925 16:07:09.923594    4273 docker_sandbox.go:394] failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod "kubernetes-metrics-scraper-747b4fd5cd-f2gtr_kube-system": CNI failed to retrieve network namespace path: cannot find network namespace for the terminated container "15b9965fb59f8b9f590ff1f01c7bf8c71746da8dcb0163a7277b7d1664f09b52"
Sep 25 16:07:09 node1 kubelet[4273]: W0925 16:07:09.928100    4273 docker_sandbox.go:394] failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod "kubernetes-dashboard-77475cf576-xc6br_kube-system": CNI failed to retrieve network namespace path: cannot find network namespace for the terminated container "3652b305aec8a60b50602b5ca12e9335a8e35c270cbcfcb7896139de97c73ace"
Sep 25 16:07:09 node1 kubelet[4273]: W0925 16:07:09.944457    4273 docker_sandbox.go:394] failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod "dns-autoscaler-85f898cd5c-jv6cb_kube-system": CNI failed to retrieve network namespace path: cannot find network namespace for the terminated container "55ee03cc56e650563ab445bccab7f327c7b8c66e086efbeb7d93bfb816aba838"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.947817    4273 kubelet_node_status.go:294] Setting node annotation to enable volume controller attach/detach
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.947980    4273 setters.go:73] Using node IP: "10.34.12.170"
Sep 25 16:07:09 node1
Sep 25 16:07:09 node1
Sep 25 16:07:09 node1
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.949354    4273 cpu_manager.go:173] [cpumanager] starting with none policy
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.949359    4273 cpu_manager.go:174] [cpumanager] reconciling every 10s
Sep 25 16:07:09 node1 kubelet[4273]: W0925 16:07:09.949360    4273 docker_sandbox.go:394] failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod "coredns-76798d84dd-6bmth_kube-system": CNI failed to retrieve network namespace path: cannot find network namespace for the terminated container "bde82b4692238cdcc8bb0860d9773ddbf57fc09d59388f655757224ef463ff1a"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.949365    4273 policy_none.go:43] [cpumanager] none policy: Start
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.950266    4273 manager.go:226] Starting Device Plugin manager
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.950956    4273 manager.go:268] Serving device plugin registration server on "/var/lib/kubelet/device-plugins/kubelet.sock"
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.951022    4273 plugin_watcher.go:54] Plugin Watcher Start at /var/lib/kubelet/plugins_registry
Sep 25 16:07:09 node1 kubelet[4273]: I0925 16:07:09.951107    4273 plugin_manager.go:112] The desired_state_of_world populator (plugin watcher) starts

Basically, it tires to talk to apiserver unsuccessfully. What I have observed and become very confused that neither apiserver container nor cni/calico container (docker ps -a) never get get started. Is not it kubelet suppose to start those, if so, why it attemps to talk to api server without trying to start it? If no kubelete then what is starting the apiserver, the restart policy on the api server container is no.

Yuki
  • 3,857
  • 5
  • 25
  • 43

0 Answers0