K8S部署---故障处理
一、kubeadm部署K8S集群故障排查
问题1:kubeadm初始化master01失败?
[root@k8s-master01 ~]# kubeadm init --config=new-kubeadm-config.yaml  --upload-certs |tee kubeadm-init.log
W0819 14:05:46.581489   81698 strict.go:47] unknown configuration schema.GroupVersionKind{Group:"kubeadm.k8s.io", Version:"v1beta3", Kind:"InitConfiguration"} for scheme definitions in "k8s.io/kubernetes/cmd/kubeadm/app/apis/kubeadm/scheme/scheme.go:31" and "k8s.io/kubernetes/cmd/kubeadm/app/componentconfigs/scheme.go:28"
no kind "InitConfiguration" is registered for version "kubeadm.k8s.io/v1beta3" in scheme "k8s.io/kubernetes/cmd/kubeadm/app/apis/kubeadm/scheme/scheme.go:31"
To see the stack trace of this error execute with --v=5 or higher
原因分析:yaml文件中apiVersion版本错误。

解决方法:修改yaml文件中apiVersion版本即可。

问题2:kubeadm初始化master01成功后,没有出现加入其他master节点的提示命令?

原因分析:初始化master的yaml文件存在问题。
解决方法:修改适合k8s版本的yaml添加“controlPlaneEndpoint: 192.168.1.236:16443”即可。


问题3:kubeadm方式部署K8S集群时,node加入K8S集群时卡顿并失败?
[root@k8s-node01 ~]# kubeadm join 192.168.1.201:6443 --token 1qo7ms.7atall1jcecf10qz --discovery-token-ca-cert-hash sha256:d1d102ceb6241a3617777f6156cd4e86dc9f9edd9e1d6d73266d6ca7f6280890
[preflight] Running pre-flight checks
原因分析:初始化主机点后发现部分组件异常;集群时间不同步;
[root@k8s-master01 ~]# kubectl get pod -n kube-system
NAME READY STATUS RESTARTS AGE
coredns-54d67798b7-28w5q 0/1 Pending 0 3m39s
coredns-54d67798b7-sxqpm 0/1 Pending 0 3m39s
etcd-k8s-master01 1/1 Running 0 3m53s
kube-apiserver-k8s-master01 1/1 Running 0 3m53s
kube-controller-manager-k8s-master01 1/1 Running 0 3m53s
kube-proxy-rvj6w 0/1 CrashLoopBackOff 5 3m40s
kube-scheduler-k8s-master01 1/1 Running 0 3m53s
解决方法:同步时间后,重新进行初始化主节点,再行加入其他节点。
###所有节点执行如下内容
ntpdate ntp1.aliyun.com;hwclock --systohc
date;hwclock
kubeadm reset -f;ipvsadm --clear;rm -rf ./.kube
###首个节点初始化
kubeadm init --config=new-kubeadm-config.yaml --upload-certs |tee kubeadm-init.log
###其他节点加入集群

问题3:kubeadm方式部署K8S集群时,node加入K8S集群失败?
[root@k8s-node01 ~]# kubeadm join 192.168.1.201:6443 --token 2g9k0a.tsm6xe31rdb7jbo8 --discovery-token-ca-cert-hash sha256:d1d102ceb6241a3617777f6156cd4e86dc9f9edd9e1d6d73266d6ca7f6280890
[preflight] Running pre-flight checks
[preflight] Reading configuration from the cluster...
[preflight] FYI: You can look at this config file with 'kubectl -n kube-system get cm kubeadm-config -o yaml'
error execution phase preflight: unable to fetch the kubeadm-config ConfigMap: failed to get config map: Unauthorized
To see the stack trace of this error execute with --v=5 or higher
原因分析:token过期了;
解决方法:重新生成不过期的token即可;
[root@k8s-master01 ~]# kubeadm token create --ttl 0 --print-join-command
W0819 12:00:27.541838    7855 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io]
kubeadm join 192.168.1.201:6443 --token 6xyv8a.cueltqmpe9qa8nxu     --discovery-token-ca-cert-hash sha256:bd78dfd370e47dfca742b5f6934c21014792168fa4dc19c9fa63bfdd87270097

问题3:kubeadm方式部署K8S集群时,node状态为NotReady,且coredns的pod状态为pending,仅初始化的master节点kubeproxy为running?
[root@k8s-master01 ~]# kubectl get nodes NAME STATUS ROLES AGE VERSION k8s-master01 NotReady control-plane,master 35m v1.22.0 k8s-master02 NotReady control-plane,master 32m v1.22.0 k8s-master03 NotReady control-plane,master 22m v1.22.0 k8s-node01 NotReady node 32m v1.22.0 k8s-node02 NotReady node 32m v1.22.0 [root@k8s-master01 ~]# kubectl get pod -n kube-system NAME READY STATUS RESTARTS AGE coredns-7d89d9b6b8-ddjjp 0/1 Pending 0 35m coredns-7d89d9b6b8-sbd8k 0/1 Pending 0 35m etcd-k8s-master01 1/1 Running 3 (18m ago) 35m etcd-k8s-master02 1/1 Running 0 32m etcd-k8s-master03 0/1 Running 1 (18m ago) 22m kube-apiserver-k8s-master01 0/1 Running 2 34m kube-apiserver-k8s-master02 0/1 Running 0 32m kube-apiserver-k8s-master03 0/1 Running 0 22m kube-controller-manager-k8s-master01 1/1 Running 4 (19m ago) 34m kube-controller-manager-k8s-master02 1/1 Running 0 32m kube-controller-manager-k8s-master03 1/1 Running 0 22m kube-proxy-6wxkt 0/1 Error 8 (21m ago) 32m kube-proxy-9286f 1/1 Running 8 (21m ago) 32m kube-proxy-h9jcv 0/1 CrashLoopBackOff 5 (19m ago) 22m kube-proxy-v944j 0/1 Error 8 (21m ago) 32m kube-proxy-zfrrk 0/1 Error 8 (23m ago) 35m kube-scheduler-k8s-master01 1/1 Running 3 (32m ago) 35m kube-scheduler-k8s-master02 1/1 Running 1 (19m ago) 32m kube-scheduler-k8s-master03 1/1 Running 0 22m [root@k8s-master01 ~]# kubectl get svc NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE kubernetes ClusterIP 10.96.0.1 <none> 443/TCP 35m
原因分析:网络插件没有安装好。
解决方法:安装网络插件,如flannel、calico等。
问题4、kubeadm方式部署K8S集群时,部署calico组件失败?
[root@k8s-master01 calico]# kubectl get pod -A -o wide | grep calico kube-system calico-kube-controllers-cdd5755b9-phg94 0/1 Running 0 6m5s 192.168.1.203 k8s-master03 <none> <none> kube-system calico-node-cb6c6 0/1 CrashLoopBackOff 5 (99s ago) 6m5s 192.168.1.202 k8s-master02 <none> <none> kube-system calico-node-knr8g 0/1 CrashLoopBackOff 5 (78s ago) 6m5s 192.168.1.203 k8s-master03 <none> <none> kube-system calico-node-l4fp9 0/1 CrashLoopBackOff 5 (98s ago) 6m5s 192.168.1.204 k8s-node01 <none> <none> kube-system calico-node-mwqlt 0/1 CrashLoopBackOff 5 (96s ago) 6m5s 192.168.1.205 k8s-node02 <none> <none> kube-system calico-node-zjkz6 0/1 CrashLoopBackOff 5 (93s ago) 6m5s 192.168.1.201 k8s-master01 <none> <none>
原因分析:calico-node的yaml文件存在问题。
[root@k8s-master01 ~]# kubectl describe pod calico-node-cb6c6  -n kube-system
Name:                 calico-node-cb6c6
Namespace:            kube-system
Priority:             2000001000
Priority Class Name:  system-node-critical
Node:                 k8s-master02/192.168.1.202
Start Time:           Tue, 14 Sep 2021 17:14:28 +0800
Labels:               controller-revision-hash=b5f966fbb
                      k8s-app=calico-node
                      pod-template-generation=1
Annotations:          <none>
Status:               Running
IP:                   192.168.1.202
IPs:
  IP:           192.168.1.202
Controlled By:  DaemonSet/calico-node
Init Containers:
  install-cni:
    Container ID:  docker://3e7d03495a0daa469b0f6b15acb451fa228c8d97a7d41fda0642d733cb21de9c
    Image:         registry.cn-beijing.aliyuncs.com/dotbalo/cni:v3.15.3
    Image ID:      docker-pullable://registry.cn-beijing.aliyuncs.com/dotbalo/cni@sha256:a559d264c7a75a7528560d11778dba2d3b55c588228aed99be401fd2baa9b607
    Port:          <none>
    Host Port:     <none>
    Command:
      /install-cni.sh
    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 14 Sep 2021 17:14:51 +0800
      Finished:     Tue, 14 Sep 2021 17:14:51 +0800
    Ready:          True
    Restart Count:  0
    Environment:
      CNI_CONF_NAME:       10-calico.conflist
      CNI_NETWORK_CONFIG:  <set to the key 'cni_network_config' of config map 'calico-config'>  Optional: false
      ETCD_ENDPOINTS:      <set to the key 'etcd_endpoints' of config map 'calico-config'>      Optional: false
      CNI_MTU:             <set to the key 'veth_mtu' of config map 'calico-config'>            Optional: false
      SLEEP:               false
    Mounts:
      /calico-secrets from etcd-certs (rw)
      /host/etc/cni/net.d from cni-net-dir (rw)
      /host/opt/cni/bin from cni-bin-dir (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-tfw5v (ro)
  flexvol-driver:
    Container ID:   docker://c8fe47bd4805c83fa3fad054470c3f3eb9dbccc1e64455cae085f050b11897c1
    Image:          registry.cn-beijing.aliyuncs.com/dotbalo/pod2daemon-flexvol:v3.15.3
    Image ID:       docker-pullable://registry.cn-beijing.aliyuncs.com/dotbalo/pod2daemon-flexvol@sha256:6bd1246d0ea1e573a6a050902995b1666ec0852339e5bda3051f583540361b55
    Port:           <none>
    Host Port:      <none>
    State:          Terminated
      Reason:       Completed
      Exit Code:    0
      Started:      Tue, 14 Sep 2021 17:14:56 +0800
      Finished:     Tue, 14 Sep 2021 17:14:56 +0800
    Ready:          True
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /host/driver from flexvol-driver-host (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-tfw5v (ro)
Containers:
  calico-node:
    Container ID:   docker://7e209f5b426027b10aa3460b4bab2b003ad4a853f46e24353f119211abb42cee
    Image:          registry.cn-beijing.aliyuncs.com/dotbalo/node:v3.15.3
    Image ID:       docker-pullable://registry.cn-beijing.aliyuncs.com/dotbalo/node@sha256:988095dbe39d2066b1964aafaa4a302a1286b149a4a80c9a1eb85544f2a0cdd0
    Port:           <none>
    Host Port:      <none>
    State:          Waiting
      Reason:       CrashLoopBackOff
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 14 Sep 2021 17:21:43 +0800
      Finished:     Tue, 14 Sep 2021 17:21:45 +0800
    Ready:          False
    Restart Count:  6
    Requests:
      cpu:      250m
    Liveness:   exec [/bin/calico-node -felix-live -bird-live] delay=10s timeout=1s period=10s #success=1 #failure=6
    Readiness:  exec [/bin/calico-node -felix-ready -bird-ready] delay=0s timeout=1s period=10s #success=1 #failure=3
    Environment:
      ETCD_ENDPOINTS:                     <set to the key 'etcd_endpoints' of config map 'calico-config'>  Optional: false
      ETCD_CA_CERT_FILE:                  <set to the key 'etcd_ca' of config map 'calico-config'>         Optional: false
      ETCD_KEY_FILE:                      <set to the key 'etcd_key' of config map 'calico-config'>        Optional: false
      ETCD_CERT_FILE:                     <set to the key 'etcd_cert' of config map 'calico-config'>       Optional: false
      CALICO_K8S_NODE_REF:                 (v1:spec.nodeName)
      CALICO_NETWORKING_BACKEND:          <set to the key 'calico_backend' of config map 'calico-config'>  Optional: false
      CLUSTER_TYPE:                       k8s,bgp
      IP:                                 autodetect
      CALICO_IPV4POOL_IPIP:               Always
      CALICO_IPV4POOL_VXLAN:              Never
      FELIX_IPINIPMTU:                    <set to the key 'veth_mtu' of config map 'calico-config'>  Optional: false
      FELIX_VXLANMTU:                     <set to the key 'veth_mtu' of config map 'calico-config'>  Optional: false
      FELIX_WIREGUARDMTU:                 <set to the key 'veth_mtu' of config map 'calico-config'>  Optional: false
      CALICO_IPV4POOL_CIDR:               172.168.0.0/12
      CALICO_DISABLE_FILE_LOGGING:        true
      FELIX_DEFAULTENDPOINTTOHOSTACTION:  ACCEPT
      FELIX_IPV6SUPPORT:                  false
      FELIX_LOGSEVERITYSCREEN:            info
      FELIX_HEALTHENABLED:                true
    Mounts:
      /calico-secrets from etcd-certs (rw)
      /lib/modules from lib-modules (ro)
      /run/xtables.lock from xtables-lock (rw)
      /var/lib/calico from var-lib-calico (rw)
      /var/run/calico from var-run-calico (rw)
      /var/run/nodeagent from policysync (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-tfw5v (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  lib-modules:
    Type:          HostPath (bare host directory volume)
    Path:          /lib/modules
    HostPathType:  
  var-run-calico:
    Type:          HostPath (bare host directory volume)
    Path:          /var/run/calico
    HostPathType:  
  var-lib-calico:
    Type:          HostPath (bare host directory volume)
    Path:          /var/lib/calico
    HostPathType:  
  xtables-lock:
    Type:          HostPath (bare host directory volume)
    Path:          /run/xtables.lock
    HostPathType:  FileOrCreate
  cni-bin-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /opt/cni/bin
    HostPathType:  
  cni-net-dir:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/cni/net.d
    HostPathType:  
  etcd-certs:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  calico-etcd-secrets
    Optional:    false
  policysync:
    Type:          HostPath (bare host directory volume)
    Path:          /var/run/nodeagent
    HostPathType:  DirectoryOrCreate
  flexvol-driver-host:
    Type:          HostPath (bare host directory volume)
    Path:          /usr/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds
    HostPathType:  DirectoryOrCreate
  kube-api-access-tfw5v:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   Burstable
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 :NoSchedule op=Exists
                             :NoExecute op=Exists
                             CriticalAddonsOnly op=Exists
                             node.kubernetes.io/disk-pressure:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/network-unavailable:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists
                             node.kubernetes.io/pid-pressure:NoSchedule op=Exists
                             node.kubernetes.io/unreachable:NoExecute op=Exists
                             node.kubernetes.io/unschedulable:NoSchedule op=Exists
Events:
  Type     Reason     Age                    From               Message
  ----     ------     ----                   ----               -------
  Normal   Scheduled  7m19s                  default-scheduler  Successfully assigned kube-system/calico-node-cb6c6 to k8s-master02
  Normal   Pulling    7m19s                  kubelet            Pulling image "registry.cn-beijing.aliyuncs.com/dotbalo/cni:v3.15.3"
  Normal   Pulled     6m57s                  kubelet            Successfully pulled image "registry.cn-beijing.aliyuncs.com/dotbalo/cni:v3.15.3" in 22.396538132s
  Normal   Created    6m56s                  kubelet            Created container install-cni
  Normal   Started    6m56s                  kubelet            Started container install-cni
  Normal   Pulling    6m55s                  kubelet            Pulling image "registry.cn-beijing.aliyuncs.com/dotbalo/pod2daemon-flexvol:v3.15.3"
  Normal   Pulling    6m51s                  kubelet            Pulling image "registry.cn-beijing.aliyuncs.com/dotbalo/node:v3.15.3"
  Normal   Pulled     6m51s                  kubelet            Successfully pulled image "registry.cn-beijing.aliyuncs.com/dotbalo/pod2daemon-flexvol:v3.15.3" in 3.78191146s
  Normal   Created    6m51s                  kubelet            Created container flexvol-driver
  Normal   Started    6m51s                  kubelet            Started container flexvol-driver
  Normal   Pulled     6m12s                  kubelet            Successfully pulled image "registry.cn-beijing.aliyuncs.com/dotbalo/node:v3.15.3" in 39.059498113s
  Warning  Unhealthy  6m7s (x4 over 6m12s)   kubelet            Readiness probe failed: calico/node is not ready: BIRD is not ready: Failed to stat() nodename file: stat /var/lib/calico/nodename: no such file or directory
  Normal   Created    5m49s (x3 over 6m12s)  kubelet            Created container calico-node
  Normal   Started    5m49s (x3 over 6m12s)  kubelet            Started container calico-node
  Normal   Pulled     5m49s (x2 over 6m10s)  kubelet            Container image "registry.cn-beijing.aliyuncs.com/dotbalo/node:v3.15.3" already present on machine
  Warning  BackOff    2m10s (x22 over 6m7s)  kubelet            Back-off restarting failed containerp 
pod详细信息显示:Readiness probe failed导致重启pod失败。
[root@k8s-master01 ~]# kubectl logs calico-node-cb6c6  -n kube-system
2021-09-14 09:21:43.819 [INFO][8] startup/startup.go 349: Early log level set to info
2021-09-14 09:21:43.819 [INFO][8] startup/startup.go 369: Using HOSTNAME environment (lowercase) for node name
2021-09-14 09:21:43.819 [INFO][8] startup/startup.go 377: Determined node name: k8s-master02
2021-09-14 09:21:43.819 [INFO][8] startup/startup.go 109: Skipping datastore connection test
2021-09-14 09:21:43.830 [INFO][8] startup/startup.go 460: Building new node resource Name="k8s-master02"
2021-09-14 09:21:45.832 [ERROR][8] startup/startup.go 149: failed to query kubeadm's config map error=Get "https://10.96.0.1:443/api/v1/namespaces/kube-system/configmaps/kubeadm-config?timeout=2s": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
2021-09-14 09:21:45.832 [WARNING][8] startup/startup.go 1264: Terminating
Calico node failed to start
pod日志显示:未能成功访问kubeadm's config导致calico启动失败。
解决方法:calico-node默认访问kubernetes SVC的443端口,这将导致无法访问apiserver,需要在yaml文件添加apiserver的IP和端口,字段名:KUBERNETES_SERVICE_HOST、KUBERNETES_SERVICE_PORT、KUBERNETES_SERVICE_PORT_HTTPS。




问题5:kubeadm方式部署K8S集群时,安装calico网络组件成功后,但coredns依然处于ContainerCreating?
[root@k8s-master01 .kube]# kubectl get pod -A -o wide NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES kube-system calico-kube-controllers-cdd5755b9-phg94 0/1 Running 0 73m 192.168.1.203 k8s-master03 <none> <none> kube-system calico-node-5d6xr 1/1 Running 0 50m 192.168.1.202 k8s-master02 <none> <none> kube-system calico-node-k7qmf 1/1 Running 0 50m 192.168.1.204 k8s-node01 <none> <none> kube-system calico-node-p47mm 1/1 Running 0 50m 192.168.1.201 k8s-master01 <none> <none> kube-system calico-node-v7q76 1/1 Running 0 50m 192.168.1.203 k8s-master03 <none> <none> kube-system calico-node-wjw4d 1/1 Running 0 50m 192.168.1.205 k8s-node02 <none> <none> kube-system coredns-7d89d9b6b8-ddjjp 0/1 ContainerCreating 0 92m <none> k8s-master03 <none> <none> kube-system coredns-7d89d9b6b8-sbd8k 0/1 ContainerCreating 0 92m <none> k8s-master03 <none> <none>
原因分析:
[root@k8s-master01 ~]# kubectl logs coredns-7d89d9b6b8-ddjjp -n kube-system Error from server (BadRequest): container "coredns" in pod "coredns-7d89d9b6b8-ddjjp" is waiting to start: ContainerCreating
日志显示:容器创建中。
[root@k8s-master01 ~]# kubectl describe pod coredns-7d89d9b6b8-ddjjp -n kube-system
Name:                 coredns-7d89d9b6b8-ddjjp
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Node:                 k8s-master03/192.168.1.203
Start Time:           Tue, 14 Sep 2021 17:14:54 +0800
Labels:               k8s-app=kube-dns
                      pod-template-hash=7d89d9b6b8
Annotations:          <none>
Status:               Pending
IP:                   
IPs:                  <none>
Controlled By:        ReplicaSet/coredns-7d89d9b6b8
Containers:
  coredns:
    Container ID:  
    Image:         registry.cn-hangzhou.aliyuncs.com/google_containers/coredns:v1.8.4
    Image ID:      
    Ports:         53/UDP, 53/TCP, 9153/TCP
    Host Ports:    0/UDP, 0/TCP, 0/TCP
    Args:
      -conf
      /etc/coredns/Corefile
    State:          Waiting
      Reason:       ContainerCreating
    Ready:          False
    Restart Count:  0
    Limits:
      memory:  170Mi
    Requests:
      cpu:        100m
      memory:     70Mi
    Liveness:     http-get http://:8080/health delay=60s timeout=5s period=10s #success=1 #failure=5
    Readiness:    http-get http://:8181/ready delay=0s timeout=1s period=10s #success=1 #failure=3
    Environment:  <none>
    Mounts:
      /etc/coredns from config-volume (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-7m9rx (ro)
Conditions:
  Type              Status
  Initialized       True 
  Ready             False 
  ContainersReady   False 
  PodScheduled      True 
Volumes:
  config-volume:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      coredns
    Optional:  false
  kube-api-access-7m9rx:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
QoS Class:                   Burstable
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 CriticalAddonsOnly op=Exists
                             node-role.kubernetes.io/control-plane:NoSchedule
                             node-role.kubernetes.io/master:NoSchedule
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason                  Age                    From               Message
  ----     ------                  ----                   ----               -------
Normal SandboxChanged 6m21s (x2093 over 136m) kubelet Pod sandbox changed, it will be killed and re-created. Warning FailedCreatePodSandBox 82s (x2154 over 136m) kubelet (combined from similar events): Failed to create pod sandbox: rpc error: code = Unknown desc = failed to set up sandbox container "089f15d2961982244b2ce122b738afc70e29fa17eb79a4e7b88a35ce5be4868b" network for pod "coredns-7d89d9b6b8-ddjjp": networkPlugin cni failed to set up pod "coredns-7d89d9b6b8-ddjjp_kube-system" network: Get "https://[10.96.0.1]:443/api/v1/namespaces/kube-system": dial tcp 10.96.0.1:443: connect: connection refused
pod详细信息显示:调度节点失败。
[root@k8s-master01 .kube]# systemctl status kubelet
● kubelet.service - kubelet: The Kubernetes Node Agent
   Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled)
  Drop-In: /usr/lib/systemd/system/kubelet.service.d
           └─10-kubeadm.conf
   Active: active (running) since Tue 2021-09-14 16:56:06 CST; 1h 42min ago
     Docs: https://kubernetes.io/docs/
 Main PID: 28526 (kubelet)
    Tasks: 14
   Memory: 78.6M
   CGroup: /system.slice/kubelet.service
           └─28526 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --net...
Sep 14 18:37:41 k8s-master01 kubelet[28526]: I0914 18:37:41.703532   28526 scope.go:110] "RemoveContainer" containerID="a9e6b2c2df1bd9530394a4077af6c159830efd7e2b2d9156cbf1580086bf6d99"
Sep 14 18:37:41 k8s-master01 kubelet[28526]: E0914 18:37:41.703788   28526 pod_workers.go:747] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-proxy\" with ...
Sep 14 18:37:52 k8s-master01 kubelet[28526]: I0914 18:37:52.704002   28526 scope.go:110] "RemoveContainer" containerID="a9e6b2c2df1bd9530394a4077af6c159830efd7e2b2d9156cbf1580086bf6d99"
Sep 14 18:37:52 k8s-master01 kubelet[28526]: E0914 18:37:52.704281   28526 pod_workers.go:747] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-proxy\" with ...
Sep 14 18:38:07 k8s-master01 kubelet[28526]: I0914 18:38:07.704831   28526 scope.go:110] "RemoveContainer" containerID="a9e6b2c2df1bd9530394a4077af6c159830efd7e2b2d9156cbf1580086bf6d99"
Sep 14 18:38:07 k8s-master01 kubelet[28526]: E0914 18:38:07.705058   28526 pod_workers.go:747] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-proxy\" with ...
Sep 14 18:38:22 k8s-master01 kubelet[28526]: I0914 18:38:22.705015   28526 scope.go:110] "RemoveContainer" containerID="a9e6b2c2df1bd9530394a4077af6c159830efd7e2b2d9156cbf1580086bf6d99"
Sep 14 18:38:22 k8s-master01 kubelet[28526]: E0914 18:38:22.705279   28526 pod_workers.go:747] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-proxy\" with ...
Sep 14 18:38:34 k8s-master01 kubelet[28526]: I0914 18:38:34.709678   28526 scope.go:110] "RemoveContainer" containerID="a9e6b2c2df1bd9530394a4077af6c159830efd7e2b2d9156cbf1580086bf6d99"
Sep 14 18:38:34 k8s-master01 kubelet[28526]: E0914 18:38:34.711197   28526 pod_workers.go:747] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-proxy\" with ...
Hint: Some lines were ellipsized, use -l to show in full.
解决方法:修改kubelet服务配置即可。


问题5:kubeadm方式部署K8S集群时,部署flannel组件失败?
kube-flannel-ds-amd64-8cqqz 0/1 CrashLoopBackOff 3 84s 192.168.66.10 k8s-master01 <none> <none>

原因分析:查看日志,发现注册网络失败,原因在于主节点初始化时yaml文件存在问题。
kubectl logs kube-flannel-ds-amd64-8cqqz -n kubesystem
I0602 01:53:54.021093 1 main.go:514] Determining IP address of default interface
I0602 01:53:54.022514       1 main.go:527] Using interface with name ens33 and address 192.168.66.10
I0602 01:53:54.022619       1 main.go:544] Defaulting external address to interface address (192.168.66.10)
I0602 01:53:54.030311       1 kube.go:126] Waiting 10m0s for node controller to sync
I0602 01:53:54.030555       1 kube.go:309] Starting kube subnet manager
I0602 01:53:55.118656       1 kube.go:133] Node controller sync successful
I0602 01:53:55.118754       1 main.go:244] Created subnet manager: Kubernetes Subnet Manager - k8s-master01
I0602 01:53:55.118765       1 main.go:247] Installing signal handlers
I0602 01:53:55.119057       1 main.go:386] Found network config - Backend type: vxlan
I0602 01:53:55.119146       1 vxlan.go:120] VXLAN config: VNI=1 Port=0 GBP=false DirectRouting=false
E0602 01:53:55.119470       1 main.go:289] Error registering network: failed to acquire lease: node "k8s-master01" pod cidr not assigned
I0602 01:53:55.119506       1 main.go:366] Stopping shutdownHandler...

解决方法:修改kubeadm-config.yaml后重新初始化主节点即可。

问题6:K8S集群初始化主节点失败?

[init] Using Kubernetes version: v1.15.1
[preflight] Running pre-flight checks
        [WARNING SystemVerification]: this Docker version is not on the list of validated versions: 20.10.6. Latest validated version: 18.09
error execution phase preflight: [preflight] Some fatal errors occurred:
        [ERROR Port-6443]: Port 6443 is in use
        [ERROR Port-10251]: Port 10251 is in use
        [ERROR Port-10252]: Port 10252 is in use
        [ERROR FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml]: /etc/kubernetes/manifests/kube-apiserver.yaml already exists
        [ERROR FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml]: /etc/kubernetes/manifests/kube-controller-manager.yaml already exists
        [ERROR FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml]: /etc/kubernetes/manifests/kube-scheduler.yaml already exists
        [ERROR FileAvailable--etc-kubernetes-manifests-etcd.yaml]: /etc/kubernetes/manifests/etcd.yaml already exists
        [ERROR Port-10250]: Port 10250 is in use
        [ERROR Port-2379]: Port 2379 is in use
        [ERROR Port-2380]: Port 2380 is in use
        [ERROR DirAvailable--var-lib-etcd]: /var/lib/etcd is not empty
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
原因分析:K8S集群已进行过初始化主节点。
解决方法:重置K8S后重新初始化。
kubeadm reset
kubeadm init --config=kubeadm-config.yaml --upload-certs |tee kubeadm-init.log
问题7:重置K8S成功后,是否需要删除相关文件?
[reset] WARNING: Changes made to this host by 'kubeadm init' or 'kubeadm join' will be reverted.
[reset] Are you sure you want to proceed? [y/N]: y
[preflight] Running pre-flight checks
W0602 10:20:53.656954   76680 removeetcdmember.go:79] [reset] No kubeadm config, using etcd pod spec to get data directory
[reset] No etcd config found. Assuming external etcd
[reset] Please, manually reset etcd to prevent further issues
[reset] Stopping the kubelet service
[reset] Unmounting mounted directories in "/var/lib/kubelet"
[reset] Deleting contents of config directories: [/etc/kubernetes/manifests /etc/kubernetes/pki]
[reset] Deleting files: [/etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/bootstrap-kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf]
[reset] Deleting contents of stateful directories: [/var/lib/kubelet /etc/cni/net.d /var/lib/dockershim /var/run/kubernetes]
The reset process does not reset or clean up iptables rules or IPVS tables.
If you wish to reset iptables, you must do so manually.
For example:
iptables -F && iptables -t nat -F && iptables -t mangle -F && iptables -X
If your cluster was setup to utilize IPVS, run ipvsadm --clear (or similar)
to reset your system's IPVS tables.
The reset process does not clean your kubeconfig files and you must remove them manually.
Please, check the contents of the $HOME/.kube/config file.

原因分析:无。
解决方法:可根据提示删除相关文件,避免主节点初始化后引起其他问题。
问题8:主节点初始化成功后,查看节点信息失败?

Unable to connect to the server: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "kubernetes")
原因分析:初始化主节点后,未导入kubelet的config文件。
解决方法:导入config文件即可。
mkdir -p $HOME/.kube
cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
chown $(id -u):$(id -g) $HOME/.kube/config
问题9、master节点加入K8S集群失败?
[root@k8s-master02 ~]# kubeadm join 192.168.1.201:6443 --token 6xyv8a.cueltqmpe9qa8nxu --discovery-token-ca-cert-hash sha256:bd78dfd370e47dfca742b5f6934c21014792168fa4dc19c9fa63bfdd87270097 \
>        --control-plane --certificate-key b464a8d23d3313c4c0bb5b65648b039cb9b1177dddefbf46e2e296899d0e4516
[preflight] Running pre-flight checks
[preflight] Reading configuration from the cluster...
[preflight] FYI: You can look at this config file with 'kubectl -n kube-system get cm kubeadm-config -oyaml'
error execution phase preflight: 
One or more conditions for hosting a new control plane instance is not satisfied.
unable to add a new control plane instance a cluster that doesn't have a stable controlPlaneEndpoint address
Please ensure that:
* The cluster has a stable controlPlaneEndpoint address.
* The certificates that must be shared among control plane instances are provided.
To see the stack trace of this error execute with --v=5 or higher
原因分析:证书未共享。
解决方法:共享证书即可!!!
##########其他master节点执行如下内容#################################
mkdir -p /etc/kubernetes/pki/etcd/
##########master01节点执行如下内容#################################
cd /etc/kubernetes/pki/
scp ca.* front-proxy-ca.* sa.* 192.168.1.202:/etc/kubernetes/pki/
scp ca.* front-proxy-ca.* sa.* 192.168.1.203:/etc/kubernetes/pki/
##########其他master节点执行如下内容#################################
kubeadm join 192.168.1.201:6443 --token 6xyv8a.cueltqmpe9qa8nxu --discovery-token-ca-cert-hash sha256:bd78dfd370e47dfca742b5f6934c21014792168fa4dc19c9fa63bfdd87270097 --control-plane --certificate-key b464a8d23d3313c4c0bb5b65648b039cb9b1177dddefbf46e2e296899d0e4516
问题10、部署prometheus失败?


unable to recognize "0prometheus-operator-serviceMonitor.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.co m/v1"
unable to recognize "alertmanager-alertmanager.yaml": no matches for kind "Alertmanager" in version "monitoring.coreos.com/v1"
unable to recognize "alertmanager-serviceMonitor.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/v1"
unable to recognize "grafana-serviceMonitor.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/v1"
unable to recognize "kube-state-metrics-serviceMonitor.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/                                v1"
unable to recognize "node-exporter-serviceMonitor.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/v1"
unable to recognize "prometheus-prometheus.yaml": no matches for kind "Prometheus" in version "monitoring.coreos.com/v1"
unable to recognize "prometheus-rules.yaml": no matches for kind "PrometheusRule" in version "monitoring.coreos.com/v1"
unable to recognize "prometheus-serviceMonitor.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/v1"
unable to recognize "prometheus-serviceMonitorApiserver.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com                                /v1"
unable to recognize "prometheus-serviceMonitorCoreDNS.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/v                                1"
unable to recognize "prometheus-serviceMonitorKubeControllerManager.yaml": no matches for kind "ServiceMonitor" in version "monitorin                                g.coreos.com/v1"
unable to recognize "prometheus-serviceMonitorKubeScheduler.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos                                .com/v1"
unable to recognize "prometheus-serviceMonitorKubelet.yaml": no matches for kind "ServiceMonitor" in version "monitoring.coreos.com/v                                1"
原因分析:不明。
解决方法:重新执行发布命令。


问题7:无法下载K8S高可用安装包,但能直接通过浏览器下载git包?
[root@k8s-master01 install-k8s-v1.17]# git clone https://github.com/dotbalo/k8s-ha-install.git
Cloning into 'k8s-ha-install'...
error: RPC failed; result=35, HTTP code = 0
fatal: The remote end hung up unexpectedly
原因分析: git buffer 太小了。
解决方法:git config --global http.postBuffer 100M #git buffer增大。

问题8:无法下载K8S高可用安装包,但能直接通过浏览器下载git包?
[root@k8s-master01 install-k8s-v1.17]# git clone https://github.com/dotbalo/k8s-ha-install.git
Cloning into 'k8s-ha-install'...
fatal: unable to access 'https://github.com/dotbalo/k8s-ha-install.git/': Empty reply from server

原因分析:不明;
解决方法:手动下载并上传git包;
问题9:K8S查看分支失败?
[root@k8s-master01 k8s-ha-install-master]# git branch -a
fatal: Not a git repository (or any of the parent directories): .git
原因分析:缺少.git本地仓库。
解决方法:初始化git即可。
[root@k8s-master01 k8s-ha-install-master]# git init
Initialized empty Git repository in /root/install-k8s-v1.17/k8s-ha-install-master/.git/
[root@k8s-master01 k8s-ha-install-master]# git branch -a
 
问题10:K8S切换分支失败?
[root@k8s-master01 k8s-ha-install-master]# git checkout manual-installation-v1.20.x
error: pathspec 'manual-installation-v1.20.x' did not match any file(s) known to git.

原因分析:没有发现分支。
解决方法:必须通过git下载,不能通过浏览器下载zip格式的文件;或将git下载的文件打包后再次解压也不可使用。
git clone https://github.com/dotbalo/k8s-ha-install.git


[root@k8s-master01 k8s-ha-install]# git checkout manual-installation-v1.20.x
Branch manual-installation-v1.20.x set up to track remote branch manual-installation-v1.20.x from origin.
Switched to a new branch 'manual-installation-v1.20.x'
问题11:apiserver聚合证书生成失败?
[root@k8s-master01 pki]# cfssl gencert \
>     -ca=/etc/kubernetes/pki/front-proxy-ca.pem \
>     -ca-key=/etc/kubernetes/pki/front-proxy-ca-key.pem \
>     -config=ca-config.json \
>     -profile=kubernetes \
>              front-proxy-client-csr.json | cfssljson -bare /etc/kubernetes/pki/front-proxy-client
2021/07/13 16:23:31 [INFO] generate received request
2021/07/13 16:23:31 [INFO] received CSR
2021/07/13 16:23:31 [INFO] generating key: rsa-2048
2021/07/13 16:23:31 [INFO] encoded CSR
2021/07/13 16:23:31 [INFO] signed certificate with serial number 723055319564838988438867282088171857759978642034
2021/07/13 16:23:31 [WARNING] This certificate lacks a "hosts" field. This makes it unsuitable for
websites. For more information see the Baseline Requirements for the Issuance and Management
of Publicly-Trusted Certificates, v.1.1.6, from the CA/Browser Forum (https://cabforum.org);
specifically, section 10.2.3 ("Information Requirements")
原因分析:命令缺少hosts参数,不适合用于网站;但不影响apiserver组件与其他组件通信。
解决方法:无需关注。

问题12:二进制安装K8S组件后,出现警告?
[root@k8s-master01 ~]# kubectl version
Client Version: version.Info{Major:"1", Minor:"20", GitVersion:"v1.20.0", GitCommit:"af46c47ce925f4c4ad5cc8d1fca46c7b77d13b38", GitTreeState:"clean", BuildDate:"2020-12-08T17:59:43Z", GoVersion:"go1.15.5", Compiler:"gc", Platform:"linux/amd64"}
The connection to the server localhost:8080 was refused - did you specify the right host or port?
原因分析:kubectl组件安装成功,但没有成功连接kubernetes apiserver;
解决方法:安装证书并配置组件即可。
问题13:二进制安装K8S进行证书配置后,启动ETCD失败?
[root@k8s-master02 ~]# systemctl status etcd 
● etcd.service - Etcd Service
   Loaded: loaded (/usr/lib/systemd/system/etcd.service; enabled; vendor preset: disabled)
   Active: activating (start) since Wed 2021-07-21 08:51:42 CST; 1min 13s ago
     Docs: https://coreos.com/etcd/docs/latest/
 Main PID: 1424 (etcd)
    Tasks: 9
   Memory: 33.1M
   CGroup: /system.slice/etcd.service
           └─1424 /usr/local/bin/etcd --config-file=/etc/etcd/etcd.config.yml
Jul 21 08:52:55 k8s-master02 etcd[1424]: rejected connection from "192.168.0.107:38222" (error "remote error: tls: bad certificate", ServerName "")
Jul 21 08:52:55 k8s-master02 etcd[1424]: rejected connection from "192.168.0.107:38224" (error "remote error: tls: bad certificate", ServerName "")
原因分析:认证失败;可能原因:创建证书的csr文件问题导致证书错误,IP未加入hosts中。
解决方法:重新生成证书即可,处理方法详见问题21.
问题14:haproxy启动失败?
[root@k8s-master01 pki]# systemctl enable --now haproxy;systemctl status haproxy
Created symlink from /etc/systemd/system/multi-user.target.wants/haproxy.service to /usr/lib/systemd/system/haproxy.service.
● haproxy.service - HAProxy Load Balancer
   Loaded: loaded (/usr/lib/systemd/system/haproxy.service; enabled; vendor preset: disabled)
   Active: failed (Result: exit-code) since Wed 2021-07-21 10:25:35 CST; 3ms ago
  Process: 2255 ExecStart=/usr/sbin/haproxy-systemd-wrapper -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid $OPTIONS (code=exited, status=1/FAILURE)
 Main PID: 2255 (code=exited, status=1/FAILURE)
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: [ALERT] 201/102535 (2256) : parsing [/etc/haproxy/haproxy.cfg:59] : 'server 192.168.0.107:6443' : invalid address...in 'check'
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: [ALERT] 201/102535 (2256) : parsing [/etc/haproxy/haproxy.cfg:60] : 'server 192.168.0.108:6443' : invalid address...in 'check'
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: [ALERT] 201/102535 (2256) : parsing [/etc/haproxy/haproxy.cfg:61] : 'server 192.168.0.109:6443' : invalid address...in 'check'
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: [ALERT] 201/102535 (2256) : Error(s) found in configuration file : /etc/haproxy/haproxy.cfg
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: [WARNING] 201/102535 (2256) : config : frontend 'GLOBAL' has no 'bind' directive. Please declare it as a backend ... intended.
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: [ALERT] 201/102535 (2256) : Fatal errors found in configuration.
Jul 21 10:25:35 k8s-master01 systemd[1]: haproxy.service: main process exited, code=exited, status=1/FAILURE
Jul 21 10:25:35 k8s-master01 haproxy-systemd-wrapper[2255]: haproxy-systemd-wrapper: exit, haproxy RC=1
Jul 21 10:25:35 k8s-master01 systemd[1]: Unit haproxy.service entered failed state.
Jul 21 10:25:35 k8s-master01 systemd[1]: haproxy.service failed.
Hint: Some lines were ellipsized, use -l to show in full.
原因分析:配置错误。
解决方法:添加option tcp-check即可。

问题15:apiserver启动失败?
[root@k8s-master01 ~]# systemctl daemon-reload && systemctl enable --now kube-apiserver
Failed to execute operation: No such file or directory
原因分析:配置问题,etcd不能正常工作原因。

解决方法:修改配置即可。
问题16:kubelet服务启动失败?
● kubelet.service - kubelet: The Kubernetes Node Agent
   Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled)
  Drop-In: /usr/lib/systemd/system/kubelet.service.d
           └─10-kubeadm.conf
   Active: activating (auto-restart) (Result: exit-code) since Fri 2021-08-13 12:15:11 CST; 948ms ago
     Docs: https://kubernetes.io/docs/
  Process: 5600 ExecStart=/usr/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS (code=exited, status=1/FAILURE)
 Main PID: 5600 (code=exited, status=1/FAILURE)
Aug 13 12:15:11 k8s-master02 systemd[1]: Unit kubelet.service entered failed state.
Aug 13 12:15:11 k8s-master02 systemd[1]: kubelet.service failed.
原因分析:selinux未关闭,且未进行初始化;
解决方法:关闭selinux并进行初始化。

问题17:k8s主节点初始化失败?
[root@k8s-master01 ~]# kubeadm init --control-plane-endpoint "k8s-master01-lb:16443" --upload-certs
I0813 13:49:07.164695   10196 version.go:254] remote version is much newer: v1.22.0; falling back to: stable-1.20
[init] Using Kubernetes version: v1.20.10
[preflight] Running pre-flight checks
error execution phase preflight: [preflight] Some fatal errors occurred:
        [ERROR KubeletVersion]: the kubelet version is higher than the control plane version. This is not a supported version skew and may lead to a malfunctional cluster. Kubelet version: "1.22.0" Control plane version: "1.20.10"
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
To see the stack trace of this error execute with --v=5 or higher
原因分析:kubeadm版本不一致;

解决方法:重新安装kubeadm即可。
yum -y remove kubectl kubelet && yum -y install kubectl-1.20.5-0 kubelet-1.20.5-0 kubeadm-1.20.5-0
问题18:kubeadm安装k8s,初始化后无法查看node状态?
[root@k8s-master01 ~]# kubectl get node
The connection to the server localhost:8080 was refused - did you specify the right host or port?
原因分析:kubectl无法与K8S集群通信;
解决方法:.kube目录是用于存放kubectl与K8S集群交互的缓存文件及相关配置文件;
mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config

问题20:calico安装异常?
[root@k8s-master01 calico]# kubectl describe pod calico-node-dtqc8 -n kube-system
Name:                 calico-node-dtqc8
Namespace:            kube-system
......
Events:
  Type     Reason     Age    From               Message
  ----     ------     ----   ----               -------
  Normal   Scheduled  10m    default-scheduler  Successfully assigned kube-system/calico-node-dtqc8 to k8s-master01
  Normal   Pulling    10m    kubelet            Pulling image "registry.cn-beijing.aliyuncs.com/dotbalo/cni:v3.15.3"
  Normal   Pulled     10m    kubelet            Successfully pulled image "registry.cn-beijing.aliyuncs.com/dotbalo/cni:v3.15.3" in 44.529354646s
  Normal   Created    10m    kubelet            Created container install-cni
  Normal   Started    10m    kubelet            Started container install-cni
  Normal   Pulling    10m    kubelet            Pulling image "registry.cn-beijing.aliyuncs.com/dotbalo/pod2daemon-flexvol:v3.15.3"
  Normal   Pulled     9m45s  kubelet            Successfully pulled image "registry.cn-beijing.aliyuncs.com/dotbalo/pod2daemon-flexvol:v3.15.3" in 27.031642012s
  Normal   Created    9m45s  kubelet            Created container flexvol-driver
  Normal   Started    9m44s  kubelet            Started container flexvol-driver
  Normal   Pulling    9m43s  kubelet            Pulling image "registry.cn-beijing.aliyuncs.com/dotbalo/node:v3.15.3"
  Normal   Pulled     8m42s  kubelet            Successfully pulled image "registry.cn-beijing.aliyuncs.com/dotbalo/node:v3.15.3" in 1m1.78614215s
  Normal   Created    8m41s  kubelet            Created container calico-node
  Normal   Started    8m41s  kubelet            Started container calico-node
  Warning  Unhealthy  8m39s  kubelet            Readiness probe failed: calico/node is not ready: BIRD is not ready: Error querying BIRD: unable to connect to BIRDv4 socket: dial unix /var/run/calico/bird.ctl: connect: connection refused
原因分析:不明。
解决方法:暂无影响,可不理会。

问题21:kubectl工具不能使用?
[root@k8s-master01 ~]# kubectl get nodes
The connection to the server 192.168.1.201:6443 was refused - did you specify the right host or port?

 
原因分析:未设置环境变量;
解决方法:
cat <<eof>> /root/.bashrc
export KUBECONFIG=/etc/kubernetes/admin.conf
eof
source /root/.bashrc

问题22:第一个master节点初始化失败?
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[kubelet-check] Initial timeout of 40s passed.
        Unfortunately, an error has occurred:
                timed out waiting for the condition
        This error is likely caused by:
                - The kubelet is not running
                - The kubelet is unhealthy due to a misconfiguration of the node in some way (required cgroups disabled)
        If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
                - 'systemctl status kubelet'
                - 'journalctl -xeu kubelet'
        Additionally, a control plane component may have crashed or exited when started by the container runtime.
        To troubleshoot, list all containers using your preferred container runtimes CLI.
        Here is one example how you may list all Kubernetes containers running in docker:
                - 'docker ps -a | grep kube | grep -v pause'
                Once you have found the failing container, you can inspect its logs with:
                - 'docker logs CONTAINERID'
error execution phase wait-control-plane: couldn't initialize a Kubernetes cluster
To see the stack trace of this error execute with --v=5 or higher
原因分析:不明;



解决方法:待解决;
问题23、helm安装ingress失败?
[root@k8s-master01 ingress-nginx]# helm install ingress-nginx -n ingress-nginx .
Error: cannot re-use a name that is still in use
原因分析:安装失败后,缓存记录了ingress-nginx;
解决方法:删除记录即可。
[root@k8s-master01 ingress-nginx]# helm uninstall ingress-nginx -n ingress-nginx
Error: uninstall: Release not loaded: ingress-nginx: release: not found
[root@k8s-master01 ingress-nginx]# 
[root@k8s-master01 ingress-nginx]# 
[root@k8s-master01 ingress-nginx]# helm install ingress-nginx -n ingress-nginx .
NAME: ingress-nginx
LAST DEPLOYED: Tue Aug 31 16:00:45 2021
NAMESPACE: ingress-nginx
STATUS: deployed
REVISION: 1
TEST SUITE: None
NOTES:
The ingress-nginx controller has been installed.
Get the application URL by running these commands:
  export POD_NAME=$(kubectl --namespace ingress-nginx get pods -o jsonpath="{.items[0].metadata.name}" -l "app=ingress-nginx,component=,release=ingress-nginx")
  kubectl --namespace ingress-nginx port-forward $POD_NAME 8080:80
  echo "Visit http://127.0.0.1:8080 to access your application."
An example Ingress that makes use of the controller:
  apiVersion: networking.k8s.io/v1beta1
  kind: Ingress
  metadata:
    annotations:
      kubernetes.io/ingress.class: nginx
    name: example
    namespace: foo
  spec:
    rules:
      - host: www.example.com
        http:
          paths:
            - backend:
                serviceName: exampleService
                servicePort: 80
              path: /
    # This section is only required if TLS is to be enabled for the Ingress
    tls:
        - hosts:
            - www.example.com
          secretName: example-tls
If TLS is enabled for the Ingress, a Secret containing the certificate and key must also be provided:
  apiVersion: v1
  kind: Secret
  metadata:
    name: example-tls
    namespace: foo
  data:
    tls.crt: <base64 encoded cert>
    tls.key: <base64 encoded key>
  type: kubernetes.io/tls
问题24:helm安装ingress失败,进入调试模式查看原因?
[root@k8s-master01 ingress-nginx]# helm install ingress-nginx -n ingress-nginx . --debug
install.go:148: [debug] Original chart version: ""
install.go:165: [debug] CHART PATH: /root/ingress-nginx
Error: could not get apiVersions from Kubernetes: unable to retrieve the complete list of server APIs: metrics.k8s.io/v1beta1: the server is currently unable to handle the request
helm.go:76: [debug] unable to retrieve the complete list of server APIs: metrics.k8s.io/v1beta1: the server is currently unable to handle the request
could not get apiVersions from Kubernetes
helm.sh/helm/v3/pkg/action.(*Configuration).getCapabilities
        /home/circleci/helm.sh/helm/pkg/action/action.go:114
helm.sh/helm/v3/pkg/action.(*Install).Run
        /home/circleci/helm.sh/helm/pkg/action/install.go:195
main.runInstall
        /home/circleci/helm.sh/helm/cmd/helm/install.go:209
main.newInstallCmd.func1
        /home/circleci/helm.sh/helm/cmd/helm/install.go:115
github.com/spf13/cobra.(*Command).execute
        /go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:826
github.com/spf13/cobra.(*Command).ExecuteC
        /go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:914
github.com/spf13/cobra.(*Command).Execute
        /go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:864
main.main
        /home/circleci/helm.sh/helm/cmd/helm/helm.go:75
runtime.main
        /usr/local/go/src/runtime/proc.go:203
runtime.goexit
        /usr/local/go/src/runtime/asm_amd64.s:1357
原因分析:存在错误的apiservice;
解决方法:删除错误的apiservice即可。

问题25:helm安装ingress失败,进入调试模式查看原因?
[root@k8s-master01 ingress-nginx]# helm install ingress-nginx -n ingress-nginx . --debug
install.go:148: [debug] Original chart version: ""
install.go:165: [debug] CHART PATH: /root/ingress-nginx
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: kind: ConfigMap, namespace: ingress-nginx, name: ingress-nginx-controller
helm.go:76: [debug] existing resource conflict: kind: ConfigMap, namespace: ingress-nginx, name: ingress-nginx-controller
rendered manifests contain a resource that already exists. Unable to continue with install
helm.sh/helm/v3/pkg/action.(*Install).Run
        /home/circleci/helm.sh/helm/pkg/action/install.go:242
main.runInstall
        /home/circleci/helm.sh/helm/cmd/helm/install.go:209
main.newInstallCmd.func1
        /home/circleci/helm.sh/helm/cmd/helm/install.go:115
github.com/spf13/cobra.(*Command).execute
        /go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:826
github.com/spf13/cobra.(*Command).ExecuteC
        /go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:914
github.com/spf13/cobra.(*Command).Execute
        /go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:864
main.main
        /home/circleci/helm.sh/helm/cmd/helm/helm.go:75
runtime.main
        /usr/local/go/src/runtime/proc.go:203
runtime.goexit
        /usr/local/go/src/runtime/asm_amd64.s:1357
[root@k8s-master01 ingress-nginx]# helm install ingress-nginx -n ingress-nginx . --debug
install.go:148: [debug] Original chart version: ""
install.go:165: [debug] CHART PATH: /root/ingress-nginx
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: kind: ServiceAccount, namespace: ingress-nginx, name: ingress-nginx
。。。
Error: rendered manifests contain a resource that already exists. Unable to continue with install: existing resource conflict: kind: ClusterRole, namespace: , name: ingress-nginx
helm.go:76: [debug] existing resource conflict: kind: ClusterRole, namespace: , name: ingress-nginx
rendered manifests contain a resource that already exists. Unable to continue with install
。。。
原因分析:已存在configmap等资源;

解决方法:删除configmap等资源即可;
kubectl delete ConfigMap ingress-nginx-controller -n ingress-nginx
kubectl delete ServiceAccount ingress-nginx -n ingress-nginx
kubectl delete ClusterRole ingress-nginx
kubectl delete ClusterRoleBinding ingress-nginx
kubectl delete Role ingress-nginx -n ingress-nginx
kubectl delete RoleBinding ingress-nginx -n ingress-nginx
kubectl delete svc ingress-nginx-controller ingress-nginx-controller-admission -n ingress-nginx
kubectl delete DaemonSet ingress-nginx-controller -n ingress-nginx
kubectl delete ValidatingWebhookConfiguration ingress-nginx-admission
 问题26:etcd集群启动失败?
[root@k8s-master01 ~]# systemctl status etcd 
● etcd.service - Etcd Service
   Loaded: loaded (/usr/lib/systemd/system/etcd.service; enabled; vendor preset: disabled)
   Active: activating (start) since Mon 2021-09-06 20:03:22 CST; 47s ago
     Docs: https://coreos.com/etcd/docs/latest/
 Main PID: 1497 (etcd)
    Tasks: 8
   Memory: 23.2M
   CGroup: /system.slice/etcd.service
           └─1497 /usr/local/bin/etcd --config-file=/etc/etcd/etcd.config.yml
Sep 06 20:04:07 k8s-master01 etcd[1497]: raft2021/09/06 20:04:07 INFO: c05e87400fe17139 is starting a new election at term 928
Sep 06 20:04:07 k8s-master01 etcd[1497]: raft2021/09/06 20:04:07 INFO: c05e87400fe17139 became candidate at term 929
Sep 06 20:04:07 k8s-master01 etcd[1497]: raft2021/09/06 20:04:07 INFO: c05e87400fe17139 received MsgVoteResp from c05e87400fe17139 at term 929
Sep 06 20:04:07 k8s-master01 etcd[1497]: raft2021/09/06 20:04:07 INFO: c05e87400fe17139 [logterm: 1, index: 3] sent MsgVote request to 208ae89f369427bb at term 929
Sep 06 20:04:07 k8s-master01 etcd[1497]: raft2021/09/06 20:04:07 INFO: c05e87400fe17139 [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 929
Sep 06 20:04:09 k8s-master01 etcd[1497]: raft2021/09/06 20:04:09 INFO: c05e87400fe17139 is starting a new election at term 929
Sep 06 20:04:09 k8s-master01 etcd[1497]: raft2021/09/06 20:04:09 INFO: c05e87400fe17139 became candidate at term 930
Sep 06 20:04:09 k8s-master01 etcd[1497]: raft2021/09/06 20:04:09 INFO: c05e87400fe17139 received MsgVoteResp from c05e87400fe17139 at term 930
Sep 06 20:04:09 k8s-master01 etcd[1497]: raft2021/09/06 20:04:09 INFO: c05e87400fe17139 [logterm: 1, index: 3] sent MsgVote request to 208ae89f369427bb at term 930
Sep 06 20:04:09 k8s-master01 etcd[1497]: raft2021/09/06 20:04:09 INFO: c05e87400fe17139 [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 930
[root@k8s-master02 ~]# systemctl status etcd
● etcd.service - Etcd Service
   Loaded: loaded (/usr/lib/systemd/system/etcd.service; enabled; vendor preset: disabled)
   Active: activating (auto-restart) (Result: timeout) since Mon 2021-09-06 20:06:34 CST; 6s ago
     Docs: https://coreos.com/etcd/docs/latest/
  Process: 1531 ExecStart=/usr/local/bin/etcd --config-file=/etc/etcd/etcd.config.yml (code=killed, signal=TERM)
 Main PID: 1531 (code=killed, signal=TERM)
Sep 06 20:06:34 k8s-master02 systemd[1]: Failed to start Etcd Service.
Sep 06 20:06:34 k8s-master02 systemd[1]: Unit etcd.service entered failed state.
Sep 06 20:06:34 k8s-master02 systemd[1]: etcd.service failed.
[root@k8s-master03 ~]# systemctl status etcd
● etcd.service - Etcd Service
   Loaded: loaded (/usr/lib/systemd/system/etcd.service; enabled; vendor preset: disabled)
   Active: inactive (dead) since Mon 2021-09-06 20:13:50 CST; 1s ago
     Docs: https://coreos.com/etcd/docs/latest/
  Process: 1781 ExecStart=/usr/local/bin/etcd --config-file=/etc/etcd/etcd.config.yml (code=killed, signal=TERM)
 Main PID: 1781 (code=killed, signal=TERM)
Sep 06 20:13:49 k8s-master03 etcd[1781]: raft2021/09/06 20:13:49 INFO: d2cb62beee1ff93f became candidate at term 1247
Sep 06 20:13:49 k8s-master03 etcd[1781]: raft2021/09/06 20:13:49 INFO: d2cb62beee1ff93f received MsgVoteResp from d2cb62beee1ff93f at term 1247
Sep 06 20:13:49 k8s-master03 etcd[1781]: raft2021/09/06 20:13:49 INFO: d2cb62beee1ff93f [logterm: 1, index: 3] sent MsgVote request to 208ae89f369427bb at term 1247
Sep 06 20:13:49 k8s-master03 etcd[1781]: raft2021/09/06 20:13:49 INFO: d2cb62beee1ff93f [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 1247
Sep 06 20:13:49 k8s-master03 etcd[1781]: health check for peer 208ae89f369427bb could not connect: dial tcp 192.168.1.208:2380: i/o timeout
Sep 06 20:13:49 k8s-master03 etcd[1781]: health check for peer c05e87400fe17139 could not connect: dial tcp 192.168.1.207:2380: i/o timeout
Sep 06 20:13:49 k8s-master03 etcd[1781]: health check for peer 208ae89f369427bb could not connect: dial tcp 192.168.1.208:2380: i/o timeout
Sep 06 20:13:49 k8s-master03 etcd[1781]: health check for peer c05e87400fe17139 could not connect: dial tcp 192.168.1.207:2380: i/o timeout
Sep 06 20:13:50 k8s-master03 etcd[1781]: publish error: etcdserver: request timed out
原因分析:证书与配置出现问题,开启了高可用服务;同时etcd2|etcd3未启动,导致etcd1启动失败。
Sep 6 20:46:33 k8s-master02 Keepalived_vrrp[1036]: /etc/keepalived/check_apiserver.sh exited with status 2
Sep  6 20:46:33 k8s-master02 etcd: raft2021/09/06 20:46:33 INFO: 208ae89f369427bb is starting a new election at term 2231
Sep  6 20:46:33 k8s-master02 etcd: raft2021/09/06 20:46:33 INFO: 208ae89f369427bb became candidate at term 2232
Sep  6 20:46:33 k8s-master02 etcd: raft2021/09/06 20:46:33 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2232
Sep  6 20:46:33 k8s-master02 etcd: raft2021/09/06 20:46:33 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2232
Sep  6 20:46:33 k8s-master02 etcd: raft2021/09/06 20:46:33 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2232
Sep  6 20:46:34 k8s-master02 etcd: raft2021/09/06 20:46:34 INFO: 208ae89f369427bb is starting a new election at term 2232
Sep  6 20:46:34 k8s-master02 etcd: raft2021/09/06 20:46:34 INFO: 208ae89f369427bb became candidate at term 2233
Sep  6 20:46:34 k8s-master02 etcd: raft2021/09/06 20:46:34 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2233
Sep  6 20:46:34 k8s-master02 etcd: raft2021/09/06 20:46:34 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2233
Sep  6 20:46:34 k8s-master02 etcd: raft2021/09/06 20:46:34 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2233
Sep  6 20:46:36 k8s-master02 etcd: raft2021/09/06 20:46:36 INFO: 208ae89f369427bb is starting a new election at term 2233
Sep  6 20:46:36 k8s-master02 etcd: raft2021/09/06 20:46:36 INFO: 208ae89f369427bb became candidate at term 2234
Sep  6 20:46:36 k8s-master02 etcd: raft2021/09/06 20:46:36 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2234
Sep  6 20:46:36 k8s-master02 etcd: raft2021/09/06 20:46:36 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2234
Sep  6 20:46:36 k8s-master02 etcd: raft2021/09/06 20:46:36 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2234
Sep  6 20:46:37 k8s-master02 etcd: raft2021/09/06 20:46:37 INFO: 208ae89f369427bb is starting a new election at term 2234
Sep  6 20:46:37 k8s-master02 etcd: raft2021/09/06 20:46:37 INFO: 208ae89f369427bb became candidate at term 2235
Sep  6 20:46:37 k8s-master02 etcd: raft2021/09/06 20:46:37 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2235
Sep  6 20:46:37 k8s-master02 etcd: raft2021/09/06 20:46:37 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2235
Sep  6 20:46:37 k8s-master02 etcd: raft2021/09/06 20:46:37 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2235
Sep  6 20:46:38 k8s-master02 etcd: health check for peer c05e87400fe17139 could not connect: dial tcp 192.168.1.207:2380: connect: no route to host
Sep  6 20:46:38 k8s-master02 etcd: health check for peer c05e87400fe17139 could not connect: dial tcp 192.168.1.207:2380: connect: no route to host
Sep  6 20:46:38 k8s-master02 etcd: health check for peer d2cb62beee1ff93f could not connect: dial tcp 192.168.1.209:2380: connect: no route to host
Sep  6 20:46:38 k8s-master02 etcd: health check for peer d2cb62beee1ff93f could not connect: dial tcp 192.168.1.209:2380: connect: no route to host
Sep  6 20:46:38 k8s-master02 Keepalived_vrrp[1036]: /etc/keepalived/check_apiserver.sh exited with status 2
Sep  6 20:46:39 k8s-master02 etcd: raft2021/09/06 20:46:39 INFO: 208ae89f369427bb is starting a new election at term 2235
Sep  6 20:46:39 k8s-master02 etcd: raft2021/09/06 20:46:39 INFO: 208ae89f369427bb became candidate at term 2236
Sep  6 20:46:39 k8s-master02 etcd: raft2021/09/06 20:46:39 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2236
Sep  6 20:46:39 k8s-master02 etcd: raft2021/09/06 20:46:39 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2236
Sep  6 20:46:39 k8s-master02 etcd: raft2021/09/06 20:46:39 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2236
Sep  6 20:46:40 k8s-master02 etcd: publish error: etcdserver: request timed out
Sep  6 20:46:41 k8s-master02 etcd: raft2021/09/06 20:46:41 INFO: 208ae89f369427bb is starting a new election at term 2236
Sep  6 20:46:41 k8s-master02 etcd: raft2021/09/06 20:46:41 INFO: 208ae89f369427bb became candidate at term 2237
Sep  6 20:46:41 k8s-master02 etcd: raft2021/09/06 20:46:41 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2237
Sep  6 20:46:41 k8s-master02 etcd: raft2021/09/06 20:46:41 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2237
Sep  6 20:46:41 k8s-master02 etcd: raft2021/09/06 20:46:41 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2237
Sep  6 20:46:42 k8s-master02 etcd: raft2021/09/06 20:46:42 INFO: 208ae89f369427bb is starting a new election at term 2237
Sep  6 20:46:42 k8s-master02 etcd: raft2021/09/06 20:46:42 INFO: 208ae89f369427bb became candidate at term 2238
Sep  6 20:46:42 k8s-master02 etcd: raft2021/09/06 20:46:42 INFO: 208ae89f369427bb received MsgVoteResp from 208ae89f369427bb at term 2238
Sep  6 20:46:42 k8s-master02 etcd: raft2021/09/06 20:46:42 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to c05e87400fe17139 at term 2238
Sep  6 20:46:42 k8s-master02 etcd: raft2021/09/06 20:46:42 INFO: 208ae89f369427bb [logterm: 1, index: 3] sent MsgVote request to d2cb62beee1ff93f at term 2238
Sep  6 20:46:43 k8s-master02 etcd: health check for peer c05e87400fe17139 could not connect: dial tcp 192.168.1.207:2380: i/o timeout
Sep  6 20:46:43 k8s-master02 etcd: health check for peer c05e87400fe17139 could not connect: dial tcp 192.168.1.207:2380: i/o timeout
Sep  6 20:46:43 k8s-master02 etcd: health check for peer d2cb62beee1ff93f could not connect: dial tcp 192.168.1.209:2380: i/o timeout
Sep  6 20:46:43 k8s-master02 etcd: health check for peer d2cb62beee1ff93f could not connect: dial tcp 192.168.1.209:2380: i/o timeout
Sep  6 20:46:43 k8s-master02 Keepalived_vrrp[1036]: /etc/keepalived/check_apiserver.sh exited with status 2
Sep  6 20:46:44 k8s-master02 etcd: raft2021/09/06 20:46:44 INFO: 208ae89f369427bb is starting a new election at term 2238

Sep 6 21:22:49 k8s-master03 etcd: raft2021/09/06 21:22:49 INFO: 9eb1f144b60ad8c6 became candidate at term 64
Sep  6 21:22:49 k8s-master03 etcd: raft2021/09/06 21:22:49 INFO: 9eb1f144b60ad8c6 received MsgVoteResp from 9eb1f144b60ad8c6 at term 64
Sep  6 21:22:49 k8s-master03 etcd: raft2021/09/06 21:22:49 INFO: 9eb1f144b60ad8c6 [logterm: 1, index: 3] sent MsgVote request to 7f890ac3cab00125 at term 64
Sep  6 21:22:49 k8s-master03 etcd: raft2021/09/06 21:22:49 INFO: 9eb1f144b60ad8c6 [logterm: 1, index: 3] sent MsgVote request to ef155f75af77a2e6 at term 64
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37322" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37324" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37330" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37332" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37338" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37340" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37346" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:49 k8s-master03 etcd: rejected connection from "192.168.1.202:37348" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37354" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37356" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37364" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37362" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37370" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37372" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37380" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37382" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: health check for peer 7f890ac3cab00125 could not connect: dial tcp 192.168.1.201:2380: connect: connection refused
Sep  6 21:22:50 k8s-master03 etcd: health check for peer ef155f75af77a2e6 could not connect: x509: certificate is valid for 127.0.0.1, 192.168.1.206, 192.168.1.207, 192.168.1.208, not 192.168.1.202
Sep  6 21:22:50 k8s-master03 etcd: health check for peer 7f890ac3cab00125 could not connect: dial tcp 192.168.1.201:2380: connect: connection refused
Sep  6 21:22:50 k8s-master03 etcd: health check for peer ef155f75af77a2e6 could not connect: x509: certificate is valid for 127.0.0.1, 192.168.1.206, 192.168.1.207, 192.168.1.208, not 192.168.1.202
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37388" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 etcd: rejected connection from "192.168.1.202:37390" (error "remote error: tls: bad certificate", ServerName "")
Sep  6 21:22:50 k8s-master03 systemd: etcd.service start operation timed out. Terminating.
Sep  6 21:22:50 k8s-master03 systemd: Failed to start Etcd Service.
Sep  6 21:22:50 k8s-master03 systemd: Unit etcd.service entered failed state.
Sep  6 21:22:50 k8s-master03 systemd: etcd.service failed.
解决方法:
重新生成证书,如下图:
cfssl gencert \
    -ca=/etc/etcd/ssl/etcd-ca.pem \
    -ca-key=/etc/etcd/ssl/etcd-ca-key.pem \
    -config=ca-config.json \
    -hostname=127.0.0.1,k8s-master01,k8s-master02,k8s-master03,192.168.1.201,192.168.1.202,192.168.1.203,192.168.1.206,192.168.1.207,192.168.1.208,192.168.1.209 \
    -profile=kubernetes \
             etcd-csr.json | cfssljson -bare /etc/etcd/ssl/etcd    #ip与域名对应,都必须写;
systemctl stop etcd && rm -rf /var/lib/etcd/* && ll /var/lib/etcd/ #停止etcd服务,删除etcd运行时产生的文件;同时严禁高可用服务开启;

tail /var/log/messages
Sep 6 22:21:41 k8s-master01 systemd: Started Etcd Service.
Sep  6 22:21:41 k8s-master01 etcd: serving client requests on 127.0.0.1:2379
Sep  6 22:21:41 k8s-master01 etcd: ready to serve client requests
Sep  6 22:21:41 k8s-master01 etcd: serving client requests on 192.168.1.201:2379
Sep  6 22:21:41 k8s-master01 etcd: established a TCP streaming connection with peer ef155f75af77a2e6 (stream Message writer)
Sep  6 22:21:41 k8s-master01 etcd: established a TCP streaming connection with peer ef155f75af77a2e6 (stream MsgApp v2 writer)
Sep  6 22:21:44 k8s-master01 etcd: updated the cluster version from 3.0 to 3.4
Sep  6 22:21:44 k8s-master01 etcd: enabled capabilities for version 3.4
Sep  6 22:30:01 k8s-master01 systemd: Started Session 16 of user root.
问题27:某个node的etcd服务正常运行,但提示rpc error?
[root@k8s-master01 pki]# ETCDCTL_API=3 etcdctl member list
{"level":"warn","ts":"2021-09-06T22:58:58.455+0800","caller":"clientv3/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"endpoint://client-40bd4115-87ef-4e13-8d82-28ede2ed9ff3/127.0.0.1:2379","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = latest balancer error: all SubConns are in TransientFailure, latest connection error: connection closed"}
Error: context deadline exceeded

原因分析:127.0.0.12379远程过程调用error不影响etcd集群使用,但会降低性能。
解决方法:可不理会。

问题28:kube-apiserver无法启动?
[root@k8s-master01 ~]# systemctl enable --now kube-apiserver
Failed to execute operation: No such file or directory
原因分析:service文件放错位置。

I0907 10:13:14.991428 6008 store.go:1376] Monitoring limitranges count at <storage-prefix>//limitranges
I0907 10:13:14.997097    6008 cacher.go:405] cacher (*core.LimitRange): initialized
I0907 10:13:14.999869    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:14.999947    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.008518    6008 store.go:1376] Monitoring resourcequotas count at <storage-prefix>//resourcequotas
I0907 10:13:15.009517    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.009542    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.027691    6008 store.go:1376] Monitoring secrets count at <storage-prefix>//secrets
I0907 10:13:15.028338    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.028366    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.031897    6008 cacher.go:405] cacher (*core.ResourceQuota): initialized
I0907 10:13:15.032423    6008 cacher.go:405] cacher (*core.Secret): initialized
I0907 10:13:15.040312    6008 store.go:1376] Monitoring persistentvolumes count at <storage-prefix>//persistentvolumes
I0907 10:13:15.041293    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.041340    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.049139    6008 cacher.go:405] cacher (*core.PersistentVolume): initialized
I0907 10:13:15.049286    6008 store.go:1376] Monitoring persistentvolumeclaims count at <storage-prefix>//persistentvolumeclaims
I0907 10:13:15.050411    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.050509    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.052529    6008 cacher.go:405] cacher (*core.PersistentVolumeClaim): initialized
I0907 10:13:15.057764    6008 store.go:1376] Monitoring configmaps count at <storage-prefix>//configmaps
I0907 10:13:15.058496    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.058550    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.065377    6008 cacher.go:405] cacher (*core.ConfigMap): initialized
I0907 10:13:15.067289    6008 store.go:1376] Monitoring namespaces count at <storage-prefix>//namespaces
I0907 10:13:15.068480    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.068510    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.077229    6008 cacher.go:405] cacher (*core.Namespace): initialized
I0907 10:13:15.077465    6008 store.go:1376] Monitoring endpoints count at <storage-prefix>//services/endpoints
I0907 10:13:15.078124    6008 client.go:360] parsed scheme: "endpoint"
I0907 10:13:15.078166    6008 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 10:13:15.084988    6008 cacher.go:405] cacher (*core.Endpoints): initialized
I0907 10:13:15.087716    6008 store.go:1376] Monitoring nodes count at <storage-prefix>//minions
Error: error building core storage: open /etc/kubenetes/pki/apiserver-key.pem: no such file or directory

解决方法:修改配置文件位置即可。


问题29:kube-apiserver无法启动?

I0907 11:10:10.104401 7577 dynamic_cafile_content.go:129] Loaded a new CA Bundle and Verifier for "client-ca-bundle::/etc/kubernetes/pki/ca.pem"
I0907 11:10:10.104462    7577 dynamic_cafile_content.go:129] Loaded a new CA Bundle and Verifier for "request-header::/etc/kubernetes/pki/front-proxy-ca.pem"
I0907 11:10:10.105150    7577 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook.
I0907 11:10:10.105167    7577 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota.
I0907 11:10:10.108787    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.108840    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.118716    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.118776    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.127421    7577 store.go:1376] Monitoring customresourcedefinitions.apiextensions.k8s.io count at <storage-prefix>//apiextensions.k8s.io/customresourcedefinitions
I0907 11:10:10.128585    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.128648    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.128859    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:10:10.128920    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.201:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:10:10.129014    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:10:10.129173    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:10:10.129187    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.202:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:10:10.129191    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:10:10.129357    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:10:10.129378    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.203:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:10:10.129384    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:10:10.130065    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc000609350, {CONNECTING <nil>}
I0907 11:10:10.130106    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc0006095c0, {CONNECTING <nil>}
I0907 11:10:10.130158    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc000609810, {CONNECTING <nil>}
I0907 11:10:10.141244    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc0006095c0, {READY <nil>}
I0907 11:10:10.141914    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc000609810, {READY <nil>}
I0907 11:10:10.142132    7577 store.go:1376] Monitoring customresourcedefinitions.apiextensions.k8s.io count at <storage-prefix>//apiextensions.k8s.io/customresourcedefinitions
I0907 11:10:10.143041    7577 cacher.go:405] cacher (*apiextensions.CustomResourceDefinition): initialized
I0907 11:10:10.144261    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:10:10.145268    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc000609350, {READY <nil>}
I0907 11:10:10.145679    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:10:10.159351    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:10:10.159498    7577 cacher.go:405] cacher (*apiextensions.CustomResourceDefinition): initialized
I0907 11:10:10.173307    7577 instance.go:289] Using reconciler: lease
I0907 11:10:10.173948    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.173992    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.184871    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.184916    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.193173    7577 store.go:1376] Monitoring podtemplates count at <storage-prefix>//podtemplates
I0907 11:10:10.193829    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.193864    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.200520    7577 cacher.go:405] cacher (*core.PodTemplate): initialized
I0907 11:10:10.204003    7577 store.go:1376] Monitoring events count at <storage-prefix>//events
I0907 11:10:10.204885    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.204931    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.211519    7577 store.go:1376] Monitoring limitranges count at <storage-prefix>//limitranges
I0907 11:10:10.214437    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.214499    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.217852    7577 cacher.go:405] cacher (*core.LimitRange): initialized
I0907 11:10:10.225868    7577 store.go:1376] Monitoring resourcequotas count at <storage-prefix>//resourcequotas
I0907 11:10:10.228182    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.228286    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.232981    7577 cacher.go:405] cacher (*core.ResourceQuota): initialized
I0907 11:10:10.240524    7577 store.go:1376] Monitoring secrets count at <storage-prefix>//secrets
I0907 11:10:10.241317    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.241338    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.247825    7577 store.go:1376] Monitoring persistentvolumes count at <storage-prefix>//persistentvolumes
I0907 11:10:10.248290    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.248314    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.250441    7577 cacher.go:405] cacher (*core.Secret): initialized
I0907 11:10:10.258956    7577 cacher.go:405] cacher (*core.PersistentVolume): initialized
I0907 11:10:10.261346    7577 store.go:1376] Monitoring persistentvolumeclaims count at <storage-prefix>//persistentvolumeclaims
I0907 11:10:10.261934    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.261958    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.264535    7577 cacher.go:405] cacher (*core.PersistentVolumeClaim): initialized
I0907 11:10:10.268862    7577 store.go:1376] Monitoring configmaps count at <storage-prefix>//configmaps
I0907 11:10:10.269788    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.269837    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.273873    7577 cacher.go:405] cacher (*core.ConfigMap): initialized
I0907 11:10:10.276406    7577 store.go:1376] Monitoring namespaces count at <storage-prefix>//namespaces
I0907 11:10:10.277404    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.277433    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.282949    7577 cacher.go:405] cacher (*core.Namespace): initialized
I0907 11:10:10.286760    7577 store.go:1376] Monitoring endpoints count at <storage-prefix>//services/endpoints
I0907 11:10:10.287791    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.287876    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.292339    7577 cacher.go:405] cacher (*core.Endpoints): initialized
I0907 11:10:10.297625    7577 store.go:1376] Monitoring nodes count at <storage-prefix>//minions
I0907 11:10:10.298802    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.298853    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.309092    7577 cacher.go:405] cacher (*core.Node): initialized
I0907 11:10:10.309260    7577 store.go:1376] Monitoring pods count at <storage-prefix>//pods
I0907 11:10:10.310157    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.310201    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.318518    7577 cacher.go:405] cacher (*core.Pod): initialized
I0907 11:10:10.319040    7577 store.go:1376] Monitoring serviceaccounts count at <storage-prefix>//serviceaccounts
I0907 11:10:10.322302    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.322364    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.328407    7577 cacher.go:405] cacher (*core.ServiceAccount): initialized
I0907 11:10:10.329588    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.329617    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.341627    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.341712    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.349861    7577 store.go:1376] Monitoring replicationcontrollers count at <storage-prefix>//controllers
I0907 11:10:10.350976    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.351033    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.359374    7577 cacher.go:405] cacher (*core.ReplicationController): initialized
I0907 11:10:10.362334    7577 store.go:1376] Monitoring services count at <storage-prefix>//services/specs
I0907 11:10:10.362421    7577 rest.go:131] the default service ipfamily for this cluster is: IPv4
I0907 11:10:10.369497    7577 cacher.go:405] cacher (*core.Service): initialized
I0907 11:10:10.433239    7577 instance.go:594] Skipping disabled API group "internal.apiserver.k8s.io".
I0907 11:10:10.433337    7577 instance.go:615] Enabling API group "authentication.k8s.io".
I0907 11:10:10.433365    7577 instance.go:615] Enabling API group "authorization.k8s.io".
I0907 11:10:10.434320    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.434415    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.442103    7577 store.go:1376] Monitoring horizontalpodautoscalers.autoscaling count at <storage-prefix>//horizontalpodautoscalers
I0907 11:10:10.443432    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.443525    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.446393    7577 cacher.go:405] cacher (*autoscaling.HorizontalPodAutoscaler): initialized
I0907 11:10:10.451431    7577 store.go:1376] Monitoring horizontalpodautoscalers.autoscaling count at <storage-prefix>//horizontalpodautoscalers
I0907 11:10:10.452156    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.452192    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.455263    7577 cacher.go:405] cacher (*autoscaling.HorizontalPodAutoscaler): initialized
I0907 11:10:10.459332    7577 store.go:1376] Monitoring horizontalpodautoscalers.autoscaling count at <storage-prefix>//horizontalpodautoscalers
I0907 11:10:10.459400    7577 instance.go:615] Enabling API group "autoscaling".
I0907 11:10:10.460460    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.460512    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.463882    7577 cacher.go:405] cacher (*autoscaling.HorizontalPodAutoscaler): initialized
I0907 11:10:10.467647    7577 store.go:1376] Monitoring jobs.batch count at <storage-prefix>//jobs
I0907 11:10:10.468669    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.468755    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.476015    7577 store.go:1376] Monitoring cronjobs.batch count at <storage-prefix>//cronjobs
I0907 11:10:10.476054    7577 instance.go:615] Enabling API group "batch".
I0907 11:10:10.476619    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.476644    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.478226    7577 cacher.go:405] cacher (*batch.Job): initialized
I0907 11:10:10.486063    7577 store.go:1376] Monitoring certificatesigningrequests.certificates.k8s.io count at <storage-prefix>//certificatesigningrequests
I0907 11:10:10.486597    7577 cacher.go:405] cacher (*batch.CronJob): initialized
I0907 11:10:10.487418    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.487473    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.490073    7577 cacher.go:405] cacher (*certificates.CertificateSigningRequest): initialized
I0907 11:10:10.495265    7577 store.go:1376] Monitoring certificatesigningrequests.certificates.k8s.io count at <storage-prefix>//certificatesigningrequests
I0907 11:10:10.495400    7577 instance.go:615] Enabling API group "certificates.k8s.io".
I0907 11:10:10.495926    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.495961    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.505839    7577 cacher.go:405] cacher (*certificates.CertificateSigningRequest): initialized
I0907 11:10:10.506269    7577 store.go:1376] Monitoring leases.coordination.k8s.io count at <storage-prefix>//leases
I0907 11:10:10.506752    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.506779    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.515924    7577 cacher.go:405] cacher (*coordination.Lease): initialized
I0907 11:10:10.518790    7577 store.go:1376] Monitoring leases.coordination.k8s.io count at <storage-prefix>//leases
I0907 11:10:10.518856    7577 instance.go:615] Enabling API group "coordination.k8s.io".
I0907 11:10:10.519538    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.519581    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.529463    7577 store.go:1376] Monitoring endpointslices.discovery.k8s.io count at <storage-prefix>//endpointslices
I0907 11:10:10.529610    7577 instance.go:615] Enabling API group "discovery.k8s.io".
I0907 11:10:10.530428    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.530480    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.534466    7577 cacher.go:405] cacher (*coordination.Lease): initialized
I0907 11:10:10.540060    7577 cacher.go:405] cacher (*discovery.EndpointSlice): initialized
I0907 11:10:10.540688    7577 store.go:1376] Monitoring ingresses.networking.k8s.io count at <storage-prefix>//ingress
I0907 11:10:10.540744    7577 instance.go:615] Enabling API group "extensions".
I0907 11:10:10.543053    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.543091    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.544632    7577 cacher.go:405] cacher (*networking.Ingress): initialized
I0907 11:10:10.552559    7577 store.go:1376] Monitoring networkpolicies.networking.k8s.io count at <storage-prefix>//networkpolicies
I0907 11:10:10.553464    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.553588    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.558500    7577 cacher.go:405] cacher (*networking.NetworkPolicy): initialized
I0907 11:10:10.563785    7577 store.go:1376] Monitoring ingresses.networking.k8s.io count at <storage-prefix>//ingress
I0907 11:10:10.564430    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.564455    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.571907    7577 cacher.go:405] cacher (*networking.Ingress): initialized
I0907 11:10:10.575443    7577 store.go:1376] Monitoring ingressclasses.networking.k8s.io count at <storage-prefix>//ingressclasses
I0907 11:10:10.576185    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.576222    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.585044    7577 cacher.go:405] cacher (*networking.IngressClass): initialized
I0907 11:10:10.585767    7577 store.go:1376] Monitoring ingresses.networking.k8s.io count at <storage-prefix>//ingress
I0907 11:10:10.586920    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.586983    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.594465    7577 store.go:1376] Monitoring ingressclasses.networking.k8s.io count at <storage-prefix>//ingressclasses
I0907 11:10:10.594507    7577 instance.go:615] Enabling API group "networking.k8s.io".
I0907 11:10:10.595218    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.595254    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.595676    7577 cacher.go:405] cacher (*networking.Ingress): initialized
I0907 11:10:10.605089    7577 cacher.go:405] cacher (*networking.IngressClass): initialized
I0907 11:10:10.605463    7577 store.go:1376] Monitoring runtimeclasses.node.k8s.io count at <storage-prefix>//runtimeclasses
I0907 11:10:10.606735    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.606793    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.608653    7577 cacher.go:405] cacher (*node.RuntimeClass): initialized
I0907 11:10:10.614156    7577 store.go:1376] Monitoring runtimeclasses.node.k8s.io count at <storage-prefix>//runtimeclasses
I0907 11:10:10.614220    7577 instance.go:615] Enabling API group "node.k8s.io".
I0907 11:10:10.614835    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.614868    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.624391    7577 cacher.go:405] cacher (*node.RuntimeClass): initialized
I0907 11:10:10.628409    7577 store.go:1376] Monitoring poddisruptionbudgets.policy count at <storage-prefix>//poddisruptionbudgets
I0907 11:10:10.629370    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.629407    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.640131    7577 cacher.go:405] cacher (*policy.PodDisruptionBudget): initialized
I0907 11:10:10.640673    7577 store.go:1376] Monitoring podsecuritypolicies.policy count at <storage-prefix>//podsecuritypolicy
I0907 11:10:10.640700    7577 instance.go:615] Enabling API group "policy".
I0907 11:10:10.641229    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.641272    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.645399    7577 cacher.go:405] cacher (*policy.PodSecurityPolicy): initialized
I0907 11:10:10.655313    7577 store.go:1376] Monitoring roles.rbac.authorization.k8s.io count at <storage-prefix>//roles
I0907 11:10:10.656420    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.656452    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.665228    7577 store.go:1376] Monitoring rolebindings.rbac.authorization.k8s.io count at <storage-prefix>//rolebindings
I0907 11:10:10.665567    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.665584    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.671995    7577 cacher.go:405] cacher (*rbac.Role): initialized
I0907 11:10:10.673596    7577 cacher.go:405] cacher (*rbac.RoleBinding): initialized
I0907 11:10:10.675409    7577 store.go:1376] Monitoring clusterroles.rbac.authorization.k8s.io count at <storage-prefix>//clusterroles
I0907 11:10:10.676170    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.676204    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.681365    7577 cacher.go:405] cacher (*rbac.ClusterRole): initialized
I0907 11:10:10.685026    7577 store.go:1376] Monitoring clusterrolebindings.rbac.authorization.k8s.io count at <storage-prefix>//clusterrolebindings
I0907 11:10:10.685852    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.685915    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.694700    7577 store.go:1376] Monitoring roles.rbac.authorization.k8s.io count at <storage-prefix>//roles
I0907 11:10:10.695418    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.695452    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.697954    7577 cacher.go:405] cacher (*rbac.ClusterRoleBinding): initialized
I0907 11:10:10.704331    7577 store.go:1376] Monitoring rolebindings.rbac.authorization.k8s.io count at <storage-prefix>//rolebindings
I0907 11:10:10.705273    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.705328    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.707115    7577 cacher.go:405] cacher (*rbac.Role): initialized
I0907 11:10:10.714130    7577 store.go:1376] Monitoring clusterroles.rbac.authorization.k8s.io count at <storage-prefix>//clusterroles
I0907 11:10:10.714280    7577 cacher.go:405] cacher (*rbac.RoleBinding): initialized
I0907 11:10:10.715533    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.715568    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.722210    7577 cacher.go:405] cacher (*rbac.ClusterRole): initialized
I0907 11:10:10.724814    7577 store.go:1376] Monitoring clusterrolebindings.rbac.authorization.k8s.io count at <storage-prefix>//clusterrolebindings
I0907 11:10:10.724868    7577 instance.go:615] Enabling API group "rbac.authorization.k8s.io".
I0907 11:10:10.727455    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.727504    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.734196    7577 cacher.go:405] cacher (*rbac.ClusterRoleBinding): initialized
I0907 11:10:10.734678    7577 store.go:1376] Monitoring priorityclasses.scheduling.k8s.io count at <storage-prefix>//priorityclasses
I0907 11:10:10.735336    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.735367    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.746599    7577 cacher.go:405] cacher (*scheduling.PriorityClass): initialized
I0907 11:10:10.747241    7577 store.go:1376] Monitoring priorityclasses.scheduling.k8s.io count at <storage-prefix>//priorityclasses
I0907 11:10:10.747274    7577 instance.go:615] Enabling API group "scheduling.k8s.io".
I0907 11:10:10.748135    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.749535    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.756577    7577 store.go:1376] Monitoring storageclasses.storage.k8s.io count at <storage-prefix>//storageclasses
I0907 11:10:10.757360    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.757411    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.760090    7577 cacher.go:405] cacher (*scheduling.PriorityClass): initialized
I0907 11:10:10.761445    7577 cacher.go:405] cacher (*storage.StorageClass): initialized
I0907 11:10:10.767222    7577 store.go:1376] Monitoring volumeattachments.storage.k8s.io count at <storage-prefix>//volumeattachments
I0907 11:10:10.768289    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.768357    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.774868    7577 cacher.go:405] cacher (*storage.VolumeAttachment): initialized
I0907 11:10:10.776209    7577 store.go:1376] Monitoring csinodes.storage.k8s.io count at <storage-prefix>//csinodes
I0907 11:10:10.776692    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.776708    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.784385    7577 cacher.go:405] cacher (*storage.CSINode): initialized
I0907 11:10:10.785952    7577 store.go:1376] Monitoring csidrivers.storage.k8s.io count at <storage-prefix>//csidrivers
I0907 11:10:10.786543    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.786571    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.795441    7577 cacher.go:405] cacher (*storage.CSIDriver): initialized
I0907 11:10:10.795719    7577 store.go:1376] Monitoring storageclasses.storage.k8s.io count at <storage-prefix>//storageclasses
I0907 11:10:10.796371    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.796406    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.800397    7577 cacher.go:405] cacher (*storage.StorageClass): initialized
I0907 11:10:10.804087    7577 store.go:1376] Monitoring volumeattachments.storage.k8s.io count at <storage-prefix>//volumeattachments
I0907 11:10:10.804770    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.804813    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.814822    7577 cacher.go:405] cacher (*storage.VolumeAttachment): initialized
I0907 11:10:10.815314    7577 store.go:1376] Monitoring csinodes.storage.k8s.io count at <storage-prefix>//csinodes
I0907 11:10:10.815972    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.816002    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.823372    7577 store.go:1376] Monitoring csidrivers.storage.k8s.io count at <storage-prefix>//csidrivers
I0907 11:10:10.823449    7577 instance.go:615] Enabling API group "storage.k8s.io".
I0907 11:10:10.824268    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.824286    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.824776    7577 cacher.go:405] cacher (*storage.CSINode): initialized
I0907 11:10:10.835613    7577 cacher.go:405] cacher (*storage.CSIDriver): initialized
I0907 11:10:10.836114    7577 store.go:1376] Monitoring flowschemas.flowcontrol.apiserver.k8s.io count at <storage-prefix>//flowschemas
I0907 11:10:10.836751    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.838994    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.848549    7577 store.go:1376] Monitoring prioritylevelconfigurations.flowcontrol.apiserver.k8s.io count at <storage-prefix>//prioritylevelconfigurations
I0907 11:10:10.848594    7577 instance.go:615] Enabling API group "flowcontrol.apiserver.k8s.io".
I0907 11:10:10.849819    7577 cacher.go:405] cacher (*flowcontrol.FlowSchema): initialized
I0907 11:10:10.849904    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.849940    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.857339    7577 store.go:1376] Monitoring deployments.apps count at <storage-prefix>//deployments
I0907 11:10:10.857508    7577 cacher.go:405] cacher (*flowcontrol.PriorityLevelConfiguration): initialized
I0907 11:10:10.858030    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.858053    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.864831    7577 store.go:1376] Monitoring statefulsets.apps count at <storage-prefix>//statefulsets
I0907 11:10:10.865201    7577 cacher.go:405] cacher (*apps.Deployment): initialized
I0907 11:10:10.865754    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.865792    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.876153    7577 cacher.go:405] cacher (*apps.StatefulSet): initialized
I0907 11:10:10.876223    7577 store.go:1376] Monitoring daemonsets.apps count at <storage-prefix>//daemonsets
I0907 11:10:10.877053    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.877083    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.888313    7577 store.go:1376] Monitoring replicasets.apps count at <storage-prefix>//replicasets
I0907 11:10:10.889675    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.889731    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.893058    7577 cacher.go:405] cacher (*apps.DaemonSet): initialized
I0907 11:10:10.900221    7577 cacher.go:405] cacher (*apps.ReplicaSet): initialized
I0907 11:10:10.900235    7577 store.go:1376] Monitoring controllerrevisions.apps count at <storage-prefix>//controllerrevisions
I0907 11:10:10.900316    7577 instance.go:615] Enabling API group "apps".
I0907 11:10:10.901237    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.901307    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.905216    7577 cacher.go:405] cacher (*apps.ControllerRevision): initialized
I0907 11:10:10.910966    7577 store.go:1376] Monitoring validatingwebhookconfigurations.admissionregistration.k8s.io count at <storage-prefix>//validatingwebhookconfigurations
I0907 11:10:10.911510    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.911540    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.921253    7577 cacher.go:405] cacher (*admissionregistration.ValidatingWebhookConfiguration): initialized
I0907 11:10:10.921432    7577 store.go:1376] Monitoring mutatingwebhookconfigurations.admissionregistration.k8s.io count at <storage-prefix>//mutatingwebhookconfigurations
I0907 11:10:10.922011    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.922098    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.931306    7577 cacher.go:405] cacher (*admissionregistration.MutatingWebhookConfiguration): initialized
I0907 11:10:10.931434    7577 store.go:1376] Monitoring validatingwebhookconfigurations.admissionregistration.k8s.io count at <storage-prefix>//validatingwebhookconfigurations
I0907 11:10:10.932370    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.932417    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.938320    7577 cacher.go:405] cacher (*admissionregistration.ValidatingWebhookConfiguration): initialized
I0907 11:10:10.940181    7577 store.go:1376] Monitoring mutatingwebhookconfigurations.admissionregistration.k8s.io count at <storage-prefix>//mutatingwebhookconfigurations
I0907 11:10:10.940215    7577 instance.go:615] Enabling API group "admissionregistration.k8s.io".
I0907 11:10:10.940809    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.940853    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.944728    7577 cacher.go:405] cacher (*admissionregistration.MutatingWebhookConfiguration): initialized
I0907 11:10:10.953665    7577 store.go:1376] Monitoring events count at <storage-prefix>//events
I0907 11:10:10.956627    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:10.956676    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:10.971189    7577 store.go:1376] Monitoring events count at <storage-prefix>//events
I0907 11:10:10.971225    7577 instance.go:615] Enabling API group "events.k8s.io".
W0907 11:10:11.073103    7577 genericapiserver.go:419] Skipping API batch/v2alpha1 because it has no resources.
W0907 11:10:11.082739    7577 genericapiserver.go:419] Skipping API discovery.k8s.io/v1alpha1 because it has no resources.
W0907 11:10:11.093612    7577 genericapiserver.go:419] Skipping API node.k8s.io/v1alpha1 because it has no resources.
W0907 11:10:11.101685    7577 genericapiserver.go:419] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources.
I0907 11:10:11.104788    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:11.104939    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
W0907 11:10:11.107101    7577 genericapiserver.go:419] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources.
W0907 11:10:11.114524    7577 genericapiserver.go:419] Skipping API storage.k8s.io/v1alpha1 because it has no resources.
W0907 11:10:11.118164    7577 genericapiserver.go:419] Skipping API flowcontrol.apiserver.k8s.io/v1alpha1 because it has no resources.
W0907 11:10:11.124193    7577 genericapiserver.go:419] Skipping API apps/v1beta2 because it has no resources.
W0907 11:10:11.124281    7577 genericapiserver.go:419] Skipping API apps/v1beta1 because it has no resources.
I0907 11:10:11.142748    7577 plugins.go:158] Loaded 12 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,MutatingAdmissionWebhook.
I0907 11:10:11.142776    7577 plugins.go:161] Loaded 10 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,CertificateSubjectRestriction,ValidatingAdmissionWebhook,ResourceQuota.
I0907 11:10:11.151123    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:11.151160    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:11.163588    7577 store.go:1376] Monitoring apiservices.apiregistration.k8s.io count at <storage-prefix>//apiregistration.k8s.io/apiservices
I0907 11:10:11.164601    7577 client.go:360] parsed scheme: "endpoint"
I0907 11:10:11.164626    7577 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://192.168.1.201:2379  <nil> 0 <nil>} {https://192.168.1.202:2379  <nil> 0 <nil>} {https://192.168.1.203:2379  <nil> 0 <nil>}]
I0907 11:10:11.179263    7577 store.go:1376] Monitoring apiservices.apiregistration.k8s.io count at <storage-prefix>//apiregistration.k8s.io/apiservices
I0907 11:10:11.181001    7577 dynamic_serving_content.go:111] Loaded a new cert/key pair for "aggregator-proxy-cert::/etc/kubernetes/pki/front-proxy-client.pem::/etc/kubernetes/pki/front-proxy-client-key.pem"
I0907 11:10:11.181300    7577 cacher.go:405] cacher (*apiregistration.APIService): initialized
I0907 11:10:11.188700    7577 cacher.go:405] cacher (*apiregistration.APIService): initialized
I0907 11:10:12.892235    7577 aggregator.go:109] Building initial OpenAPI spec
I0907 11:10:13.528646    7577 aggregator.go:112] Finished initial OpenAPI spec generation after 636.366725ms
I0907 11:10:13.529062    7577 dynamic_cafile_content.go:167] Starting request-header::/etc/kubernetes/pki/front-proxy-ca.pem
I0907 11:10:13.529096    7577 dynamic_cafile_content.go:167] Starting client-ca-bundle::/etc/kubernetes/pki/ca.pem
I0907 11:10:13.529147    7577 tlsconfig.go:178] loaded client CA [0/"client-ca-bundle::/etc/kubernetes/pki/ca.pem,request-header::/etc/kubernetes/pki/front-proxy-ca.pem"]: "kubernetes" [] groups=[Kubernetes] issuer="<self>" (2021-09-06 05:53:00 +0000 UTC to 2121-08-13 05:53:00 +0000 UTC (now=2021-09-07 03:10:13.529132692 +0000 UTC))
I0907 11:10:13.529162    7577 tlsconfig.go:178] loaded client CA [1/"client-ca-bundle::/etc/kubernetes/pki/ca.pem,request-header::/etc/kubernetes/pki/front-proxy-ca.pem"]: "kubernetes" [] issuer="<self>" (2021-09-06 06:11:00 +0000 UTC to 2026-09-05 06:11:00 +0000 UTC (now=2021-09-07 03:10:13.5291582 +0000 UTC))
I0907 11:10:13.529378    7577 tlsconfig.go:200] loaded serving cert ["serving-cert::/etc/kubernetes/pki/apiserver.pem::/etc/kubernetes/pki/apiserver-key.pem"]: "kube-apiserver" [serving,client] groups=[Kubernetes] validServingFor=[127.0.0.1,10.96.0.1,192.168.0.211,192.168.1.201,192.168.1.202,192.168.1.203,192.168.1.206,192.168.1.207,192.168.1.208,kubernetes,kubernetes.default,kubernetes.default.svc,kubernetes.default.svc.cluster,kubernetes.default.svc.cluster.local] issuer="kubernetes" (2021-09-06 05:59:00 +0000 UTC to 2121-08-13 05:59:00 +0000 UTC (now=2021-09-07 03:10:13.529368866 +0000 UTC))
I0907 11:10:13.529457    7577 dynamic_serving_content.go:130] Starting serving-cert::/etc/kubernetes/pki/apiserver.pem::/etc/kubernetes/pki/apiserver-key.pem
I0907 11:10:13.529566    7577 named_certificates.go:53] loaded SNI cert [0/"self-signed loopback"]: "apiserver-loopback-client@1630984210" [serving] validServingFor=[apiserver-loopback-client] issuer="apiserver-loopback-client-ca@1630984209" (2021-09-07 02:10:09 +0000 UTC to 2022-09-07 02:10:09 +0000 UTC (now=2021-09-07 03:10:13.529559333 +0000 UTC))
I0907 11:10:13.529585    7577 secure_serving.go:197] Serving securely on [::]:6443
I0907 11:10:13.529614    7577 controller.go:83] Starting OpenAPI AggregationController
I0907 11:10:13.529627    7577 tlsconfig.go:240] Starting DynamicServingCertificateController
I0907 11:10:13.529692    7577 available_controller.go:475] Starting AvailableConditionController
I0907 11:10:13.529697    7577 cache.go:32] Waiting for caches to sync for AvailableConditionController controller
I0907 11:10:13.529709    7577 dynamic_serving_content.go:130] Starting aggregator-proxy-cert::/etc/kubernetes/pki/front-proxy-client.pem::/etc/kubernetes/pki/front-proxy-client-key.pem
I0907 11:10:13.529739    7577 apiservice_controller.go:97] Starting APIServiceRegistrationController
I0907 11:10:13.529741    7577 cache.go:32] Waiting for caches to sync for APIServiceRegistrationController controller
I0907 11:10:13.529772    7577 customresource_discovery_controller.go:209] Starting DiscoveryController
I0907 11:10:13.529799    7577 apf_controller.go:249] Starting API Priority and Fairness config controller
I0907 11:10:13.529936    7577 reflector.go:219] Starting reflector *v1.CustomResourceDefinition (5m0s) from k8s.io/apiextensions-apiserver/pkg/client/informers/externalversions/factory.go:117
I0907 11:10:13.530485    7577 controller.go:86] Starting OpenAPI controller
I0907 11:10:13.530500    7577 naming_controller.go:291] Starting NamingConditionController
I0907 11:10:13.530508    7577 establishing_controller.go:76] Starting EstablishingController
I0907 11:10:13.530519    7577 nonstructuralschema_controller.go:192] Starting NonStructuralSchemaConditionController
I0907 11:10:13.530529    7577 apiapproval_controller.go:186] Starting KubernetesAPIApprovalPolicyConformantConditionController
I0907 11:10:13.530540    7577 crd_finalizer.go:266] Starting CRDFinalizer
I0907 11:10:13.534971    7577 autoregister_controller.go:141] Starting autoregister controller
I0907 11:10:13.534989    7577 cache.go:32] Waiting for caches to sync for autoregister controller
I0907 11:10:13.535796    7577 cluster_authentication_trust_controller.go:440] Starting cluster_authentication_trust_controller controller
I0907 11:10:13.535822    7577 shared_informer.go:240] Waiting for caches to sync for cluster_authentication_trust_controller
I0907 11:10:13.536050    7577 reflector.go:219] Starting reflector *v1beta1.PriorityLevelConfiguration (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.536667    7577 reflector.go:219] Starting reflector *v1.APIService (30s) from k8s.io/kube-aggregator/pkg/client/informers/externalversions/factory.go:117
I0907 11:10:13.542294    7577 crdregistration_controller.go:111] Starting crd-autoregister controller
I0907 11:10:13.542345    7577 shared_informer.go:240] Waiting for caches to sync for crd-autoregister
I0907 11:10:13.542439    7577 dynamic_cafile_content.go:167] Starting client-ca-bundle::/etc/kubernetes/pki/ca.pem
I0907 11:10:13.542468    7577 dynamic_cafile_content.go:167] Starting request-header::/etc/kubernetes/pki/front-proxy-ca.pem
I0907 11:10:13.542661    7577 reflector.go:219] Starting reflector *v1.ConfigMap (12h0m0s) from k8s.io/kubernetes/pkg/controlplane/controller/clusterauthenticationtrust/cluster_authentication_trust_controller.go:444
I0907 11:10:13.543102    7577 reflector.go:219] Starting reflector *v1.Secret (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.543321    7577 reflector.go:219] Starting reflector *v1.Pod (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.543472    7577 reflector.go:219] Starting reflector *v1.VolumeAttachment (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.543619    7577 reflector.go:219] Starting reflector *v1.Role (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.543761    7577 reflector.go:219] Starting reflector *v1.RoleBinding (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.543925    7577 reflector.go:219] Starting reflector *v1.Endpoints (10m0s) from k8s.io/client-go/informers/factory.go:134
E0907 11:10:13.544000    7577 controller.go:152] Unable to remove old endpoints from kubernetes service: StorageError: key not found, Code: 1, Key: /registry/masterleases/192.168.1.201, ResourceVersion: 0, AdditionalErrorMsg: 
I0907 11:10:13.544079    7577 reflector.go:219] Starting reflector *v1.ServiceAccount (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.544255    7577 reflector.go:219] Starting reflector *v1.PersistentVolume (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.544458    7577 reflector.go:219] Starting reflector *v1.ClusterRole (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.544634    7577 reflector.go:219] Starting reflector *v1.Service (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.544821    7577 reflector.go:219] Starting reflector *v1.StorageClass (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.544976    7577 reflector.go:219] Starting reflector *v1.RuntimeClass (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.545182    7577 reflector.go:219] Starting reflector *v1.Node (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.545330    7577 reflector.go:219] Starting reflector *v1.ClusterRoleBinding (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.545479    7577 reflector.go:219] Starting reflector *v1.Namespace (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.545648    7577 reflector.go:219] Starting reflector *v1.MutatingWebhookConfiguration (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.545817    7577 reflector.go:219] Starting reflector *v1beta1.FlowSchema (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.545972    7577 reflector.go:219] Starting reflector *v1.LimitRange (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.546139    7577 reflector.go:219] Starting reflector *v1.PriorityClass (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.546358    7577 reflector.go:219] Starting reflector *v1.IngressClass (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.546515    7577 reflector.go:219] Starting reflector *v1.ValidatingWebhookConfiguration (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.546660    7577 reflector.go:219] Starting reflector *v1.ResourceQuota (10m0s) from k8s.io/client-go/informers/factory.go:134
I0907 11:10:13.583371    7577 healthz.go:244] poststarthook/start-apiextensions-controllers,poststarthook/crd-informer-synced,poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes,poststarthook/apiservice-registration-controller check failed: readyz
[-]poststarthook/start-apiextensions-controllers failed: not finished
[-]poststarthook/crd-informer-synced failed: not finished
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
[-]poststarthook/apiservice-registration-controller failed: not finished
I0907 11:10:13.604242    7577 shared_informer.go:247] Caches are synced for node_authorizer 
I0907 11:10:13.629834    7577 apf_controller.go:253] Running API Priority and Fairness config worker
I0907 11:10:13.629839    7577 cache.go:39] Caches are synced for AvailableConditionController controller
I0907 11:10:13.629855    7577 cache.go:39] Caches are synced for APIServiceRegistrationController controller
I0907 11:10:13.638389    7577 shared_informer.go:247] Caches are synced for cluster_authentication_trust_controller 
I0907 11:10:13.638653    7577 cache.go:39] Caches are synced for autoregister controller
I0907 11:10:13.642879    7577 shared_informer.go:247] Caches are synced for crd-autoregister 
I0907 11:10:13.688937    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:13.787157    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:13.889949    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:13.988980    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:14.087720    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:14.188777    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:14.293555    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:14.387192    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:14.487020    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles,poststarthook/scheduling/bootstrap-system-priority-classes check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0907 11:10:14.528928    7577 controller.go:132] OpenAPI AggregationController: action for item : Nothing (removed from the queue).
I0907 11:10:14.529050    7577 controller.go:132] OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue).
I0907 11:10:14.541140    7577 storage_scheduling.go:148] all system priority classes are created successfully or already exist.
I0907 11:10:14.589159    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
I0907 11:10:14.687935    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
I0907 11:10:14.787275    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
I0907 11:10:14.890847    7577 healthz.go:244] poststarthook/rbac/bootstrap-roles check failed: readyz
[-]poststarthook/rbac/bootstrap-roles failed: not finished
I0907 11:10:42.761046    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:10:42.761119    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.201:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:10:42.761132    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:10:42.761347    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00ec66420, {CONNECTING <nil>}
I0907 11:10:42.768677    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00ec66420, {READY <nil>}
I0907 11:10:42.769865    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:10:52.208425    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:10:52.208522    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.203:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:10:52.208542    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:10:52.208728    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00ed6daa0, {CONNECTING <nil>}
I0907 11:10:52.216190    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00ed6daa0, {READY <nil>}
I0907 11:10:52.217462    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:10:52.279861    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:10:52.279938    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.202:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:10:52.279952    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:10:52.280099    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00ee5cca0, {CONNECTING <nil>}
I0907 11:10:52.290254    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00ee5cca0, {READY <nil>}
I0907 11:10:52.291364    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:11:25.425875    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:11:25.426010    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.201:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:11:25.426026    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:11:25.426411    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00f0c2bd0, {CONNECTING <nil>}
I0907 11:11:25.432251    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00f0c2bd0, {READY <nil>}
I0907 11:11:25.433261    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
I0907 11:11:30.467306    7577 client.go:360] parsed scheme: "passthrough"
I0907 11:11:30.467362    7577 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{https://192.168.1.202:2379  <nil> 0 <nil>}] <nil> <nil>}
I0907 11:11:30.467371    7577 clientconn.go:948] ClientConn switching balancer to "pick_first"
I0907 11:11:30.467490    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00f18cb10, {CONNECTING <nil>}
I0907 11:11:30.477647    7577 balancer_conn_wrappers.go:78] pickfirstBalancer: HandleSubConnStateChange: 0xc00f18cb10, {READY <nil>}
I0907 11:11:30.479113    7577 controlbuf.go:508] transport: loopyWriter.run returning. connection error: desc = "transport is closing"
原因分析:I0907表示为imformation,不是报错,error: desc = "transport is closing"不会影响使用,但会影响性能;只有E开头(error)的才是报错;
解决方法:服务正常运行,暂不处理。
问题30:bootstrap.secret.yaml创建secret等资源失败?
[root@k8s-master01 bootstrap]# kubectl create -f bootstrap.secret.yaml
Unable to connect to the server: dial tcp 192.168.1.211:8443: connect: no route to host
原因分析:ip无法访问;

解决方法:修改证书的apiserver地址即可。
mkdir -p /root/.kube;cp /etc/kubernetes/admin.kubeconfig /root/.kube/config


问题31、telnet高可用VIP:8443端口,立即异常退出,RS IP需按ctrl+]才会退出端口测试?


[root@k8s-master02 ~]# telnet 192.168.1.236 8443
Trying 192.168.1.236...
Connected to 192.168.1.236.
Escape character is '^]'.
Connection closed by foreign host.
[root@k8s-master02 ~]# telnet 192.168.1.201 6443
Trying 192.168.1.201...
Connected to 192.168.1.201.
Escape character is '^]'.
原因分析:VIP:8443不提供具体服务,而是由RS:6443提供具体服务;
解决方法:无需处理。
问题32、K8S集群无法查询node资源?
[root@k8s-master01 bootstrap]# kubectl get node
No resources found
原因分析:kubelet未安装,所以无法查询node信息;
解决方法:无需处理。
问题33、kubelet启动失败?
[root@k8s-master01 kubernetes]# systemctl status kubelet.service
● kubelet.service - kubernetes kubelet
   Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled)
  Drop-In: /etc/systemd/system/kubelet.service.d
           └─kubelet.conf
   Active: activating (auto-restart) (Result: exit-code) since Wed 2021-09-08 15:47:44 CST; 435ms ago
     Docs: https://github.com/kubernetes/kubernetes
  Process: 18448 ExecStart=/usr/local/bin/kubelet (code=exited, status=255)
 Main PID: 18448 (code=exited, status=255)
Sep 08 15:47:44 k8s-master01 systemd[1]: Unit kubelet.service entered failed state.
Sep 08 15:47:44 k8s-master01 systemd[1]: kubelet.service failed.
原因分析:日志显示,配置错误。
Sep  8 15:56:56 k8s-master01 kubelet: I0908 15:56:56.003430   22603 docker_service.go:240] Hairpin mode set to "hairpin-veth"
Sep  8 15:56:56 k8s-master01 kubelet: W0908 15:56:56.003518   22603 cni.go:239] Unable to update cni config: no networks found in /etc/cni/net.d
Sep  8 15:56:56 k8s-master01 kubelet: I0908 15:56:56.009817   22603 docker_service.go:255] Docker cri networking managed by kubernetes.io/no-op
Sep  8 15:56:56 k8s-master01 kubelet: I0908 15:56:56.021591   22603 docker_service.go:260] Docker Info: &{ID:BKPM:2C3W:J2KC:TFPT:DS5J:ANQ3:UVLF:KYLC:6WL6:32YK:3HTR:ZAWJ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:0 Driver:overlay2 DriverStatus:[[Backing Filesystem xfs] [Supports d_type true] [Native Overlay Diff true]] SystemStatus:[] Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:[] Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:true KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6tables:true Debug:false NFd:22 OomKillDisable:true NGoroutines:35 SystemTime:2021-09-08T15:56:56.011370377+08:00 LoggingDriver:json-file CgroupDriver:systemd NEventsListener:0 KernelVersion:4.19.12-1.el7.elrepo.x86_64 OperatingSystem:CentOS Linux 7 (Core) OSType:linux Architecture:x86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:0xc000972380 NCPU:2 MemTotal:4111605760 GenericResources:[] DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:k8s-master01 Labels:[] ExperimentalBuild:false ServerVersion:19.03.11 ClusterStore: ClusterAdvertise: Runtimes:map[runc:{Path:runc Args:[]}] DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:[] Nodes:0 Managers:0 Cluster:<nil> Warnings:[]} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:e25210fe30a0a703442421b0f60afac609f950a3 Expected:e25210fe30a0a703442421b0f60afac609f950a3} RuncCommit:{ID:v1.0.1-0-g4144b63 Expected:v1.0.1-0-g4144b63} InitCommit:{ID:fec3683 Expected:fec3683} SecurityOptions:[name=seccomp,profile=default] ProductLicense: Warnings:[]}
Sep  8 15:56:56 k8s-master01 kubelet: F0908 15:56:56.021746   22603 server.go:269] failed to run Kubelet: misconfiguration: kubelet cgroup driver: "cgroupfs" is different from docker cgroup driver: "systemd"
Sep  8 15:56:56 k8s-master01 kubelet: goroutine 1 [running]:
Sep  8 15:56:56 k8s-master01 kubelet: k8s.io/kubernetes/vendor/k8s.io/klog/v2.stacks(0xc000010001, 0xc00000bf00, 0xaa, 0xfc)
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:1026 +0xb9
Sep  8 15:56:56 k8s-master01 kubelet: k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).output(0x70c7460, 0xc000000003, 0x0, 0x0, 0xc0009723f0, 0x6f32a12, 0x9, 0x10d, 0x411b00)
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:975 +0x19b
Sep  8 15:56:56 k8s-master01 kubelet: k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).printDepth(0x70c7460, 0xc000000003, 0x0, 0x0, 0x0, 0x0, 0x1, 0xc00092f2a0, 0x1, 0x1)
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:732 +0x16f
Sep  8 15:56:56 k8s-master01 kubelet: k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).print(...)
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:714
Sep  8 15:56:56 k8s-master01 kubelet: k8s.io/kubernetes/vendor/k8s.io/klog/v2.Fatal(...)
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:1482
...
Sep  8 15:56:56 k8s-master01 systemd: kubelet.service: main process exited, code=exited, status=255/n/a
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/net.go:182 +0x8e
Sep  8 15:56:56 k8s-master01 kubelet: net/http.(*persistConn).Read(0xc0007f8fc0, 0xc000873000, 0x1000, 0x1000, 0xc00009c420, 0xc00007dc58, 0x409095)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:1887 +0x77
Sep  8 15:56:56 k8s-master01 kubelet: bufio.(*Reader).fill(0xc0007e9da0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/bufio/bufio.go:101 +0x105
Sep  8 15:56:56 k8s-master01 kubelet: bufio.(*Reader).Peek(0xc0007e9da0, 0x1, 0x0, 0x0, 0x1, 0x0, 0xc000142720)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/bufio/bufio.go:139 +0x4f
Sep  8 15:56:56 k8s-master01 kubelet: net/http.(*persistConn).readLoop(0xc0007f8fc0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:2040 +0x1a8
Sep  8 15:56:56 k8s-master01 kubelet: created by net/http.(*Transport).dialConn
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:1708 +0xcb7
Sep  8 15:56:56 k8s-master01 kubelet: goroutine 104 [select]:
Sep  8 15:56:56 k8s-master01 kubelet: net/http.(*persistConn).writeLoop(0xc0007f8fc0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:2340 +0x11c
Sep  8 15:56:56 k8s-master01 kubelet: created by net/http.(*Transport).dialConn
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:1709 +0xcdc
Sep  8 15:56:56 k8s-master01 kubelet: goroutine 106 [chan receive]:
Sep  8 15:56:56 k8s-master01 kubelet: k8s.io/kubernetes/vendor/k8s.io/client-go/tools/record.(*eventBroadcasterImpl).StartEventWatcher.func1(0x4f1b800, 0xc00090ea20, 0xc0003f8b70)
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/client-go/tools/record/event.go:301 +0xaa
Sep  8 15:56:56 k8s-master01 kubelet: created by k8s.io/kubernetes/vendor/k8s.io/client-go/tools/record.(*eventBroadcasterImpl).StartEventWatcher
Sep  8 15:56:56 k8s-master01 kubelet: /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/client-go/tools/record/event.go:299 +0x6e
Sep  8 15:56:56 k8s-master01 kubelet: goroutine 108 [IO wait]:
Sep  8 15:56:56 k8s-master01 kubelet: internal/poll.runtime_pollWait(0x7f8f9916cd58, 0x72, 0x4f0ee60)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/runtime/netpoll.go:222 +0x55
Sep  8 15:56:56 k8s-master01 kubelet: internal/poll.(*pollDesc).wait(0xc00087c498, 0x72, 0x4f0ee00, 0x6fd5e28, 0x0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/internal/poll/fd_poll_runtime.go:87 +0x45
Sep  8 15:56:56 k8s-master01 kubelet: internal/poll.(*pollDesc).waitRead(...)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/internal/poll/fd_poll_runtime.go:92
Sep  8 15:56:56 k8s-master01 kubelet: internal/poll.(*FD).Read(0xc00087c480, 0xc000705000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/internal/poll/fd_unix.go:159 +0x1a5
Sep  8 15:56:56 k8s-master01 kubelet: net.(*netFD).Read(0xc00087c480, 0xc000705000, 0x1000, 0x1000, 0x43de5c, 0xc000079b58, 0x46b2e0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/fd_posix.go:55 +0x4f
Sep  8 15:56:56 k8s-master01 systemd: Unit kubelet.service entered failed state.
Sep  8 15:56:56 k8s-master01 kubelet: net.(*conn).Read(0xc00014e248, 0xc000705000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/net.go:182 +0x8e
Sep  8 15:56:56 k8s-master01 kubelet: net/http.(*persistConn).Read(0xc000a44c60, 0xc000705000, 0x1000, 0x1000, 0xc00009cd20, 0xc000079c58, 0x409095)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:1887 +0x77
Sep  8 15:56:56 k8s-master01 kubelet: bufio.(*Reader).fill(0xc000019440)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/bufio/bufio.go:101 +0x105
Sep  8 15:56:56 k8s-master01 kubelet: bufio.(*Reader).Peek(0xc000019440, 0x1, 0x0, 0x0, 0x1, 0x0, 0xc000143380)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/bufio/bufio.go:139 +0x4f
Sep  8 15:56:56 k8s-master01 kubelet: net/http.(*persistConn).readLoop(0xc000a44c60)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:2040 +0x1a8
Sep  8 15:56:56 k8s-master01 kubelet: created by net/http.(*Transport).dialConn
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:1708 +0xcb7
Sep  8 15:56:56 k8s-master01 kubelet: goroutine 109 [select]:
Sep  8 15:56:56 k8s-master01 kubelet: net/http.(*persistConn).writeLoop(0xc000a44c60)
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:2340 +0x11c
Sep  8 15:56:56 k8s-master01 kubelet: created by net/http.(*Transport).dialConn
Sep  8 15:56:56 k8s-master01 kubelet: /usr/local/go/src/net/http/transport.go:1709 +0xcdc
Sep  8 15:56:56 k8s-master01 systemd: kubelet.service failed.

解决方法:修改配置即可。
cat <<eof> /etc/docker/daemon.json 
{
    "exec-opts":["native.cgroupdriver=cgroupfs"],
    "log-driver":"json-file",
    "log-opts":{
        "max-size":"100m"
    }
}
eof
systemctl restart docker
systemctl status docker


问题34、calico-etcd部署失败?
[root@k8s-master01 calico]# kubectl apply -f calico-etcd.yaml
secret/calico-etcd-secrets created
error: error parsing calico-etcd.yaml: error converting YAML to JSON: yaml: line 12: did not find expected key
原因分析:查看12行,没有错误。实际错误发生在18行。

 
解决方法:修改配置即可。

问题34、创建pod失败?
[root@k8s-master01 ~]# cat myapp.yaml
apiVersion: v1          
kind: Pod               
metadata:               
    name: myapp-pod
    labels:
        app: myapp
spec:                   
    containers:
    - name: myapp-container     
      image: busybox            
      command: ['sh','-c','echo The app is running! && sleep 3600']
[root@k8s-master01 ~]# 
[root@k8s-master01 ~]# kubectl apply -f myapp.yaml
Error from server (Forbidden): error when creating "myapp.yaml": pods "myapp-pod" is forbidden: error looking up service account default/default: serviceaccount "default" not found
原因分析:没有default的SA;

解决方法:创建default/default的sa即可。
问题35、创建pod失败?
[root@k8s-master01 ~]# kubectl apply -f myapp.yaml
Error from server (ServerTimeout): error when creating "myapp.yaml": No API token found for service account "default", retry after the token is automatically created and added to the service account
原因分析:SA没有API token;
[root@k8s-master01 ~]# kubectl get sa
NAME      SECRETS   AGE
default   0         4s
解决方法:待处理;
问题36、kubelet.service启动出现异常?
[root@k8s-master01 ~]# systemctl status kubelet.service
● kubelet.service - kubernetes kubelet
   Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled)
  Drop-In: /etc/systemd/system/kubelet.service.d
           └─10-kubelet.conf
   Active: active (running) since Fri 2021-09-10 10:08:26 CST; 19min ago
     Docs: https://github.com/kubernetes/kubernetes
 Main PID: 21876 (kubelet)
   CGroup: /system.slice/kubelet.service
           └─21876 /usr/local/bin/kubelet
Sep 10 10:08:27 k8s-master01 kubelet[21876]: I0910 10:08:27.353822   21876 policy_none.go:43] [cpumanager] none policy: Start
Sep 10 10:08:27 k8s-master01 kubelet[21876]: W0910 10:08:27.354453   21876 manager.go:594] Failed to retrieve checkpoint for "kubelet_internal_checkpoint": checkpoint is not found
Sep 10 10:08:27 k8s-master01 kubelet[21876]: I0910 10:08:27.354919   21876 plugin_manager.go:114] Starting Kubelet Plugin Manager
Sep 10 10:08:27 k8s-master01 kubelet[21876]: I0910 10:08:27.444279   21876 reconciler.go:157] Reconciler: start to sync state
Sep 10 10:10:21 k8s-master01 systemd[1]: [/etc/systemd/system/kubelet.service.d/10-kubelet.conf:5] Trailing garbage, ignoring.
Sep 10 10:10:21 k8s-master01 systemd[1]: [/etc/systemd/system/kubelet.service.d/10-kubelet.conf:5] Trailing garbage, ignoring.
Sep 10 10:10:51 k8s-master01 systemd[1]: [/etc/systemd/system/kubelet.service.d/10-kubelet.conf:5] Trailing garbage, ignoring.
Sep 10 10:10:51 k8s-master01 systemd[1]: [/etc/systemd/system/kubelet.service.d/10-kubelet.conf:5] Trailing garbage, ignoring.
Sep 10 10:10:51 k8s-master01 systemd[1]: [/etc/systemd/system/kubelet.service.d/10-kubelet.conf:5] Trailing garbage, ignoring.
Sep 10 10:10:51 k8s-master01 systemd[1]: [/etc/systemd/system/kubelet.service.d/10-kubelet.conf:5] Trailing garbage, ignoring.
[root@k8s-master01 ~]# kubectl get nodes
No resources found
[root@k8s-master01 ~]# 
原因分析:配置产生垃圾;
解决方法:清除垃圾即可;
[root@k8s-master01 ~]# systemctl unmask kubelet.service
[root@k8s-master01 ~]# 
[root@k8s-master01 ~]# systemctl restart kubelet.service
systemctl status kubelet.service
[root@k8s-master01 ~]# systemctl status kubelet.service
● kubelet.service - kubernetes kubelet
   Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled)
  Drop-In: /etc/systemd/system/kubelet.service.d
           └─10-kubelet.conf
   Active: active (running) since Fri 2021-09-10 10:31:42 CST; 31ms ago
     Docs: https://github.com/kubernetes/kubernetes
 Main PID: 27518 (kubelet)
    Tasks: 6
   Memory: 7.1M
   CGroup: /system.slice/kubelet.service
           └─27518 /usr/local/bin/kubelet
Sep 10 10:31:42 k8s-master01 systemd[1]: Started kubernetes kubelet.
问题37、kubelet.service启动出现异常?
[root@k8s-master02 ~]# kubeadm config images pull --config /root/new-kubeadm-config.yaml
[config/images] Pulled registry.aliyuncs.com/google_containers/kube-apiserver:v1.22.0
[config/images] Pulled registry.aliyuncs.com/google_containers/kube-controller-manager:v1.22.0
[config/images] Pulled registry.aliyuncs.com/google_containers/kube-scheduler:v1.22.0
[config/images] Pulled registry.aliyuncs.com/google_containers/kube-proxy:v1.22.0
[config/images] Pulled registry.aliyuncs.com/google_containers/pause:3.5
[config/images] Pulled registry.aliyuncs.com/google_containers/etcd:3.5.0-0
failed to pull image "registry.aliyuncs.com/google_containers/coredns:v1.8.4": output: Error response from daemon: manifest for registry.aliyuncs.com/google_containers/coredns:v1.8.4 not found: manifest unknown: manifest unknown
, error: exit status 1
To see the stack trace of this error execute with --v=5 or higher
原因分析:源站镜像tag格式不同。
解决方法:使用docker拉取镜像就行;
docker pull registry.aliyuncs.com/google_containers/coredns:1.8.4

 
                    
                     
                    
                 
                    
                 
                
            
         
         浙公网安备 33010602011771号
浙公网安备 33010602011771号