文章是按https://blog.csdn.net/Excairun/article/details/88962769,来进行操作并记录相关结果

版本:k8s V14.0,docker-ce 18.09.03

1.环境准备

内核及版本

[root@k8s-node1 ~]# cat /etc/redhat-release
CentOS Linux release 7.6.1810 (Core) 
[root@k8s-node1 ~]# uname -r
3.10.0-957.el7.x86_64
[root@k8s-node1 ~]# uname -a
Linux k8s-node1 3.10.0-957.el7.x86_64 #1 SMP Thu Nov 8 23:39:32 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
[root@k8s-node1 ~]# 

 

修改主机名配置hosts

[root@localhost ~]# hostname
localhost.localdomain
[root@localhost ~]# hostname k8s-master ##临时生效
localhost.localdomain
[root@localhost ~]# hostnamectl set-hostname k8s-master ##重启后永久生效
[root@localhost ~]# reboot

 

[root@localhost ~]# cat /etc/hosts
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1         localhost localhost.localdomain localhost6 localhost6.localdomain6
128.192.111.130 k8s-master
128.192.111.131 k8s-node1
[root@localhost ~]# 

同步时间

yum install -y ntpdate
ntpdate -u ntp.api.bz

 


关闭防火墙,selinux,swap,桥接网络配置

# 所有主机:基本系统配置
 
# 关闭Selinux/firewalld
systemctl stop firewalld
systemctl disable firewalld
setenforce 0
sed -i "s/SELINUX=enforcing/SELINUX=disabled/g" /etc/selinux/config
 
# 关闭交换分区
swapoff -a
yes | cp /etc/fstab /etc/fstab_bak
cat /etc/fstab_bak |grep -v swap > /etc/fstab
 
# 设置网桥包经IPTables,core文件生成路径

modprobe br_netfilter
cat >/etc/sysctl.conf <<EOF
vm.swappiness = 0
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
EOF

sysctl -p

2.docker安装

#安装yum操作的基本服务,如果已安装可跳过此步骤
yum install -y net-tools epel-release 
yum install -y vim  yum-utils device-mapper-persistent-data lvm2
#添加阿里云的docker-ce源
yum-config-manager --add-repo http://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo
#查看docker可安装版本
yum list docker-ce.x86_64  --showduplicates |sort -r
#安装docker如果不指定版本则会安装最新可安装版本
yum install docker-ce-18.09.3-3.el7
#设置开机启动 systemctl enable docker #启动服务 systemctl start docker


3.kubernetes相关配置

配置k8s走阿里镜像

[root@localhost ~]# cat /etc/yum.repos.d/kubernetes.repo
[kubernetes]
name=Kubernetes
baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64/
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg https://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg

安装kubectl,kubelet,kubernetes-cni,kubeadm

yum install -y kubectl-1.14.0 kubelet-1.14.0 kubernetes-cni-1.14.0 kubeadm-1.14.0
systemctl enable kubelet #开机启动


设置docker Cgroup Driver为systemd,达到与kubelet相一致的要求

cat > /etc/docker/daemon.json <<EOF
{
  "exec-opts": ["native.cgroupdriver=systemd"],
  "log-driver": "json-file",
  "log-opts": {
    "max-size": "100m"
  },
  "storage-driver": "overlay2",
  "storage-opts": [
    "overlay2.override_kernel_check=true"
  ]
}
EOF

 

重启docker

[root@localhost ~]# mkdir -p /etc/systemd/system/docker.service.d
[root@localhost ~]# 
[root@localhost ~]# systemctl daemon-reload
[root@localhost ~]# systemctl restart docker

查看kubeadm v1.14.0要求的版本

[root@localhost ~]# kubeadm config images list
I0402 13:55:29.358043    8471 version.go:96] could not fetch a Kubernetes version from the internet: unable to get URL "https://dl.k8s.io/release/stable-1.txt": Get https://dl.k8s.io/release/stable-1.txt: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
I0402 13:55:29.358209    8471 version.go:97] falling back to the local client version: v1.14.0
k8s.gcr.io/kube-apiserver:v1.14.0
k8s.gcr.io/kube-controller-manager:v1.14.0
k8s.gcr.io/kube-scheduler:v1.14.0
k8s.gcr.io/kube-proxy:v1.14.0
k8s.gcr.io/pause:3.1
k8s.gcr.io/etcd:3.3.10
k8s.gcr.io/coredns:1.3.1

 


拉取国内镜像

[root@localhost ~]# cat pullK8sImages.sh 
#!/bin/bash
KUBE_VERSION=v1.14.0
KUBE_PAUSE_VERSION=3.1
ETCD_VERSION=3.3.10
DNS_VERSION=1.3.1
DASHBOARD_VERSION=v1.10.1
username=registry.cn-hangzhou.aliyuncs.com/google_containers
#多了一个dashboard组件是为后续安装做准备
images=(
        kube-proxy-amd64:${KUBE_VERSION}
        kube-scheduler-amd64:${KUBE_VERSION}
        kube-controller-manager-amd64:${KUBE_VERSION}
        kube-apiserver-amd64:${KUBE_VERSION}
        pause:${KUBE_PAUSE_VERSION}
        etcd-amd64:${ETCD_VERSION}
        coredns:${DNS_VERSION}
        kubernetes-dashboard-amd64:DASHBOARD_VERSION 
        )
for image in ${images[@]}
do
NEW_IMAGE=`echo ${image}|awk '{gsub(/-amd64/,"",$0);print}'`
echo ${NEW_IMAGE}
docker pull ${username}/${image}
docker tag ${username}/${image} k8s.gcr.io/${NEW_IMAGE}
docker rmi ${username}/${image}
done
[root@localhost ~]# sh pullK8sImages.sh 


拉取结果

[root@localhost ~]# docker images
REPOSITORY                           TAG                 IMAGE ID            CREATED             SIZE
k8s.gcr.io/kube-proxy                v1.14.0             5cd54e388aba        7 days ago          82.1MB
k8s.gcr.io/kube-scheduler            v1.14.0             00638a24688b        7 days ago          81.6MB
k8s.gcr.io/kube-controller-manager   v1.14.0             b95b1efa0436        7 days ago          158MB
k8s.gcr.io/kube-apiserver            v1.14.0             ecf910f40d6e        7 days ago          210MB
k8s.gcr.io/coredns                   1.3.1               eb516548c180        2 months ago        40.3MB
k8s.gcr.io/etcd                      3.3.10              2c4adeb21b4f        4 months ago        258MB
k8s.gcr.io/pause                     3.1                 da86e6ba6ca1        15 months ago       742kB
[root@localhost ~]# 

 4.kubeadm(仅master执行)安装

kubeadm init   --kubernetes-version=v1.14.0   --pod-network-cidr=10.244.0.0/16  --apiserver-advertise-address=192.168.111.130  --token-ttl 0  --ignore-preflight-errors=Swap

安装结果

To start using your cluster, you need to run the following as a regular user:

  mkdir -p $HOME/.kube
  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
  sudo chown $(id -u):$(id -g) $HOME/.kube/config

You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
  https://kubernetes.io/docs/concepts/cluster-administration/addons/

Then you can join any number of worker nodes by running the following on each as root:

kubeadm join 192.168.111.130:6443 --token ymtl8s.933t59qfezi9gjcq \
    --discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7 

出现最后一行,表示安装成功,若报错,可执行

kubeadm reset

rm -rf $HOME/.kube/config

调整后,再执行kubeadm init
为了在点上使用kubectl,需作以下设置

[root@localhost ~]# kubectl version
Client Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.0", GitCommit:"641856db18352033a0d96dbc99153fa3b27298e5", GitTreeState:"clean", BuildDate:"2019-03-25T15:53:57Z", GoVersion:"go1.12.1", Compiler:"gc", Platform:"linux/amd64"}
The connection to the server localhost:8080 was refused - did you specify the right host or port?
[root@localhost ~]# hostname
k8s-master
[root@localhost ~]# mkdir -p $HOME/.kube
[root@localhost ~]# cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
[root@localhost ~]# chown $(id -u):$(id -g) $HOME/.kube/config
[root@localhost ~]# kubectl version
Client Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.0", GitCommit:"641856db18352033a0d96dbc99153fa3b27298e5", GitTreeState:"clean", BuildDate:"2019-03-25T15:53:57Z", GoVersion:"go1.12.1", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.0", GitCommit:"641856db18352033a0d96dbc99153fa3b27298e5", GitTreeState:"clean", BuildDate:"2019-03-25T15:45:25Z", GoVersion:"go1.12.1", Compiler:"gc", Platform:"linux/amd64"}
[root@localhost ~]# 

[root@localhost ~]# kubectl cluster-info Kubernetes master is running at https://192.168.111.130:6443 KubeDNS is running at https://192.168.111.130:6443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy

To further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.

 5.配置flannel网络(仅master节点执行)

不成功可多次尝试

[root@localhost ~]# yum -y install wget
[root@localhost ~]# wget https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml
[root@localhost ~]# kubectl  apply -f kube-flannel.yml

 6.添加子节点(节点机器上操作)

修改节点机器名称 linux centos7.5修改主机名和ip永久生效

##临时生效,重启失效
[root@localhost ~]# hostname
localhost.localdomain

[root@localhost ~]#hostname
k8s-node1

##永久有效,重启后生效
[root@localhost ~]#hostnamectl set-hostname k8s-node1 [root@localhost ~]# reboot

 

 

kubeadm join 192.168.111.130:6443 --token ymtl8s.933t59qfezi9gjcq \
    --discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7 

执行结果

[root@localhost ~]# kubeadm join 192.168.111.130:6443 --token ymtl8s.933t59qfezi9gjcq     --discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7 
[preflight] Running pre-flight checks
[preflight] Reading configuration from the cluster...
[preflight] FYI: You can look at this config file with 'kubectl -n kube-system get cm kubeadm-config -oyaml'
[kubelet-start] Downloading configuration for the kubelet from the "kubelet-config-1.14" ConfigMap in the kube-system namespace
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Activating the kubelet service
[kubelet-start] Waiting for the kubelet to perform the TLS Bootstrap...

This node has joined the cluster:
* Certificate signing request was sent to apiserver and a response was received.
* The Kubelet was informed of the new secure connection details.

Run 'kubectl get nodes' on the control-plane to see this node join the cluster.

[root@localhost ~]# 

报错可执行对应操作,后再执行kubeadm

[root@localhost ~]# kubeadm join 192.168.111.130:6443 --token ymtl8s.933t59qfezi9gjcq \
>     --discovery-token-ca-cert-hash sha256:7816d0b2572e6c569ed8e63ece15a7a08d06ed3fc89698245bf2aaa6acc345d7 
[preflight] Running pre-flight checks
error execution phase preflight: [preflight] Some fatal errors occurred:
        [ERROR FileContent--proc-sys-net-ipv4-ip_forward]: /proc/sys/net/ipv4/ip_forward contents are not set to 1
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
[root@localhost ~]# echo 1 > /proc/sys/net/ipv4/ip_forward
[root@localhost ~]# cat /proc/sys/net/ipv4/ip_forward
1

 (master)查看要点信息

[root@localhost ~]# kubectl get nodes
NAME         STATUS   ROLES    AGE     VERSION
k8s-master   Ready    master   40m     v1.14.0
k8s-node1    Ready    <none>   4m31s   v1.14.0

查看节点详情

[root@localhost ~]# kubectl describe nodes k8s-node1
Name:               k8s-node1
Roles:              <none>
Labels:             beta.kubernetes.io/arch=amd64
                    beta.kubernetes.io/os=linux
                    kubernetes.io/arch=amd64
                    kubernetes.io/hostname=k8s-node1
                    kubernetes.io/os=linux
Annotations:        flannel.alpha.coreos.com/backend-data: {"VtepMAC":"ca:bc:84:14:09:94"}
                    flannel.alpha.coreos.com/backend-type: vxlan
                    flannel.alpha.coreos.com/kube-subnet-manager: true
                    flannel.alpha.coreos.com/public-ip: 192.168.111.131
                    kubeadm.alpha.kubernetes.io/cri-socket: /var/run/dockershim.sock
                    node.alpha.kubernetes.io/ttl: 0
                    volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp:  Tue, 02 Apr 2019 14:47:33 +0800
Taints:             <none>
Unschedulable:      false
Conditions:
  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message
  ----             ------  -----------------                 ------------------                ------                       -------
  MemoryPressure   False   Tue, 02 Apr 2019 14:54:34 +0800   Tue, 02 Apr 2019 14:47:33 +0800   KubeletHasSufficientMemory   kubelet has sufficient memory available
  DiskPressure     False   Tue, 02 Apr 2019 14:54:34 +0800   Tue, 02 Apr 2019 14:47:33 +0800   KubeletHasNoDiskPressure     kubelet has no disk pressure
  PIDPressure      False   Tue, 02 Apr 2019 14:54:34 +0800   Tue, 02 Apr 2019 14:47:33 +0800   KubeletHasSufficientPID      kubelet has sufficient PID available
  Ready            True    Tue, 02 Apr 2019 14:54:34 +0800   Tue, 02 Apr 2019 14:48:24 +0800   KubeletReady                 kubelet is posting ready status
Addresses:
  InternalIP:  192.168.111.131
  Hostname:    k8s-node1
Capacity:
 cpu:                1
 ephemeral-storage:  17394Mi
 hugepages-1Gi:      0
 hugepages-2Mi:      0
 memory:             995896Ki
 pods:               110
Allocatable:
 cpu:                1
 ephemeral-storage:  16415037823
 hugepages-1Gi:      0
 hugepages-2Mi:      0
 memory:             893496Ki
 pods:               110
System Info:
 Machine ID:                 a5a43f5916c643bf83d6f99425a4b9d2
 System UUID:                FCCE4D56-202D-568C-8377-7A69D9ADF401
 Boot ID:                    68e3c38d-d1d6-4197-9af5-01a9699ce00c
 Kernel Version:             3.10.0-957.el7.x86_64
 OS Image:                   CentOS Linux 7 (Core)
 Operating System:           linux
 Architecture:               amd64
 Container Runtime Version:  docker://18.9.3
 Kubelet Version:            v1.14.0
 Kube-Proxy Version:         v1.14.0
PodCIDR:                     10.244.1.0/24
Non-terminated Pods:         (2 in total)
  Namespace                  Name                           CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE
  ---------                  ----                           ------------  ----------  ---------------  -------------  ---
  kube-system                kube-flannel-ds-amd64-bsrkx    100m (10%)    100m (10%)  50Mi (5%)        50Mi (5%)      7m19s
  kube-system                kube-proxy-2mj4q               0 (0%)        0 (0%)      0 (0%)           0 (0%)         7m19s
Allocated resources:
  (Total limits may be over 100 percent, i.e., overcommitted.)
  Resource           Requests    Limits
  --------           --------    ------
  cpu                100m (10%)  100m (10%)
  memory             50Mi (5%)   50Mi (5%)
  ephemeral-storage  0 (0%)      0 (0%)
Events:
  Type    Reason                   Age                    From                   Message
  ----    ------                   ----                   ----                   -------
  Normal  Starting                 7m20s                  kubelet, k8s-node1     Starting kubelet.
  Normal  NodeHasSufficientMemory  7m20s (x2 over 7m20s)  kubelet, k8s-node1     Node k8s-node1 status is now: NodeHasSufficientMemory
  Normal  NodeHasNoDiskPressure    7m20s (x2 over 7m20s)  kubelet, k8s-node1     Node k8s-node1 status is now: NodeHasNoDiskPressure
  Normal  NodeHasSufficientPID     7m20s (x2 over 7m20s)  kubelet, k8s-node1     Node k8s-node1 status is now: NodeHasSufficientPID
  Normal  NodeAllocatableEnforced  7m19s                  kubelet, k8s-node1     Updated Node Allocatable limit across pods
  Normal  Starting                 7m16s                  kube-proxy, k8s-node1  Starting kube-proxy.
  Normal  NodeReady                6m29s                  kubelet, k8s-node1     Node k8s-node1 status is now: NodeReady
[root@localhost ~]# 

kubectl查看系统pod信息

[root@k8s-master ~]# kubectl get namespaces ##get 可查看services,deployments,pods,replicasets
NAME              STATUS   AGE
default           Active   3h17m
kube-node-lease   Active   3h17m
kube-public       Active   3h17m
kube-system       Active   3h17m

 

root@k8s-master ~]# kubectl get po --all-namespaces -o wide ##-o查看pod运行在那个node上,--namespace=kube-system查看系统pod
NAMESPACE     NAME                                 READY   STATUS    RESTARTS   AGE     IP                NODE         NOMINATED NODE   READINESS GATES
kube-system   coredns-fb8b8dccf-ltlx4              1/1     Running   0          3h12m   10.244.0.4        k8s-master   <none>           <none>
kube-system   coredns-fb8b8dccf-q949f              1/1     Running   0          3h12m   10.244.0.5        k8s-master   <none>           <none>
kube-system   etcd-k8s-master                      1/1     Running   2          3h11m   192.168.111.130   k8s-master   <none>           <none>
kube-system   kube-apiserver-k8s-master            1/1     Running   2          3h12m   192.168.111.130   k8s-master   <none>           <none>
kube-system   kube-controller-manager-k8s-master   1/1     Running   2          3h11m   192.168.111.130   k8s-master   <none>           <none>
kube-system   kube-flannel-ds-amd64-2gr2v          1/1     Running   0          177m    192.168.111.130   k8s-master   <none>           <none>
kube-system   kube-flannel-ds-amd64-bsrkx          1/1     Running   0          157m    192.168.111.131   k8s-node1    <none>           <none>
kube-system   kube-flannel-ds-amd64-xdg5p          1/1     Running   0          15m     192.168.111.132   k8s-node2    <none>           <none>
kube-system   kube-proxy-2mj4q                     1/1     Running   0          157m    192.168.111.131   k8s-node1    <none>           <none>
kube-system   kube-proxy-ffd8s                     1/1     Running   0          15m     192.168.111.132   k8s-node2    <none>           <none>
kube-system   kube-proxy-qp5k7                     1/1     Running   0          3h12m   192.168.111.130   k8s-master   <none>           <none>
kube-system   kube-scheduler-k8s-master            1/1     Running   2          3h12m   192.168.111.130   k8s-master   <none>           <none>
[root@k8s-master ~]#

 docker ps -a查看container

[root@k8s-master docker]# docker ps -a
CONTAINER ID        IMAGE                  COMMAND                  CREATED             STATUS                       PORTS                     NAMES
c8a00a7aab85        tomcat:8               "catalina.sh run"        22 seconds ago      Up 20 seconds                0.0.0.0:58080->8080/tcp   mytomcat_1
75bde1b990e7        b95b1efa0436           "kube-controller-man…"   5 minutes ago       Up 5 minutes                                           k8s_kube-controller-manager_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_4
0073cac306f0        ff281650a721           "/opt/bin/flanneld -…"   5 minutes ago       Up 5 minutes                                           k8s_kube-flannel_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962-5510-11e9-a96d-000c291ae345_1
f769c27066dc        ff281650a721           "cp -f /etc/kube-fla…"   5 minutes ago       Exited (0) 5 minutes ago                               k8s_install-cni_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962-5510-11e9-a96d-000c291ae345_1
b54ff4716e39        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962-5510-11e9-a96d-000c291ae345_1
408f7e4b42e8        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_4
fecb42b177df        eb516548c180           "/coredns -conf /etc…"   5 minutes ago       Up 5 minutes                                           k8s_coredns_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_1
e5dbdf2d4af3        eb516548c180           "/coredns -conf /etc…"   5 minutes ago       Up 5 minutes                                           k8s_coredns_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_1
de96a90cd8c8        ecf910f40d6e           "kube-apiserver --ad…"   5 minutes ago       Up 5 minutes                                           k8s_kube-apiserver_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_3
2a5b2f1eff20        5cd54e388aba           "/usr/local/bin/kube…"   5 minutes ago       Up 5 minutes                                           k8s_kube-proxy_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_1
c9de0db0ce55        00638a24688b           "kube-scheduler --bi…"   5 minutes ago       Up 5 minutes                                           k8s_kube-scheduler_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_3
97364e091073        2c4adeb21b4f           "etcd --advertise-cl…"   5 minutes ago       Up 5 minutes                                           k8s_etcd_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_3
bd0c87c3d533        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_4
08f894c8c252        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_4
e0e7875598de        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_3
461eaa7b2899        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_1
5233aca7eca5        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_3
841be3a3cc7d        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Up 5 minutes                                           k8s_POD_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_4
4698599e362c        b95b1efa0436           "kube-controller-man…"   5 minutes ago       Exited (2) 5 minutes ago                               k8s_kube-controller-manager_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_3
f0f87dd886f9        k8s.gcr.io/pause:3.1   "/pause"                 5 minutes ago       Exited (0) 5 minutes ago                               k8s_POD_kube-controller-manager-k8s-master_kube-system_0ff88c9b6e64cded3762e51ff18bce90_3
a0c24081b996        eb516548c180           "/coredns -conf /etc…"   About an hour ago   Exited (0) 5 minutes ago                               k8s_coredns_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_0
b8b4efead3b4        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_coredns-fb8b8dccf-q949f_kube-system_2f8c9997-550e-11e9-a96d-000c291ae345_3
96a92d82c0dc        eb516548c180           "/coredns -conf /etc…"   About an hour ago   Exited (0) 5 minutes ago                               k8s_coredns_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_0
c2f167c6e9c8        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_coredns-fb8b8dccf-ltlx4_kube-system_2f8b4ffb-550e-11e9-a96d-000c291ae345_3
cb1e6e3458ee        ff281650a721           "/opt/bin/flanneld -…"   About an hour ago   Exited (137) 5 minutes ago                             k8s_kube-flannel_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962-5510-11e9-a96d-000c291ae345_0
99acbfbb1e68        5cd54e388aba           "/usr/local/bin/kube…"   About an hour ago   Exited (2) 5 minutes ago                               k8s_kube-proxy_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_0
09195555e12f        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_kube-proxy-qp5k7_kube-system_2f1a1559-550e-11e9-a96d-000c291ae345_0
ed47844f108f        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_kube-flannel-ds-amd64-2gr2v_kube-system_4cf85962-5510-11e9-a96d-000c291ae345_0
f09401726136        00638a24688b           "kube-scheduler --bi…"   About an hour ago   Exited (2) 5 minutes ago                               k8s_kube-scheduler_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_2
a03a28a6ef98        ecf910f40d6e           "kube-apiserver --ad…"   About an hour ago   Exited (1) 5 minutes ago                               k8s_kube-apiserver_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_2
4769961090db        2c4adeb21b4f           "etcd --advertise-cl…"   About an hour ago   Exited (0) 5 minutes ago                               k8s_etcd_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_2
191090440ed3        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_etcd-k8s-master_kube-system_804ba6a1bef952d18f2040a1ff90dbc3_2
89001ab3e457        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_kube-apiserver-k8s-master_kube-system_078b26a5af4c34641521cf85bb8b5ee7_2
4e93b1057da0        k8s.gcr.io/pause:3.1   "/pause"                 About an hour ago   Exited (0) 5 minutes ago                               k8s_POD_kube-scheduler-k8s-master_kube-system_58272442e226c838b193bbba4c44091e_2

 

至此,kubernetes安装完成

posted on 2019-04-02 14:00  pu20065226  阅读(1314)  评论(0编辑  收藏  举报