如何搭建属于你自己的kubernetes高可用集群

系统架构图

安装docker(每个节点)

# install-docker.sh
# system: centos7.5+
yum update -y yum install -y \ yum-utils \ device-mapper-persistent-data \ lvm2 yum-config-manager -y \ --add-repo \ https://download.docker.com/linux/centos/docker-ce.repo yum update -y yum install -y docker-ce systemctl enable docker systemctl start docker systemctl status docker

更改hostname(每个节点)

编辑/etc/hostname 文件,将每个节点的hostname 修改的不一样

编辑/etc/hosts,在其中将hostname 和 每个节点的ip地址对应起来,如果节点数量太多,使用dns或者不配置也可

修改内核参数(每个节点)

cat > /etc/sysctl.d/k8s.conf <<EOF
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.ip_nonlocal_bind = 1
net.ipv4.ip_forward = 1
vm.swappiness=0
EOF
sysctl --system

安装并启用ipvs 和管理工具 ipset ipvsadm(每个节点)

cat > /etc/sysconfig/modules/ipvs.modules <<EOF
#!/bin/bash
modprobe -- ip_vs
modprobe -- ip_vs_rr
modprobe -- ip_vs_wrr
modprobe -- ip_vs_sh
modprobe -- nf_conntrack_ipv4
EOF
#执行脚本
chmod 755 /etc/sysconfig/modules/ipvs.modules && bash /etc/sysconfig/modules/ipvs.modules && lsmod | grep -e ip_vs -e nf_conntrack_ipv4
#安装相关管理工具
yum install ipset ipvsadm -y

 

启动haproxy和keepalived

在所有节点上运行start_haproxy.sh

# start_haproxy.sh
#!/bin/bash MasterIP1=192.168.1.81 MasterIP2=192.168.1.82 MasterIP3=192.168.1.83 MasterPort=6443 docker run -d --restart=always --name HAProxy-K8S -p 6444:6444 \ -e MasterIP1=$MasterIP1 \ -e MasterIP2=$MasterIP2 \ -e MasterIP3=$MasterIP3 \ -e MasterPort=$MasterPort \
wise2c/haproxy-k8s 

在所有master节点上运行start_keepalived.sh

#start_keepalived.sh
#!/bin/bash VIRTUAL_IP=192.168.1.90 INTERFACE=ens192 NETMASK_BIT=24 CHECK_PORT=6444 RID=10 VRID=90 MCAST_GROUP=224.0.0.18 docker run -itd --restart=always --name=Keepalived-K8S \ --net=host --cap-add=NET_ADMIN \ -e VIRTUAL_IP=$VIRTUAL_IP \ -e INTERFACE=$INTERFACE \ -e CHECK_PORT=$CHECK_PORT \ -e RID=$RID \ -e VRID=$VRID \ -e NETMASK_BIT=$NETMASK_BIT \ -e MCAST_GROUP=$MCAST_GROUP \ wise2c/keepalived-k8s

 

 关闭swap(每个节点)

swapoff -a
yes | cp /etc/fstab /etc/fstab_bak
cat /etc/fstab_bak | grep -v swap > /etc/fstab

 

安装kubeadm

#众所周知的原因,这里使用阿里云yum源进行替换:
cat <<EOF > /etc/yum.repos.d/kubernetes.repo
[kubernetes]
name=Kubernetes
baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg https://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg
EOF

#安装kubeadm、kubelet、kubectl,注意这里默认安装当前最新版本
yum install -y kubeadm kubelet kubectl
systemctl enable kubelet && systemctl start kubelet

 

初始化master节点

配置文件(kubeadm-config.yaml)

apiVersion: kubeadm.k8s.io/v1beta2
bootstrapTokens:
- groups:
  - system:bootstrappers:kubeadm:default-node-token
  token: abcdef.0123456789abcdef
  ttl: 24h0m0s
  usages:
  - signing
  - authentication
kind: InitConfiguration
localAPIEndpoint:
  advertiseAddress: 192.168.1.x
  bindPort: 6443
nodeRegistration:
  criSocket: /var/run/dockershim.sock
  name: node81
  taints:
  - effect: NoSchedule
    key: node-role.kubernetes.io/master
---
apiServer:
  timeoutForControlPlane: 4m0s
apiVersion: kubeadm.k8s.io/v1beta2
certificatesDir: /etc/kubernetes/pki
clusterName: kubernetes
controlPlaneEndpoint: "192.168.1.90:6444"
controllerManager: {}
dns:
  type: CoreDNS
etcd:
  local:
    dataDir: /var/lib/etcd
imageRepository:  registry.aliyuncs.com/google_containers
kind: ClusterConfiguration
kubernetesVersion: v1.18.0
networking:
  dnsDomain: cluster.local
  serviceSubnet: 10.96.0.0/12    
scheduler: {}

---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
featureGates:
  SupportIPVSProxyMode: true
mode: ipvs

说明:  

  • advertiseAddress: 此处为你运行第一个kubeadm 的内网地址
  • bindPort: 端口号默认6443
  • controlPlaneEndpoint:   配置虚拟ip的haproxy的前端端口号
  • imageRepository:   配置镜像拉取的地址,应为国内无法访问默认地址,此处配置为阿里云的地址,如果你自己有私有的镜像仓库并且其中有所有镜像,可以写你自己的
  • kubernetesVersion: 版本拉取镜像时候用
  • serviceSubnet:  calico子网的大小,决定你能运行多少pod
  • 最后的配置,是让kubenetes知道使用的网络模式

初始化kubenetesmaster 节点

kubeadm init --config=kubeadm-config.yaml --upload-certs > kubeadm-init.log

说明:

  • --upload-certs 上传证书,这样添加其他节点的时候会自动拉取,不用手动复制了
  • --config   编辑完成的配置文件的路径和名称

 配置kubectl

# root用户
cat << EOF >> ~/.bashrc export KUBECONFIG=/etc/kubernetes/admin.conf EOF source ~/.bashrc
# 普通用户
mkdir -p $HOME/.kube sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config sudo chown $(id -u):$(id -g) $HOME/.kube/config

 

安装网路组件calico

配置文件 (下载地址https://docs.projectcalico.org/manifests/calico.yaml)

 部署 calico

kubectl apply -f  calico.yaml

 

添加其他master节点 & 添加slave节点

在初始化过程中产生的logs文件中有添加文件的命令找到并在相应的节点上运行

添加slave 

Then you can join any number of worker nodes by running the following on each as root:

kubeadm join 192.168.1.90:6444 --token abcdef.012345678aaaaaaaa \
    --discovery-token-ca-cert-hash sha256:2bde1f5f2e9b47f90acd2be8442f6ca7abcaaaaaaaaaaaaaa

添加master

You can now join any number of the control-plane node running the following command on each as root:

  kubeadm join 192.168.1.90:6444 --token abcdef.0123456789abcdef \
    --discovery-token-ca-cert-hash sha256:2bde1f5f2e9b47f90acd2be8442f6ca7abcfd1cc16135fe77aaaaaaaaaa \
    --control-plane --certificate-key 2b41e901aac5fd70da7ef9371758d59c32a61f52cb65bdcf7aaaaa

 

参考:

  https://blog.csdn.net/networken/article/details/89599004#t9/

https://kubernetes.io/docs/setup/production-environment/tools/kubeadm/high-availability/

 

建议: 

  在管理的节点比较少,应用结构相对简单的情况下,可以使用docker swarm 学习搭建的成本很低,性价比高;

  当要管理的节点上升到100+的时候,这个时候kubenetes的优势会比较明显; 

posted @ 2020-07-21 18:16  TiDao  阅读(237)  评论(0)    收藏  举报