Kubernetes集群部署-kubernetes 1.21.0 集群部署

Kubernetes集群部署-kubernetes 1.21.0 集群部署1 集群软件版本说明 kubeadm kubelet kubectl 版本 1 21 0 1 21 0 1 21 0 安装位置 集群所有主机 集群所有主机 集群所有主机 作用 初始化集群 管理集群等 用于接收 api server 指令 对 pod 生命周期进行管理 集群应用命令行管理工具 2 kubernetes YUM 源准备 在 etc yum repos

大家好,我是讯享网,很高兴认识大家。

1. 集群软件版本说明

kubeadm kubelet kubectl
版本 1.21.0 1.21.0 1.21.0
安装位置 集群所有主机 集群所有主机 集群所有主机
作用 初始化集群、管理集群等 用于接收api-server指令,对pod生命周期进行管理 集群应用命令行管理工具

2. kubernetes YUM源准备

 在/etc/yum.repos.d/目录中创建k8s.repo文件,把下面内容复制进去即可。

2.1 谷歌YUM源

[kubernetes]
name=Kubernetes
baseurl=https://packages.cloud.google.com/yum/repos/kubernetes-el7-x86_64
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://packages.cloud.google.com/yum/doc/yum-key.gpg
        https://packages.cloud.google.com/yum/doc/rpm-package-key.gpg

2.2 阿里云YUM源

 [kubernetes]
name=Kubernetes
baseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64/
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg https://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg

 查看yum源 

 yum repolist

 3. 集群软件安装

查看指定版本
yum list kubeadm.x86_64 --showduplicates | sort -r
yum list kubelet.x86_64 --showduplicates | sort -r
yum list kubectl.x86_64 --showduplicates | sort -r

4. 配置kubelet

为了实现docker使用的cgroupdriver与kubelet使用的cgroup的一致性,建议修改如下文件内容。

增加配置:
# vim /etc/sysconfig/kubelet
KUBELET_EXTRA_ARGS="--cgroup-driver=systemd"

查看 kubelet是否自动启动

设置kubelet为开机自启动即可,由于没有生成配置文件,集群初始化后自动启动
systemctl enable kubelet

5. 集群镜像准备

5.1 查看 

#kubeadm config images list --kubernetes-version=v1.21.0


k8s.gcr.io/kube-apiserver:v1.21.0
k8s.gcr.io/kube-controller-manager:v1.21.0
k8s.gcr.io/kube-scheduler:v1.21.0
k8s.gcr.io/kube-proxy:v1.21.0
k8s.gcr.io/pause:3.4.1
k8s.gcr.io/etcd:3.4.13-0
k8s.gcr.io/coredns/coredns:v1.8.0

5.2 编辑下载脚本

vim image_download.sh

#将下面脚本cpoy
#!/bin/bash
images_list='
k8s.gcr.io/kube-apiserver:v1.21.0
k8s.gcr.io/kube-controller-manager:v1.21.0
k8s.gcr.io/kube-scheduler:v1.21.0
k8s.gcr.io/kube-proxy:v1.21.0
k8s.gcr.io/pause:3.4.1
k8s.gcr.io/etcd:3.4.13-0
k8s.gcr.io/coredns/coredns:v1.8.0'

for i in $images_list
do
        docker pull $i
done

docker save -o k8s-1-21-0.tar $images_list

#执行脚本即可

连不上使用下面

vim image_download.sh

#将下面脚本cpoy
#!/bin/bash
images_list='
aiotceo/kube-apiserver:v1.21.0
aiotceo/kube-controller-manager:v1.21.0
aiotceo/kube-scheduler:v1.21.0
aiotceo/kube-proxy:v1.21.0
aiotceo/pause:3.4.1
aiotceo/etcd:3.4.13-0
aiotceo/coredns/coredns:v1.8.0'

for i in $images_list
do
        docker pull $i
done

docker save -o k8s-1-21-0.tar $images_list

#执行脚本即可

images_list='
registry.aliyuncs.com/google_containers/kube-apiserver:v1.21.0
registry.aliyuncs.com/google_containers/kube-controller-manager:v1.21.0
registry.aliyuncs.com/google_containers/kube-scheduler:v1.21.0
registry.aliyuncs.com/google_containers/kube-proxy:v1.21.0
registry.aliyuncs.com/google_containers/pause:3.4.1
registry.aliyuncs.com/google_containers/etcd:3.4.13-0
registry.aliyuncs.com/google_containers/coredns/coredns:v1.8.0' 

 最后会在目下生成一个 k8s-1-21-0.tar 文件

注意:如果coredns报错请使用下面命令

docker pull coredns/coredns:1.8.0

docker tag coredns/coredns:1.8.0 registry.aliyuncs.com/google_containers/coredns/coredns:v1.8.0

kubeadm config images pull --image-repository registry.aliyuncs.com/google_containers --cri-socket unix:///var/run/cri-dockerd.sock

6. 集群初始化

kubeadm init   --image-repository=registry.aliyuncs.com/google_containers --kubernetes-version=v1.21.0 --pod-network-cidr=10.244.0.0/16 --apiserver-advertise-address=192.168.80.200

kubeadm init 命令

 --apiserver-advertise-address string 设置 apiserver 绑定的 IP. --apiserver-bind-port int32 设置apiserver 监听的端口. (默认 6443) --apiserver-cert-extra-sans strings api证书中指定额外的Subject Alternative Names (SANs) 可以是IP 也可以是DNS名称。 证书是和SAN绑定的。 --cert-dir string 证书存放的目录 (默认 "/etc/kubernetes/pki") --certificate-key string kubeadm-cert secret 中 用于加密 control-plane 证书的key --config string kubeadm 配置文件的路径. --cri-socket string CRI socket 文件路径,如果为空 kubeadm 将自动发现相关的socket文件; 只有当机器中存在多个 CRI socket 或者 存在非标准 CRI socket 时才指定. --dry-run 测试,并不真正执行;输出运行后的结果. --feature-gates string 指定启用哪些额外的feature 使用 key=value 对的形式。 -h, --help 帮助文档 --ignore-preflight-errors strings 忽略前置检查错误,被忽略的错误将被显示为警告. 例子: 'IsPrivilegedUser,Swap'. Value 'all' ignores errors from all checks. --image-repository string 选择拉取 control plane images 的镜像repo (default "k8s.gcr.io") --kubernetes-version string 选择K8S版本. (default "stable-1") --node-name string 指定node的名称,默认使用 node 的 hostname. --pod-network-cidr string 指定 pod 的网络, control plane 会自动将 网络发布到其他节点的node,让其上启动的容器使用此网络 --service-cidr string 指定service 的IP 范围. (default "10.96.0.0/12") --service-dns-domain string 指定 service 的 dns 后缀, e.g. "myorg.internal". (default "cluster.local") --skip-certificate-key-print 不打印 control-plane 用于加密证书的key. --skip-phases strings 跳过指定的阶段(phase) --skip-token-print 不打印 kubeadm init 生成的 default bootstrap token --token string 指定 node 和control plane 之间,简历双向认证的token ,格式为 [a-z0-9]{6}\.[a-z0-9]{16} - e.g. abcdef.0abcdef --token-ttl duration token 自动删除的时间间隔。 (e.g. 1s, 2m, 3h). 如果设置为 '0', token 永不过期 (default 24h0m0s) --upload-certs 上传 control-plane 证书到 kubeadm-certs Secret.

讯享网

执行报错1


讯享网

解决方法:

1. 首先清空上次初始化产生的文件

rm -rf /etc/kubernetes/*
rm -rf ~/.kube/*
rm -rf /var/lib/etcd/*

2.  然后下载个包,这个包是用来停用被占用的端口的 

yum install lsof

3.  然后停用被占用的端口号,端口号替换成自己报错的端口

 docker ps -a | grep kube | grep -v pause



lsof -i :10259|grep -v "PID"|awk '{print "kill -9",$2}'|sh

lsof -i :10250|grep -v "PID"|awk '{print "kill -9",$2}'|sh

lsof -i :10257|grep -v "PID"|awk '{print "kill -9",$2}'|sh

lsof -i :6443|grep -v "PID"|awk '{print "kill -9",$2}'|sh

4. 重启 

kubeadm reset 

5. 重新执行init 命令

6.如果安装报错了,可以清理一下

swapoff -a && kubeadm reset  && systemctl daemon-reload && systemctl restart kubelet  && iptables -F && iptables -t nat -F && iptables -t mangle -F && iptables -X

安装成功

[root@master01 member]# kubeadm init    --image-repository=registry.aliyuncs.com/google_containers --kubernetes-version=v1.21.0 --pod-network-cidr=10.244.0.0/16 --apiserver-advertise-address=192.168.80.200 --v=5
I1230 17:20:05.   33175 initconfiguration.go:115] detected and using CRI socket: /var/run/dockershim.sock
[init] Using Kubernetes version: v1.21.0
[preflight] Running pre-flight checks
I1230 17:20:05.   33175 checks.go:582] validating Kubernetes and kubeadm version
I1230 17:20:05.   33175 checks.go:167] validating if the firewall is enabled and active
I1230 17:20:05.   33175 checks.go:202] validating availability of port 6443
I1230 17:20:05.   33175 checks.go:202] validating availability of port 10259
I1230 17:20:05.   33175 checks.go:202] validating availability of port 10257
I1230 17:20:05.   33175 checks.go:287] validating the existence of file /etc/kubernetes/manifests/kube-apiserver.yaml
I1230 17:20:05.   33175 checks.go:287] validating the existence of file /etc/kubernetes/manifests/kube-controller-manager.yaml
I1230 17:20:05.   33175 checks.go:287] validating the existence of file /etc/kubernetes/manifests/kube-scheduler.yaml
I1230 17:20:05.   33175 checks.go:287] validating the existence of file /etc/kubernetes/manifests/etcd.yaml
I1230 17:20:05.   33175 checks.go:437] validating if the connectivity type is via proxy or direct
I1230 17:20:05.   33175 checks.go:476] validating http connectivity to first IP address in the CIDR
I1230 17:20:05.   33175 checks.go:476] validating http connectivity to first IP address in the CIDR
I1230 17:20:05.   33175 checks.go:103] validating the container runtime
I1230 17:20:05.   33175 checks.go:129] validating if the "docker" service is enabled and active
I1230 17:20:05.   33175 checks.go:336] validating the contents of file /proc/sys/net/bridge/bridge-nf-call-iptables
I1230 17:20:05.   33175 checks.go:336] validating the contents of file /proc/sys/net/ipv4/ip_forward
I1230 17:20:05.   33175 checks.go:654] validating whether swap is enabled or not
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable conntrack
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable ip
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable iptables
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable mount
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable nsenter
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable ebtables
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable ethtool
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable socat
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable tc
I1230 17:20:05.   33175 checks.go:377] validating the presence of executable touch
I1230 17:20:05.   33175 checks.go:525] running all checks
I1230 17:20:05.   33175 checks.go:408] checking whether the given node name is valid and reachable using net.LookupHost
I1230 17:20:05.   33175 checks.go:623] validating kubelet version
I1230 17:20:06.058585   33175 checks.go:129] validating if the "kubelet" service is enabled and active
I1230 17:20:06.071287   33175 checks.go:202] validating availability of port 10250
I1230 17:20:06.071938   33175 checks.go:202] validating availability of port 2379
I1230 17:20:06.071979   33175 checks.go:202] validating availability of port 2380
I1230 17:20:06.072018   33175 checks.go:250] validating the existence and emptiness of directory /var/lib/etcd
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
I1230 17:20:06.092436   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/kube-apiserver:v1.21.0
I1230 17:20:06.   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/kube-controller-manager:v1.21.0
I1230 17:20:06.   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/kube-scheduler:v1.21.0
I1230 17:20:06.   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/kube-proxy:v1.21.0
I1230 17:20:06.   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/pause:3.4.1
I1230 17:20:06.   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/etcd:3.4.13-0
I1230 17:20:06.   33175 checks.go:844] image exists: registry.aliyuncs.com/google_containers/coredns/coredns:v1.8.0
[certs] Using certificateDir folder "/etc/kubernetes/pki"
I1230 17:20:06.   33175 certs.go:110] creating a new certificate authority for ca
[certs] Generating "ca" certificate and key
I1230 17:20:06.   33175 certs.go:487] validating certificate period for ca certificate
[certs] Generating "apiserver" certificate and key
[certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local master01] and IPs [10.96.0.1 192.168.80.200]
[certs] Generating "apiserver-kubelet-client" certificate and key
I1230 17:20:06.   33175 certs.go:110] creating a new certificate authority for front-proxy-ca
[certs] Generating "front-proxy-ca" certificate and key
I1230 17:20:07.   33175 certs.go:487] validating certificate period for front-proxy-ca certificate
[certs] Generating "front-proxy-client" certificate and key
I1230 17:20:07.   33175 certs.go:110] creating a new certificate authority for etcd-ca
[certs] Generating "etcd/ca" certificate and key
I1230 17:20:07.   33175 certs.go:487] validating certificate period for etcd/ca certificate
[certs] Generating "etcd/server" certificate and key
[certs] etcd/server serving cert is signed for DNS names [localhost master01] and IPs [192.168.80.200 127.0.0.1 ::1]
[certs] Generating "etcd/peer" certificate and key
[certs] etcd/peer serving cert is signed for DNS names [localhost master01] and IPs [192.168.80.200 127.0.0.1 ::1]
[certs] Generating "etcd/healthcheck-client" certificate and key
[certs] Generating "apiserver-etcd-client" certificate and key
I1230 17:20:08.   33175 certs.go:76] creating new public/private key files for signing service account users
[certs] Generating "sa" key and public key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I1230 17:20:08.   33175 kubeconfig.go:101] creating kubeconfig file for admin.conf
[kubeconfig] Writing "admin.conf" kubeconfig file
I1230 17:20:08.   33175 kubeconfig.go:101] creating kubeconfig file for kubelet.conf
[kubeconfig] Writing "kubelet.conf" kubeconfig file
I1230 17:20:08.   33175 kubeconfig.go:101] creating kubeconfig file for controller-manager.conf
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
I1230 17:20:09.   33175 kubeconfig.go:101] creating kubeconfig file for scheduler.conf
[kubeconfig] Writing "scheduler.conf" kubeconfig file
I1230 17:20:09.   33175 kubelet.go:63] Stopping the kubelet
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
I1230 17:20:09.   33175 manifests.go:96] [control-plane] getting StaticPodSpecs
I1230 17:20:09.   33175 certs.go:487] validating certificate period for CA certificate
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "ca-certs" for component "kube-apiserver"
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "etc-pki" for component "kube-apiserver"
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "k8s-certs" for component "kube-apiserver"
I1230 17:20:09.   33175 manifests.go:126] [control-plane] wrote static Pod manifest for component "kube-apiserver" to "/etc/kubernetes/manifests/kube-apiserver.yaml"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
I1230 17:20:09.   33175 manifests.go:96] [control-plane] getting StaticPodSpecs
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "ca-certs" for component "kube-controller-manager"
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "etc-pki" for component "kube-controller-manager"
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "flexvolume-dir" for component "kube-controller-manager"
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "k8s-certs" for component "kube-controller-manager"
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "kubeconfig" for component "kube-controller-manager"
I1230 17:20:09.   33175 manifests.go:126] [control-plane] wrote static Pod manifest for component "kube-controller-manager" to "/etc/kubernetes/manifests/kube-controller-manager.yaml"
[control-plane] Creating static Pod manifest for "kube-scheduler"
I1230 17:20:09.   33175 manifests.go:96] [control-plane] getting StaticPodSpecs
I1230 17:20:09.   33175 manifests.go:109] [control-plane] adding volume "kubeconfig" for component "kube-scheduler"
I1230 17:20:09.   33175 manifests.go:126] [control-plane] wrote static Pod manifest for component "kube-scheduler" to "/etc/kubernetes/manifests/kube-scheduler.yaml"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I1230 17:20:09.   33175 local.go:74] [etcd] wrote Static Pod manifest for a local etcd member to "/etc/kubernetes/manifests/etcd.yaml"
I1230 17:20:09.   33175 waitcontrolplane.go:87] [wait-control-plane] Waiting for the API server to be healthy
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[kubelet-check] Initial timeout of 40s passed.
[apiclient] All control plane components are healthy after 58. seconds
I1230 17:21:08.   33175 uploadconfig.go:108] [upload-config] Uploading the kubeadm ClusterConfiguration to a ConfigMap
[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1230 17:21:08.   33175 uploadconfig.go:122] [upload-config] Uploading the kubelet component config to a ConfigMap
[kubelet] Creating a ConfigMap "kubelet-config-1.21" in namespace kube-system with the configuration for the kubelets in the cluster
I1230 17:21:08.   33175 uploadconfig.go:127] [upload-config] Preserving the CRISocket information for the control-plane node
I1230 17:21:08.   33175 patchnode.go:30] [patchnode] Uploading the CRI Socket information "/var/run/dockershim.sock" to the Node API object "master01" as an annotation
[upload-certs] Skipping phase. Please see --upload-certs
[mark-control-plane] Marking the node master01 as control-plane by adding the labels: [node-role.kubernetes.io/master(deprecated) node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
[mark-control-plane] Marking the node master01 as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]
[bootstrap-token] Using token: 874v22.cuk7cmslc4j950y0
[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodes
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1230 17:21:09.   33175 clusterinfo.go:45] [bootstrap-token] loading admin kubeconfig
I1230 17:21:09.   33175 clusterinfo.go:56] [bootstrap-token] copying the cluster from admin.conf to the bootstrap kubeconfig
I1230 17:21:09.   33175 clusterinfo.go:68] [bootstrap-token] creating/updating ConfigMap in kube-public namespace
I1230 17:21:09.   33175 clusterinfo.go:82] creating the RBAC rules for exposing the cluster-info ConfigMap in the kube-public namespace
I1230 17:21:09.   33175 kubeletfinalize.go:88] [kubelet-finalize] Assuming that kubelet client certificate rotation is enabled: found "/var/lib/kubelet/pki/kubelet-client-current.pem"
[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1230 17:21:09.   33175 kubeletfinalize.go:132] [kubelet-finalize] Restarting the kubelet to enable client certificate rotation
[addons] Applied essential addon: CoreDNS
I1230 17:21:09.   33175 request.go:600] Waited for 138.76319ms due to client-side throttling, not priority and fairness, request: POST:https://192.168.80.200:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s
[addons] Applied essential addon: kube-proxy

Your Kubernetes control-plane has initialized successfully!

To start using your cluster, you need to run the following as a regular user:

  mkdir -p $HOME/.kube
  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
  sudo chown $(id -u):$(id -g) $HOME/.kube/config

Alternatively, if you are the root user, you can run:

  export KUBECONFIG=/etc/kubernetes/admin.conf

You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
  https://kubernetes.io/docs/concepts/cluster-administration/addons/

Then you can join any number of worker nodes by running the following on each as root:

kubeadm join 192.168.80.200:6443 --token 874v22.cuk7cmslc4j950y0 \
    --discovery-token-ca-cert-hash sha256:fb7763cda54e9a3fc012c1f0b88fbce7de08d92c9164f49d4fb 
[root@master01 member]# 
 

7. 集群应用客户端管理集群文件准备

  mkdir -p $HOME/.kube
  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
  sudo chown $(id -u):$(id -g) $HOME/.kube/config

  export KUBECONFIG=/etc/kubernetes/admin.conf

8 集群网络来配置

使用calico插件

使用calico部署集群网络

安装参考网址:About Calico | Calico Documentation

8.1 calicao安装

1. 创建calico自定义类型命令

kubectl create -f https://raw.githubusercontent.com/projectcalico/calico/v3.27.0/manifests/tigera-operator.yaml

2. 配置calico的核心资源配置文件

# This section includes base Calico installation configuration.
# For more information, see: https://docs.tigera.io/calico/latest/reference/installation/api#operator.tigera.io/v1.Installation
apiVersion: operator.tigera.io/v1
kind: Installation
metadata:
  name: default
spec:
  # Configures Calico networking.
  calicoNetwork:
    # Note: The ipPools section cannot be modified post-install.
    ipPools:
    - blockSize: 26

       #这个需要改成你自己的iip
      cidr: 10.244.0.0/16
      encapsulation: VXLANCrossSubnet
      natOutgoing: Enabled
      nodeSelector: all()

---

# This section configures the Calico API server.
# For more information, see: https://docs.tigera.io/calico/latest/reference/installation/api#operator.tigera.io/v1.APIServer
apiVersion: operator.tigera.io/v1
kind: APIServer
metadata:
  name: default
spec: {}

 3. 执行命令

kubectl apply -f custom-resources.yaml

 4. 检查是否安装成功

已经全部运行
# kubectl get pods -n calico-system
NAME                                      READY   STATUS    RESTARTS   AGE
calico-kube-controllers-666bb9949-dzp68   1/1     Running   0          11m
calico-node-jhcf4                         1/1     Running   4          11m
calico-typha-68b96d8d9c-7qfq7             1/1     Running   2          11m

8.2 calico 客户端安装

查看添加权限后文件
# ls /usr/bin/calicoctl
/usr/bin/calicoctl

查看calicoctl版本
# calicoctl  version
Client Version:    v3.21.4
Git commit:        220d04c94
Cluster Version:   v3.21.4
Cluster Type:      typha,kdd,k8s,operator,bgp,kubeadm

通过~/.kube/config连接kubernetes集群,查看已运行节点
# DATASTORE_TYPE=kubernetes KUBECONFIG=~/.kube/config calicoctl get nodes
NAME
master01 

9. 集群工作节点添加

要使用你初始化master节点最下面的命令

在master节点上执行,检查网络节点是否添加成功

DATASTORE_TYPE=kubernetes KUBECONFIG=~/.kube/config calicoctl get nodes 

[root@master01 calico]# DATASTORE_TYPE=kubernetes KUBECONFIG=~/.kube/config calicoctl get nodes
NAME       
master01   
worker01   
worker02   
 

10. 检查集群可用性

查看所有的节点

[root@master01 ~]# kubectl get nodes
NAME       STATUS   ROLES                  AGE    VERSION
master01   Ready    control-plane,master   169m   v1.21.0
worker01   Ready    <none>                 28m    v1.21.0
worker02   Ready    <none>                 28m    v1.21.0

 查看集群健康情况,理想状态

 [root@master01 ~]# kubectl get cs
NAME                 STATUS    MESSAGE             ERROR
controller-manager   Healthy   ok
scheduler            Healthy   ok
etcd-0               Healthy   {"health":"true"}

有问题的情况

[root@master01 calico]# kubectl get cs
Warning: v1 ComponentStatus is deprecated in v1.19+
NAME                 STATUS      MESSAGE                                                                                       ERROR
controller-manager   Unhealthy   Get "http://127.0.0.1:10252/healthz": dial tcp 127.0.0.1:10252: connect: connection refused   
scheduler            Unhealthy   Get "http://127.0.0.1:10251/healthz": dial tcp 127.0.0.1:10251: connect: connection refused   
etcd-0               Healthy     {"health":"true"}  

 查看kubernetes集群pod运行情况

[root@master01 ~]# kubectl get pods -n kube-system
NAME                               READY   STATUS    RESTARTS   AGE
coredns-558bd4d5db-4jbdv           1/1     Running   1          169m
coredns-558bd4d5db-pw5x5           1/1     Running   1          169m
etcd-master01                      1/1     Running   1          170m
kube-apiserver-master01            1/1     Running   1          170m
kube-controller-manager-master01   1/1     Running   14         170m
kube-proxy-kbx4z                   1/1     Running   1          169m
kube-proxy-rgtr8                   1/1     Running   0          29m
kube-proxy-sq9xv                   1/1     Running   0          29m
kube-scheduler-master01            1/1     Running   11         170m

再次查看calico-system命名空间中pod运行情况。
[root@master01 ~]# kubectl get pods -n calico-system
NAME                                      READY   STATUS    RESTARTS   AGE
calico-kube-controllers-666bb9949-dzp68   1/1     Running   3          70m
calico-node-jhcf4                         1/1     Running   15         70m
calico-node-jxq9p                         1/1     Running   0          30m
calico-node-kf78q                         1/1     Running   0          30m
calico-typha-68b96d8d9c-7qfq7             1/1     Running   13         70m
calico-typha-68b96d8d9c-wz2zj             1/1     Running   0          20m 

 kubectl get pods -n calico-system -o wide

10. 

小讯
上一篇 2025-01-14 23:34
下一篇 2025-02-14 12:09

相关推荐

版权声明:本文内容由互联网用户自发贡献,该文观点仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容,请联系我们,一经查实,本站将立刻删除。
如需转载请保留出处:https://51itzy.com/kjqy/48763.html