【k8s集群搭建(一):基于虚拟机的linux的k8s集群搭建_超详细_解决并记录全过程步骤以及自己的踩坑记录】

2 篇文章 0 订阅

虚拟机准备3台Linux系统

在这里插入图片描述

在这里插入图片描述

k8s集群安装

每一台机器需要安装以下内容:
docker:容器运行环境
kubelet:控制机器中所有资源
bubelctl:命令行
kubeladm:初始化集群的工具

Docker安装

安装一些必要的包,yum-util 提供yum-config-manager功能,另两个是devicemapper驱动依赖

#yum install -y yum-utils
yum install -y yum-utils device-mapper-persistent-data lvm2

添加dcoker仓库

yum-config-manager --add-repo http://download.docker.com/linux/centos/docker-ce.repo(中央仓库)

yum-config-manager --add-repo http://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo(阿里仓库)

选择一个版本并安装:yum install -y docker-ce-版本号

yum install -y docker-ce-20.10.7 docker-ce-cli-20.10.7 containerd.io-1.4.6

既要现在启动也要开机启动

systemctl enable docker --now

有可能出现如下问题

Job for docker.service failed because the control process exited with error code. See "systemctl status docker.service" and "journalctl -xe" for details.

通过systemctl status docker.servicejournalctl -xe查看问题

[root@localhost docker]# systemctl status docker.service
● docker.service - Docker Application Container Engine
   Loaded: loaded (/usr/lib/systemd/system/docker.service; enabled; vendor preset: disabled)
   Active: failed (Result: start-limit) since 六 2023-11-11 23:34:34 CST; 15s ago
     Docs: https://docs.docker.com
  Process: 33144 ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock (code=exited, status=1/FAILURE)
 Main PID: 33144 (code=exited, status=1/FAILURE)

1111 23:34:32 localhost.localdomain systemd[1]: docker.service: main process exited, code=exited, status=1/FAILURE
1111 23:34:32 localhost.localdomain systemd[1]: Failed to start Docker Application Container Engine.
1111 23:34:32 localhost.localdomain systemd[1]: Unit docker.service entered failed state.
1111 23:34:32 localhost.localdomain systemd[1]: docker.service failed.
1111 23:34:34 localhost.localdomain systemd[1]: docker.service holdoff time over, scheduling restart.
1111 23:34:34 localhost.localdomain systemd[1]: Stopped Docker Application Container Engine.
1111 23:34:34 localhost.localdomain systemd[1]: start request repeated too quickly for docker.service
1111 23:34:34 localhost.localdomain systemd[1]: Failed to start Docker Application Container Engine.
1111 23:34:34 localhost.localdomain systemd[1]: Unit docker.service entered failed state.
1111 23:34:34 localhost.localdomain systemd[1]: docker.service failed.
[root@localhost docker]# journalctl -xe
-- 
-- Unit docker.service has finished shutting down.
1111 23:34:32 localhost.localdomain systemd[1]: Starting Docker Application Container Engine...
-- Subject: Unit docker.service has begun start-up
-- Defined-By: systemd
-- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel
-- 
-- Unit docker.service has begun starting up.
1111 23:34:32 localhost.localdomain dockerd[33144]: unable to configure the Docker daemon with file /etc/docker/daemon.json: invalid character '"' after object key:value pair
1111 23:34:32 localhost.localdomain systemd[1]: docker.service: main process exited, code=exited, status=1/FAILURE
1111 23:34:32 localhost.localdomain systemd[1]: Failed to start Docker Application Container Engine.

或者

[root@localhost ~]# journalctl -xe
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.210236454+08:00" level=info msg="Starting up"
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.211453996+08:00" level=info msg="parsed scheme: \"unix\"" module=grpc
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.211480009+08:00" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.211493531+08:00" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///run/containerd/containerd.sock  <nil> 
11月 11 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.211499438+08:00" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
11月 11 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.213840454+08:00" level=info msg="parsed scheme: \"unix\"" module=grpc
11月 11 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.213851730+08:00" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc
11月 11 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.213863519+08:00" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///run/containerd/containerd.sock  <nil> 
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.213867949+08:00" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.225245436+08:00" level=warning msg="Usage of loopback devices is strongly discouraged for production use. Please use `--sto
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.244578994+08:00" level=warning msg="Base device already exists and has filesystem xfs on it. User specified filesystem  wil
11月 11 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.269766614+08:00" level=info msg="[graphdriver] using prior storage driver: devicemapper"
11月 11 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.269788540+08:00" level=warning msg="[graphdriver] WARNING: the devicemapper storage-driver is deprecated, and will be remov
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.272569072+08:00" level=warning msg="Unable to find pids cgroup in mounts"
1111 23:52:06 localhost.localdomain dockerd[30814]: time="2023-11-11T23:52:06.272709009+08:00" level=info msg="Loading containers: start."
1111 23:52:06 localhost.localdomain firewalld[1011]: 2023-11-11 23:52:06 ERROR: INVALID_TYPE: structure size mismatch 16 != 13
1111 23:52:06 localhost.localdomain firewalld[1011]: 2023-11-11 23:52:06 ERROR: COMMAND_FAILED: '/sbin/iptables -w2 -t filter -C FORWARD -j DOCKER-ISOLATION' failed: iptables v1.4.21: Couldn't load tar
                                                         
                                                         Try `iptables -h' or 'iptables --help' for more information.

问题产生原因如下:json格式问题
在这里插入图片描述
防火墙问题:参考链接
在这里插入图片描述

解决方法

进入目录:

cd /etc/docker

看看有没有文件daemon.json,没有的话创建

touch daemon.json

编辑

vi daemon.json
{
	"storage-driver": "devicemapper",
        "iptables": false
}

重新启动:既要现在启动也要开机启动

systemctl enable docker --now

配置镜像加速器

在这里插入图片描述

sudo mkdir -p /etc/docker
sudo tee /etc/docker/daemon.json <<-'EOF'
{
  "registry-mirrors": ["你的阿里云镜像加速器地址"],
  "exec-opts": ["native.cgroupdriver=systemd"],
  "log-driver": "json-file",
  "log-opts": {
	"max-size": "100m"
  },
  "storage-driver": "overlay2"
}
EOF
sudo systemctl daemon-reload
sudo systemctl restart docker

但是上面会覆盖之前写的/etc/docker/daemon.json配置导致和之前一样的错误,而"storage-driver": "overlay2"会导致报错

failed to start daemon: error initializing graphdriver: overlay2: the backing xfs filesystem is formatted without d_type support, which leads to inco

所以暂时先把"storage-driver": "overlay2"这一行删除,重新执行如下命令:

sudo mkdir -p /etc/docker
sudo tee /etc/docker/daemon.json <<-'EOF'
{
  "storage-driver": "devicemapper",
  "iptables": false,
  "registry-mirrors": ["你的阿里云镜像加速器地址"],
  "exec-opts": ["native.cgroupdriver=systemd"],
  "log-driver": "json-file",
  "log-opts": {
	"max-size": "100m"
  }
}
EOF
sudo systemctl daemon-reload
sudo systemctl restart docker

这一问题通过这一操作,暂时解决。

修改各节点名称,方便secureCRT观察是哪个节点

参考链接

vi /etc/passwd

在这里插入图片描述

vi /etc/shadow

在这里插入图片描述
无法保存退出,采用强制退出(需root权限参考链接
在这里插入图片描述

:wq!
vi /etc/sudoers

在这里插入图片描述
同上需要强制保存退出。
修改之后连接不上!
查找原因,参考链接
在这里插入图片描述

Last login: Sun Nov 12 10:11:12 2023 from 192.168.139.1
Could not chdir to home directory /master-root: No such file or directory
-bash-4.2# vi /etc/passwd
-bash-4.2# 

在这里插入图片描述
但是虚拟机开机有问题,因此这种方法先不用,先不使用root账户,修改各个机器的普通用户名。
在这里插入图片描述
修改之后普通用户连接不上,但是root可以连接上
在这里插入图片描述
观察之后发现控制台中的用户名还没有被修改过来
在这里插入图片描述
查看文件:cat /etc/passwd,发现这些地方没有改
在这里插入图片描述
修改之后发现虚拟机登不上了,提示密码错误。

vi /etc/shadow

在这里插入图片描述
这时候可以登录了,但是登陆成功又返回到登录界面!

最终解决

修改用户名2个需要修改的地方

 vi /etc/passwd

在这里插入图片描述

vi /etc/shadow

在这里插入图片描述

vi /etc/sudoers

在这里插入图片描述

从以上可以推测出前面修改root用户名的错误的原因,也可以得到有效解决!
修改用户名完成。
在这里插入图片描述
看看docker,出现如下错误:

[node2@localhost ~]$ docker ps
Got permission denied while trying to connect to the Docker daemon socket at unix:///var/run/docker.sock: Get http://%2Fvar%2Frun%2Fdocker.sock/v1.24/containers/json: dial unix /var/run/docker.sock: connect: permission denied

以上命令只需要如下操作,即可

[node2@localhost ~]$ sudo gpasswd -a $USER docker
正在将用户“node2”加入到“docker”组中
[node2@localhost ~]$ newgrp docker

参考文章:链接1链接2

[node2@localhost ~]$ docker ps
Got permission denied while trying to connect to the Docker daemon socket at unix:///var/run/docker.sock: Get http://%2Fvar%2Frun%2Fdocker.sock/v1.24/containers/json: dial unix /var/run/docker.sock: connect: permission denied
[node2@localhost ~]$ sudo groupadd docker
[sudo] password for node2: 
groupadd:“docker”组已存在
[node2@localhost ~]$ sudo gpasswd -a $USER docker
正在将用户“node2”加入到“docker”组中
[node2@localhost ~]$ newgrp docker
[node2@localhost ~]$ docker ps
Cannot connect to the Docker daemon at unix:///var/run/docker.sock. Is the docker daemon running?
[node2@localhost ~]$ sudo service docker restart
Redirecting to /bin/systemctl restart  docker.service
Job for docker.service failed because start of the service was attempted too often. See "systemctl status docker.service" and "journalctl -xe" for details.
To force a start use "systemctl reset-failed docker.service" followed by "systemctl start docker.service" again.
[node2@localhost ~]$ systemctl status docker.service
● docker.service - Docker Application Container Engine
   Loaded: loaded (/usr/lib/systemd/system/docker.service; enabled; vendor preset: disabled)
   Active: failed (Result: start-limit) since 日 2023-11-12 15:12:31 CST; 1min 16s ago
     Docs: https://docs.docker.com
  Process: 10231 ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock (code=exited, status=1/FAILURE)
 Main PID: 10231 (code=exited, status=1/FAILURE)
[node2@localhost ~]$ journalctl -xe
Hint: You are currently not seeing messages from other users and the system.
      Users in the 'systemd-journal' group can see all messages. Pass -q to
      turn off this notice.
No journal files were opened due to insufficient permissions.

在这里插入图片描述
发现是之前的问题,重新执行以下前面的配置阿里云镜像加速器命令就可以了。

B站视频链接

设置主机名,防止机器之间重复

hostnamectl set-hostname 主机名

在这里插入图片描述
在这里插入图片描述

关闭交换分区

在这里插入图片描述
下面是一些安全设置

# 查看 交换分区
free -m

# 将 SELinux 设置为 permissive 模式(相当于将其禁用)  第一行是临时禁用,第二行是永久禁用
setenforce 0
sed -i 's/^SELINUX=enforcing$/SELINUX=permissive/' /etc/selinux/config

# 关闭swap;第一行是临时禁用,第二行是永久禁用
swapoff -a  
sed -ri 's/.*swap.*/#&/' /etc/fstab

# 允许 iptables 检查桥接流量 (K8s 官方要求)
cat <<EOF | sudo tee /etc/modules-load.d/k8s.conf
br_netfilter
EOF
cat <<EOF | sudo tee /etc/sysctl.d/k8s.conf
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
EOF

# 让配置生效
sysctl --system

安装k8s三大件

安装 kubelet、kebeadm、kubectl;注意版本 (1.20.9)

# 配置 k8s 的 yum 源地址
cat <<EOF | sudo tee /etc/yum.repos.d/kubernetes.repo
[kubernetes]
name=Kubernetes
baseurl=http://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64
enabled=1
gpgcheck=0
repo_gpgcheck=0
gpgkey=http://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg
   http://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg
EOF
# 安装 kubelet、kubeadm、kubectl
yum install -y kubelet-1.20.9 kubeadm-1.20.9 kubectl-1.20.9 --disableexcludes=kubernetes

# 启动kubelet
systemctl enable --now kubelet

# 查看 kubelet 状态:一会停止 一会运行。 这个状态是对的,kubelet 等待 kubeadm 发号指令。
systemctl status kubelet

下载镜像

# 配置镜像,生成 images.sh
sudo tee ./images.sh <<-'EOF'
#!/bin/bash
images=(
kube-apiserver:v1.20.9
kube-proxy:v1.20.9
kube-controller-manager:v1.20.9
kube-scheduler:v1.20.9
coredns:1.7.0
etcd:3.4.13-0
pause:3.2
)
for imageName in ${images[@]} ; do
docker pull registry.cn-hangzhou.aliyuncs.com/lfy_k8s_images/$imageName
done
EOF

# 拉取镜像
chmod +x ./images.sh && ./images.sh

初始化主节点第一步

# 所有机器添加 master 域名映射,以下 IP 为 master 的 IP;
# 访问 cluster-endpoint 即 访问 192.168.139.131
echo "192.168.139.131  cluster-endpoint" >> /etc/hosts

在这里插入图片描述
在其他节点上ping一下

ping cluster-endpoint

在这里插入图片描述
以上让所有节点知道了master节点在哪。

初始化主节点第二步

# 主节点初始化 (只在 master 服务器执行, 其他 node 不用)
# --apiserver-advertise-address: master 的 IP
# --control-plane-endpoint: master 的域名
# --service-cidr 和 --pod-network-cidr 是网络范围,雷神 建议不要改。要改的话 2 个cidr 和 vps(172.31.x.x) 的,3 个网络互相不能重叠;还要修改 calico.yaml的 IP(下图有写)。
kubeadm init \
--apiserver-advertise-address=192.168.139.131 \
--control-plane-endpoint=cluster-endpoint \
--image-repository registry.cn-hangzhou.aliyuncs.com/lfy_k8s_images \
--kubernetes-version v1.20.9 \
--service-cidr=10.96.0.0/16 \
--pod-network-cidr=10.244.0.0/16

参考链接:参考
出现如下错误:

CGROUPS_FREEZER: enabled
CGROUPS_MEMORY: enabled
CGROUPS_PIDS: missing
CGROUPS_HUGETLB: enabled
        [WARNING SystemVerification]: this Docker version is not on the list of validated versions: 20.10.7. Latest validated version: 19.03
        [WARNING Hostname]: hostname "k8s-master" could not be reached
        [WARNING Hostname]: hostname "k8s-master": lookup k8s-master on 192.168.139.2:53: no such host
error execution phase preflight: [preflight] Some fatal errors occurred:
        [ERROR SystemVerification]: missing required cgroups: pids
[preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
To see the stack trace of this error execute with --v=5 or higher

解决过程:
参考文章:链接1链接2

[root@k8s-master ~]# cd /boot/
[root@k8s-master boot]# ls
config-3.10.0-327.el7.x86_64                             initramfs-3.10.0-327.el7.x86_64.img       System.map-3.10.0-327.el7.x86_64
grub                                                     initramfs-3.10.0-327.el7.x86_64kdump.img  vmlinuz-0-rescue-5e285834986d407b84e8ba7a4b2e2a83
grub2                                                    initrd-plymouth.img                       vmlinuz-3.10.0-327.el7.x86_64
initramfs-0-rescue-5e285834986d407b84e8ba7a4b2e2a83.img  symvers-3.10.0-327.el7.x86_64.gz
[root@k8s-master boot]# vi config-3.10.0-327.el7.x86_64

在这里插入图片描述

在这里插入图片描述
升级内核

rpm --import https://www.elrepo.org/RPM-GPG-KEY-elrepo.org
rpm -Uvh http://www.elrepo.org/elrepo-release-7.0-2.el7.elrepo.noarch.rpm
yum --disablerepo="*" --enablerepo="elrepo-kernel" list available
yum --enablerepo=elrepo-kernel install kernel-ml
cp /etc/default/grub  /etc/default/grub_bak
vi /etc/default/grub
grub2-mkconfig -o /boot/grub2/grub.cfg
 systemctl enable docker.service
reboot

在这里插入图片描述
参考文章:参考文章1参考文章2
开机界面选择新内核:

重新执行,初始化主节点成功!

kubeadm init \
--apiserver-advertise-address=192.168.139.131 \
--control-plane-endpoint=cluster-endpoint \
--image-repository registry.aliyuncs.com/google_containers \
--kubernetes-version v1.20.9 \
--service-cidr=10.96.0.0/12 \
--pod-network-cidr=10.244.0.0/16

在这里插入图片描述

Your Kubernetes control-plane has initialized successfully!

To start using your cluster, you need to run the following as a regular user:

  mkdir -p $HOME/.kube
  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
  sudo chown $(id -u):$(id -g) $HOME/.kube/config

Alternatively, if you are the root user, you can run:

  export KUBECONFIG=/etc/kubernetes/admin.conf

You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
  https://kubernetes.io/docs/concepts/cluster-administration/addons/

You can now join any number of control-plane nodes by copying certificate authorities
and service account keys on each node and then running the following as root:

  kubeadm join cluster-endpoint:6443 --token 99sc8r.rv1dsi2mgvhqshvb \
    --discovery-token-ca-cert-hash sha256:0fe5bf32c6f0f818d91fbf6c096fb275e6d21776a9d462e5fce3a12c08a5f745 \
    --control-plane 

Then you can join any number of worker nodes by running the following on each as root:

kubeadm join cluster-endpoint:6443 --token 99sc8r.rv1dsi2mgvhqshvb \
    --discovery-token-ca-cert-hash sha256:0fe5bf32c6f0f818d91fbf6c096fb275e6d21776a9d462e5fce3a12c08a5f745 

按照提示,然后依次运行如下命令

mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config
[root@k8s-master ~]# kubectl get nodes
NAME         STATUS     ROLES                  AGE   VERSION
k8s-master   NotReady   control-plane,master   14m   v1.20.9

NotReady,是因为还需要部署网络插件把节点之间的网络打通

安装网络组件

master运行

# 下载 calico.yaml
curl https://docs.projectcalico.org/manifests/calico.yaml -O

# 加载配置
kubectl apply -f calico.yaml

出现问题:

[root@k8s-master ~]# curl https://docs.projectcalico.org/manifests/calico.yaml -O
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
curl: (60) The certificate issuer's certificate has expired.  Check your system date and time.
More details here: http://curl.haxx.se/docs/sslcerts.html

curl performs SSL certificate verification by default, using a "bundle"
 of Certificate Authority (CA) public keys (CA certs). If the default
 bundle file isn't adequate, you can specify an alternate file
 using the --cacert option.
If this HTTPS server uses a certificate signed by a CA represented in
 the bundle, the certificate verification probably failed due to a
 problem with the certificate (it might be expired, or the name might
 not match the domain name in the URL).
If you'd like to turn off curl's verification of the certificate, use
 the -k (or --insecure) option.

解决办法:参考博客

sudo yum install ca-certificates
sudo update-ca-trust extract

在这里插入图片描述
出现错误:

[root@k8s-master ~]# kubectl apply -f calico.yaml
error: error validating "calico.yaml": error validating data: invalid object to validate; if you choose to ignore these errors, turn validation off with --validate=false

解决办法:参考博客
只需要换一个版本就行了,执行下面这条命令,再执行文件就不会报错了

curl https://docs.projectcalico.org/v3.20/manifests/calico.yaml -O

在这里插入图片描述
如果修改了 初始化主节点中的--pod-network-cidr=192.168.0.0/16
将 calico.yaml 的配置, # 去掉,IP 写 改的 IP,参考链接。这里我忘记把注释#去掉了,下文会提到!

sed -i "s#192\.168\.0\.0/16#10.244.0.0/16#" calico.yaml
[root@k8s-master ~]# cat calico.yaml | grep "192.168"
[root@k8s-master ~]# cat calico.yaml | grep "10.244.0.0/16"
            #   value: "10.244.0.0/16"

查看集群中的所有资源

 kubectl get pod -A

在这里插入图片描述
网络组件配置好后,STATUS变为Ready

[root@k8s-master ~]# kubectl get nodes
NAME         STATUS   ROLES                  AGE   VERSION
k8s-master   Ready    control-plane,master   70m   v1.20.9

Worker节点加入集群

更具之前的提示内容,在worker节点运行:

kubeadm join cluster-endpoint:6443 --token 99sc8r.rv1dsi2mgvhqshvb \
    --discovery-token-ca-cert-hash sha256:0fe5bf32c6f0f818d91fbf6c096fb275e6d21776a9d462e5fce3a12c08a5f745 

出现和之前一样的错误,即升级内核。
之后重新运行以上命令,但是一直卡在那,是因为防火墙的原因,用firewall-cmd --state查看一下防火墙状态。

firewall-cmd --state

关闭防火墙,参考文章

systemctl stop firewalld.service

但以上关闭防火墙还是不能解决我的问题,在命令之后加上--v=5查看日志:
在这里插入图片描述

[root@k8s-node2 ~]# kubeadm join cluster-endpoint:6443 --token 8lc3nb.vszlli4xic9w2ueh     --discovery-token-ca-cert-hash sha256:0fe5bf32c6f0f818d91fbf6c096fb275e6d21776a9d462e5fce3a12c08a5f745  --v=5
I1113 22:38:32.855410   14399 join.go:395] [preflight] found NodeName empty; using OS hostname as NodeName
I1113 22:38:32.855757   14399 initconfiguration.go:104] detected and using CRI socket: /var/run/dockershim.sock
[preflight] Running pre-flight checks
I1113 22:38:32.855857   14399 preflight.go:90] [preflight] Running general checks
I1113 22:38:32.855990   14399 checks.go:249] validating the existence and emptiness of directory /etc/kubernetes/manifests
I1113 22:38:32.856034   14399 checks.go:286] validating the existence of file /etc/kubernetes/kubelet.conf
I1113 22:38:32.856044   14399 checks.go:286] validating the existence of file /etc/kubernetes/bootstrap-kubelet.conf
I1113 22:38:32.856072   14399 checks.go:102] validating the container runtime
I1113 22:38:32.965468   14399 checks.go:128] validating if the "docker" service is enabled and active
I1113 22:38:33.042133   14399 checks.go:335] validating the contents of file /proc/sys/net/bridge/bridge-nf-call-iptables
I1113 22:38:33.042272   14399 checks.go:335] validating the contents of file /proc/sys/net/ipv4/ip_forward
I1113 22:38:33.042302   14399 checks.go:649] validating whether swap is enabled or not
I1113 22:38:33.042558   14399 checks.go:376] validating the presence of executable conntrack
I1113 22:38:33.042648   14399 checks.go:376] validating the presence of executable ip
I1113 22:38:33.042722   14399 checks.go:376] validating the presence of executable iptables
I1113 22:38:33.042857   14399 checks.go:376] validating the presence of executable mount
I1113 22:38:33.042899   14399 checks.go:376] validating the presence of executable nsenter
I1113 22:38:33.042919   14399 checks.go:376] validating the presence of executable ebtables
I1113 22:38:33.042934   14399 checks.go:376] validating the presence of executable ethtool
I1113 22:38:33.042981   14399 checks.go:376] validating the presence of executable socat
I1113 22:38:33.042996   14399 checks.go:376] validating the presence of executable tc
I1113 22:38:33.043007   14399 checks.go:376] validating the presence of executable touch
I1113 22:38:33.043022   14399 checks.go:520] running all checks
        [WARNING SystemVerification]: this Docker version is not on the list of validated versions: 20.10.7. Latest validated version: 19.03
I1113 22:38:33.144313   14399 checks.go:406] checking whether the given node name is reachable using net.LookupHost
I1113 22:38:35.196501   14399 checks.go:618] validating kubelet version
I1113 22:38:35.248496   14399 checks.go:128] validating if the "kubelet" service is enabled and active
I1113 22:38:35.254997   14399 checks.go:201] validating availability of port 10250
I1113 22:38:35.255319   14399 checks.go:286] validating the existence of file /etc/kubernetes/pki/ca.crt
I1113 22:38:35.255346   14399 checks.go:432] validating if the connectivity type is via proxy or direct
I1113 22:38:35.255381   14399 join.go:465] [preflight] Discovering cluster-info
I1113 22:38:35.255403   14399 token.go:78] [discovery] Created cluster-info discovery client, requesting info from "cluster-endpoint:6443"
I1113 22:38:35.256704   14399 token.go:215] [discovery] Failed to request cluster-info, will try again: Get "https://cluster-endpoint:6443/api/v1/namespaces/kube-public/configmaps/cluster-info?timeout=10s": dial tcp 192.168.139.131:6443: connect: no route to host
I1113 22:38:41.255239   14399 token.go:215] [discovery] Failed to request cluster-info, will try again: Get "https://cluster-endpoint:6443/api/v1/namespaces/kube-public/configmaps/cluster-info?timeout=10s": dial tcp 192.168.139.131:6443: connect: no route to host

相似问题:参考链接,没有解决我的问题
在这里插入图片描述

突然发现之前的“#”没有去除,通过字符串查找10.244.0.0/16,把#号去掉
在这里插入图片描述
这个前面的#也要去掉,后来才发现!
在这里插入图片描述

[root@k8s-master ~]# cat calico.yaml | grep "192.168"
[root@k8s-master ~]# cat calico.yaml | grep "10.244.0.0/16"
            #   value: "10.244.0.0/16"
[root@k8s-master ~]# vi calico.yaml
[root@k8s-master ~]# cat calico.yaml | grep "10.244.0.0/16"
              value: "10.244.0.0/16"
[root@k8s-master ~]#

然后重新执行,master重新产生token:参考链接1参考链接2

kubectl apply -f calico.yaml
kubeadm token create --print-join-command

在这里插入图片描述
woker节点加入:
在这里插入图片描述
回到master 中查看:

[root@k8s-master ~]# kubectl get nodes
NAME         STATUS     ROLES                  AGE     VERSION
k8s-master   Ready      control-plane,master   2d1h    v1.20.9
k8s-node1    Ready      <none>                 3m49s   v1.20.9
k8s-node2    NotReady   <none>                 119s    v1.20.9
[root@k8s-master ~]# 

节点部署完成!

文章参考链接:参考链接

  • 7
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 1
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值