CKA考试预配环境

目录

第一题:RBAC

第二题:指定node设置为不可用

第三题:升级kubernetes节点

第四题:etcd 备份还原

第五题:创建 NetworkPolicy

第六题:创建 svc

第七题:创建 ingress 资源

第八题:扩展 deployment

第九题:将 pod 部署到指定 node 节点上

第十题:检查有多少 node 节点是健康状态

第十一题:创建多个 container 的 Pod

第十二题:创建 Persistent Volume

第十三题:创建 PVC

第十四题:监控 pod 的日志

第十五题:添加 sidecar container

第十六题:查看最高 CPU 使用率的 Pod

第十七题:集群故障排查


 

第一题:RBAC

# 配置 namespace,名称为 app-team1
# 编写 namespace 资源清单
user1@k8s-master:~/cka-lab-init/1$ cat ns-app-team1.yaml
apiVersion: v1
kind: Namespace
metadata:
  name: app-team1
 
# 部署 namespace 资源
user1@k8s-master:~/cka-lab-init/1$ kubectl apply -f ns-app-team1.yaml
namespace/app-team1 created
 
# 查看部署的资源
user1@k8s-master:~/cka-lab-init/1$ kubectl get namespaces app-team1
NAME        STATUS   AGE
app-team1   Active   17s

第二题:指定node设置为不可用

# 配置一个2副本的 deployments,分别调度到 k8s-node-1 和 k8s-node-1 上
# 编写 deployments 资源清单
user1@k8s-master:~/cka-lab-init/2$ cat deployment-nginx-pod-nodes.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: nginx-pod-nodes
spec:
  replicas: 2
  selector:
    matchLabels:
      app: nginx-pod-nodes
  template:
    metadata:
      labels:
        app: nginx-pod-nodes
    spec:
      containers:
      - name: nginx
        image: nginx:1.18.0
        imagePullPolicy: IfNotPresent 
        ports:
        - containerPort: 80
 
# 部署 deployments 资源
user1@k8s-master:~/cka-lab-init/2$ kubectl apply -f deployment-nginx-pod-nodes.yaml
deployment.apps/nginx-pod-nodes created
 
# 查看部署的资源,分别在 k8s-node-1 和 k8s-node-1 上
user1@k8s-master:~/cka-lab-init/2$ kubectl get pod -o wide
NAME                              READY  STATUS   RESTARTS  AGE   IP          NODE        NOMINATED NODE  READINESS GATES
nginx-pod-nodes-859fcdcc9d-65bwd  1/1    Running  0        115s   10.244.2.2  k8s-node-2   <none>          <none>
nginx-pod-nodes-859fcdcc9d-8jvdx  1/1    Running  0        115s   10.244.1.2  k8s-node-1   <none>          <none>

第三题:升级kubernetes节点

无需预配。

 

第四题:etcd 备份还原

# 预配置解决证书问题,就是把 /etc/kubernetes/pki/etcd 目录下的证书,copy 到 /opt/KUIN000601/ 目录
# 创建证书的目录
user1@k8s-master:~$ sudo mkdir /opt/KUIN00601/
 
# 把etcd的证书copy到上面创建的目录中,并且修改名字
user1@k8s-master:~$ cd /etc/kubernetes/pki/etcd
user1@k8s-master:/etc/kubernetes/pki/etcd$ sudo cp ca.crt /opt/KUIN00601/ca.crt
user1@k8s-master:/etc/kubernetes/pki/etcd$ sudo cp peer.crt /opt/KUIN00601/etcd-client.crt
user1@k8s-master:/etc/kubernetes/pki/etcd$ sudo cp peer.key /opt/KUIN00601/etcd-client.key
 
# 查看 etcd 的证书
user1@k8s-master:~$ ls -l /opt/KUIN00601/
total 12
-rw-r--r-- 1 root root 1017 1月   9 20:38 ca.crt
-rw-r--r-- 1 root root 1135 1月   9 20:38 etcd-client.crt
-rw------- 1 root root 1679 1月   9 20:38 etcd-client.key
 
# 注:如果没有 etcdctl 命令的话,可以从 docker 中复制 etcdctl 命令,查看 etcd 的 docker 名字
user1@k8s-master:~$ sudo docker ps | grep etcd
9c1e60ade311   303ce5db0e90                                "etcd --advertise-cl…"   3 hours ago   Up 3 hours             k8s_etcd_etcd-k8s-master_kube-system_4ffc056bc44b10088ef49648e403a455_1
 
# 从 etcd 的 docker 中把 etcdctl 文件 copy 出来
user1@k8s-master:~$ sudo docker cp 9c1e60ade311:/usr/local/bin/etcdctl /usr/local/bin/etcdctl
 
# 确认有 etcdctl 命令
user1@k8s-master:~$ etcdctl version 
etcdctl version: 3.4.3
API version: 3.4
 
# 创建一个备份到目录
user1@k8s-master:~$ sudo mkdir /var/lib/backup
user1@k8s-master:~$ sudo mkdir /srv/data
 
# 把备份到copy为etcd-snapshot-previous.db,用于恢复到实验(要等前面3题完成时的备份)
user1@k8s-master:~$ sudo cp /srv/data/etcd-snapshot.db /var/lib/backup/etcd-snapshot-previous.db

第五题:创建 NetworkPolicy

# 预配要增加一个 internal 的 namespace
# 编写 namespace 资源清单
user1@k8s-master:~/cka-lab-init/5$ cat ns-internal.yaml
apiVersion: v1
kind: Namespace
metadata:
  name: internal
 
# 部署 namespace 资源
user1@k8s-master:~/cka-lab-init/5$ kubectl apply -f ns-internal.yaml
namespace/internal created
 
# 查看部署的资源
user1@k8s-master:~/cka-lab-init/5$ kubectl get ns internal
NAME       STATUS   AGE
internal   Active   18s
 
# 预配置在 internal 的 namespace 里面增加2个 pod,用于测试
user1@k8s-master:~/cka-lab-init/5$ cat nginx-app-pod-80.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: nginx-pod
  namespace: internal
spec:
  replicas: 1
  selector:
    matchLabels:
      app: nginx
  template:
    metadata:
      labels:
        app: nginx
    spec:
      containers:
      - name: nginx
        image: nginx
        imagePullPolicy: IfNotPresent
---
apiVersion: v1
kind: Pod
metadata:
  name: client-pod-1
  namespace: internal
  labels:
    app: client-pod-1
spec:
  containers:
    - name: client-pod
      image: centos:7
      command: ["/bin/bash","-c","sleep 3600"]
       
# 创建2个 pod 用于测试
user1@k8s-master:~/cka-lab-init/5$ kubectl apply -f nginx-app-pod-80.yaml
deployment.apps/nginx-pod created
pod/client-pod-1 created
 
# 查看 pod 资源
user1@k8s-master:~/cka-lab-init/5$ kubectl get pod -o wide -n internal
NAME                         READY   STATUS    RESTARTS   AGE   IP            NODE         NOMINATED NODE   READINESS GATES
client-pod-1                 1/1     Running   0          57s   10.244.2.10   k8s-node-2   <none>           <none>
nginx-pod-7cf7d6dbc8-2tgz2   1/1     Running   0          57s   10.244.2.9    k8s-node-2   <none>           <none>
 
# 测试 ping 访问
user1@k8s-master:~/cka-lab-init/5$ kubectl exec -n internal client-pod-1 -- ping 10.244.2.9
PING 10.244.2.9 (10.244.2.9) 56(84) bytes of data.
64 bytes from 10.244.2.9: icmp_seq=1 ttl=63 time=0.071 ms
 
# 测试 http 访问
user1@k8s-master:~/cka-lab-init/5$ kubectl exec -n internal client-pod-1 -- curl 10.244.2.9
<!DOCTYPE html>
<html>
<head>
<title>Welcome to nginx!</title>
<style>

第六题:创建 svc

# 预配置创建一个 deployment 名称为 front-end。
user1@k8s-master:~/cka-lab-init/6$ cat front-end.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: front-end
spec:
  replicas: 1
  selector:
    matchLabels:
      app: nginx-front-end
  template:
    metadata:
      labels:
        app: nginx-front-end
    spec:
      containers:
      - name: nginx
        image: nginx:1.18.0
        imagePullPolicy: IfNotPresent
 
# 部署 front-end。
user1@k8s-master:~/cka-lab-init/6$ kubectl apply -f front-end.yaml
deployment.apps/front-end created
 
### 查看 front-end。
user1@k8s-master:~/cka-lab-init/6$ kubectl get deployments front-end
NAME        READY   UP-TO-DATE   AVAILABLE   AGE
front-end   1/1     1            1           63s
 
### 查看 pod。
user1@k8s-master:~/cka-lab-init/6$ kubectl get pod front-end-7bd47f9bb4-vhlx2 -o wide
NAME                         READY   STATUS    RESTARTS   AGE   IP            NODE         NOMINATED NODE   READINESS GATES
front-end-7bd47f9bb4-vhlx2   1/1     Running   0          39s   10.244.2.11   k8s-node-2   <none>           <none>

第七题:创建 ingress 资源

# 预配置创建一个 ing-internal 的 namespace
# 编写 namespace 资源清单
user1@k8s-master:~/cka-lab-init/7$ sudo cat ns-ing-internal.yaml
apiVersion: v1
kind: Namespace
metadata:
  name: ing-internal
  
# 部署 namespace 资源
user1@k8s-master:~/cka-lab-init/7$ kubectl apply -f ns-ing-internal.yaml
namespace/ing-internal created
 
# 编写 ingress 控制器的资源清单
user1@k8s-master:~/cka-lab-init/7$ sudo cat deploy.yaml
 
apiVersion: v1
kind: Namespace
metadata:
  name: ingress-nginx
  labels:
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
 
---
# Source: ingress-nginx/templates/controller-serviceaccount.yaml
apiVersion: v1
kind: ServiceAccount
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx
  namespace: ingress-nginx
---
# Source: ingress-nginx/templates/controller-configmap.yaml
apiVersion: v1
kind: ConfigMap
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx-controller
  namespace: ingress-nginx
data:
---
# Source: ingress-nginx/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
  name: ingress-nginx
  namespace: ingress-nginx
rules:
  - apiGroups:
      - ''
    resources:
      - configmaps
      - endpoints
      - nodes
      - pods
      - secrets
    verbs:
      - list
      - watch
  - apiGroups:
      - ''
    resources:
      - nodes
    verbs:
      - get
  - apiGroups:
      - ''
    resources:
      - services
    verbs:
      - get
      - list
      - update
      - watch
  - apiGroups:
      - extensions
      - networking.k8s.io   # k8s 1.14+
    resources:
      - ingresses
    verbs:
      - get
      - list
      - watch
  - apiGroups:
      - ''
    resources:
      - events
    verbs:
      - create
      - patch
  - apiGroups:
      - extensions
      - networking.k8s.io   # k8s 1.14+
    resources:
      - ingresses/status
    verbs:
      - update
  - apiGroups:
      - networking.k8s.io   # k8s 1.14+
    resources:
      - ingressclasses
    verbs:
      - get
      - list
      - watch
---
# Source: ingress-nginx/templates/clusterrolebinding.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
  name: ingress-nginx
  namespace: ingress-nginx
roleRef:
  apiGroup: rbac.authorization.k8s.io
  kind: ClusterRole
  name: ingress-nginx
subjects:
  - kind: ServiceAccount
    name: ingress-nginx
    namespace: ingress-nginx
---
# Source: ingress-nginx/templates/controller-role.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx
  namespace: ingress-nginx
rules:
  - apiGroups:
      - ''
    resources:
      - namespaces
    verbs:
      - get
  - apiGroups:
      - ''
    resources:
      - configmaps
      - pods
      - secrets
      - endpoints
    verbs:
      - get
      - list
      - watch
  - apiGroups:
      - ''
    resources:
      - services
    verbs:
      - get
      - list
      - update
      - watch
  - apiGroups:
      - extensions
      - networking.k8s.io   # k8s 1.14+
    resources:
      - ingresses
    verbs:
      - get
      - list
      - watch
  - apiGroups:
      - extensions
      - networking.k8s.io   # k8s 1.14+
    resources:
      - ingresses/status
    verbs:
      - update
  - apiGroups:
      - networking.k8s.io   # k8s 1.14+
    resources:
      - ingressclasses
    verbs:
      - get
      - list
      - watch
  - apiGroups:
      - ''
    resources:
      - configmaps
    resourceNames:
      - ingress-controller-leader-nginx
    verbs:
      - get
      - update
  - apiGroups:
      - ''
    resources:
      - configmaps
    verbs:
      - create
  - apiGroups:
      - ''
    resources:
      - endpoints
    verbs:
      - create
      - get
      - update
  - apiGroups:
      - ''
    resources:
      - events
    verbs:
      - create
      - patch
---
# Source: ingress-nginx/templates/controller-rolebinding.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx
  namespace: ingress-nginx
roleRef:
  apiGroup: rbac.authorization.k8s.io
  kind: Role
  name: ingress-nginx
subjects:
  - kind: ServiceAccount
    name: ingress-nginx
    namespace: ingress-nginx
---
# Source: ingress-nginx/templates/controller-service-webhook.yaml
apiVersion: v1
kind: Service
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx-controller-admission
  namespace: ingress-nginx
spec:
  type: ClusterIP
  ports:
    - name: https-webhook
      port: 443
      targetPort: webhook
  selector:
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/component: controller
---
# Source: ingress-nginx/templates/controller-service.yaml
apiVersion: v1
kind: Service
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx-controller
  namespace: ingress-nginx
spec:
  type: NodePort
  ports:
    - name: http
      port: 80
      protocol: TCP
      targetPort: http
    - name: https
      port: 443
      protocol: TCP
      targetPort: https
  selector:
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/component: controller
---
# Source: ingress-nginx/templates/controller-deployment.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: controller
  name: ingress-nginx-controller
  namespace: ingress-nginx
spec:
  replicas: 1
  selector:
    matchLabels:
      app.kubernetes.io/name: ingress-nginx
      app.kubernetes.io/instance: ingress-nginx
      app.kubernetes.io/component: controller
  revisionHistoryLimit: 10
  minReadySeconds: 0
  template:
    metadata:
      labels:
        app.kubernetes.io/name: ingress-nginx
        app.kubernetes.io/instance: ingress-nginx
        app.kubernetes.io/component: controller
    spec:
      dnsPolicy: ClusterFirst
      containers:
        - name: controller
          image: quay.io/kubernetes-ingress-controller/nginx-ingress-controller:0.33.0
          imagePullPolicy: IfNotPresent
          lifecycle:
            preStop:
              exec:
                command:
                  - /wait-shutdown
          args:
            - /nginx-ingress-controller
            - --election-id=ingress-controller-leader
            - --ingress-class=nginx
            - --configmap=ingress-nginx/ingress-nginx-controller
            - --validating-webhook=:8443
            - --validating-webhook-certificate=/usr/local/certificates/cert
            - --validating-webhook-key=/usr/local/certificates/key
          securityContext:
            capabilities:
              drop:
                - ALL
              add:
                - NET_BIND_SERVICE
            runAsUser: 101
            allowPrivilegeEscalation: true
          env:
            - name: POD_NAME
              valueFrom:
                fieldRef:
                  fieldPath: metadata.name
            - name: POD_NAMESPACE
              valueFrom:
                fieldRef:
                  fieldPath: metadata.namespace
          livenessProbe:
            httpGet:
              path: /healthz
              port: 10254
              scheme: HTTP
            initialDelaySeconds: 10
            periodSeconds: 10
            timeoutSeconds: 1
            successThreshold: 1
            failureThreshold: 3
          readinessProbe:
            httpGet:
              path: /healthz
              port: 10254
              scheme: HTTP
            initialDelaySeconds: 10
            periodSeconds: 10
            timeoutSeconds: 1
            successThreshold: 1
            failureThreshold: 3
          ports:
            - name: http
              containerPort: 80
              protocol: TCP
            - name: https
              containerPort: 443
              protocol: TCP
            - name: webhook
              containerPort: 8443
              protocol: TCP
          volumeMounts:
            - name: webhook-cert
              mountPath: /usr/local/certificates/
              readOnly: true
          resources:
            requests:
              cpu: 100m
              memory: 90Mi
      serviceAccountName: ingress-nginx
      terminationGracePeriodSeconds: 300
      volumes:
        - name: webhook-cert
          secret:
            secretName: ingress-nginx-admission
---
# Source: ingress-nginx/templates/admission-webhooks/validating-webhook.yaml
apiVersion: admissionregistration.k8s.io/v1beta1
kind: ValidatingWebhookConfiguration
metadata:
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  name: ingress-nginx-admission
  namespace: ingress-nginx
webhooks:
  - name: validate.nginx.ingress.kubernetes.io
    rules:
      - apiGroups:
          - extensions
          - networking.k8s.io
        apiVersions:
          - v1beta1
        operations:
          - CREATE
          - UPDATE
        resources:
          - ingresses
    failurePolicy: Fail
    clientConfig:
      service:
        namespace: ingress-nginx
        name: ingress-nginx-controller-admission
        path: /extensions/v1beta1/ingresses
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
  name: ingress-nginx-admission
  annotations:
    helm.sh/hook: pre-install,pre-upgrade,post-install,post-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
rules:
  - apiGroups:
      - admissionregistration.k8s.io
    resources:
      - validatingwebhookconfigurations
    verbs:
      - get
      - update
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/clusterrolebinding.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
  name: ingress-nginx-admission
  annotations:
    helm.sh/hook: pre-install,pre-upgrade,post-install,post-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
roleRef:
  apiGroup: rbac.authorization.k8s.io
  kind: ClusterRole
  name: ingress-nginx-admission
subjects:
  - kind: ServiceAccount
    name: ingress-nginx-admission
    namespace: ingress-nginx
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/job-createSecret.yaml
apiVersion: batch/v1
kind: Job
metadata:
  name: ingress-nginx-admission-create
  annotations:
    helm.sh/hook: pre-install,pre-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
spec:
  template:
    metadata:
      name: ingress-nginx-admission-create
      labels:
        helm.sh/chart: ingress-nginx-2.4.0
        app.kubernetes.io/name: ingress-nginx
        app.kubernetes.io/instance: ingress-nginx
        app.kubernetes.io/version: 0.33.0
        app.kubernetes.io/managed-by: Helm
        app.kubernetes.io/component: admission-webhook
    spec:
      containers:
        - name: create
          image: jettech/kube-webhook-certgen:v1.2.0
          imagePullPolicy: IfNotPresent
          args:
            - create
            - --host=ingress-nginx-controller-admission,ingress-nginx-controller-admission.ingress-nginx.svc
            - --namespace=ingress-nginx
            - --secret-name=ingress-nginx-admission
      restartPolicy: OnFailure
      serviceAccountName: ingress-nginx-admission
      securityContext:
        runAsNonRoot: true
        runAsUser: 2000
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/job-patchWebhook.yaml
apiVersion: batch/v1
kind: Job
metadata:
  name: ingress-nginx-admission-patch
  annotations:
    helm.sh/hook: post-install,post-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
spec:
  template:
    metadata:
      name: ingress-nginx-admission-patch
      labels:
        helm.sh/chart: ingress-nginx-2.4.0
        app.kubernetes.io/name: ingress-nginx
        app.kubernetes.io/instance: ingress-nginx
        app.kubernetes.io/version: 0.33.0
        app.kubernetes.io/managed-by: Helm
        app.kubernetes.io/component: admission-webhook
    spec:
      containers:
        - name: patch
          image: jettech/kube-webhook-certgen:v1.2.0
          imagePullPolicy: IfNotPresent
          args:
            - patch
            - --webhook-name=ingress-nginx-admission
            - --namespace=ingress-nginx
            - --patch-mutating=false
            - --secret-name=ingress-nginx-admission
            - --patch-failure-policy=Fail
      restartPolicy: OnFailure
      serviceAccountName: ingress-nginx-admission
      securityContext:
        runAsNonRoot: true
        runAsUser: 2000
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/role.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
  name: ingress-nginx-admission
  annotations:
    helm.sh/hook: pre-install,pre-upgrade,post-install,post-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
rules:
  - apiGroups:
      - ''
    resources:
      - secrets
    verbs:
      - get
      - create
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/rolebinding.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
  name: ingress-nginx-admission
  annotations:
    helm.sh/hook: pre-install,pre-upgrade,post-install,post-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
roleRef:
  apiGroup: rbac.authorization.k8s.io
  kind: Role
  name: ingress-nginx-admission
subjects:
  - kind: ServiceAccount
    name: ingress-nginx-admission
    namespace: ingress-nginx
---
# Source: ingress-nginx/templates/admission-webhooks/job-patch/serviceaccount.yaml
apiVersion: v1
kind: ServiceAccount
metadata:
  name: ingress-nginx-admission
  annotations:
    helm.sh/hook: pre-install,pre-upgrade,post-install,post-upgrade
    helm.sh/hook-delete-policy: before-hook-creation,hook-succeeded
  labels:
    helm.sh/chart: ingress-nginx-2.4.0
    app.kubernetes.io/name: ingress-nginx
    app.kubernetes.io/instance: ingress-nginx
    app.kubernetes.io/version: 0.33.0
    app.kubernetes.io/managed-by: Helm
    app.kubernetes.io/component: admission-webhook
  namespace: ingress-nginx
 
# 创建 ingress 控制器。
user1@k8s-master:~/cka-lab-init/7$ kubectl apply -f deploy.yaml
 
# 查看创建的 SVC。
user1@k8s-master:~/cka-lab-init/7$ kubectl get svc -n ingress-nginx
NAME                                 TYPE        CLUSTER-IP      EXTERNAL-IP   PORT(S)                      AGE
ingress-nginx-controller             NodePort    10.103.24.70    <none>        80:32401/TCP,443:30961/TCP   43s
ingress-nginx-controller-admission   ClusterIP   10.101.98.116   <none>        443/TCP                      43s
 
### 查看创建的 pod。
user1@k8s-master:~/cka-lab-init/7$ kubectl get pod -o wide -n ingress-nginx
NAME                                        READY   STATUS      RESTARTS   AGE     IP            NODE         NOMINATED NODE   READINESS GATES
ingress-nginx-admission-create-tz7jq        0/1     Completed   0          6m15s   10.244.2.12   k8s-node-2   <none>           <none>
ingress-nginx-admission-patch-7lklk         0/1     Completed   0          6m15s   10.244.2.13   k8s-node-2   <none>           <none>
ingress-nginx-controller-7fcdd96bcf-hrfg4   1/1     Running     0          6m25s   10.244.2.16   k8s-node-2   <none>           <none>
 
### 查看 ingress-controller 的 endpoint。
user1@k8s-master:~/cka-lab-init/7$ kubectl get endpoints -n ingress-nginx
NAME                                 ENDPOINTS                        AGE
ingress-nginx-controller             10.244.2.16:443,10.244.2.16:80   6m54s
ingress-nginx-controller-admission   10.244.2.16:8443                 6m54s
 
### 查看 ingress-controller 的详细信息。
user1@k8s-master:~/cka-lab-init/7$ kubectl describe endpoints -n ingress-nginx ingress-nginx-controller
Name:         ingress-nginx-controller
Namespace:    ingress-nginx
Labels:       app.kubernetes.io/component=controller
              app.kubernetes.io/instance=ingress-nginx
              app.kubernetes.io/managed-by=Helm
              app.kubernetes.io/name=ingress-nginx
              app.kubernetes.io/version=0.33.0
              helm.sh/chart=ingress-nginx-2.4.0
Annotations:  endpoints.kubernetes.io/last-change-trigger-time: 2021-01-09T14:02:16Z
Subsets:
  Addresses:          10.244.2.16
  NotReadyAddresses:  <none>
  Ports:
    Name   Port  Protocol
    ----   ----  --------
    https  443   TCP
    http   80    TCP
 
### 在 k8s-node-2 上创建文件
user1@k8s-node-2:/tmp/7/html$ pwd
/tmp/7/html
user1@k8s-node-2:/tmp/7/html$ cat index.html
hi
 
# 编写 SVC 为 hi 的资源清单。指定在 k8s-node-2 节点,因为 hostpath 存储在 node-2 节点。
user1@k8s-master:~/cka-lab-init/7$ sudo cat nginx-app-service.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: nginx-app
  namespace: ing-internal
spec:
  replicas: 1
  selector:
    matchLabels:
      app: nginx-app
  template:
    metadata:
      labels:
        app: nginx-app
    spec:
      nodeName: k8s-node-2
      containers:
      - name: nginx
        image: nginx:1.18.0
        imagePullPolicy: IfNotPresent
        volumeMounts:
        - name: html-hi
          mountPath: /usr/share/nginx/html/
      volumes:
      - name: html-hi
        hostPath:
          path: /tmp/7/html
---
apiVersion: v1
kind: Service
metadata:
  name: hi
  namespace: ing-internal
spec:
  selector:
    app: nginx-app
  ports:
  - name: web
    port: 5678
    targetPort: 80
   
# 部署 SVC 和 pod
user1@k8s-master:~/cka-lab-init/7$ kubectl apply -f nginx-app-service.yaml
deployment.apps/nginx-app created
service/hi created
 
# 查看部署的 SVC。
user1@k8s-master:~/cka-lab-init/7$ kubectl get svc -n ing-internal
NAME   TYPE        CLUSTER-IP     EXTERNAL-IP   PORT(S)    AGE
hi     ClusterIP   10.96.192.19   <none>        5678/TCP   40s
 
### 查看部署的 pod。
user1@k8s-master:~/cka-lab-init/7$ kubectl get pod -n ing-internal -o wide
NAME                         READY   STATUS    RESTARTS   AGE   IP            NODE         NOMINATED NODE   READINESS GATES
nginx-app-778d9f8cbc-kfvxm   1/1     Running   0          62s   10.244.2.17   k8s-node-2   <none>           <none>
 
# 访问 SVC 和 pod
user1@k8s-master:~/cka-lab-init/7$ sudo curl 10.96.192.19:5678 
hi
user1@k8s-master:~/cka-lab-init/7$ sudo curl 10.244.2.17
hi

第八题:扩展 deployment

# 预配置要创建一个名为 webserver 的 deployment
user1@k8s-master:~/cka-lab-init/8$ sudo cat deploy-webserver.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: webserver
spec:
  replicas: 1
  selector:
    matchLabels:
      app: webserver
  template:
    metadata:
      labels:
        app: webserver
    spec:
      containers:
      - name: nginx
        image: nginx:1.18.0
        imagePullPolicy: IfNotPresent
 
# 部署 webserver 应用。
user1@k8s-master:~/cka-lab-init/8$ kubectl apply -f deploy-webserver.yaml
deployment.apps/webserver created
 
# 查看 webserver 的 pod 数量。
user1@k8s-master:~/cka-lab-init/8$ kubectl get pod webserver-7b7cfc759-7gk5r
NAME                        READY   STATUS    RESTARTS   AGE
webserver-7b7cfc759-7gk5r   1/1     Running   0          22s

第九题:将 pod 部署到指定 node 节点上

# 预配置给节点打标签
user1@k8s-master:~$ kubectl label nodes k8s-node-2 disk=ssd
node/k8s-node-2 labeled
 
# 查看标签
user1@k8s-master:~$ kubectl get nodes --show-labels | grep ssd
k8s-node-2   Ready                      <none>   4h27m   v1.20.0   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,disk=ssd,kubernetes.io/arch=amd64,kubernetes.io/hostname=k8s-node-2,kubernetes.io/os=linux

第十题:检查有多少 node 节点是健康状态


# 预配置目录和文件。
user1@k8s-master:~$ sudo mkdir /opt/KUSC00402/
user1@k8s-master:~$ sudo touch /opt/KUSC00402/kusc00402.txt
user1@k8s-master:~$ ls /opt/KUSC00402/kusc00402.txt
/opt/KUSC00402/kusc00402.txt

第十一题:创建多个 container 的 Pod

第十二题:创建 Persistent Volume


# 预配置目录 /srv/app-data
user1@k8s-master:~$ sudo mkdir -p /srv/app-data
user1@k8s-master:~$ ls -d /srv/app-data/
/srv/app-data/

 

第十三题:创建 PVC

# 预配置 storageClass
# 创建NFS服务器。
user1@k8s-master:~$ sudo apt-get install -y nfs-kernel-server
 
# 配置NFS文件共享。
user1@k8s-master:~$ sudo mkdir /nfs-server
user1@k8s-master:~$ cat /etc/exports
/nfs-server *(rw,sync,no_root_squash)
user1@k8s-master:~$ sudo chmod 700 /nfs-server/
user1@k8s-master:~$ sudo service nfs-kernel-server restart
user1@k8s-master:~$ sudo service nfs-kernel-server status
 
# 验证NFS服务。
user1@k8s-master:~$ sudo showmount -e 127.0.0.1
Export list for 127.0.0.1:
/nfs-server *
 
# 在 node2 节点也要安装 nfs client 软件。并且测试一下 nfs 存储。
user1@k8s-node-2:~$ sudo apt-get install -y nfs-kernel-server
user1@k8s-node-2:~$ sudo showmount -e k8s-master
Export list for k8s-master:
/nfs-server *
 
# 创建ServiceAccount解决权限问题。
# 编写 RBAC 资源清单文件
user1@k8s-master:~/cka-lab-init/13$ sudo cat rbac.yaml
apiVersion: v1
kind: ServiceAccount
metadata:
  name: nfs-client-provisioner
  namespace: default
---
kind: ClusterRole
apiVersion: rbac.authorization.k8s.io/v1
metadata:
  name: nfs-client-provisioner-runner
rules:
  - apiGroups: [""]
    resources: ["persistentvolumes"]
    verbs: ["get", "list", "watch", "create", "delete"]
  - apiGroups: [""]
    resources: ["persistentvolumeclaims"]
    verbs: ["get", "list", "watch", "update"]
  - apiGroups: ["storage.k8s.io"]
    resources: ["storageclasses"]
    verbs: ["get", "list", "watch"]
  - apiGroups: [""]
    resources: ["events"]
    verbs: ["create", "update", "patch"]
---
kind: ClusterRoleBinding
apiVersion: rbac.authorization.k8s.io/v1
metadata:
  name: run-nfs-client-provisioner
subjects:
  - kind: ServiceAccount
    name: nfs-client-provisioner
    namespace: default
roleRef:
  kind: ClusterRole
  name: nfs-client-provisioner-runner
  apiGroup: rbac.authorization.k8s.io
---
kind: Role
apiVersion: rbac.authorization.k8s.io/v1
metadata:
  name: leader-locking-nfs-client-provisioner
  namespace: default
rules:
  - apiGroups: [""]
    resources: ["endpoints"]
    verbs: ["get", "list", "watch", "create", "update", "patch"]
---
kind: RoleBinding
apiVersion: rbac.authorization.k8s.io/v1
metadata:
  name: leader-locking-nfs-client-provisioner
subjects:
  - kind: ServiceAccount
    name: nfs-client-provisioner
    namespace: default
roleRef:
  kind: Role
  name: leader-locking-nfs-client-provisioner
  apiGroup: rbac.authorization.k8s.io
 
# 创建 RBAC 资源
user1@k8s-master:~/cka-lab-init/13$ kubectl apply -f rbac.yaml
serviceaccount/nfs-client-provisioner created
clusterrole.rbac.authorization.k8s.io/nfs-client-provisioner-runner created
clusterrolebinding.rbac.authorization.k8s.io/run-nfs-client-provisioner created
role.rbac.authorization.k8s.io/leader-locking-nfs-client-provisioner created
rolebinding.rbac.authorization.k8s.io/leader-locking-nfs-client-provisioner created
 
# 编写 nfs-client 的资源清单文件
user1@k8s-master:~/cka-lab-init/13$ sudo cat nfs-client-provisioner.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: nfs-client-provisioner
  labels:
    app: nfs-client-provisioner
  namespace: default
spec:
  replicas: 1
  selector:
    matchLabels:
      app: nfs-client-provisioner
  strategy:
    type: Recreate
  selector:
    matchLabels:
      app: nfs-client-provisioner
  template:
    metadata:
      labels:
        app: nfs-client-provisioner
    spec:
      serviceAccountName: nfs-client-provisioner
      containers:
        - name: nfs-client-provisioner
          image: quay.io/external_storage/nfs-client-provisioner:latest
          volumeMounts:
            - name: nfs-client-root
              mountPath: /persistentvolumes
          env:
            - name: PROVISIONER_NAME
              value: nfs-storage
            - name: NFS_SERVER
              value: k8s-master
            - name: NFS_PATH 
              value: /nfs-server
      volumes:
        - name: nfs-client-root
          nfs:
            server: k8s-master
            path: /nfs-server
 
# 创建 nfs-client 资源
user1@k8s-master:~/cka-lab-init/13$ kubectl apply -f nfs-client-provisioner.yaml
deployment.apps/nfs-client-provisioner created
 
# 查看 nfs-client 的 pod
user1@k8s-master:~/cka-lab-init/13$ kubectl get pod nfs-client-provisioner-6546c4b76-xj42f
NAME                                     READY   STATUS    RESTARTS   AGE
nfs-client-provisioner-6546c4b76-xj42f   1/1     Running   0          30s
 
# 编写 storageclass 资源清单
user1@k8s-master:~/cka-lab-init/13$ sudo cat managed-nfs-storageclass.yaml
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
  name: csi-hostpath-sc
provisioner: nfs-storage
allowVolumeExpansion: true
 
# 创建 storageclass 资源
user1@k8s-master:~/cka-lab-init/13$ kubectl apply -f managed-nfs-storageclass.yaml
storageclass.storage.k8s.io/csi-hostpath-sc created
 
# 查看 storageclass 资源
user1@k8s-master:~/cka-lab-init/13$ kubectl get storageclasses.storage.k8s.io
NAME              PROVISIONER   RECLAIMPOLICY   VOLUMEBINDINGMODE   ALLOWVOLUMEEXPANSION   AGE
csi-hostpath-sc   nfs-storage   Delete          Immediate           true                   35s

第十四题:监控 pod 的日志

# 预配置
# 编写 pod 为 foobar 的资源清单
user1@k8s-master:~/cka-lab-init/14$ sudo cat foobar-pod.yaml
apiVersion: v1
kind: Pod
metadata:
  name: foobar
  labels:
    app: foobar
spec:
  containers:
  - name: nginx
    image: nginx
    imagePullPolicy: IfNotPresent
    ports:
    - containerPort: 80
 
# 创建 pod 资源
user1@k8s-master:~/cka-lab-init/14$ kubectl apply -f foobar-pod.yaml
pod/foobar created
 
# 查看 pod。
user1@k8s-master:~/cka-lab-init/14$ kubectl get pod foobar
NAME     READY   STATUS    RESTARTS   AGE
foobar   1/1     Running   0          23s
 
# 创建 /opt/KUTR00101
user1@k8s-master:~$ sudo mkdir /opt/KUTR00101

第十五题:添加 sidecar container

# 编写 pod 资源清单
user1@k8s-master:~/cka-lab-init/15$ sudo cat big-corp-app.yaml
apiVersion: v1
kind: Pod
metadata:
  name: big-corp-app
spec:
  containers:
  - name: count
    image: busybox
    args:
    - /bin/sh
    - -c
    - >
      i=0;
      while true;
      do
        echo "$i: $(date)" >> /var/log/big-corp-app.log;
        sleep 1;
      done
    volumeMounts:
    - name: varlog
      mountPath: /var/log
  volumes:
  - name: varlog
    emptyDir: {}
 
### 创建 pod
user1@k8s-master:~/cka-lab-init/15$ kubectl apply -f big-corp-app.yaml
pod/big-corp-app created
 
### 查看 pod
user1@k8s-master:~/cka-lab-init/15$ kubectl get pod
NAME                                     READY   STATUS    RESTARTS   AGE
big-corp-app                             1/1     Running   0          23s

第十六题:查看最高 CPU 使用率的 Pod

# 预配置 metrics-server
# 1、下载yaml文件。
user1@k8s-master:~/cka-lab-init/16$ sudo wget https://github.com/kubernetes-sigs/metrics-server/releases/download/v0.3.6/components.yaml
 
# 2、修改编码yaml文件。增加如下命令。
user1@k8s-master:~/cka-lab-init/16$ sudo vim components.yaml
......
kind: Deployment
    command:
    - /metrics-server
    - --kubelet-preferred-address-types=InternalIP
    - --kubelet-insecure-tls
    registry.aliyuncs.com/google_containers  #修改源
 
# 3、修改api-server的yaml文件。
user1@k8s-master:~/cka-lab-init/16$ sudo vim /etc/kubernetes/manifests/kube-apiserver.yaml
......
spec:
  containers:
  - command
    - --enable-aggregator-routing=true
     
# 4、创建metrics-server,使用yaml文件。
user1@k8s-master:~/cka-lab-init/16$ kubectl apply -f components.yaml
 
# 5、查看metrics-server的deployments。
user1@k8s-master:~/cka-lab-init/16$ kubectl get deployments.apps -n kube-system metrics-server
NAME             READY   UP-TO-DATE   AVAILABLE   AGE
metrics-server   1/1     1            1           28s
 
# 6、查看metrics-server的pod。
user1@k8s-master:~/cka-lab-init/16$ kubectl get pod -n kube-system metrics-server-5f5f46797c-l9vs8
NAME                              READY   STATUS    RESTARTS   AGE
metrics-server-5f5f46797c-l9vs8   1/1     Running   0          57s
 
# 7、查看api是否已添加metrics。
user1@k8s-master:~/cka-lab-init/16$ kubectl api-versions | grep metrics
metrics.k8s.io/v1beta1
 
# 8、检查监控指标的可用性。
user1@k8s-master:~/cka-lab-init/16$ kubectl get --raw "/apis/metrics.k8s.io/v1beta1/nodes"
user1@k8s-master:~/cka-lab-init/16$ kubectl get --raw "/apis/metrics.k8s.io/v1beta1/pods"
 
# 9、监控pod的资源信息。
user1@k8s-master:~/cka-lab-init/16$ kubectl top pod
NAME                                     CPU(cores)   MEMORY(bytes)  
big-corp-app                             1m           1Mi                             
 
# 编写一个 pod 资源清单,标签为 name=cpu-user
user1@k8s-master:~/cka-lab-init/16$ sudo cat cpu-pod.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: cpu-pod
spec:
  replicas: 3
  selector:
    matchLabels:
      name: cpu-user
  template:
    metadata:
      labels:
        name: cpu-user
    spec:
      containers:
      - name: nginx
        image: nginx:1.18.0
        imagePullPolicy: IfNotPresent
 
# 创建 pod 资源
user1@k8s-master:~/cka-lab-init/16$ kubectl apply -f cpu-pod.yaml
deployment.apps/cpu-pod created
 
# 查看标签为 name=cpu-user 的 pod
user1@k8s-master:~/cka-lab-init/16$ kubectl get pod -l name=cpu-user -A
NAMESPACE   NAME                       READY   STATUS    RESTARTS   AGE
default     cpu-pod-7bfb667b98-8v2gx   1/1     Running   0          21s
default     cpu-pod-7bfb667b98-qjfmm   1/1     Running   0          21s
default     cpu-pod-7bfb667b98-w6mm6   1/1     Running   0          21s
 
# 创建目录/opt/KUTR00401
user1@k8s-master:~$ sudo mkdir /opt/KUTR00401

第十七题:集群故障排查

# 预配 关闭 node-3 节点的 kubelet
# 查看 kubelet 状态
user1@k8s-node-3:~$ sudo systemctl status kubelet.service
● kubelet.service - kubelet: The Kubernetes Node Agent
   Loaded: loaded (/lib/systemd/system/kubelet.service; enabled; vendor preset: enabled)
  Drop-In: /etc/systemd/system/kubelet.service.d
           └─10-kubeadm.conf
   Active: active (running) since Sat 2021-01-09 18:05:24 CST; 5h 9min ago
 
# 关闭 kubelet
user1@k8s-node-3:~$ sudo systemctl stop kubelet.service
 
# 查看节点状态
user1@k8s-master:~$ kubectl get node
NAME         STATUS                     ROLES    AGE     VERSION
k8s-master   Ready                      master   5h39m   v1.19.0
k8s-node-1   Ready,SchedulingDisabled   <none>   5h24m   v1.18.8
k8s-node-2   Ready                      <none>   5h24m   v1.18.8
k8s-node-3   NotReady                   <none>   5h24m   v1.18.8

 

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

云原生运维社区

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值