k8s 部署redis

参考:

K8S搭建Redis集群_172.31.0.82-CSDN博客

Docker国内设置下载镜像 - 简书

  • 首先创建nfs存储,当redis迁移之后,仍然能够从nfs中获取之前的数据

需要在机器上安装nfs服务,这里我已经安装好了,可以参考这篇文章k8s Persistent Volume-CSDN博客进行安装

  • 之后设置共享的路径

$ sudo mkdir -p /usr/local/kubernetes/redis/pv{1..6}
$ sudo chmod 777 /usr/local/kubernetes/redis/pv{1..6}

$ ls -la /usr/local/kubernetes/redis
total 32
drwxr-xr-x 8 root root 4096 Aug  7 15:54 .
drwxr-xr-x 3 root root 4096 Aug  7 15:54 ..
drwxrwxrwx 2 root root 4096 Aug  7 15:54 pv1
drwxrwxrwx 2 root root 4096 Aug  7 15:54 pv2
drwxrwxrwx 2 root root 4096 Aug  7 15:54 pv3
drwxrwxrwx 2 root root 4096 Aug  7 15:54 pv4
drwxrwxrwx 2 root root 4096 Aug  7 15:54 pv5
drwxrwxrwx 2 root root 4096 Aug  7 15:54 pv6
  • 然后设置这些目录的共享

编辑/etc/exports

$ sudo vi /etc/exports


插入以下代码
#================================================================================================
/usr/local/kubernetes/redis/pv1 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv2 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv3 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv4 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv5 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv6 *(rw,sync,no_root_squash)
#================================================================================================
~

之后重启服务

$ sudo service nfs-kernel-server restart
$ sudo systemctl restart rpcbind

然后通过命令查看共享情况

$ showmount -e
Export list for xxxxxxx
/usr/local/kubernetes/redis/pv6 *
/usr/local/kubernetes/redis/pv5 *
/usr/local/kubernetes/redis/pv4 *
/usr/local/kubernetes/redis/pv3 *
/usr/local/kubernetes/redis/pv2 *
/usr/local/kubernetes/redis/pv1 *


$ sudo exportfs -v
/usr/local/kubernetes/redis/pv1
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv2
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv3
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv4
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv5
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv6
                <world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
  • 基于上述的nfs共享目录,创建pv

$ cat redis-pv.yaml
apiVersion: v1
kind: PersistentVolume
metadata:
  name: nfs-pv1
spec:
  storageClassName: redis
  capacity:
    storage: 1Gi
  accessModes:
    - ReadWriteMany
  nfs:
    server: xxxxxxx
    path: "/usr/local/kubernetes/redis/pv1"

---
apiVersion: v1
kind: PersistentVolume
metadata:
  name: nfs-pv2
spec:
  storageClassName: redis
  capacity:
    storage: 1Gi
  accessModes:
    - ReadWriteMany
  nfs:
    server: xxxxxxx
    path: "/usr/local/kubernetes/redis/pv2"

---
apiVersion: v1
kind: PersistentVolume
metadata:
  name: nfs-pv3
spec:
  storageClassName: redis
  capacity:
    storage: 1Gi
  accessModes:
    - ReadWriteMany
  nfs:
    server: xxxxxxx
    path: "/usr/local/kubernetes/redis/pv3"

---
apiVersion: v1
kind: PersistentVolume
metadata:
  name: nfs-pv4
spec:
  storageClassName: redis
  capacity:
    storage: 1Gi
  accessModes:
    - ReadWriteMany
  nfs:
    server: xxxxxxx
    path: "/usr/local/kubernetes/redis/pv4"

---
apiVersion: v1
kind: PersistentVolume
metadata:
  name: nfs-pv5
spec:
  storageClassName: redis
  capacity:
    storage: 1Gi
  accessModes:
    - ReadWriteMany
  nfs:
    server: xxxxxxx
    path: "/usr/local/kubernetes/redis/pv5"

---
apiVersion: v1
kind: PersistentVolume
metadata:
  name: nfs-pv6
spec:
  storageClassName: redis
  capacity:
    storage: 1Gi
  accessModes:
    - ReadWriteMany
  nfs:
    server: xxxxxxx
    path: "/usr/local/kubernetes/redis/pv6"

 然后创建pv

$ kubectl create -f redis-pv.yaml
persistentvolume/nfs-pv1 created
persistentvolume/nfs-pv2 created
persistentvolume/nfs-pv3 created
persistentvolume/nfs-pv4 created
persistentvolume/nfs-pv5 created
persistentvolume/nfs-pv6 created


$ kubectl get pv
NAME      CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS      CLAIM          STORAGECLASS   REASON   AGE
nfs-pv1   1Gi        RWX            Retain           Available                  redis                   7s
nfs-pv2   1Gi        RWX            Retain           Available                  redis                   7s
nfs-pv3   1Gi        RWX            Retain           Available                  redis                   6s
nfs-pv4   1Gi        RWX            Retain           Available                  redis                   6s
nfs-pv5   1Gi        RWX            Retain           Available                  redis                   6s
nfs-pv6   1Gi        RWX            Retain           Available                  redis                   6s
pv1       1Gi        RWO            Retain           Bound       default/pvc1   nfs                     2d

  • 创建configMap并将其转化为configmap文件

$ cat redis.conf
protected-mode no
appendonly yes
cluster-enabled yes
cluster-config-file /var/lib/redis/nodes.conf
cluster-node-timeout 5000
dir /var/lib/redis
port 6379


#将reids.conf转为configMap文件
kubectl create configmap redis-conf --from-file=redis.conf

$ kubectl get cm
NAME               DATA   AGE
kube-root-ca.crt   1      47d
redis-conf         1      68m


$ kubectl get cm
NAME               DATA   AGE
kube-root-ca.crt   1      47d
redis-conf         1      68m
kevin@kevin-pc:~/deploy-tests/redis$ kubectl describe  cm redis-conf
Name:         redis-conf
Namespace:    default
Labels:       <none>
Annotations:  <none>

Data
====
redis.conf:
----
protected-mode no
appendonly yes
cluster-enabled yes
cluster-config-file /var/lib/redis/nodes.conf
cluster-node-timeout 5000
dir /var/lib/redis
port 6379


BinaryData
====

Events:  <none>
  •  创建redis的service

$ cat redis-service.yaml
apiVersion: v1
kind: Service
metadata:
  name: redis-service
  labels:
    app: redis
spec:
  ports:
  - name: redis-port
    port: 6379
  clusterIP: None
  selector:
    app: redis
    appCluster: redis-cluster


$ kubectl create -f redis-service.yaml

$ kubectl get svc
NAME                TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)        AGE
redis-service       ClusterIP   None             <none>        6379/TCP       67m

$ kubectl describe svc redis-service
Name:              redis-service
Namespace:         default
Labels:            app=redis
Annotations:       <none>
Selector:          app=redis,appCluster=redis-cluster
Type:              ClusterIP
IP Family Policy:  SingleStack
IP Families:       IPv4
IP:                None
IPs:               None
Port:              redis-port  6379/TCP
TargetPort:        6379/TCP
Endpoints:         <none>
Session Affinity:  None
Events:            <none>

  •  创建redis的statefulSet

在 Kubernetes 上,通过 Deployment 资源来部署 Redis 是一种通常用于无状态(stateless)应用的方法。尽管 Redis 通常被认为是有状态的应用,但如果你只是用于开发或测试目的,使用 Deployment 也可以。如果是在生产环境中使用,建议使用 StatefulSet,因为 StatefulSet 更适合管理有状态的应用。—— from ChatGPT

apiVersion: apps/v1
kind: StatefulSet
metadata:
  name: redis-app
spec:
  serviceName: "redis-service"
  replicas: 6
  selector:
    matchLabels:
      app: redis
      appCluster: redis-cluster
  template:
    metadata:
      labels:
        app: redis
        appCluster: redis-cluster
    spec:
      containers:
      - name: redis
        image: redis
        imagePullPolicy: IfNotPresent
        command: ["/bin/bash", "-ce", "tail -f /dev/null"]
        command: ["redis-server"]
        args:
          - "/etc/redis/redis.conf"
          - "--protected-mode"
          - "no"
        ports:
            - name: redis
              containerPort: 6379
              protocol: "TCP"
            - name: cluster
              containerPort: 16379
              protocol: "TCP"
        volumeMounts:
          - name: "redis-conf"
            mountPath: "/etc/redis"
          - name: "redis-data"
            mountPath: "/var/lib/redis"
      volumes:
      - name: "redis-conf"
        configMap:
          name: "redis-conf"
          items:
            - key: "redis.conf"
              path: "redis.conf"
  volumeClaimTemplates:
  - metadata:
      name: redis-data
    spec:
      accessModes: [ "ReadWriteMany" ]
      storageClassName: "redis"
      resources:
        requests:
          storage: 1Gi


// 启动服务
$ sudo kubectl create -f redis-stateful.yaml

// 可以看到redis服务已经起来了
$ kubectl get pods
NAME                                 READY   STATUS    RESTARTS       AGE
redis-app-0                          1/1     Running   0              27m
redis-app-1                          1/1     Running   0              27m
redis-app-2                          1/1     Running   0              27m
redis-app-3                          1/1     Running   0              27m
redis-app-4                          1/1     Running   0              27m
redis-app-5                          1/1     Running   0              27m

$ kubectl get svc
NAME                TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)        AGE
redis-service       ClusterIP   None             <none>        6379/TCP       5d


  • 创建redis集群

使用redis-trib工具创建redis集群,通过docker构建

$ sudo docker pull inem0o/redis-trib

$ kubectl describe pod redis-app | grep IP
                  cni.projectcalico.org/podIP: 172.16.43.41/32
                  cni.projectcalico.org/podIPs: 172.16.43.41/32
IP:               172.16.43.41
IPs:
  IP:           172.16.43.41
                  cni.projectcalico.org/podIP: 172.16.43.40/32
                  cni.projectcalico.org/podIPs: 172.16.43.40/32
IP:               172.16.43.40
IPs:
  IP:           172.16.43.40
                  cni.projectcalico.org/podIP: 172.16.43.18/32
                  cni.projectcalico.org/podIPs: 172.16.43.18/32
IP:               172.16.43.18
IPs:
  IP:           172.16.43.18
                  cni.projectcalico.org/podIP: 172.16.43.20/32
                  cni.projectcalico.org/podIPs: 172.16.43.20/32
IP:               172.16.43.20
IPs:
  IP:           172.16.43.20
                  cni.projectcalico.org/podIP: 172.16.43.52/32
                  cni.projectcalico.org/podIPs: 172.16.43.52/32
IP:               172.16.43.52
IPs:
  IP:           172.16.43.52
                  cni.projectcalico.org/podIP: 172.16.43.53/32
                  cni.projectcalico.org/podIPs: 172.16.43.53/32
IP:               172.16.43.53
IPs:
  IP:           172.16.43.53


$ sudo docker run --rm -ti inem0o/redis-trib create --replicas 1 172.16.43.41:6379 172.16.43.40:6379 172.16.43.18:6379 172.16.43.20:6379 172.16.43.52:6379 172.16.43.53:6379

>>> Creating cluster
>>> Performing hash slots allocation on 6 nodes...
Using 3 masters:
172.16.43.41:6379
172.16.43.40:6379
172.16.43.18:6379
Adding replica 172.16.43.20:6379 to 172.16.43.41:6379
Adding replica 172.16.43.52:6379 to 172.16.43.40:6379
Adding replica 172.16.43.53:6379 to 172.16.43.18:6379
M: 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f 172.16.43.41:6379
   slots:0-5460 (5461 slots) master
M: f5d617c0ed655dd6afa32c5d4ec6260713668639 172.16.43.40:6379
   slots:5461-10922 (5462 slots) master
M: 808de7e00f10fe17a5582cd76a533159a25006d8 172.16.43.18:6379
   slots:10923-16383 (5461 slots) master
S: 2f5860e62f03ea17d398bbe447a6f1d428ae8698 172.16.43.20:6379
   replicates 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f
S: 44ac042b99b9b73051b05d1be3d98cf475f67f0a 172.16.43.52:6379
   replicates f5d617c0ed655dd6afa32c5d4ec6260713668639
S: 8db8f89b7b28d0ce098de275340e3c4679fd342d 172.16.43.53:6379
   replicates 808de7e00f10fe17a5582cd76a533159a25006d8
Can I set the above configuration? (type 'yes' to accept): yes
>>> Nodes configuration updated
>>> Assign a different config epoch to each node
>>> Sending CLUSTER MEET messages to join the cluster
Waiting for the cluster to join.
>>> Performing Cluster Check (using node 172.16.43.41:6379)
M: 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f 172.16.43.41:6379
   slots:0-5460 (5461 slots) master
   1 additional replica(s)
S: 44ac042b99b9b73051b05d1be3d98cf475f67f0a 172.16.43.52:6379@16379
   slots: (0 slots) slave
   replicates f5d617c0ed655dd6afa32c5d4ec6260713668639
M: f5d617c0ed655dd6afa32c5d4ec6260713668639 172.16.43.40:6379@16379
   slots:5461-10922 (5462 slots) master
   1 additional replica(s)
S: 2f5860e62f03ea17d398bbe447a6f1d428ae8698 172.16.43.20:6379@16379
   slots: (0 slots) slave
   replicates 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f
M: 808de7e00f10fe17a5582cd76a533159a25006d8 172.16.43.18:6379@16379
   slots:10923-16383 (5461 slots) master
   1 additional replica(s)
S: 8db8f89b7b28d0ce098de275340e3c4679fd342d 172.16.43.53:6379@16379
   slots: (0 slots) slave
   replicates 808de7e00f10fe17a5582cd76a533159a25006d8
[OK] All nodes agree about slots configuration.
>>> Check for open slots...
>>> Check slots coverage...
[OK] All 16384 slots covered.

至此集群构建成功。

  • 测试

  • 登录redis的pod0,查看节点状态
$ kubectl exec -it redis-app-0 -- redis-cli
127.0.0.1:6379> cluster info
cluster_state:ok
cluster_slots_assigned:16384
cluster_slots_ok:16384
cluster_slots_pfail:0
cluster_slots_fail:0
cluster_known_nodes:6
cluster_size:3
cluster_current_epoch:6
cluster_my_epoch:1
cluster_stats_messages_ping_sent:1537
cluster_stats_messages_pong_sent:1570
cluster_stats_messages_sent:3107
cluster_stats_messages_ping_received:1565
cluster_stats_messages_pong_received:1537
cluster_stats_messages_meet_received:5
cluster_stats_messages_received:3107
total_cluster_links_buffer_limit_exceeded:0
127.0.0.1:6379> role
1) "master"
2) (integer) 1204
3) 1) 1) "172.16.43.20"
      2) "6379"
      3) "1204"
  •  创建key-value值
// 在开启集群后,redis-cli用普通用户登录无法操作集群中的数据,需要加上-c 用集群模式登录才可进行操作。
$ kubectl exec -it redis-app-0 -- redis-cli -c
172.16.43.18:6379> set test 111
-> Redirected to slot [6918] located at 172.16.43.40:6379
OK
172.16.43.40:6379> get test
"111"


// 然后登录另一个节点,可以看到数据同步过去了
$ kubectl exec -it redis-app-1 -- redis-cli -c
127.0.0.1:6379> get test
"111"

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值