参考:
K8S搭建Redis集群_172.31.0.82-CSDN博客
-
首先创建nfs存储,当redis迁移之后,仍然能够从nfs中获取之前的数据
需要在机器上安装nfs服务,这里我已经安装好了,可以参考这篇文章k8s Persistent Volume-CSDN博客进行安装
-
之后设置共享的路径
$ sudo mkdir -p /usr/local/kubernetes/redis/pv{1..6}
$ sudo chmod 777 /usr/local/kubernetes/redis/pv{1..6}
$ ls -la /usr/local/kubernetes/redis
total 32
drwxr-xr-x 8 root root 4096 Aug 7 15:54 .
drwxr-xr-x 3 root root 4096 Aug 7 15:54 ..
drwxrwxrwx 2 root root 4096 Aug 7 15:54 pv1
drwxrwxrwx 2 root root 4096 Aug 7 15:54 pv2
drwxrwxrwx 2 root root 4096 Aug 7 15:54 pv3
drwxrwxrwx 2 root root 4096 Aug 7 15:54 pv4
drwxrwxrwx 2 root root 4096 Aug 7 15:54 pv5
drwxrwxrwx 2 root root 4096 Aug 7 15:54 pv6
-
然后设置这些目录的共享
编辑/etc/exports
$ sudo vi /etc/exports
插入以下代码
#================================================================================================
/usr/local/kubernetes/redis/pv1 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv2 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv3 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv4 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv5 *(rw,sync,no_root_squash)
/usr/local/kubernetes/redis/pv6 *(rw,sync,no_root_squash)
#================================================================================================
~
之后重启服务
$ sudo service nfs-kernel-server restart
$ sudo systemctl restart rpcbind
然后通过命令查看共享情况
$ showmount -e
Export list for xxxxxxx
/usr/local/kubernetes/redis/pv6 *
/usr/local/kubernetes/redis/pv5 *
/usr/local/kubernetes/redis/pv4 *
/usr/local/kubernetes/redis/pv3 *
/usr/local/kubernetes/redis/pv2 *
/usr/local/kubernetes/redis/pv1 *
$ sudo exportfs -v
/usr/local/kubernetes/redis/pv1
<world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv2
<world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv3
<world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv4
<world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv5
<world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
/usr/local/kubernetes/redis/pv6
<world>(sync,wdelay,hide,no_subtree_check,sec=sys,rw,secure,no_root_squash,no_all_squash)
-
基于上述的nfs共享目录,创建pv
$ cat redis-pv.yaml
apiVersion: v1
kind: PersistentVolume
metadata:
name: nfs-pv1
spec:
storageClassName: redis
capacity:
storage: 1Gi
accessModes:
- ReadWriteMany
nfs:
server: xxxxxxx
path: "/usr/local/kubernetes/redis/pv1"
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: nfs-pv2
spec:
storageClassName: redis
capacity:
storage: 1Gi
accessModes:
- ReadWriteMany
nfs:
server: xxxxxxx
path: "/usr/local/kubernetes/redis/pv2"
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: nfs-pv3
spec:
storageClassName: redis
capacity:
storage: 1Gi
accessModes:
- ReadWriteMany
nfs:
server: xxxxxxx
path: "/usr/local/kubernetes/redis/pv3"
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: nfs-pv4
spec:
storageClassName: redis
capacity:
storage: 1Gi
accessModes:
- ReadWriteMany
nfs:
server: xxxxxxx
path: "/usr/local/kubernetes/redis/pv4"
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: nfs-pv5
spec:
storageClassName: redis
capacity:
storage: 1Gi
accessModes:
- ReadWriteMany
nfs:
server: xxxxxxx
path: "/usr/local/kubernetes/redis/pv5"
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: nfs-pv6
spec:
storageClassName: redis
capacity:
storage: 1Gi
accessModes:
- ReadWriteMany
nfs:
server: xxxxxxx
path: "/usr/local/kubernetes/redis/pv6"
然后创建pv
$ kubectl create -f redis-pv.yaml
persistentvolume/nfs-pv1 created
persistentvolume/nfs-pv2 created
persistentvolume/nfs-pv3 created
persistentvolume/nfs-pv4 created
persistentvolume/nfs-pv5 created
persistentvolume/nfs-pv6 created
$ kubectl get pv
NAME CAPACITY ACCESS MODES RECLAIM POLICY STATUS CLAIM STORAGECLASS REASON AGE
nfs-pv1 1Gi RWX Retain Available redis 7s
nfs-pv2 1Gi RWX Retain Available redis 7s
nfs-pv3 1Gi RWX Retain Available redis 6s
nfs-pv4 1Gi RWX Retain Available redis 6s
nfs-pv5 1Gi RWX Retain Available redis 6s
nfs-pv6 1Gi RWX Retain Available redis 6s
pv1 1Gi RWO Retain Bound default/pvc1 nfs 2d
-
创建configMap并将其转化为configmap文件
$ cat redis.conf
protected-mode no
appendonly yes
cluster-enabled yes
cluster-config-file /var/lib/redis/nodes.conf
cluster-node-timeout 5000
dir /var/lib/redis
port 6379
#将reids.conf转为configMap文件
kubectl create configmap redis-conf --from-file=redis.conf
$ kubectl get cm
NAME DATA AGE
kube-root-ca.crt 1 47d
redis-conf 1 68m
$ kubectl get cm
NAME DATA AGE
kube-root-ca.crt 1 47d
redis-conf 1 68m
kevin@kevin-pc:~/deploy-tests/redis$ kubectl describe cm redis-conf
Name: redis-conf
Namespace: default
Labels: <none>
Annotations: <none>
Data
====
redis.conf:
----
protected-mode no
appendonly yes
cluster-enabled yes
cluster-config-file /var/lib/redis/nodes.conf
cluster-node-timeout 5000
dir /var/lib/redis
port 6379
BinaryData
====
Events: <none>
-
创建redis的service
$ cat redis-service.yaml
apiVersion: v1
kind: Service
metadata:
name: redis-service
labels:
app: redis
spec:
ports:
- name: redis-port
port: 6379
clusterIP: None
selector:
app: redis
appCluster: redis-cluster
$ kubectl create -f redis-service.yaml
$ kubectl get svc
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
redis-service ClusterIP None <none> 6379/TCP 67m
$ kubectl describe svc redis-service
Name: redis-service
Namespace: default
Labels: app=redis
Annotations: <none>
Selector: app=redis,appCluster=redis-cluster
Type: ClusterIP
IP Family Policy: SingleStack
IP Families: IPv4
IP: None
IPs: None
Port: redis-port 6379/TCP
TargetPort: 6379/TCP
Endpoints: <none>
Session Affinity: None
Events: <none>
-
创建redis的statefulSet
在 Kubernetes 上,通过 Deployment 资源来部署 Redis 是一种通常用于无状态(stateless)应用的方法。尽管 Redis 通常被认为是有状态的应用,但如果你只是用于开发或测试目的,使用 Deployment 也可以。如果是在生产环境中使用,建议使用 StatefulSet,因为 StatefulSet 更适合管理有状态的应用。—— from ChatGPT
apiVersion: apps/v1
kind: StatefulSet
metadata:
name: redis-app
spec:
serviceName: "redis-service"
replicas: 6
selector:
matchLabels:
app: redis
appCluster: redis-cluster
template:
metadata:
labels:
app: redis
appCluster: redis-cluster
spec:
containers:
- name: redis
image: redis
imagePullPolicy: IfNotPresent
command: ["/bin/bash", "-ce", "tail -f /dev/null"]
command: ["redis-server"]
args:
- "/etc/redis/redis.conf"
- "--protected-mode"
- "no"
ports:
- name: redis
containerPort: 6379
protocol: "TCP"
- name: cluster
containerPort: 16379
protocol: "TCP"
volumeMounts:
- name: "redis-conf"
mountPath: "/etc/redis"
- name: "redis-data"
mountPath: "/var/lib/redis"
volumes:
- name: "redis-conf"
configMap:
name: "redis-conf"
items:
- key: "redis.conf"
path: "redis.conf"
volumeClaimTemplates:
- metadata:
name: redis-data
spec:
accessModes: [ "ReadWriteMany" ]
storageClassName: "redis"
resources:
requests:
storage: 1Gi
// 启动服务
$ sudo kubectl create -f redis-stateful.yaml
// 可以看到redis服务已经起来了
$ kubectl get pods
NAME READY STATUS RESTARTS AGE
redis-app-0 1/1 Running 0 27m
redis-app-1 1/1 Running 0 27m
redis-app-2 1/1 Running 0 27m
redis-app-3 1/1 Running 0 27m
redis-app-4 1/1 Running 0 27m
redis-app-5 1/1 Running 0 27m
$ kubectl get svc
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
redis-service ClusterIP None <none> 6379/TCP 5d
-
创建redis集群
使用redis-trib工具创建redis集群,通过docker构建
$ sudo docker pull inem0o/redis-trib
$ kubectl describe pod redis-app | grep IP
cni.projectcalico.org/podIP: 172.16.43.41/32
cni.projectcalico.org/podIPs: 172.16.43.41/32
IP: 172.16.43.41
IPs:
IP: 172.16.43.41
cni.projectcalico.org/podIP: 172.16.43.40/32
cni.projectcalico.org/podIPs: 172.16.43.40/32
IP: 172.16.43.40
IPs:
IP: 172.16.43.40
cni.projectcalico.org/podIP: 172.16.43.18/32
cni.projectcalico.org/podIPs: 172.16.43.18/32
IP: 172.16.43.18
IPs:
IP: 172.16.43.18
cni.projectcalico.org/podIP: 172.16.43.20/32
cni.projectcalico.org/podIPs: 172.16.43.20/32
IP: 172.16.43.20
IPs:
IP: 172.16.43.20
cni.projectcalico.org/podIP: 172.16.43.52/32
cni.projectcalico.org/podIPs: 172.16.43.52/32
IP: 172.16.43.52
IPs:
IP: 172.16.43.52
cni.projectcalico.org/podIP: 172.16.43.53/32
cni.projectcalico.org/podIPs: 172.16.43.53/32
IP: 172.16.43.53
IPs:
IP: 172.16.43.53
$ sudo docker run --rm -ti inem0o/redis-trib create --replicas 1 172.16.43.41:6379 172.16.43.40:6379 172.16.43.18:6379 172.16.43.20:6379 172.16.43.52:6379 172.16.43.53:6379
>>> Creating cluster
>>> Performing hash slots allocation on 6 nodes...
Using 3 masters:
172.16.43.41:6379
172.16.43.40:6379
172.16.43.18:6379
Adding replica 172.16.43.20:6379 to 172.16.43.41:6379
Adding replica 172.16.43.52:6379 to 172.16.43.40:6379
Adding replica 172.16.43.53:6379 to 172.16.43.18:6379
M: 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f 172.16.43.41:6379
slots:0-5460 (5461 slots) master
M: f5d617c0ed655dd6afa32c5d4ec6260713668639 172.16.43.40:6379
slots:5461-10922 (5462 slots) master
M: 808de7e00f10fe17a5582cd76a533159a25006d8 172.16.43.18:6379
slots:10923-16383 (5461 slots) master
S: 2f5860e62f03ea17d398bbe447a6f1d428ae8698 172.16.43.20:6379
replicates 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f
S: 44ac042b99b9b73051b05d1be3d98cf475f67f0a 172.16.43.52:6379
replicates f5d617c0ed655dd6afa32c5d4ec6260713668639
S: 8db8f89b7b28d0ce098de275340e3c4679fd342d 172.16.43.53:6379
replicates 808de7e00f10fe17a5582cd76a533159a25006d8
Can I set the above configuration? (type 'yes' to accept): yes
>>> Nodes configuration updated
>>> Assign a different config epoch to each node
>>> Sending CLUSTER MEET messages to join the cluster
Waiting for the cluster to join.
>>> Performing Cluster Check (using node 172.16.43.41:6379)
M: 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f 172.16.43.41:6379
slots:0-5460 (5461 slots) master
1 additional replica(s)
S: 44ac042b99b9b73051b05d1be3d98cf475f67f0a 172.16.43.52:6379@16379
slots: (0 slots) slave
replicates f5d617c0ed655dd6afa32c5d4ec6260713668639
M: f5d617c0ed655dd6afa32c5d4ec6260713668639 172.16.43.40:6379@16379
slots:5461-10922 (5462 slots) master
1 additional replica(s)
S: 2f5860e62f03ea17d398bbe447a6f1d428ae8698 172.16.43.20:6379@16379
slots: (0 slots) slave
replicates 57d9f345d23e7bf7dd2f331e14d9d7143aa9617f
M: 808de7e00f10fe17a5582cd76a533159a25006d8 172.16.43.18:6379@16379
slots:10923-16383 (5461 slots) master
1 additional replica(s)
S: 8db8f89b7b28d0ce098de275340e3c4679fd342d 172.16.43.53:6379@16379
slots: (0 slots) slave
replicates 808de7e00f10fe17a5582cd76a533159a25006d8
[OK] All nodes agree about slots configuration.
>>> Check for open slots...
>>> Check slots coverage...
[OK] All 16384 slots covered.
至此集群构建成功。
-
测试
- 登录redis的pod0,查看节点状态
$ kubectl exec -it redis-app-0 -- redis-cli
127.0.0.1:6379> cluster info
cluster_state:ok
cluster_slots_assigned:16384
cluster_slots_ok:16384
cluster_slots_pfail:0
cluster_slots_fail:0
cluster_known_nodes:6
cluster_size:3
cluster_current_epoch:6
cluster_my_epoch:1
cluster_stats_messages_ping_sent:1537
cluster_stats_messages_pong_sent:1570
cluster_stats_messages_sent:3107
cluster_stats_messages_ping_received:1565
cluster_stats_messages_pong_received:1537
cluster_stats_messages_meet_received:5
cluster_stats_messages_received:3107
total_cluster_links_buffer_limit_exceeded:0
127.0.0.1:6379> role
1) "master"
2) (integer) 1204
3) 1) 1) "172.16.43.20"
2) "6379"
3) "1204"
- 创建key-value值
// 在开启集群后,redis-cli用普通用户登录无法操作集群中的数据,需要加上-c 用集群模式登录才可进行操作。
$ kubectl exec -it redis-app-0 -- redis-cli -c
172.16.43.18:6379> set test 111
-> Redirected to slot [6918] located at 172.16.43.40:6379
OK
172.16.43.40:6379> get test
"111"
// 然后登录另一个节点,可以看到数据同步过去了
$ kubectl exec -it redis-app-1 -- redis-cli -c
127.0.0.1:6379> get test
"111"