ceph更换硬盘

ceph -s  //检查故障

k8s70132:~$ ceph -s
  cluster:
    id:     d10b3028-b78d-4b23-bacb-ca26c0a35c99
    health: HEALTH_WARN
            1 daemons have recently crashed
 
  services:
    mon: 5 daemons, quorum k8sceph70201,k8sceph70202,k8sceph70203,k8sceph70204,k8sceph70205 (age 4d)
    mgr: k8sceph70204(active, since 11w), standbys: k8sceph70201, k8sceph70205
    mds: cephfs:1 {0=k8sceph70204=up:active} 2 up:standby
    osd: 23 osds: 22 up (since 4d), 22 in (since 4d)
    rgw: 3 daemons active (k8sceph70201, k8sceph70204, k8sceph70205)
 
  task status:
 
  data:
    pools:   11 pools, 281 pgs
    objects: 809.23k objects, 24 GiB
    usage:   159 GiB used, 38 TiB / 38 TiB avail
    pgs:     281 active+clean
 
  io:
    client:   47 KiB/s wr, 0 op/s rd, 2 op/s wr

这里health有告警,查看故障点:

ceph crush ls

k8s70132:~$ ceph crash ls
ID                                                                ENTITY  NEW  
2023-04-30T19:46:01.008208Z_26692ab3-ba90-4129-9929-2ad8f29f0acb  osd.1        
2023-09-03T07:42:42.451722Z_d0e7268f-0da6-4d59-b706-35c49ee8617b  osd.2        
2023-12-13T22:17:16.706091Z_8dbfc488-1309-4e9a-b4b9-c1eadeb3016e  osd.0    * 

ceph osd tree

k8s70132:~$ ceph osd tree
ID   CLASS  WEIGHT    TYPE NAME              STATUS  REWEIGHT  PRI-AFF
 -1         42.02631  root default                                    
 -3         10.91606      host k8sceph70201                           
  0    hdd   3.63869          osd.0            down         0  1.00000
  3    hdd   3.63869          osd.3              up   1.00000  1.00000
  6    hdd   3.63869          osd.6              up   1.00000  1.00000
 -5          2.72910      host k8sceph70202                           
  1    hdd   0.90970          osd.1              up   1.00000  1.00000
  4    hdd   0.90970          osd.4              up   1.00000  1.00000
  7    hdd   0.90970          osd.7              up   1.00000  1.00000
 -7         10.91606      host k8sceph70203                           
  2    hdd   3.63869          osd.2              up   1.00000  1.00000
  5    hdd   3.63869          osd.5              up   1.00000  1.00000
  8    hdd   3.63869          osd.8              up   1.00000  1.00000

到k8sceph70201查看

sudo systmectl |grep ceph-osd

k8sceph70201:~$ sudo systemctl |grep ceph-osd
  var-lib-ceph-osd-ceph\x2d0.mount                                                               loaded active mounted   /var/lib/ceph/osd/ceph-0                                                      
  var-lib-ceph-osd-ceph\x2d3.mount                                                               loaded active mounted   /var/lib/ceph/osd/ceph-3                                                      
  var-lib-ceph-osd-ceph\x2d6.mount                                                               loaded active mounted   /var/lib/ceph/osd/ceph-6                                                      
● ceph-osd@0.service                                                                             loaded failed failed    Ceph object storage daemon osd.0                                              
  ceph-osd@3.service                                                                             loaded active running   Ceph object storage daemon osd.3                                              
  ceph-osd@6.service                                                                             loaded active running   Ceph object storage daemon osd.6                                              
  ceph-osd.target                                                                                loaded active active    ceph target allowing to start/stop all ceph-osd@.service instances 

删除故障硬盘

//在k8sceph70201上关停服务
sudo systemctl stop ceph-osd@0.service

//在ceph客户端操作如下
 ceph osd down osd.0
//回显内容 osd.0 is already down. 
ceph osd out osd.0
//回显内容 osd.0 is already out. 
ceph osd crush remove osd.0
//回显内容 removed item id 0 name 'osd.0' from crush map
ceph osd rm 0
//回显内容 removed osd.0
ceph auth del osd.0
//回显内容 updated
ceph osd tree
//回显内容 
ID   CLASS  WEIGHT    TYPE NAME              STATUS  REWEIGHT  PRI-AFF
 -1         38.38762  root default                                    
 -3          7.27737      host k8sceph70201                           
  3    hdd   3.63869          osd.3              up   1.00000  1.00000
  6    hdd   3.63869          osd.6              up   1.00000  1.00000
 -5          2.72910      host k8sceph70202                           
  1    hdd   0.90970          osd.1              up   1.00000  1.00000
  4    hdd   0.90970          osd.4              up   1.00000  1.00000
  7    hdd   0.90970          osd.7              up   1.00000  1.00000
 -7         10.91606      host k8sceph70203                           
  2    hdd   3.63869          osd.2              up   1.00000  1.00000
  5    hdd   3.63869          osd.5              up   1.00000  1.00000
  8    hdd   3.63869          osd.8            down   1.00000  1.00000

添加硬盘

ceph-deploy disk zap  /dev/sdb 
ceph-deploy --overwrite-conf  osd create  /dev/sdb
Ceph是一个可靠的、数据自动重均衡、自动恢复的SDS(软件定义存储)分布式存储系统,功能主要有三大块:块存储、对象存储、文件系统。Ceph不但是提供了统一存储,并且同时还充分利用了客户端的计算能力,在存储每一个数据时,都会通过计算得出该数据存储的位置,尽量将数据分布均衡,同时由于Ceph的设计,采用了CRUSH算法、HASH环等方法,使得它不存在传统的单点故障的问题,且随着规模的扩大性能并不会受到影响。不管是私有云还是公有云,随着云架构的发展,Ceph 成为了OpenStack、Proxmox标配的后端存储,不光如此,就连目前最火的容器编排 Kubernetes 持久存储也都支持 Ceph ,让 Ceph 成为当前主流的存储系统,而对 Ceph 存储的学习也就变的刻不容缓。 该课程属于比较基础的 Ceph 存储课程,旨在让大家能快速上手对 Ceph 集群的部署以及操作,主要讲解以下几个方面: 1.     架构的介绍2.     集群的多种部署方式3.     块存储部署,客户端使用及快照、克隆和镜像4.     对象存储的部署及简单使用5.     Ceph FS 文件系统的部署、客户端使用及导出为NFS6.     集群的操作和管理,包括服务,集群扩展,缩减和集群维护7.     Ceph 基础知识的介绍,包括集群map,PG和身份验证8.     Ceph 集群的监控
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值