ceph (luminous 版) data disk 故障测试

本文介绍了在Ceph Luminous环境中,如何进行数据磁盘故障的模拟与恢复过程。首先,展示了故障前的系统状态,然后详细描述了模拟故障的步骤,包括查看错误日志和删除OSD授权。接着,详述了恢复流程,如移除OSD、初始化磁盘和重新创建OSD。最后,通过检查系统状态和OSD树验证了恢复效果。
摘要由CSDN通过智能技术生成

目的

模拟 ceph (luminous 版) data disk 故障
修复上述问题

环境

参考手动部署 ceph 环境说明 (luminous 版)

参考当前 ceph 环境

ceph -s

  cluster:
    id:     c45b752d-5d4d-4d3a-a3b2-04e73eff4ccd
    health: HEALTH_OK

  services:
    mon: 3 daemons, quorum hh-ceph-128040,hh-ceph-128214,hh-ceph-128215
    mgr: openstack(active)
    osd: 36 osds: 36 up, 36 in

  data:
    pools:   1 pools, 2048 pgs
    objects: 28024 objects, 109 GB
    usage:   331 GB used, 196 TB / 196 TB avail
    pgs:     2048 active+clean

osd tree (取部分)

[root@hh-ceph-128214 ceph]# ceph osd tree
ID  CLASS WEIGHT    TYPE NAME                   STATUS REWEIGHT PRI-AFF
 -1       216.00000 root default
-10        72.00000     rack racka07
 -3        72.00000         host hh-ceph-128214
 12   hdd   6.00000             osd.12              up  1.00000 1.00000
 13   hdd   6.00000             osd.13              up  1.00000 1.00000
 14   hdd   6.00000             osd.14              up  1.00000 1.00000
 15   hdd   6.00000             osd.15              up  1.00000 1.00000
 16   hdd   6.00000             osd.16              up  1.00000 1.00000
 17   hdd   6.00000             osd.17              up  1.00000 1.00000
 18   hdd   6.00000             osd.18              up  1.00000 1.00000
 19   hdd   6.00000             osd.19              up  1.00000 1.00000
 20   hdd   6.00000             osd.20              up  1.00000 1.00000
 21   hdd   6.00000             osd.21              up  1.00000 1.00000
 22   hdd   6.00000             osd.22              up  1.00000 1.00000
 23   hdd   6.00000             osd.23              up  1.00000 1.00000
 -9        72.00000     rack racka12
 -2        72.00000         host hh-ceph-128040
  0   hdd   6.00000             osd.0               up  1.00000 0.50000
  1   hdd   6.00000             osd.1               up  1.00000 1.00000
  2   hdd   6.00000             osd.2               up  1.00000 1.00000
  3   hdd   6.00000             osd.3               up  1.00000 1.00000

故障模拟

[root@hh-ceph-128214 ceph]# df -h | grep ceph-14
/dev/sdc1       5.5T  8.8G  5.5T    1% /var/lib/ceph/osd/ceph-14
/dev/sdn3       4.7G  2.1G  2.7G   44% /var/lib/ceph/journal/ceph-14
[root@hh-ceph-128214 ceph]# rm -rf  /var/lib/ceph/osd/ceph-14/*
[root@hh-ceph-128214 ceph]# ls /var/lib/ceph/osd/ceph-14/

查询当前状态

  cluster:
    id:     c45b752d-5d4d-4d3a-a3b2-04e73eff4ccd
    health: HEALTH_WARN
            1 osds down
            Degraded data redundancy: 3246/121608 objects degraded (2.669%), 124 pgs unclean, 155 pgs degraded

  services:
    mon: 3 daemons, quorum hh-ceph-128040,hh-ceph-128214,hh-ceph-128215
    mgr: openstack(active)
    osd: 36 osds: 35 up, 36 in

  data:
    pools:   1 pools, 2048 pgs
    objects: 40536 objects, 157 GB
    usage:   493 GB used, 195 TB / 196 TB avail
    pgs:     3246/121608 objects degraded (2.669%)
             1893 active+clean
             155  active+undersized+degraded

  io:
    client:   132 kB/s rd, 177 MB/s wr, 165 op/s rd, 175 op/s wr

参考 osd tree

[root@hh-ceph-128214 ceph]# ceph osd tree
ID  CLASS WEIGHT    TYPE NAME                   STATUS REWEIGHT PRI-AFF
 -1       216.00000 root default
-10        72.00000     rack racka07
 -3        72.00000         host hh-ceph-128214
 12   hdd   6.00000             osd.12              up  1.00000 1.00000
 13   hdd  
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

Terry_Tsang

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值