103) ceph更换故障OSD/添加新OSD步骤
作者:互联网
1- 更换故障OSD
1.1- 故障现象
$ ceph health detail
......
OSD_SCRUB_ERRORS 31 scrub errors
PG_DAMAGED Possible data damage: 5 pgs inconsistent
pg 41.33 is active+clean+inconsistent, acting [35,33]
pg 41.42 is active+clean+inconsistent, acting [29,35]
pg 51.24 is active+clean+inconsistent, acting [35,43]
pg 51.77 is active+clean+inconsistent, acting [28,35]
pg 51.7b is active+clean+inconsistent, acting [35,46]
1.2- 临时解决办法
执行 ceph pg repair 解决,此时由于磁盘坏道造成不可读的数据会拷贝到其他位置。但这不能从根本上解决问题,磁盘损坏会持续报出类似的错误。
$ ceph pg repair 41.33
$ ceph pg repair 41.42
$ ceph pg repair 51.24
$ ceph pg repair 51.77
$ ceph pg repair 51.7b
1.3- 获取磁盘错误信息
- 定位磁盘:
apt install -y smartmontools
# yum install -y smartmontools
smartctl -a /dev/sdc
smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-121-generic] (local build)
Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Device Model: TOSHIBA MG04ACA600E
Serial Number: 57J6KA41F6CD
LU WWN Device Id: 5 000039 7cb9822be
Firmware Version: FS1K
User Capacity: 6,001,175,126,016 bytes [6.00 TB]
Sector Sizes: 512 bytes logical, 4096 bytes physical
Rotation Rate: 7200 rpm
Form Factor: 3.5 inches
Device is: Not in smartctl database [for details use: -P showall]
ATA Version is: ATA8-ACS (minor revision not indicated)
SATA Version is: SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s)
Local Time is: Tue Aug 7 14:46:45 2018 CST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
1.4- 关闭ceph集群数据迁移:
osd硬盘故障,状态变为down。在经过mod osd down out interval 设定的时间间隔后,ceph将其标记为out,并开始进行数据迁移恢复。为了降低ceph进行数据恢复或scrub等操作对性能的影响,可以先将其暂时关闭,待硬盘更换完成且osd恢复后再开启:
1.5- 进入osd故障的节点,卸载osd挂载目录
umount /var/lib/ceph/osd/ceph-5
1.6- 从crush map 中移除osd
ceph osd crush remove osd.5
1.7- 删除故障osd的密钥
ceph auth del osd.5
1.8- 删除故障osd
ceph osd rm 5
1.9- 更换完新硬盘后,注意新硬盘的盘符,并创建osd
ceph-deploy osd create --data /dev/sdd node3
1.10- 待新osd添加crush map后,重新开启集群禁用标志
for i in noout nobackfill norecover noscrub nodeep-scrub;do ceph osd unset $i;done
1.11- ceph集群经过一段时间的数据迁移后,恢复active+clean状态
2- 添加新OSD
- 选择一个osd节点,添加好新的硬盘
- 擦净节点磁盘
ceph-deploy disk zap [node_name] /dev/sdb
- 准备Object Storage Daemon
ceph-deploy osd prepare [node_name]:/var/lib/ceph/osd1
- 激活Object Storage Daemon
ceph-deploy osd activate [node_name]:/var/lib/ceph/osd1
3- 删除OSD
- 把 OSD 踢出集群
ceph osd out osd.4
- 在相应的节点,停止ceph-osd服务
systemctl stop ceph-osd@4.service
systemctl disable ceph-osd@4.service
- 删除 CRUSH 图的对应 OSD 条目,它就不再接收数据了
ceph osd crush remove osd.4
- 删除 OSD 认证密钥
ceph auth del osd.4
- 删除osd.4
ceph osd rm osd.4
标签:repair,osd,ceph,pg,103,OSD,inconsistent 来源: https://www.cnblogs.com/lemanlai/p/13798948.html