维护ceph集群
- 人工智能
- 2025-09-02 07:33:01

1. set: 设置标志位
# ceph osd set <flag_name> # ceph osd set noout # ceph osd set nodown # ceph osd set norecover2. unset: 清除标志位
# ceph osd unset <flag_name> # ceph osd unset noout # ceph osd unset nodown # ceph osd unset norecover3. 标志位
noout: 该标志位将使 ceph 集群不会将任何 OSD 标记为 out(集群外),无论其实际状态如何。这将会把所有的 OSD 保留在 ceph 集群中。
nodown: 该标志位将使得 ceph 集群不会将任何 OSD 标记为 down(服务已停止),无论其实际状态如何。这将会使集群中的所有 OSD 保持 UP(服务运行中)状态,面不会是 DOWN 状态。
noup:
4. "Too many repaired reads on 1 OSDs" 告警处理
# ceph -s cluster: id: dfcdf8de-f388-4c84-adc2-ee721da8df84 health: HEALTH_WARN nodeep-scrub flag(s) set Too many repaired reads on 1 OSDs 3 pgs not deep-scrubbed in time services: mon: 1 daemons, quorum server (age 9w) mgr: server(active, since 9w) osd: 4 osds: 3 up (since 9w), 3 in (since 4M) flags nodeep-scrub data: pools: 3 pools, 320 pgs objects: 155.80k objects, 605 GiB usage: 1.2 TiB used, 3.4 TiB / 4.6 TiB avail pgs: 320 active+clean io: client: 0 B/s rd, 19 KiB/s wr, 1 op/s rd, 1 op/s wr # ceph health detail HEALTH_WARN nodeep-scrub flag(s) set; Too many repaired reads on 1 OSDs; 3 pgs not deep-scrubbed in time OSDMAP_FLAGS nodeep-scrub flag(s) set OSD_TOO_MANY_REPAIRS Too many repaired reads on 1 OSDs osd.3 had 13 reads repaired PG_NOT_DEEP_SCRUBBED 3 pgs not deep-scrubbed in time pg 4.48 not deep-scrubbed since 2025-02-05 07:05:25.334392 pg 5.44 not deep-scrubbed since 2025-02-05 07:33:33.616573 pg 6.29 not deep-scrubbed since 2025-02-05 01:01:12.492269 # ceph tell osd.3 clear_shards_repaired[待续]
上一篇
训练营3,