ceph 分布式存储出现告警 22 pgs not deep-scrubbed in time 解决办法:
$ ceph -scluster:
id: ce68aab8-8f46-11ed-88c0-ac1f6b3a30b9
health: HEALTH_ERR
3 failed cephadm daemon(s)
failed to probe daemons or devices
2 mgr modules have failed
mon gm268-1 is low on available space
22 pgs not deep-scrubbed in time
1 slow ops, oldest one blocked for 5580 sec, mon.gm268-3 has slow ops
services:
mon: 3 daemons, quorum gm268-2,gm268-3,gm268-1 (age 86m)
mgr: gm268-2.zttohs(active, since 97m), standbys: gm268-3.sjagqo, gm268-1.jgdvxs
mds: cephfs:1 {0=cephfs.gm268-3.ppyjrl=up:active} 1 up:standby
osd: 41 osds: 41 up (since 93m), 41 in (since 21h); 182 remapped pgs
data:
pools: 5 pools, 11265 pgs
objects: 42.48M objects, 115 TiB
usage: 232 TiB used, 365 TiB / 597 TiB avail
pgs: 785971/84967110 objects misplaced (0.925%)
11083 active+clean
181 active+remapped+backfill_wait
1 active+remapped+backfilling
io:
recovery: 20 MiB/s, 6 objects/s
解决办法:
$ ceph config set global osd_deep_scrub_interval 3628800
$ ceph config getmon
WHO MASKLEVEL OPTION VALUE RO
mon advancedauth_allow_insecure_global_id_reclaimfalse
global basic container_image quay.io/ceph/ceph:v15*
global advancedmon_max_pg_per_osd 800
global advancedosd_deep_scrub_interval 3628800.000000
global advancedosd_pool_default_pg_autoscale_mode off
mon advancedpublic_network 10.12.0.0/15 *
$ ceph config getosd
WHO MASKLEVEL OPTION VALUE RO
global basic container_image quay.io/ceph/ceph:v15*
global advancedmon_max_pg_per_osd 800
global advancedosd_deep_scrub_interval 3628800.000000
global advancedosd_pool_default_pg_autoscale_modeoff
$ ceph -s
cluster:
id: ce68aab8-8f46-11ed-88c0-ac1f6b3a30b9
health: HEALTH_ERR
3 failed cephadm daemon(s)
failed to probe daemons or devices
2 mgr modules have failed
mon gm268-1 is low on available space
1 slow ops, oldest one blocked for 5791 sec, mon.gm268-3 has slow ops
services:
mon: 3 daemons, quorum gm268-2,gm268-3,gm268-1 (age 89m)
mgr: gm268-2.zttohs(active, since 100m), standbys: gm268-3.sjagqo, gm268-1.jgdvxs
mds: cephfs:1 {0=cephfs.gm268-3.ppyjrl=up:active} 1 up:standby
osd: 41 osds: 41 up (since 96m), 41 in (since 21h); 181 remapped pgs
data:
pools: 5 pools, 11265 pgs
objects: 42.48M objects, 115 TiB
usage: 232 TiB used, 365 TiB / 597 TiB avail
pgs: 784873/84967110 objects misplaced (0.924%)
11084 active+clean
180 active+remapped+backfill_wait
1 active+remapped+backfilling
io:
recovery: 20 MiB/s, 6 objects/s
页:
[1]