Alert rule: CephPGsHighPerOSD
Please consider opening a PR to improve this runbook if you gain new information about causes of the alert, or how to debug or resolve the alert. Click "Edit this Page" in the top right corner to create a PR directly on GitHub.
The number of placement groups per OSD is too high (exceeds the mon_max_pg_per_osd setting). Autoscale might be disabled or there are too few nodes in the cluster.
$ ceph_cluster_ns=syn-rook-ceph-cluster # List pools POOL SIZE [...] BIAS PG_NUM NEW PG_NUM AUTOSCALE PROFILE fspool-data0 0 [...] 1.0 32 on scale-down (1) device_health_metrics 937.0k [...] 1.0 8 on scale-down fspool-metadata 23363k [...] 4.0 32 on scale-down storagepool 20199 [...] 1.0 32 on scale-down
|1||Autoscale should be on for the pool.|