Ceph operation, maintenance and repair
Jump to navigation
Jump to search
How to see the status of the cluster
On any cluster node, you can run the ceph health
, ceph health detail
or ceph status
commands to get an increasingly-detailed overview of the cluster's status.
ceph health
[root@vmfram1 ~]# ceph health HEALTH_WARN Low space hindering backfill (add storage if this doesn't resolve itself): 2 pgs backfill_toofull [root@vmfram1 ~]#
ceph health detail
[root@vmfram1 ~]# ceph health detail HEALTH_WARN Low space hindering backfill (add storage if this doesn't resolve itself): 2 pgs backfill_toofull PG_BACKFILL_FULL Low space hindering backfill (add storage if this doesn't resolve itself): 2 pgs backfill_toofull pg 9.5 is active+remapped+backfill_wait+backfill_toofull, acting [305,406,103] pg 9.e is active+remapped+backfill_wait+backfill_toofull, acting [306,406,104] [root@vmfram1 ~]#
ceph status
root@storage00:~# ceph status cluster: id: db5b6a5a-1080-46d2-974a-80fe8274c8ba health: HEALTH_OK services: mon: 3 daemons, quorum storage00,storage01,compute01 (age 12d) mgr: storage01(active, since 12d), standbys: storage00 mds: vm:1 {0=storage01=up:active} 1 up:standby osd: 10 osds: 8 up (since 12d), 8 in (since 3M) data: pools: 4 pools, 448 pgs objects: 15.60k objects, 59 GiB usage: 177 GiB used, 14 TiB / 14 TiB avail pgs: 448 active+clean io: client: 341 B/s wr, 0 op/s rd, 0 op/s wr root@storage00:~#