X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fdoc%2Frados%2Foperations%2Fpg-states.rst;fp=src%2Fceph%2Fdoc%2Frados%2Foperations%2Fpg-states.rst;h=0fbd3dcf0b0494bb94154c1e63e4a4c44de87ced;hb=812ff6ca9fcd3e629e49d4328905f33eee8ca3f5;hp=0000000000000000000000000000000000000000;hpb=15280273faafb77777eab341909a3f495cf248d9;p=stor4nfv.git diff --git a/src/ceph/doc/rados/operations/pg-states.rst b/src/ceph/doc/rados/operations/pg-states.rst new file mode 100644 index 0000000..0fbd3dc --- /dev/null +++ b/src/ceph/doc/rados/operations/pg-states.rst @@ -0,0 +1,80 @@ +======================== + Placement Group States +======================== + +When checking a cluster's status (e.g., running ``ceph -w`` or ``ceph -s``), +Ceph will report on the status of the placement groups. A placement group has +one or more states. The optimum state for placement groups in the placement group +map is ``active + clean``. + +*Creating* + Ceph is still creating the placement group. + +*Active* + Ceph will process requests to the placement group. + +*Clean* + Ceph replicated all objects in the placement group the correct number of times. + +*Down* + A replica with necessary data is down, so the placement group is offline. + +*Scrubbing* + Ceph is checking the placement group for inconsistencies. + +*Degraded* + Ceph has not replicated some objects in the placement group the correct number of times yet. + +*Inconsistent* + Ceph detects inconsistencies in the one or more replicas of an object in the placement group + (e.g. objects are the wrong size, objects are missing from one replica *after* recovery finished, etc.). + +*Peering* + The placement group is undergoing the peering process + +*Repair* + Ceph is checking the placement group and repairing any inconsistencies it finds (if possible). + +*Recovering* + Ceph is migrating/synchronizing objects and their replicas. + +*Forced-Recovery* + High recovery priority of that PG is enforced by user. + +*Backfill* + Ceph is scanning and synchronizing the entire contents of a placement group + instead of inferring what contents need to be synchronized from the logs of + recent operations. *Backfill* is a special case of recovery. + +*Forced-Backfill* + High backfill priority of that PG is enforced by user. + +*Wait-backfill* + The placement group is waiting in line to start backfill. + +*Backfill-toofull* + A backfill operation is waiting because the destination OSD is over its + full ratio. + +*Incomplete* + Ceph detects that a placement group is missing information about + writes that may have occurred, or does not have any healthy + copies. If you see this state, try to start any failed OSDs that may + contain the needed information. In the case of an erasure coded pool + temporarily reducing min_size may allow recovery. + +*Stale* + The placement group is in an unknown state - the monitors have not received + an update for it since the placement group mapping changed. + +*Remapped* + The placement group is temporarily mapped to a different set of OSDs from what + CRUSH specified. + +*Undersized* + The placement group fewer copies than the configured pool replication level. + +*Peered* + The placement group has peered, but cannot serve client IO due to not having + enough copies to reach the pool's configured min_size parameter. Recovery + may occur in this state, so the pg may heal up to min_size eventually.