summaryrefslogtreecommitdiffstats
path: root/src/ceph/doc/rados/operations/pg-states.rst
diff options
context:
space:
mode:
Diffstat (limited to 'src/ceph/doc/rados/operations/pg-states.rst')
-rw-r--r--src/ceph/doc/rados/operations/pg-states.rst80
1 files changed, 0 insertions, 80 deletions
diff --git a/src/ceph/doc/rados/operations/pg-states.rst b/src/ceph/doc/rados/operations/pg-states.rst
deleted file mode 100644
index 0fbd3dc..0000000
--- a/src/ceph/doc/rados/operations/pg-states.rst
+++ /dev/null
@@ -1,80 +0,0 @@
-========================
- Placement Group States
-========================
-
-When checking a cluster's status (e.g., running ``ceph -w`` or ``ceph -s``),
-Ceph will report on the status of the placement groups. A placement group has
-one or more states. The optimum state for placement groups in the placement group
-map is ``active + clean``.
-
-*Creating*
- Ceph is still creating the placement group.
-
-*Active*
- Ceph will process requests to the placement group.
-
-*Clean*
- Ceph replicated all objects in the placement group the correct number of times.
-
-*Down*
- A replica with necessary data is down, so the placement group is offline.
-
-*Scrubbing*
- Ceph is checking the placement group for inconsistencies.
-
-*Degraded*
- Ceph has not replicated some objects in the placement group the correct number of times yet.
-
-*Inconsistent*
- Ceph detects inconsistencies in the one or more replicas of an object in the placement group
- (e.g. objects are the wrong size, objects are missing from one replica *after* recovery finished, etc.).
-
-*Peering*
- The placement group is undergoing the peering process
-
-*Repair*
- Ceph is checking the placement group and repairing any inconsistencies it finds (if possible).
-
-*Recovering*
- Ceph is migrating/synchronizing objects and their replicas.
-
-*Forced-Recovery*
- High recovery priority of that PG is enforced by user.
-
-*Backfill*
- Ceph is scanning and synchronizing the entire contents of a placement group
- instead of inferring what contents need to be synchronized from the logs of
- recent operations. *Backfill* is a special case of recovery.
-
-*Forced-Backfill*
- High backfill priority of that PG is enforced by user.
-
-*Wait-backfill*
- The placement group is waiting in line to start backfill.
-
-*Backfill-toofull*
- A backfill operation is waiting because the destination OSD is over its
- full ratio.
-
-*Incomplete*
- Ceph detects that a placement group is missing information about
- writes that may have occurred, or does not have any healthy
- copies. If you see this state, try to start any failed OSDs that may
- contain the needed information. In the case of an erasure coded pool
- temporarily reducing min_size may allow recovery.
-
-*Stale*
- The placement group is in an unknown state - the monitors have not received
- an update for it since the placement group mapping changed.
-
-*Remapped*
- The placement group is temporarily mapped to a different set of OSDs from what
- CRUSH specified.
-
-*Undersized*
- The placement group fewer copies than the configured pool replication level.
-
-*Peered*
- The placement group has peered, but cannot serve client IO due to not having
- enough copies to reach the pool's configured min_size parameter. Recovery
- may occur in this state, so the pg may heal up to min_size eventually.