diff options
author | Qiaowei Ren <qiaowei.ren@intel.com> | 2018-03-01 14:38:11 +0800 |
---|---|---|
committer | Qiaowei Ren <qiaowei.ren@intel.com> | 2018-03-01 14:38:11 +0800 |
commit | 7da45d65be36d36b880cc55c5036e96c24b53f00 (patch) | |
tree | d4f944eb4f8f8de50a9a7584ffa408dc3a3185b2 /src/ceph/doc/man/8/ceph.rst | |
parent | 691462d09d0987b47e112d6ee8740375df3c51b2 (diff) |
remove ceph code
This patch removes initial ceph code, due to license issue.
Change-Id: I092d44f601cdf34aed92300fe13214925563081c
Signed-off-by: Qiaowei Ren <qiaowei.ren@intel.com>
Diffstat (limited to 'src/ceph/doc/man/8/ceph.rst')
-rw-r--r-- | src/ceph/doc/man/8/ceph.rst | 1550 |
1 files changed, 0 insertions, 1550 deletions
diff --git a/src/ceph/doc/man/8/ceph.rst b/src/ceph/doc/man/8/ceph.rst deleted file mode 100644 index 1a18000..0000000 --- a/src/ceph/doc/man/8/ceph.rst +++ /dev/null @@ -1,1550 +0,0 @@ -:orphan: - -================================== - ceph -- ceph administration tool -================================== - -.. program:: ceph - -Synopsis -======== - -| **ceph** **auth** [ *add* \| *caps* \| *del* \| *export* \| *get* \| *get-key* \| *get-or-create* \| *get-or-create-key* \| *import* \| *list* \| *print-key* \| *print_key* ] ... - -| **ceph** **compact** - -| **ceph** **config-key** [ *del* | *exists* | *get* | *list* | *dump* | *put* ] ... - -| **ceph** **daemon** *<name>* \| *<path>* *<command>* ... - -| **ceph** **daemonperf** *<name>* \| *<path>* [ *interval* [ *count* ] ] - -| **ceph** **df** *{detail}* - -| **ceph** **fs** [ *ls* \| *new* \| *reset* \| *rm* ] ... - -| **ceph** **fsid** - -| **ceph** **health** *{detail}* - -| **ceph** **heap** [ *dump* \| *start_profiler* \| *stop_profiler* \| *release* \| *stats* ] ... - -| **ceph** **injectargs** *<injectedargs>* [ *<injectedargs>*... ] - -| **ceph** **log** *<logtext>* [ *<logtext>*... ] - -| **ceph** **mds** [ *compat* \| *deactivate* \| *fail* \| *rm* \| *rmfailed* \| *set_state* \| *stat* \| *tell* ] ... - -| **ceph** **mon** [ *add* \| *dump* \| *getmap* \| *remove* \| *stat* ] ... - -| **ceph** **mon_status** - -| **ceph** **osd** [ *blacklist* \| *blocked-by* \| *create* \| *new* \| *deep-scrub* \| *df* \| *down* \| *dump* \| *erasure-code-profile* \| *find* \| *getcrushmap* \| *getmap* \| *getmaxosd* \| *in* \| *lspools* \| *map* \| *metadata* \| *ok-to-stop* \| *out* \| *pause* \| *perf* \| *pg-temp* \| *force-create-pg* \| *primary-affinity* \| *primary-temp* \| *repair* \| *reweight* \| *reweight-by-pg* \| *rm* \| *destroy* \| *purge* \| *safe-to-destroy* \| *scrub* \| *set* \| *setcrushmap* \| *setmaxosd* \| *stat* \| *tree* \| *unpause* \| *unset* ] ... - -| **ceph** **osd** **crush** [ *add* \| *add-bucket* \| *create-or-move* \| *dump* \| *get-tunable* \| *link* \| *move* \| *remove* \| *rename-bucket* \| *reweight* \| *reweight-all* \| *reweight-subtree* \| *rm* \| *rule* \| *set* \| *set-tunable* \| *show-tunables* \| *tunables* \| *unlink* ] ... - -| **ceph** **osd** **pool** [ *create* \| *delete* \| *get* \| *get-quota* \| *ls* \| *mksnap* \| *rename* \| *rmsnap* \| *set* \| *set-quota* \| *stats* ] ... - -| **ceph** **osd** **tier** [ *add* \| *add-cache* \| *cache-mode* \| *remove* \| *remove-overlay* \| *set-overlay* ] ... - -| **ceph** **pg** [ *debug* \| *deep-scrub* \| *dump* \| *dump_json* \| *dump_pools_json* \| *dump_stuck* \| *force_create_pg* \| *getmap* \| *ls* \| *ls-by-osd* \| *ls-by-pool* \| *ls-by-primary* \| *map* \| *repair* \| *scrub* \| *set_full_ratio* \| *set_nearfull_ratio* \| *stat* ] ... - -| **ceph** **quorum** [ *enter* \| *exit* ] - -| **ceph** **quorum_status** - -| **ceph** **report** { *<tags>* [ *<tags>...* ] } - -| **ceph** **scrub** - -| **ceph** **status** - -| **ceph** **sync** **force** {--yes-i-really-mean-it} {--i-know-what-i-am-doing} - -| **ceph** **tell** *<name (type.id)> <args> [<args>...]* - -| **ceph** **version** - -Description -=========== - -:program:`ceph` is a control utility which is used for manual deployment and maintenance -of a Ceph cluster. It provides a diverse set of commands that allows deployment of -monitors, OSDs, placement groups, MDS and overall maintenance, administration -of the cluster. - -Commands -======== - -auth ----- - -Manage authentication keys. It is used for adding, removing, exporting -or updating of authentication keys for a particular entity such as a monitor or -OSD. It uses some additional subcommands. - -Subcommand ``add`` adds authentication info for a particular entity from input -file, or random key if no input is given and/or any caps specified in the command. - -Usage:: - - ceph auth add <entity> {<caps> [<caps>...]} - -Subcommand ``caps`` updates caps for **name** from caps specified in the command. - -Usage:: - - ceph auth caps <entity> <caps> [<caps>...] - -Subcommand ``del`` deletes all caps for ``name``. - -Usage:: - - ceph auth del <entity> - -Subcommand ``export`` writes keyring for requested entity, or master keyring if -none given. - -Usage:: - - ceph auth export {<entity>} - -Subcommand ``get`` writes keyring file with requested key. - -Usage:: - - ceph auth get <entity> - -Subcommand ``get-key`` displays requested key. - -Usage:: - - ceph auth get-key <entity> - -Subcommand ``get-or-create`` adds authentication info for a particular entity -from input file, or random key if no input given and/or any caps specified in the -command. - -Usage:: - - ceph auth get-or-create <entity> {<caps> [<caps>...]} - -Subcommand ``get-or-create-key`` gets or adds key for ``name`` from system/caps -pairs specified in the command. If key already exists, any given caps must match -the existing caps for that key. - -Usage:: - - ceph auth get-or-create-key <entity> {<caps> [<caps>...]} - -Subcommand ``import`` reads keyring from input file. - -Usage:: - - ceph auth import - -Subcommand ``ls`` lists authentication state. - -Usage:: - - ceph auth ls - -Subcommand ``print-key`` displays requested key. - -Usage:: - - ceph auth print-key <entity> - -Subcommand ``print_key`` displays requested key. - -Usage:: - - ceph auth print_key <entity> - - -compact -------- - -Causes compaction of monitor's leveldb storage. - -Usage:: - - ceph compact - - -config-key ----------- - -Manage configuration key. It uses some additional subcommands. - -Subcommand ``del`` deletes configuration key. - -Usage:: - - ceph config-key del <key> - -Subcommand ``exists`` checks for configuration keys existence. - -Usage:: - - ceph config-key exists <key> - -Subcommand ``get`` gets the configuration key. - -Usage:: - - ceph config-key get <key> - -Subcommand ``list`` lists configuration keys. - -Usage:: - - ceph config-key ls - -Subcommand ``dump`` dumps configuration keys and values. - -Usage:: - - ceph config-key dump - -Subcommand ``set`` puts configuration key and value. - -Usage:: - - ceph config-key set <key> {<val>} - - -daemon ------- - -Submit admin-socket commands. - -Usage:: - - ceph daemon {daemon_name|socket_path} {command} ... - -Example:: - - ceph daemon osd.0 help - - -daemonperf ----------- - -Watch performance counters from a Ceph daemon. - -Usage:: - - ceph daemonperf {daemon_name|socket_path} [{interval} [{count}]] - - -df --- - -Show cluster's free space status. - -Usage:: - - ceph df {detail} - -.. _ceph features: - -features --------- - -Show the releases and features of all connected daemons and clients connected -to the cluster, along with the numbers of them in each bucket grouped by the -corresponding features/releases. Each release of Ceph supports a different set -of features, expressed by the features bitmask. New cluster features require -that clients support the feature, or else they are not allowed to connect to -these new features. As new features or capabilities are enabled after an -upgrade, older clients are prevented from connecting. - -Usage:: - - ceph features - -fs --- - -Manage cephfs filesystems. It uses some additional subcommands. - -Subcommand ``ls`` to list filesystems - -Usage:: - - ceph fs ls - -Subcommand ``new`` to make a new filesystem using named pools <metadata> and <data> - -Usage:: - - ceph fs new <fs_name> <metadata> <data> - -Subcommand ``reset`` is used for disaster recovery only: reset to a single-MDS map - -Usage:: - - ceph fs reset <fs_name> {--yes-i-really-mean-it} - -Subcommand ``rm`` to disable the named filesystem - -Usage:: - - ceph fs rm <fs_name> {--yes-i-really-mean-it} - - -fsid ----- - -Show cluster's FSID/UUID. - -Usage:: - - ceph fsid - - -health ------- - -Show cluster's health. - -Usage:: - - ceph health {detail} - - -heap ----- - -Show heap usage info (available only if compiled with tcmalloc) - -Usage:: - - ceph heap dump|start_profiler|stop_profiler|release|stats - - -injectargs ----------- - -Inject configuration arguments into monitor. - -Usage:: - - ceph injectargs <injected_args> [<injected_args>...] - - -log ---- - -Log supplied text to the monitor log. - -Usage:: - - ceph log <logtext> [<logtext>...] - - -mds ---- - -Manage metadata server configuration and administration. It uses some -additional subcommands. - -Subcommand ``compat`` manages compatible features. It uses some additional -subcommands. - -Subcommand ``rm_compat`` removes compatible feature. - -Usage:: - - ceph mds compat rm_compat <int[0-]> - -Subcommand ``rm_incompat`` removes incompatible feature. - -Usage:: - - ceph mds compat rm_incompat <int[0-]> - -Subcommand ``show`` shows mds compatibility settings. - -Usage:: - - ceph mds compat show - -Subcommand ``deactivate`` stops mds. - -Usage:: - - ceph mds deactivate <who> - -Subcommand ``fail`` forces mds to status fail. - -Usage:: - - ceph mds fail <who> - -Subcommand ``rm`` removes inactive mds. - -Usage:: - - ceph mds rm <int[0-]> <name> (type.id)> - -Subcommand ``rmfailed`` removes failed mds. - -Usage:: - - ceph mds rmfailed <int[0-]> - -Subcommand ``set_state`` sets mds state of <gid> to <numeric-state>. - -Usage:: - - ceph mds set_state <int[0-]> <int[0-20]> - -Subcommand ``stat`` shows MDS status. - -Usage:: - - ceph mds stat - -Subcommand ``tell`` sends command to particular mds. - -Usage:: - - ceph mds tell <who> <args> [<args>...] - -mon ---- - -Manage monitor configuration and administration. It uses some additional -subcommands. - -Subcommand ``add`` adds new monitor named <name> at <addr>. - -Usage:: - - ceph mon add <name> <IPaddr[:port]> - -Subcommand ``dump`` dumps formatted monmap (optionally from epoch) - -Usage:: - - ceph mon dump {<int[0-]>} - -Subcommand ``getmap`` gets monmap. - -Usage:: - - ceph mon getmap {<int[0-]>} - -Subcommand ``remove`` removes monitor named <name>. - -Usage:: - - ceph mon remove <name> - -Subcommand ``stat`` summarizes monitor status. - -Usage:: - - ceph mon stat - -mon_status ----------- - -Reports status of monitors. - -Usage:: - - ceph mon_status - -mgr ---- - -Ceph manager daemon configuration and management. - -Subcommand ``dump`` dumps the latest MgrMap, which describes the active -and standby manager daemons. - -Usage:: - - ceph mgr dump - -Subcommand ``fail`` will mark a manager daemon as failed, removing it -from the manager map. If it is the active manager daemon a standby -will take its place. - -Usage:: - - ceph mgr fail <name> - -Subcommand ``module ls`` will list currently enabled manager modules (plugins). - -Usage:: - - ceph mgr module ls - -Subcommand ``module enable`` will enable a manager module. Available modules are included in MgrMap and visible via ``mgr dump``. - -Usage:: - - ceph mgr module enable <module> - -Subcommand ``module disable`` will disable an active manager module. - -Usage:: - - ceph mgr module disable <module> - -Subcommand ``metadata`` will report metadata about all manager daemons or, if the name is specified, a single manager daemon. - -Usage:: - - ceph mgr metadata [name] - -Subcommand ``versions`` will report a count of running daemon versions. - -Usage:: - - ceph mgr versions - -Subcommand ``count-metadata`` will report a count of any daemon metadata field. - -Usage:: - - ceph mgr count-metadata <field> - - -osd ---- - -Manage OSD configuration and administration. It uses some additional -subcommands. - -Subcommand ``blacklist`` manage blacklisted clients. It uses some additional -subcommands. - -Subcommand ``add`` add <addr> to blacklist (optionally until <expire> seconds -from now) - -Usage:: - - ceph osd blacklist add <EntityAddr> {<float[0.0-]>} - -Subcommand ``ls`` show blacklisted clients - -Usage:: - - ceph osd blacklist ls - -Subcommand ``rm`` remove <addr> from blacklist - -Usage:: - - ceph osd blacklist rm <EntityAddr> - -Subcommand ``blocked-by`` prints a histogram of which OSDs are blocking their peers - -Usage:: - - ceph osd blocked-by - -Subcommand ``create`` creates new osd (with optional UUID and ID). - -This command is DEPRECATED as of the Luminous release, and will be removed in -a future release. - -Subcommand ``new`` should instead be used. - -Usage:: - - ceph osd create {<uuid>} {<id>} - -Subcommand ``new`` can be used to create a new OSD or to recreate a previously -destroyed OSD with a specific *id*. The new OSD will have the specified *uuid*, -and the command expects a JSON file containing the base64 cephx key for auth -entity *client.osd.<id>*, as well as optional base64 cepx key for dm-crypt -lockbox access and a dm-crypt key. Specifying a dm-crypt requires specifying -the accompanying lockbox cephx key. - -Usage:: - - ceph osd new {<uuid>} {<id>} -i {<secrets.json>} - -The secrets JSON file is optional but if provided, is expected to maintain -a form of the following format:: - - { - "cephx_secret": "AQBWtwhZdBO5ExAAIDyjK2Bh16ZXylmzgYYEjg==" - } - -Or:: - - { - "cephx_secret": "AQBWtwhZdBO5ExAAIDyjK2Bh16ZXylmzgYYEjg==", - "cephx_lockbox_secret": "AQDNCglZuaeVCRAAYr76PzR1Anh7A0jswkODIQ==", - "dmcrypt_key": "<dm-crypt key>" - } - - -Subcommand ``crush`` is used for CRUSH management. It uses some additional -subcommands. - -Subcommand ``add`` adds or updates crushmap position and weight for <name> with -<weight> and location <args>. - -Usage:: - - ceph osd crush add <osdname (id|osd.id)> <float[0.0-]> <args> [<args>...] - -Subcommand ``add-bucket`` adds no-parent (probably root) crush bucket <name> of -type <type>. - -Usage:: - - ceph osd crush add-bucket <name> <type> - -Subcommand ``create-or-move`` creates entry or moves existing entry for <name> -<weight> at/to location <args>. - -Usage:: - - ceph osd crush create-or-move <osdname (id|osd.id)> <float[0.0-]> <args> - [<args>...] - -Subcommand ``dump`` dumps crush map. - -Usage:: - - ceph osd crush dump - -Subcommand ``get-tunable`` get crush tunable straw_calc_version - -Usage:: - - ceph osd crush get-tunable straw_calc_version - -Subcommand ``link`` links existing entry for <name> under location <args>. - -Usage:: - - ceph osd crush link <name> <args> [<args>...] - -Subcommand ``move`` moves existing entry for <name> to location <args>. - -Usage:: - - ceph osd crush move <name> <args> [<args>...] - -Subcommand ``remove`` removes <name> from crush map (everywhere, or just at -<ancestor>). - -Usage:: - - ceph osd crush remove <name> {<ancestor>} - -Subcommand ``rename-bucket`` renames buchket <srcname> to <stname> - -Usage:: - - ceph osd crush rename-bucket <srcname> <dstname> - -Subcommand ``reweight`` change <name>'s weight to <weight> in crush map. - -Usage:: - - ceph osd crush reweight <name> <float[0.0-]> - -Subcommand ``reweight-all`` recalculate the weights for the tree to -ensure they sum correctly - -Usage:: - - ceph osd crush reweight-all - -Subcommand ``reweight-subtree`` changes all leaf items beneath <name> -to <weight> in crush map - -Usage:: - - ceph osd crush reweight-subtree <name> <weight> - -Subcommand ``rm`` removes <name> from crush map (everywhere, or just at -<ancestor>). - -Usage:: - - ceph osd crush rm <name> {<ancestor>} - -Subcommand ``rule`` is used for creating crush rules. It uses some additional -subcommands. - -Subcommand ``create-erasure`` creates crush rule <name> for erasure coded pool -created with <profile> (default default). - -Usage:: - - ceph osd crush rule create-erasure <name> {<profile>} - -Subcommand ``create-simple`` creates crush rule <name> to start from <root>, -replicate across buckets of type <type>, using a choose mode of <firstn|indep> -(default firstn; indep best for erasure pools). - -Usage:: - - ceph osd crush rule create-simple <name> <root> <type> {firstn|indep} - -Subcommand ``dump`` dumps crush rule <name> (default all). - -Usage:: - - ceph osd crush rule dump {<name>} - -Subcommand ``ls`` lists crush rules. - -Usage:: - - ceph osd crush rule ls - -Subcommand ``rm`` removes crush rule <name>. - -Usage:: - - ceph osd crush rule rm <name> - -Subcommand ``set`` used alone, sets crush map from input file. - -Usage:: - - ceph osd crush set - -Subcommand ``set`` with osdname/osd.id update crushmap position and weight -for <name> to <weight> with location <args>. - -Usage:: - - ceph osd crush set <osdname (id|osd.id)> <float[0.0-]> <args> [<args>...] - -Subcommand ``set-tunable`` set crush tunable <tunable> to <value>. The only -tunable that can be set is straw_calc_version. - -Usage:: - - ceph osd crush set-tunable straw_calc_version <value> - -Subcommand ``show-tunables`` shows current crush tunables. - -Usage:: - - ceph osd crush show-tunables - -Subcommand ``tree`` shows the crush buckets and items in a tree view. - -Usage:: - - ceph osd crush tree - -Subcommand ``tunables`` sets crush tunables values to <profile>. - -Usage:: - - ceph osd crush tunables legacy|argonaut|bobtail|firefly|hammer|optimal|default - -Subcommand ``unlink`` unlinks <name> from crush map (everywhere, or just at -<ancestor>). - -Usage:: - - ceph osd crush unlink <name> {<ancestor>} - -Subcommand ``df`` shows OSD utilization - -Usage:: - - ceph osd df {plain|tree} - -Subcommand ``deep-scrub`` initiates deep scrub on specified osd. - -Usage:: - - ceph osd deep-scrub <who> - -Subcommand ``down`` sets osd(s) <id> [<id>...] down. - -Usage:: - - ceph osd down <ids> [<ids>...] - -Subcommand ``dump`` prints summary of OSD map. - -Usage:: - - ceph osd dump {<int[0-]>} - -Subcommand ``erasure-code-profile`` is used for managing the erasure code -profiles. It uses some additional subcommands. - -Subcommand ``get`` gets erasure code profile <name>. - -Usage:: - - ceph osd erasure-code-profile get <name> - -Subcommand ``ls`` lists all erasure code profiles. - -Usage:: - - ceph osd erasure-code-profile ls - -Subcommand ``rm`` removes erasure code profile <name>. - -Usage:: - - ceph osd erasure-code-profile rm <name> - -Subcommand ``set`` creates erasure code profile <name> with [<key[=value]> ...] -pairs. Add a --force at the end to override an existing profile (IT IS RISKY). - -Usage:: - - ceph osd erasure-code-profile set <name> {<profile> [<profile>...]} - -Subcommand ``find`` find osd <id> in the CRUSH map and shows its location. - -Usage:: - - ceph osd find <int[0-]> - -Subcommand ``getcrushmap`` gets CRUSH map. - -Usage:: - - ceph osd getcrushmap {<int[0-]>} - -Subcommand ``getmap`` gets OSD map. - -Usage:: - - ceph osd getmap {<int[0-]>} - -Subcommand ``getmaxosd`` shows largest OSD id. - -Usage:: - - ceph osd getmaxosd - -Subcommand ``in`` sets osd(s) <id> [<id>...] in. - -Usage:: - - ceph osd in <ids> [<ids>...] - -Subcommand ``lost`` marks osd as permanently lost. THIS DESTROYS DATA IF NO -MORE REPLICAS EXIST, BE CAREFUL. - -Usage:: - - ceph osd lost <int[0-]> {--yes-i-really-mean-it} - -Subcommand ``ls`` shows all OSD ids. - -Usage:: - - ceph osd ls {<int[0-]>} - -Subcommand ``lspools`` lists pools. - -Usage:: - - ceph osd lspools {<int>} - -Subcommand ``map`` finds pg for <object> in <pool>. - -Usage:: - - ceph osd map <poolname> <objectname> - -Subcommand ``metadata`` fetches metadata for osd <id>. - -Usage:: - - ceph osd metadata {int[0-]} (default all) - -Subcommand ``out`` sets osd(s) <id> [<id>...] out. - -Usage:: - - ceph osd out <ids> [<ids>...] - -Subcommand ``ok-to-stop`` checks whether the list of OSD(s) can be -stopped without immediately making data unavailable. That is, all -data should remain readable and writeable, although data redundancy -may be reduced as some PGs may end up in a degraded (but active) -state. It will return a success code if it is okay to stop the -OSD(s), or an error code and informative message if it is not or if no -conclusion can be drawn at the current time. - -Usage:: - - ceph osd ok-to-stop <id> [<ids>...] - -Subcommand ``pause`` pauses osd. - -Usage:: - - ceph osd pause - -Subcommand ``perf`` prints dump of OSD perf summary stats. - -Usage:: - - ceph osd perf - -Subcommand ``pg-temp`` set pg_temp mapping pgid:[<id> [<id>...]] (developers -only). - -Usage:: - - ceph osd pg-temp <pgid> {<id> [<id>...]} - -Subcommand ``force-create-pg`` forces creation of pg <pgid>. - -Usage:: - - ceph osd force-create-pg <pgid> - - -Subcommand ``pool`` is used for managing data pools. It uses some additional -subcommands. - -Subcommand ``create`` creates pool. - -Usage:: - - ceph osd pool create <poolname> <int[0-]> {<int[0-]>} {replicated|erasure} - {<erasure_code_profile>} {<ruleset>} {<int>} - -Subcommand ``delete`` deletes pool. - -Usage:: - - ceph osd pool delete <poolname> {<poolname>} {--yes-i-really-really-mean-it} - -Subcommand ``get`` gets pool parameter <var>. - -Usage:: - - ceph osd pool get <poolname> size|min_size|crash_replay_interval|pg_num| - pgp_num|crush_ruleset|auid|write_fadvise_dontneed - -Only for tiered pools:: - - ceph osd pool get <poolname> hit_set_type|hit_set_period|hit_set_count|hit_set_fpp| - target_max_objects|target_max_bytes|cache_target_dirty_ratio|cache_target_dirty_high_ratio| - cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age| - min_read_recency_for_promote|hit_set_grade_decay_rate|hit_set_search_last_n - -Only for erasure coded pools:: - - ceph osd pool get <poolname> erasure_code_profile - -Use ``all`` to get all pool parameters that apply to the pool's type:: - - ceph osd pool get <poolname> all - -Subcommand ``get-quota`` obtains object or byte limits for pool. - -Usage:: - - ceph osd pool get-quota <poolname> - -Subcommand ``ls`` list pools - -Usage:: - - ceph osd pool ls {detail} - -Subcommand ``mksnap`` makes snapshot <snap> in <pool>. - -Usage:: - - ceph osd pool mksnap <poolname> <snap> - -Subcommand ``rename`` renames <srcpool> to <destpool>. - -Usage:: - - ceph osd pool rename <poolname> <poolname> - -Subcommand ``rmsnap`` removes snapshot <snap> from <pool>. - -Usage:: - - ceph osd pool rmsnap <poolname> <snap> - -Subcommand ``set`` sets pool parameter <var> to <val>. - -Usage:: - - ceph osd pool set <poolname> size|min_size|crash_replay_interval|pg_num| - pgp_num|crush_ruleset|hashpspool|nodelete|nopgchange|nosizechange| - hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|debug_fake_ec_pool| - target_max_bytes|target_max_objects|cache_target_dirty_ratio| - cache_target_dirty_high_ratio| - cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid| - min_read_recency_for_promote|write_fadvise_dontneed|hit_set_grade_decay_rate| - hit_set_search_last_n - <val> {--yes-i-really-mean-it} - -Subcommand ``set-quota`` sets object or byte limit on pool. - -Usage:: - - ceph osd pool set-quota <poolname> max_objects|max_bytes <val> - -Subcommand ``stats`` obtain stats from all pools, or from specified pool. - -Usage:: - - ceph osd pool stats {<name>} - -Subcommand ``primary-affinity`` adjust osd primary-affinity from 0.0 <=<weight> -<= 1.0 - -Usage:: - - ceph osd primary-affinity <osdname (id|osd.id)> <float[0.0-1.0]> - -Subcommand ``primary-temp`` sets primary_temp mapping pgid:<id>|-1 (developers -only). - -Usage:: - - ceph osd primary-temp <pgid> <id> - -Subcommand ``repair`` initiates repair on a specified osd. - -Usage:: - - ceph osd repair <who> - -Subcommand ``reweight`` reweights osd to 0.0 < <weight> < 1.0. - -Usage:: - - osd reweight <int[0-]> <float[0.0-1.0]> - -Subcommand ``reweight-by-pg`` reweight OSDs by PG distribution -[overload-percentage-for-consideration, default 120]. - -Usage:: - - ceph osd reweight-by-pg {<int[100-]>} {<poolname> [<poolname...]} - {--no-increasing} - -Subcommand ``reweight-by-utilization`` reweight OSDs by utilization -[overload-percentage-for-consideration, default 120]. - -Usage:: - - ceph osd reweight-by-utilization {<int[100-]>} - {--no-increasing} - -Subcommand ``rm`` removes osd(s) <id> [<id>...] from the OSD map. - - -Usage:: - - ceph osd rm <ids> [<ids>...] - -Subcommand ``destroy`` marks OSD *id* as *destroyed*, removing its cephx -entity's keys and all of its dm-crypt and daemon-private config key -entries. - -This command will not remove the OSD from crush, nor will it remove the -OSD from the OSD map. Instead, once the command successfully completes, -the OSD will show marked as *destroyed*. - -In order to mark an OSD as destroyed, the OSD must first be marked as -**lost**. - -Usage:: - - ceph osd destroy <id> {--yes-i-really-mean-it} - - -Subcommand ``purge`` performs a combination of ``osd destroy``, -``osd rm`` and ``osd crush remove``. - -Usage:: - - ceph osd purge <id> {--yes-i-really-mean-it} - -Subcommand ``safe-to-destroy`` checks whether it is safe to remove or -destroy an OSD without reducing overall data redundancy or durability. -It will return a success code if it is definitely safe, or an error -code and informative message if it is not or if no conclusion can be -drawn at the current time. - -Usage:: - - ceph osd safe-to-destroy <id> [<ids>...] - -Subcommand ``scrub`` initiates scrub on specified osd. - -Usage:: - - ceph osd scrub <who> - -Subcommand ``set`` sets <key>. - -Usage:: - - ceph osd set full|pause|noup|nodown|noout|noin|nobackfill| - norebalance|norecover|noscrub|nodeep-scrub|notieragent - -Subcommand ``setcrushmap`` sets crush map from input file. - -Usage:: - - ceph osd setcrushmap - -Subcommand ``setmaxosd`` sets new maximum osd value. - -Usage:: - - ceph osd setmaxosd <int[0-]> - -Subcommand ``set-require-min-compat-client`` enforces the cluster to be backward -compatible with the specified client version. This subcommand prevents you from -making any changes (e.g., crush tunables, or using new features) that -would violate the current setting. Please note, This subcommand will fail if -any connected daemon or client is not compatible with the features offered by -the given <version>. To see the features and releases of all clients connected -to cluster, please see `ceph features`_. - -Usage:: - - ceph osd set-require-min-compat-client <version> - -Subcommand ``stat`` prints summary of OSD map. - -Usage:: - - ceph osd stat - -Subcommand ``tier`` is used for managing tiers. It uses some additional -subcommands. - -Subcommand ``add`` adds the tier <tierpool> (the second one) to base pool <pool> -(the first one). - -Usage:: - - ceph osd tier add <poolname> <poolname> {--force-nonempty} - -Subcommand ``add-cache`` adds a cache <tierpool> (the second one) of size <size> -to existing pool <pool> (the first one). - -Usage:: - - ceph osd tier add-cache <poolname> <poolname> <int[0-]> - -Subcommand ``cache-mode`` specifies the caching mode for cache tier <pool>. - -Usage:: - - ceph osd tier cache-mode <poolname> none|writeback|forward|readonly| - readforward|readproxy - -Subcommand ``remove`` removes the tier <tierpool> (the second one) from base pool -<pool> (the first one). - -Usage:: - - ceph osd tier remove <poolname> <poolname> - -Subcommand ``remove-overlay`` removes the overlay pool for base pool <pool>. - -Usage:: - - ceph osd tier remove-overlay <poolname> - -Subcommand ``set-overlay`` set the overlay pool for base pool <pool> to be -<overlaypool>. - -Usage:: - - ceph osd tier set-overlay <poolname> <poolname> - -Subcommand ``tree`` prints OSD tree. - -Usage:: - - ceph osd tree {<int[0-]>} - -Subcommand ``unpause`` unpauses osd. - -Usage:: - - ceph osd unpause - -Subcommand ``unset`` unsets <key>. - -Usage:: - - ceph osd unset full|pause|noup|nodown|noout|noin|nobackfill| - norebalance|norecover|noscrub|nodeep-scrub|notieragent - - -pg --- - -It is used for managing the placement groups in OSDs. It uses some -additional subcommands. - -Subcommand ``debug`` shows debug info about pgs. - -Usage:: - - ceph pg debug unfound_objects_exist|degraded_pgs_exist - -Subcommand ``deep-scrub`` starts deep-scrub on <pgid>. - -Usage:: - - ceph pg deep-scrub <pgid> - -Subcommand ``dump`` shows human-readable versions of pg map (only 'all' valid -with plain). - -Usage:: - - ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief} [{all|summary|sum|delta|pools|osds|pgs|pgs_brief...]} - -Subcommand ``dump_json`` shows human-readable version of pg map in json only. - -Usage:: - - ceph pg dump_json {all|summary|sum|delta|pools|osds|pgs|pgs_brief} [{all|summary|sum|delta|pools|osds|pgs|pgs_brief...]} - -Subcommand ``dump_pools_json`` shows pg pools info in json only. - -Usage:: - - ceph pg dump_pools_json - -Subcommand ``dump_stuck`` shows information about stuck pgs. - -Usage:: - - ceph pg dump_stuck {inactive|unclean|stale|undersized|degraded [inactive|unclean|stale|undersized|degraded...]} - {<int>} - -Subcommand ``getmap`` gets binary pg map to -o/stdout. - -Usage:: - - ceph pg getmap - -Subcommand ``ls`` lists pg with specific pool, osd, state - -Usage:: - - ceph pg ls {<int>} {active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale| remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized [active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale|remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized...]} - -Subcommand ``ls-by-osd`` lists pg on osd [osd] - -Usage:: - - ceph pg ls-by-osd <osdname (id|osd.id)> {<int>} - {active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale| remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized [active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale|remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized...]} - -Subcommand ``ls-by-pool`` lists pg with pool = [poolname] - -Usage:: - - ceph pg ls-by-pool <poolstr> {<int>} {active| - clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale| remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized [active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale|remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized...]} - -Subcommand ``ls-by-primary`` lists pg with primary = [osd] - -Usage:: - - ceph pg ls-by-primary <osdname (id|osd.id)> {<int>} - {active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale| remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized [active|clean|down|replay|splitting| - scrubbing|scrubq|degraded|inconsistent|peering|repair| - recovery|backfill_wait|incomplete|stale|remapped| - deep_scrub|backfill|backfill_toofull|recovery_wait| - undersized...]} - -Subcommand ``map`` shows mapping of pg to osds. - -Usage:: - - ceph pg map <pgid> - -Subcommand ``repair`` starts repair on <pgid>. - -Usage:: - - ceph pg repair <pgid> - -Subcommand ``scrub`` starts scrub on <pgid>. - -Usage:: - - ceph pg scrub <pgid> - -Subcommand ``set_full_ratio`` sets ratio at which pgs are considered full. - -Usage:: - - ceph pg set_full_ratio <float[0.0-1.0]> - -Subcommand ``set_backfillfull_ratio`` sets ratio at which pgs are considered too full to backfill. - -Usage:: - - ceph pg set_backfillfull_ratio <float[0.0-1.0]> - -Subcommand ``set_nearfull_ratio`` sets ratio at which pgs are considered nearly -full. - -Usage:: - - ceph pg set_nearfull_ratio <float[0.0-1.0]> - -Subcommand ``stat`` shows placement group status. - -Usage:: - - ceph pg stat - - -quorum ------- - -Cause MON to enter or exit quorum. - -Usage:: - - ceph quorum enter|exit - -Note: this only works on the MON to which the ``ceph`` command is connected. -If you want a specific MON to enter or exit quorum, use this syntax:: - - ceph tell mon.<id> quorum enter|exit - -quorum_status -------------- - -Reports status of monitor quorum. - -Usage:: - - ceph quorum_status - - -report ------- - -Reports full status of cluster, optional title tag strings. - -Usage:: - - ceph report {<tags> [<tags>...]} - - -scrub ------ - -Scrubs the monitor stores. - -Usage:: - - ceph scrub - - -status ------- - -Shows cluster status. - -Usage:: - - ceph status - - -sync force ----------- - -Forces sync of and clear monitor store. - -Usage:: - - ceph sync force {--yes-i-really-mean-it} {--i-know-what-i-am-doing} - - -tell ----- - -Sends a command to a specific daemon. - -Usage:: - - ceph tell <name (type.id)> <args> [<args>...] - - -List all available commands. - -Usage:: - - ceph tell <name (type.id)> help - -version -------- - -Show mon daemon version - -Usage:: - - ceph version - -Options -======= - -.. option:: -i infile - - will specify an input file to be passed along as a payload with the - command to the monitor cluster. This is only used for specific - monitor commands. - -.. option:: -o outfile - - will write any payload returned by the monitor cluster with its - reply to outfile. Only specific monitor commands (e.g. osd getmap) - return a payload. - -.. option:: -c ceph.conf, --conf=ceph.conf - - Use ceph.conf configuration file instead of the default - ``/etc/ceph/ceph.conf`` to determine monitor addresses during startup. - -.. option:: --id CLIENT_ID, --user CLIENT_ID - - Client id for authentication. - -.. option:: --name CLIENT_NAME, -n CLIENT_NAME - - Client name for authentication. - -.. option:: --cluster CLUSTER - - Name of the Ceph cluster. - -.. option:: --admin-daemon ADMIN_SOCKET, daemon DAEMON_NAME - - Submit admin-socket commands via admin sockets in /var/run/ceph. - -.. option:: --admin-socket ADMIN_SOCKET_NOPE - - You probably mean --admin-daemon - -.. option:: -s, --status - - Show cluster status. - -.. option:: -w, --watch - - Watch live cluster changes. - -.. option:: --watch-debug - - Watch debug events. - -.. option:: --watch-info - - Watch info events. - -.. option:: --watch-sec - - Watch security events. - -.. option:: --watch-warn - - Watch warning events. - -.. option:: --watch-error - - Watch error events. - -.. option:: --version, -v - - Display version. - -.. option:: --verbose - - Make verbose. - -.. option:: --concise - - Make less verbose. - -.. option:: -f {json,json-pretty,xml,xml-pretty,plain}, --format - - Format of output. - -.. option:: --connect-timeout CLUSTER_TIMEOUT - - Set a timeout for connecting to the cluster. - -.. option:: --no-increasing - - ``--no-increasing`` is off by default. So increasing the osd weight is allowed - using the ``reweight-by-utilization`` or ``test-reweight-by-utilization`` commands. - If this option is used with these commands, it will help not to increase osd weight - even the osd is under utilized. - - -Availability -============ - -:program:`ceph` is part of Ceph, a massively scalable, open-source, distributed storage system. Please refer to -the Ceph documentation at http://ceph.com/docs for more information. - - -See also -======== - -:doc:`ceph-mon <ceph-mon>`\(8), -:doc:`ceph-osd <ceph-osd>`\(8), -:doc:`ceph-mds <ceph-mds>`\(8) |