diff options
Diffstat (limited to 'src/ceph/doc/install/upgrading-ceph.rst')
-rw-r--r-- | src/ceph/doc/install/upgrading-ceph.rst | 746 |
1 files changed, 0 insertions, 746 deletions
diff --git a/src/ceph/doc/install/upgrading-ceph.rst b/src/ceph/doc/install/upgrading-ceph.rst deleted file mode 100644 index 962b90c..0000000 --- a/src/ceph/doc/install/upgrading-ceph.rst +++ /dev/null @@ -1,746 +0,0 @@ -================ - Upgrading Ceph -================ - -Each release of Ceph may have additional steps. Refer to the release-specific -sections in this document and the `release notes`_ document to identify -release-specific procedures for your cluster before using the upgrade -procedures. - - -Summary -======= - -You can upgrade daemons in your Ceph cluster while the cluster is online and in -service! Certain types of daemons depend upon others. For example, Ceph Metadata -Servers and Ceph Object Gateways depend upon Ceph Monitors and Ceph OSD Daemons. -We recommend upgrading in this order: - -#. `Ceph Deploy`_ -#. Ceph Monitors -#. Ceph OSD Daemons -#. Ceph Metadata Servers -#. Ceph Object Gateways - -As a general rule, we recommend upgrading all the daemons of a specific type -(e.g., all ``ceph-mon`` daemons, all ``ceph-osd`` daemons, etc.) to ensure that -they are all on the same release. We also recommend that you upgrade all the -daemons in your cluster before you try to exercise new functionality in a -release. - -The `Upgrade Procedures`_ are relatively simple, but please look at -distribution-specific sections before upgrading. The basic process involves -three steps: - -#. Use ``ceph-deploy`` on your admin node to upgrade the packages for - multiple hosts (using the ``ceph-deploy install`` command), or login to each - host and upgrade the Ceph package `manually`_. For example, when - `Upgrading Monitors`_, the ``ceph-deploy`` syntax might look like this:: - - ceph-deploy install --release {release-name} ceph-node1[ ceph-node2] - ceph-deploy install --release firefly mon1 mon2 mon3 - - **Note:** The ``ceph-deploy install`` command will upgrade the packages - in the specified node(s) from the old release to the release you specify. - There is no ``ceph-deploy upgrade`` command. - -#. Login in to each Ceph node and restart each Ceph daemon. - See `Operating a Cluster`_ for details. - -#. Ensure your cluster is healthy. See `Monitoring a Cluster`_ for details. - -.. important:: Once you upgrade a daemon, you cannot downgrade it. - - -Ceph Deploy -=========== - -Before upgrading Ceph daemons, upgrade the ``ceph-deploy`` tool. :: - - sudo pip install -U ceph-deploy - -Or:: - - sudo apt-get install ceph-deploy - -Or:: - - sudo yum install ceph-deploy python-pushy - - -Argonaut to Bobtail -=================== - -When upgrading from Argonaut to Bobtail, you need to be aware of several things: - -#. Authentication now defaults to **ON**, but used to default to **OFF**. -#. Monitors use a new internal on-wire protocol. -#. RBD ``format2`` images require upgrading all OSDs before using it. - -Ensure that you update package repository paths. For example:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-bobtail/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -See the following sections for additional details. - -Authentication --------------- - -The Ceph Bobtail release enables authentication by default. Bobtail also has -finer-grained authentication configuration settings. In previous versions of -Ceph (i.e., actually v 0.55 and earlier), you could simply specify:: - - auth supported = [cephx | none] - -This option still works, but is deprecated. New releases support -``cluster``, ``service`` and ``client`` authentication settings as -follows:: - - auth cluster required = [cephx | none] # default cephx - auth service required = [cephx | none] # default cephx - auth client required = [cephx | none] # default cephx,none - -.. important:: If your cluster does not currently have an ``auth - supported`` line that enables authentication, you must explicitly - turn it off in Bobtail using the settings below.:: - - auth cluster required = none - auth service required = none - - This will disable authentication on the cluster, but still leave - clients with the default configuration where they can talk to a - cluster that does enable it, but do not require it. - -.. important:: If your cluster already has an ``auth supported`` option defined in - the configuration file, no changes are necessary. - -See `User Management - Backward Compatibility`_ for details. - - -Monitor On-wire Protocol ------------------------- - -We recommend upgrading all monitors to Bobtail. A mixture of Bobtail and -Argonaut monitors will not be able to use the new on-wire protocol, as the -protocol requires all monitors to be Bobtail or greater. Upgrading only a -majority of the nodes (e.g., two out of three) may expose the cluster to a -situation where a single additional failure may compromise availability (because -the non-upgraded daemon cannot participate in the new protocol). We recommend -not waiting for an extended period of time between ``ceph-mon`` upgrades. - - -RBD Images ----------- - -The Bobtail release supports ``format 2`` images! However, you should not create -or use ``format 2`` RBD images until after all ``ceph-osd`` daemons have been -upgraded. Note that ``format 1`` is still the default. You can use the new -``ceph osd ls`` and ``ceph tell osd.N version`` commands to doublecheck your -cluster. ``ceph osd ls`` will give a list of all OSD IDs that are part of the -cluster, and you can use that to write a simple shell loop to display all the -OSD version strings: :: - - for i in $(ceph osd ls); do - ceph tell osd.${i} version - done - - -Argonaut to Cuttlefish -====================== - -To upgrade your cluster from Argonaut to Cuttlefish, please read this -section, and the sections on upgrading from Argonaut to Bobtail and -upgrading from Bobtail to Cuttlefish carefully. When upgrading from -Argonaut to Cuttlefish, **YOU MUST UPGRADE YOUR MONITORS FROM ARGONAUT -TO BOBTAIL v0.56.5 FIRST!!!**. All other Ceph daemons can upgrade from -Argonaut to Cuttlefish without the intermediate upgrade to Bobtail. - -.. important:: Ensure that the repository specified points to Bobtail, not - Cuttlefish. - -For example:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-bobtail/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -We recommend upgrading all monitors to Bobtail before proceeding with the -upgrade of the monitors to Cuttlefish. A mixture of Bobtail and Argonaut -monitors will not be able to use the new on-wire protocol, as the protocol -requires all monitors to be Bobtail or greater. Upgrading only a majority of the -nodes (e.g., two out of three) may expose the cluster to a situation where a -single additional failure may compromise availability (because the non-upgraded -daemon cannot participate in the new protocol). We recommend not waiting for an -extended period of time between ``ceph-mon`` upgrades. See `Upgrading -Monitors`_ for details. - -.. note:: See the `Authentication`_ section and the - `User Management - Backward Compatibility`_ for additional information - on authentication backward compatibility settings for Bobtail. - -Once you complete the upgrade of your monitors from Argonaut to -Bobtail, and have restarted the monitor daemons, you must upgrade the -monitors from Bobtail to Cuttlefish. Ensure that you have a quorum -before beginning this upgrade procedure. Before upgrading, remember to -replace the reference to the Bobtail repository with a reference to -the Cuttlefish repository. For example:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-cuttlefish/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -See `Upgrading Monitors`_ for details. - -The architecture of the monitors changed significantly from Argonaut to -Cuttlefish. See `Monitor Config Reference`_ and `Joao's blog post`_ for details. -Once you complete the monitor upgrade, you can upgrade the OSD daemons and the -MDS daemons using the generic procedures. See `Upgrading an OSD`_ and `Upgrading -a Metadata Server`_ for details. - - -Bobtail to Cuttlefish -===================== - -Upgrading your cluster from Bobtail to Cuttlefish has a few important -considerations. First, the monitor uses a new architecture, so you should -upgrade the full set of monitors to use Cuttlefish. Second, if you run multiple -metadata servers in a cluster, ensure the metadata servers have unique names. -See the following sections for details. - -Replace any ``apt`` reference to older repositories with a reference to the -Cuttlefish repository. For example:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-cuttlefish/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - - -Monitor -------- - -The architecture of the monitors changed significantly from Bobtail to -Cuttlefish. See `Monitor Config Reference`_ and `Joao's blog post`_ for -details. This means that v0.59 and pre-v0.59 monitors do not talk to each other -(Cuttlefish is v.0.61). When you upgrade each monitor, it will convert its -local data store to the new format. Once you upgrade a majority of monitors, -the monitors form a quorum using the new protocol and the old monitors will be -blocked until they get upgraded. For this reason, we recommend upgrading the -monitors in immediate succession. - -.. important:: Do not run a mixed-version cluster for an extended period. - - -MDS Unique Names ----------------- - -The monitor now enforces that MDS names be unique. If you have multiple metadata -server daemons that start with the same ID (e.g., mds.a) the second -metadata server will implicitly mark the first metadata server as ``failed``. -Multi-MDS configurations with identical names must be adjusted accordingly to -give daemons unique names. If you run your cluster with one metadata server, -you can disregard this notice for now. - - -ceph-deploy ------------ - -The ``ceph-deploy`` tool is now the preferred method of provisioning new clusters. -For existing clusters created via the obsolete ``mkcephfs`` tool that would like to transition to the -new tool, there is a migration path, documented at `Transitioning to ceph-deploy`_. - -Cuttlefish to Dumpling -====================== - -When upgrading from Cuttlefish (v0.61-v0.61.7) you may perform a rolling -upgrade. However, there are a few important considerations. First, you must -upgrade the ``ceph`` command line utility, because it has changed significantly. -Second, you must upgrade the full set of monitors to use Dumpling, because of a -protocol change. - -Replace any reference to older repositories with a reference to the -Dumpling repository. For example, with ``apt`` perform the following:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-dumpling/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -With CentOS/Red Hat distributions, remove the old repository. :: - - sudo rm /etc/yum.repos.d/ceph.repo - -Then add a new ``ceph.repo`` repository entry with the following contents. - -.. code-block:: ini - - [ceph] - name=Ceph Packages and Backports $basearch - baseurl=http://download.ceph.com/rpm/el6/$basearch - enabled=1 - gpgcheck=1 - gpgkey=https://download.ceph.com/keys/release.asc - - -.. note:: Ensure you use the correct URL for your distribution. Check the - http://download.ceph.com/rpm directory for your distribution. - -.. note:: Since you can upgrade using ``ceph-deploy`` you will only need to add - the repository on Ceph Client nodes where you use the ``ceph`` command line - interface or the ``ceph-deploy`` tool. - - -Dumpling to Emperor -=================== - -When upgrading from Dumpling (v0.64) you may perform a rolling -upgrade. - -Replace any reference to older repositories with a reference to the -Emperor repository. For example, with ``apt`` perform the following:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-emperor/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -With CentOS/Red Hat distributions, remove the old repository. :: - - sudo rm /etc/yum.repos.d/ceph.repo - -Then add a new ``ceph.repo`` repository entry with the following contents and -replace ``{distro}`` with your distribution (e.g., ``el6``, ``rhel6``, etc). - -.. code-block:: ini - - [ceph] - name=Ceph Packages and Backports $basearch - baseurl=http://download.ceph.com/rpm-emperor/{distro}/$basearch - enabled=1 - gpgcheck=1 - gpgkey=https://download.ceph.com/keys/release.asc - - -.. note:: Ensure you use the correct URL for your distribution. Check the - http://download.ceph.com/rpm directory for your distribution. - -.. note:: Since you can upgrade using ``ceph-deploy`` you will only need to add - the repository on Ceph Client nodes where you use the ``ceph`` command line - interface or the ``ceph-deploy`` tool. - - -Command Line Utility --------------------- - -In V0.65, the ``ceph`` commandline interface (CLI) utility changed -significantly. You will not be able to use the old CLI with Dumpling. This means -that you must upgrade the ``ceph-common`` library on all nodes that access the -Ceph Storage Cluster with the ``ceph`` CLI before upgrading Ceph daemons. :: - - sudo apt-get update && sudo apt-get install ceph-common - -Ensure that you have the latest version (v0.67 or later). If you do not, -you may need to uninstall, auto remove dependencies and reinstall. - -See `v0.65`_ for details on the new command line interface. - -.. _v0.65: http://docs.ceph.com/docs/master/release-notes/#v0-65 - - -Monitor -------- - -Dumpling (v0.67) ``ceph-mon`` daemons have an internal protocol change. This -means that v0.67 daemons cannot talk to v0.66 or older daemons. Once you -upgrade a majority of monitors, the monitors form a quorum using the new -protocol and the old monitors will be blocked until they get upgraded. For this -reason, we recommend upgrading all monitors at once (or in relatively quick -succession) to minimize the possibility of downtime. - -.. important:: Do not run a mixed-version cluster for an extended period. - - - -Dumpling to Firefly -=================== - -If your existing cluster is running a version older than v0.67 Dumpling, please -first upgrade to the latest Dumpling release before upgrading to v0.80 Firefly. - - -Monitor -------- - -Dumpling (v0.67) ``ceph-mon`` daemons have an internal protocol change. This -means that v0.67 daemons cannot talk to v0.66 or older daemons. Once you -upgrade a majority of monitors, the monitors form a quorum using the new -protocol and the old monitors will be blocked until they get upgraded. For this -reason, we recommend upgrading all monitors at once (or in relatively quick -succession) to minimize the possibility of downtime. - -.. important:: Do not run a mixed-version cluster for an extended period. - - -Ceph Config File Changes ------------------------- - -We recommand adding the following to the ``[mon]`` section of your -``ceph.conf`` prior to upgrade:: - - mon warn on legacy crush tunables = false - -This will prevent health warnings due to the use of legacy CRUSH placement. -Although it is possible to rebalance existing data across your cluster, we do -not normally recommend it for production environments as a large amount of data -will move and there is a significant performance impact from the rebalancing. - - -Command Line Utility --------------------- - -In V0.65, the ``ceph`` commandline interface (CLI) utility changed -significantly. You will not be able to use the old CLI with Firefly. This means -that you must upgrade the ``ceph-common`` library on all nodes that access the -Ceph Storage Cluster with the ``ceph`` CLI before upgrading Ceph daemons. - -For Debian/Ubuntu, execute:: - - sudo apt-get update && sudo apt-get install ceph-common - -For CentOS/RHEL, execute:: - - sudo yum install ceph-common - -Ensure that you have the latest version. If you do not, -you may need to uninstall, auto remove dependencies and reinstall. - -See `v0.65`_ for details on the new command line interface. - -.. _v0.65: http://docs.ceph.com/docs/master/release-notes/#v0-65 - - -Upgrade Sequence ----------------- - -Replace any reference to older repositories with a reference to the -Firely repository. For example, with ``apt`` perform the following:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-firefly/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -With CentOS/Red Hat distributions, remove the old repository. :: - - sudo rm /etc/yum.repos.d/ceph.repo - -Then add a new ``ceph.repo`` repository entry with the following contents and -replace ``{distro}`` with your distribution (e.g., ``el6``, ``rhel6``, -``rhel7``, etc.). - -.. code-block:: ini - - [ceph] - name=Ceph Packages and Backports $basearch - baseurl=http://download.ceph.com/rpm-firefly/{distro}/$basearch - enabled=1 - gpgcheck=1 - gpgkey=https://download.ceph.com/keys/release.asc - - -Upgrade daemons in the following order: - -#. **Monitors:** If the ``ceph-mon`` daemons are not restarted prior to the - ``ceph-osd`` daemons, the monitors will not correctly register their new - capabilities with the cluster and new features may not be usable until - the monitors are restarted a second time. - -#. **OSDs** - -#. **MDSs:** If the ``ceph-mds`` daemon is restarted first, it will wait until - all OSDs have been upgraded before finishing its startup sequence. - -#. **Gateways:** Upgrade ``radosgw`` daemons together. There is a subtle change - in behavior for multipart uploads that prevents a multipart request that - was initiated with a new ``radosgw`` from being completed by an old - ``radosgw``. - -.. note:: Make sure you upgrade your **ALL** of your Ceph monitors **AND** - restart them **BEFORE** upgrading and restarting OSDs, MDSs, and gateways! - - -Emperor to Firefly -================== - -If your existing cluster is running a version older than v0.67 Dumpling, please -first upgrade to the latest Dumpling release before upgrading to v0.80 Firefly. -Please refer to `Cuttlefish to Dumpling`_ and the `Firefly release notes`_ for -details. To upgrade from a post-Emperor point release, see the `Firefly release -notes`_ for details. - - -Ceph Config File Changes ------------------------- - -We recommand adding the following to the ``[mon]`` section of your -``ceph.conf`` prior to upgrade:: - - mon warn on legacy crush tunables = false - -This will prevent health warnings due to the use of legacy CRUSH placement. -Although it is possible to rebalance existing data across your cluster, we do -not normally recommend it for production environments as a large amount of data -will move and there is a significant performance impact from the rebalancing. - - -Upgrade Sequence ----------------- - -Replace any reference to older repositories with a reference to the -Firefly repository. For example, with ``apt`` perform the following:: - - sudo rm /etc/apt/sources.list.d/ceph.list - echo deb http://download.ceph.com/debian-firefly/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - -With CentOS/Red Hat distributions, remove the old repository. :: - - sudo rm /etc/yum.repos.d/ceph.repo - -Then add a new ``ceph.repo`` repository entry with the following contents, but -replace ``{distro}`` with your distribution (e.g., ``el6``, ``rhel6``, -``rhel7``, etc.). - -.. code-block:: ini - - [ceph] - name=Ceph Packages and Backports $basearch - baseurl=http://download.ceph.com/rpm/{distro}/$basearch - enabled=1 - gpgcheck=1 - gpgkey=https://download.ceph.com/keys/release.asc - - -.. note:: Ensure you use the correct URL for your distribution. Check the - http://download.ceph.com/rpm directory for your distribution. - -.. note:: Since you can upgrade using ``ceph-deploy`` you will only need to add - the repository on Ceph Client nodes where you use the ``ceph`` command line - interface or the ``ceph-deploy`` tool. - - -Upgrade daemons in the following order: - -#. **Monitors:** If the ``ceph-mon`` daemons are not restarted prior to the - ``ceph-osd`` daemons, the monitors will not correctly register their new - capabilities with the cluster and new features may not be usable until - the monitors are restarted a second time. - -#. **OSDs** - -#. **MDSs:** If the ``ceph-mds`` daemon is restarted first, it will wait until - all OSDs have been upgraded before finishing its startup sequence. - -#. **Gateways:** Upgrade ``radosgw`` daemons together. There is a subtle change - in behavior for multipart uploads that prevents a multipart request that - was initiated with a new ``radosgw`` from being completed by an old - ``radosgw``. - - -Upgrade Procedures -================== - -The following sections describe the upgrade process. - -.. important:: Each release of Ceph may have some additional steps. Refer to - release-specific sections for details **BEFORE** you begin upgrading daemons. - - -Upgrading Monitors ------------------- - -To upgrade monitors, perform the following steps: - -#. Upgrade the Ceph package for each daemon instance. - - You may use ``ceph-deploy`` to address all monitor nodes at once. - For example:: - - ceph-deploy install --release {release-name} ceph-node1[ ceph-node2] - ceph-deploy install --release hammer mon1 mon2 mon3 - - You may also use the package manager for your Linux distribution on - each individual node. To upgrade packages manually on each Debian/Ubuntu - host, perform the following steps . :: - - ssh {mon-host} - sudo apt-get update && sudo apt-get install ceph - - On CentOS/Red Hat hosts, perform the following steps:: - - ssh {mon-host} - sudo yum update && sudo yum install ceph - - -#. Restart each monitor. For Ubuntu distributions, use:: - - sudo restart ceph-mon id={hostname} - - For CentOS/Red Hat/Debian distributions, use:: - - sudo /etc/init.d/ceph restart {mon-id} - - For CentOS/Red Hat distributions deployed with ``ceph-deploy``, - the monitor ID is usually ``mon.{hostname}``. - -#. Ensure each monitor has rejoined the quorum. :: - - ceph mon stat - -Ensure that you have completed the upgrade cycle for all of your Ceph Monitors. - - -Upgrading an OSD ----------------- - -To upgrade a Ceph OSD Daemon, perform the following steps: - -#. Upgrade the Ceph OSD Daemon package. - - You may use ``ceph-deploy`` to address all Ceph OSD Daemon nodes at - once. For example:: - - ceph-deploy install --release {release-name} ceph-node1[ ceph-node2] - ceph-deploy install --release hammer osd1 osd2 osd3 - - You may also use the package manager on each node to upgrade packages - manually. For Debian/Ubuntu hosts, perform the following steps on each - host. :: - - ssh {osd-host} - sudo apt-get update && sudo apt-get install ceph - - For CentOS/Red Hat hosts, perform the following steps:: - - ssh {osd-host} - sudo yum update && sudo yum install ceph - - -#. Restart the OSD, where ``N`` is the OSD number. For Ubuntu, use:: - - sudo restart ceph-osd id=N - - For multiple OSDs on a host, you may restart all of them with Upstart. :: - - sudo restart ceph-osd-all - - For CentOS/Red Hat/Debian distributions, use:: - - sudo /etc/init.d/ceph restart N - - -#. Ensure each upgraded Ceph OSD Daemon has rejoined the cluster:: - - ceph osd stat - -Ensure that you have completed the upgrade cycle for all of your -Ceph OSD Daemons. - - -Upgrading a Metadata Server ---------------------------- - -To upgrade a Ceph Metadata Server, perform the following steps: - -#. Upgrade the Ceph Metadata Server package. You may use ``ceph-deploy`` to - address all Ceph Metadata Server nodes at once, or use the package manager - on each node. For example:: - - ceph-deploy install --release {release-name} ceph-node1 - ceph-deploy install --release hammer mds1 - - To upgrade packages manually, perform the following steps on each - Debian/Ubuntu host. :: - - ssh {mon-host} - sudo apt-get update && sudo apt-get install ceph-mds - - Or the following steps on CentOS/Red Hat hosts:: - - ssh {mon-host} - sudo yum update && sudo yum install ceph-mds - - -#. Restart the metadata server. For Ubuntu, use:: - - sudo restart ceph-mds id={hostname} - - For CentOS/Red Hat/Debian distributions, use:: - - sudo /etc/init.d/ceph restart mds.{hostname} - - For clusters deployed with ``ceph-deploy``, the name is usually either - the name you specified on creation or the hostname. - -#. Ensure the metadata server is up and running:: - - ceph mds stat - - -Upgrading a Client ------------------- - -Once you have upgraded the packages and restarted daemons on your Ceph -cluster, we recommend upgrading ``ceph-common`` and client libraries -(``librbd1`` and ``librados2``) on your client nodes too. - -#. Upgrade the package:: - - ssh {client-host} - apt-get update && sudo apt-get install ceph-common librados2 librbd1 python-rados python-rbd - -#. Ensure that you have the latest version:: - - ceph --version - -If you do not have the latest version, you may need to uninstall, auto remove -dependencies and reinstall. - - -Transitioning to ceph-deploy -============================ - -If you have an existing cluster that you deployed with ``mkcephfs`` (usually -Argonaut or Bobtail releases), you will need to make a few changes to your -configuration to ensure that your cluster will work with ``ceph-deploy``. - - -Monitor Keyring ---------------- - -You will need to add ``caps mon = "allow *"`` to your monitor keyring if it is -not already in the keyring. By default, the monitor keyring is located under -``/var/lib/ceph/mon/ceph-$id/keyring``. When you have added the ``caps`` -setting, your monitor keyring should look something like this:: - - [mon.] - key = AQBJIHhRuHCwDRAAZjBTSJcIBIoGpdOR9ToiyQ== - caps mon = "allow *" - -Adding ``caps mon = "allow *"`` will ease the transition from ``mkcephfs`` to -``ceph-deploy`` by allowing ``ceph-create-keys`` to use the ``mon.`` keyring -file in ``$mon_data`` and get the caps it needs. - - -Use Default Paths ------------------ - -Under the ``/var/lib/ceph`` directory, the ``mon`` and ``osd`` directories need -to use the default paths. - -- **OSDs**: The path should be ``/var/lib/ceph/osd/ceph-$id`` -- **MON**: The path should be ``/var/lib/ceph/mon/ceph-$id`` - -Under those directories, the keyring should be in a file named ``keyring``. - - - - -.. _Monitor Config Reference: ../../rados/configuration/mon-config-ref -.. _Joao's blog post: http://ceph.com/dev-notes/cephs-new-monitor-changes -.. _User Management - Backward Compatibility: ../../rados/configuration/auth-config-ref/#backward-compatibility -.. _manually: ../install-storage-cluster/ -.. _Operating a Cluster: ../../rados/operations/operating -.. _Monitoring a Cluster: ../../rados/operations/monitoring -.. _Firefly release notes: ../../release-notes/#v0-80-firefly -.. _release notes: ../../release-notes |