pveceph: Reorganize TOC for new sections

Put the previous added sections into subsection for a better outline of
the TOC.

With the rearrangement of the first level titles to second level, the
general descriptions of a service needs to move into the new first level
titles. And add/corrects some statements of those descriptions.

Signed-off-by: Alwin Antreich <a.antreich@proxmox.com>
This commit is contained in:
Alwin Antreich 2019-11-06 15:09:08 +01:00 committed by Thomas Lamprecht
parent c1f38fe3af
commit b3338e29cd

View File

@ -212,8 +212,8 @@ This sets up an `apt` package repository in
`/etc/apt/sources.list.d/ceph.list` and installs the required software. `/etc/apt/sources.list.d/ceph.list` and installs the required software.
Creating initial Ceph configuration Create initial Ceph configuration
----------------------------------- ---------------------------------
[thumbnail="screenshot/gui-ceph-config.png"] [thumbnail="screenshot/gui-ceph-config.png"]
@ -234,11 +234,8 @@ configuration file.
[[pve_ceph_monitors]] [[pve_ceph_monitors]]
Creating Ceph Monitors Ceph Monitor
---------------------- -----------
[thumbnail="screenshot/gui-ceph-monitor.png"]
The Ceph Monitor (MON) The Ceph Monitor (MON)
footnote:[Ceph Monitor http://docs.ceph.com/docs/luminous/start/intro/] footnote:[Ceph Monitor http://docs.ceph.com/docs/luminous/start/intro/]
maintains a master copy of the cluster map. For high availability you need to maintains a master copy of the cluster map. For high availability you need to
@ -247,6 +244,12 @@ used the installation wizard. You won't need more than 3 monitors as long
as your cluster is small to midsize, only really large clusters will as your cluster is small to midsize, only really large clusters will
need more than that. need more than that.
Create Monitors
~~~~~~~~~~~~~~~
[thumbnail="screenshot/gui-ceph-monitor.png"]
On each node where you want to place a monitor (three monitors are recommended), On each node where you want to place a monitor (three monitors are recommended),
create it by using the 'Ceph -> Monitor' tab in the GUI or run. create it by using the 'Ceph -> Monitor' tab in the GUI or run.
@ -256,12 +259,9 @@ create it by using the 'Ceph -> Monitor' tab in the GUI or run.
pveceph mon create pveceph mon create
---- ----
This will also install the needed Ceph Manager ('ceph-mgr') by default. If you
do not want to install a manager, specify the '-exclude-manager' option.
Destroy Monitors
Destroying Ceph Monitor ~~~~~~~~~~~~~~~~
----------------------
To remove a Ceph Monitor via the GUI first select a node in the tree view and To remove a Ceph Monitor via the GUI first select a node in the tree view and
go to the **Ceph -> Monitor** panel. Select the MON and click the **Destroy** go to the **Ceph -> Monitor** panel. Select the MON and click the **Destroy**
@ -278,14 +278,17 @@ NOTE: At least three Monitors are needed for quorum.
[[pve_ceph_manager]] [[pve_ceph_manager]]
Creating Ceph Manager Ceph Manager
---------------------- ------------
The Manager daemon runs alongside the monitors. It provides an interface to
monitor the cluster. Since the Ceph luminous release at least one ceph-mgr
footnote:[Ceph Manager http://docs.ceph.com/docs/luminous/mgr/] daemon is
required.
The Manager daemon runs alongside the monitors, providing an interface for Create Manager
monitoring the cluster. Since the Ceph luminous release the ~~~~~~~~~~~~~~
ceph-mgr footnote:[Ceph Manager http://docs.ceph.com/docs/luminous/mgr/] daemon
is required. During monitor installation the ceph manager will be installed as Multiple Managers can be installed, but at any time only one Manager is active.
well.
[source,bash] [source,bash]
---- ----
@ -296,8 +299,8 @@ NOTE: It is recommended to install the Ceph Manager on the monitor nodes. For
high availability install more then one manager. high availability install more then one manager.
Destroying Ceph Manager Destroy Manager
---------------------- ~~~~~~~~~~~~~~~
To remove a Ceph Manager via the GUI first select a node in the tree view and To remove a Ceph Manager via the GUI first select a node in the tree view and
go to the **Ceph -> Monitor** panel. Select the Manager and click the go to the **Ceph -> Monitor** panel. Select the Manager and click the
@ -315,8 +318,15 @@ the cluster status or usage require a running Manager.
[[pve_ceph_osds]] [[pve_ceph_osds]]
Creating Ceph OSDs Ceph OSDs
------------------ ---------
Ceph **O**bject **S**torage **D**aemons are storing objects for Ceph over the
network. It is recommended to use one OSD per physical disk.
NOTE: By default an object is 4 MiB in size.
Create OSDs
~~~~~~~~~~~
[thumbnail="screenshot/gui-ceph-osd-status.png"] [thumbnail="screenshot/gui-ceph-osd-status.png"]
@ -327,8 +337,8 @@ via GUI or via CLI as follows:
pveceph osd create /dev/sd[X] pveceph osd create /dev/sd[X]
---- ----
TIP: We recommend a Ceph cluster size, starting with 12 OSDs, distributed evenly TIP: We recommend a Ceph cluster size, starting with 12 OSDs, distributed
among your, at least three nodes (4 OSDs on each node). evenly among your, at least three nodes (4 OSDs on each node).
If the disk was used before (eg. ZFS/RAID/OSD), to remove partition table, boot If the disk was used before (eg. ZFS/RAID/OSD), to remove partition table, boot
sector and any OSD leftover the following command should be sufficient. sector and any OSD leftover the following command should be sufficient.
@ -340,8 +350,7 @@ ceph-volume lvm zap /dev/sd[X] --destroy
WARNING: The above command will destroy data on the disk! WARNING: The above command will destroy data on the disk!
Ceph Bluestore .Ceph Bluestore
~~~~~~~~~~~~~~
Starting with the Ceph Kraken release, a new Ceph OSD storage type was Starting with the Ceph Kraken release, a new Ceph OSD storage type was
introduced, the so called Bluestore introduced, the so called Bluestore
@ -356,8 +365,8 @@ pveceph osd create /dev/sd[X]
.Block.db and block.wal .Block.db and block.wal
If you want to use a separate DB/WAL device for your OSDs, you can specify it If you want to use a separate DB/WAL device for your OSDs, you can specify it
through the '-db_dev' and '-wal_dev' options. The WAL is placed with the DB, if not through the '-db_dev' and '-wal_dev' options. The WAL is placed with the DB, if
specified separately. not specified separately.
[source,bash] [source,bash]
---- ----
@ -380,8 +389,7 @@ internal journal or write-ahead log. It is recommended to use a fast SSD or
NVRAM for better performance. NVRAM for better performance.
Ceph Filestore .Ceph Filestore
~~~~~~~~~~~~~~
Before Ceph Luminous, Filestore was used as default storage type for Ceph OSDs. Before Ceph Luminous, Filestore was used as default storage type for Ceph OSDs.
Starting with Ceph Nautilus, {pve} does not support creating such OSDs with Starting with Ceph Nautilus, {pve} does not support creating such OSDs with
@ -393,8 +401,8 @@ Starting with Ceph Nautilus, {pve} does not support creating such OSDs with
ceph-volume lvm create --filestore --data /dev/sd[X] --journal /dev/sd[Y] ceph-volume lvm create --filestore --data /dev/sd[X] --journal /dev/sd[Y]
---- ----
Destroying Ceph OSDs Destroy OSDs
-------------------- ~~~~~~~~~~~~
To remove an OSD via the GUI first select a {PVE} node in the tree view and go To remove an OSD via the GUI first select a {PVE} node in the tree view and go
to the **Ceph -> OSD** panel. Select the OSD to destroy. Next click the **OUT** to the **Ceph -> OSD** panel. Select the OSD to destroy. Next click the **OUT**
@ -422,14 +430,17 @@ WARNING: The above command will destroy data on the disk!
[[pve_ceph_pools]] [[pve_ceph_pools]]
Creating Ceph Pools Ceph Pools
------------------- ----------
[thumbnail="screenshot/gui-ceph-pools.png"]
A pool is a logical group for storing objects. It holds **P**lacement A pool is a logical group for storing objects. It holds **P**lacement
**G**roups (`PG`, `pg_num`), a collection of objects. **G**roups (`PG`, `pg_num`), a collection of objects.
Create Pools
~~~~~~~~~~~~
[thumbnail="screenshot/gui-ceph-pools.png"]
When no options are given, we set a default of **128 PGs**, a **size of 3 When no options are given, we set a default of **128 PGs**, a **size of 3
replicas** and a **min_size of 2 replicas** for serving objects in a degraded replicas** and a **min_size of 2 replicas** for serving objects in a degraded
state. state.
@ -461,8 +472,8 @@ http://docs.ceph.com/docs/luminous/rados/operations/pools/]
manual. manual.
Destroying Ceph Pools Destroy Pools
--------------------- ~~~~~~~~~~~~~
To destroy a pool via the GUI select a node in the tree view and go to the To destroy a pool via the GUI select a node in the tree view and go to the
**Ceph -> Pools** panel. Select the pool to destroy and click the **Destroy** **Ceph -> Pools** panel. Select the pool to destroy and click the **Destroy**
@ -553,8 +564,8 @@ ceph osd pool set <pool-name> crush_rule <rule-name>
---- ----
TIP: If the pool already contains objects, all of these have to be moved TIP: If the pool already contains objects, all of these have to be moved
accordingly. Depending on your setup this may introduce a big performance hit on accordingly. Depending on your setup this may introduce a big performance hit
your cluster. As an alternative, you can create a new pool and move disks on your cluster. As an alternative, you can create a new pool and move disks
separately. separately.