mirror of
https://git.proxmox.com/git/pve-docs
synced 2025-06-14 11:07:37 +00:00
pveceph: annotate code blocks as bash
to have them set to the same highlighting scheme throughout the chapter Signed-off-by: Aaron Lauterer <a.lauterer@proxmox.com>
This commit is contained in:
parent
41292dab6e
commit
c49d3f328c
24
pveceph.adoc
24
pveceph.adoc
@ -802,14 +802,14 @@ accommodate the need for easy ruleset generation.
|
|||||||
The device classes can be seen in the 'ceph osd tree' output. These classes
|
The device classes can be seen in the 'ceph osd tree' output. These classes
|
||||||
represent their own root bucket, which can be seen with the below command.
|
represent their own root bucket, which can be seen with the below command.
|
||||||
|
|
||||||
[source, bash]
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph osd crush tree --show-shadow
|
ceph osd crush tree --show-shadow
|
||||||
----
|
----
|
||||||
|
|
||||||
Example output form the above command:
|
Example output form the above command:
|
||||||
|
|
||||||
[source, bash]
|
[source,bash]
|
||||||
----
|
----
|
||||||
ID CLASS WEIGHT TYPE NAME
|
ID CLASS WEIGHT TYPE NAME
|
||||||
-16 nvme 2.18307 root default~nvme
|
-16 nvme 2.18307 root default~nvme
|
||||||
@ -831,7 +831,7 @@ ID CLASS WEIGHT TYPE NAME
|
|||||||
To instruct a pool to only distribute objects on a specific device class, you
|
To instruct a pool to only distribute objects on a specific device class, you
|
||||||
first need to create a ruleset for the device class:
|
first need to create a ruleset for the device class:
|
||||||
|
|
||||||
[source, bash]
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph osd crush rule create-replicated <rule-name> <root> <failure-domain> <class>
|
ceph osd crush rule create-replicated <rule-name> <root> <failure-domain> <class>
|
||||||
----
|
----
|
||||||
@ -846,7 +846,7 @@ ceph osd crush rule create-replicated <rule-name> <root> <failure-domain> <class
|
|||||||
|
|
||||||
Once the rule is in the CRUSH map, you can tell a pool to use the ruleset.
|
Once the rule is in the CRUSH map, you can tell a pool to use the ruleset.
|
||||||
|
|
||||||
[source, bash]
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph osd pool set <pool-name> crush_rule <rule-name>
|
ceph osd pool set <pool-name> crush_rule <rule-name>
|
||||||
----
|
----
|
||||||
@ -908,6 +908,7 @@ CephFS needs at least one Metadata Server to be configured and running, in order
|
|||||||
to function. You can create an MDS through the {pve} web GUI's `Node
|
to function. You can create an MDS through the {pve} web GUI's `Node
|
||||||
-> CephFS` panel or from the command line with:
|
-> CephFS` panel or from the command line with:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
pveceph mds create
|
pveceph mds create
|
||||||
----
|
----
|
||||||
@ -919,6 +920,7 @@ You can speed up the handover between the active and standby MDS by using
|
|||||||
the 'hotstandby' parameter option on creation, or if you have already created it
|
the 'hotstandby' parameter option on creation, or if you have already created it
|
||||||
you may set/add:
|
you may set/add:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
mds standby replay = true
|
mds standby replay = true
|
||||||
----
|
----
|
||||||
@ -959,6 +961,7 @@ After this is complete, you can simply create a CephFS through
|
|||||||
either the Web GUI's `Node -> CephFS` panel or the command-line tool `pveceph`,
|
either the Web GUI's `Node -> CephFS` panel or the command-line tool `pveceph`,
|
||||||
for example:
|
for example:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
pveceph fs create --pg_num 128 --add-storage
|
pveceph fs create --pg_num 128 --add-storage
|
||||||
----
|
----
|
||||||
@ -988,6 +991,7 @@ necessary:
|
|||||||
want to destroy.
|
want to destroy.
|
||||||
* Unmount the CephFS storages on all cluster nodes manually with
|
* Unmount the CephFS storages on all cluster nodes manually with
|
||||||
+
|
+
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
umount /mnt/pve/<STORAGE-NAME>
|
umount /mnt/pve/<STORAGE-NAME>
|
||||||
----
|
----
|
||||||
@ -999,12 +1003,14 @@ Where `<STORAGE-NAME>` is the name of the CephFS storage in your {PVE}.
|
|||||||
interface or via the command-line interface, for the latter you would issue
|
interface or via the command-line interface, for the latter you would issue
|
||||||
the following command:
|
the following command:
|
||||||
+
|
+
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
pveceph stop --service mds.NAME
|
pveceph stop --service mds.NAME
|
||||||
----
|
----
|
||||||
+
|
+
|
||||||
to stop them, or
|
to stop them, or
|
||||||
+
|
+
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
pveceph mds destroy NAME
|
pveceph mds destroy NAME
|
||||||
----
|
----
|
||||||
@ -1017,6 +1023,7 @@ servers.
|
|||||||
|
|
||||||
* Now you can destroy the CephFS with
|
* Now you can destroy the CephFS with
|
||||||
+
|
+
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
pveceph fs destroy NAME --remove-storages --remove-pools
|
pveceph fs destroy NAME --remove-storages --remove-pools
|
||||||
----
|
----
|
||||||
@ -1094,6 +1101,7 @@ Once all clients, VMs and containers are off or not accessing the Ceph cluster
|
|||||||
anymore, verify that the Ceph cluster is in a healthy state. Either via the Web UI
|
anymore, verify that the Ceph cluster is in a healthy state. Either via the Web UI
|
||||||
or the CLI:
|
or the CLI:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph -s
|
ceph -s
|
||||||
----
|
----
|
||||||
@ -1102,6 +1110,7 @@ To disable all self-healing actions, and to pause any client IO in the Ceph
|
|||||||
cluster, enable the following OSD flags in the **Ceph -> OSD** panel or via the
|
cluster, enable the following OSD flags in the **Ceph -> OSD** panel or via the
|
||||||
CLI:
|
CLI:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph osd set noout
|
ceph osd set noout
|
||||||
ceph osd set norecover
|
ceph osd set norecover
|
||||||
@ -1118,6 +1127,7 @@ When powering on the cluster, start the nodes with monitors (MONs) first. Once
|
|||||||
all nodes are up and running, confirm that all Ceph services are up and running
|
all nodes are up and running, confirm that all Ceph services are up and running
|
||||||
before you unset the OSD flags again:
|
before you unset the OSD flags again:
|
||||||
|
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph osd unset pause
|
ceph osd unset pause
|
||||||
ceph osd unset nodown
|
ceph osd unset nodown
|
||||||
@ -1145,11 +1155,13 @@ below will also give you an overview of the current events and actions to take.
|
|||||||
To stop their execution, press CTRL-C.
|
To stop their execution, press CTRL-C.
|
||||||
|
|
||||||
Continuously watch the cluster status:
|
Continuously watch the cluster status:
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
watch ceph --status
|
watch ceph --status
|
||||||
----
|
----
|
||||||
|
|
||||||
Print the cluster status once (not being updated) and continuously append lines of status events:
|
Print the cluster status once (not being updated) and continuously append lines of status events:
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph --watch
|
ceph --watch
|
||||||
----
|
----
|
||||||
@ -1169,6 +1181,7 @@ footnote:[Ceph troubleshooting {cephdocs-url}/rados/troubleshooting/].
|
|||||||
* xref:disk_health_monitoring[Disk Health Monitoring]
|
* xref:disk_health_monitoring[Disk Health Monitoring]
|
||||||
* __System -> System Log__ or via the CLI, for example of the last 2 days:
|
* __System -> System Log__ or via the CLI, for example of the last 2 days:
|
||||||
+
|
+
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
journalctl --since "2 days ago"
|
journalctl --since "2 days ago"
|
||||||
----
|
----
|
||||||
@ -1176,11 +1189,13 @@ journalctl --since "2 days ago"
|
|||||||
|
|
||||||
Ceph service crashes can be listed and viewed in detail by running the following
|
Ceph service crashes can be listed and viewed in detail by running the following
|
||||||
commands:
|
commands:
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph crash ls
|
ceph crash ls
|
||||||
ceph crash info <crash_id>
|
ceph crash info <crash_id>
|
||||||
----
|
----
|
||||||
Crashes marked as new can be acknowledged by running:
|
Crashes marked as new can be acknowledged by running:
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph crash archive-all
|
ceph crash archive-all
|
||||||
----
|
----
|
||||||
@ -1220,6 +1235,7 @@ interface, go to __Any node -> Ceph -> OSD__, select the OSD and click
|
|||||||
on **Start**, **In** and **Reload**. When using the shell, run following
|
on **Start**, **In** and **Reload**. When using the shell, run following
|
||||||
command on the affected node:
|
command on the affected node:
|
||||||
+
|
+
|
||||||
|
[source,bash]
|
||||||
----
|
----
|
||||||
ceph-volume lvm activate --all
|
ceph-volume lvm activate --all
|
||||||
----
|
----
|
||||||
|
Loading…
Reference in New Issue
Block a user