mirror of
https://git.proxmox.com/git/pve-docs
synced 2025-04-30 01:15:02 +00:00
pveceph: troubleshooting maintenance: rework to have CLI commands in blocks
having CLI commands in their own blocks instead of inline makes them stand out quickly and a lot easier to copy & paste. Signed-off-by: Aaron Lauterer <a.lauterer@proxmox.com>
This commit is contained in:
parent
9676a0d867
commit
41292dab6e
46
pveceph.adoc
46
pveceph.adoc
@ -1052,7 +1052,11 @@ same type and size.
|
|||||||
|
|
||||||
. After automatic rebalancing, the cluster status should switch back
|
. After automatic rebalancing, the cluster status should switch back
|
||||||
to `HEALTH_OK`. Any still listed crashes can be acknowledged by
|
to `HEALTH_OK`. Any still listed crashes can be acknowledged by
|
||||||
running, for example, `ceph crash archive-all`.
|
running the following command:
|
||||||
|
[source,bash]
|
||||||
|
----
|
||||||
|
ceph crash archive-all
|
||||||
|
----
|
||||||
|
|
||||||
Trim/Discard
|
Trim/Discard
|
||||||
~~~~~~~~~~~~
|
~~~~~~~~~~~~
|
||||||
@ -1140,13 +1144,14 @@ The following Ceph commands can be used to see if the cluster is healthy
|
|||||||
below will also give you an overview of the current events and actions to take.
|
below will also give you an overview of the current events and actions to take.
|
||||||
To stop their execution, press CTRL-C.
|
To stop their execution, press CTRL-C.
|
||||||
|
|
||||||
|
Continuously watch the cluster status:
|
||||||
|
----
|
||||||
|
watch ceph --status
|
||||||
----
|
----
|
||||||
# Continuously watch the cluster status
|
|
||||||
pve# watch ceph --status
|
|
||||||
|
|
||||||
# Print the cluster status once (not being updated)
|
Print the cluster status once (not being updated) and continuously append lines of status events:
|
||||||
# and continuously append lines of status events
|
----
|
||||||
pve# ceph --watch
|
ceph --watch
|
||||||
----
|
----
|
||||||
|
|
||||||
[[pve_ceph_ts]]
|
[[pve_ceph_ts]]
|
||||||
@ -1162,14 +1167,23 @@ footnote:[Ceph troubleshooting {cephdocs-url}/rados/troubleshooting/].
|
|||||||
.Relevant Logs on Affected Node
|
.Relevant Logs on Affected Node
|
||||||
|
|
||||||
* xref:disk_health_monitoring[Disk Health Monitoring]
|
* xref:disk_health_monitoring[Disk Health Monitoring]
|
||||||
* __System -> System Log__ (or, for example,
|
* __System -> System Log__ or via the CLI, for example of the last 2 days:
|
||||||
`journalctl --since "2 days ago"`)
|
+
|
||||||
|
----
|
||||||
|
journalctl --since "2 days ago"
|
||||||
|
----
|
||||||
* IPMI and RAID controller logs
|
* IPMI and RAID controller logs
|
||||||
|
|
||||||
Ceph service crashes can be listed and viewed in detail by running
|
Ceph service crashes can be listed and viewed in detail by running the following
|
||||||
`ceph crash ls` and `ceph crash info <crash_id>`. Crashes marked as
|
commands:
|
||||||
new can be acknowledged by running, for example,
|
----
|
||||||
`ceph crash archive-all`.
|
ceph crash ls
|
||||||
|
ceph crash info <crash_id>
|
||||||
|
----
|
||||||
|
Crashes marked as new can be acknowledged by running:
|
||||||
|
----
|
||||||
|
ceph crash archive-all
|
||||||
|
----
|
||||||
|
|
||||||
To get a more detailed view, every Ceph service has a log file under
|
To get a more detailed view, every Ceph service has a log file under
|
||||||
`/var/log/ceph/`. If more detail is required, the log level can be
|
`/var/log/ceph/`. If more detail is required, the log level can be
|
||||||
@ -1203,8 +1217,12 @@ A faulty OSD will be reported as `down` and mostly (auto) `out` 10
|
|||||||
minutes later. Depending on the cause, it can also automatically
|
minutes later. Depending on the cause, it can also automatically
|
||||||
become `up` and `in` again. To try a manual activation via web
|
become `up` and `in` again. To try a manual activation via web
|
||||||
interface, go to __Any node -> Ceph -> OSD__, select the OSD and click
|
interface, go to __Any node -> Ceph -> OSD__, select the OSD and click
|
||||||
on **Start**, **In** and **Reload**. When using the shell, run on the
|
on **Start**, **In** and **Reload**. When using the shell, run following
|
||||||
affected node `ceph-volume lvm activate --all`.
|
command on the affected node:
|
||||||
|
+
|
||||||
|
----
|
||||||
|
ceph-volume lvm activate --all
|
||||||
|
----
|
||||||
+
|
+
|
||||||
To activate a failed OSD, it may be necessary to
|
To activate a failed OSD, it may be necessary to
|
||||||
xref:ha_manager_node_maintenance[safely reboot] the respective node
|
xref:ha_manager_node_maintenance[safely reboot] the respective node
|
||||||
|
Loading…
Reference in New Issue
Block a user