Commit Graph

5742 Commits

Author SHA1 Message Date
Oguz Bektas
4befa0abe0 pvereport: get pressure stall information
PSI can be queried in /proc/pressure/{cpu,io,memory} for the
corresponding resources. this helps us track down disruptions caused
by resource overcommitment.

Signed-off-by: Oguz Bektas <o.bektas@proxmox.com>
2021-05-11 14:07:25 +02:00
Lorenz Stechauner
230c4416f2 api: pools: fix typo/syntax error
Signed-off-by: Lorenz Stechauner <l.stechauner@proxmox.com>
2021-05-10 17:12:17 +02:00
Thomas Lamprecht
274a0e2394 api: pools: whitespace, indentation and code cleanups
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-09 18:58:42 +02:00
Thomas Lamprecht
926ce2b8d9 api: pools: whitespace, indentation and code cleanups
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-09 18:52:28 +02:00
Thomas Lamprecht
400d6ce399 bump version to 6.4-6
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-07 18:19:43 +02:00
Stoiko Ivanov
36ad2b3c0e proxy: fix wildcard address use
This patch fixes a regression for hosts disabling ipv6 via kernel
commandline ('ipv6.disable=1')introduced in commit
e224b7d2e6
(disabling IPv6 via sysctl did not exhibit these problems)

by hardcoding the address to '::', pveproxy and spiceproxy failed to
start with:
'unable to create socket - Address family not supported by protocol'

This patch depends on the commit in pve-common, which tries first
binding to '::' and then falling back to '0.0.0.0', and needs a
versioned dependency bump on libpve-common-perl.

With this patch the listening addresses are (`ss -tlnp |grep 8006` output)
* ipv6 disabled via kernel cmdline: '0.0.0.0:8006'
* sysctl net.ipv6.conf.all.disable_ipv6=1: '*:8006'
* sysctl net.ipv6.bindv6only=1: '[::]:8006'
* else: '*:8006'

Signed-off-by: Stoiko Ivanov <s.ivanov@proxmox.com>
2021-05-07 18:15:43 +02:00
Thomas Lamprecht
55871e9e34 ui: storage status: use SI units for usage for consitency with RRD chart
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-07 18:15:05 +02:00
Thomas Lamprecht
87fd494fe5 update aplinfo.dat
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-07 10:05:47 +02:00
Thomas Lamprecht
96d4c3b4a6 api: acme: fix API index entry for challenge-schema
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-05 15:52:49 +02:00
Thomas Lamprecht
6110ed0348 api: nodes: whitespace cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-04 11:58:55 +02:00
Thomas Lamprecht
73bf05bb0a bump version to 6.4-5
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-03 15:06:52 +02:00
Thomas Lamprecht
aa396ea6b3 Revert "fix #2745: ui: backup: allow specifying remove parameter for manual backup"
Pushed this one out by mistake when commit a fixup, so do revert now
as I did not fully reviewed it and saw some UI changes I'd like to
do.

This reverts commit cfc6d15ed0.
2021-05-03 14:41:15 +02:00
Thomas Lamprecht
37bf860e3e pveceph: fixup syntax issue
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-03 14:38:16 +02:00
Fabian Ebner
cfc6d15ed0 fix #2745: ui: backup: allow specifying remove parameter for manual backup
and also show the retention options that will be used for a given storage. A
user with Datastore.AllocateSpace and VM.Backup can already remove backups from
the GUI manually, so it shouldn't be a problem if they can set the remove flag
when starting a manual backup in the GUI.

Signed-off-by: Fabian Ebner <f.ebner@proxmox.com>
2021-05-03 14:36:59 +02:00
Fabian Ebner
fddb041217 ui: backup: fill in some of the configured vzdump defaults
Do not fill in the default for compression, because the initial default for the
backend is to not compress, while the current default for the UI is zstd, which
is preferable.

The 'defaults' API call expects the user to have permissions on the storage,
because retention options are storage-dependent. Use a flag initialDefaults to
make sure storage-independent properties are only set once, so they are not
reset when a user changes the storage after editing them.

Signed-off-by: Fabian Ebner <f.ebner@proxmox.com>
2021-05-03 14:36:59 +02:00
Fabian Ebner
5b9a4030dd api: vzdump: add call to get currently configured vzdump defaults
on a given node (and storage).

There is no datacenter/storage fallback for the bandwidth limit, so the default
can just be returned as is. While the bandwidth limit is a root-only option when
executing the backup, it still makes sense to return it for all users, so they
can see what's going to be used.

Signed-off-by: Fabian Ebner <f.ebner@proxmox.com>
2021-05-03 14:36:59 +02:00
Thomas Lamprecht
f8b7b9e143 ui: ceph code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-03 14:35:43 +02:00
Thomas Lamprecht
ece49b3c90 ceph: reload API daemons after installation
To make them load the updated librados2, as else they may potentially
not be able to communicate with the potentially newer ceph monitors,
as Debian 10 ships Jewel (12.2) by default...

While we could do some more fancy signaling to the workers to reload
the lib, that is rather a PITA and complex solution for something
that happens once in a blue moon.

We may want to add a trigger in ceph for this on updates though, that
would effectively fix this too - but needs to be thought out better.
So for now lets go with the simplest solution.

Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-03 14:32:50 +02:00
Thomas Lamprecht
51498a2664 ceph: code/indentation cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-05-03 14:03:32 +02:00
Thomas Lamprecht
c0d51f284a bump version to 6.4-4
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-27 13:03:58 +02:00
Thomas Lamprecht
0dd48804e1 api: ceph/monitor: automatically disable insecure global ID reclaim after creating first monitor
nautilus 14.2.20 and octopus 15.2.11 fixed a security issue with
reclaiming the global ID auth (CVE-2021-20288). As fixing this issue
means that older client won't be able to connect anymore, the fix was
done behind a switch, with a HEALTH warning if it was not active
(i.e., disallowed connection from older clients).

New installations have this switch also at the insecure level, for
compat reasons, so lets deactivate it ourself after monitor creation
to avoid the health warning and slightly insecure setup (in default
PVE ceph the whole issue was of rather low impact/risk). But, only do
so when creating the first monitor of a ceph cluster, to avoid
breaking existing setups by accident.

An admin can always switch it back again, e.g., if they're recovering
from some failure and need to setup fresh monitors but have still old
clients.

Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-27 12:35:34 +02:00
Thomas Lamprecht
d26556c07b pveceph: octopus is OK and wanted to be installed/used
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-27 12:27:32 +02:00
Thomas Lamprecht
7e4adde160 bump version to 6.4-3
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 20:20:00 +02:00
Thomas Lamprecht
97edb617c4 ui: rbd: move namespace to advanced section
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 20:16:45 +02:00
Aaron Lauterer
2d6f898b2a ui: RBDStorage: add field for RBD namespace
Makes it possible to configure the RBD namespace via the GUI.

RBD namespaces must be configured manually. The most likely use case is
when connecting to an external Ceph cluster as this makes it possible to
separate client PVE clusters by namespace, not by pool.

Signed-off-by: Aaron Lauterer <a.lauterer@proxmox.com>
2021-04-26 20:10:47 +02:00
Thomas Lamprecht
36094b108b ui: sdn vnets: disable subnet create button if no vnet is selected
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 20:08:03 +02:00
Thomas Lamprecht
69908c5106 ui: sdn: eslint fixes and code cleanups
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 20:07:31 +02:00
Thomas Lamprecht
4966ba74f3 ui: sdn: eslint auto fixes
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 15:44:57 +02:00
Alexandre Derumier
81a6afa87d sdn: controllers: bgp: use node as controllerid
Signed-off-by: Alexandre Derumier <aderumier@odiso.com>
2021-04-26 15:43:47 +02:00
Alexandre Derumier
bdcd66ed3e sdn: controller: bgp: add ebgp-multihop option
Signed-off-by: Alexandre Derumier <aderumier@odiso.com>
2021-04-26 15:43:47 +02:00
Alexandre Derumier
ddf0e952ff sdn: move mac option from vnet to evpn zone plugin
Signed-off-by: Alexandre Derumier <aderumier@odiso.com>
2021-04-26 15:43:47 +02:00
Alexandre Derumier
1d9643f67c sdn: add subnet/ipam/sdn management
Signed-off-by: Alexandre Derumier <aderumier@odiso.com>
2021-04-26 15:43:47 +02:00
Thomas Lamprecht
d7eb3e2282 ui: ceph: pool edit: set emptyValue to 0 for target-size field
ensure it gets "deleted" (ceph resets it on setting this to zero)
when modified to empty.

Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 14:25:02 +02:00
Thomas Lamprecht
b34b1187ee ui: size field: allow to specifiy emptyValue
I.e., the value the empty field should resolve for the submit value

Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 14:25:02 +02:00
Thomas Lamprecht
6bcb88cf53 ui: ceph: pool edit: drop allowZero from numberfield
not a thing there, so do not be misleading and suggest this is
required

Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 14:25:02 +02:00
Thomas Lamprecht
3527924585 ui: ceph status: use two column layout for 1600+ window width
Looks already OK at that size, and one gets a better overview.

We have a slightly complex layout here (to columns which should be
above each other) so we cannot just use the generic helper, but
that's OK here - it *is* a special view.

Note, not all people use full-sized windows all the time, so the
widths here must not only be considered in terms of display
resolutions...

Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 12:31:11 +02:00
Thomas Lamprecht
f9f81a01dc ui: tasks: add ceph edit-pool and set-flags descriptions
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 12:27:48 +02:00
Thomas Lamprecht
301f52b172 ui: break overly long line
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-26 12:27:14 +02:00
Thomas Lamprecht
4c42fc7422 update shipped APL info
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-25 18:22:58 +02:00
Thomas Lamprecht
e248e28455 bump version to 6.4-2
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 19:29:03 +02:00
Thomas Lamprecht
519d9f3fee ui: ceph pools: fixup class names, drop extry xtype definition
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 19:23:04 +02:00
Thomas Lamprecht
5da1b6d0db ui: ceph: pool creation: default auto-scale value to "on"
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 19:18:51 +02:00
Thomas Lamprecht
3eaf388170 ui: node config: fix internal code order for nav bar
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 19:11:46 +02:00
Thomas Lamprecht
17f4e62e32 ui: ceph: status details: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 19:00:42 +02:00
Thomas Lamprecht
10b00f3a06 ui: ceph: status: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 19:00:42 +02:00
Thomas Lamprecht
f87b8d2f3f ui: ceph: services: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 18:44:57 +02:00
Thomas Lamprecht
fb792310d1 ui: ceph: service list: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 18:44:39 +02:00
Thomas Lamprecht
261bb63356 ui: ceph: pool: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 18:44:23 +02:00
Thomas Lamprecht
d7acb90022 ui: ceph: fs: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 18:44:13 +02:00
Thomas Lamprecht
436f638d44 ui: ceph: install wizard: eslint fixes and code cleanup
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2021-04-24 18:44:01 +02:00