mirror of
https://git.proxmox.com/git/proxmox-backup
synced 2025-06-13 19:31:29 +00:00
fix #3935: datastore/api/backup: move datastore locking to '/run'
to avoid issues when removing a group or snapshot directory where two threads hold a lock to the same directory, move locking to the tmpfs backed '/run' directory. also adds double stat'ing to make it possible to remove locks without certain race condition issues. this new mechanism is only employed when we can be sure, that a reboot has occured so that all processes are using the new locking mechanism. otherwise, two separate process could assume they have exclusive rights to a group or snapshot. bumps the rust version to 1.81 so we can use `std::fs::exists` without issue. Signed-off-by: Shannon Sterz <s.sterz@proxmox.com> Acked-by: Wolfgang Bumiller <w.bumiller@proxmox.com> [TL: drop unused format_err import] Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
This commit is contained in:
parent
e2c1866b13
commit
27dd73777f
@ -13,7 +13,7 @@ authors = [
|
||||
edition = "2021"
|
||||
license = "AGPL-3"
|
||||
repository = "https://git.proxmox.com/?p=proxmox-backup.git"
|
||||
rust-version = "1.80"
|
||||
rust-version = "1.81"
|
||||
|
||||
[package]
|
||||
name = "proxmox-backup"
|
||||
|
5
debian/postinst
vendored
5
debian/postinst
vendored
@ -80,6 +80,11 @@ EOF
|
||||
update_sync_job "$prev_job"
|
||||
fi
|
||||
fi
|
||||
|
||||
if dpkg --compare-versions "$2" 'lt' '3.3.4~'; then
|
||||
# ensure old locking is used by the daemon until a reboot happened
|
||||
touch "/run/proxmox-backup/old-locking"
|
||||
fi
|
||||
fi
|
||||
;;
|
||||
|
||||
|
2
debian/rules
vendored
2
debian/rules
vendored
@ -22,6 +22,8 @@ ifneq ("$(wildcard .repoid)","")
|
||||
endif
|
||||
|
||||
%:
|
||||
@echo "fix locking version in postinst"
|
||||
false
|
||||
dh $@ --with=bash-completion
|
||||
|
||||
override_dh_auto_configure:
|
||||
|
@ -22,6 +22,8 @@ pub use config_version_cache::ConfigVersionCache;
|
||||
|
||||
use anyhow::{format_err, Error};
|
||||
use nix::unistd::{Gid, Group, Uid, User};
|
||||
use proxmox_sys::fs::DirLockGuard;
|
||||
use std::os::unix::prelude::AsRawFd;
|
||||
|
||||
pub use pbs_buildcfg::{BACKUP_GROUP_NAME, BACKUP_USER_NAME};
|
||||
|
||||
@ -46,13 +48,34 @@ pub fn backup_group() -> Result<nix::unistd::Group, Error> {
|
||||
}
|
||||
|
||||
pub struct BackupLockGuard {
|
||||
_file: Option<std::fs::File>,
|
||||
file: Option<std::fs::File>,
|
||||
// TODO: Remove `_legacy_dir` with PBS 5
|
||||
_legacy_dir: Option<DirLockGuard>,
|
||||
}
|
||||
|
||||
impl AsRawFd for BackupLockGuard {
|
||||
fn as_raw_fd(&self) -> i32 {
|
||||
self.file.as_ref().map_or(-1, |f| f.as_raw_fd())
|
||||
}
|
||||
}
|
||||
|
||||
// TODO: Remove with PBS 5
|
||||
impl From<DirLockGuard> for BackupLockGuard {
|
||||
fn from(value: DirLockGuard) -> Self {
|
||||
Self {
|
||||
file: None,
|
||||
_legacy_dir: Some(value),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[doc(hidden)]
|
||||
/// Note: do not use for production code, this is only intended for tests
|
||||
pub unsafe fn create_mocked_lock() -> BackupLockGuard {
|
||||
BackupLockGuard { _file: None }
|
||||
BackupLockGuard {
|
||||
file: None,
|
||||
_legacy_dir: None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Open or create a lock file owned by user "backup" and lock it.
|
||||
@ -76,7 +99,10 @@ pub fn open_backup_lockfile<P: AsRef<std::path::Path>>(
|
||||
let timeout = timeout.unwrap_or(std::time::Duration::new(10, 0));
|
||||
|
||||
let file = proxmox_sys::fs::open_file_locked(&path, timeout, exclusive, options)?;
|
||||
Ok(BackupLockGuard { _file: Some(file) })
|
||||
Ok(BackupLockGuard {
|
||||
file: Some(file),
|
||||
_legacy_dir: None,
|
||||
})
|
||||
}
|
||||
|
||||
/// Atomically write data to file owned by "root:backup" with permission "0640"
|
||||
|
@ -35,6 +35,7 @@ proxmox-lang.workspace=true
|
||||
proxmox-schema = { workspace = true, features = [ "api-macro" ] }
|
||||
proxmox-serde = { workspace = true, features = [ "serde_json" ] }
|
||||
proxmox-sys.workspace = true
|
||||
proxmox-systemd.workspace = true
|
||||
proxmox-time.workspace = true
|
||||
proxmox-uuid.workspace = true
|
||||
proxmox-worker-task.workspace = true
|
||||
|
@ -1,13 +1,15 @@
|
||||
use std::fmt;
|
||||
use std::os::unix::io::RawFd;
|
||||
use std::os::unix::io::{AsRawFd, RawFd};
|
||||
use std::os::unix::prelude::OsStrExt;
|
||||
use std::path::Path;
|
||||
use std::path::PathBuf;
|
||||
use std::sync::Arc;
|
||||
use std::sync::{Arc, LazyLock};
|
||||
use std::time::Duration;
|
||||
|
||||
use anyhow::{bail, format_err, Context, Error};
|
||||
|
||||
use proxmox_sys::fs::{
|
||||
lock_dir_noblock, lock_dir_noblock_shared, replace_file, CreateOptions, DirLockGuard,
|
||||
};
|
||||
use proxmox_sys::fs::{lock_dir_noblock, lock_dir_noblock_shared, replace_file, CreateOptions};
|
||||
use proxmox_systemd::escape_unit;
|
||||
|
||||
use pbs_api_types::{
|
||||
Authid, BackupGroupDeleteStats, BackupNamespace, BackupType, GroupFilter, VerifyState,
|
||||
@ -18,6 +20,18 @@ use pbs_config::{open_backup_lockfile, BackupLockGuard};
|
||||
use crate::manifest::{BackupManifest, MANIFEST_LOCK_NAME};
|
||||
use crate::{DataBlob, DataStore};
|
||||
|
||||
pub const DATASTORE_LOCKS_DIR: &str = "/run/proxmox-backup/locks";
|
||||
|
||||
// TODO: Remove with PBS 5
|
||||
// Note: The `expect()` call here will only happen if we can neither confirm nor deny the existence
|
||||
// of the file. this should only happen if a user messes with the `/run/proxmox-backup` directory.
|
||||
// if that happens, a lot more should fail as we rely on the existence of the directory throughout
|
||||
// the code. so just panic with a reasonable message.
|
||||
static OLD_LOCKING: LazyLock<bool> = LazyLock::new(|| {
|
||||
std::fs::exists("/run/proxmox-backup/old-locking")
|
||||
.expect("cannot read `/run/proxmox-backup`, please check permissions")
|
||||
});
|
||||
|
||||
/// BackupGroup is a directory containing a list of BackupDir
|
||||
#[derive(Clone)]
|
||||
pub struct BackupGroup {
|
||||
@ -225,6 +239,7 @@ impl BackupGroup {
|
||||
delete_stats.increment_removed_groups();
|
||||
}
|
||||
|
||||
let _ = std::fs::remove_file(self.lock_path());
|
||||
Ok(delete_stats)
|
||||
}
|
||||
|
||||
@ -241,13 +256,34 @@ impl BackupGroup {
|
||||
.set_owner(&self.ns, self.as_ref(), auth_id, force)
|
||||
}
|
||||
|
||||
/// Lock a group exclusively
|
||||
pub fn lock(&self) -> Result<DirLockGuard, Error> {
|
||||
/// Returns a file name for locking a group.
|
||||
///
|
||||
/// The lock file will be located in:
|
||||
/// `${DATASTORE_LOCKS_DIR}/${datastore name}/${lock_file_path_helper(rpath)}`
|
||||
/// where `rpath` is the relative path of the group.
|
||||
fn lock_path(&self) -> PathBuf {
|
||||
let path = Path::new(DATASTORE_LOCKS_DIR).join(self.store.name());
|
||||
|
||||
let rpath = Path::new(self.group.ty.as_str()).join(&self.group.id);
|
||||
|
||||
path.join(lock_file_path_helper(&self.ns, rpath))
|
||||
}
|
||||
|
||||
/// Locks a group exclusively.
|
||||
pub fn lock(&self) -> Result<BackupLockGuard, Error> {
|
||||
if *OLD_LOCKING {
|
||||
lock_dir_noblock(
|
||||
&self.full_group_path(),
|
||||
"backup group",
|
||||
"possible running backup",
|
||||
"possible runing backup, group is in use",
|
||||
)
|
||||
.map(BackupLockGuard::from)
|
||||
} else {
|
||||
lock_helper(self.store.name(), &self.lock_path(), |p| {
|
||||
open_backup_lockfile(p, Some(Duration::from_secs(0)), true)
|
||||
.with_context(|| format!("unable to acquire backup group lock {p:?}"))
|
||||
})
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@ -454,22 +490,53 @@ impl BackupDir {
|
||||
.map_err(|err| format_err!("unable to acquire manifest lock {:?} - {}", &path, err))
|
||||
}
|
||||
|
||||
/// Lock this snapshot exclusively
|
||||
pub fn lock(&self) -> Result<DirLockGuard, Error> {
|
||||
/// Returns a file name for locking a snapshot.
|
||||
///
|
||||
/// The lock file will be located in:
|
||||
/// `${DATASTORE_LOCKS_DIR}/${datastore name}/${lock_file_path_helper(rpath)}`
|
||||
/// where `rpath` is the relative path of the snapshot.
|
||||
fn lock_path(&self) -> PathBuf {
|
||||
let path = Path::new(DATASTORE_LOCKS_DIR).join(self.store.name());
|
||||
|
||||
let rpath = Path::new(self.dir.group.ty.as_str())
|
||||
.join(&self.dir.group.id)
|
||||
.join(&self.backup_time_string);
|
||||
|
||||
path.join(lock_file_path_helper(&self.ns, rpath))
|
||||
}
|
||||
|
||||
/// Locks a snapshot exclusively.
|
||||
pub fn lock(&self) -> Result<BackupLockGuard, Error> {
|
||||
if *OLD_LOCKING {
|
||||
lock_dir_noblock(
|
||||
&self.full_path(),
|
||||
"snapshot",
|
||||
"backup is running or snapshot is in use",
|
||||
)
|
||||
.map(BackupLockGuard::from)
|
||||
} else {
|
||||
lock_helper(self.store.name(), &self.lock_path(), |p| {
|
||||
open_backup_lockfile(p, Some(Duration::from_secs(0)), true)
|
||||
.with_context(|| format!("unable to acquire snapshot lock {p:?}"))
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
/// Acquire a shared lock on this snapshot
|
||||
pub fn lock_shared(&self) -> Result<DirLockGuard, Error> {
|
||||
/// Acquires a shared lock on a snapshot.
|
||||
pub fn lock_shared(&self) -> Result<BackupLockGuard, Error> {
|
||||
if *OLD_LOCKING {
|
||||
lock_dir_noblock_shared(
|
||||
&self.full_path(),
|
||||
"snapshot",
|
||||
"backup is running or snapshot is in use, could not acquire shared lock",
|
||||
)
|
||||
.map(BackupLockGuard::from)
|
||||
} else {
|
||||
lock_helper(self.store.name(), &self.lock_path(), |p| {
|
||||
open_backup_lockfile(p, Some(Duration::from_secs(0)), false)
|
||||
.with_context(|| format!("unable to acquire shared snapshot lock {p:?}"))
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
/// Destroy the whole snapshot, bails if it's protected
|
||||
@ -494,11 +561,13 @@ impl BackupDir {
|
||||
format_err!("removing backup snapshot {:?} failed - {}", full_path, err,)
|
||||
})?;
|
||||
|
||||
// the manifest doesn't exist anymore, no need to keep the lock (already done by guard?)
|
||||
// remove no longer needed lock files
|
||||
if let Ok(path) = self.manifest_lock_path() {
|
||||
let _ = std::fs::remove_file(path); // ignore errors
|
||||
}
|
||||
|
||||
let _ = std::fs::remove_file(self.lock_path()); // ignore errors
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@ -692,3 +761,75 @@ fn list_backup_files<P: ?Sized + nix::NixPath>(
|
||||
|
||||
Ok(files)
|
||||
}
|
||||
|
||||
/// Creates a path to a lock file depending on the relative path of an object (snapshot, group,
|
||||
/// manifest) in a datastore. First all namespaces will be concatenated with a colon (ns-folder).
|
||||
/// Then the actual file name will depend on the length of the relative path without namespaces. If
|
||||
/// it is shorter than 255 characters in its unit encoded form, than the unit encoded form will be
|
||||
/// used directly. If not, the file name will consist of the first 80 character, the last 80
|
||||
/// characters and the hash of the unit encoded relative path without namespaces. It will also be
|
||||
/// placed into a "hashed" subfolder in the namespace folder.
|
||||
///
|
||||
/// Examples:
|
||||
///
|
||||
/// - vm-100
|
||||
/// - vm-100-2022\x2d05\x2d02T08\x3a11\x3a33Z
|
||||
/// - ns1:ns2:ns3:ns4:ns5:ns6:ns7/vm-100-2022\x2d05\x2d02T08\x3a11\x3a33Z
|
||||
///
|
||||
/// A "hashed" lock file would look like this:
|
||||
/// - ns1:ns2:ns3/hashed/$first_eighty...$last_eighty-$hash
|
||||
fn lock_file_path_helper(ns: &BackupNamespace, path: PathBuf) -> PathBuf {
|
||||
let to_return = PathBuf::from(
|
||||
ns.components()
|
||||
.map(String::from)
|
||||
.reduce(|acc, n| format!("{acc}:{n}"))
|
||||
.unwrap_or_default(),
|
||||
);
|
||||
|
||||
let path_bytes = path.as_os_str().as_bytes();
|
||||
|
||||
let enc = escape_unit(path_bytes, true);
|
||||
|
||||
if enc.len() < 255 {
|
||||
return to_return.join(enc);
|
||||
}
|
||||
|
||||
let to_return = to_return.join("hashed");
|
||||
|
||||
let first_eigthy = &enc[..80];
|
||||
let last_eighty = &enc[enc.len() - 80..];
|
||||
let hash = hex::encode(openssl::sha::sha256(path_bytes));
|
||||
|
||||
to_return.join(format!("{first_eigthy}...{last_eighty}-{hash}"))
|
||||
}
|
||||
|
||||
/// Helps implement the double stat'ing procedure. It avoids certain race conditions upon lock
|
||||
/// deletion.
|
||||
///
|
||||
/// It also creates the base directory for lock files.
|
||||
fn lock_helper<F>(
|
||||
store_name: &str,
|
||||
path: &std::path::Path,
|
||||
lock_fn: F,
|
||||
) -> Result<BackupLockGuard, Error>
|
||||
where
|
||||
F: Fn(&std::path::Path) -> Result<BackupLockGuard, Error>,
|
||||
{
|
||||
let mut lock_dir = Path::new(DATASTORE_LOCKS_DIR).join(store_name);
|
||||
|
||||
if let Some(parent) = path.parent() {
|
||||
lock_dir = lock_dir.join(parent);
|
||||
};
|
||||
|
||||
std::fs::create_dir_all(&lock_dir)?;
|
||||
|
||||
let lock = lock_fn(path)?;
|
||||
|
||||
let inode = nix::sys::stat::fstat(lock.as_raw_fd())?.st_ino;
|
||||
|
||||
if nix::sys::stat::stat(path).map_or(true, |st| inode != st.st_ino) {
|
||||
bail!("could not acquire lock, another thread modified the lock file");
|
||||
}
|
||||
|
||||
Ok(lock)
|
||||
}
|
||||
|
@ -13,7 +13,6 @@ use proxmox_human_byte::HumanByte;
|
||||
use proxmox_schema::ApiType;
|
||||
|
||||
use proxmox_sys::error::SysError;
|
||||
use proxmox_sys::fs::DirLockGuard;
|
||||
use proxmox_sys::fs::{file_read_optional_string, replace_file, CreateOptions};
|
||||
use proxmox_sys::linux::procfs::MountInfo;
|
||||
use proxmox_sys::process_locker::ProcessLockSharedGuard;
|
||||
@ -24,6 +23,7 @@ use pbs_api_types::{
|
||||
DataStoreConfig, DatastoreFSyncLevel, DatastoreTuning, GarbageCollectionStatus,
|
||||
MaintenanceMode, MaintenanceType, Operation, UPID,
|
||||
};
|
||||
use pbs_config::BackupLockGuard;
|
||||
|
||||
use crate::backup_info::{BackupDir, BackupGroup};
|
||||
use crate::chunk_store::ChunkStore;
|
||||
@ -778,7 +778,7 @@ impl DataStore {
|
||||
ns: &BackupNamespace,
|
||||
backup_group: &pbs_api_types::BackupGroup,
|
||||
auth_id: &Authid,
|
||||
) -> Result<(Authid, DirLockGuard), Error> {
|
||||
) -> Result<(Authid, BackupLockGuard), Error> {
|
||||
let backup_group = self.backup_group(ns.clone(), backup_group.clone());
|
||||
|
||||
// create intermediate path first
|
||||
@ -816,7 +816,7 @@ impl DataStore {
|
||||
self: &Arc<Self>,
|
||||
ns: &BackupNamespace,
|
||||
backup_dir: &pbs_api_types::BackupDir,
|
||||
) -> Result<(PathBuf, bool, DirLockGuard), Error> {
|
||||
) -> Result<(PathBuf, bool, BackupLockGuard), Error> {
|
||||
let backup_dir = self.backup_dir(ns.clone(), backup_dir.clone())?;
|
||||
let relative_path = backup_dir.relative_path();
|
||||
|
||||
|
@ -6,6 +6,9 @@ use std::sync::Arc;
|
||||
|
||||
use anyhow::{bail, Context, Error};
|
||||
use nix::dir::Dir;
|
||||
use nix::fcntl::OFlag;
|
||||
use nix::sys::stat::Mode;
|
||||
use pbs_config::BackupLockGuard;
|
||||
|
||||
use pbs_api_types::{
|
||||
print_store_and_ns, ArchiveType, BackupNamespace, Operation, CLIENT_LOG_BLOB_NAME,
|
||||
@ -26,6 +29,10 @@ pub struct SnapshotReader {
|
||||
datastore_name: String,
|
||||
file_list: Vec<String>,
|
||||
locked_dir: Dir,
|
||||
|
||||
// while this is never read, the lock needs to be kept until the
|
||||
// reader is dropped to ensure valid locking semantics
|
||||
_lock: BackupLockGuard,
|
||||
}
|
||||
|
||||
impl SnapshotReader {
|
||||
@ -46,10 +53,13 @@ impl SnapshotReader {
|
||||
bail!("snapshot {} does not exist!", snapshot.dir());
|
||||
}
|
||||
|
||||
let locked_dir = snapshot
|
||||
let lock = snapshot
|
||||
.lock_shared()
|
||||
.with_context(|| format!("while trying to read snapshot '{snapshot:?}'"))?;
|
||||
|
||||
let locked_dir = Dir::open(&snapshot_path, OFlag::O_RDONLY, Mode::empty())
|
||||
.with_context(|| format!("unable to open snapshot directory {snapshot_path:?}"))?;
|
||||
|
||||
let datastore_name = datastore.name().to_string();
|
||||
let manifest = match snapshot.load_manifest() {
|
||||
Ok((manifest, _)) => manifest,
|
||||
@ -79,6 +89,7 @@ impl SnapshotReader {
|
||||
datastore_name,
|
||||
file_list,
|
||||
locked_dir,
|
||||
_lock: lock,
|
||||
})
|
||||
}
|
||||
|
||||
|
@ -1,5 +1,6 @@
|
||||
use anyhow::{bail, format_err, Context, Error};
|
||||
use nix::dir::Dir;
|
||||
use pbs_config::BackupLockGuard;
|
||||
|
||||
use std::collections::HashMap;
|
||||
use std::sync::{Arc, Mutex};
|
||||
use tracing::info;
|
||||
@ -635,7 +636,7 @@ impl BackupEnvironment {
|
||||
/// If verify-new is set on the datastore, this will run a new verify task
|
||||
/// for the backup. If not, this will return and also drop the passed lock
|
||||
/// immediately.
|
||||
pub fn verify_after_complete(&self, excl_snap_lock: Dir) -> Result<(), Error> {
|
||||
pub fn verify_after_complete(&self, excl_snap_lock: BackupLockGuard) -> Result<(), Error> {
|
||||
self.ensure_finished()?;
|
||||
|
||||
if !self.datastore.verify_new() {
|
||||
|
@ -1,10 +1,10 @@
|
||||
use pbs_config::BackupLockGuard;
|
||||
use std::collections::HashSet;
|
||||
use std::sync::atomic::{AtomicUsize, Ordering};
|
||||
use std::sync::{Arc, Mutex};
|
||||
use std::time::Instant;
|
||||
|
||||
use anyhow::{bail, Error};
|
||||
use proxmox_sys::fs::DirLockGuard;
|
||||
use tracing::{error, info, warn};
|
||||
|
||||
use proxmox_worker_task::WorkerTaskContext;
|
||||
@ -330,7 +330,7 @@ pub fn verify_backup_dir_with_lock(
|
||||
backup_dir: &BackupDir,
|
||||
upid: UPID,
|
||||
filter: Option<&dyn Fn(&BackupManifest) -> bool>,
|
||||
_snap_lock: DirLockGuard,
|
||||
_snap_lock: BackupLockGuard,
|
||||
) -> Result<bool, Error> {
|
||||
let datastore_name = verify_worker.datastore.name();
|
||||
let backup_dir_name = backup_dir.dir();
|
||||
|
@ -10,7 +10,7 @@ use std::time::Duration;
|
||||
use anyhow::{bail, format_err, Context, Error};
|
||||
use futures::{future::FutureExt, select};
|
||||
use hyper::http::StatusCode;
|
||||
use proxmox_sys::fs::DirLockGuard;
|
||||
use pbs_config::BackupLockGuard;
|
||||
use serde_json::json;
|
||||
use tracing::{info, warn};
|
||||
|
||||
@ -106,7 +106,7 @@ pub(crate) struct RemoteSourceReader {
|
||||
}
|
||||
|
||||
pub(crate) struct LocalSourceReader {
|
||||
pub(crate) _dir_lock: Arc<Mutex<DirLockGuard>>,
|
||||
pub(crate) _dir_lock: Arc<Mutex<BackupLockGuard>>,
|
||||
pub(crate) path: PathBuf,
|
||||
pub(crate) datastore: Arc<DataStore>,
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user