qemu-server/PVE/VZDump/QemuServer.pm
Wolfgang Bumiller 8009da73f1 fix #6317: backup: use correct cleanup_fleecing_images helper
The local one is specific for `allocate_fleecing_images` and has a
comment stating to use the one from `PVE::QemuConfig` in all other
cases.

The `cleanup` sub already called this, but only if the VM was running.
We do allocate fleecing images for previously-stopped VMs as well,
though, so we also need to do the cleanup.

As for the `detach_fleecing_images()` call: while could have stayed in
the `vm_running_locall()` branch, it also performs this check and this
way the entire fleecing cleanup stays together in one place.

Signed-off-by: Wolfgang Bumiller <w.bumiller@proxmox.com>
2025-04-16 12:39:34 +02:00

1612 lines
48 KiB
Perl

package PVE::VZDump::QemuServer;
use strict;
use warnings;
use Fcntl qw(:mode);
use File::Basename;
use File::Path qw(make_path remove_tree);
use File::stat qw();
use IO::File;
use IPC::Open3;
use JSON;
use POSIX qw(EINTR EAGAIN);
use Time::HiRes qw(usleep);
use PVE::Cluster qw(cfs_read_file);
use PVE::INotify;
use PVE::IPCC;
use PVE::JSONSchema;
use PVE::PBSClient;
use PVE::RESTEnvironment qw(log_warn);
use PVE::QMPClient;
use PVE::Storage::Plugin;
use PVE::Storage::PBSPlugin;
use PVE::Storage;
use PVE::Tools qw(run_command);
use PVE::VZDump;
use PVE::Format qw(render_duration render_bytes);
use PVE::QemuConfig;
use PVE::QemuServer;
use PVE::QemuServer::Drive qw(checked_volume_format);
use PVE::QemuServer::Helpers;
use PVE::QemuServer::Machine;
use PVE::QemuServer::Monitor qw(mon_cmd);
use PVE::QemuServer::QMPHelpers;
use base qw (PVE::VZDump::Plugin);
sub new {
my ($class, $vzdump) = @_;
PVE::VZDump::check_bin('qm');
my $self = bless { vzdump => $vzdump }, $class;
$self->{vmlist} = PVE::QemuServer::vzlist();
$self->{storecfg} = PVE::Storage::config();
return $self;
};
sub type {
return 'qemu';
}
sub vmlist {
my ($self) = @_;
return [ keys %{$self->{vmlist}} ];
}
sub prepare {
my ($self, $task, $vmid, $mode) = @_;
my $running = PVE::QemuServer::Helpers::vm_running_locally($vmid);
if ($running && (my $status = mon_cmd($vmid, 'query-backup'))) {
if ($status->{status} && $status->{status} eq 'active') {
$self->log('warn', "left-over backup job still running inside QEMU - canceling now");
mon_cmd($vmid, 'backup-cancel');
}
}
$task->{disks} = [];
my $conf = $self->{vmlist}->{$vmid} = PVE::QemuConfig->load_config($vmid);
$self->loginfo("VM Name: $conf->{name}")
if defined($conf->{name});
$self->{vm_was_running} = $running ? 1 : 0;
$self->{vm_was_paused} = 0;
if ($running && PVE::QemuServer::vm_is_paused($vmid, 0)) {
# Do not treat a suspended VM as paused, as it would cause us to skip
# fs-freeze even if the VM wakes up before we reach qga_fs_freeze.
$self->{vm_was_paused} = 1;
}
$task->{hostname} = $conf->{name};
my $hostname = PVE::INotify::nodename();
my $vollist = [];
my $drivehash = {};
my $backup_volumes = PVE::QemuConfig->get_backup_volumes($conf);
foreach my $volume (@{$backup_volumes}) {
my $name = $volume->{key};
my $volume_config = $volume->{volume_config};
my $volid = $volume_config->{file};
if (!$volume->{included}) {
$self->loginfo("exclude disk '$name' '$volid' ($volume->{reason})");
next;
} else {
my $log = "include disk '$name' '$volid'";
if (defined(my $size = $volume_config->{size})) {
my $readable_size = PVE::JSONSchema::format_size($size);
$log .= " $readable_size";
}
$self->loginfo($log);
}
my ($storeid, $volname) = PVE::Storage::parse_volume_id($volid, 1);
push @$vollist, $volid if $storeid;
$drivehash->{$name} = $volume->{volume_config};
}
PVE::Storage::activate_volumes($self->{storecfg}, $vollist);
foreach my $ds (sort keys %$drivehash) {
my $drive = $drivehash->{$ds};
my $volid = $drive->{file};
my ($storeid, $volname) = PVE::Storage::parse_volume_id($volid, 1);
my $path = $volid;
if ($storeid) {
$path = PVE::Storage::path($self->{storecfg}, $volid);
}
next if !$path;
my ($size, $format);
if ($storeid) {
# The call in list context can be expensive for certain plugins like RBD, just get size
$size = eval { PVE::Storage::volume_size_info($self->{storecfg}, $volid, 5) };
die "cannot determine size of volume '$volid' - $@\n" if $@;
$format = checked_volume_format($self->{storecfg}, $volid);
} else {
($size, $format) = eval {
PVE::Storage::volume_size_info($self->{storecfg}, $volid, 5);
};
die "cannot determine size and format of volume '$volid' - $@\n" if $@;
}
my $diskinfo = {
path => $path,
volid => $volid,
storeid => $storeid,
size => $size,
format => $format,
virtdev => $ds,
qmdevice => "drive-$ds",
};
if ($ds eq 'tpmstate0') {
# TPM drive only exists for backup, which is reflected in the name
$diskinfo->{qmdevice} = 'drive-tpmstate0-backup';
$task->{tpmpath} = $path;
}
if (-b $path) {
$diskinfo->{type} = 'block';
} else {
$diskinfo->{type} = 'file';
}
push @{$task->{disks}}, $diskinfo;
}
}
sub vm_status {
my ($self, $vmid) = @_;
my $running = PVE::QemuServer::check_running($vmid) ? 1 : 0;
return wantarray ? ($running, $running ? 'running' : 'stopped') : $running;
}
sub lock_vm {
my ($self, $vmid) = @_;
PVE::QemuConfig->set_lock($vmid, 'backup');
}
sub unlock_vm {
my ($self, $vmid) = @_;
PVE::QemuConfig->remove_lock($vmid, 'backup');
}
sub stop_vm {
my ($self, $task, $vmid) = @_;
my $opts = $self->{vzdump}->{opts};
my $wait = $opts->{stopwait} * 60;
# send shutdown and wait
$self->cmd ("qm shutdown $vmid --skiplock --keepActive --timeout $wait");
}
sub start_vm {
my ($self, $task, $vmid) = @_;
$self->cmd ("qm start $vmid --skiplock");
}
sub suspend_vm {
my ($self, $task, $vmid) = @_;
return if $self->{vm_was_paused};
$self->cmd ("qm suspend $vmid --skiplock");
}
sub resume_vm {
my ($self, $task, $vmid) = @_;
return if $self->{vm_was_paused};
$self->cmd ("qm resume $vmid --skiplock");
}
sub assemble {
my ($self, $task, $vmid) = @_;
my $conffile = PVE::QemuConfig->config_file($vmid);
my $outfile = "$task->{tmpdir}/qemu-server.conf";
my $firewall_src = "/etc/pve/firewall/$vmid.fw";
my $firewall_dest = "$task->{tmpdir}/qemu-server.fw";
my $outfd = IO::File->new(">$outfile") or die "unable to open '$outfile' - $!\n";
my $conffd = IO::File->new($conffile, 'r') or die "unable to open '$conffile' - $!\n";
my $found_snapshot;
my $found_pending;
my $found_special;
while (defined (my $line = <$conffd>)) {
next if $line =~ m/^\#vzdump\#/; # just to be sure
next if $line =~ m/^\#qmdump\#/; # just to be sure
if ($line =~ m/^\[(.*)\]\s*$/) {
if ($1 =~ m/^PENDING$/i) {
$found_pending = 1;
} elsif ($1 =~ m/^special:.*$/) {
$found_special = 1;
} else {
$found_snapshot = 1;
}
}
# skip all snapshots, pending changes and special sections
next if $found_snapshot || $found_pending || $found_special;
if ($line =~ m/^unused\d+:\s*(\S+)\s*/) {
$self->loginfo("skip unused drive '$1' (not included into backup)");
next;
}
next if $line =~ m/^lock:/ || $line =~ m/^parent:/;
print $outfd $line;
}
foreach my $di (@{$task->{disks}}) {
if ($di->{type} eq 'block' || $di->{type} eq 'file') {
my $storeid = $di->{storeid} || '';
my $format = $di->{format} || '';
print $outfd "#qmdump#map:$di->{virtdev}:$di->{qmdevice}:$storeid:$format:\n";
} else {
die "internal error";
}
}
if ($found_special) {
$self->loginfo("special config section found (not included into backup)");
}
if ($found_snapshot) {
$self->loginfo("snapshots found (not included into backup)");
}
if ($found_pending) {
$self->loginfo("pending configuration changes found (not included into backup)");
}
PVE::Tools::file_copy($firewall_src, $firewall_dest) if -f $firewall_src;
}
sub archive {
my ($self, $task, $vmid, $filename, $comp) = @_;
my $opts = $self->{vzdump}->{opts};
my $scfg = $opts->{scfg};
if ($self->{vzdump}->{opts}->{pbs}) {
$self->archive_pbs($task, $vmid);
} elsif ($self->{vzdump}->{'backup-provider'}) {
$self->archive_external($task, $vmid);
} else {
$self->archive_vma($task, $vmid, $filename, $comp);
}
}
my $bitmap_action_to_human = sub {
my ($self, $info) = @_;
my $action = $info->{action};
if ($action eq "not-used") {
return "disabled (no support)";
} elsif ($action eq "not-used-removed") {
return "disabled (old bitmap cleared)";
} elsif ($action eq "new") {
return "created new";
} elsif ($action eq "used") {
if ($info->{dirty} == 0) {
return "OK (drive clean)";
} else {
my $size = render_bytes($info->{size}, 1);
my $dirty = render_bytes($info->{dirty}, 1);
return "OK ($dirty of $size dirty)";
}
} elsif ($action eq "invalid") {
return "existing bitmap was invalid and has been cleared";
} elsif ($action eq "missing-recreated") {
# Lie about the TPM state, because it is newly attached each time.
return "created new" if $info->{drive} eq 'drive-tpmstate0-backup';
return "expected bitmap was missing and has been recreated";
} else {
return "unknown";
}
};
my $query_backup_status_loop = sub {
my ($self, $vmid, $job_uuid, $qemu_support) = @_;
my $starttime = time ();
my $last_time = $starttime;
my ($last_percent, $last_total, $last_target, $last_zero, $last_transferred) = (-1, 0, 0, 0, 0);
my ($transferred, $reused);
my $get_mbps = sub {
my ($mb, $delta) = @_;
return "0 B/s" if $mb <= 0;
my $bw = int(($mb / $delta));
return render_bytes($bw, 1) . "/s";
};
my $target = 0;
my $last_reused = 0;
my $has_query_bitmap = $qemu_support && $qemu_support->{'query-bitmap-info'};
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
if ($has_query_bitmap) {
my $total = 0;
my $bitmap_info = mon_cmd($vmid, 'query-pbs-bitmap-info');
for my $info (sort { $a->{drive} cmp $b->{drive} } @$bitmap_info) {
if (!$is_template) {
my $text = $bitmap_action_to_human->($self, $info);
my $drive = $info->{drive};
$drive =~ s/^drive-//; # for consistency
$self->loginfo("$drive: dirty-bitmap status: $text");
}
$target += $info->{dirty};
$total += $info->{size};
$last_reused += $info->{size} - $info->{dirty};
}
if ($target < $total) {
my $total_h = render_bytes($total, 1);
my $target_h = render_bytes($target, 1);
$self->loginfo("using fast incremental mode (dirty-bitmap), $target_h dirty of $total_h total");
}
}
my $last_finishing = 0;
while(1) {
my $status = mon_cmd($vmid, 'query-backup');
my $total = $status->{total} || 0;
my $dirty = $status->{dirty};
$target = (defined($dirty) && $dirty < $total) ? $dirty : $total if !$has_query_bitmap;
$transferred = $status->{transferred} || 0;
$reused = $status->{reused};
my $percent = $target ? int(($transferred * 100)/$target) : 100;
my $zero = $status->{'zero-bytes'} || 0;
die "got unexpected uuid\n" if !$status->{uuid} || ($status->{uuid} ne $job_uuid);
my $ctime = time();
my $duration = $ctime - $starttime;
my $rbytes = $transferred - $last_transferred;
my $wbytes;
if ($reused) {
# reused includes zero bytes for PBS
$wbytes = $rbytes - ($reused - $last_reused);
} else {
$wbytes = $rbytes - ($zero - $last_zero);
}
my $timediff = ($ctime - $last_time) || 1; # fixme
my $mbps_read = $get_mbps->($rbytes, $timediff);
my $mbps_write = $get_mbps->($wbytes, $timediff);
my $target_h = render_bytes($target, 1);
my $transferred_h = render_bytes($transferred, 1);
my $statusline = sprintf("%3d%% ($transferred_h of $target_h) in %s"
.", read: $mbps_read, write: $mbps_write", $percent, render_duration($duration));
my $res = $status->{status} || 'unknown';
if ($res ne 'active') {
if ($last_percent < 100) {
$self->loginfo($statusline);
}
if ($res ne 'done') {
die (($status->{errmsg} || "unknown error") . "\n") if $res eq 'error';
die "got unexpected status '$res'\n";
}
$last_target = $target if $target;
$last_total = $total if $total;
$last_zero = $zero if $zero;
$last_transferred = $transferred if $transferred;
last;
}
if ($percent != $last_percent && ($timediff > 2)) {
$self->loginfo($statusline);
$last_percent = $percent;
$last_target = $target if $target;
$last_total = $total if $total;
$last_zero = $zero if $zero;
$last_transferred = $transferred if $transferred;
$last_time = $ctime;
$last_reused = $reused;
if (!$last_finishing && $status->{finishing}) {
$self->loginfo("Waiting for server to finish backup validation...");
}
$last_finishing = $status->{finishing};
}
sleep(1);
}
my $duration = time() - $starttime;
if ($last_zero) {
my $zero_per = $last_target ? int(($last_zero * 100)/$last_target) : 0;
my $zero_h = render_bytes($last_zero);
$self->loginfo("backup is sparse: $zero_h (${zero_per}%) total zero data");
}
if ($reused) {
my $reused_h = render_bytes($reused);
my $reuse_per = int($reused * 100 / $last_total);
$self->loginfo("backup was done incrementally, reused $reused_h (${reuse_per}%)");
}
if ($transferred) {
my $transferred_h = render_bytes($transferred);
if ($duration) {
my $mbps = $get_mbps->($transferred, $duration);
$self->loginfo("transferred $transferred_h in $duration seconds ($mbps)");
} else {
$self->loginfo("transferred $transferred_h in <1 seconds");
}
}
return {
total => $last_total,
reused => $reused,
};
};
my $attach_tpmstate_drive = sub {
my ($self, $task, $vmid) = @_;
return if !$task->{tpmpath};
# unconditionally try to remove the tpmstate-named drive - it only exists
# for backing up, and avoids errors if left over from some previous event
eval { PVE::QemuServer::qemu_drivedel($vmid, "tpmstate0-backup"); };
$self->loginfo('attaching TPM drive to QEMU for backup');
my $drive = "file=$task->{tpmpath},if=none,read-only=on,id=drive-tpmstate0-backup";
$drive =~ s/\\/\\\\/g;
my $ret = PVE::QemuServer::Monitor::hmp_cmd($vmid, "drive_add auto \"$drive\"", 60);
die "attaching TPM drive failed - $ret\n" if $ret !~ m/OK/s;
};
my $detach_tpmstate_drive = sub {
my ($task, $vmid) = @_;
return if !$task->{tpmpath} || !PVE::QemuServer::check_running($vmid);
eval { PVE::QemuServer::qemu_drivedel($vmid, "tpmstate0-backup"); };
};
my sub add_backup_performance_options {
my ($qmp_param, $perf, $qemu_support) = @_;
return if !$perf || scalar(keys $perf->%*) == 0;
if (!$qemu_support) {
my $settings_string = join(', ', sort keys $perf->%*);
log_warn("ignoring setting(s): $settings_string - issue checking if supported");
return;
}
if (defined($perf->{'max-workers'})) {
if ($qemu_support->{'backup-max-workers'}) {
$qmp_param->{'max-workers'} = int($perf->{'max-workers'});
} else {
log_warn("ignoring 'max-workers' setting - not supported by running QEMU");
}
}
}
sub get_and_check_pbs_encryption_config {
my ($self) = @_;
my $opts = $self->{vzdump}->{opts};
my $scfg = $opts->{scfg};
my $keyfile = PVE::Storage::PBSPlugin::pbs_encryption_key_file_name($scfg, $opts->{storage});
my $master_keyfile = PVE::Storage::PBSPlugin::pbs_master_pubkey_file_name($scfg, $opts->{storage});
if (-e $keyfile) {
if (-e $master_keyfile) {
$self->loginfo("enabling encryption with master key feature");
return ($keyfile, $master_keyfile);
} elsif ($scfg->{'master-pubkey'}) {
die "master public key configured but no key file found\n";
} else {
$self->loginfo("enabling encryption");
return ($keyfile, undef);
}
} else {
my $encryption_fp = $scfg->{'encryption-key'};
die "encryption configured ('$encryption_fp') but no encryption key file found!\n"
if $encryption_fp;
if (-e $master_keyfile) {
$self->log(
'warn',
"backup target storage is configured with master-key, but no encryption key set!"
." Ignoring master key settings and creating unencrypted backup."
);
}
return (undef, undef);
}
die "internal error - unhandled case for getting & checking PBS encryption ($keyfile, $master_keyfile)!";
}
# Helper is intended to be called from allocate_fleecing_images() only. Otherwise, fleecing volids
# have already been recorded in the configuration and PVE::QemuConfig::cleanup_fleecing_images()
# should be used instead.
my sub cleanup_fleecing_images {
my ($self, $vmid, $disks) = @_;
my $failed = [];
for my $di ($disks->@*) {
if (my $volid = $di->{'fleece-volid'}) {
eval { PVE::Storage::vdisk_free($self->{storecfg}, $volid); };
if (my $err = $@) {
$self->log('warn', "error removing fleecing image '$volid' - $err");
push $failed->@*, $volid;
}
}
}
PVE::QemuConfig::record_fleecing_images($vmid, $failed);
}
my sub allocate_fleecing_images {
my ($self, $disks, $vmid, $fleecing_storeid, $format, $all_images) = @_;
die "internal error - no fleecing storage specified\n" if !$fleecing_storeid;
my $fleece_volids = [];
eval {
my $n = 0; # counter for fleecing image names
for my $di ($disks->@*) {
# EFI/TPM are usually too small to be worth it, but it's required for external providers
next if !$all_images && $di->{virtdev} =~ m/^(?:tpmstate|efidisk)\d$/;
if ($di->{type} eq 'block' || $di->{type} eq 'file') {
my $scfg = PVE::Storage::storage_config($self->{storecfg}, $fleecing_storeid);
my $name = "vm-$vmid-fleece-$n";
$name .= ".$format" if $scfg->{path};
my $size;
if ($format ne 'raw') {
# Since non-raw images cannot be attached with an explicit 'size' parameter to
# QEMU later, pass the exact size to the storage layer. This makes qcow2
# fleecing images work for non-1KiB-aligned source images.
$size = $di->{'block-node-size'}/1024;
} else {
$size = PVE::Tools::convert_size($di->{'block-node-size'}, 'b' => 'kb');
}
$di->{'fleece-volid'} = PVE::Storage::vdisk_alloc(
$self->{storecfg}, $fleecing_storeid, $vmid, $format, $name, $size);
push $fleece_volids->@*, $di->{'fleece-volid'};
$n++;
} else {
die "implement me (type '$di->{type}')";
}
}
};
if (my $err = $@) {
cleanup_fleecing_images($self, $vmid, $disks);
die $err;
}
PVE::QemuConfig::record_fleecing_images($vmid, $fleece_volids);
}
my sub detach_fleecing_images {
my ($disks, $vmid) = @_;
return if !PVE::QemuServer::Helpers::vm_running_locally($vmid);
for my $di ($disks->@*) {
if (my $volid = $di->{'fleece-volid'}) {
my $devid = "$di->{qmdevice}-fleecing";
$devid =~ s/^drive-//; # re-added by qemu_drivedel()
eval { PVE::QemuServer::qemu_drivedel($vmid, $devid) };
}
}
}
my sub attach_fleecing_images {
my ($self, $disks, $vmid, $format) = @_;
# unconditionally try to remove potential left-overs from a previous backup
detach_fleecing_images($disks, $vmid);
my $vollist = [ map { $_->{'fleece-volid'} } grep { $_->{'fleece-volid'} } $disks->@* ];
PVE::Storage::activate_volumes($self->{storecfg}, $vollist);
for my $di ($disks->@*) {
if (my $volid = $di->{'fleece-volid'}) {
$self->loginfo("$di->{qmdevice}: attaching fleecing image $volid to QEMU");
my $path = PVE::Storage::path($self->{storecfg}, $volid);
my $devid = "$di->{qmdevice}-fleecing";
my $drive = "file=$path,if=none,id=$devid,format=$format,discard=unmap";
# Specify size explicitly, to make it work if storage backend rounded up size for
# fleecing image when allocating.
$drive .= ",size=$di->{'block-node-size'}" if $format eq 'raw';
$drive =~ s/\\/\\\\/g;
my $ret = PVE::QemuServer::Monitor::hmp_cmd($vmid, "drive_add auto \"$drive\"", 60);
die "attaching fleecing image $volid failed - $ret\n" if $ret !~ m/OK/s;
}
}
}
my sub check_and_prepare_fleecing {
my ($self, $vmid, $fleecing_opts, $disks, $is_template, $qemu_support, $all_images) = @_;
# Even if the VM was started specifically for fleecing, it's possible that the VM is resumed and
# then starts doing IO. For VMs that are not resumed the fleecing images will just stay empty,
# so there is no big cost.
my $use_fleecing = $fleecing_opts && $fleecing_opts->{enabled} && !$is_template;
if ($use_fleecing && !$qemu_support->{'backup-fleecing'}) {
$self->log(
'warn',
"running QEMU version does not support backup fleecing - continuing without",
);
$use_fleecing = 0;
}
# clean up potential left-overs from a previous attempt
eval {
PVE::QemuConfig::cleanup_fleecing_images(
$vmid, $self->{storecfg}, sub { $self->log($_[0], $_[1]); });
};
$self->log('warn', "attempt to clean up left-over fleecing images failed - $@") if $@;
if ($use_fleecing) {
$self->query_block_node_sizes($vmid, $disks);
my ($default_format, $valid_formats) = PVE::Storage::storage_default_format(
$self->{storecfg}, $fleecing_opts->{storage});
my $format = scalar(grep { $_ eq 'qcow2' } $valid_formats->@*) ? 'qcow2' : 'raw';
allocate_fleecing_images(
$self, $disks, $vmid, $fleecing_opts->{storage}, $format, $all_images);
attach_fleecing_images($self, $disks, $vmid, $format);
}
return $use_fleecing;
}
sub archive_pbs {
my ($self, $task, $vmid) = @_;
my $conffile = "$task->{tmpdir}/qemu-server.conf";
my $firewall = "$task->{tmpdir}/qemu-server.fw";
my $opts = $self->{vzdump}->{opts};
my $scfg = $opts->{scfg};
my $starttime = time();
my $fingerprint = $scfg->{fingerprint};
my $repo = PVE::PBSClient::get_repository($scfg);
my $password = PVE::Storage::PBSPlugin::pbs_get_password($scfg, $opts->{storage});
my ($keyfile, $master_keyfile) = $self->get_and_check_pbs_encryption_config();
my $diskcount = scalar(@{$task->{disks}});
# proxmox-backup-client can only handle raw files and block devs, so only use it (directly) for
# disk-less VMs
if (!$diskcount) {
$self->loginfo("backup contains no disks");
local $ENV{PBS_PASSWORD} = $password;
local $ENV{PBS_FINGERPRINT} = $fingerprint if defined($fingerprint);
my $cmd = [
'/usr/bin/proxmox-backup-client',
'backup',
'--repository', $repo,
'--backup-type', 'vm',
'--backup-id', "$vmid",
'--backup-time', $task->{backup_time},
];
if (defined(my $ns = $scfg->{namespace})) {
push @$cmd, '--ns', $ns;
}
if (defined($keyfile)) {
push @$cmd, '--keyfile', $keyfile;
push @$cmd, '--master-pubkey-file', $master_keyfile if defined($master_keyfile);
}
push @$cmd, "qemu-server.conf:$conffile";
push @$cmd, "fw.conf:$firewall" if -e $firewall;
$self->loginfo("starting diskless backup");
$self->loginfo(join(' ', @$cmd));
$self->cmd($cmd);
return;
}
# get list early so we die on unknown drive types before doing anything
my $devlist = _get_task_devlist($task);
my $backup_job_uuid;
eval {
$SIG{INT} = $SIG{TERM} = $SIG{QUIT} = $SIG{HUP} = $SIG{PIPE} = sub {
die "interrupted by signal\n";
};
$self->enforce_vm_running_for_backup($vmid);
$self->{qmeventd_fh} = PVE::QemuServer::register_qmeventd_handle($vmid);
my $qemu_support = eval { mon_cmd($vmid, "query-proxmox-support") };
my $err = $@;
if (!$qemu_support || $err) {
die "query-proxmox-support returned empty value\n" if !$err;
if ($err =~ m/The command query-proxmox-support has not been found/) {
die "PBS backups are not supported by the running QEMU version. Please make "
. "sure you've installed the latest version and the VM has been restarted.\n";
} else {
die "QMP command query-proxmox-support failed - $err\n";
}
}
# pve-qemu supports it since 5.2.0-1 (PVE 6.4), so safe to die since PVE 8
die "master key configured but running QEMU version does not support master keys\n"
if !$qemu_support->{'pbs-masterkey'} && defined($master_keyfile);
$attach_tpmstate_drive->($self, $task, $vmid);
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
$task->{'use-fleecing'} = check_and_prepare_fleecing(
$self, $vmid, $opts->{fleecing}, $task->{disks}, $is_template, $qemu_support, 0);
my $fs_frozen = $self->qga_fs_freeze($task, $vmid);
my $params = {
format => "pbs",
'backup-file' => $repo,
'backup-id' => "$vmid",
'backup-time' => $task->{backup_time},
password => $password,
devlist => $devlist,
'config-file' => $conffile,
};
$params->{fleecing} = JSON::true if $task->{'use-fleecing'};
if (defined(my $ns = $scfg->{namespace})) {
$params->{'backup-ns'} = $ns;
}
$params->{speed} = $opts->{bwlimit}*1024 if $opts->{bwlimit};
add_backup_performance_options($params, $opts->{performance}, $qemu_support);
$params->{fingerprint} = $fingerprint if defined($fingerprint);
$params->{'firewall-file'} = $firewall if -e $firewall;
$params->{encrypt} = defined($keyfile) ? JSON::true : JSON::false;
if (defined($keyfile)) {
$params->{keyfile} = $keyfile;
$params->{"master-keyfile"} = $master_keyfile if defined($master_keyfile);
}
$params->{'use-dirty-bitmap'} = JSON::true
if $qemu_support->{'pbs-dirty-bitmap'} && !$is_template;
$params->{timeout} = 125; # give some time to connect to the backup server
my $res = eval { mon_cmd($vmid, "backup", %$params) };
my $qmperr = $@;
$backup_job_uuid = $res->{UUID} if $res;
if ($fs_frozen) {
$self->qga_fs_thaw($vmid);
}
die $qmperr if $qmperr;
die "got no uuid for backup task\n" if !defined($backup_job_uuid);
$self->loginfo("started backup task '$backup_job_uuid'");
$self->resume_vm_after_job_start($task, $vmid);
my $stat = $query_backup_status_loop->($self, $vmid, $backup_job_uuid, $qemu_support);
$task->{size} = $stat->{total};
};
my $err = $@;
if ($err) {
$self->logerr($err);
$self->mon_backup_cancel($vmid);
$self->resume_vm_after_job_start($task, $vmid);
}
$self->restore_vm_power_state($vmid);
die $err if $err;
}
my $fork_compressor_pipe = sub {
my ($self, $comp, $outfileno) = @_;
my @pipefd = POSIX::pipe();
my $cpid = fork();
die "unable to fork worker - $!" if !defined($cpid) || $cpid < 0;
if ($cpid == 0) {
eval {
POSIX::close($pipefd[1]);
# redirect STDIN
my $fd = fileno(STDIN);
close STDIN;
POSIX::close(0) if $fd != 0;
die "unable to redirect STDIN - $!"
if !open(STDIN, "<&", $pipefd[0]);
# redirect STDOUT
$fd = fileno(STDOUT);
close STDOUT;
POSIX::close (1) if $fd != 1;
die "unable to redirect STDOUT - $!"
if !open(STDOUT, ">&", $outfileno);
exec($comp);
die "fork compressor '$comp' failed\n";
};
if (my $err = $@) {
$self->logerr($err);
POSIX::_exit(1);
}
POSIX::_exit(0);
kill(-9, $$);
} else {
POSIX::close($pipefd[0]);
$outfileno = $pipefd[1];
}
return ($cpid, $outfileno);
};
sub archive_vma {
my ($self, $task, $vmid, $filename, $comp) = @_;
my $conffile = "$task->{tmpdir}/qemu-server.conf";
my $firewall = "$task->{tmpdir}/qemu-server.fw";
my $opts = $self->{vzdump}->{opts};
my $starttime = time();
my $speed = 0;
if ($opts->{bwlimit}) {
$speed = $opts->{bwlimit}*1024;
}
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
my $diskcount = scalar(@{$task->{disks}});
if (!$diskcount) {
$self->loginfo("backup contains no disks");
my $outcmd;
if ($comp) {
$outcmd = "exec:$comp";
} else {
$outcmd = "exec:cat";
}
$outcmd .= " > $filename" if !$opts->{stdout};
my $cmd = ['/usr/bin/vma', 'create', '-v', '-c', $conffile];
push @$cmd, '-c', $firewall if -e $firewall;
push @$cmd, $outcmd;
$self->loginfo("starting diskless backup");
$self->loginfo(join(' ', @$cmd));
if ($opts->{stdout}) {
$self->cmd($cmd, output => ">&" . fileno($opts->{stdout}));
} else {
$self->cmd($cmd);
}
return;
}
my $devlist = _get_task_devlist($task);
my $cpid;
my $backup_job_uuid;
eval {
$SIG{INT} = $SIG{TERM} = $SIG{QUIT} = $SIG{HUP} = $SIG{PIPE} = sub {
die "interrupted by signal\n";
};
$self->enforce_vm_running_for_backup($vmid);
$self->{qmeventd_fh} = PVE::QemuServer::register_qmeventd_handle($vmid);
# Currently, failing to determine Proxmox support is not critical here, because it's only
# used for performance settings like 'max-workers'.
my $qemu_support = eval { mon_cmd($vmid, "query-proxmox-support") };
log_warn($@) if $@;
$attach_tpmstate_drive->($self, $task, $vmid);
$task->{'use-fleecing'} = check_and_prepare_fleecing(
$self, $vmid, $opts->{fleecing}, $task->{disks}, $is_template, $qemu_support, 0);
my $outfh;
if ($opts->{stdout}) {
$outfh = $opts->{stdout};
} else {
$outfh = IO::File->new($filename, "w") ||
die "unable to open file '$filename' - $!\n";
}
my $outfileno = fileno($outfh);
if ($comp) {
($cpid, $outfileno) = $fork_compressor_pipe->($self, $comp, $outfileno);
}
my $qmpclient = PVE::QMPClient->new();
my $backup_cb = sub {
my ($vmid, $resp) = @_;
$backup_job_uuid = $resp->{return}->{UUID};
};
my $add_fd_cb = sub {
my ($vmid, $resp) = @_;
my $params = {
'backup-file' => "/dev/fdname/backup",
speed => $speed,
'config-file' => $conffile,
devlist => $devlist
};
$params->{'firewall-file'} = $firewall if -e $firewall;
$params->{fleecing} = JSON::true if $task->{'use-fleecing'};
add_backup_performance_options($params, $opts->{performance}, $qemu_support);
$qmpclient->queue_cmd($vmid, $backup_cb, 'backup', %$params);
};
$qmpclient->queue_cmd($vmid, $add_fd_cb, 'getfd', fd => $outfileno, fdname => "backup");
my $fs_frozen = $self->qga_fs_freeze($task, $vmid);
eval { $qmpclient->queue_execute(30) };
my $qmperr = $@;
if ($fs_frozen) {
$self->qga_fs_thaw($vmid);
}
die $qmperr if $qmperr;
die $qmpclient->{errors}->{$vmid} if $qmpclient->{errors}->{$vmid};
if ($cpid) {
POSIX::close($outfileno) == 0 ||
die "close output file handle failed\n";
}
die "got no uuid for backup task\n" if !defined($backup_job_uuid);
$self->loginfo("started backup task '$backup_job_uuid'");
$self->resume_vm_after_job_start($task, $vmid);
$query_backup_status_loop->($self, $vmid, $backup_job_uuid);
};
my $err = $@;
if ($err) {
$self->logerr($err);
$self->mon_backup_cancel($vmid);
$self->resume_vm_after_job_start($task, $vmid);
}
$self->restore_vm_power_state($vmid);
if ($err) {
if ($cpid) {
kill(9, $cpid);
waitpid($cpid, 0);
}
die $err;
}
if ($cpid && (waitpid($cpid, 0) > 0)) {
my $stat = $?;
my $ec = $stat >> 8;
my $signal = $stat & 127;
if ($ec || $signal) {
die "$comp failed - wrong exit status $ec" .
($signal ? " (signal $signal)\n" : "\n");
}
}
}
sub _get_task_devlist {
my ($task) = @_;
my $devlist = '';
foreach my $di (@{$task->{disks}}) {
if ($di->{type} eq 'block' || $di->{type} eq 'file') {
$devlist .= ',' if $devlist;
$devlist .= $di->{qmdevice};
} else {
die "implement me (type '$di->{type}')";
}
}
return $devlist;
}
sub qga_fs_freeze {
my ($self, $task, $vmid) = @_;
return if !$self->{vmlist}->{$vmid}->{agent} || $task->{mode} eq 'stop' || !$self->{vm_was_running} || $self->{vm_was_paused};
if (!PVE::QemuServer::qga_check_running($vmid, 1)) {
$self->loginfo("skipping guest-agent 'fs-freeze', agent configured but not running?");
return;
}
my $freeze = PVE::QemuServer::get_qga_key($self->{vmlist}->{$vmid}, 'freeze-fs-on-backup') // 1;
if (!$freeze) {
$self->loginfo("skipping guest-agent 'fs-freeze', disabled in VM options");
return;
}
$self->loginfo("issuing guest-agent 'fs-freeze' command");
eval { mon_cmd($vmid, "guest-fsfreeze-freeze") };
$self->logerr($@) if $@;
return 1; # even on mon command error, ensure we always thaw again
}
# only call if fs_freeze return 1
sub qga_fs_thaw {
my ($self, $vmid) = @_;
$self->loginfo("issuing guest-agent 'fs-thaw' command");
eval { mon_cmd($vmid, "guest-fsfreeze-thaw") };
$self->logerr($@) if $@;
}
# The size for fleecing images needs to be exactly the same size as QEMU sees. E.g. EFI disk can bex
# attached with a smaller size then the underyling image on the storage.
sub query_block_node_sizes {
my ($self, $vmid, $disks) = @_;
my $block_info = mon_cmd($vmid, "query-block");
$block_info = { map { $_->{device} => $_ } $block_info->@* };
for my $diskinfo ($disks->@*) {
my $drive_key = $diskinfo->{virtdev};
$drive_key .= "-backup" if $drive_key eq 'tpmstate0';
my $block_node_size =
eval { $block_info->{"drive-$drive_key"}->{inserted}->{image}->{'virtual-size'}; };
if (!$block_node_size) {
$self->loginfo(
"could not determine block node size of drive '$drive_key' - using fallback");
$block_node_size = $diskinfo->{size}
or die "could not determine size of drive '$drive_key'\n";
}
$diskinfo->{'block-node-size'} = $block_node_size;
}
return;
}
# we need a running QEMU/KVM process for backup, starts a paused (prelaunch)
# one if VM isn't already running
sub enforce_vm_running_for_backup {
my ($self, $vmid) = @_;
if (PVE::QemuServer::check_running($vmid)) {
$self->{vm_was_running} = 1;
return;
}
eval {
$self->loginfo("starting kvm to execute backup task");
# start with skiplock
my $params = {
skiplock => 1,
skiptemplate => 1,
paused => 1,
};
PVE::QemuServer::vm_start($self->{storecfg}, $vmid, $params);
};
die $@ if $@;
}
# resume VM again once in a clear state (stop mode backup of running VM)
sub resume_vm_after_job_start {
my ($self, $task, $vmid) = @_;
return if !$self->{vm_was_running} || $self->{vm_was_paused};
if (my $stoptime = $task->{vmstoptime}) {
my $delay = time() - $task->{vmstoptime};
$task->{vmstoptime} = undef; # avoid printing 'online after ..' twice
$self->loginfo("resuming VM again after $delay seconds");
} else {
$self->loginfo("resuming VM again");
}
mon_cmd($vmid, 'cont', timeout => 45);
}
# stop again if VM was not running before
sub restore_vm_power_state {
my ($self, $vmid) = @_;
# we always let VMs keep running
return if $self->{vm_was_running};
eval {
my $resp = mon_cmd($vmid, 'query-status');
my $status = $resp && $resp->{status} ? $resp->{status} : 'unknown';
if ($status eq 'prelaunch') {
$self->loginfo("stopping kvm after backup task");
PVE::QemuServer::vm_stop($self->{storecfg}, $vmid, 1);
} else {
$self->loginfo("kvm status changed after backup ('$status') - keep VM running");
}
};
warn $@ if $@;
}
sub mon_backup_cancel {
my ($self, $vmid) = @_;
$self->loginfo("aborting backup job");
eval { mon_cmd($vmid, 'backup-cancel') };
$self->logerr($@) if $@;
}
sub snapshot {
my ($self, $task, $vmid) = @_;
# nothing to do
}
my sub cleanup_file_handles {
my ($self, $file_handles) = @_;
for my $file_handle ($file_handles->@*) {
close($file_handle) or $self->log('warn', "unable to close file handle - $!");
}
}
my sub cleanup_nbd_mounts {
my ($self, $info) = @_;
for my $mount_point (keys $info->%*) {
my $pid_file = delete($info->{$mount_point}->{'pid-file'});
unlink($pid_file) or $self->log('warn', "unable to unlink '$pid_file' - $!");
# Do a lazy unmount, because the target might still be busy even if the file handle was
# already closed.
eval { run_command(['fusermount', '-z', '-u', $mount_point ]); };
if (my $err = $@) {
delete $info->{$mount_point};
$self->log('warn', "unable to unmount NBD backup source '$mount_point' - $err");
}
}
# Wait for the unmount before cleaning up child PIDs to avoid 'nbdfuse' processes being
# interrupted by the signals issued there.
my $waited;
my $wait_limit = 50; # 5 seconds
for ($waited = 0; $waited < $wait_limit && scalar(keys $info->%*); $waited++) {
for my $mount_point (keys $info->%*) {
delete($info->{$mount_point}) if !-e $info->{$mount_point}->{'virtual-file'};
eval { remove_tree($mount_point); };
}
usleep(100_000);
}
# just informational, remaining child processes will be killed afterwards
$self->loginfo("unable to gracefully cleanup NBD fuse mounts") if scalar(keys $info->%*) != 0;
}
my sub cleanup_child_processes {
my ($self, $cpids) = @_;
my $waited;
my $wait_limit = 5;
for ($waited = 0; $waited < $wait_limit && scalar(keys $cpids->%*); $waited++) {
for my $cpid (keys $cpids->%*) {
delete($cpids->{$cpid}) if waitpid($cpid, POSIX::WNOHANG) > 0;
}
if ($waited == 0) {
kill 15, $_ for keys $cpids->%*;
}
sleep 1;
}
if ($waited == $wait_limit && scalar(keys $cpids->%*)) {
kill 9, $_ for keys $cpids->%*;
sleep 1;
for my $cpid (keys $cpids->%*) {
delete($cpids->{$cpid}) if waitpid($cpid, POSIX::WNOHANG) > 0;
}
$self->log('warn', "unable to collect child process '$_'") for keys $cpids->%*;
}
}
sub cleanup {
my ($self, $task, $vmid) = @_;
# If VM was started only for backup, it is already stopped now.
if (PVE::QemuServer::Helpers::vm_running_locally($vmid)) {
if ($task->{cleanup}->{'nbd-stop'}) {
eval { PVE::QemuServer::QMPHelpers::nbd_stop($vmid); };
$self->logerr($@) if $@;
}
if (my $info = $task->{cleanup}->{'backup-access-teardown'}) {
my $params = {
'target-id' => $info->{'target-id'},
timeout => 60,
success => $info->{success} ? JSON::true : JSON::false,
};
$self->loginfo("tearing down backup-access");
eval { mon_cmd($vmid, "backup-access-teardown", $params->%*) };
$self->logerr($@) if $@;
}
$detach_tpmstate_drive->($task, $vmid);
}
if ($task->{'use-fleecing'}) {
eval {
detach_fleecing_images($task->{disks}, $vmid);
PVE::QemuConfig::cleanup_fleecing_images(
$vmid, $self->{storecfg}, sub { $self->log($_[0], $_[1]); });
};
$self->log('warn', "attempt to clean up fleecing images failed - $@") if $@;
}
if ($self->{qmeventd_fh}) {
close($self->{qmeventd_fh});
}
cleanup_file_handles($self, $task->{cleanup}->{'file-handles'})
if $task->{cleanup}->{'file-handles'};
cleanup_nbd_mounts($self, $task->{cleanup}->{'nbd-mounts'})
if $task->{cleanup}->{'nbd-mounts'};
cleanup_child_processes($self, $task->{cleanup}->{'child-pids'})
if $task->{cleanup}->{'child-pids'};
if (my $dir = $task->{'backup-access-root-dir'}) {
eval { remove_tree($dir) };
$self->log('warn', "unable to cleanup directory $dir - $@") if $@;
}
}
my sub virtual_file_backup_prepare {
my ($self, $vmid, $task, $device_name, $size, $nbd_path, $bitmap_name) = @_;
my $cleanup = $task->{cleanup};
my $nbd_uri = "nbd+unix:///${device_name}?socket=${nbd_path}";
my $error_fh;
my $next_dirty_region;
# If there is no dirty bitmap, it can be treated as if there's a full dirty one. The output of
# nbdinfo is a list of tuples with offset, length, type, description. The first bit of 'type' is
# set when the bitmap is dirty, see QEMU's docs/interop/nbd.txt
my $dirty_bitmap = [];
if ($bitmap_name) {
my $input = IO::File->new();
my $info = IO::File->new();
$error_fh = IO::File->new();
my $nbdinfo_cmd = ["nbdinfo", $nbd_uri, "--map=qemu:dirty-bitmap:${bitmap_name}"];
my $cpid = open3($input, $info, $error_fh, $nbdinfo_cmd->@*)
or die "failed to spawn nbdinfo child - $!\n";
$cleanup->{'child-pids'}->{$cpid} = 1;
$next_dirty_region = sub {
my ($offset, $length, $type);
do {
my $line = <$info>;
return if !$line;
die "unexpected output from nbdinfo - $line\n"
if $line !~ m/^\s*(\d+)\s*(\d+)\s*(\d+)/; # also untaints
($offset, $length, $type) = ($1, $2, $3);
} while (($type & 0x1) == 0); # not dirty
return ($offset, $length);
};
} else {
my $done = 0;
$next_dirty_region = sub {
return if $done;
$done = 1;
return (0, $size);
};
}
my $mount_point = $task->{'backup-access-root-dir'}
."/${vmid}-nbd.backup-access.${device_name}.$$";
make_path($mount_point) or die "unable to create directory $mount_point\n";
$cleanup->{'nbd-mounts'}->{$mount_point} = {};
# Note that nbdfuse requires "$dir/$file". A single name would be treated as a dir and the file
# would be named "$dir/nbd" then
my $virtual_file = "${mount_point}/${device_name}";
$cleanup->{'nbd-mounts'}->{$mount_point}->{'virtual-file'} = $virtual_file;
my $pid_file = "${mount_point}.pid";
PVE::Tools::file_set_contents($pid_file, '', 0600);
$cleanup->{'nbd-mounts'}->{$mount_point}->{'pid-file'} = $pid_file;
my $cpid = fork() // die "fork failed: $!\n";
if (!$cpid) {
# By default, access will be restricted to the current user, because the allow_other fuse
# mount option is not used.
eval {
run_command(
["nbdfuse", '--pidfile', $pid_file, $virtual_file, $nbd_uri],
logfunc => sub { $self->loginfo("nbdfuse '$virtual_file': $_[0]") },
);
};
if (my $err = $@) {
eval { $self->loginfo($err); };
POSIX::_exit(1);
}
POSIX::_exit(0);
}
$cleanup->{'child-pids'}->{$cpid} = 1;
my ($virtual_file_ready, $waited) = (0, 0);
while (!$virtual_file_ready && $waited < 30) { # 3 seconds
my $pid = PVE::Tools::file_read_firstline($pid_file);
if ($pid) {
$virtual_file_ready = 1;
} else {
usleep(100_000);
$waited++;
}
}
die "timeout setting up virtual file '$virtual_file'" if !$virtual_file_ready;
$self->loginfo("provided NBD export as a virtual file '$virtual_file'");
# NOTE O_DIRECT, because each block should be read exactly once and also because fuse will try
# to read ahead otherwise, which would produce warning messages if the next block is not
# mapped/allocated for the NBD export in case of incremental backup. Open as writable to support
# discard.
my $fh = IO::File->new($virtual_file, O_RDWR | O_DIRECT)
or die "unable to open backup source '$virtual_file' - $!\n";
push $cleanup->{'file-handles'}->@*, $fh;
return ($fh, $next_dirty_region);
}
my sub backup_access_to_volume_info {
my ($self, $vmid, $task, $backup_access_info, $mechanism, $nbd_path) = @_;
my $bitmap_action_to_status = {
'not-used' => 'none',
'not-used-removed' => 'none',
'new' => 'new',
'used' => 'reuse',
'invalid' => 'new',
'missing-recreated' => 'new',
};
my $volumes = {};
for my $info ($backup_access_info->@*) {
my $bitmap_status = 'none';
my $bitmap_name;
if (my $bitmap_action = $info->{'bitmap-action'}) {
$bitmap_status = $bitmap_action_to_status->{$bitmap_action}
or die "got unexpected bitmap action '$bitmap_action'\n";
$bitmap_name = $info->{'bitmap-name'} or die "bitmap-name is not present\n";
}
my ($device, $size) = $info->@{qw(device size)};
$volumes->{$device}->{'bitmap-mode'} = $bitmap_status;
$volumes->{$device}->{size} = $size;
if ($mechanism eq 'file-handle') {
my ($fh, $next_dirty_region) = virtual_file_backup_prepare(
$self, $vmid, $task, $device, $size, $nbd_path, $bitmap_name);
$volumes->{$device}->{'file-handle'} = $fh;
$volumes->{$device}->{'next-dirty-region'} = $next_dirty_region;
} elsif ($mechanism eq 'nbd') {
$volumes->{$device}->{'nbd-path'} = $nbd_path;
$volumes->{$device}->{'bitmap-name'} = $bitmap_name;
} else {
die "internal error - unkown mechanism '$mechanism'";
}
}
return $volumes;
}
sub archive_external {
my ($self, $task, $vmid) = @_;
$task->{'backup-access-root-dir'} = "/run/qemu-server/${vmid}.backup-access.$$/";
make_path($task->{'backup-access-root-dir'})
or die "unable to create directory $task->{'backup-access-root-dir'}\n";
chmod(0700, $task->{'backup-access-root-dir'})
or die "unable to chmod directory $task->{'backup-access-root-dir'}\n";
my $guest_config = PVE::Tools::file_get_contents("$task->{tmpdir}/qemu-server.conf");
my $firewall_file = "$task->{tmpdir}/qemu-server.fw";
my $opts = $self->{vzdump}->{opts};
my $backup_provider = $self->{vzdump}->{'backup-provider'};
$self->loginfo("starting external backup via " . $backup_provider->provider_name());
my $starttime = time();
$self->enforce_vm_running_for_backup($vmid);
$self->{qmeventd_fh} = PVE::QemuServer::register_qmeventd_handle($vmid);
eval {
$SIG{INT} = $SIG{TERM} = $SIG{QUIT} = $SIG{HUP} = $SIG{PIPE} = sub {
die "interrupted by signal\n";
};
my $qemu_support = mon_cmd($vmid, "query-proxmox-support");
if (!$qemu_support->{'backup-access-api'}) {
die "backups access API required for external provider backup is not supported by"
." the running QEMU version. Please make sure you've installed the latest "
." version and the VM has been restarted.\n";
}
$attach_tpmstate_drive->($self, $task, $vmid);
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
my $fleecing = check_and_prepare_fleecing(
$self, $vmid, $opts->{fleecing}, $task->{disks}, $is_template, $qemu_support, 1);
die "cannot setup backup access without fleecing\n" if !$fleecing;
$task->{'use-fleecing'} = 1;
my $target_id = "snapshot-access:$opts->{storage}";
my $mechanism = $backup_provider->backup_get_mechanism($vmid, 'qemu');
die "mechanism '$mechanism' requested by backup provider is not supported for VMs\n"
if $mechanism ne 'file-handle' && $mechanism ne 'nbd';
$self->loginfo("using backup mechanism '$mechanism'");
if ($mechanism eq 'file-handle') {
# For mechanism 'file-handle', the nbdfuse binary is required. Also, the bitmap needs
# to be passed to the provider. The bitmap cannot be dumped via QMP and doing it via
# qemu-img is experimental, so use nbdinfo. Both are in libnbd-bin.
die "need 'nbdfuse' binary from package libnbd-bin\n" if !-e "/usr/bin/nbdfuse";
}
my $devices = {};
for my $di ($task->{disks}->@*) {
my $device_name = $di->{qmdevice};
die "implement me (type '$di->{type}')"
if $di->{type} ne 'block' && $di->{type} ne 'file';
$devices->{$device_name}->{size} = $di->{'block-node-size'};
}
my $incremental_info = $backup_provider->backup_vm_query_incremental($vmid, $devices);
my $qmp_devices = [];
for my $device (sort keys $devices->%*) {
my $qmp_device = { device => $device };
if (defined(my $mode = $incremental_info->{$device})) {
if ($mode eq 'new' || $mode eq 'use' || $mode eq 'none') {
$qmp_device->{'bitmap-mode'} = $mode;
} else {
die "invalid incremental mode '$mode' returned by backup provider plugin\n";
}
}
push($qmp_devices->@*, $qmp_device);
}
my $params = {
'target-id' => $target_id,
devices => $qmp_devices,
timeout => 60,
};
my $fs_frozen = $self->qga_fs_freeze($task, $vmid);
$self->loginfo("setting up snapshot-access for backup");
$task->{cleanup}->{'backup-access-teardown'} = { 'target-id' => $target_id, success => 0 };
my $backup_access_info = eval { mon_cmd($vmid, "backup-access-setup", $params->%*) };
my $qmperr = $@;
if ($fs_frozen) {
$self->qga_fs_thaw($vmid);
}
die $qmperr if $qmperr;
$self->resume_vm_after_job_start($task, $vmid);
my $bitmap_info = mon_cmd($vmid, 'query-pbs-bitmap-info');
for my $info (sort { $a->{drive} cmp $b->{drive} } $bitmap_info->@*) {
my $text = $bitmap_action_to_human->($self, $info);
my $drive = $info->{drive};
$drive =~ s/^drive-//; # for consistency
$self->loginfo("$drive: dirty-bitmap status: $text");
}
$self->loginfo("starting NBD server");
my $nbd_path = "$task->{'backup-access-root-dir'}/${vmid}-nbd.backup-access";
mon_cmd(
$vmid, "nbd-server-start", addr => { type => 'unix', data => { path => $nbd_path } } );
$task->{cleanup}->{'nbd-stop'} = 1;
for my $info ($backup_access_info->@*) {
$self->loginfo("adding NBD export for $info->{device}");
my $export_params = {
id => $info->{device},
'node-name' => $info->{'node-name'},
writable => JSON::true, # for discard
type => "nbd",
name => $info->{device}, # NBD export name
};
if ($info->{'bitmap-name'}) {
$export_params->{bitmaps} = [{
node => $info->{'bitmap-node-name'},
name => $info->{'bitmap-name'},
}];
}
mon_cmd($vmid, "block-export-add", $export_params->%*);
}
my $volumes = backup_access_to_volume_info(
$self, $vmid, $task, $backup_access_info, $mechanism, $nbd_path);
my $param = {};
$param->{'bandwidth-limit'} = $opts->{bwlimit} * 1024 if $opts->{bwlimit};
$param->{'firewall-config'} = PVE::Tools::file_get_contents($firewall_file)
if -e $firewall_file;
$backup_provider->backup_vm($vmid, $guest_config, $volumes, $param);
};
my $err = $@;
if ($err) {
$self->logerr($err);
$self->resume_vm_after_job_start($task, $vmid);
} else {
$task->{cleanup}->{'backup-access-teardown'}->{success} = 1;
}
$self->restore_vm_power_state($vmid);
die $err if $err;
}
1;