mirror of
https://git.proxmox.com/git/qemu-server
synced 2025-04-28 09:58:12 +00:00

The local one is specific for `allocate_fleecing_images` and has a comment stating to use the one from `PVE::QemuConfig` in all other cases. The `cleanup` sub already called this, but only if the VM was running. We do allocate fleecing images for previously-stopped VMs as well, though, so we also need to do the cleanup. As for the `detach_fleecing_images()` call: while could have stayed in the `vm_running_locall()` branch, it also performs this check and this way the entire fleecing cleanup stays together in one place. Signed-off-by: Wolfgang Bumiller <w.bumiller@proxmox.com>
1612 lines
48 KiB
Perl
1612 lines
48 KiB
Perl
package PVE::VZDump::QemuServer;
|
|
|
|
use strict;
|
|
use warnings;
|
|
|
|
use Fcntl qw(:mode);
|
|
use File::Basename;
|
|
use File::Path qw(make_path remove_tree);
|
|
use File::stat qw();
|
|
use IO::File;
|
|
use IPC::Open3;
|
|
use JSON;
|
|
use POSIX qw(EINTR EAGAIN);
|
|
use Time::HiRes qw(usleep);
|
|
|
|
use PVE::Cluster qw(cfs_read_file);
|
|
use PVE::INotify;
|
|
use PVE::IPCC;
|
|
use PVE::JSONSchema;
|
|
use PVE::PBSClient;
|
|
use PVE::RESTEnvironment qw(log_warn);
|
|
use PVE::QMPClient;
|
|
use PVE::Storage::Plugin;
|
|
use PVE::Storage::PBSPlugin;
|
|
use PVE::Storage;
|
|
use PVE::Tools qw(run_command);
|
|
use PVE::VZDump;
|
|
use PVE::Format qw(render_duration render_bytes);
|
|
|
|
use PVE::QemuConfig;
|
|
use PVE::QemuServer;
|
|
use PVE::QemuServer::Drive qw(checked_volume_format);
|
|
use PVE::QemuServer::Helpers;
|
|
use PVE::QemuServer::Machine;
|
|
use PVE::QemuServer::Monitor qw(mon_cmd);
|
|
use PVE::QemuServer::QMPHelpers;
|
|
|
|
use base qw (PVE::VZDump::Plugin);
|
|
|
|
sub new {
|
|
my ($class, $vzdump) = @_;
|
|
|
|
PVE::VZDump::check_bin('qm');
|
|
|
|
my $self = bless { vzdump => $vzdump }, $class;
|
|
|
|
$self->{vmlist} = PVE::QemuServer::vzlist();
|
|
$self->{storecfg} = PVE::Storage::config();
|
|
|
|
return $self;
|
|
};
|
|
|
|
sub type {
|
|
return 'qemu';
|
|
}
|
|
|
|
sub vmlist {
|
|
my ($self) = @_;
|
|
return [ keys %{$self->{vmlist}} ];
|
|
}
|
|
|
|
sub prepare {
|
|
my ($self, $task, $vmid, $mode) = @_;
|
|
|
|
my $running = PVE::QemuServer::Helpers::vm_running_locally($vmid);
|
|
|
|
if ($running && (my $status = mon_cmd($vmid, 'query-backup'))) {
|
|
if ($status->{status} && $status->{status} eq 'active') {
|
|
$self->log('warn', "left-over backup job still running inside QEMU - canceling now");
|
|
mon_cmd($vmid, 'backup-cancel');
|
|
}
|
|
}
|
|
|
|
$task->{disks} = [];
|
|
|
|
my $conf = $self->{vmlist}->{$vmid} = PVE::QemuConfig->load_config($vmid);
|
|
|
|
$self->loginfo("VM Name: $conf->{name}")
|
|
if defined($conf->{name});
|
|
|
|
$self->{vm_was_running} = $running ? 1 : 0;
|
|
$self->{vm_was_paused} = 0;
|
|
if ($running && PVE::QemuServer::vm_is_paused($vmid, 0)) {
|
|
# Do not treat a suspended VM as paused, as it would cause us to skip
|
|
# fs-freeze even if the VM wakes up before we reach qga_fs_freeze.
|
|
$self->{vm_was_paused} = 1;
|
|
}
|
|
|
|
$task->{hostname} = $conf->{name};
|
|
|
|
my $hostname = PVE::INotify::nodename();
|
|
|
|
my $vollist = [];
|
|
my $drivehash = {};
|
|
my $backup_volumes = PVE::QemuConfig->get_backup_volumes($conf);
|
|
|
|
foreach my $volume (@{$backup_volumes}) {
|
|
my $name = $volume->{key};
|
|
my $volume_config = $volume->{volume_config};
|
|
my $volid = $volume_config->{file};
|
|
|
|
if (!$volume->{included}) {
|
|
$self->loginfo("exclude disk '$name' '$volid' ($volume->{reason})");
|
|
next;
|
|
} else {
|
|
my $log = "include disk '$name' '$volid'";
|
|
if (defined(my $size = $volume_config->{size})) {
|
|
my $readable_size = PVE::JSONSchema::format_size($size);
|
|
$log .= " $readable_size";
|
|
}
|
|
$self->loginfo($log);
|
|
}
|
|
|
|
my ($storeid, $volname) = PVE::Storage::parse_volume_id($volid, 1);
|
|
push @$vollist, $volid if $storeid;
|
|
$drivehash->{$name} = $volume->{volume_config};
|
|
}
|
|
|
|
PVE::Storage::activate_volumes($self->{storecfg}, $vollist);
|
|
|
|
foreach my $ds (sort keys %$drivehash) {
|
|
my $drive = $drivehash->{$ds};
|
|
|
|
my $volid = $drive->{file};
|
|
my ($storeid, $volname) = PVE::Storage::parse_volume_id($volid, 1);
|
|
|
|
my $path = $volid;
|
|
if ($storeid) {
|
|
$path = PVE::Storage::path($self->{storecfg}, $volid);
|
|
}
|
|
next if !$path;
|
|
|
|
my ($size, $format);
|
|
if ($storeid) {
|
|
# The call in list context can be expensive for certain plugins like RBD, just get size
|
|
$size = eval { PVE::Storage::volume_size_info($self->{storecfg}, $volid, 5) };
|
|
die "cannot determine size of volume '$volid' - $@\n" if $@;
|
|
|
|
$format = checked_volume_format($self->{storecfg}, $volid);
|
|
} else {
|
|
($size, $format) = eval {
|
|
PVE::Storage::volume_size_info($self->{storecfg}, $volid, 5);
|
|
};
|
|
die "cannot determine size and format of volume '$volid' - $@\n" if $@;
|
|
}
|
|
|
|
my $diskinfo = {
|
|
path => $path,
|
|
volid => $volid,
|
|
storeid => $storeid,
|
|
size => $size,
|
|
format => $format,
|
|
virtdev => $ds,
|
|
qmdevice => "drive-$ds",
|
|
};
|
|
|
|
if ($ds eq 'tpmstate0') {
|
|
# TPM drive only exists for backup, which is reflected in the name
|
|
$diskinfo->{qmdevice} = 'drive-tpmstate0-backup';
|
|
$task->{tpmpath} = $path;
|
|
}
|
|
|
|
if (-b $path) {
|
|
$diskinfo->{type} = 'block';
|
|
} else {
|
|
$diskinfo->{type} = 'file';
|
|
}
|
|
|
|
push @{$task->{disks}}, $diskinfo;
|
|
}
|
|
}
|
|
|
|
sub vm_status {
|
|
my ($self, $vmid) = @_;
|
|
|
|
my $running = PVE::QemuServer::check_running($vmid) ? 1 : 0;
|
|
|
|
return wantarray ? ($running, $running ? 'running' : 'stopped') : $running;
|
|
}
|
|
|
|
sub lock_vm {
|
|
my ($self, $vmid) = @_;
|
|
|
|
PVE::QemuConfig->set_lock($vmid, 'backup');
|
|
}
|
|
|
|
sub unlock_vm {
|
|
my ($self, $vmid) = @_;
|
|
|
|
PVE::QemuConfig->remove_lock($vmid, 'backup');
|
|
}
|
|
|
|
sub stop_vm {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
my $opts = $self->{vzdump}->{opts};
|
|
|
|
my $wait = $opts->{stopwait} * 60;
|
|
# send shutdown and wait
|
|
$self->cmd ("qm shutdown $vmid --skiplock --keepActive --timeout $wait");
|
|
}
|
|
|
|
sub start_vm {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
$self->cmd ("qm start $vmid --skiplock");
|
|
}
|
|
|
|
sub suspend_vm {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
return if $self->{vm_was_paused};
|
|
|
|
$self->cmd ("qm suspend $vmid --skiplock");
|
|
}
|
|
|
|
sub resume_vm {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
return if $self->{vm_was_paused};
|
|
|
|
$self->cmd ("qm resume $vmid --skiplock");
|
|
}
|
|
|
|
sub assemble {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
my $conffile = PVE::QemuConfig->config_file($vmid);
|
|
|
|
my $outfile = "$task->{tmpdir}/qemu-server.conf";
|
|
my $firewall_src = "/etc/pve/firewall/$vmid.fw";
|
|
my $firewall_dest = "$task->{tmpdir}/qemu-server.fw";
|
|
|
|
my $outfd = IO::File->new(">$outfile") or die "unable to open '$outfile' - $!\n";
|
|
my $conffd = IO::File->new($conffile, 'r') or die "unable to open '$conffile' - $!\n";
|
|
|
|
my $found_snapshot;
|
|
my $found_pending;
|
|
my $found_special;
|
|
while (defined (my $line = <$conffd>)) {
|
|
next if $line =~ m/^\#vzdump\#/; # just to be sure
|
|
next if $line =~ m/^\#qmdump\#/; # just to be sure
|
|
if ($line =~ m/^\[(.*)\]\s*$/) {
|
|
if ($1 =~ m/^PENDING$/i) {
|
|
$found_pending = 1;
|
|
} elsif ($1 =~ m/^special:.*$/) {
|
|
$found_special = 1;
|
|
} else {
|
|
$found_snapshot = 1;
|
|
}
|
|
}
|
|
# skip all snapshots, pending changes and special sections
|
|
next if $found_snapshot || $found_pending || $found_special;
|
|
|
|
if ($line =~ m/^unused\d+:\s*(\S+)\s*/) {
|
|
$self->loginfo("skip unused drive '$1' (not included into backup)");
|
|
next;
|
|
}
|
|
next if $line =~ m/^lock:/ || $line =~ m/^parent:/;
|
|
|
|
print $outfd $line;
|
|
}
|
|
|
|
foreach my $di (@{$task->{disks}}) {
|
|
if ($di->{type} eq 'block' || $di->{type} eq 'file') {
|
|
my $storeid = $di->{storeid} || '';
|
|
my $format = $di->{format} || '';
|
|
print $outfd "#qmdump#map:$di->{virtdev}:$di->{qmdevice}:$storeid:$format:\n";
|
|
} else {
|
|
die "internal error";
|
|
}
|
|
}
|
|
|
|
if ($found_special) {
|
|
$self->loginfo("special config section found (not included into backup)");
|
|
}
|
|
if ($found_snapshot) {
|
|
$self->loginfo("snapshots found (not included into backup)");
|
|
}
|
|
if ($found_pending) {
|
|
$self->loginfo("pending configuration changes found (not included into backup)");
|
|
}
|
|
|
|
PVE::Tools::file_copy($firewall_src, $firewall_dest) if -f $firewall_src;
|
|
}
|
|
|
|
sub archive {
|
|
my ($self, $task, $vmid, $filename, $comp) = @_;
|
|
|
|
my $opts = $self->{vzdump}->{opts};
|
|
my $scfg = $opts->{scfg};
|
|
|
|
if ($self->{vzdump}->{opts}->{pbs}) {
|
|
$self->archive_pbs($task, $vmid);
|
|
} elsif ($self->{vzdump}->{'backup-provider'}) {
|
|
$self->archive_external($task, $vmid);
|
|
} else {
|
|
$self->archive_vma($task, $vmid, $filename, $comp);
|
|
}
|
|
}
|
|
|
|
my $bitmap_action_to_human = sub {
|
|
my ($self, $info) = @_;
|
|
|
|
my $action = $info->{action};
|
|
|
|
if ($action eq "not-used") {
|
|
return "disabled (no support)";
|
|
} elsif ($action eq "not-used-removed") {
|
|
return "disabled (old bitmap cleared)";
|
|
} elsif ($action eq "new") {
|
|
return "created new";
|
|
} elsif ($action eq "used") {
|
|
if ($info->{dirty} == 0) {
|
|
return "OK (drive clean)";
|
|
} else {
|
|
my $size = render_bytes($info->{size}, 1);
|
|
my $dirty = render_bytes($info->{dirty}, 1);
|
|
return "OK ($dirty of $size dirty)";
|
|
}
|
|
} elsif ($action eq "invalid") {
|
|
return "existing bitmap was invalid and has been cleared";
|
|
} elsif ($action eq "missing-recreated") {
|
|
# Lie about the TPM state, because it is newly attached each time.
|
|
return "created new" if $info->{drive} eq 'drive-tpmstate0-backup';
|
|
return "expected bitmap was missing and has been recreated";
|
|
} else {
|
|
return "unknown";
|
|
}
|
|
};
|
|
|
|
my $query_backup_status_loop = sub {
|
|
my ($self, $vmid, $job_uuid, $qemu_support) = @_;
|
|
|
|
my $starttime = time ();
|
|
my $last_time = $starttime;
|
|
my ($last_percent, $last_total, $last_target, $last_zero, $last_transferred) = (-1, 0, 0, 0, 0);
|
|
my ($transferred, $reused);
|
|
|
|
my $get_mbps = sub {
|
|
my ($mb, $delta) = @_;
|
|
return "0 B/s" if $mb <= 0;
|
|
my $bw = int(($mb / $delta));
|
|
return render_bytes($bw, 1) . "/s";
|
|
};
|
|
|
|
my $target = 0;
|
|
my $last_reused = 0;
|
|
my $has_query_bitmap = $qemu_support && $qemu_support->{'query-bitmap-info'};
|
|
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
|
|
if ($has_query_bitmap) {
|
|
my $total = 0;
|
|
my $bitmap_info = mon_cmd($vmid, 'query-pbs-bitmap-info');
|
|
for my $info (sort { $a->{drive} cmp $b->{drive} } @$bitmap_info) {
|
|
if (!$is_template) {
|
|
my $text = $bitmap_action_to_human->($self, $info);
|
|
my $drive = $info->{drive};
|
|
$drive =~ s/^drive-//; # for consistency
|
|
$self->loginfo("$drive: dirty-bitmap status: $text");
|
|
}
|
|
$target += $info->{dirty};
|
|
$total += $info->{size};
|
|
$last_reused += $info->{size} - $info->{dirty};
|
|
}
|
|
if ($target < $total) {
|
|
my $total_h = render_bytes($total, 1);
|
|
my $target_h = render_bytes($target, 1);
|
|
$self->loginfo("using fast incremental mode (dirty-bitmap), $target_h dirty of $total_h total");
|
|
}
|
|
}
|
|
|
|
my $last_finishing = 0;
|
|
while(1) {
|
|
my $status = mon_cmd($vmid, 'query-backup');
|
|
|
|
my $total = $status->{total} || 0;
|
|
my $dirty = $status->{dirty};
|
|
$target = (defined($dirty) && $dirty < $total) ? $dirty : $total if !$has_query_bitmap;
|
|
$transferred = $status->{transferred} || 0;
|
|
$reused = $status->{reused};
|
|
my $percent = $target ? int(($transferred * 100)/$target) : 100;
|
|
my $zero = $status->{'zero-bytes'} || 0;
|
|
|
|
die "got unexpected uuid\n" if !$status->{uuid} || ($status->{uuid} ne $job_uuid);
|
|
|
|
my $ctime = time();
|
|
my $duration = $ctime - $starttime;
|
|
|
|
my $rbytes = $transferred - $last_transferred;
|
|
my $wbytes;
|
|
if ($reused) {
|
|
# reused includes zero bytes for PBS
|
|
$wbytes = $rbytes - ($reused - $last_reused);
|
|
} else {
|
|
$wbytes = $rbytes - ($zero - $last_zero);
|
|
}
|
|
|
|
my $timediff = ($ctime - $last_time) || 1; # fixme
|
|
my $mbps_read = $get_mbps->($rbytes, $timediff);
|
|
my $mbps_write = $get_mbps->($wbytes, $timediff);
|
|
my $target_h = render_bytes($target, 1);
|
|
my $transferred_h = render_bytes($transferred, 1);
|
|
|
|
my $statusline = sprintf("%3d%% ($transferred_h of $target_h) in %s"
|
|
.", read: $mbps_read, write: $mbps_write", $percent, render_duration($duration));
|
|
|
|
my $res = $status->{status} || 'unknown';
|
|
if ($res ne 'active') {
|
|
if ($last_percent < 100) {
|
|
$self->loginfo($statusline);
|
|
}
|
|
if ($res ne 'done') {
|
|
die (($status->{errmsg} || "unknown error") . "\n") if $res eq 'error';
|
|
die "got unexpected status '$res'\n";
|
|
}
|
|
$last_target = $target if $target;
|
|
$last_total = $total if $total;
|
|
$last_zero = $zero if $zero;
|
|
$last_transferred = $transferred if $transferred;
|
|
last;
|
|
}
|
|
if ($percent != $last_percent && ($timediff > 2)) {
|
|
$self->loginfo($statusline);
|
|
$last_percent = $percent;
|
|
$last_target = $target if $target;
|
|
$last_total = $total if $total;
|
|
$last_zero = $zero if $zero;
|
|
$last_transferred = $transferred if $transferred;
|
|
$last_time = $ctime;
|
|
$last_reused = $reused;
|
|
|
|
if (!$last_finishing && $status->{finishing}) {
|
|
$self->loginfo("Waiting for server to finish backup validation...");
|
|
}
|
|
$last_finishing = $status->{finishing};
|
|
}
|
|
sleep(1);
|
|
}
|
|
|
|
my $duration = time() - $starttime;
|
|
|
|
if ($last_zero) {
|
|
my $zero_per = $last_target ? int(($last_zero * 100)/$last_target) : 0;
|
|
my $zero_h = render_bytes($last_zero);
|
|
$self->loginfo("backup is sparse: $zero_h (${zero_per}%) total zero data");
|
|
}
|
|
if ($reused) {
|
|
my $reused_h = render_bytes($reused);
|
|
my $reuse_per = int($reused * 100 / $last_total);
|
|
$self->loginfo("backup was done incrementally, reused $reused_h (${reuse_per}%)");
|
|
}
|
|
if ($transferred) {
|
|
my $transferred_h = render_bytes($transferred);
|
|
if ($duration) {
|
|
my $mbps = $get_mbps->($transferred, $duration);
|
|
$self->loginfo("transferred $transferred_h in $duration seconds ($mbps)");
|
|
} else {
|
|
$self->loginfo("transferred $transferred_h in <1 seconds");
|
|
}
|
|
}
|
|
|
|
return {
|
|
total => $last_total,
|
|
reused => $reused,
|
|
};
|
|
};
|
|
|
|
my $attach_tpmstate_drive = sub {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
return if !$task->{tpmpath};
|
|
|
|
# unconditionally try to remove the tpmstate-named drive - it only exists
|
|
# for backing up, and avoids errors if left over from some previous event
|
|
eval { PVE::QemuServer::qemu_drivedel($vmid, "tpmstate0-backup"); };
|
|
|
|
$self->loginfo('attaching TPM drive to QEMU for backup');
|
|
|
|
my $drive = "file=$task->{tpmpath},if=none,read-only=on,id=drive-tpmstate0-backup";
|
|
$drive =~ s/\\/\\\\/g;
|
|
my $ret = PVE::QemuServer::Monitor::hmp_cmd($vmid, "drive_add auto \"$drive\"", 60);
|
|
die "attaching TPM drive failed - $ret\n" if $ret !~ m/OK/s;
|
|
};
|
|
|
|
my $detach_tpmstate_drive = sub {
|
|
my ($task, $vmid) = @_;
|
|
return if !$task->{tpmpath} || !PVE::QemuServer::check_running($vmid);
|
|
eval { PVE::QemuServer::qemu_drivedel($vmid, "tpmstate0-backup"); };
|
|
};
|
|
|
|
my sub add_backup_performance_options {
|
|
my ($qmp_param, $perf, $qemu_support) = @_;
|
|
|
|
return if !$perf || scalar(keys $perf->%*) == 0;
|
|
|
|
if (!$qemu_support) {
|
|
my $settings_string = join(', ', sort keys $perf->%*);
|
|
log_warn("ignoring setting(s): $settings_string - issue checking if supported");
|
|
return;
|
|
}
|
|
|
|
if (defined($perf->{'max-workers'})) {
|
|
if ($qemu_support->{'backup-max-workers'}) {
|
|
$qmp_param->{'max-workers'} = int($perf->{'max-workers'});
|
|
} else {
|
|
log_warn("ignoring 'max-workers' setting - not supported by running QEMU");
|
|
}
|
|
}
|
|
}
|
|
|
|
sub get_and_check_pbs_encryption_config {
|
|
my ($self) = @_;
|
|
|
|
my $opts = $self->{vzdump}->{opts};
|
|
my $scfg = $opts->{scfg};
|
|
|
|
my $keyfile = PVE::Storage::PBSPlugin::pbs_encryption_key_file_name($scfg, $opts->{storage});
|
|
my $master_keyfile = PVE::Storage::PBSPlugin::pbs_master_pubkey_file_name($scfg, $opts->{storage});
|
|
|
|
if (-e $keyfile) {
|
|
if (-e $master_keyfile) {
|
|
$self->loginfo("enabling encryption with master key feature");
|
|
return ($keyfile, $master_keyfile);
|
|
} elsif ($scfg->{'master-pubkey'}) {
|
|
die "master public key configured but no key file found\n";
|
|
} else {
|
|
$self->loginfo("enabling encryption");
|
|
return ($keyfile, undef);
|
|
}
|
|
} else {
|
|
my $encryption_fp = $scfg->{'encryption-key'};
|
|
die "encryption configured ('$encryption_fp') but no encryption key file found!\n"
|
|
if $encryption_fp;
|
|
if (-e $master_keyfile) {
|
|
$self->log(
|
|
'warn',
|
|
"backup target storage is configured with master-key, but no encryption key set!"
|
|
." Ignoring master key settings and creating unencrypted backup."
|
|
);
|
|
}
|
|
return (undef, undef);
|
|
}
|
|
die "internal error - unhandled case for getting & checking PBS encryption ($keyfile, $master_keyfile)!";
|
|
}
|
|
|
|
# Helper is intended to be called from allocate_fleecing_images() only. Otherwise, fleecing volids
|
|
# have already been recorded in the configuration and PVE::QemuConfig::cleanup_fleecing_images()
|
|
# should be used instead.
|
|
my sub cleanup_fleecing_images {
|
|
my ($self, $vmid, $disks) = @_;
|
|
|
|
my $failed = [];
|
|
|
|
for my $di ($disks->@*) {
|
|
if (my $volid = $di->{'fleece-volid'}) {
|
|
eval { PVE::Storage::vdisk_free($self->{storecfg}, $volid); };
|
|
if (my $err = $@) {
|
|
$self->log('warn', "error removing fleecing image '$volid' - $err");
|
|
push $failed->@*, $volid;
|
|
}
|
|
}
|
|
}
|
|
|
|
PVE::QemuConfig::record_fleecing_images($vmid, $failed);
|
|
}
|
|
|
|
my sub allocate_fleecing_images {
|
|
my ($self, $disks, $vmid, $fleecing_storeid, $format, $all_images) = @_;
|
|
|
|
die "internal error - no fleecing storage specified\n" if !$fleecing_storeid;
|
|
|
|
my $fleece_volids = [];
|
|
|
|
eval {
|
|
my $n = 0; # counter for fleecing image names
|
|
|
|
for my $di ($disks->@*) {
|
|
# EFI/TPM are usually too small to be worth it, but it's required for external providers
|
|
next if !$all_images && $di->{virtdev} =~ m/^(?:tpmstate|efidisk)\d$/;
|
|
if ($di->{type} eq 'block' || $di->{type} eq 'file') {
|
|
my $scfg = PVE::Storage::storage_config($self->{storecfg}, $fleecing_storeid);
|
|
my $name = "vm-$vmid-fleece-$n";
|
|
$name .= ".$format" if $scfg->{path};
|
|
|
|
my $size;
|
|
if ($format ne 'raw') {
|
|
# Since non-raw images cannot be attached with an explicit 'size' parameter to
|
|
# QEMU later, pass the exact size to the storage layer. This makes qcow2
|
|
# fleecing images work for non-1KiB-aligned source images.
|
|
$size = $di->{'block-node-size'}/1024;
|
|
} else {
|
|
$size = PVE::Tools::convert_size($di->{'block-node-size'}, 'b' => 'kb');
|
|
}
|
|
|
|
$di->{'fleece-volid'} = PVE::Storage::vdisk_alloc(
|
|
$self->{storecfg}, $fleecing_storeid, $vmid, $format, $name, $size);
|
|
|
|
push $fleece_volids->@*, $di->{'fleece-volid'};
|
|
|
|
$n++;
|
|
} else {
|
|
die "implement me (type '$di->{type}')";
|
|
}
|
|
}
|
|
};
|
|
if (my $err = $@) {
|
|
cleanup_fleecing_images($self, $vmid, $disks);
|
|
die $err;
|
|
}
|
|
|
|
PVE::QemuConfig::record_fleecing_images($vmid, $fleece_volids);
|
|
}
|
|
|
|
my sub detach_fleecing_images {
|
|
my ($disks, $vmid) = @_;
|
|
|
|
return if !PVE::QemuServer::Helpers::vm_running_locally($vmid);
|
|
|
|
for my $di ($disks->@*) {
|
|
if (my $volid = $di->{'fleece-volid'}) {
|
|
my $devid = "$di->{qmdevice}-fleecing";
|
|
$devid =~ s/^drive-//; # re-added by qemu_drivedel()
|
|
eval { PVE::QemuServer::qemu_drivedel($vmid, $devid) };
|
|
}
|
|
}
|
|
}
|
|
|
|
my sub attach_fleecing_images {
|
|
my ($self, $disks, $vmid, $format) = @_;
|
|
|
|
# unconditionally try to remove potential left-overs from a previous backup
|
|
detach_fleecing_images($disks, $vmid);
|
|
|
|
my $vollist = [ map { $_->{'fleece-volid'} } grep { $_->{'fleece-volid'} } $disks->@* ];
|
|
PVE::Storage::activate_volumes($self->{storecfg}, $vollist);
|
|
|
|
for my $di ($disks->@*) {
|
|
if (my $volid = $di->{'fleece-volid'}) {
|
|
$self->loginfo("$di->{qmdevice}: attaching fleecing image $volid to QEMU");
|
|
|
|
my $path = PVE::Storage::path($self->{storecfg}, $volid);
|
|
my $devid = "$di->{qmdevice}-fleecing";
|
|
my $drive = "file=$path,if=none,id=$devid,format=$format,discard=unmap";
|
|
# Specify size explicitly, to make it work if storage backend rounded up size for
|
|
# fleecing image when allocating.
|
|
$drive .= ",size=$di->{'block-node-size'}" if $format eq 'raw';
|
|
$drive =~ s/\\/\\\\/g;
|
|
my $ret = PVE::QemuServer::Monitor::hmp_cmd($vmid, "drive_add auto \"$drive\"", 60);
|
|
die "attaching fleecing image $volid failed - $ret\n" if $ret !~ m/OK/s;
|
|
}
|
|
}
|
|
}
|
|
|
|
my sub check_and_prepare_fleecing {
|
|
my ($self, $vmid, $fleecing_opts, $disks, $is_template, $qemu_support, $all_images) = @_;
|
|
|
|
# Even if the VM was started specifically for fleecing, it's possible that the VM is resumed and
|
|
# then starts doing IO. For VMs that are not resumed the fleecing images will just stay empty,
|
|
# so there is no big cost.
|
|
|
|
my $use_fleecing = $fleecing_opts && $fleecing_opts->{enabled} && !$is_template;
|
|
|
|
if ($use_fleecing && !$qemu_support->{'backup-fleecing'}) {
|
|
$self->log(
|
|
'warn',
|
|
"running QEMU version does not support backup fleecing - continuing without",
|
|
);
|
|
$use_fleecing = 0;
|
|
}
|
|
|
|
# clean up potential left-overs from a previous attempt
|
|
eval {
|
|
PVE::QemuConfig::cleanup_fleecing_images(
|
|
$vmid, $self->{storecfg}, sub { $self->log($_[0], $_[1]); });
|
|
};
|
|
$self->log('warn', "attempt to clean up left-over fleecing images failed - $@") if $@;
|
|
|
|
if ($use_fleecing) {
|
|
$self->query_block_node_sizes($vmid, $disks);
|
|
|
|
my ($default_format, $valid_formats) = PVE::Storage::storage_default_format(
|
|
$self->{storecfg}, $fleecing_opts->{storage});
|
|
my $format = scalar(grep { $_ eq 'qcow2' } $valid_formats->@*) ? 'qcow2' : 'raw';
|
|
|
|
allocate_fleecing_images(
|
|
$self, $disks, $vmid, $fleecing_opts->{storage}, $format, $all_images);
|
|
attach_fleecing_images($self, $disks, $vmid, $format);
|
|
}
|
|
|
|
return $use_fleecing;
|
|
}
|
|
|
|
sub archive_pbs {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
my $conffile = "$task->{tmpdir}/qemu-server.conf";
|
|
my $firewall = "$task->{tmpdir}/qemu-server.fw";
|
|
|
|
my $opts = $self->{vzdump}->{opts};
|
|
my $scfg = $opts->{scfg};
|
|
|
|
my $starttime = time();
|
|
|
|
my $fingerprint = $scfg->{fingerprint};
|
|
my $repo = PVE::PBSClient::get_repository($scfg);
|
|
my $password = PVE::Storage::PBSPlugin::pbs_get_password($scfg, $opts->{storage});
|
|
my ($keyfile, $master_keyfile) = $self->get_and_check_pbs_encryption_config();
|
|
|
|
my $diskcount = scalar(@{$task->{disks}});
|
|
# proxmox-backup-client can only handle raw files and block devs, so only use it (directly) for
|
|
# disk-less VMs
|
|
if (!$diskcount) {
|
|
$self->loginfo("backup contains no disks");
|
|
|
|
local $ENV{PBS_PASSWORD} = $password;
|
|
local $ENV{PBS_FINGERPRINT} = $fingerprint if defined($fingerprint);
|
|
my $cmd = [
|
|
'/usr/bin/proxmox-backup-client',
|
|
'backup',
|
|
'--repository', $repo,
|
|
'--backup-type', 'vm',
|
|
'--backup-id', "$vmid",
|
|
'--backup-time', $task->{backup_time},
|
|
];
|
|
if (defined(my $ns = $scfg->{namespace})) {
|
|
push @$cmd, '--ns', $ns;
|
|
}
|
|
if (defined($keyfile)) {
|
|
push @$cmd, '--keyfile', $keyfile;
|
|
push @$cmd, '--master-pubkey-file', $master_keyfile if defined($master_keyfile);
|
|
}
|
|
|
|
push @$cmd, "qemu-server.conf:$conffile";
|
|
push @$cmd, "fw.conf:$firewall" if -e $firewall;
|
|
|
|
$self->loginfo("starting diskless backup");
|
|
$self->loginfo(join(' ', @$cmd));
|
|
|
|
$self->cmd($cmd);
|
|
|
|
return;
|
|
}
|
|
|
|
# get list early so we die on unknown drive types before doing anything
|
|
my $devlist = _get_task_devlist($task);
|
|
|
|
my $backup_job_uuid;
|
|
eval {
|
|
$SIG{INT} = $SIG{TERM} = $SIG{QUIT} = $SIG{HUP} = $SIG{PIPE} = sub {
|
|
die "interrupted by signal\n";
|
|
};
|
|
|
|
$self->enforce_vm_running_for_backup($vmid);
|
|
$self->{qmeventd_fh} = PVE::QemuServer::register_qmeventd_handle($vmid);
|
|
|
|
my $qemu_support = eval { mon_cmd($vmid, "query-proxmox-support") };
|
|
my $err = $@;
|
|
if (!$qemu_support || $err) {
|
|
die "query-proxmox-support returned empty value\n" if !$err;
|
|
if ($err =~ m/The command query-proxmox-support has not been found/) {
|
|
die "PBS backups are not supported by the running QEMU version. Please make "
|
|
. "sure you've installed the latest version and the VM has been restarted.\n";
|
|
} else {
|
|
die "QMP command query-proxmox-support failed - $err\n";
|
|
}
|
|
}
|
|
|
|
# pve-qemu supports it since 5.2.0-1 (PVE 6.4), so safe to die since PVE 8
|
|
die "master key configured but running QEMU version does not support master keys\n"
|
|
if !$qemu_support->{'pbs-masterkey'} && defined($master_keyfile);
|
|
|
|
$attach_tpmstate_drive->($self, $task, $vmid);
|
|
|
|
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
|
|
|
|
$task->{'use-fleecing'} = check_and_prepare_fleecing(
|
|
$self, $vmid, $opts->{fleecing}, $task->{disks}, $is_template, $qemu_support, 0);
|
|
|
|
my $fs_frozen = $self->qga_fs_freeze($task, $vmid);
|
|
|
|
my $params = {
|
|
format => "pbs",
|
|
'backup-file' => $repo,
|
|
'backup-id' => "$vmid",
|
|
'backup-time' => $task->{backup_time},
|
|
password => $password,
|
|
devlist => $devlist,
|
|
'config-file' => $conffile,
|
|
};
|
|
$params->{fleecing} = JSON::true if $task->{'use-fleecing'};
|
|
|
|
if (defined(my $ns = $scfg->{namespace})) {
|
|
$params->{'backup-ns'} = $ns;
|
|
}
|
|
|
|
$params->{speed} = $opts->{bwlimit}*1024 if $opts->{bwlimit};
|
|
add_backup_performance_options($params, $opts->{performance}, $qemu_support);
|
|
|
|
$params->{fingerprint} = $fingerprint if defined($fingerprint);
|
|
$params->{'firewall-file'} = $firewall if -e $firewall;
|
|
|
|
$params->{encrypt} = defined($keyfile) ? JSON::true : JSON::false;
|
|
if (defined($keyfile)) {
|
|
$params->{keyfile} = $keyfile;
|
|
$params->{"master-keyfile"} = $master_keyfile if defined($master_keyfile);
|
|
}
|
|
|
|
$params->{'use-dirty-bitmap'} = JSON::true
|
|
if $qemu_support->{'pbs-dirty-bitmap'} && !$is_template;
|
|
|
|
$params->{timeout} = 125; # give some time to connect to the backup server
|
|
|
|
my $res = eval { mon_cmd($vmid, "backup", %$params) };
|
|
my $qmperr = $@;
|
|
$backup_job_uuid = $res->{UUID} if $res;
|
|
|
|
if ($fs_frozen) {
|
|
$self->qga_fs_thaw($vmid);
|
|
}
|
|
|
|
die $qmperr if $qmperr;
|
|
die "got no uuid for backup task\n" if !defined($backup_job_uuid);
|
|
|
|
$self->loginfo("started backup task '$backup_job_uuid'");
|
|
|
|
$self->resume_vm_after_job_start($task, $vmid);
|
|
|
|
my $stat = $query_backup_status_loop->($self, $vmid, $backup_job_uuid, $qemu_support);
|
|
$task->{size} = $stat->{total};
|
|
};
|
|
my $err = $@;
|
|
if ($err) {
|
|
$self->logerr($err);
|
|
$self->mon_backup_cancel($vmid);
|
|
$self->resume_vm_after_job_start($task, $vmid);
|
|
}
|
|
$self->restore_vm_power_state($vmid);
|
|
|
|
die $err if $err;
|
|
}
|
|
|
|
my $fork_compressor_pipe = sub {
|
|
my ($self, $comp, $outfileno) = @_;
|
|
|
|
my @pipefd = POSIX::pipe();
|
|
my $cpid = fork();
|
|
die "unable to fork worker - $!" if !defined($cpid) || $cpid < 0;
|
|
if ($cpid == 0) {
|
|
eval {
|
|
POSIX::close($pipefd[1]);
|
|
# redirect STDIN
|
|
my $fd = fileno(STDIN);
|
|
close STDIN;
|
|
POSIX::close(0) if $fd != 0;
|
|
die "unable to redirect STDIN - $!"
|
|
if !open(STDIN, "<&", $pipefd[0]);
|
|
|
|
# redirect STDOUT
|
|
$fd = fileno(STDOUT);
|
|
close STDOUT;
|
|
POSIX::close (1) if $fd != 1;
|
|
|
|
die "unable to redirect STDOUT - $!"
|
|
if !open(STDOUT, ">&", $outfileno);
|
|
|
|
exec($comp);
|
|
die "fork compressor '$comp' failed\n";
|
|
};
|
|
if (my $err = $@) {
|
|
$self->logerr($err);
|
|
POSIX::_exit(1);
|
|
}
|
|
POSIX::_exit(0);
|
|
kill(-9, $$);
|
|
} else {
|
|
POSIX::close($pipefd[0]);
|
|
$outfileno = $pipefd[1];
|
|
}
|
|
|
|
return ($cpid, $outfileno);
|
|
};
|
|
|
|
sub archive_vma {
|
|
my ($self, $task, $vmid, $filename, $comp) = @_;
|
|
|
|
my $conffile = "$task->{tmpdir}/qemu-server.conf";
|
|
my $firewall = "$task->{tmpdir}/qemu-server.fw";
|
|
|
|
my $opts = $self->{vzdump}->{opts};
|
|
|
|
my $starttime = time();
|
|
|
|
my $speed = 0;
|
|
if ($opts->{bwlimit}) {
|
|
$speed = $opts->{bwlimit}*1024;
|
|
}
|
|
|
|
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
|
|
|
|
my $diskcount = scalar(@{$task->{disks}});
|
|
if (!$diskcount) {
|
|
$self->loginfo("backup contains no disks");
|
|
|
|
my $outcmd;
|
|
if ($comp) {
|
|
$outcmd = "exec:$comp";
|
|
} else {
|
|
$outcmd = "exec:cat";
|
|
}
|
|
|
|
$outcmd .= " > $filename" if !$opts->{stdout};
|
|
|
|
my $cmd = ['/usr/bin/vma', 'create', '-v', '-c', $conffile];
|
|
push @$cmd, '-c', $firewall if -e $firewall;
|
|
push @$cmd, $outcmd;
|
|
|
|
$self->loginfo("starting diskless backup");
|
|
$self->loginfo(join(' ', @$cmd));
|
|
|
|
if ($opts->{stdout}) {
|
|
$self->cmd($cmd, output => ">&" . fileno($opts->{stdout}));
|
|
} else {
|
|
$self->cmd($cmd);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
my $devlist = _get_task_devlist($task);
|
|
|
|
my $cpid;
|
|
my $backup_job_uuid;
|
|
|
|
eval {
|
|
$SIG{INT} = $SIG{TERM} = $SIG{QUIT} = $SIG{HUP} = $SIG{PIPE} = sub {
|
|
die "interrupted by signal\n";
|
|
};
|
|
|
|
$self->enforce_vm_running_for_backup($vmid);
|
|
$self->{qmeventd_fh} = PVE::QemuServer::register_qmeventd_handle($vmid);
|
|
|
|
# Currently, failing to determine Proxmox support is not critical here, because it's only
|
|
# used for performance settings like 'max-workers'.
|
|
my $qemu_support = eval { mon_cmd($vmid, "query-proxmox-support") };
|
|
log_warn($@) if $@;
|
|
|
|
$attach_tpmstate_drive->($self, $task, $vmid);
|
|
|
|
$task->{'use-fleecing'} = check_and_prepare_fleecing(
|
|
$self, $vmid, $opts->{fleecing}, $task->{disks}, $is_template, $qemu_support, 0);
|
|
|
|
my $outfh;
|
|
if ($opts->{stdout}) {
|
|
$outfh = $opts->{stdout};
|
|
} else {
|
|
$outfh = IO::File->new($filename, "w") ||
|
|
die "unable to open file '$filename' - $!\n";
|
|
}
|
|
my $outfileno = fileno($outfh);
|
|
|
|
if ($comp) {
|
|
($cpid, $outfileno) = $fork_compressor_pipe->($self, $comp, $outfileno);
|
|
}
|
|
|
|
my $qmpclient = PVE::QMPClient->new();
|
|
my $backup_cb = sub {
|
|
my ($vmid, $resp) = @_;
|
|
$backup_job_uuid = $resp->{return}->{UUID};
|
|
};
|
|
my $add_fd_cb = sub {
|
|
my ($vmid, $resp) = @_;
|
|
|
|
my $params = {
|
|
'backup-file' => "/dev/fdname/backup",
|
|
speed => $speed,
|
|
'config-file' => $conffile,
|
|
devlist => $devlist
|
|
};
|
|
$params->{'firewall-file'} = $firewall if -e $firewall;
|
|
$params->{fleecing} = JSON::true if $task->{'use-fleecing'};
|
|
add_backup_performance_options($params, $opts->{performance}, $qemu_support);
|
|
|
|
$qmpclient->queue_cmd($vmid, $backup_cb, 'backup', %$params);
|
|
};
|
|
|
|
$qmpclient->queue_cmd($vmid, $add_fd_cb, 'getfd', fd => $outfileno, fdname => "backup");
|
|
|
|
my $fs_frozen = $self->qga_fs_freeze($task, $vmid);
|
|
|
|
eval { $qmpclient->queue_execute(30) };
|
|
my $qmperr = $@;
|
|
|
|
if ($fs_frozen) {
|
|
$self->qga_fs_thaw($vmid);
|
|
}
|
|
|
|
die $qmperr if $qmperr;
|
|
die $qmpclient->{errors}->{$vmid} if $qmpclient->{errors}->{$vmid};
|
|
|
|
if ($cpid) {
|
|
POSIX::close($outfileno) == 0 ||
|
|
die "close output file handle failed\n";
|
|
}
|
|
|
|
die "got no uuid for backup task\n" if !defined($backup_job_uuid);
|
|
|
|
$self->loginfo("started backup task '$backup_job_uuid'");
|
|
|
|
$self->resume_vm_after_job_start($task, $vmid);
|
|
|
|
$query_backup_status_loop->($self, $vmid, $backup_job_uuid);
|
|
};
|
|
my $err = $@;
|
|
if ($err) {
|
|
$self->logerr($err);
|
|
$self->mon_backup_cancel($vmid);
|
|
$self->resume_vm_after_job_start($task, $vmid);
|
|
}
|
|
|
|
$self->restore_vm_power_state($vmid);
|
|
|
|
if ($err) {
|
|
if ($cpid) {
|
|
kill(9, $cpid);
|
|
waitpid($cpid, 0);
|
|
}
|
|
die $err;
|
|
}
|
|
|
|
if ($cpid && (waitpid($cpid, 0) > 0)) {
|
|
my $stat = $?;
|
|
my $ec = $stat >> 8;
|
|
my $signal = $stat & 127;
|
|
if ($ec || $signal) {
|
|
die "$comp failed - wrong exit status $ec" .
|
|
($signal ? " (signal $signal)\n" : "\n");
|
|
}
|
|
}
|
|
}
|
|
|
|
sub _get_task_devlist {
|
|
my ($task) = @_;
|
|
|
|
my $devlist = '';
|
|
foreach my $di (@{$task->{disks}}) {
|
|
if ($di->{type} eq 'block' || $di->{type} eq 'file') {
|
|
$devlist .= ',' if $devlist;
|
|
$devlist .= $di->{qmdevice};
|
|
} else {
|
|
die "implement me (type '$di->{type}')";
|
|
}
|
|
}
|
|
return $devlist;
|
|
}
|
|
|
|
sub qga_fs_freeze {
|
|
my ($self, $task, $vmid) = @_;
|
|
return if !$self->{vmlist}->{$vmid}->{agent} || $task->{mode} eq 'stop' || !$self->{vm_was_running} || $self->{vm_was_paused};
|
|
|
|
if (!PVE::QemuServer::qga_check_running($vmid, 1)) {
|
|
$self->loginfo("skipping guest-agent 'fs-freeze', agent configured but not running?");
|
|
return;
|
|
}
|
|
|
|
my $freeze = PVE::QemuServer::get_qga_key($self->{vmlist}->{$vmid}, 'freeze-fs-on-backup') // 1;
|
|
if (!$freeze) {
|
|
$self->loginfo("skipping guest-agent 'fs-freeze', disabled in VM options");
|
|
return;
|
|
}
|
|
|
|
$self->loginfo("issuing guest-agent 'fs-freeze' command");
|
|
eval { mon_cmd($vmid, "guest-fsfreeze-freeze") };
|
|
$self->logerr($@) if $@;
|
|
|
|
return 1; # even on mon command error, ensure we always thaw again
|
|
}
|
|
|
|
# only call if fs_freeze return 1
|
|
sub qga_fs_thaw {
|
|
my ($self, $vmid) = @_;
|
|
|
|
$self->loginfo("issuing guest-agent 'fs-thaw' command");
|
|
eval { mon_cmd($vmid, "guest-fsfreeze-thaw") };
|
|
$self->logerr($@) if $@;
|
|
}
|
|
|
|
# The size for fleecing images needs to be exactly the same size as QEMU sees. E.g. EFI disk can bex
|
|
# attached with a smaller size then the underyling image on the storage.
|
|
sub query_block_node_sizes {
|
|
my ($self, $vmid, $disks) = @_;
|
|
|
|
my $block_info = mon_cmd($vmid, "query-block");
|
|
$block_info = { map { $_->{device} => $_ } $block_info->@* };
|
|
|
|
for my $diskinfo ($disks->@*) {
|
|
my $drive_key = $diskinfo->{virtdev};
|
|
$drive_key .= "-backup" if $drive_key eq 'tpmstate0';
|
|
my $block_node_size =
|
|
eval { $block_info->{"drive-$drive_key"}->{inserted}->{image}->{'virtual-size'}; };
|
|
if (!$block_node_size) {
|
|
$self->loginfo(
|
|
"could not determine block node size of drive '$drive_key' - using fallback");
|
|
$block_node_size = $diskinfo->{size}
|
|
or die "could not determine size of drive '$drive_key'\n";
|
|
}
|
|
$diskinfo->{'block-node-size'} = $block_node_size;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
# we need a running QEMU/KVM process for backup, starts a paused (prelaunch)
|
|
# one if VM isn't already running
|
|
sub enforce_vm_running_for_backup {
|
|
my ($self, $vmid) = @_;
|
|
|
|
if (PVE::QemuServer::check_running($vmid)) {
|
|
$self->{vm_was_running} = 1;
|
|
return;
|
|
}
|
|
|
|
eval {
|
|
$self->loginfo("starting kvm to execute backup task");
|
|
# start with skiplock
|
|
my $params = {
|
|
skiplock => 1,
|
|
skiptemplate => 1,
|
|
paused => 1,
|
|
};
|
|
PVE::QemuServer::vm_start($self->{storecfg}, $vmid, $params);
|
|
};
|
|
die $@ if $@;
|
|
}
|
|
|
|
# resume VM again once in a clear state (stop mode backup of running VM)
|
|
sub resume_vm_after_job_start {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
return if !$self->{vm_was_running} || $self->{vm_was_paused};
|
|
|
|
if (my $stoptime = $task->{vmstoptime}) {
|
|
my $delay = time() - $task->{vmstoptime};
|
|
$task->{vmstoptime} = undef; # avoid printing 'online after ..' twice
|
|
$self->loginfo("resuming VM again after $delay seconds");
|
|
} else {
|
|
$self->loginfo("resuming VM again");
|
|
}
|
|
mon_cmd($vmid, 'cont', timeout => 45);
|
|
}
|
|
|
|
# stop again if VM was not running before
|
|
sub restore_vm_power_state {
|
|
my ($self, $vmid) = @_;
|
|
|
|
# we always let VMs keep running
|
|
return if $self->{vm_was_running};
|
|
|
|
eval {
|
|
my $resp = mon_cmd($vmid, 'query-status');
|
|
my $status = $resp && $resp->{status} ? $resp->{status} : 'unknown';
|
|
if ($status eq 'prelaunch') {
|
|
$self->loginfo("stopping kvm after backup task");
|
|
PVE::QemuServer::vm_stop($self->{storecfg}, $vmid, 1);
|
|
} else {
|
|
$self->loginfo("kvm status changed after backup ('$status') - keep VM running");
|
|
}
|
|
};
|
|
warn $@ if $@;
|
|
}
|
|
|
|
sub mon_backup_cancel {
|
|
my ($self, $vmid) = @_;
|
|
|
|
$self->loginfo("aborting backup job");
|
|
eval { mon_cmd($vmid, 'backup-cancel') };
|
|
$self->logerr($@) if $@;
|
|
}
|
|
|
|
sub snapshot {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
# nothing to do
|
|
}
|
|
|
|
my sub cleanup_file_handles {
|
|
my ($self, $file_handles) = @_;
|
|
|
|
for my $file_handle ($file_handles->@*) {
|
|
close($file_handle) or $self->log('warn', "unable to close file handle - $!");
|
|
}
|
|
}
|
|
|
|
my sub cleanup_nbd_mounts {
|
|
my ($self, $info) = @_;
|
|
|
|
for my $mount_point (keys $info->%*) {
|
|
my $pid_file = delete($info->{$mount_point}->{'pid-file'});
|
|
unlink($pid_file) or $self->log('warn', "unable to unlink '$pid_file' - $!");
|
|
# Do a lazy unmount, because the target might still be busy even if the file handle was
|
|
# already closed.
|
|
eval { run_command(['fusermount', '-z', '-u', $mount_point ]); };
|
|
if (my $err = $@) {
|
|
delete $info->{$mount_point};
|
|
$self->log('warn', "unable to unmount NBD backup source '$mount_point' - $err");
|
|
}
|
|
}
|
|
|
|
# Wait for the unmount before cleaning up child PIDs to avoid 'nbdfuse' processes being
|
|
# interrupted by the signals issued there.
|
|
my $waited;
|
|
my $wait_limit = 50; # 5 seconds
|
|
for ($waited = 0; $waited < $wait_limit && scalar(keys $info->%*); $waited++) {
|
|
for my $mount_point (keys $info->%*) {
|
|
delete($info->{$mount_point}) if !-e $info->{$mount_point}->{'virtual-file'};
|
|
eval { remove_tree($mount_point); };
|
|
}
|
|
usleep(100_000);
|
|
}
|
|
# just informational, remaining child processes will be killed afterwards
|
|
$self->loginfo("unable to gracefully cleanup NBD fuse mounts") if scalar(keys $info->%*) != 0;
|
|
}
|
|
|
|
my sub cleanup_child_processes {
|
|
my ($self, $cpids) = @_;
|
|
|
|
my $waited;
|
|
my $wait_limit = 5;
|
|
for ($waited = 0; $waited < $wait_limit && scalar(keys $cpids->%*); $waited++) {
|
|
for my $cpid (keys $cpids->%*) {
|
|
delete($cpids->{$cpid}) if waitpid($cpid, POSIX::WNOHANG) > 0;
|
|
}
|
|
if ($waited == 0) {
|
|
kill 15, $_ for keys $cpids->%*;
|
|
}
|
|
sleep 1;
|
|
}
|
|
if ($waited == $wait_limit && scalar(keys $cpids->%*)) {
|
|
kill 9, $_ for keys $cpids->%*;
|
|
sleep 1;
|
|
for my $cpid (keys $cpids->%*) {
|
|
delete($cpids->{$cpid}) if waitpid($cpid, POSIX::WNOHANG) > 0;
|
|
}
|
|
$self->log('warn', "unable to collect child process '$_'") for keys $cpids->%*;
|
|
}
|
|
}
|
|
|
|
sub cleanup {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
# If VM was started only for backup, it is already stopped now.
|
|
if (PVE::QemuServer::Helpers::vm_running_locally($vmid)) {
|
|
if ($task->{cleanup}->{'nbd-stop'}) {
|
|
eval { PVE::QemuServer::QMPHelpers::nbd_stop($vmid); };
|
|
$self->logerr($@) if $@;
|
|
}
|
|
|
|
if (my $info = $task->{cleanup}->{'backup-access-teardown'}) {
|
|
my $params = {
|
|
'target-id' => $info->{'target-id'},
|
|
timeout => 60,
|
|
success => $info->{success} ? JSON::true : JSON::false,
|
|
};
|
|
|
|
$self->loginfo("tearing down backup-access");
|
|
eval { mon_cmd($vmid, "backup-access-teardown", $params->%*) };
|
|
$self->logerr($@) if $@;
|
|
}
|
|
|
|
$detach_tpmstate_drive->($task, $vmid);
|
|
}
|
|
|
|
if ($task->{'use-fleecing'}) {
|
|
eval {
|
|
detach_fleecing_images($task->{disks}, $vmid);
|
|
PVE::QemuConfig::cleanup_fleecing_images(
|
|
$vmid, $self->{storecfg}, sub { $self->log($_[0], $_[1]); });
|
|
};
|
|
$self->log('warn', "attempt to clean up fleecing images failed - $@") if $@;
|
|
}
|
|
|
|
if ($self->{qmeventd_fh}) {
|
|
close($self->{qmeventd_fh});
|
|
}
|
|
|
|
cleanup_file_handles($self, $task->{cleanup}->{'file-handles'})
|
|
if $task->{cleanup}->{'file-handles'};
|
|
|
|
cleanup_nbd_mounts($self, $task->{cleanup}->{'nbd-mounts'})
|
|
if $task->{cleanup}->{'nbd-mounts'};
|
|
|
|
cleanup_child_processes($self, $task->{cleanup}->{'child-pids'})
|
|
if $task->{cleanup}->{'child-pids'};
|
|
|
|
if (my $dir = $task->{'backup-access-root-dir'}) {
|
|
eval { remove_tree($dir) };
|
|
$self->log('warn', "unable to cleanup directory $dir - $@") if $@;
|
|
}
|
|
}
|
|
|
|
my sub virtual_file_backup_prepare {
|
|
my ($self, $vmid, $task, $device_name, $size, $nbd_path, $bitmap_name) = @_;
|
|
|
|
my $cleanup = $task->{cleanup};
|
|
|
|
my $nbd_uri = "nbd+unix:///${device_name}?socket=${nbd_path}";
|
|
|
|
my $error_fh;
|
|
my $next_dirty_region;
|
|
|
|
# If there is no dirty bitmap, it can be treated as if there's a full dirty one. The output of
|
|
# nbdinfo is a list of tuples with offset, length, type, description. The first bit of 'type' is
|
|
# set when the bitmap is dirty, see QEMU's docs/interop/nbd.txt
|
|
my $dirty_bitmap = [];
|
|
if ($bitmap_name) {
|
|
my $input = IO::File->new();
|
|
my $info = IO::File->new();
|
|
$error_fh = IO::File->new();
|
|
my $nbdinfo_cmd = ["nbdinfo", $nbd_uri, "--map=qemu:dirty-bitmap:${bitmap_name}"];
|
|
my $cpid = open3($input, $info, $error_fh, $nbdinfo_cmd->@*)
|
|
or die "failed to spawn nbdinfo child - $!\n";
|
|
$cleanup->{'child-pids'}->{$cpid} = 1;
|
|
|
|
$next_dirty_region = sub {
|
|
my ($offset, $length, $type);
|
|
do {
|
|
my $line = <$info>;
|
|
return if !$line;
|
|
die "unexpected output from nbdinfo - $line\n"
|
|
if $line !~ m/^\s*(\d+)\s*(\d+)\s*(\d+)/; # also untaints
|
|
($offset, $length, $type) = ($1, $2, $3);
|
|
} while (($type & 0x1) == 0); # not dirty
|
|
return ($offset, $length);
|
|
};
|
|
} else {
|
|
my $done = 0;
|
|
$next_dirty_region = sub {
|
|
return if $done;
|
|
$done = 1;
|
|
return (0, $size);
|
|
};
|
|
}
|
|
|
|
my $mount_point = $task->{'backup-access-root-dir'}
|
|
."/${vmid}-nbd.backup-access.${device_name}.$$";
|
|
make_path($mount_point) or die "unable to create directory $mount_point\n";
|
|
$cleanup->{'nbd-mounts'}->{$mount_point} = {};
|
|
|
|
# Note that nbdfuse requires "$dir/$file". A single name would be treated as a dir and the file
|
|
# would be named "$dir/nbd" then
|
|
my $virtual_file = "${mount_point}/${device_name}";
|
|
$cleanup->{'nbd-mounts'}->{$mount_point}->{'virtual-file'} = $virtual_file;
|
|
|
|
my $pid_file = "${mount_point}.pid";
|
|
PVE::Tools::file_set_contents($pid_file, '', 0600);
|
|
$cleanup->{'nbd-mounts'}->{$mount_point}->{'pid-file'} = $pid_file;
|
|
|
|
my $cpid = fork() // die "fork failed: $!\n";
|
|
if (!$cpid) {
|
|
# By default, access will be restricted to the current user, because the allow_other fuse
|
|
# mount option is not used.
|
|
eval {
|
|
run_command(
|
|
["nbdfuse", '--pidfile', $pid_file, $virtual_file, $nbd_uri],
|
|
logfunc => sub { $self->loginfo("nbdfuse '$virtual_file': $_[0]") },
|
|
);
|
|
};
|
|
if (my $err = $@) {
|
|
eval { $self->loginfo($err); };
|
|
POSIX::_exit(1);
|
|
}
|
|
POSIX::_exit(0);
|
|
}
|
|
$cleanup->{'child-pids'}->{$cpid} = 1;
|
|
|
|
my ($virtual_file_ready, $waited) = (0, 0);
|
|
while (!$virtual_file_ready && $waited < 30) { # 3 seconds
|
|
my $pid = PVE::Tools::file_read_firstline($pid_file);
|
|
if ($pid) {
|
|
$virtual_file_ready = 1;
|
|
} else {
|
|
usleep(100_000);
|
|
$waited++;
|
|
}
|
|
}
|
|
die "timeout setting up virtual file '$virtual_file'" if !$virtual_file_ready;
|
|
|
|
$self->loginfo("provided NBD export as a virtual file '$virtual_file'");
|
|
|
|
# NOTE O_DIRECT, because each block should be read exactly once and also because fuse will try
|
|
# to read ahead otherwise, which would produce warning messages if the next block is not
|
|
# mapped/allocated for the NBD export in case of incremental backup. Open as writable to support
|
|
# discard.
|
|
my $fh = IO::File->new($virtual_file, O_RDWR | O_DIRECT)
|
|
or die "unable to open backup source '$virtual_file' - $!\n";
|
|
push $cleanup->{'file-handles'}->@*, $fh;
|
|
|
|
return ($fh, $next_dirty_region);
|
|
}
|
|
|
|
my sub backup_access_to_volume_info {
|
|
my ($self, $vmid, $task, $backup_access_info, $mechanism, $nbd_path) = @_;
|
|
|
|
my $bitmap_action_to_status = {
|
|
'not-used' => 'none',
|
|
'not-used-removed' => 'none',
|
|
'new' => 'new',
|
|
'used' => 'reuse',
|
|
'invalid' => 'new',
|
|
'missing-recreated' => 'new',
|
|
};
|
|
|
|
my $volumes = {};
|
|
|
|
for my $info ($backup_access_info->@*) {
|
|
my $bitmap_status = 'none';
|
|
my $bitmap_name;
|
|
if (my $bitmap_action = $info->{'bitmap-action'}) {
|
|
$bitmap_status = $bitmap_action_to_status->{$bitmap_action}
|
|
or die "got unexpected bitmap action '$bitmap_action'\n";
|
|
|
|
$bitmap_name = $info->{'bitmap-name'} or die "bitmap-name is not present\n";
|
|
}
|
|
|
|
my ($device, $size) = $info->@{qw(device size)};
|
|
|
|
$volumes->{$device}->{'bitmap-mode'} = $bitmap_status;
|
|
$volumes->{$device}->{size} = $size;
|
|
|
|
if ($mechanism eq 'file-handle') {
|
|
my ($fh, $next_dirty_region) = virtual_file_backup_prepare(
|
|
$self, $vmid, $task, $device, $size, $nbd_path, $bitmap_name);
|
|
$volumes->{$device}->{'file-handle'} = $fh;
|
|
$volumes->{$device}->{'next-dirty-region'} = $next_dirty_region;
|
|
} elsif ($mechanism eq 'nbd') {
|
|
$volumes->{$device}->{'nbd-path'} = $nbd_path;
|
|
$volumes->{$device}->{'bitmap-name'} = $bitmap_name;
|
|
} else {
|
|
die "internal error - unkown mechanism '$mechanism'";
|
|
}
|
|
}
|
|
|
|
return $volumes;
|
|
}
|
|
|
|
sub archive_external {
|
|
my ($self, $task, $vmid) = @_;
|
|
|
|
$task->{'backup-access-root-dir'} = "/run/qemu-server/${vmid}.backup-access.$$/";
|
|
make_path($task->{'backup-access-root-dir'})
|
|
or die "unable to create directory $task->{'backup-access-root-dir'}\n";
|
|
chmod(0700, $task->{'backup-access-root-dir'})
|
|
or die "unable to chmod directory $task->{'backup-access-root-dir'}\n";
|
|
|
|
my $guest_config = PVE::Tools::file_get_contents("$task->{tmpdir}/qemu-server.conf");
|
|
my $firewall_file = "$task->{tmpdir}/qemu-server.fw";
|
|
|
|
my $opts = $self->{vzdump}->{opts};
|
|
|
|
my $backup_provider = $self->{vzdump}->{'backup-provider'};
|
|
|
|
$self->loginfo("starting external backup via " . $backup_provider->provider_name());
|
|
|
|
my $starttime = time();
|
|
|
|
$self->enforce_vm_running_for_backup($vmid);
|
|
$self->{qmeventd_fh} = PVE::QemuServer::register_qmeventd_handle($vmid);
|
|
|
|
eval {
|
|
$SIG{INT} = $SIG{TERM} = $SIG{QUIT} = $SIG{HUP} = $SIG{PIPE} = sub {
|
|
die "interrupted by signal\n";
|
|
};
|
|
|
|
my $qemu_support = mon_cmd($vmid, "query-proxmox-support");
|
|
|
|
if (!$qemu_support->{'backup-access-api'}) {
|
|
die "backups access API required for external provider backup is not supported by"
|
|
." the running QEMU version. Please make sure you've installed the latest "
|
|
." version and the VM has been restarted.\n";
|
|
}
|
|
|
|
$attach_tpmstate_drive->($self, $task, $vmid);
|
|
|
|
my $is_template = PVE::QemuConfig->is_template($self->{vmlist}->{$vmid});
|
|
|
|
my $fleecing = check_and_prepare_fleecing(
|
|
$self, $vmid, $opts->{fleecing}, $task->{disks}, $is_template, $qemu_support, 1);
|
|
die "cannot setup backup access without fleecing\n" if !$fleecing;
|
|
|
|
$task->{'use-fleecing'} = 1;
|
|
|
|
my $target_id = "snapshot-access:$opts->{storage}";
|
|
|
|
my $mechanism = $backup_provider->backup_get_mechanism($vmid, 'qemu');
|
|
die "mechanism '$mechanism' requested by backup provider is not supported for VMs\n"
|
|
if $mechanism ne 'file-handle' && $mechanism ne 'nbd';
|
|
|
|
$self->loginfo("using backup mechanism '$mechanism'");
|
|
|
|
if ($mechanism eq 'file-handle') {
|
|
# For mechanism 'file-handle', the nbdfuse binary is required. Also, the bitmap needs
|
|
# to be passed to the provider. The bitmap cannot be dumped via QMP and doing it via
|
|
# qemu-img is experimental, so use nbdinfo. Both are in libnbd-bin.
|
|
die "need 'nbdfuse' binary from package libnbd-bin\n" if !-e "/usr/bin/nbdfuse";
|
|
}
|
|
|
|
my $devices = {};
|
|
for my $di ($task->{disks}->@*) {
|
|
my $device_name = $di->{qmdevice};
|
|
die "implement me (type '$di->{type}')"
|
|
if $di->{type} ne 'block' && $di->{type} ne 'file';
|
|
$devices->{$device_name}->{size} = $di->{'block-node-size'};
|
|
}
|
|
|
|
my $incremental_info = $backup_provider->backup_vm_query_incremental($vmid, $devices);
|
|
|
|
my $qmp_devices = [];
|
|
for my $device (sort keys $devices->%*) {
|
|
my $qmp_device = { device => $device };
|
|
if (defined(my $mode = $incremental_info->{$device})) {
|
|
if ($mode eq 'new' || $mode eq 'use' || $mode eq 'none') {
|
|
$qmp_device->{'bitmap-mode'} = $mode;
|
|
} else {
|
|
die "invalid incremental mode '$mode' returned by backup provider plugin\n";
|
|
}
|
|
}
|
|
push($qmp_devices->@*, $qmp_device);
|
|
}
|
|
|
|
my $params = {
|
|
'target-id' => $target_id,
|
|
devices => $qmp_devices,
|
|
timeout => 60,
|
|
};
|
|
|
|
my $fs_frozen = $self->qga_fs_freeze($task, $vmid);
|
|
|
|
$self->loginfo("setting up snapshot-access for backup");
|
|
|
|
$task->{cleanup}->{'backup-access-teardown'} = { 'target-id' => $target_id, success => 0 };
|
|
|
|
my $backup_access_info = eval { mon_cmd($vmid, "backup-access-setup", $params->%*) };
|
|
my $qmperr = $@;
|
|
|
|
if ($fs_frozen) {
|
|
$self->qga_fs_thaw($vmid);
|
|
}
|
|
|
|
die $qmperr if $qmperr;
|
|
|
|
$self->resume_vm_after_job_start($task, $vmid);
|
|
|
|
my $bitmap_info = mon_cmd($vmid, 'query-pbs-bitmap-info');
|
|
for my $info (sort { $a->{drive} cmp $b->{drive} } $bitmap_info->@*) {
|
|
my $text = $bitmap_action_to_human->($self, $info);
|
|
my $drive = $info->{drive};
|
|
$drive =~ s/^drive-//; # for consistency
|
|
$self->loginfo("$drive: dirty-bitmap status: $text");
|
|
}
|
|
|
|
$self->loginfo("starting NBD server");
|
|
|
|
my $nbd_path = "$task->{'backup-access-root-dir'}/${vmid}-nbd.backup-access";
|
|
mon_cmd(
|
|
$vmid, "nbd-server-start", addr => { type => 'unix', data => { path => $nbd_path } } );
|
|
$task->{cleanup}->{'nbd-stop'} = 1;
|
|
|
|
for my $info ($backup_access_info->@*) {
|
|
$self->loginfo("adding NBD export for $info->{device}");
|
|
|
|
my $export_params = {
|
|
id => $info->{device},
|
|
'node-name' => $info->{'node-name'},
|
|
writable => JSON::true, # for discard
|
|
type => "nbd",
|
|
name => $info->{device}, # NBD export name
|
|
};
|
|
|
|
if ($info->{'bitmap-name'}) {
|
|
$export_params->{bitmaps} = [{
|
|
node => $info->{'bitmap-node-name'},
|
|
name => $info->{'bitmap-name'},
|
|
}];
|
|
}
|
|
|
|
mon_cmd($vmid, "block-export-add", $export_params->%*);
|
|
}
|
|
|
|
my $volumes = backup_access_to_volume_info(
|
|
$self, $vmid, $task, $backup_access_info, $mechanism, $nbd_path);
|
|
|
|
my $param = {};
|
|
$param->{'bandwidth-limit'} = $opts->{bwlimit} * 1024 if $opts->{bwlimit};
|
|
$param->{'firewall-config'} = PVE::Tools::file_get_contents($firewall_file)
|
|
if -e $firewall_file;
|
|
|
|
$backup_provider->backup_vm($vmid, $guest_config, $volumes, $param);
|
|
};
|
|
my $err = $@;
|
|
|
|
if ($err) {
|
|
$self->logerr($err);
|
|
$self->resume_vm_after_job_start($task, $vmid);
|
|
} else {
|
|
$task->{cleanup}->{'backup-access-teardown'}->{success} = 1;
|
|
}
|
|
$self->restore_vm_power_state($vmid);
|
|
|
|
die $err if $err;
|
|
}
|
|
|
|
1;
|