mirror of
https://git.proxmox.com/git/qemu-server
synced 2025-05-02 16:35:30 +00:00

fixing the following two issues: - the legacy code path was never converted to the new fork_tunnel signature (which probably means that nothing triggers it in practice anymore?) - the NBD Unix socket got forwarded multiple times if more than one disk was migrated via NBD (this is harmless, but wrong) for the second issue I opted to keep the code compatible with the possibility that Qemu starts supporting multiple NBD servers in the future (and the target node could thus return multiple UNIX socket paths). currently we can only start one NBD server on one socket, and each drive-mirror simply starts a new connection over that single socket. I took the liberty of renaming the variables/keys since I found 'tunnel_addr' and 'sock_addr' rather confusing. Reviewed-By: Mira Limbeck <m.limbeck@proxmox.com> Tested-By: Mira Limbeck <m.limbeck@proxmox.com> Signed-off-by: Fabian Grünbichler <f.gruenbichler@proxmox.com>
1307 lines
40 KiB
Perl
1307 lines
40 KiB
Perl
package PVE::QemuMigrate;
|
|
|
|
use strict;
|
|
use warnings;
|
|
|
|
use IO::File;
|
|
use IPC::Open2;
|
|
use POSIX qw( WNOHANG );
|
|
use Time::HiRes qw( usleep );
|
|
|
|
use PVE::Cluster;
|
|
use PVE::INotify;
|
|
use PVE::RPCEnvironment;
|
|
use PVE::Replication;
|
|
use PVE::ReplicationConfig;
|
|
use PVE::ReplicationState;
|
|
use PVE::Storage;
|
|
use PVE::Tools;
|
|
|
|
use PVE::QemuConfig;
|
|
use PVE::QemuServer::CPUConfig;
|
|
use PVE::QemuServer::Drive;
|
|
use PVE::QemuServer::Helpers qw(min_version);
|
|
use PVE::QemuServer::Machine;
|
|
use PVE::QemuServer::Monitor qw(mon_cmd);
|
|
use PVE::QemuServer;
|
|
|
|
use PVE::AbstractMigrate;
|
|
use base qw(PVE::AbstractMigrate);
|
|
|
|
sub fork_command_pipe {
|
|
my ($self, $cmd) = @_;
|
|
|
|
my $reader = IO::File->new();
|
|
my $writer = IO::File->new();
|
|
|
|
my $orig_pid = $$;
|
|
|
|
my $cpid;
|
|
|
|
eval { $cpid = open2($reader, $writer, @$cmd); };
|
|
|
|
my $err = $@;
|
|
|
|
# catch exec errors
|
|
if ($orig_pid != $$) {
|
|
$self->log('err', "can't fork command pipe\n");
|
|
POSIX::_exit(1);
|
|
kill('KILL', $$);
|
|
}
|
|
|
|
die $err if $err;
|
|
|
|
return { writer => $writer, reader => $reader, pid => $cpid };
|
|
}
|
|
|
|
sub finish_command_pipe {
|
|
my ($self, $cmdpipe, $timeout) = @_;
|
|
|
|
my $cpid = $cmdpipe->{pid};
|
|
return if !defined($cpid);
|
|
|
|
my $writer = $cmdpipe->{writer};
|
|
my $reader = $cmdpipe->{reader};
|
|
|
|
$writer->close();
|
|
$reader->close();
|
|
|
|
my $collect_child_process = sub {
|
|
my $res = waitpid($cpid, WNOHANG);
|
|
if (defined($res) && ($res == $cpid)) {
|
|
delete $cmdpipe->{cpid};
|
|
return 1;
|
|
} else {
|
|
return 0;
|
|
}
|
|
};
|
|
|
|
if ($timeout) {
|
|
for (my $i = 0; $i < $timeout; $i++) {
|
|
return if &$collect_child_process();
|
|
sleep(1);
|
|
}
|
|
}
|
|
|
|
$self->log('info', "ssh tunnel still running - terminating now with SIGTERM\n");
|
|
kill(15, $cpid);
|
|
|
|
# wait again
|
|
for (my $i = 0; $i < 10; $i++) {
|
|
return if &$collect_child_process();
|
|
sleep(1);
|
|
}
|
|
|
|
$self->log('info', "ssh tunnel still running - terminating now with SIGKILL\n");
|
|
kill 9, $cpid;
|
|
sleep 1;
|
|
|
|
$self->log('err', "ssh tunnel child process (PID $cpid) couldn't be collected\n")
|
|
if !&$collect_child_process();
|
|
}
|
|
|
|
sub read_tunnel {
|
|
my ($self, $tunnel, $timeout) = @_;
|
|
|
|
$timeout = 60 if !defined($timeout);
|
|
|
|
my $reader = $tunnel->{reader};
|
|
|
|
my $output;
|
|
eval {
|
|
PVE::Tools::run_with_timeout($timeout, sub { $output = <$reader>; });
|
|
};
|
|
die "reading from tunnel failed: $@\n" if $@;
|
|
|
|
chomp $output;
|
|
|
|
return $output;
|
|
}
|
|
|
|
sub write_tunnel {
|
|
my ($self, $tunnel, $timeout, $command) = @_;
|
|
|
|
$timeout = 60 if !defined($timeout);
|
|
|
|
my $writer = $tunnel->{writer};
|
|
|
|
eval {
|
|
PVE::Tools::run_with_timeout($timeout, sub {
|
|
print $writer "$command\n";
|
|
$writer->flush();
|
|
});
|
|
};
|
|
die "writing to tunnel failed: $@\n" if $@;
|
|
|
|
if ($tunnel->{version} && $tunnel->{version} >= 1) {
|
|
my $res = eval { $self->read_tunnel($tunnel, 10); };
|
|
die "no reply to command '$command': $@\n" if $@;
|
|
|
|
if ($res eq 'OK') {
|
|
return;
|
|
} else {
|
|
die "tunnel replied '$res' to command '$command'\n";
|
|
}
|
|
}
|
|
}
|
|
|
|
sub fork_tunnel {
|
|
my ($self, $ssh_forward_info) = @_;
|
|
|
|
my @localtunnelinfo = ();
|
|
foreach my $addr (@$ssh_forward_info) {
|
|
push @localtunnelinfo, '-L', $addr;
|
|
}
|
|
|
|
my $cmd = [@{$self->{rem_ssh}}, '-o ExitOnForwardFailure=yes', @localtunnelinfo, '/usr/sbin/qm', 'mtunnel' ];
|
|
|
|
my $tunnel = $self->fork_command_pipe($cmd);
|
|
|
|
eval {
|
|
my $helo = $self->read_tunnel($tunnel, 60);
|
|
die "no reply\n" if !$helo;
|
|
die "no quorum on target node\n" if $helo =~ m/^no quorum$/;
|
|
die "got strange reply from mtunnel ('$helo')\n"
|
|
if $helo !~ m/^tunnel online$/;
|
|
};
|
|
my $err = $@;
|
|
|
|
eval {
|
|
my $ver = $self->read_tunnel($tunnel, 10);
|
|
if ($ver =~ /^ver (\d+)$/) {
|
|
$tunnel->{version} = $1;
|
|
$self->log('info', "ssh tunnel $ver\n");
|
|
} else {
|
|
$err = "received invalid tunnel version string '$ver'\n" if !$err;
|
|
}
|
|
};
|
|
|
|
if ($err) {
|
|
$self->finish_command_pipe($tunnel);
|
|
die "can't open migration tunnel - $err";
|
|
}
|
|
return $tunnel;
|
|
}
|
|
|
|
sub finish_tunnel {
|
|
my ($self, $tunnel) = @_;
|
|
|
|
eval { $self->write_tunnel($tunnel, 30, 'quit'); };
|
|
my $err = $@;
|
|
|
|
$self->finish_command_pipe($tunnel, 30);
|
|
|
|
if (my $unix_sockets = $tunnel->{unix_sockets}) {
|
|
# ssh does not clean up on local host
|
|
my $cmd = ['rm', '-f', @$unix_sockets];
|
|
PVE::Tools::run_command($cmd);
|
|
|
|
# .. and just to be sure check on remote side
|
|
unshift @{$cmd}, @{$self->{rem_ssh}};
|
|
PVE::Tools::run_command($cmd);
|
|
}
|
|
|
|
die $err if $err;
|
|
}
|
|
|
|
sub lock_vm {
|
|
my ($self, $vmid, $code, @param) = @_;
|
|
|
|
return PVE::QemuConfig->lock_config($vmid, $code, @param);
|
|
}
|
|
|
|
sub prepare {
|
|
my ($self, $vmid) = @_;
|
|
|
|
my $online = $self->{opts}->{online};
|
|
|
|
$self->{storecfg} = PVE::Storage::config();
|
|
|
|
# test if VM exists
|
|
my $conf = $self->{vmconf} = PVE::QemuConfig->load_config($vmid);
|
|
|
|
PVE::QemuConfig->check_lock($conf);
|
|
|
|
my $running = 0;
|
|
if (my $pid = PVE::QemuServer::check_running($vmid)) {
|
|
die "can't migrate running VM without --online\n" if !$online;
|
|
$running = $pid;
|
|
|
|
$self->{forcemachine} = PVE::QemuServer::Machine::qemu_machine_pxe($vmid, $conf);
|
|
|
|
# To support custom CPU types, we keep QEMU's "-cpu" parameter intact.
|
|
# Since the parameter itself contains no reference to a custom model,
|
|
# this makes migration independent of changes to "cpu-models.conf".
|
|
if ($conf->{cpu}) {
|
|
my $cpuconf = PVE::QemuServer::CPUConfig::parse_cpu_conf_basic($conf->{cpu});
|
|
if ($cpuconf && PVE::QemuServer::CPUConfig::is_custom_model($cpuconf->{cputype})) {
|
|
$self->{forcecpu} = PVE::QemuServer::CPUConfig::get_cpu_from_running_vm($pid);
|
|
}
|
|
}
|
|
}
|
|
|
|
my $loc_res = PVE::QemuServer::check_local_resources($conf, 1);
|
|
if (scalar @$loc_res) {
|
|
if ($self->{running} || !$self->{opts}->{force}) {
|
|
die "can't migrate VM which uses local devices: " . join(", ", @$loc_res) . "\n";
|
|
} else {
|
|
$self->log('info', "migrating VM which uses local devices");
|
|
}
|
|
}
|
|
|
|
my $vollist = PVE::QemuServer::get_vm_volumes($conf);
|
|
|
|
my $need_activate = [];
|
|
foreach my $volid (@$vollist) {
|
|
my ($sid, $volname) = PVE::Storage::parse_volume_id($volid, 1);
|
|
|
|
# check if storage is available on both nodes
|
|
my $targetsid = PVE::QemuServer::map_storage($self->{opts}->{storagemap}, $sid);
|
|
|
|
my $scfg = PVE::Storage::storage_check_node($self->{storecfg}, $sid);
|
|
PVE::Storage::storage_check_node($self->{storecfg}, $targetsid, $self->{node});
|
|
|
|
if ($scfg->{shared}) {
|
|
# PVE::Storage::activate_storage checks this for non-shared storages
|
|
my $plugin = PVE::Storage::Plugin->lookup($scfg->{type});
|
|
warn "Used shared storage '$sid' is not online on source node!\n"
|
|
if !$plugin->check_connection($sid, $scfg);
|
|
} else {
|
|
# only activate if not shared
|
|
next if ($volid =~ m/vm-\d+-cloudinit/);
|
|
push @$need_activate, $volid;
|
|
}
|
|
}
|
|
|
|
# activate volumes
|
|
PVE::Storage::activate_volumes($self->{storecfg}, $need_activate);
|
|
|
|
# test ssh connection
|
|
my $cmd = [ @{$self->{rem_ssh}}, '/bin/true' ];
|
|
eval { $self->cmd_quiet($cmd); };
|
|
die "Can't connect to destination address using public key\n" if $@;
|
|
|
|
return $running;
|
|
}
|
|
|
|
sub sync_disks {
|
|
my ($self, $vmid) = @_;
|
|
|
|
my $conf = $self->{vmconf};
|
|
|
|
# local volumes which have been copied
|
|
# and their old_id => new_id pairs
|
|
$self->{volumes} = [];
|
|
$self->{volume_map} = {};
|
|
|
|
my $storecfg = $self->{storecfg};
|
|
eval {
|
|
|
|
# found local volumes and their origin
|
|
my $local_volumes = {};
|
|
my $local_volumes_errors = {};
|
|
my $other_errors = [];
|
|
my $abort = 0;
|
|
|
|
my $log_error = sub {
|
|
my ($msg, $volid) = @_;
|
|
|
|
if (defined($volid)) {
|
|
$local_volumes_errors->{$volid} = $msg;
|
|
} else {
|
|
push @$other_errors, $msg;
|
|
}
|
|
$abort = 1;
|
|
};
|
|
|
|
my @sids = PVE::Storage::storage_ids($storecfg);
|
|
foreach my $storeid (@sids) {
|
|
my $scfg = PVE::Storage::storage_config($storecfg, $storeid);
|
|
next if $scfg->{shared};
|
|
next if !PVE::Storage::storage_check_enabled($storecfg, $storeid, undef, 1);
|
|
|
|
# get list from PVE::Storage (for unused volumes)
|
|
my $dl = PVE::Storage::vdisk_list($storecfg, $storeid, $vmid);
|
|
|
|
next if @{$dl->{$storeid}} == 0;
|
|
|
|
my $targetsid = PVE::QemuServer::map_storage($self->{opts}->{storagemap}, $storeid);
|
|
# check if storage is available on target node
|
|
PVE::Storage::storage_check_node($storecfg, $targetsid, $self->{node});
|
|
|
|
# grandfather in existing mismatches
|
|
if ($targetsid ne $storeid) {
|
|
my $target_scfg = PVE::Storage::storage_config($storecfg, $targetsid);
|
|
die "content type 'images' is not available on storage '$targetsid'\n"
|
|
if !$target_scfg->{content}->{images};
|
|
}
|
|
|
|
PVE::Storage::foreach_volid($dl, sub {
|
|
my ($volid, $sid, $volinfo) = @_;
|
|
|
|
$local_volumes->{$volid}->{ref} = 'storage';
|
|
|
|
# If with_snapshots is not set for storage migrate, it tries to use
|
|
# a raw+size stream, but on-the-fly conversion from qcow2 to raw+size
|
|
# back to qcow2 is currently not possible.
|
|
$local_volumes->{$volid}->{snapshots} = ($volinfo->{format} =~ /^(?:qcow2|vmdk)$/);
|
|
$local_volumes->{$volid}->{format} = $volinfo->{format};
|
|
});
|
|
}
|
|
|
|
my $rep_cfg = PVE::ReplicationConfig->new();
|
|
my $replication_jobcfg = $rep_cfg->find_local_replication_job($vmid, $self->{node});
|
|
my $replicatable_volumes = !$replication_jobcfg ? {}
|
|
: PVE::QemuConfig->get_replicatable_volumes($storecfg, $vmid, $conf, 0, 1);
|
|
|
|
my $test_volid = sub {
|
|
my ($volid, $attr) = @_;
|
|
|
|
if ($volid =~ m|^/|) {
|
|
return if $attr->{shared};
|
|
$local_volumes->{$volid}->{ref} = 'config';
|
|
die "local file/device\n";
|
|
}
|
|
|
|
my $snaprefs = $attr->{referenced_in_snapshot};
|
|
|
|
if ($attr->{cdrom}) {
|
|
if ($volid eq 'cdrom') {
|
|
my $msg = "can't migrate local cdrom drive";
|
|
if (defined($snaprefs) && !$attr->{referenced_in_config}) {
|
|
my $snapnames = join(', ', sort keys %$snaprefs);
|
|
$msg .= " (referenced in snapshot - $snapnames)";
|
|
}
|
|
&$log_error("$msg\n");
|
|
return;
|
|
}
|
|
return if $volid eq 'none';
|
|
}
|
|
|
|
my ($sid, $volname) = PVE::Storage::parse_volume_id($volid);
|
|
|
|
my $targetsid = PVE::QemuServer::map_storage($self->{opts}->{storagemap}, $sid);
|
|
# check if storage is available on both nodes
|
|
my $scfg = PVE::Storage::storage_check_node($storecfg, $sid);
|
|
PVE::Storage::storage_check_node($storecfg, $targetsid, $self->{node});
|
|
|
|
return if $scfg->{shared};
|
|
|
|
$local_volumes->{$volid}->{ref} = $attr->{referenced_in_config} ? 'config' : 'snapshot';
|
|
$local_volumes->{$volid}->{ref} = 'storage' if $attr->{is_unused};
|
|
|
|
$local_volumes->{$volid}->{is_vmstate} = $attr->{is_vmstate} ? 1 : 0;
|
|
|
|
if ($attr->{cdrom}) {
|
|
if ($volid =~ /vm-\d+-cloudinit/) {
|
|
$local_volumes->{$volid}->{ref} = 'generated';
|
|
return;
|
|
}
|
|
die "local cdrom image\n";
|
|
}
|
|
|
|
my ($path, $owner) = PVE::Storage::path($storecfg, $volid);
|
|
|
|
die "owned by other VM (owner = VM $owner)\n"
|
|
if !$owner || ($owner != $vmid);
|
|
|
|
return if $attr->{is_vmstate};
|
|
|
|
if (defined($snaprefs)) {
|
|
$local_volumes->{$volid}->{snapshots} = 1;
|
|
|
|
# we cannot migrate shapshots on local storage
|
|
# exceptions: 'zfspool' or 'qcow2' files (on directory storage)
|
|
|
|
die "online storage migration not possible if snapshot exists\n" if $self->{running};
|
|
if (!($scfg->{type} eq 'zfspool' || $local_volumes->{$volid}->{format} eq 'qcow2')) {
|
|
die "non-migratable snapshot exists\n";
|
|
}
|
|
}
|
|
|
|
die "referenced by linked clone(s)\n"
|
|
if PVE::Storage::volume_is_base_and_used($storecfg, $volid);
|
|
};
|
|
|
|
PVE::QemuServer::foreach_volid($conf, sub {
|
|
my ($volid, $attr) = @_;
|
|
eval { $test_volid->($volid, $attr); };
|
|
if (my $err = $@) {
|
|
&$log_error($err, $volid);
|
|
}
|
|
});
|
|
|
|
foreach my $vol (sort keys %$local_volumes) {
|
|
my $type = $replicatable_volumes->{$vol} ? 'local, replicated' : 'local';
|
|
my $ref = $local_volumes->{$vol}->{ref};
|
|
if ($ref eq 'storage') {
|
|
$self->log('info', "found $type disk '$vol' (via storage)\n");
|
|
} elsif ($ref eq 'config') {
|
|
&$log_error("can't live migrate attached local disks without with-local-disks option\n", $vol)
|
|
if $self->{running} && !$self->{opts}->{"with-local-disks"};
|
|
$self->log('info', "found $type disk '$vol' (in current VM config)\n");
|
|
} elsif ($ref eq 'snapshot') {
|
|
$self->log('info', "found $type disk '$vol' (referenced by snapshot(s))\n");
|
|
} elsif ($ref eq 'generated') {
|
|
$self->log('info', "found generated disk '$vol' (in current VM config)\n");
|
|
} else {
|
|
$self->log('info', "found $type disk '$vol'\n");
|
|
}
|
|
}
|
|
|
|
foreach my $vol (sort keys %$local_volumes_errors) {
|
|
$self->log('warn', "can't migrate local disk '$vol': $local_volumes_errors->{$vol}");
|
|
}
|
|
foreach my $err (@$other_errors) {
|
|
$self->log('warn', "$err");
|
|
}
|
|
|
|
if ($abort) {
|
|
die "can't migrate VM - check log\n";
|
|
}
|
|
|
|
# additional checks for local storage
|
|
foreach my $volid (keys %$local_volumes) {
|
|
my ($sid, $volname) = PVE::Storage::parse_volume_id($volid);
|
|
my $scfg = PVE::Storage::storage_config($storecfg, $sid);
|
|
|
|
my $migratable = $scfg->{type} =~ /^(?:dir|zfspool|lvmthin|lvm)$/;
|
|
|
|
die "can't migrate '$volid' - storage type '$scfg->{type}' not supported\n"
|
|
if !$migratable;
|
|
|
|
# image is a linked clone on local storage, se we can't migrate.
|
|
if (my $basename = (PVE::Storage::parse_volname($storecfg, $volid))[3]) {
|
|
die "can't migrate '$volid' as it's a clone of '$basename'";
|
|
}
|
|
}
|
|
|
|
if ($replication_jobcfg) {
|
|
if ($self->{running}) {
|
|
|
|
my $version = PVE::QemuServer::kvm_user_version();
|
|
if (!min_version($version, 4, 2)) {
|
|
die "can't live migrate VM with replicated volumes, pve-qemu to old (< 4.2)!\n"
|
|
}
|
|
|
|
my $live_replicatable_volumes = {};
|
|
PVE::QemuConfig->foreach_volume($conf, sub {
|
|
my ($ds, $drive) = @_;
|
|
|
|
my $volid = $drive->{file};
|
|
$live_replicatable_volumes->{$ds} = $volid
|
|
if defined($replicatable_volumes->{$volid});
|
|
});
|
|
foreach my $drive (keys %$live_replicatable_volumes) {
|
|
my $volid = $live_replicatable_volumes->{$drive};
|
|
|
|
my $bitmap = "repl_$drive";
|
|
|
|
# start tracking before replication to get full delta + a few duplicates
|
|
$self->log('info', "$drive: start tracking writes using block-dirty-bitmap '$bitmap'");
|
|
mon_cmd($vmid, 'block-dirty-bitmap-add', node => "drive-$drive", name => $bitmap);
|
|
|
|
# other info comes from target node in phase 2
|
|
$self->{target_drive}->{$drive}->{bitmap} = $bitmap;
|
|
}
|
|
}
|
|
$self->log('info', "replicating disk images");
|
|
|
|
my $start_time = time();
|
|
my $logfunc = sub { $self->log('info', shift) };
|
|
$self->{replicated_volumes} = PVE::Replication::run_replication(
|
|
'PVE::QemuConfig', $replication_jobcfg, $start_time, $start_time, $logfunc);
|
|
}
|
|
|
|
# sizes in config have to be accurate for remote node to correctly
|
|
# allocate disks, rescan to be sure
|
|
my $volid_hash = PVE::QemuServer::scan_volids($storecfg, $vmid);
|
|
PVE::QemuConfig->foreach_volume($conf, sub {
|
|
my ($key, $drive) = @_;
|
|
return if $key eq 'efidisk0'; # skip efidisk, will be handled later
|
|
return if !defined($local_volumes->{$key}); # only update sizes for local volumes
|
|
|
|
my ($updated, $old_size, $new_size) = PVE::QemuServer::Drive::update_disksize($drive, $volid_hash);
|
|
if (defined($updated)) {
|
|
$conf->{$key} = PVE::QemuServer::print_drive($updated);
|
|
$self->log('info', "size of disk '$updated->{file}' ($key) updated from $old_size to $new_size\n");
|
|
}
|
|
});
|
|
|
|
# we want to set the efidisk size in the config to the size of the
|
|
# real OVMF_VARS.fd image, else we can create a too big image, which does not work
|
|
if (defined($conf->{efidisk0})) {
|
|
PVE::QemuServer::update_efidisk_size($conf);
|
|
}
|
|
|
|
$self->log('info', "copying local disk images") if scalar(%$local_volumes);
|
|
|
|
foreach my $volid (keys %$local_volumes) {
|
|
my ($sid, $volname) = PVE::Storage::parse_volume_id($volid);
|
|
my $targetsid = PVE::QemuServer::map_storage($self->{opts}->{storagemap}, $sid);
|
|
my $ref = $local_volumes->{$volid}->{ref};
|
|
if ($self->{running} && $ref eq 'config') {
|
|
push @{$self->{online_local_volumes}}, $volid;
|
|
} elsif ($ref eq 'generated') {
|
|
die "can't live migrate VM with local cloudinit disk. use a shared storage instead\n" if $self->{running};
|
|
# skip all generated volumes but queue them for deletion in phase3_cleanup
|
|
push @{$self->{volumes}}, $volid;
|
|
next;
|
|
} else {
|
|
next if $self->{replicated_volumes}->{$volid};
|
|
push @{$self->{volumes}}, $volid;
|
|
my $opts = $self->{opts};
|
|
# use 'migrate' limit for transfer to other node
|
|
my $bwlimit = PVE::Storage::get_bandwidth_limit('migration', [$targetsid, $sid], $opts->{bwlimit});
|
|
# JSONSchema and get_bandwidth_limit use kbps - storage_migrate bps
|
|
$bwlimit = $bwlimit * 1024 if defined($bwlimit);
|
|
|
|
my $storage_migrate_opts = {
|
|
'bwlimit' => $bwlimit,
|
|
'insecure' => $opts->{migration_type} eq 'insecure',
|
|
'with_snapshots' => $local_volumes->{$volid}->{snapshots},
|
|
'allow_rename' => !$local_volumes->{$volid}->{is_vmstate},
|
|
};
|
|
|
|
my $logfunc = sub { $self->log('info', $_[0]); };
|
|
my $new_volid = eval {
|
|
PVE::Storage::storage_migrate($storecfg, $volid, $self->{ssh_info},
|
|
$targetsid, $storage_migrate_opts, $logfunc);
|
|
};
|
|
if (my $err = $@) {
|
|
die "storage migration for '$volid' to storage '$targetsid' failed - $err\n";
|
|
}
|
|
|
|
$self->{volume_map}->{$volid} = $new_volid;
|
|
$self->log('info', "volume '$volid' is '$new_volid' on the target\n");
|
|
}
|
|
}
|
|
};
|
|
die "Failed to sync data - $@" if $@;
|
|
}
|
|
|
|
sub cleanup_remotedisks {
|
|
my ($self) = @_;
|
|
|
|
foreach my $target_drive (keys %{$self->{target_drive}}) {
|
|
my $drivestr = $self->{target_drive}->{$target_drive}->{drivestr};
|
|
next if !defined($drivestr);
|
|
|
|
my $drive = PVE::QemuServer::parse_drive($target_drive, $drivestr);
|
|
|
|
# don't clean up replicated disks!
|
|
next if defined($self->{replicated_volumes}->{$drive->{file}});
|
|
|
|
my ($storeid, $volname) = PVE::Storage::parse_volume_id($drive->{file});
|
|
|
|
my $cmd = [@{$self->{rem_ssh}}, 'pvesm', 'free', "$storeid:$volname"];
|
|
|
|
eval{ PVE::Tools::run_command($cmd, outfunc => sub {}, errfunc => sub {}) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
sub cleanup_bitmaps {
|
|
my ($self) = @_;
|
|
foreach my $drive (keys %{$self->{target_drive}}) {
|
|
my $bitmap = $self->{target_drive}->{$drive}->{bitmap};
|
|
next if !$bitmap;
|
|
$self->log('info', "$drive: removing block-dirty-bitmap '$bitmap'");
|
|
mon_cmd($self->{vmid}, 'block-dirty-bitmap-remove', node => "drive-$drive", name => $bitmap);
|
|
}
|
|
}
|
|
|
|
sub phase1 {
|
|
my ($self, $vmid) = @_;
|
|
|
|
$self->log('info', "starting migration of VM $vmid to node '$self->{node}' ($self->{nodeip})");
|
|
|
|
my $conf = $self->{vmconf};
|
|
|
|
# set migrate lock in config file
|
|
$conf->{lock} = 'migrate';
|
|
PVE::QemuConfig->write_config($vmid, $conf);
|
|
|
|
sync_disks($self, $vmid);
|
|
|
|
# sync_disks fixes disk sizes to match their actual size, write changes so
|
|
# target allocates correct volumes
|
|
PVE::QemuConfig->write_config($vmid, $conf);
|
|
};
|
|
|
|
sub phase1_cleanup {
|
|
my ($self, $vmid, $err) = @_;
|
|
|
|
$self->log('info', "aborting phase 1 - cleanup resources");
|
|
|
|
my $conf = $self->{vmconf};
|
|
delete $conf->{lock};
|
|
eval { PVE::QemuConfig->write_config($vmid, $conf) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
}
|
|
|
|
if ($self->{volumes}) {
|
|
foreach my $volid (@{$self->{volumes}}) {
|
|
$self->log('err', "found stale volume copy '$volid' on node '$self->{node}'");
|
|
# fixme: try to remove ?
|
|
}
|
|
}
|
|
|
|
eval { $self->cleanup_bitmaps() };
|
|
if (my $err =$@) {
|
|
$self->log('err', $err);
|
|
}
|
|
|
|
}
|
|
|
|
sub phase2 {
|
|
my ($self, $vmid) = @_;
|
|
|
|
my $conf = $self->{vmconf};
|
|
|
|
$self->log('info', "starting VM $vmid on remote node '$self->{node}'");
|
|
|
|
my $raddr;
|
|
my $rport;
|
|
my $ruri; # the whole migration dst. URI (protocol:address[:port])
|
|
my $nodename = PVE::INotify::nodename();
|
|
|
|
## start on remote node
|
|
my $cmd = [@{$self->{rem_ssh}}];
|
|
|
|
my $spice_ticket;
|
|
if (PVE::QemuServer::vga_conf_has_spice($conf->{vga})) {
|
|
my $res = mon_cmd($vmid, 'query-spice');
|
|
$spice_ticket = $res->{ticket};
|
|
}
|
|
|
|
push @$cmd , 'qm', 'start', $vmid, '--skiplock', '--migratedfrom', $nodename;
|
|
|
|
my $migration_type = $self->{opts}->{migration_type};
|
|
|
|
push @$cmd, '--migration_type', $migration_type;
|
|
|
|
push @$cmd, '--migration_network', $self->{opts}->{migration_network}
|
|
if $self->{opts}->{migration_network};
|
|
|
|
if ($migration_type eq 'insecure') {
|
|
push @$cmd, '--stateuri', 'tcp';
|
|
} else {
|
|
push @$cmd, '--stateuri', 'unix';
|
|
}
|
|
|
|
if ($self->{forcemachine}) {
|
|
push @$cmd, '--machine', $self->{forcemachine};
|
|
}
|
|
|
|
if ($self->{forcecpu}) {
|
|
push @$cmd, '--force-cpu', $self->{forcecpu};
|
|
}
|
|
|
|
if ($self->{online_local_volumes}) {
|
|
push @$cmd, '--targetstorage', ($self->{opts}->{targetstorage} // '1');
|
|
}
|
|
|
|
my $spice_port;
|
|
my $unix_socket_info = {};
|
|
# version > 0 for unix socket support
|
|
my $nbd_protocol_version = 1;
|
|
# TODO change to 'spice_ticket: <ticket>\n' in 7.0
|
|
my $input = $spice_ticket ? "$spice_ticket\n" : "\n";
|
|
$input .= "nbd_protocol_version: $nbd_protocol_version\n";
|
|
|
|
my $number_of_online_replicated_volumes = 0;
|
|
|
|
# prevent auto-vivification
|
|
if ($self->{online_local_volumes}) {
|
|
foreach my $volid (@{$self->{online_local_volumes}}) {
|
|
next if !$self->{replicated_volumes}->{$volid};
|
|
$number_of_online_replicated_volumes++;
|
|
$input .= "replicated_volume: $volid\n";
|
|
}
|
|
}
|
|
|
|
my $target_replicated_volumes = {};
|
|
|
|
# Note: We try to keep $spice_ticket secret (do not pass via command line parameter)
|
|
# instead we pipe it through STDIN
|
|
my $exitcode = PVE::Tools::run_command($cmd, input => $input, outfunc => sub {
|
|
my $line = shift;
|
|
|
|
if ($line =~ m/^migration listens on tcp:(localhost|[\d\.]+|\[[\d\.:a-fA-F]+\]):(\d+)$/) {
|
|
$raddr = $1;
|
|
$rport = int($2);
|
|
$ruri = "tcp:$raddr:$rport";
|
|
}
|
|
elsif ($line =~ m!^migration listens on unix:(/run/qemu-server/(\d+)\.migrate)$!) {
|
|
$raddr = $1;
|
|
die "Destination UNIX sockets VMID does not match source VMID" if $vmid ne $2;
|
|
$ruri = "unix:$raddr";
|
|
}
|
|
elsif ($line =~ m/^migration listens on port (\d+)$/) {
|
|
$raddr = "localhost";
|
|
$rport = int($1);
|
|
$ruri = "tcp:$raddr:$rport";
|
|
}
|
|
elsif ($line =~ m/^spice listens on port (\d+)$/) {
|
|
$spice_port = int($1);
|
|
}
|
|
elsif ($line =~ m/^storage migration listens on nbd:(localhost|[\d\.]+|\[[\d\.:a-fA-F]+\]):(\d+):exportname=(\S+) volume:(\S+)$/) {
|
|
my $drivestr = $4;
|
|
my $nbd_uri = "nbd:$1:$2:exportname=$3";
|
|
my $targetdrive = $3;
|
|
$targetdrive =~ s/drive-//g;
|
|
|
|
$self->{stopnbd} = 1;
|
|
$self->{target_drive}->{$targetdrive}->{drivestr} = $drivestr;
|
|
$self->{target_drive}->{$targetdrive}->{nbd_uri} = $nbd_uri;
|
|
} elsif ($line =~ m!^storage migration listens on nbd:unix:(/run/qemu-server/(\d+)_nbd\.migrate):exportname=(\S+) volume:(\S+)$!) {
|
|
my $drivestr = $4;
|
|
die "Destination UNIX socket's VMID does not match source VMID" if $vmid ne $2;
|
|
my $nbd_unix_addr = $1;
|
|
my $nbd_uri = "nbd:unix:$nbd_unix_addr:exportname=$3";
|
|
my $targetdrive = $3;
|
|
$targetdrive =~ s/drive-//g;
|
|
|
|
$self->{stopnbd} = 1;
|
|
$self->{target_drive}->{$targetdrive}->{drivestr} = $drivestr;
|
|
$self->{target_drive}->{$targetdrive}->{nbd_uri} = $nbd_uri;
|
|
$unix_socket_info->{$nbd_unix_addr} = 1;
|
|
} elsif ($line =~ m/^re-using replicated volume: (\S+) - (.*)$/) {
|
|
my $drive = $1;
|
|
my $volid = $2;
|
|
$target_replicated_volumes->{$volid} = $drive;
|
|
} elsif ($line =~ m/^QEMU: (.*)$/) {
|
|
$self->log('info', "[$self->{node}] $1\n");
|
|
}
|
|
}, errfunc => sub {
|
|
my $line = shift;
|
|
$self->log('info', "[$self->{node}] $line");
|
|
}, noerr => 1);
|
|
|
|
die "remote command failed with exit code $exitcode\n" if $exitcode;
|
|
|
|
die "unable to detect remote migration address\n" if !$raddr;
|
|
|
|
if (scalar(keys %$target_replicated_volumes) != $number_of_online_replicated_volumes) {
|
|
die "number of replicated disks on source and target node do not match - target node too old?\n"
|
|
}
|
|
|
|
$self->log('info', "start remote tunnel");
|
|
|
|
if ($migration_type eq 'secure') {
|
|
|
|
if ($ruri =~ /^unix:/) {
|
|
my $ssh_forward_info = ["$raddr:$raddr"];
|
|
$unix_socket_info->{$raddr} = 1;
|
|
|
|
my $unix_sockets = [ keys %$unix_socket_info ];
|
|
for my $sock (@$unix_sockets) {
|
|
push @$ssh_forward_info, "$sock:$sock";
|
|
unlink $sock;
|
|
}
|
|
|
|
$self->{tunnel} = $self->fork_tunnel($ssh_forward_info);
|
|
|
|
my $unix_socket_try = 0; # wait for the socket to become ready
|
|
while ($unix_socket_try <= 100) {
|
|
$unix_socket_try++;
|
|
my $available = 0;
|
|
foreach my $sock (@$unix_sockets) {
|
|
if (-S $sock) {
|
|
$available++;
|
|
}
|
|
}
|
|
|
|
if ($available == @$unix_sockets) {
|
|
last;
|
|
}
|
|
|
|
usleep(50000);
|
|
}
|
|
if ($unix_socket_try > 100) {
|
|
$self->{errors} = 1;
|
|
$self->finish_tunnel($self->{tunnel});
|
|
die "Timeout, migration socket $ruri did not get ready";
|
|
}
|
|
$self->{tunnel}->{unix_sockets} = $unix_sockets if (@$unix_sockets);
|
|
|
|
} elsif ($ruri =~ /^tcp:/) {
|
|
my $ssh_forward_info = [];
|
|
if ($raddr eq "localhost") {
|
|
# for backwards compatibility with older qemu-server versions
|
|
my $pfamily = PVE::Tools::get_host_address_family($nodename);
|
|
my $lport = PVE::Tools::next_migrate_port($pfamily);
|
|
push @$ssh_forward_info, "$lport:localhost:$rport";
|
|
}
|
|
|
|
$self->{tunnel} = $self->fork_tunnel($ssh_forward_info);
|
|
|
|
} else {
|
|
die "unsupported protocol in migration URI: $ruri\n";
|
|
}
|
|
} else {
|
|
#fork tunnel for insecure migration, to send faster commands like resume
|
|
$self->{tunnel} = $self->fork_tunnel();
|
|
}
|
|
my $start = time();
|
|
|
|
my $opt_bwlimit = $self->{opts}->{bwlimit};
|
|
|
|
if (defined($self->{online_local_volumes})) {
|
|
$self->{storage_migration} = 1;
|
|
$self->{storage_migration_jobs} = {};
|
|
$self->log('info', "starting storage migration");
|
|
|
|
die "The number of local disks does not match between the source and the destination.\n"
|
|
if (scalar(keys %{$self->{target_drive}}) != scalar @{$self->{online_local_volumes}});
|
|
foreach my $drive (keys %{$self->{target_drive}}){
|
|
my $target = $self->{target_drive}->{$drive};
|
|
my $nbd_uri = $target->{nbd_uri};
|
|
|
|
my $source_drive = PVE::QemuServer::parse_drive($drive, $conf->{$drive});
|
|
my $target_drive = PVE::QemuServer::parse_drive($drive, $target->{drivestr});
|
|
|
|
my $source_volid = $source_drive->{file};
|
|
my $target_volid = $target_drive->{file};
|
|
|
|
my $source_sid = PVE::Storage::Plugin::parse_volume_id($source_volid);
|
|
my $target_sid = PVE::Storage::Plugin::parse_volume_id($target_volid);
|
|
|
|
my $bwlimit = PVE::Storage::get_bandwidth_limit('migration', [$source_sid, $target_sid], $opt_bwlimit);
|
|
my $bitmap = $target->{bitmap};
|
|
|
|
$self->log('info', "$drive: start migration to $nbd_uri");
|
|
PVE::QemuServer::qemu_drive_mirror($vmid, $drive, $nbd_uri, $vmid, undef, $self->{storage_migration_jobs}, 'skip', undef, $bwlimit, $bitmap);
|
|
|
|
$self->{volume_map}->{$source_volid} = $target_volid;
|
|
$self->log('info', "volume '$source_volid' is '$target_volid' on the target\n");
|
|
}
|
|
}
|
|
|
|
$self->log('info', "starting online/live migration on $ruri");
|
|
$self->{livemigration} = 1;
|
|
|
|
# load_defaults
|
|
my $defaults = PVE::QemuServer::load_defaults();
|
|
|
|
$self->log('info', "set migration_caps");
|
|
eval {
|
|
PVE::QemuServer::set_migration_caps($vmid);
|
|
};
|
|
warn $@ if $@;
|
|
|
|
my $qemu_migrate_params = {};
|
|
|
|
# migrate speed can be set via bwlimit (datacenter.cfg and API) and via the
|
|
# migrate_speed parameter in qm.conf - take the lower of the two.
|
|
my $bwlimit = PVE::Storage::get_bandwidth_limit('migration', undef, $opt_bwlimit) // 0;
|
|
my $migrate_speed = $conf->{migrate_speed} // $bwlimit;
|
|
# migrate_speed is in MB/s, bwlimit in KB/s
|
|
$migrate_speed *= 1024;
|
|
|
|
$migrate_speed = ($bwlimit < $migrate_speed) ? $bwlimit : $migrate_speed;
|
|
|
|
# always set migrate speed (overwrite kvm default of 32m) we set a very high
|
|
# default of 8192m which is basically unlimited
|
|
$migrate_speed ||= ($defaults->{migrate_speed} || 8192) * 1024;
|
|
|
|
# qmp takes migrate_speed in B/s.
|
|
$migrate_speed *= 1024;
|
|
$self->log('info', "migration speed limit: $migrate_speed B/s");
|
|
$qemu_migrate_params->{'max-bandwidth'} = int($migrate_speed);
|
|
|
|
my $migrate_downtime = $defaults->{migrate_downtime};
|
|
$migrate_downtime = $conf->{migrate_downtime} if defined($conf->{migrate_downtime});
|
|
# migrate-set-parameters expects limit in ms
|
|
$migrate_downtime *= 1000;
|
|
$self->log('info', "migration downtime limit: $migrate_downtime ms");
|
|
$qemu_migrate_params->{'downtime-limit'} = int($migrate_downtime);
|
|
|
|
# set cachesize to 10% of the total memory
|
|
my $memory = $conf->{memory} || $defaults->{memory};
|
|
my $cachesize = int($memory * 1048576 / 10);
|
|
$cachesize = round_powerof2($cachesize);
|
|
|
|
$self->log('info', "migration cachesize: $cachesize B");
|
|
$qemu_migrate_params->{'xbzrle-cache-size'} = int($cachesize);
|
|
|
|
$self->log('info', "set migration parameters");
|
|
eval {
|
|
mon_cmd($vmid, "migrate-set-parameters", %{$qemu_migrate_params});
|
|
};
|
|
$self->log('info', "migrate-set-parameters error: $@") if $@;
|
|
|
|
if (PVE::QemuServer::vga_conf_has_spice($conf->{vga})) {
|
|
my $rpcenv = PVE::RPCEnvironment::get();
|
|
my $authuser = $rpcenv->get_user();
|
|
|
|
my (undef, $proxyticket) = PVE::AccessControl::assemble_spice_ticket($authuser, $vmid, $self->{node});
|
|
|
|
my $filename = "/etc/pve/nodes/$self->{node}/pve-ssl.pem";
|
|
my $subject = PVE::AccessControl::read_x509_subject_spice($filename);
|
|
|
|
$self->log('info', "spice client_migrate_info");
|
|
|
|
eval {
|
|
mon_cmd($vmid, "client_migrate_info", protocol => 'spice',
|
|
hostname => $proxyticket, 'port' => 0, 'tls-port' => $spice_port,
|
|
'cert-subject' => $subject);
|
|
};
|
|
$self->log('info', "client_migrate_info error: $@") if $@;
|
|
|
|
}
|
|
|
|
$self->log('info', "start migrate command to $ruri");
|
|
eval {
|
|
mon_cmd($vmid, "migrate", uri => $ruri);
|
|
};
|
|
my $merr = $@;
|
|
$self->log('info', "migrate uri => $ruri failed: $merr") if $merr;
|
|
|
|
my $lstat = 0;
|
|
my $usleep = 1000000;
|
|
my $i = 0;
|
|
my $err_count = 0;
|
|
my $lastrem = undef;
|
|
my $downtimecounter = 0;
|
|
while (1) {
|
|
$i++;
|
|
my $avglstat = $lstat/$i if $lstat;
|
|
|
|
usleep($usleep);
|
|
my $stat;
|
|
eval {
|
|
$stat = mon_cmd($vmid, "query-migrate");
|
|
};
|
|
if (my $err = $@) {
|
|
$err_count++;
|
|
warn "query migrate failed: $err\n";
|
|
$self->log('info', "query migrate failed: $err");
|
|
if ($err_count <= 5) {
|
|
usleep(1000000);
|
|
next;
|
|
}
|
|
die "too many query migrate failures - aborting\n";
|
|
}
|
|
|
|
if (defined($stat->{status}) && $stat->{status} =~ m/^(setup)$/im) {
|
|
sleep(1);
|
|
next;
|
|
}
|
|
|
|
if (defined($stat->{status}) && $stat->{status} =~ m/^(active|completed|failed|cancelled)$/im) {
|
|
$merr = undef;
|
|
$err_count = 0;
|
|
if ($stat->{status} eq 'completed') {
|
|
my $delay = time() - $start;
|
|
if ($delay > 0) {
|
|
my $mbps = sprintf "%.2f", $memory / $delay;
|
|
my $downtime = $stat->{downtime} || 0;
|
|
$self->log('info', "migration speed: $mbps MB/s - downtime $downtime ms");
|
|
}
|
|
}
|
|
|
|
if ($stat->{status} eq 'failed' || $stat->{status} eq 'cancelled') {
|
|
$self->log('info', "migration status error: $stat->{status}");
|
|
die "aborting\n"
|
|
}
|
|
|
|
if ($stat->{status} ne 'active') {
|
|
$self->log('info', "migration status: $stat->{status}");
|
|
last;
|
|
}
|
|
|
|
if ($stat->{ram}->{transferred} ne $lstat) {
|
|
my $trans = $stat->{ram}->{transferred} || 0;
|
|
my $rem = $stat->{ram}->{remaining} || 0;
|
|
my $total = $stat->{ram}->{total} || 0;
|
|
my $xbzrlecachesize = $stat->{"xbzrle-cache"}->{"cache-size"} || 0;
|
|
my $xbzrlebytes = $stat->{"xbzrle-cache"}->{"bytes"} || 0;
|
|
my $xbzrlepages = $stat->{"xbzrle-cache"}->{"pages"} || 0;
|
|
my $xbzrlecachemiss = $stat->{"xbzrle-cache"}->{"cache-miss"} || 0;
|
|
my $xbzrleoverflow = $stat->{"xbzrle-cache"}->{"overflow"} || 0;
|
|
# reduce sleep if remainig memory is lower than the average transfer speed
|
|
$usleep = 100000 if $avglstat && $rem < $avglstat;
|
|
|
|
$self->log('info', "migration status: $stat->{status} (transferred ${trans}, " .
|
|
"remaining ${rem}), total ${total})");
|
|
|
|
if (${xbzrlecachesize}) {
|
|
$self->log('info', "migration xbzrle cachesize: ${xbzrlecachesize} transferred ${xbzrlebytes} pages ${xbzrlepages} cachemiss ${xbzrlecachemiss} overflow ${xbzrleoverflow}");
|
|
}
|
|
|
|
if (($lastrem && $rem > $lastrem ) || ($rem == 0)) {
|
|
$downtimecounter++;
|
|
}
|
|
$lastrem = $rem;
|
|
|
|
if ($downtimecounter > 5) {
|
|
$downtimecounter = 0;
|
|
$migrate_downtime *= 2;
|
|
$self->log('info', "auto-increased downtime to continue migration: $migrate_downtime ms");
|
|
eval {
|
|
# migrate-set-parameters does not touch values not
|
|
# specified, so this only changes downtime-limit
|
|
mon_cmd($vmid, "migrate-set-parameters", 'downtime-limit' => int($migrate_downtime));
|
|
};
|
|
$self->log('info', "migrate-set-parameters error: $@") if $@;
|
|
}
|
|
|
|
}
|
|
|
|
|
|
$lstat = $stat->{ram}->{transferred};
|
|
|
|
} else {
|
|
die $merr if $merr;
|
|
die "unable to parse migration status '$stat->{status}' - aborting\n";
|
|
}
|
|
}
|
|
}
|
|
|
|
sub phase2_cleanup {
|
|
my ($self, $vmid, $err) = @_;
|
|
|
|
return if !$self->{errors};
|
|
$self->{phase2errors} = 1;
|
|
|
|
$self->log('info', "aborting phase 2 - cleanup resources");
|
|
|
|
$self->log('info', "migrate_cancel");
|
|
eval {
|
|
mon_cmd($vmid, "migrate_cancel");
|
|
};
|
|
$self->log('info', "migrate_cancel error: $@") if $@;
|
|
|
|
my $conf = $self->{vmconf};
|
|
delete $conf->{lock};
|
|
eval { PVE::QemuConfig->write_config($vmid, $conf) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
}
|
|
|
|
# cleanup ressources on target host
|
|
if ($self->{storage_migration}) {
|
|
eval { PVE::QemuServer::qemu_blockjobs_cancel($vmid, $self->{storage_migration_jobs}) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
}
|
|
}
|
|
|
|
eval { $self->cleanup_bitmaps() };
|
|
if (my $err =$@) {
|
|
$self->log('err', $err);
|
|
}
|
|
|
|
my $nodename = PVE::INotify::nodename();
|
|
|
|
my $cmd = [@{$self->{rem_ssh}}, 'qm', 'stop', $vmid, '--skiplock', '--migratedfrom', $nodename];
|
|
eval{ PVE::Tools::run_command($cmd, outfunc => sub {}, errfunc => sub {}) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
|
|
# cleanup after stopping, otherwise disks might be in-use by target VM!
|
|
eval { PVE::QemuMigrate::cleanup_remotedisks($self) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
}
|
|
|
|
|
|
if ($self->{tunnel}) {
|
|
eval { finish_tunnel($self, $self->{tunnel}); };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
sub phase3 {
|
|
my ($self, $vmid) = @_;
|
|
|
|
my $volids = $self->{volumes};
|
|
return if $self->{phase2errors};
|
|
|
|
# destroy local copies
|
|
foreach my $volid (@$volids) {
|
|
eval { PVE::Storage::vdisk_free($self->{storecfg}, $volid); };
|
|
if (my $err = $@) {
|
|
$self->log('err', "removing local copy of '$volid' failed - $err");
|
|
$self->{errors} = 1;
|
|
last if $err =~ /^interrupted by signal$/;
|
|
}
|
|
}
|
|
}
|
|
|
|
sub phase3_cleanup {
|
|
my ($self, $vmid, $err) = @_;
|
|
|
|
my $conf = $self->{vmconf};
|
|
return if $self->{phase2errors};
|
|
|
|
my $tunnel = $self->{tunnel};
|
|
|
|
if ($self->{storage_migration}) {
|
|
# finish block-job with block-job-cancel, to disconnect source VM from NBD
|
|
# to avoid it trying to re-establish it. We are in blockjob ready state,
|
|
# thus, this command changes to it to blockjob complete (see qapi docs)
|
|
eval { PVE::QemuServer::qemu_drive_mirror_monitor($vmid, undef, $self->{storage_migration_jobs}, 'cancel'); };
|
|
|
|
if (my $err = $@) {
|
|
eval { PVE::QemuServer::qemu_blockjobs_cancel($vmid, $self->{storage_migration_jobs}) };
|
|
eval { PVE::QemuMigrate::cleanup_remotedisks($self) };
|
|
die "Failed to complete storage migration: $err\n";
|
|
}
|
|
}
|
|
|
|
if ($self->{volume_map}) {
|
|
my $target_drives = $self->{target_drive};
|
|
|
|
# FIXME: for NBD storage migration we now only update the volid, and
|
|
# not the full drivestr from the target node. Workaround that until we
|
|
# got some real rescan, to avoid things like wrong format in the drive
|
|
delete $conf->{$_} for keys %$target_drives;
|
|
PVE::QemuConfig->update_volume_ids($conf, $self->{volume_map});
|
|
|
|
for my $drive (keys %$target_drives) {
|
|
$conf->{$drive} = $target_drives->{$drive}->{drivestr};
|
|
}
|
|
PVE::QemuConfig->write_config($vmid, $conf);
|
|
}
|
|
|
|
# transfer replication state before move config
|
|
$self->transfer_replication_state() if $self->{replicated_volumes};
|
|
|
|
# move config to remote node
|
|
my $conffile = PVE::QemuConfig->config_file($vmid);
|
|
my $newconffile = PVE::QemuConfig->config_file($vmid, $self->{node});
|
|
|
|
die "Failed to move config to node '$self->{node}' - rename failed: $!\n"
|
|
if !rename($conffile, $newconffile);
|
|
|
|
$self->switch_replication_job_target() if $self->{replicated_volumes};
|
|
|
|
if ($self->{livemigration}) {
|
|
if ($self->{stopnbd}) {
|
|
$self->log('info', "stopping NBD storage migration server on target.");
|
|
# stop nbd server on remote vm - requirement for resume since 2.9
|
|
my $cmd = [@{$self->{rem_ssh}}, 'qm', 'nbdstop', $vmid];
|
|
|
|
eval{ PVE::Tools::run_command($cmd, outfunc => sub {}, errfunc => sub {}) };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
}
|
|
|
|
# config moved and nbd server stopped - now we can resume vm on target
|
|
if ($tunnel && $tunnel->{version} && $tunnel->{version} >= 1) {
|
|
eval {
|
|
$self->write_tunnel($tunnel, 30, "resume $vmid");
|
|
};
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
} else {
|
|
my $cmd = [@{$self->{rem_ssh}}, 'qm', 'resume', $vmid, '--skiplock', '--nocheck'];
|
|
my $logf = sub {
|
|
my $line = shift;
|
|
$self->log('err', $line);
|
|
};
|
|
eval { PVE::Tools::run_command($cmd, outfunc => sub {}, errfunc => $logf); };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
}
|
|
|
|
if ($self->{storage_migration} && PVE::QemuServer::parse_guest_agent($conf)->{fstrim_cloned_disks} && $self->{running}) {
|
|
my $cmd = [@{$self->{rem_ssh}}, 'qm', 'guest', 'cmd', $vmid, 'fstrim'];
|
|
eval{ PVE::Tools::run_command($cmd, outfunc => sub {}, errfunc => sub {}) };
|
|
}
|
|
}
|
|
|
|
# close tunnel on successful migration, on error phase2_cleanup closed it
|
|
if ($tunnel) {
|
|
eval { finish_tunnel($self, $tunnel); };
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
}
|
|
|
|
eval {
|
|
my $timer = 0;
|
|
if (PVE::QemuServer::vga_conf_has_spice($conf->{vga}) && $self->{running}) {
|
|
$self->log('info', "Waiting for spice server migration");
|
|
while (1) {
|
|
my $res = mon_cmd($vmid, 'query-spice');
|
|
last if int($res->{'migrated'}) == 1;
|
|
last if $timer > 50;
|
|
$timer ++;
|
|
usleep(200000);
|
|
}
|
|
}
|
|
};
|
|
|
|
# always stop local VM
|
|
eval { PVE::QemuServer::vm_stop($self->{storecfg}, $vmid, 1, 1); };
|
|
if (my $err = $@) {
|
|
$self->log('err', "stopping vm failed - $err");
|
|
$self->{errors} = 1;
|
|
}
|
|
|
|
# always deactivate volumes - avoid lvm LVs to be active on several nodes
|
|
eval {
|
|
my $vollist = PVE::QemuServer::get_vm_volumes($conf);
|
|
PVE::Storage::deactivate_volumes($self->{storecfg}, $vollist);
|
|
};
|
|
if (my $err = $@) {
|
|
$self->log('err', $err);
|
|
$self->{errors} = 1;
|
|
}
|
|
|
|
if($self->{storage_migration}) {
|
|
# destroy local copies
|
|
my $volids = $self->{online_local_volumes};
|
|
|
|
foreach my $volid (@$volids) {
|
|
# keep replicated volumes!
|
|
next if $self->{replicated_volumes}->{$volid};
|
|
|
|
eval { PVE::Storage::vdisk_free($self->{storecfg}, $volid); };
|
|
if (my $err = $@) {
|
|
$self->log('err', "removing local copy of '$volid' failed - $err");
|
|
$self->{errors} = 1;
|
|
last if $err =~ /^interrupted by signal$/;
|
|
}
|
|
}
|
|
|
|
}
|
|
|
|
# clear migrate lock
|
|
my $cmd = [ @{$self->{rem_ssh}}, 'qm', 'unlock', $vmid ];
|
|
$self->cmd_logerr($cmd, errmsg => "failed to clear migrate lock");
|
|
}
|
|
|
|
sub final_cleanup {
|
|
my ($self, $vmid) = @_;
|
|
|
|
# nothing to do
|
|
}
|
|
|
|
sub round_powerof2 {
|
|
return 1 if $_[0] < 2;
|
|
return 2 << int(log($_[0]-1)/log(2));
|
|
}
|
|
|
|
1;
|