mirror of
https://git.proxmox.com/git/mirror_ubuntu-kernels.git
synced 2025-11-18 15:59:35 +00:00
- Add support for building the kernel using PC-relative addressing on Power10.
- Allow HV KVM guests on Power10 to use prefixed instructions.
- Unify support for the P2020 CPU (85xx) into a single machine description.
- Always build the 64-bit kernel with 128-bit long double.
- Drop support for several obsolete 2000's era development boards as
identified by Paul Gortmaker.
- A series fixing VFIO on Power since some generic changes.
- Various other small features and fixes.
Thanks to: Alexey Kardashevskiy, Andrew Donnellan, Benjamin Gray, Bo Liu,
Christophe Leroy, Dan Carpenter, David Binderman, Ira Weiny, Joel Stanley,
Kajol Jain, Kautuk Consul, Liang He, Luis Chamberlain, Masahiro Yamada, Michael
Neuling, Nathan Chancellor, Nathan Lynch, Nicholas Miehlbradt, Nicholas Piggin,
Nick Desaulniers, Nysal Jan K.A, Pali Rohár, Paul Gortmaker, Paul Mackerras,
Petr Vaněk, Randy Dunlap, Rob Herring, Sachin Sant, Sean Christopherson, Segher
Boessenkool, Timothy Pearson.
-----BEGIN PGP SIGNATURE-----
iQJHBAABCAAxFiEEJFGtCPCthwEv2Y/bUevqPMjhpYAFAmRLdD8THG1wZUBlbGxl
cm1hbi5pZC5hdQAKCRBR6+o8yOGlgPrED/93Hwi1/8udXYV6OAFBnLLLDX9RZWNx
sj6W7PC/yY7MGUTIGcayrAURt5xFfQZMBdq9nhhH46Wyd8pSUe1IlpXIEgqeH64Y
5uaSe6u4OZ5dDmiYz8bM+Y4Ixkfq1xMO0Rj27FIRJyU4Pp6gyMQ6/8W+iPU2vYIb
jl4frVO8PpmCbW8euOyT/b9YB2h79+5nLgZT4RvmYJblKQwgzRZWaiCAU0wYf+xT
xYsMQcqJlPstizTnXIr+GC08VrMPQR51kJCurnNUMXoAQY6toEXebveWRNZ3sH39
K0BRQ036NNGPS4GlHVbjgLIdoWr4pUEROZ48Jy9WeiDh6OwO2vb2zrm7ZLtKlGXI
LCQ08T9diPbAAJyVJaKjpsNXhTffuLPRhIOr1o2vqGvY+Fqbw96bPQAyFbxpJtrw
rGTTc5e93fEdZV+eaGR8kfdNM75WM6lgiteaIbUnpirYTh/0j6YEO1Ijc4d9e5ux
aoHN2eXQDjMm9foZf1D6vaCTRyN8LwLa9hcydKCn6+qULUQzoZ2r4cRt6eSD7+fx
Wni1Zg7vD6xx294nVmhg5dWuD4qVIAZj3BZPFHHR2SPqy+UbEJLk/NKgznmrhhgQ
HBcZAEFqIBZkA4e0w6LJKh/1j1rxpZUokgjMotOJq4VRivq82W1P7SioFDY3/6w5
UvGtIgGq4Qsa2Q==
=89WH
-----END PGP SIGNATURE-----
Merge tag 'powerpc-6.4-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux
Pull powerpc updates from Michael Ellerman:
- Add support for building the kernel using PC-relative addressing on
Power10.
- Allow HV KVM guests on Power10 to use prefixed instructions.
- Unify support for the P2020 CPU (85xx) into a single machine
description.
- Always build the 64-bit kernel with 128-bit long double.
- Drop support for several obsolete 2000's era development boards as
identified by Paul Gortmaker.
- A series fixing VFIO on Power since some generic changes.
- Various other small features and fixes.
Thanks to Alexey Kardashevskiy, Andrew Donnellan, Benjamin Gray, Bo Liu,
Christophe Leroy, Dan Carpenter, David Binderman, Ira Weiny, Joel
Stanley, Kajol Jain, Kautuk Consul, Liang He, Luis Chamberlain, Masahiro
Yamada, Michael Neuling, Nathan Chancellor, Nathan Lynch, Nicholas
Miehlbradt, Nicholas Piggin, Nick Desaulniers, Nysal Jan K.A, Pali
Rohár, Paul Gortmaker, Paul Mackerras, Petr Vaněk, Randy Dunlap, Rob
Herring, Sachin Sant, Sean Christopherson, Segher Boessenkool, and
Timothy Pearson.
* tag 'powerpc-6.4-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux: (156 commits)
powerpc/64s: Disable pcrel code model on Clang
powerpc: Fix merge conflict between pcrel and copy_thread changes
powerpc/configs/powernv: Add IGB=y
powerpc/configs/64s: Drop JFS Filesystem
powerpc/configs/64s: Use EXT4 to mount EXT2 filesystems
powerpc/configs: Make pseries_defconfig an alias for ppc64le_guest
powerpc/configs: Make pseries_le an alias for ppc64le_guest
powerpc/configs: Incorporate generic kvm_guest.config into guest configs
powerpc/configs: Add IBMVETH=y and IBMVNIC=y to guest configs
powerpc/configs/64s: Enable Device Mapper options
powerpc/configs/64s: Enable PSTORE
powerpc/configs/64s: Enable VLAN support
powerpc/configs/64s: Enable BLK_DEV_NVME
powerpc/configs/64s: Drop REISERFS
powerpc/configs/64s: Use SHA512 for module signatures
powerpc/configs/64s: Enable IO_STRICT_DEVMEM
powerpc/configs/64s: Enable SCHEDSTATS
powerpc/configs/64s: Enable DEBUG_VM & other options
powerpc/configs/64s: Enable EMULATED_STATS
powerpc/configs/64s: Enable KUNIT and most tests
...
830 lines
19 KiB
C
830 lines
19 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Support for Partition Mobility/Migration
|
|
*
|
|
* Copyright (C) 2010 Nathan Fontenot
|
|
* Copyright (C) 2010 IBM Corporation
|
|
*/
|
|
|
|
|
|
#define pr_fmt(fmt) "mobility: " fmt
|
|
|
|
#include <linux/cpu.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kobject.h>
|
|
#include <linux/nmi.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/stat.h>
|
|
#include <linux/stop_machine.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/device.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/stringify.h>
|
|
|
|
#include <asm/machdep.h>
|
|
#include <asm/rtas.h>
|
|
#include "pseries.h"
|
|
#include "vas.h" /* vas_migration_handler() */
|
|
#include "../../kernel/cacheinfo.h"
|
|
|
|
static struct kobject *mobility_kobj;
|
|
|
|
struct update_props_workarea {
|
|
__be32 phandle;
|
|
__be32 state;
|
|
__be64 reserved;
|
|
__be32 nprops;
|
|
} __packed;
|
|
|
|
#define NODE_ACTION_MASK 0xff000000
|
|
#define NODE_COUNT_MASK 0x00ffffff
|
|
|
|
#define DELETE_DT_NODE 0x01000000
|
|
#define UPDATE_DT_NODE 0x02000000
|
|
#define ADD_DT_NODE 0x03000000
|
|
|
|
#define MIGRATION_SCOPE (1)
|
|
#define PRRN_SCOPE -2
|
|
|
|
#ifdef CONFIG_PPC_WATCHDOG
|
|
static unsigned int nmi_wd_lpm_factor = 200;
|
|
|
|
#ifdef CONFIG_SYSCTL
|
|
static struct ctl_table nmi_wd_lpm_factor_ctl_table[] = {
|
|
{
|
|
.procname = "nmi_wd_lpm_factor",
|
|
.data = &nmi_wd_lpm_factor,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = proc_douintvec_minmax,
|
|
},
|
|
{}
|
|
};
|
|
|
|
static int __init register_nmi_wd_lpm_factor_sysctl(void)
|
|
{
|
|
register_sysctl("kernel", nmi_wd_lpm_factor_ctl_table);
|
|
|
|
return 0;
|
|
}
|
|
device_initcall(register_nmi_wd_lpm_factor_sysctl);
|
|
#endif /* CONFIG_SYSCTL */
|
|
#endif /* CONFIG_PPC_WATCHDOG */
|
|
|
|
static int mobility_rtas_call(int token, char *buf, s32 scope)
|
|
{
|
|
int rc;
|
|
|
|
spin_lock(&rtas_data_buf_lock);
|
|
|
|
memcpy(rtas_data_buf, buf, RTAS_DATA_BUF_SIZE);
|
|
rc = rtas_call(token, 2, 1, NULL, rtas_data_buf, scope);
|
|
memcpy(buf, rtas_data_buf, RTAS_DATA_BUF_SIZE);
|
|
|
|
spin_unlock(&rtas_data_buf_lock);
|
|
return rc;
|
|
}
|
|
|
|
static int delete_dt_node(struct device_node *dn)
|
|
{
|
|
struct device_node *pdn;
|
|
bool is_platfac;
|
|
|
|
pdn = of_get_parent(dn);
|
|
is_platfac = of_node_is_type(dn, "ibm,platform-facilities") ||
|
|
of_node_is_type(pdn, "ibm,platform-facilities");
|
|
of_node_put(pdn);
|
|
|
|
/*
|
|
* The drivers that bind to nodes in the platform-facilities
|
|
* hierarchy don't support node removal, and the removal directive
|
|
* from firmware is always followed by an add of an equivalent
|
|
* node. The capability (e.g. RNG, encryption, compression)
|
|
* represented by the node is never interrupted by the migration.
|
|
* So ignore changes to this part of the tree.
|
|
*/
|
|
if (is_platfac) {
|
|
pr_notice("ignoring remove operation for %pOFfp\n", dn);
|
|
return 0;
|
|
}
|
|
|
|
pr_debug("removing node %pOFfp\n", dn);
|
|
dlpar_detach_node(dn);
|
|
return 0;
|
|
}
|
|
|
|
static int update_dt_property(struct device_node *dn, struct property **prop,
|
|
const char *name, u32 vd, char *value)
|
|
{
|
|
struct property *new_prop = *prop;
|
|
int more = 0;
|
|
|
|
/* A negative 'vd' value indicates that only part of the new property
|
|
* value is contained in the buffer and we need to call
|
|
* ibm,update-properties again to get the rest of the value.
|
|
*
|
|
* A negative value is also the two's compliment of the actual value.
|
|
*/
|
|
if (vd & 0x80000000) {
|
|
vd = ~vd + 1;
|
|
more = 1;
|
|
}
|
|
|
|
if (new_prop) {
|
|
/* partial property fixup */
|
|
char *new_data = kzalloc(new_prop->length + vd, GFP_KERNEL);
|
|
if (!new_data)
|
|
return -ENOMEM;
|
|
|
|
memcpy(new_data, new_prop->value, new_prop->length);
|
|
memcpy(new_data + new_prop->length, value, vd);
|
|
|
|
kfree(new_prop->value);
|
|
new_prop->value = new_data;
|
|
new_prop->length += vd;
|
|
} else {
|
|
new_prop = kzalloc(sizeof(*new_prop), GFP_KERNEL);
|
|
if (!new_prop)
|
|
return -ENOMEM;
|
|
|
|
new_prop->name = kstrdup(name, GFP_KERNEL);
|
|
if (!new_prop->name) {
|
|
kfree(new_prop);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
new_prop->length = vd;
|
|
new_prop->value = kzalloc(new_prop->length, GFP_KERNEL);
|
|
if (!new_prop->value) {
|
|
kfree(new_prop->name);
|
|
kfree(new_prop);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memcpy(new_prop->value, value, vd);
|
|
*prop = new_prop;
|
|
}
|
|
|
|
if (!more) {
|
|
pr_debug("updating node %pOF property %s\n", dn, name);
|
|
of_update_property(dn, new_prop);
|
|
*prop = NULL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int update_dt_node(struct device_node *dn, s32 scope)
|
|
{
|
|
struct update_props_workarea *upwa;
|
|
struct property *prop = NULL;
|
|
int i, rc, rtas_rc;
|
|
char *prop_data;
|
|
char *rtas_buf;
|
|
int update_properties_token;
|
|
u32 nprops;
|
|
u32 vd;
|
|
|
|
update_properties_token = rtas_function_token(RTAS_FN_IBM_UPDATE_PROPERTIES);
|
|
if (update_properties_token == RTAS_UNKNOWN_SERVICE)
|
|
return -EINVAL;
|
|
|
|
rtas_buf = kzalloc(RTAS_DATA_BUF_SIZE, GFP_KERNEL);
|
|
if (!rtas_buf)
|
|
return -ENOMEM;
|
|
|
|
upwa = (struct update_props_workarea *)&rtas_buf[0];
|
|
upwa->phandle = cpu_to_be32(dn->phandle);
|
|
|
|
do {
|
|
rtas_rc = mobility_rtas_call(update_properties_token, rtas_buf,
|
|
scope);
|
|
if (rtas_rc < 0)
|
|
break;
|
|
|
|
prop_data = rtas_buf + sizeof(*upwa);
|
|
nprops = be32_to_cpu(upwa->nprops);
|
|
|
|
/* On the first call to ibm,update-properties for a node the
|
|
* first property value descriptor contains an empty
|
|
* property name, the property value length encoded as u32,
|
|
* and the property value is the node path being updated.
|
|
*/
|
|
if (*prop_data == 0) {
|
|
prop_data++;
|
|
vd = be32_to_cpu(*(__be32 *)prop_data);
|
|
prop_data += vd + sizeof(vd);
|
|
nprops--;
|
|
}
|
|
|
|
for (i = 0; i < nprops; i++) {
|
|
char *prop_name;
|
|
|
|
prop_name = prop_data;
|
|
prop_data += strlen(prop_name) + 1;
|
|
vd = be32_to_cpu(*(__be32 *)prop_data);
|
|
prop_data += sizeof(vd);
|
|
|
|
switch (vd) {
|
|
case 0x00000000:
|
|
/* name only property, nothing to do */
|
|
break;
|
|
|
|
case 0x80000000:
|
|
of_remove_property(dn, of_find_property(dn,
|
|
prop_name, NULL));
|
|
prop = NULL;
|
|
break;
|
|
|
|
default:
|
|
rc = update_dt_property(dn, &prop, prop_name,
|
|
vd, prop_data);
|
|
if (rc) {
|
|
pr_err("updating %s property failed: %d\n",
|
|
prop_name, rc);
|
|
}
|
|
|
|
prop_data += vd;
|
|
break;
|
|
}
|
|
|
|
cond_resched();
|
|
}
|
|
|
|
cond_resched();
|
|
} while (rtas_rc == 1);
|
|
|
|
kfree(rtas_buf);
|
|
return 0;
|
|
}
|
|
|
|
static int add_dt_node(struct device_node *parent_dn, __be32 drc_index)
|
|
{
|
|
struct device_node *dn;
|
|
int rc;
|
|
|
|
dn = dlpar_configure_connector(drc_index, parent_dn);
|
|
if (!dn)
|
|
return -ENOENT;
|
|
|
|
/*
|
|
* Since delete_dt_node() ignores this node type, this is the
|
|
* necessary counterpart. We also know that a platform-facilities
|
|
* node returned from dlpar_configure_connector() has children
|
|
* attached, and dlpar_attach_node() only adds the parent, leaking
|
|
* the children. So ignore these on the add side for now.
|
|
*/
|
|
if (of_node_is_type(dn, "ibm,platform-facilities")) {
|
|
pr_notice("ignoring add operation for %pOF\n", dn);
|
|
dlpar_free_cc_nodes(dn);
|
|
return 0;
|
|
}
|
|
|
|
rc = dlpar_attach_node(dn, parent_dn);
|
|
if (rc)
|
|
dlpar_free_cc_nodes(dn);
|
|
|
|
pr_debug("added node %pOFfp\n", dn);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int pseries_devicetree_update(s32 scope)
|
|
{
|
|
char *rtas_buf;
|
|
__be32 *data;
|
|
int update_nodes_token;
|
|
int rc;
|
|
|
|
update_nodes_token = rtas_function_token(RTAS_FN_IBM_UPDATE_NODES);
|
|
if (update_nodes_token == RTAS_UNKNOWN_SERVICE)
|
|
return 0;
|
|
|
|
rtas_buf = kzalloc(RTAS_DATA_BUF_SIZE, GFP_KERNEL);
|
|
if (!rtas_buf)
|
|
return -ENOMEM;
|
|
|
|
do {
|
|
rc = mobility_rtas_call(update_nodes_token, rtas_buf, scope);
|
|
if (rc && rc != 1)
|
|
break;
|
|
|
|
data = (__be32 *)rtas_buf + 4;
|
|
while (be32_to_cpu(*data) & NODE_ACTION_MASK) {
|
|
int i;
|
|
u32 action = be32_to_cpu(*data) & NODE_ACTION_MASK;
|
|
u32 node_count = be32_to_cpu(*data) & NODE_COUNT_MASK;
|
|
|
|
data++;
|
|
|
|
for (i = 0; i < node_count; i++) {
|
|
struct device_node *np;
|
|
__be32 phandle = *data++;
|
|
__be32 drc_index;
|
|
|
|
np = of_find_node_by_phandle(be32_to_cpu(phandle));
|
|
if (!np) {
|
|
pr_warn("Failed lookup: phandle 0x%x for action 0x%x\n",
|
|
be32_to_cpu(phandle), action);
|
|
continue;
|
|
}
|
|
|
|
switch (action) {
|
|
case DELETE_DT_NODE:
|
|
delete_dt_node(np);
|
|
break;
|
|
case UPDATE_DT_NODE:
|
|
update_dt_node(np, scope);
|
|
break;
|
|
case ADD_DT_NODE:
|
|
drc_index = *data++;
|
|
add_dt_node(np, drc_index);
|
|
break;
|
|
}
|
|
|
|
of_node_put(np);
|
|
cond_resched();
|
|
}
|
|
}
|
|
|
|
cond_resched();
|
|
} while (rc == 1);
|
|
|
|
kfree(rtas_buf);
|
|
return rc;
|
|
}
|
|
|
|
void post_mobility_fixup(void)
|
|
{
|
|
int rc;
|
|
|
|
rtas_activate_firmware();
|
|
|
|
/*
|
|
* We don't want CPUs to go online/offline while the device
|
|
* tree is being updated.
|
|
*/
|
|
cpus_read_lock();
|
|
|
|
/*
|
|
* It's common for the destination firmware to replace cache
|
|
* nodes. Release all of the cacheinfo hierarchy's references
|
|
* before updating the device tree.
|
|
*/
|
|
cacheinfo_teardown();
|
|
|
|
rc = pseries_devicetree_update(MIGRATION_SCOPE);
|
|
if (rc)
|
|
pr_err("device tree update failed: %d\n", rc);
|
|
|
|
cacheinfo_rebuild();
|
|
|
|
cpus_read_unlock();
|
|
|
|
/* Possibly switch to a new L1 flush type */
|
|
pseries_setup_security_mitigations();
|
|
|
|
/* Reinitialise system information for hv-24x7 */
|
|
read_24x7_sys_info();
|
|
|
|
return;
|
|
}
|
|
|
|
static int poll_vasi_state(u64 handle, unsigned long *res)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long hvrc;
|
|
int ret;
|
|
|
|
hvrc = plpar_hcall(H_VASI_STATE, retbuf, handle);
|
|
switch (hvrc) {
|
|
case H_SUCCESS:
|
|
ret = 0;
|
|
*res = retbuf[0];
|
|
break;
|
|
case H_PARAMETER:
|
|
ret = -EINVAL;
|
|
break;
|
|
case H_FUNCTION:
|
|
ret = -EOPNOTSUPP;
|
|
break;
|
|
case H_HARDWARE:
|
|
default:
|
|
pr_err("unexpected H_VASI_STATE result %ld\n", hvrc);
|
|
ret = -EIO;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int wait_for_vasi_session_suspending(u64 handle)
|
|
{
|
|
unsigned long state;
|
|
int ret;
|
|
|
|
/*
|
|
* Wait for transition from H_VASI_ENABLED to
|
|
* H_VASI_SUSPENDING. Treat anything else as an error.
|
|
*/
|
|
while (true) {
|
|
ret = poll_vasi_state(handle, &state);
|
|
|
|
if (ret != 0 || state == H_VASI_SUSPENDING) {
|
|
break;
|
|
} else if (state == H_VASI_ENABLED) {
|
|
ssleep(1);
|
|
} else {
|
|
pr_err("unexpected H_VASI_STATE result %lu\n", state);
|
|
ret = -EIO;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Proceed even if H_VASI_STATE is unavailable. If H_JOIN or
|
|
* ibm,suspend-me are also unimplemented, we'll recover then.
|
|
*/
|
|
if (ret == -EOPNOTSUPP)
|
|
ret = 0;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void wait_for_vasi_session_completed(u64 handle)
|
|
{
|
|
unsigned long state = 0;
|
|
int ret;
|
|
|
|
pr_info("waiting for memory transfer to complete...\n");
|
|
|
|
/*
|
|
* Wait for transition from H_VASI_RESUMED to H_VASI_COMPLETED.
|
|
*/
|
|
while (true) {
|
|
ret = poll_vasi_state(handle, &state);
|
|
|
|
/*
|
|
* If the memory transfer is already complete and the migration
|
|
* has been cleaned up by the hypervisor, H_PARAMETER is return,
|
|
* which is translate in EINVAL by poll_vasi_state().
|
|
*/
|
|
if (ret == -EINVAL || (!ret && state == H_VASI_COMPLETED)) {
|
|
pr_info("memory transfer completed.\n");
|
|
break;
|
|
}
|
|
|
|
if (ret) {
|
|
pr_err("H_VASI_STATE return error (%d)\n", ret);
|
|
break;
|
|
}
|
|
|
|
if (state != H_VASI_RESUMED) {
|
|
pr_err("unexpected H_VASI_STATE result %lu\n", state);
|
|
break;
|
|
}
|
|
|
|
msleep(500);
|
|
}
|
|
}
|
|
|
|
static void prod_single(unsigned int target_cpu)
|
|
{
|
|
long hvrc;
|
|
int hwid;
|
|
|
|
hwid = get_hard_smp_processor_id(target_cpu);
|
|
hvrc = plpar_hcall_norets(H_PROD, hwid);
|
|
if (hvrc == H_SUCCESS)
|
|
return;
|
|
pr_err_ratelimited("H_PROD of CPU %u (hwid %d) error: %ld\n",
|
|
target_cpu, hwid, hvrc);
|
|
}
|
|
|
|
static void prod_others(void)
|
|
{
|
|
unsigned int cpu;
|
|
|
|
for_each_online_cpu(cpu) {
|
|
if (cpu != smp_processor_id())
|
|
prod_single(cpu);
|
|
}
|
|
}
|
|
|
|
static u16 clamp_slb_size(void)
|
|
{
|
|
#ifdef CONFIG_PPC_64S_HASH_MMU
|
|
u16 prev = mmu_slb_size;
|
|
|
|
slb_set_size(SLB_MIN_SIZE);
|
|
|
|
return prev;
|
|
#else
|
|
return 0;
|
|
#endif
|
|
}
|
|
|
|
static int do_suspend(void)
|
|
{
|
|
u16 saved_slb_size;
|
|
int status;
|
|
int ret;
|
|
|
|
pr_info("calling ibm,suspend-me on CPU %i\n", smp_processor_id());
|
|
|
|
/*
|
|
* The destination processor model may have fewer SLB entries
|
|
* than the source. We reduce mmu_slb_size to a safe minimum
|
|
* before suspending in order to minimize the possibility of
|
|
* programming non-existent entries on the destination. If
|
|
* suspend fails, we restore it before returning. On success
|
|
* the OF reconfig path will update it from the new device
|
|
* tree after resuming on the destination.
|
|
*/
|
|
saved_slb_size = clamp_slb_size();
|
|
|
|
ret = rtas_ibm_suspend_me(&status);
|
|
if (ret != 0) {
|
|
pr_err("ibm,suspend-me error: %d\n", status);
|
|
slb_set_size(saved_slb_size);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* struct pseries_suspend_info - State shared between CPUs for join/suspend.
|
|
* @counter: Threads are to increment this upon resuming from suspend
|
|
* or if an error is received from H_JOIN. The thread which performs
|
|
* the first increment (i.e. sets it to 1) is responsible for
|
|
* waking the other threads.
|
|
* @done: False if join/suspend is in progress. True if the operation is
|
|
* complete (successful or not).
|
|
*/
|
|
struct pseries_suspend_info {
|
|
atomic_t counter;
|
|
bool done;
|
|
};
|
|
|
|
static int do_join(void *arg)
|
|
{
|
|
struct pseries_suspend_info *info = arg;
|
|
atomic_t *counter = &info->counter;
|
|
long hvrc;
|
|
int ret;
|
|
|
|
retry:
|
|
/* Must ensure MSR.EE off for H_JOIN. */
|
|
hard_irq_disable();
|
|
hvrc = plpar_hcall_norets(H_JOIN);
|
|
|
|
switch (hvrc) {
|
|
case H_CONTINUE:
|
|
/*
|
|
* All other CPUs are offline or in H_JOIN. This CPU
|
|
* attempts the suspend.
|
|
*/
|
|
ret = do_suspend();
|
|
break;
|
|
case H_SUCCESS:
|
|
/*
|
|
* The suspend is complete and this cpu has received a
|
|
* prod, or we've received a stray prod from unrelated
|
|
* code (e.g. paravirt spinlocks) and we need to join
|
|
* again.
|
|
*
|
|
* This barrier orders the return from H_JOIN above vs
|
|
* the load of info->done. It pairs with the barrier
|
|
* in the wakeup/prod path below.
|
|
*/
|
|
smp_mb();
|
|
if (READ_ONCE(info->done) == false) {
|
|
pr_info_ratelimited("premature return from H_JOIN on CPU %i, retrying",
|
|
smp_processor_id());
|
|
goto retry;
|
|
}
|
|
ret = 0;
|
|
break;
|
|
case H_BAD_MODE:
|
|
case H_HARDWARE:
|
|
default:
|
|
ret = -EIO;
|
|
pr_err_ratelimited("H_JOIN error %ld on CPU %i\n",
|
|
hvrc, smp_processor_id());
|
|
break;
|
|
}
|
|
|
|
if (atomic_inc_return(counter) == 1) {
|
|
pr_info("CPU %u waking all threads\n", smp_processor_id());
|
|
WRITE_ONCE(info->done, true);
|
|
/*
|
|
* This barrier orders the store to info->done vs subsequent
|
|
* H_PRODs to wake the other CPUs. It pairs with the barrier
|
|
* in the H_SUCCESS case above.
|
|
*/
|
|
smp_mb();
|
|
prod_others();
|
|
}
|
|
/*
|
|
* Execution may have been suspended for several seconds, so reset
|
|
* the watchdogs. touch_nmi_watchdog() also touches the soft lockup
|
|
* watchdog.
|
|
*/
|
|
rcu_cpu_stall_reset();
|
|
touch_nmi_watchdog();
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Abort reason code byte 0. We use only the 'Migrating partition' value.
|
|
*/
|
|
enum vasi_aborting_entity {
|
|
ORCHESTRATOR = 1,
|
|
VSP_SOURCE = 2,
|
|
PARTITION_FIRMWARE = 3,
|
|
PLATFORM_FIRMWARE = 4,
|
|
VSP_TARGET = 5,
|
|
MIGRATING_PARTITION = 6,
|
|
};
|
|
|
|
static void pseries_cancel_migration(u64 handle, int err)
|
|
{
|
|
u32 reason_code;
|
|
u32 detail;
|
|
u8 entity;
|
|
long hvrc;
|
|
|
|
entity = MIGRATING_PARTITION;
|
|
detail = abs(err) & 0xffffff;
|
|
reason_code = (entity << 24) | detail;
|
|
|
|
hvrc = plpar_hcall_norets(H_VASI_SIGNAL, handle,
|
|
H_VASI_SIGNAL_CANCEL, reason_code);
|
|
if (hvrc)
|
|
pr_err("H_VASI_SIGNAL error: %ld\n", hvrc);
|
|
}
|
|
|
|
static int pseries_suspend(u64 handle)
|
|
{
|
|
const unsigned int max_attempts = 5;
|
|
unsigned int retry_interval_ms = 1;
|
|
unsigned int attempt = 1;
|
|
int ret;
|
|
|
|
while (true) {
|
|
struct pseries_suspend_info info;
|
|
unsigned long vasi_state;
|
|
int vasi_err;
|
|
|
|
info = (struct pseries_suspend_info) {
|
|
.counter = ATOMIC_INIT(0),
|
|
.done = false,
|
|
};
|
|
|
|
ret = stop_machine(do_join, &info, cpu_online_mask);
|
|
if (ret == 0)
|
|
break;
|
|
/*
|
|
* Encountered an error. If the VASI stream is still
|
|
* in Suspending state, it's likely a transient
|
|
* condition related to some device in the partition
|
|
* and we can retry in the hope that the cause has
|
|
* cleared after some delay.
|
|
*
|
|
* A better design would allow drivers etc to prepare
|
|
* for the suspend and avoid conditions which prevent
|
|
* the suspend from succeeding. For now, we have this
|
|
* mitigation.
|
|
*/
|
|
pr_notice("Partition suspend attempt %u of %u error: %d\n",
|
|
attempt, max_attempts, ret);
|
|
|
|
if (attempt == max_attempts)
|
|
break;
|
|
|
|
vasi_err = poll_vasi_state(handle, &vasi_state);
|
|
if (vasi_err == 0) {
|
|
if (vasi_state != H_VASI_SUSPENDING) {
|
|
pr_notice("VASI state %lu after failed suspend\n",
|
|
vasi_state);
|
|
break;
|
|
}
|
|
} else if (vasi_err != -EOPNOTSUPP) {
|
|
pr_err("VASI state poll error: %d", vasi_err);
|
|
break;
|
|
}
|
|
|
|
pr_notice("Will retry partition suspend after %u ms\n",
|
|
retry_interval_ms);
|
|
|
|
msleep(retry_interval_ms);
|
|
retry_interval_ms *= 10;
|
|
attempt++;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int pseries_migrate_partition(u64 handle)
|
|
{
|
|
int ret;
|
|
unsigned int factor = 0;
|
|
|
|
#ifdef CONFIG_PPC_WATCHDOG
|
|
factor = nmi_wd_lpm_factor;
|
|
#endif
|
|
/*
|
|
* When the migration is initiated, the hypervisor changes VAS
|
|
* mappings to prepare before OS gets the notification and
|
|
* closes all VAS windows. NX generates continuous faults during
|
|
* this time and the user space can not differentiate these
|
|
* faults from the migration event. So reduce this time window
|
|
* by closing VAS windows at the beginning of this function.
|
|
*/
|
|
vas_migration_handler(VAS_SUSPEND);
|
|
|
|
ret = wait_for_vasi_session_suspending(handle);
|
|
if (ret)
|
|
goto out;
|
|
|
|
if (factor)
|
|
watchdog_nmi_set_timeout_pct(factor);
|
|
|
|
ret = pseries_suspend(handle);
|
|
if (ret == 0) {
|
|
post_mobility_fixup();
|
|
/*
|
|
* Wait until the memory transfer is complete, so that the user
|
|
* space process returns from the syscall after the transfer is
|
|
* complete. This allows the user hooks to be executed at the
|
|
* right time.
|
|
*/
|
|
wait_for_vasi_session_completed(handle);
|
|
} else
|
|
pseries_cancel_migration(handle, ret);
|
|
|
|
if (factor)
|
|
watchdog_nmi_set_timeout_pct(0);
|
|
|
|
out:
|
|
vas_migration_handler(VAS_RESUME);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int rtas_syscall_dispatch_ibm_suspend_me(u64 handle)
|
|
{
|
|
return pseries_migrate_partition(handle);
|
|
}
|
|
|
|
static ssize_t migration_store(const struct class *class,
|
|
const struct class_attribute *attr, const char *buf,
|
|
size_t count)
|
|
{
|
|
u64 streamid;
|
|
int rc;
|
|
|
|
rc = kstrtou64(buf, 0, &streamid);
|
|
if (rc)
|
|
return rc;
|
|
|
|
rc = pseries_migrate_partition(streamid);
|
|
if (rc)
|
|
return rc;
|
|
|
|
return count;
|
|
}
|
|
|
|
/*
|
|
* Used by drmgr to determine the kernel behavior of the migration interface.
|
|
*
|
|
* Version 1: Performs all PAPR requirements for migration including
|
|
* firmware activation and device tree update.
|
|
*/
|
|
#define MIGRATION_API_VERSION 1
|
|
|
|
static CLASS_ATTR_WO(migration);
|
|
static CLASS_ATTR_STRING(api_version, 0444, __stringify(MIGRATION_API_VERSION));
|
|
|
|
static int __init mobility_sysfs_init(void)
|
|
{
|
|
int rc;
|
|
|
|
mobility_kobj = kobject_create_and_add("mobility", kernel_kobj);
|
|
if (!mobility_kobj)
|
|
return -ENOMEM;
|
|
|
|
rc = sysfs_create_file(mobility_kobj, &class_attr_migration.attr);
|
|
if (rc)
|
|
pr_err("unable to create migration sysfs file (%d)\n", rc);
|
|
|
|
rc = sysfs_create_file(mobility_kobj, &class_attr_api_version.attr.attr);
|
|
if (rc)
|
|
pr_err("unable to create api_version sysfs file (%d)\n", rc);
|
|
|
|
return 0;
|
|
}
|
|
machine_device_initcall(pseries, mobility_sysfs_init);
|