mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
synced 2025-09-01 15:14:52 +00:00

Reduce system call overhead time (round trip time for invoking a non-existent system call) by 25%. With the removal of set_fs() [1] lazy control register handling was removed in order to keep kernel entry and exit simple. However this made system calls slower. With the conversion to generic entry [2] and numerous follow up changes which simplified the entry code significantly, adding support for lazy asce handling doesn't add much complexity to the entry code anymore. In particular this means: - On kernel entry the primary asce is not modified and contains the user asce - Kernel accesses which require secondary-space mode (for example futex operations) are surrounded by enable_sacf_uaccess() and disable_sacf_uaccess() calls. enable_sacf_uaccess() sets the primary asce to kernel asce so that the sacf instruction can be used to switch to secondary-space mode. The primary asce is changed back to user asce with disable_sacf_uaccess(). The state of the control register which contains the primary asce is reflected with a new TIF_ASCE_PRIMARY bit. This is required on context switch so that the correct asce is restored for the scheduled in process. In result address spaces are now setup like this: CPU running in | %cr1 ASCE | %cr7 ASCE | %cr13 ASCE -----------------------------|-----------|-----------|----------- user space | user | user | kernel kernel (no sacf) | user | user | kernel kernel (during sacf uaccess) | kernel | user | kernel kernel (kvm guest execution) | guest | user | kernel In result cr1 control register content is not changed except for: - futex system calls - legacy s390 PCI system calls - the kvm specific cmpxchg_user_key() uaccess helper This leads to faster system call execution. [1]87d5986345
("s390/mm: remove set_fs / rework address space handling") [2]56e62a7370
("s390: convert to generic entry") Reviewed-by: Alexander Gordeev <agordeev@linux.ibm.com> Signed-off-by: Heiko Carstens <hca@linux.ibm.com>
103 lines
2.8 KiB
C
103 lines
2.8 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _ASM_S390_FUTEX_H
|
|
#define _ASM_S390_FUTEX_H
|
|
|
|
#include <linux/instrumented.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/futex.h>
|
|
#include <asm/asm-extable.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/errno.h>
|
|
|
|
#define FUTEX_OP_FUNC(name, insn) \
|
|
static uaccess_kmsan_or_inline int \
|
|
__futex_atomic_##name(int oparg, int *old, u32 __user *uaddr) \
|
|
{ \
|
|
bool sacf_flag; \
|
|
int rc, new; \
|
|
\
|
|
instrument_copy_from_user_before(old, uaddr, sizeof(*old)); \
|
|
sacf_flag = enable_sacf_uaccess(); \
|
|
asm_inline volatile( \
|
|
" sacf 256\n" \
|
|
"0: l %[old],%[uaddr]\n" \
|
|
"1:"insn \
|
|
"2: cs %[old],%[new],%[uaddr]\n" \
|
|
"3: jl 1b\n" \
|
|
" lhi %[rc],0\n" \
|
|
"4: sacf 768\n" \
|
|
EX_TABLE_UA_FAULT(0b, 4b, %[rc]) \
|
|
EX_TABLE_UA_FAULT(1b, 4b, %[rc]) \
|
|
EX_TABLE_UA_FAULT(2b, 4b, %[rc]) \
|
|
EX_TABLE_UA_FAULT(3b, 4b, %[rc]) \
|
|
: [rc] "=d" (rc), [old] "=&d" (*old), \
|
|
[new] "=&d" (new), [uaddr] "+Q" (*uaddr) \
|
|
: [oparg] "d" (oparg) \
|
|
: "cc"); \
|
|
disable_sacf_uaccess(sacf_flag); \
|
|
if (!rc) \
|
|
instrument_copy_from_user_after(old, uaddr, sizeof(*old), 0); \
|
|
return rc; \
|
|
}
|
|
|
|
FUTEX_OP_FUNC(set, "lr %[new],%[oparg]\n")
|
|
FUTEX_OP_FUNC(add, "lr %[new],%[old]\n ar %[new],%[oparg]\n")
|
|
FUTEX_OP_FUNC(or, "lr %[new],%[old]\n or %[new],%[oparg]\n")
|
|
FUTEX_OP_FUNC(and, "lr %[new],%[old]\n nr %[new],%[oparg]\n")
|
|
FUTEX_OP_FUNC(xor, "lr %[new],%[old]\n xr %[new],%[oparg]\n")
|
|
|
|
static inline
|
|
int arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
|
|
{
|
|
int old, rc;
|
|
|
|
switch (op) {
|
|
case FUTEX_OP_SET:
|
|
rc = __futex_atomic_set(oparg, &old, uaddr);
|
|
break;
|
|
case FUTEX_OP_ADD:
|
|
rc = __futex_atomic_add(oparg, &old, uaddr);
|
|
break;
|
|
case FUTEX_OP_OR:
|
|
rc = __futex_atomic_or(oparg, &old, uaddr);
|
|
break;
|
|
case FUTEX_OP_ANDN:
|
|
rc = __futex_atomic_and(~oparg, &old, uaddr);
|
|
break;
|
|
case FUTEX_OP_XOR:
|
|
rc = __futex_atomic_xor(oparg, &old, uaddr);
|
|
break;
|
|
default:
|
|
rc = -ENOSYS;
|
|
}
|
|
if (!rc)
|
|
*oval = old;
|
|
return rc;
|
|
}
|
|
|
|
static uaccess_kmsan_or_inline
|
|
int futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, u32 oldval, u32 newval)
|
|
{
|
|
bool sacf_flag;
|
|
int rc;
|
|
|
|
instrument_copy_from_user_before(uval, uaddr, sizeof(*uval));
|
|
sacf_flag = enable_sacf_uaccess();
|
|
asm_inline volatile(
|
|
" sacf 256\n"
|
|
"0: cs %[old],%[new],%[uaddr]\n"
|
|
"1: lhi %[rc],0\n"
|
|
"2: sacf 768\n"
|
|
EX_TABLE_UA_FAULT(0b, 2b, %[rc])
|
|
EX_TABLE_UA_FAULT(1b, 2b, %[rc])
|
|
: [rc] "=d" (rc), [old] "+d" (oldval), [uaddr] "+Q" (*uaddr)
|
|
: [new] "d" (newval)
|
|
: "cc", "memory");
|
|
disable_sacf_uaccess(sacf_flag);
|
|
*uval = oldval;
|
|
instrument_copy_from_user_after(uval, uaddr, sizeof(*uval), 0);
|
|
return rc;
|
|
}
|
|
|
|
#endif /* _ASM_S390_FUTEX_H */
|