mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
synced 2025-08-28 00:19:36 +00:00

A VMM may send a non-fatal signal to its threads, including vCPU tasks, at any time, and thus may signal vCPU tasks during KVM_RUN. If a vCPU task receives the signal while its trying to spawn the huge page recovery vhost task, then KVM_RUN will fail due to copy_process() returning -ERESTARTNOINTR. Rework call_once() to mark the call complete if and only if the called function succeeds, and plumb the function's true error code back to the call_once() invoker. This provides userspace with the correct, non-fatal error code so that the VMM doesn't terminate the VM on -ENOMEM, and allows subsequent KVM_RUN a succeed by virtue of retrying creation of the NX huge page task. Co-developed-by: Sean Christopherson <seanjc@google.com> Signed-off-by: Sean Christopherson <seanjc@google.com> [implemented the kvm user side] Signed-off-by: Keith Busch <kbusch@kernel.org> Message-ID: <20250227230631.303431-3-kbusch@meta.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
67 lines
1.7 KiB
C
67 lines
1.7 KiB
C
#ifndef _LINUX_CALL_ONCE_H
|
|
#define _LINUX_CALL_ONCE_H
|
|
|
|
#include <linux/types.h>
|
|
#include <linux/mutex.h>
|
|
|
|
#define ONCE_NOT_STARTED 0
|
|
#define ONCE_RUNNING 1
|
|
#define ONCE_COMPLETED 2
|
|
|
|
struct once {
|
|
atomic_t state;
|
|
struct mutex lock;
|
|
};
|
|
|
|
static inline void __once_init(struct once *once, const char *name,
|
|
struct lock_class_key *key)
|
|
{
|
|
atomic_set(&once->state, ONCE_NOT_STARTED);
|
|
__mutex_init(&once->lock, name, key);
|
|
}
|
|
|
|
#define once_init(once) \
|
|
do { \
|
|
static struct lock_class_key __key; \
|
|
__once_init((once), #once, &__key); \
|
|
} while (0)
|
|
|
|
/*
|
|
* call_once - Ensure a function has been called exactly once
|
|
*
|
|
* @once: Tracking struct
|
|
* @cb: Function to be called
|
|
*
|
|
* If @once has never completed successfully before, call @cb and, if
|
|
* it returns a zero or positive value, mark @once as completed. Return
|
|
* the value returned by @cb
|
|
*
|
|
* If @once has completed succesfully before, return 0.
|
|
*
|
|
* The call to @cb is implicitly surrounded by a mutex, though for
|
|
* efficiency the * function avoids taking it after the first call.
|
|
*/
|
|
static inline int call_once(struct once *once, int (*cb)(struct once *))
|
|
{
|
|
int r, state;
|
|
|
|
/* Pairs with atomic_set_release() below. */
|
|
if (atomic_read_acquire(&once->state) == ONCE_COMPLETED)
|
|
return 0;
|
|
|
|
guard(mutex)(&once->lock);
|
|
state = atomic_read(&once->state);
|
|
if (unlikely(state != ONCE_NOT_STARTED))
|
|
return WARN_ON_ONCE(state != ONCE_COMPLETED) ? -EINVAL : 0;
|
|
|
|
atomic_set(&once->state, ONCE_RUNNING);
|
|
r = cb(once);
|
|
if (r < 0)
|
|
atomic_set(&once->state, ONCE_NOT_STARTED);
|
|
else
|
|
atomic_set_release(&once->state, ONCE_COMPLETED);
|
|
return r;
|
|
}
|
|
|
|
#endif /* _LINUX_CALL_ONCE_H */
|