linux/drivers/gpu/drm/xe/xe_pt_walk.h
Matthew Brost ae482ec8cd
drm/xe: Add staging tree for VM binds
Concurrent VM bind staging and zapping of PTEs from a userptr notifier
do not work because the view of PTEs is not stable. VM binds cannot
acquire the notifier lock during staging, as memory allocations are
required. To resolve this race condition, use a staging tree for VM
binds that is committed only under the userptr notifier lock during the
final step of the bind. This ensures a consistent view of the PTEs in
the userptr notifier.

A follow up may only use staging for VM in fault mode as this is the
only mode in which the above race exists.

v3:
 - Drop zap PTE change (Thomas)
 - s/xe_pt_entry/xe_pt_entry_staging (Thomas)

Suggested-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Cc: <stable@vger.kernel.org>
Fixes: e8babb280b ("drm/xe: Convert multiple bind ops into single job")
Fixes: a708f6501c ("drm/xe: Update PT layer with better error handling")
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
Reviewed-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20250228073058.59510-5-thomas.hellstrom@linux.intel.com
Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
(cherry picked from commit 6f39b0c5ef)
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
2025-03-05 14:25:11 -05:00

153 lines
4.5 KiB
C

/* SPDX-License-Identifier: GPL-2.0-only */
/*
* Copyright © 2022 Intel Corporation
*/
#ifndef __XE_PT_WALK__
#define __XE_PT_WALK__
#include <linux/pagewalk.h>
#include <linux/types.h>
/**
* struct xe_ptw - base class for driver pagetable subclassing.
* @children: Pointer to an array of children if any.
* @staging: Pointer to an array of staging if any.
*
* Drivers could subclass this, and if it's a page-directory, typically
* embed an array of xe_ptw pointers.
*/
struct xe_ptw {
struct xe_ptw **children;
struct xe_ptw **staging;
};
/**
* struct xe_pt_walk - Embeddable struct for walk parameters
*/
struct xe_pt_walk {
/** @ops: The walk ops used for the pagewalk */
const struct xe_pt_walk_ops *ops;
/**
* @shifts: Array of page-table entry shifts used for the
* different levels, starting out with the leaf level 0
* page-shift as the first entry. It's legal for this pointer to be
* changed during the walk.
*/
const u64 *shifts;
/** @max_level: Highest populated level in @sizes */
unsigned int max_level;
/**
* @shared_pt_mode: Whether to skip all entries that are private
* to the address range and called only for entries that are
* shared with other address ranges. Such entries are referred to
* as shared pagetables.
*/
bool shared_pt_mode;
/** @staging: Walk staging PT structure */
bool staging;
};
/**
* typedef xe_pt_entry_fn - gpu page-table-walk callback-function
* @parent: The parent page.table.
* @offset: The offset (number of entries) into the page table.
* @level: The level of @parent.
* @addr: The virtual address.
* @next: The virtual address for the next call, or end address.
* @child: Pointer to pointer to child page-table at this @offset. The
* function may modify the value pointed to if, for example, allocating a
* child page table.
* @action: The walk action to take upon return. See <linux/pagewalk.h>.
* @walk: The walk parameters.
*/
typedef int (*xe_pt_entry_fn)(struct xe_ptw *parent, pgoff_t offset,
unsigned int level, u64 addr, u64 next,
struct xe_ptw **child,
enum page_walk_action *action,
struct xe_pt_walk *walk);
/**
* struct xe_pt_walk_ops - Walk callbacks.
*/
struct xe_pt_walk_ops {
/**
* @pt_entry: Callback to be called for each page table entry prior
* to descending to the next level. The returned value of the action
* function parameter is honored.
*/
xe_pt_entry_fn pt_entry;
/**
* @pt_post_descend: Callback to be called for each page table entry
* after return from descending to the next level. The returned value
* of the action function parameter is ignored.
*/
xe_pt_entry_fn pt_post_descend;
};
int xe_pt_walk_range(struct xe_ptw *parent, unsigned int level,
u64 addr, u64 end, struct xe_pt_walk *walk);
int xe_pt_walk_shared(struct xe_ptw *parent, unsigned int level,
u64 addr, u64 end, struct xe_pt_walk *walk);
/**
* xe_pt_covers - Whether the address range covers an entire entry in @level
* @addr: Start of the range.
* @end: End of range + 1.
* @level: Page table level.
* @walk: Page table walk info.
*
* This function is a helper to aid in determining whether a leaf page table
* entry can be inserted at this @level.
*
* Return: Whether the range provided covers exactly an entry at this level.
*/
static inline bool xe_pt_covers(u64 addr, u64 end, unsigned int level,
const struct xe_pt_walk *walk)
{
u64 pt_size = 1ull << walk->shifts[level];
return end - addr == pt_size && IS_ALIGNED(addr, pt_size);
}
/**
* xe_pt_num_entries: Number of page-table entries of a given range at this
* level
* @addr: Start address.
* @end: End address.
* @level: Page table level.
* @walk: Walk info.
*
* Return: The number of page table entries at this level between @start and
* @end.
*/
static inline pgoff_t
xe_pt_num_entries(u64 addr, u64 end, unsigned int level,
const struct xe_pt_walk *walk)
{
u64 pt_size = 1ull << walk->shifts[level];
return (round_up(end, pt_size) - round_down(addr, pt_size)) >>
walk->shifts[level];
}
/**
* xe_pt_offset: Offset of the page-table entry for a given address.
* @addr: The address.
* @level: Page table level.
* @walk: Walk info.
*
* Return: The page table entry offset for the given address in a
* page table with size indicated by @level.
*/
static inline pgoff_t
xe_pt_offset(u64 addr, unsigned int level, const struct xe_pt_walk *walk)
{
if (level < walk->max_level)
addr &= ((1ull << walk->shifts[level + 1]) - 1);
return addr >> walk->shifts[level];
}
#endif