mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
synced 2025-09-10 10:08:21 +00:00
drm/amdgpu/userq: add helpers to start/stop scheduling
This will be used to stop/start user queue scheduling for example when switching between kernel and user queues when enforce isolation is enabled. v2: use idx v3: only stop compute/gfx queues Reviewed-by: Sunil Khatri <sunil.khatri@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
parent
56a0a80af0
commit
94976e7e5e
@ -1249,6 +1249,7 @@ struct amdgpu_device {
|
|||||||
|
|
||||||
struct list_head userq_mgr_list;
|
struct list_head userq_mgr_list;
|
||||||
struct mutex userq_mutex;
|
struct mutex userq_mutex;
|
||||||
|
bool userq_halt_for_enforce_isolation;
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline uint32_t amdgpu_ip_version(const struct amdgpu_device *adev,
|
static inline uint32_t amdgpu_ip_version(const struct amdgpu_device *adev,
|
||||||
|
@ -275,6 +275,7 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args)
|
|||||||
const struct amdgpu_userq_funcs *uq_funcs;
|
const struct amdgpu_userq_funcs *uq_funcs;
|
||||||
struct amdgpu_usermode_queue *queue;
|
struct amdgpu_usermode_queue *queue;
|
||||||
struct amdgpu_db_info db_info;
|
struct amdgpu_db_info db_info;
|
||||||
|
bool skip_map_queue;
|
||||||
uint64_t index;
|
uint64_t index;
|
||||||
int qid, r = 0;
|
int qid, r = 0;
|
||||||
|
|
||||||
@ -348,6 +349,7 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args)
|
|||||||
goto unlock;
|
goto unlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
qid = idr_alloc(&uq_mgr->userq_idr, queue, 1, AMDGPU_MAX_USERQ_COUNT, GFP_KERNEL);
|
qid = idr_alloc(&uq_mgr->userq_idr, queue, 1, AMDGPU_MAX_USERQ_COUNT, GFP_KERNEL);
|
||||||
if (qid < 0) {
|
if (qid < 0) {
|
||||||
DRM_ERROR("Failed to allocate a queue id\n");
|
DRM_ERROR("Failed to allocate a queue id\n");
|
||||||
@ -358,15 +360,28 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args)
|
|||||||
goto unlock;
|
goto unlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
r = uq_funcs->map(uq_mgr, queue);
|
/* don't map the queue if scheduling is halted */
|
||||||
if (r) {
|
mutex_lock(&adev->userq_mutex);
|
||||||
DRM_ERROR("Failed to map Queue\n");
|
if (adev->userq_halt_for_enforce_isolation &&
|
||||||
idr_remove(&uq_mgr->userq_idr, qid);
|
((queue->queue_type == AMDGPU_HW_IP_GFX) ||
|
||||||
amdgpu_userq_fence_driver_free(queue);
|
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)))
|
||||||
uq_funcs->mqd_destroy(uq_mgr, queue);
|
skip_map_queue = true;
|
||||||
kfree(queue);
|
else
|
||||||
goto unlock;
|
skip_map_queue = false;
|
||||||
|
if (!skip_map_queue) {
|
||||||
|
r = uq_funcs->map(uq_mgr, queue);
|
||||||
|
if (r) {
|
||||||
|
mutex_unlock(&adev->userq_mutex);
|
||||||
|
DRM_ERROR("Failed to map Queue\n");
|
||||||
|
idr_remove(&uq_mgr->userq_idr, qid);
|
||||||
|
amdgpu_userq_fence_driver_free(queue);
|
||||||
|
uq_funcs->mqd_destroy(uq_mgr, queue);
|
||||||
|
kfree(queue);
|
||||||
|
goto unlock;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
mutex_unlock(&adev->userq_mutex);
|
||||||
|
|
||||||
|
|
||||||
args->out.queue_id = qid;
|
args->out.queue_id = qid;
|
||||||
|
|
||||||
@ -733,3 +748,58 @@ int amdgpu_userq_resume(struct amdgpu_device *adev)
|
|||||||
mutex_unlock(&adev->userq_mutex);
|
mutex_unlock(&adev->userq_mutex);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int amdgpu_userq_stop_sched_for_enforce_isolation(struct amdgpu_device *adev,
|
||||||
|
u32 idx)
|
||||||
|
{
|
||||||
|
const struct amdgpu_userq_funcs *userq_funcs;
|
||||||
|
struct amdgpu_usermode_queue *queue;
|
||||||
|
struct amdgpu_userq_mgr *uqm, *tmp;
|
||||||
|
int queue_id;
|
||||||
|
int ret = 0;
|
||||||
|
|
||||||
|
mutex_lock(&adev->userq_mutex);
|
||||||
|
if (adev->userq_halt_for_enforce_isolation)
|
||||||
|
dev_warn(adev->dev, "userq scheduling already stopped!\n");
|
||||||
|
adev->userq_halt_for_enforce_isolation = true;
|
||||||
|
list_for_each_entry_safe(uqm, tmp, &adev->userq_mgr_list, list) {
|
||||||
|
cancel_delayed_work_sync(&uqm->resume_work);
|
||||||
|
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
|
||||||
|
if (((queue->queue_type == AMDGPU_HW_IP_GFX) ||
|
||||||
|
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)) &&
|
||||||
|
(queue->xcp_id == idx)) {
|
||||||
|
userq_funcs = adev->userq_funcs[queue->queue_type];
|
||||||
|
ret |= userq_funcs->unmap(uqm, queue);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
mutex_unlock(&adev->userq_mutex);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
int amdgpu_userq_start_sched_for_enforce_isolation(struct amdgpu_device *adev,
|
||||||
|
u32 idx)
|
||||||
|
{
|
||||||
|
const struct amdgpu_userq_funcs *userq_funcs;
|
||||||
|
struct amdgpu_usermode_queue *queue;
|
||||||
|
struct amdgpu_userq_mgr *uqm, *tmp;
|
||||||
|
int queue_id;
|
||||||
|
int ret = 0;
|
||||||
|
|
||||||
|
mutex_lock(&adev->userq_mutex);
|
||||||
|
if (!adev->userq_halt_for_enforce_isolation)
|
||||||
|
dev_warn(adev->dev, "userq scheduling already started!\n");
|
||||||
|
adev->userq_halt_for_enforce_isolation = false;
|
||||||
|
list_for_each_entry_safe(uqm, tmp, &adev->userq_mgr_list, list) {
|
||||||
|
idr_for_each_entry(&uqm->userq_idr, queue, queue_id) {
|
||||||
|
if (((queue->queue_type == AMDGPU_HW_IP_GFX) ||
|
||||||
|
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)) &&
|
||||||
|
(queue->xcp_id == idx)) {
|
||||||
|
userq_funcs = adev->userq_funcs[queue->queue_type];
|
||||||
|
ret |= userq_funcs->map(uqm, queue);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
mutex_unlock(&adev->userq_mutex);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
@ -115,4 +115,9 @@ uint64_t amdgpu_userqueue_get_doorbell_index(struct amdgpu_userq_mgr *uq_mgr,
|
|||||||
int amdgpu_userq_suspend(struct amdgpu_device *adev);
|
int amdgpu_userq_suspend(struct amdgpu_device *adev);
|
||||||
int amdgpu_userq_resume(struct amdgpu_device *adev);
|
int amdgpu_userq_resume(struct amdgpu_device *adev);
|
||||||
|
|
||||||
|
int amdgpu_userq_stop_sched_for_enforce_isolation(struct amdgpu_device *adev,
|
||||||
|
u32 idx);
|
||||||
|
int amdgpu_userq_start_sched_for_enforce_isolation(struct amdgpu_device *adev,
|
||||||
|
u32 idx);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
Loading…
Reference in New Issue
Block a user