mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2025-08-27 10:34:13 +00:00

Implement the EU stall sampling APIs to read() and poll() EU stall data. A work function periodically polls the EU stall data buffer write pointer registers to look for any new data and caches the write pointer. The read function compares the cached read and write pointers and copies any new data to the user space. v11: Used gt->eu_stall->stream_lock instead of stream->buf_lock. Removed read and write offsets from trace and added read size. Moved workqueue from struct xe_eu_stall_data_stream to struct xe_eu_stall_gt. v10: Used cancel_delayed_work_sync() instead of flush_delayed_work() Replaced per xecore lock with a lock for all the xecore buffers Code movement and optimizations as per review feedback v9: New patch split from the previous patch. Used *_delayed_work functions instead of hrtimer Addressed the review feedback in read and poll functions Reviewed-by: Ashutosh Dixit <ashutosh.dixit@intel.com> Signed-off-by: Harish Chegondi <harish.chegondi@intel.com> Signed-off-by: Ashutosh Dixit <ashutosh.dixit@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/369dee85a3b6bd2c08aeae89ca55e66a9a0242d2.1740533885.git.harish.chegondi@intel.com
468 lines
12 KiB
C
468 lines
12 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Copyright © 2022 Intel Corporation
|
|
*/
|
|
|
|
#undef TRACE_SYSTEM
|
|
#define TRACE_SYSTEM xe
|
|
|
|
#if !defined(_XE_TRACE_H_) || defined(TRACE_HEADER_MULTI_READ)
|
|
#define _XE_TRACE_H_
|
|
|
|
#include <linux/tracepoint.h>
|
|
#include <linux/types.h>
|
|
|
|
#include "xe_exec_queue_types.h"
|
|
#include "xe_gpu_scheduler_types.h"
|
|
#include "xe_gt_tlb_invalidation_types.h"
|
|
#include "xe_gt_types.h"
|
|
#include "xe_guc_exec_queue_types.h"
|
|
#include "xe_sched_job.h"
|
|
#include "xe_vm.h"
|
|
|
|
#define __dev_name_xe(xe) dev_name((xe)->drm.dev)
|
|
#define __dev_name_tile(tile) __dev_name_xe(tile_to_xe((tile)))
|
|
#define __dev_name_gt(gt) __dev_name_xe(gt_to_xe((gt)))
|
|
#define __dev_name_eq(q) __dev_name_gt((q)->gt)
|
|
|
|
DECLARE_EVENT_CLASS(xe_gt_tlb_invalidation_fence,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_xe(xe))
|
|
__field(struct xe_gt_tlb_invalidation_fence *, fence)
|
|
__field(int, seqno)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->fence = fence;
|
|
__entry->seqno = fence->seqno;
|
|
),
|
|
|
|
TP_printk("dev=%s, fence=%p, seqno=%d",
|
|
__get_str(dev), __entry->fence, __entry->seqno)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence, xe_gt_tlb_invalidation_fence_create,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence,
|
|
xe_gt_tlb_invalidation_fence_work_func,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence, xe_gt_tlb_invalidation_fence_cb,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence, xe_gt_tlb_invalidation_fence_send,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence, xe_gt_tlb_invalidation_fence_recv,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence, xe_gt_tlb_invalidation_fence_signal,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_gt_tlb_invalidation_fence, xe_gt_tlb_invalidation_fence_timeout,
|
|
TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_invalidation_fence *fence),
|
|
TP_ARGS(xe, fence)
|
|
);
|
|
|
|
DECLARE_EVENT_CLASS(xe_exec_queue,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_eq(q))
|
|
__field(enum xe_engine_class, class)
|
|
__field(u32, logical_mask)
|
|
__field(u8, gt_id)
|
|
__field(u16, width)
|
|
__field(u16, guc_id)
|
|
__field(u32, guc_state)
|
|
__field(u32, flags)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->class = q->class;
|
|
__entry->logical_mask = q->logical_mask;
|
|
__entry->gt_id = q->gt->info.id;
|
|
__entry->width = q->width;
|
|
__entry->guc_id = q->guc->id;
|
|
__entry->guc_state = atomic_read(&q->guc->state);
|
|
__entry->flags = q->flags;
|
|
),
|
|
|
|
TP_printk("dev=%s, %d:0x%x, gt=%d, width=%d, guc_id=%d, guc_state=0x%x, flags=0x%x",
|
|
__get_str(dev), __entry->class, __entry->logical_mask,
|
|
__entry->gt_id, __entry->width, __entry->guc_id,
|
|
__entry->guc_state, __entry->flags)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_create,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_supress_resume,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_submit,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_scheduling_enable,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_scheduling_disable,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_scheduling_done,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_register,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_deregister,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_deregister_done,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_close,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_kill,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_cleanup_entity,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_destroy,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_reset,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_memory_cat_error,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_stop,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_resubmit,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_exec_queue, xe_exec_queue_lr_cleanup,
|
|
TP_PROTO(struct xe_exec_queue *q),
|
|
TP_ARGS(q)
|
|
);
|
|
|
|
DECLARE_EVENT_CLASS(xe_sched_job,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_eq(job->q))
|
|
__field(u32, seqno)
|
|
__field(u32, lrc_seqno)
|
|
__field(u8, gt_id)
|
|
__field(u16, guc_id)
|
|
__field(u32, guc_state)
|
|
__field(u32, flags)
|
|
__field(int, error)
|
|
__field(struct dma_fence *, fence)
|
|
__field(u64, batch_addr)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->seqno = xe_sched_job_seqno(job);
|
|
__entry->lrc_seqno = xe_sched_job_lrc_seqno(job);
|
|
__entry->gt_id = job->q->gt->info.id;
|
|
__entry->guc_id = job->q->guc->id;
|
|
__entry->guc_state =
|
|
atomic_read(&job->q->guc->state);
|
|
__entry->flags = job->q->flags;
|
|
__entry->error = job->fence ? job->fence->error : 0;
|
|
__entry->fence = job->fence;
|
|
__entry->batch_addr = (u64)job->ptrs[0].batch_addr;
|
|
),
|
|
|
|
TP_printk("dev=%s, fence=%p, seqno=%u, lrc_seqno=%u, gt=%u, guc_id=%d, batch_addr=0x%012llx, guc_state=0x%x, flags=0x%x, error=%d",
|
|
__get_str(dev), __entry->fence, __entry->seqno,
|
|
__entry->lrc_seqno, __entry->gt_id, __entry->guc_id,
|
|
__entry->batch_addr, __entry->guc_state,
|
|
__entry->flags, __entry->error)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_create,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_exec,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_run,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_free,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_timedout,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_set_error,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_job, xe_sched_job_ban,
|
|
TP_PROTO(struct xe_sched_job *job),
|
|
TP_ARGS(job)
|
|
);
|
|
|
|
DECLARE_EVENT_CLASS(xe_sched_msg,
|
|
TP_PROTO(struct xe_sched_msg *msg),
|
|
TP_ARGS(msg),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_eq(((struct xe_exec_queue *)msg->private_data)))
|
|
__field(u32, opcode)
|
|
__field(u16, guc_id)
|
|
__field(u8, gt_id)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->opcode = msg->opcode;
|
|
__entry->guc_id =
|
|
((struct xe_exec_queue *)msg->private_data)->guc->id;
|
|
__entry->gt_id =
|
|
((struct xe_exec_queue *)msg->private_data)->gt->info.id;
|
|
),
|
|
|
|
TP_printk("dev=%s, gt=%u guc_id=%d, opcode=%u", __get_str(dev), __entry->gt_id, __entry->guc_id,
|
|
__entry->opcode)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_msg, xe_sched_msg_add,
|
|
TP_PROTO(struct xe_sched_msg *msg),
|
|
TP_ARGS(msg)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_sched_msg, xe_sched_msg_recv,
|
|
TP_PROTO(struct xe_sched_msg *msg),
|
|
TP_ARGS(msg)
|
|
);
|
|
|
|
DECLARE_EVENT_CLASS(xe_hw_fence,
|
|
TP_PROTO(struct xe_hw_fence *fence),
|
|
TP_ARGS(fence),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_xe(fence->xe))
|
|
__field(u64, ctx)
|
|
__field(u32, seqno)
|
|
__field(struct xe_hw_fence *, fence)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->ctx = fence->dma.context;
|
|
__entry->seqno = fence->dma.seqno;
|
|
__entry->fence = fence;
|
|
),
|
|
|
|
TP_printk("dev=%s, ctx=0x%016llx, fence=%p, seqno=%u",
|
|
__get_str(dev), __entry->ctx, __entry->fence, __entry->seqno)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_hw_fence, xe_hw_fence_create,
|
|
TP_PROTO(struct xe_hw_fence *fence),
|
|
TP_ARGS(fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_hw_fence, xe_hw_fence_signal,
|
|
TP_PROTO(struct xe_hw_fence *fence),
|
|
TP_ARGS(fence)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_hw_fence, xe_hw_fence_try_signal,
|
|
TP_PROTO(struct xe_hw_fence *fence),
|
|
TP_ARGS(fence)
|
|
);
|
|
|
|
TRACE_EVENT(xe_reg_rw,
|
|
TP_PROTO(struct xe_mmio *mmio, bool write, u32 reg, u64 val, int len),
|
|
|
|
TP_ARGS(mmio, write, reg, val, len),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_tile(mmio->tile))
|
|
__field(u64, val)
|
|
__field(u32, reg)
|
|
__field(u16, write)
|
|
__field(u16, len)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->val = val;
|
|
__entry->reg = reg;
|
|
__entry->write = write;
|
|
__entry->len = len;
|
|
),
|
|
|
|
TP_printk("dev=%s, %s reg=0x%x, len=%d, val=(0x%x, 0x%x)",
|
|
__get_str(dev), __entry->write ? "write" : "read",
|
|
__entry->reg, __entry->len,
|
|
(u32)(__entry->val & 0xffffffff),
|
|
(u32)(__entry->val >> 32))
|
|
);
|
|
|
|
DECLARE_EVENT_CLASS(xe_pm_runtime,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller),
|
|
|
|
TP_STRUCT__entry(
|
|
__string(dev, __dev_name_xe(xe))
|
|
__field(void *, caller)
|
|
),
|
|
|
|
TP_fast_assign(
|
|
__assign_str(dev);
|
|
__entry->caller = caller;
|
|
),
|
|
|
|
TP_printk("dev=%s caller_function=%pS", __get_str(dev), __entry->caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_runtime_get,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_runtime_put,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_resume,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_suspend,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_runtime_resume,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_runtime_suspend,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
DEFINE_EVENT(xe_pm_runtime, xe_pm_runtime_get_ioctl,
|
|
TP_PROTO(struct xe_device *xe, void *caller),
|
|
TP_ARGS(xe, caller)
|
|
);
|
|
|
|
TRACE_EVENT(xe_eu_stall_data_read,
|
|
TP_PROTO(u8 slice, u8 subslice,
|
|
u32 read_ptr, u32 write_ptr,
|
|
size_t read_size, size_t total_size),
|
|
TP_ARGS(slice, subslice,
|
|
read_ptr, write_ptr,
|
|
read_size, total_size),
|
|
|
|
TP_STRUCT__entry(__field(u8, slice)
|
|
__field(u8, subslice)
|
|
__field(u32, read_ptr)
|
|
__field(u32, write_ptr)
|
|
__field(size_t, read_size)
|
|
__field(size_t, total_size)
|
|
),
|
|
|
|
TP_fast_assign(__entry->slice = slice;
|
|
__entry->subslice = subslice;
|
|
__entry->read_ptr = read_ptr;
|
|
__entry->write_ptr = write_ptr;
|
|
__entry->read_size = read_size;
|
|
__entry->total_size = total_size;
|
|
),
|
|
|
|
TP_printk("slice: %u subslice: %u read ptr: 0x%x write ptr: 0x%x read size: %zu total read size: %zu",
|
|
__entry->slice, __entry->subslice,
|
|
__entry->read_ptr, __entry->write_ptr,
|
|
__entry->read_size, __entry->total_size)
|
|
);
|
|
|
|
#endif
|
|
|
|
/* This part must be outside protection */
|
|
#undef TRACE_INCLUDE_PATH
|
|
#undef TRACE_INCLUDE_FILE
|
|
#define TRACE_INCLUDE_PATH ../../drivers/gpu/drm/xe
|
|
#define TRACE_INCLUDE_FILE xe_trace
|
|
#include <trace/define_trace.h>
|