On Thu, 2025-04-24 at 10:38 +0200, Pierre-Eric Pelloux-Prayer wrote: > Its only purpose was for trace events, but jobs can already be > uniquely identified using their fence. > > The downside of using the fence is that it's only available > after 'drm_sched_job_arm' was called which is true for all trace > events that used job.id so they can safely switch to using it.
nit: in the title you use double colon :: as a namespace separator. In the kernel docu style only . (as here above) or -> are used. Commit title should be consistent with that. Other than that, nice clean up. P. > > Suggested-by: Tvrtko Ursulin <tursu...@igalia.com> > Signed-off-by: Pierre-Eric Pelloux-Prayer > <pierre-eric.pelloux-pra...@amd.com> > Reviewed-by: Christian König <christian.koe...@amd.com> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h | 18 ++++++---------- > -- > .../gpu/drm/scheduler/gpu_scheduler_trace.h | 18 ++++++---------- > -- > drivers/gpu/drm/scheduler/sched_main.c | 1 - > include/drm/gpu_scheduler.h | 3 --- > 4 files changed, 12 insertions(+), 28 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > index 11dd2e0f7979..4fd810cb5387 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h > @@ -167,7 +167,6 @@ TRACE_EVENT(amdgpu_cs_ioctl, > TP_PROTO(struct amdgpu_job *job), > TP_ARGS(job), > TP_STRUCT__entry( > - __field(uint64_t, sched_job_id) > __string(timeline, > AMDGPU_JOB_GET_TIMELINE_NAME(job)) > __field(unsigned int, context) > __field(unsigned int, seqno) > @@ -177,15 +176,14 @@ TRACE_EVENT(amdgpu_cs_ioctl, > ), > > TP_fast_assign( > - __entry->sched_job_id = job->base.id; > __assign_str(timeline); > __entry->context = job->base.s_fence- > >finished.context; > __entry->seqno = job->base.s_fence- > >finished.seqno; > __assign_str(ring); > __entry->num_ibs = job->num_ibs; > ), > - TP_printk("sched_job=%llu, timeline=%s, context=%u, > seqno=%u, ring_name=%s, num_ibs=%u", > - __entry->sched_job_id, __get_str(timeline), > __entry->context, > + TP_printk("timeline=%s, context=%u, seqno=%u, > ring_name=%s, num_ibs=%u", > + __get_str(timeline), __entry->context, > __entry->seqno, __get_str(ring), __entry- > >num_ibs) > ); > > @@ -193,7 +191,6 @@ TRACE_EVENT(amdgpu_sched_run_job, > TP_PROTO(struct amdgpu_job *job), > TP_ARGS(job), > TP_STRUCT__entry( > - __field(uint64_t, sched_job_id) > __string(timeline, > AMDGPU_JOB_GET_TIMELINE_NAME(job)) > __field(unsigned int, context) > __field(unsigned int, seqno) > @@ -202,15 +199,14 @@ TRACE_EVENT(amdgpu_sched_run_job, > ), > > TP_fast_assign( > - __entry->sched_job_id = job->base.id; > __assign_str(timeline); > __entry->context = job->base.s_fence- > >finished.context; > __entry->seqno = job->base.s_fence- > >finished.seqno; > __assign_str(ring); > __entry->num_ibs = job->num_ibs; > ), > - TP_printk("sched_job=%llu, timeline=%s, context=%u, > seqno=%u, ring_name=%s, num_ibs=%u", > - __entry->sched_job_id, __get_str(timeline), > __entry->context, > + TP_printk("timeline=%s, context=%u, seqno=%u, > ring_name=%s, num_ibs=%u", > + __get_str(timeline), __entry->context, > __entry->seqno, __get_str(ring), __entry- > >num_ibs) > ); > > @@ -551,7 +547,6 @@ TRACE_EVENT(amdgpu_ib_pipe_sync, > TP_ARGS(sched_job, fence), > TP_STRUCT__entry( > __string(ring, sched_job->base.sched- > >name) > - __field(uint64_t, id) > __field(struct dma_fence *, fence) > __field(uint64_t, ctx) > __field(unsigned, seqno) > @@ -559,13 +554,12 @@ TRACE_EVENT(amdgpu_ib_pipe_sync, > > TP_fast_assign( > __assign_str(ring); > - __entry->id = sched_job->base.id; > __entry->fence = fence; > __entry->ctx = fence->context; > __entry->seqno = fence->seqno; > ), > - TP_printk("job ring=%s, id=%llu, need pipe sync to > fence=%p, context=%llu, seq=%u", > - __get_str(ring), __entry->id, > + TP_printk("job ring=%s need pipe sync to fence=%p, > context=%llu, seq=%u", > + __get_str(ring), > __entry->fence, __entry->ctx, > __entry->seqno) > ); > diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > index 4ce53e493fef..781b20349389 100644 > --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > @@ -36,7 +36,6 @@ DECLARE_EVENT_CLASS(drm_sched_job, > TP_PROTO(struct drm_sched_job *sched_job, struct > drm_sched_entity *entity), > TP_ARGS(sched_job, entity), > TP_STRUCT__entry( > - __field(uint64_t, id) > __string(name, sched_job->sched->name) > __field(u32, job_count) > __field(int, hw_job_count) > @@ -47,7 +46,6 @@ DECLARE_EVENT_CLASS(drm_sched_job, > ), > > TP_fast_assign( > - __entry->id = sched_job->id; > __assign_str(name); > __entry->job_count = > spsc_queue_count(&entity->job_queue); > __entry->hw_job_count = atomic_read( > @@ -57,8 +55,8 @@ DECLARE_EVENT_CLASS(drm_sched_job, > __entry->fence_seqno = sched_job- > >s_fence->finished.seqno; > __entry->client_id = sched_job->s_fence- > >drm_client_id; > ), > - TP_printk("dev=%s, id=%llu, fence=%llu:%llu, ring=%s, > job count:%u, hw job count:%d, client_id:%llu", > - __get_str(dev), __entry->id, > + TP_printk("dev=%s, fence=%llu:%llu, ring=%s, job > count:%u, hw job count:%d, client_id:%llu", > + __get_str(dev), > __entry->fence_context, __entry->fence_seqno, > __get_str(name), > __entry->job_count, __entry->hw_job_count, > __entry->client_id) > ); > @@ -95,7 +93,6 @@ TRACE_EVENT(drm_sched_job_add_dep, > TP_STRUCT__entry( > __field(u64, fence_context) > __field(u64, fence_seqno) > - __field(u64, id) > __field(u64, ctx) > __field(u64, seqno) > ), > @@ -103,12 +100,11 @@ TRACE_EVENT(drm_sched_job_add_dep, > TP_fast_assign( > __entry->fence_context = sched_job->s_fence- > >finished.context; > __entry->fence_seqno = sched_job->s_fence- > >finished.seqno; > - __entry->id = sched_job->id; > __entry->ctx = fence->context; > __entry->seqno = fence->seqno; > ), > - TP_printk("fence=%llu:%llu, id=%llu depends on > fence=%llu:%llu", > - __entry->fence_context, __entry->fence_seqno, > __entry->id, > + TP_printk("fence=%llu:%llu depends on fence=%llu:%llu", > + __entry->fence_context, __entry->fence_seqno, > __entry->ctx, __entry->seqno) > ); > > @@ -118,7 +114,6 @@ TRACE_EVENT(drm_sched_job_unschedulable, > TP_STRUCT__entry( > __field(u64, fence_context) > __field(u64, fence_seqno) > - __field(uint64_t, id) > __field(u64, ctx) > __field(u64, seqno) > ), > @@ -126,12 +121,11 @@ TRACE_EVENT(drm_sched_job_unschedulable, > TP_fast_assign( > __entry->fence_context = sched_job- > >s_fence->finished.context; > __entry->fence_seqno = sched_job- > >s_fence->finished.seqno; > - __entry->id = sched_job->id; > __entry->ctx = fence->context; > __entry->seqno = fence->seqno; > ), > - TP_printk("fence=%llu:%llu, id=%llu depends on > unsignalled fence=%llu:%llu", > - __entry->fence_context, __entry->fence_seqno, > __entry->id, > + TP_printk("fence=%llu:%llu depends on unsignalled > fence=%llu:%llu", > + __entry->fence_context, __entry->fence_seqno, > __entry->ctx, __entry->seqno) > ); > > diff --git a/drivers/gpu/drm/scheduler/sched_main.c > b/drivers/gpu/drm/scheduler/sched_main.c > index 195b5f891068..dafda1803c7c 100644 > --- a/drivers/gpu/drm/scheduler/sched_main.c > +++ b/drivers/gpu/drm/scheduler/sched_main.c > @@ -852,7 +852,6 @@ void drm_sched_job_arm(struct drm_sched_job *job) > > job->sched = sched; > job->s_priority = entity->priority; > - job->id = atomic64_inc_return(&sched->job_id_count); > > drm_sched_fence_init(job->s_fence, job->entity); > } > diff --git a/include/drm/gpu_scheduler.h > b/include/drm/gpu_scheduler.h > index 6fe3b4c0cffb..48190fdf661a 100644 > --- a/include/drm/gpu_scheduler.h > +++ b/include/drm/gpu_scheduler.h > @@ -326,7 +326,6 @@ struct drm_sched_fence *to_drm_sched_fence(struct > dma_fence *f); > * @finish_cb: the callback for the finished fence. > * @credits: the number of credits this job contributes to the > scheduler > * @work: Helper to reschedule job kill to different context. > - * @id: a unique id assigned to each job scheduled on the scheduler. > * @karma: increment on every hang caused by this job. If this > exceeds the hang > * limit of the scheduler then the job is marked guilty and > will not > * be scheduled further. > @@ -339,8 +338,6 @@ struct drm_sched_fence *to_drm_sched_fence(struct > dma_fence *f); > * to schedule the job. > */ > struct drm_sched_job { > - u64 id; > - > /** > * @submit_ts: > *