Re: [PATCH 3/3] drm: fix the warnning of string style for scheduler trace.
Ping~ Seems this patch is missed. Thanks, Ray On Mon, Dec 13, 2021 at 02:34:22PM +0800, Huang, Ray wrote: > Use __string(), __assign_str() and __get_str() helpers in the TRACE_EVENT() > instead of string definitions in gpu scheduler trace. > > [ 158.890890] [ cut here ] > [ 158.890899] fmt: 'entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw > job count:%d >' current_buffer: 'Xorg-1588[001] . > 149.391136: drm_sched_job: entity=76f0d517, id=1, > fence=8dd56028, ring=' > [ 158.890910] WARNING: CPU: 6 PID: 1617 at kernel/trace/trace.c:3830 > trace_check_vprintf+0x481/0x4a0 > > Signed-off-by: Huang Rui > --- > drivers/gpu/drm/scheduler/gpu_scheduler_trace.h | 17 + > 1 file changed, 9 insertions(+), 8 deletions(-) > > diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > index 877ce9b127f1..4e397790c195 100644 > --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > @@ -38,6 +38,7 @@ TRACE_EVENT(drm_sched_job, > TP_STRUCT__entry( >__field(struct drm_sched_entity *, entity) >__field(struct dma_fence *, fence) > + __string(name, sched_job->sched->name) >__field(const char *, name) >__field(uint64_t, id) >__field(u32, job_count) > @@ -48,14 +49,14 @@ TRACE_EVENT(drm_sched_job, > __entry->entity = entity; > __entry->id = sched_job->id; > __entry->fence = _job->s_fence->finished; > -__entry->name = sched_job->sched->name; > +__assign_str(name, sched_job->sched->name); > __entry->job_count = > spsc_queue_count(>job_queue); > __entry->hw_job_count = atomic_read( > _job->sched->hw_rq_count); > ), > TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw > job count:%d", > __entry->entity, __entry->id, > - __entry->fence, __entry->name, > + __entry->fence, __get_str(name), > __entry->job_count, __entry->hw_job_count) > ); > > @@ -65,7 +66,7 @@ TRACE_EVENT(drm_run_job, > TP_STRUCT__entry( >__field(struct drm_sched_entity *, entity) >__field(struct dma_fence *, fence) > - __field(const char *, name) > + __string(name, sched_job->sched->name) >__field(uint64_t, id) >__field(u32, job_count) >__field(int, hw_job_count) > @@ -75,14 +76,14 @@ TRACE_EVENT(drm_run_job, > __entry->entity = entity; > __entry->id = sched_job->id; > __entry->fence = _job->s_fence->finished; > -__entry->name = sched_job->sched->name; > +__assign_str(name, sched_job->sched->name); > __entry->job_count = > spsc_queue_count(>job_queue); > __entry->hw_job_count = atomic_read( > _job->sched->hw_rq_count); > ), > TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw > job count:%d", > __entry->entity, __entry->id, > - __entry->fence, __entry->name, > + __entry->fence, __get_str(name), > __entry->job_count, __entry->hw_job_count) > ); > > @@ -103,7 +104,7 @@ TRACE_EVENT(drm_sched_job_wait_dep, > TP_PROTO(struct drm_sched_job *sched_job, struct dma_fence *fence), > TP_ARGS(sched_job, fence), > TP_STRUCT__entry( > - __field(const char *,name) > + __string(name, sched_job->sched->name) >__field(uint64_t, id) >__field(struct dma_fence *, fence) >__field(uint64_t, ctx) > @@ -111,14 +112,14 @@ TRACE_EVENT(drm_sched_job_wait_dep, >), > > TP_fast_assign( > -__entry->name = sched_job->sched->name; > +__assign_str(name, sched_job->sched->name); > __entry->id = sched_job->id; > __entry->fence = fence; > __entry->ctx = fence->context; > __entry->seqno = fence->seqno; > ), > TP_printk("job ring=%s, id=%llu, depends fence=%p,
Re: [PATCH 3/3] drm: fix the warnning of string style for scheduler trace.
A soft reminder. May I know any comments of this patch, just a minor warning fix? Thanks, Ray On Mon, Dec 13, 2021 at 02:34:22PM +0800, Huang, Ray wrote: > Use __string(), __assign_str() and __get_str() helpers in the TRACE_EVENT() > instead of string definitions in gpu scheduler trace. > > [ 158.890890] [ cut here ] > [ 158.890899] fmt: 'entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw > job count:%d >' current_buffer: 'Xorg-1588[001] . > 149.391136: drm_sched_job: entity=76f0d517, id=1, > fence=8dd56028, ring=' > [ 158.890910] WARNING: CPU: 6 PID: 1617 at kernel/trace/trace.c:3830 > trace_check_vprintf+0x481/0x4a0 > > Signed-off-by: Huang Rui > --- > drivers/gpu/drm/scheduler/gpu_scheduler_trace.h | 17 + > 1 file changed, 9 insertions(+), 8 deletions(-) > > diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > index 877ce9b127f1..4e397790c195 100644 > --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h > @@ -38,6 +38,7 @@ TRACE_EVENT(drm_sched_job, > TP_STRUCT__entry( >__field(struct drm_sched_entity *, entity) >__field(struct dma_fence *, fence) > + __string(name, sched_job->sched->name) >__field(const char *, name) >__field(uint64_t, id) >__field(u32, job_count) > @@ -48,14 +49,14 @@ TRACE_EVENT(drm_sched_job, > __entry->entity = entity; > __entry->id = sched_job->id; > __entry->fence = _job->s_fence->finished; > -__entry->name = sched_job->sched->name; > +__assign_str(name, sched_job->sched->name); > __entry->job_count = > spsc_queue_count(>job_queue); > __entry->hw_job_count = atomic_read( > _job->sched->hw_rq_count); > ), > TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw > job count:%d", > __entry->entity, __entry->id, > - __entry->fence, __entry->name, > + __entry->fence, __get_str(name), > __entry->job_count, __entry->hw_job_count) > ); > > @@ -65,7 +66,7 @@ TRACE_EVENT(drm_run_job, > TP_STRUCT__entry( >__field(struct drm_sched_entity *, entity) >__field(struct dma_fence *, fence) > - __field(const char *, name) > + __string(name, sched_job->sched->name) >__field(uint64_t, id) >__field(u32, job_count) >__field(int, hw_job_count) > @@ -75,14 +76,14 @@ TRACE_EVENT(drm_run_job, > __entry->entity = entity; > __entry->id = sched_job->id; > __entry->fence = _job->s_fence->finished; > -__entry->name = sched_job->sched->name; > +__assign_str(name, sched_job->sched->name); > __entry->job_count = > spsc_queue_count(>job_queue); > __entry->hw_job_count = atomic_read( > _job->sched->hw_rq_count); > ), > TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw > job count:%d", > __entry->entity, __entry->id, > - __entry->fence, __entry->name, > + __entry->fence, __get_str(name), > __entry->job_count, __entry->hw_job_count) > ); > > @@ -103,7 +104,7 @@ TRACE_EVENT(drm_sched_job_wait_dep, > TP_PROTO(struct drm_sched_job *sched_job, struct dma_fence *fence), > TP_ARGS(sched_job, fence), > TP_STRUCT__entry( > - __field(const char *,name) > + __string(name, sched_job->sched->name) >__field(uint64_t, id) >__field(struct dma_fence *, fence) >__field(uint64_t, ctx) > @@ -111,14 +112,14 @@ TRACE_EVENT(drm_sched_job_wait_dep, >), > > TP_fast_assign( > -__entry->name = sched_job->sched->name; > +__assign_str(name, sched_job->sched->name); > __entry->id = sched_job->id; > __entry->fence = fence; > __entry->ctx = fence->context; > __entry->seqno = fence->seqno; > ), > TP_printk("job
[PATCH 3/3] drm: fix the warnning of string style for scheduler trace.
Use __string(), __assign_str() and __get_str() helpers in the TRACE_EVENT() instead of string definitions in gpu scheduler trace. [ 158.890890] [ cut here ] [ 158.890899] fmt: 'entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d ' current_buffer: 'Xorg-1588[001] . 149.391136: drm_sched_job: entity=76f0d517, id=1, fence=8dd56028, ring=' [ 158.890910] WARNING: CPU: 6 PID: 1617 at kernel/trace/trace.c:3830 trace_check_vprintf+0x481/0x4a0 Signed-off-by: Huang Rui --- drivers/gpu/drm/scheduler/gpu_scheduler_trace.h | 17 + 1 file changed, 9 insertions(+), 8 deletions(-) diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h index 877ce9b127f1..4e397790c195 100644 --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h @@ -38,6 +38,7 @@ TRACE_EVENT(drm_sched_job, TP_STRUCT__entry( __field(struct drm_sched_entity *, entity) __field(struct dma_fence *, fence) +__string(name, sched_job->sched->name) __field(const char *, name) __field(uint64_t, id) __field(u32, job_count) @@ -48,14 +49,14 @@ TRACE_EVENT(drm_sched_job, __entry->entity = entity; __entry->id = sched_job->id; __entry->fence = _job->s_fence->finished; - __entry->name = sched_job->sched->name; + __assign_str(name, sched_job->sched->name); __entry->job_count = spsc_queue_count(>job_queue); __entry->hw_job_count = atomic_read( _job->sched->hw_rq_count); ), TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d", __entry->entity, __entry->id, - __entry->fence, __entry->name, + __entry->fence, __get_str(name), __entry->job_count, __entry->hw_job_count) ); @@ -65,7 +66,7 @@ TRACE_EVENT(drm_run_job, TP_STRUCT__entry( __field(struct drm_sched_entity *, entity) __field(struct dma_fence *, fence) -__field(const char *, name) +__string(name, sched_job->sched->name) __field(uint64_t, id) __field(u32, job_count) __field(int, hw_job_count) @@ -75,14 +76,14 @@ TRACE_EVENT(drm_run_job, __entry->entity = entity; __entry->id = sched_job->id; __entry->fence = _job->s_fence->finished; - __entry->name = sched_job->sched->name; + __assign_str(name, sched_job->sched->name); __entry->job_count = spsc_queue_count(>job_queue); __entry->hw_job_count = atomic_read( _job->sched->hw_rq_count); ), TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d", __entry->entity, __entry->id, - __entry->fence, __entry->name, + __entry->fence, __get_str(name), __entry->job_count, __entry->hw_job_count) ); @@ -103,7 +104,7 @@ TRACE_EVENT(drm_sched_job_wait_dep, TP_PROTO(struct drm_sched_job *sched_job, struct dma_fence *fence), TP_ARGS(sched_job, fence), TP_STRUCT__entry( -__field(const char *,name) +__string(name, sched_job->sched->name) __field(uint64_t, id) __field(struct dma_fence *, fence) __field(uint64_t, ctx) @@ -111,14 +112,14 @@ TRACE_EVENT(drm_sched_job_wait_dep, ), TP_fast_assign( - __entry->name = sched_job->sched->name; + __assign_str(name, sched_job->sched->name); __entry->id = sched_job->id; __entry->fence = fence; __entry->ctx = fence->context; __entry->seqno = fence->seqno; ), TP_printk("job ring=%s, id=%llu, depends fence=%p, context=%llu, seq=%u", - __entry->name, __entry->id, + __get_str(name), __entry->id, __entry->fence, __entry->ctx,