[Intel-xe] [PATCH v2 19/20] drm/xe/uapi: Crystal Reference Clock updates
Souza, Jose
jose.souza at intel.com
Thu Sep 21 21:04:58 UTC 2023
On Thu, 2023-09-21 at 16:49 -0400, Rodrigo Vivi wrote:
> On Thu, Sep 21, 2023 at 03:50:07PM +0000, Souza, Jose wrote:
> > On Wed, 2023-09-20 at 15:29 -0400, Rodrigo Vivi wrote:
> > > First of all, let's remove the duplication.
> > > But also, let's rename it to remove the word 'frequency'
> > > out of it. In general, the first thing people think of frequency
> > > is the frequency in which the GTs are operating to execute the
> > > GPU instructions.
> > >
> > > While this frequency here is a crystal reference clock frequency
> > > which is the base of everything else, and in this case of this
> > > uAPI it is used to calculate a better and precise timestamp.
> > >
> > > Cc: Matt Roper <matthew.d.roper at intel.com>
> > > Umesh Nerlige Ramappa <umesh.nerlige.ramappa at intel.com>
> > > Cc: Francois Dugast <francois.dugast at intel.com>
> > > Signed-off-by: Rodrigo Vivi <rodrigo.vivi at intel.com>
> > > Signed-off-by: Francois Dugast <francois.dugast at intel.com>
> > > ---
> > > drivers/gpu/drm/xe/xe_query.c | 5 ++---
> > > include/uapi/drm/xe_drm.h | 10 ++++------
> > > 2 files changed, 6 insertions(+), 9 deletions(-)
> > >
> > > diff --git a/drivers/gpu/drm/xe/xe_query.c b/drivers/gpu/drm/xe/xe_query.c
> > > index a27b10583f5a..8184df4abde1 100644
> > > --- a/drivers/gpu/drm/xe/xe_query.c
> > > +++ b/drivers/gpu/drm/xe/xe_query.c
> > > @@ -149,7 +149,7 @@ query_cs_cycles(struct xe_device *xe,
> > > if (!hwe)
> > > return -EINVAL;
> > >
> > > - resp.cs_frequency = gt->info.clock_freq;
> > > + resp.cs_reference_clock = gt->info.clock_freq;
> > >
> > > xe_device_mem_access_get(xe);
> > > xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL);
> > > @@ -167,7 +167,7 @@ query_cs_cycles(struct xe_device *xe,
> > > resp.width = 36;
> > >
> > > /* Only write to the output fields of user query */
> > > - if (put_user(resp.cs_frequency, &query_ptr->cs_frequency))
> > > + if (put_user(resp.cs_reference_clock, &query_ptr->cs_reference_clock))
> > > return -EFAULT;
> > >
> > > if (put_user(resp.cpu_timestamp, &query_ptr->cpu_timestamp))
> > > @@ -380,7 +380,6 @@ static int query_gt_list(struct xe_device *xe, struct drm_xe_device_query *query
> > > else
> > > gt_list->gt_list[id].type = XE_QUERY_GT_TYPE_MAIN;
> > > gt_list->gt_list[id].gt_id = gt->info.id;
> > > - gt_list->gt_list[id].clock_freq = gt->info.clock_freq;
> > > if (!IS_DGFX(xe))
> > > gt_list->gt_list[id].native_mem_regions = 0x1;
> > > else
> > > diff --git a/include/uapi/drm/xe_drm.h b/include/uapi/drm/xe_drm.h
> > > index 45ec20c2765d..5156bac84f20 100644
> > > --- a/include/uapi/drm/xe_drm.h
> > > +++ b/include/uapi/drm/xe_drm.h
> > > @@ -248,8 +248,8 @@ struct drm_xe_query_mem_region {
> > > * relevant GPU timestamp. clockid is used to return the specific CPU
> > > * timestamp.
> > > *
> > > - * The query returns the command streamer cycles and the frequency that can
> > > - * be used to calculate the command streamer timestamp. In addition the
> > > + * The query returns the command streamer cycles and the reference clock that
> > > + * can be used to calculate the command streamer timestamp. In addition the
> > > * query returns a set of cpu timestamps that indicate when the command
> > > * streamer cycle count was captured.
> > > */
> > > @@ -266,8 +266,8 @@ struct drm_xe_query_cs_cycles {
> > > */
> > > __u64 cs_cycles;
> > >
> > > - /** Frequency of the cs cycles in Hz. */
> > > - __u64 cs_frequency;
> > > + /** Reference Clock of the cs cycles in Hz. */
> > > + __u64 cs_reference_clock;
> > >
> > > /**
> > > * CPU timestamp in ns. The timestamp is captured before reading the
> > > @@ -381,8 +381,6 @@ struct drm_xe_query_gt {
> > > __u16 type;
> > > /** @gt_id: Unique ID of this GT within the PCI Device */
> > > __u16 gt_id;
> > > - /** @clock_freq: A clock frequency for timestamp */
> > > - __u32 clock_freq;
> >
> > This needs to be brought back, if you want to remove the duplication remove it from drm_xe_query_cs_cycles then.
> > DRM_XE_QUERY_CS_CYCLES will fail when running in SRIOV VF as it can't do mmio reads.
>
> So, does the CS_CYLES make sense then? or all that information could/should
> be inside the gt info here?
It makes sense, cs_cycles, cpu_timestamp and cpu_delta will change on every call.
cs_cycles = engine timestamp
cpu_timestamp = CPU timestamp after read cs_cycles
cpu_delta = cpu_timestamp - CPU timestamp before read cs_cycles
For some use cases Mesa can request engine timestamp from batch buffer so it will need the GT clock_freq even when DRM_XE_QUERY_CS_CYCLES is not
possible, for others case like when gathering data to do performance analysts the request trough batch buffers adds to much delay and variance.
>
> Let's take OA into consideration as well here.
> https://lore.kernel.org/all/20230919161049.2307855-17-ashutosh.dixit@intel.com
>
> Cc: Ashutosh.
> Anything from OA that we could do to unify these data?
>
> >
> > > /**
> > > * @native_mem_regions: Bit mask of instances from
> > > * drm_xe_query_mem_usage that lives on the same GPU/Tile and have
> >
More information about the Intel-xe
mailing list