[PATCH 15/32] drm/amdkfd: prepare trap workaround for gfx11

Kim, Jonathan Jonathan.Kim at amd.com
Thu Mar 23 13:50:40 UTC 2023


[Public]

> -----Original Message-----
> From: Kuehling, Felix <Felix.Kuehling at amd.com>
> Sent: Monday, March 20, 2023 5:50 PM
> To: Kim, Jonathan <Jonathan.Kim at amd.com>; amd-
> gfx at lists.freedesktop.org; dri-devel at lists.freedesktop.org
> Subject: Re: [PATCH 15/32] drm/amdkfd: prepare trap workaround for gfx11
>
>
> On 2023-01-25 14:53, Jonathan Kim wrote:
> > Due to a HW bug, waves in only half the shader arrays can enter trap.
> >
> > When starting a debug session, relocate all waves to the first shader
> > array of each shader engine and mask off the 2nd shader array as
> > unavailable.
> >
> > When ending a debug session, re-enable the 2nd shader array per
> > shader engine.
> >
> > User CU masking per queue cannot be guaranteed to remain functional
> > if requested during debugging (e.g. user cu mask requests only 2nd shader
> > array as an available resource leading to zero HW resources available)
> > nor can runtime be alerted of any of these changes during execution.
> >
> > Make user CU masking and debugging mutual exclusive with respect to
> > availability.
> >
> > If the debugger tries to attach to a process with a user cu masked
> > queue, return the runtime status as enabled but busy.
> >
> > If the debugger tries to attach and fails to reallocate queue waves to
> > the first shader array of each shader engine, return the runtime status
> > as enabled but with an error.
> >
> > In addition, like any other mutli-process debug supported devices,
> > disable trap temporary setup per-process to avoid performance impact
> from
> > setup overhead.
> >
> > Signed-off-by: Jonathan Kim <jonathan.kim at amd.com>
> > ---
> >   drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h       |  2 +
> >   drivers/gpu/drm/amd/amdgpu/mes_v11_0.c        |  7 +-
> >   drivers/gpu/drm/amd/amdkfd/kfd_chardev.c      |  2 -
> >   drivers/gpu/drm/amd/amdkfd/kfd_debug.c        | 64
> +++++++++++++++++++
> >   drivers/gpu/drm/amd/amdkfd/kfd_debug.h        |  3 +-
> >   .../drm/amd/amdkfd/kfd_device_queue_manager.c |  7 ++
> >   .../gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c  |  3 +-
> >   .../gpu/drm/amd/amdkfd/kfd_mqd_manager_v10.c  |  3 +-
> >   .../gpu/drm/amd/amdkfd/kfd_mqd_manager_v11.c  | 42 ++++++++----
> >   .../gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c   |  3 +-
> >   .../gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c   |  3 +-
> >   drivers/gpu/drm/amd/amdkfd/kfd_priv.h         |  5 +-
> >   .../amd/amdkfd/kfd_process_queue_manager.c    |  9 ++-
> >   13 files changed, 124 insertions(+), 29 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
> > index d20df0cf0d88..b5f5eed2b5ef 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
> > @@ -219,6 +219,8 @@ struct mes_add_queue_input {
> >     uint32_t        gws_size;
> >     uint64_t        tba_addr;
> >     uint64_t        tma_addr;
> > +   uint32_t        trap_en;
> > +   uint32_t        skip_process_ctx_clear;
> >     uint32_t        is_kfd_process;
> >     uint32_t        is_aql_queue;
> >     uint32_t        queue_size;
> > diff --git a/drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
> b/drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
> > index fbacdc42efac..38c7a0cbf264 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
> > @@ -197,17 +197,14 @@ static int mes_v11_0_add_hw_queue(struct
> amdgpu_mes *mes,
> >     mes_add_queue_pkt.gws_size = input->gws_size;
> >     mes_add_queue_pkt.trap_handler_addr = input->tba_addr;
> >     mes_add_queue_pkt.tma_addr = input->tma_addr;
> > +   mes_add_queue_pkt.trap_en = input->trap_en;
> > +   mes_add_queue_pkt.skip_process_ctx_clear = input-
> >skip_process_ctx_clear;
> >     mes_add_queue_pkt.is_kfd_process = input->is_kfd_process;
> >
> >     /* For KFD, gds_size is re-used for queue size (needed in MES for AQL
> queues) */
> >     mes_add_queue_pkt.is_aql_queue = input->is_aql_queue;
> >     mes_add_queue_pkt.gds_size = input->queue_size;
> >
> > -   if (!(((adev->mes.sched_version & AMDGPU_MES_VERSION_MASK) >=
> 4) &&
> > -             (adev->ip_versions[GC_HWIP][0] >= IP_VERSION(11, 0, 0))
> &&
> > -             (adev->ip_versions[GC_HWIP][0] <= IP_VERSION(11, 0, 3))))
> > -           mes_add_queue_pkt.trap_en = 1;
> > -
> >     /* For KFD, gds_size is re-used for queue size (needed in MES for AQL
> queues) */
> >     mes_add_queue_pkt.is_aql_queue = input->is_aql_queue;
> >     mes_add_queue_pkt.gds_size = input->queue_size;
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> > index ee05c2e54ef6..f5f639de28f0 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
> > @@ -530,8 +530,6 @@ static int kfd_ioctl_set_cu_mask(struct file *filp,
> struct kfd_process *p,
> >             goto out;
> >     }
> >
> > -   minfo.update_flag = UPDATE_FLAG_CU_MASK;
> > -
> >     mutex_lock(&p->mutex);
> >
> >     retval = pqm_update_mqd(&p->pqm, args->queue_id, &minfo);
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> > index f6ea6db266b4..6e99a0160275 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_debug.c
> > @@ -37,6 +37,70 @@ void debug_event_write_work_handler(struct
> work_struct *work)
> >     kernel_write(process->dbg_ev_file, &write_data, 1, &pos);
> >   }
> >
> > +static int kfd_dbg_set_queue_workaround(struct queue *q, bool enable)
> > +{
> > +   struct mqd_update_info minfo = {0};
> > +   int err;
> > +
> > +   if (!q || (!q->properties.is_dbg_wa && !enable))
>
> Should this condition be:
>
>      if (!q || q->properties.is_dbg_wa != enable)

The latter part should probably be q->properties.is_dbg_wa == enable.  q->properties.is_dbg_wa != enable would always skip a request to change the queue's current workaround state.
I think we can just drop the latter half of this test condition as a redundant queue workaround update is harmless.
It's a static call from a process wide call and the process wide call is static itself and only gets called twice, once on attach and once on detach.

Thanks,

Jon

>
>
> > +           return 0;
> > +
> > +   if (KFD_GC_VERSION(q->device) < IP_VERSION(11, 0, 0) ||
> > +                   KFD_GC_VERSION(q->device) >= IP_VERSION(12, 0, 0))
>
> Indentation. It would be more readable if the KFD_GC_VERSIONs were
> aligned.
>
>
> > +           return 0;
> > +
> > +   if (enable && q->properties.is_user_cu_masked)
> > +           return -EBUSY;
> > +
> > +   minfo.update_flag = enable ? UPDATE_FLAG_DBG_WA_ENABLE :
> UPDATE_FLAG_DBG_WA_DISABLE;
> > +
> > +   q->properties.is_dbg_wa = enable;
> > +   err = q->device->dqm->ops.update_queue(q->device->dqm, q,
> &minfo);
> > +   if (err)
> > +           q->properties.is_dbg_wa = false;
> > +
> > +   return err;
> > +}
> > +
> > +static int kfd_dbg_set_workaround(struct kfd_process *target, bool
> enable)
> > +{
> > +   struct process_queue_manager *pqm = &target->pqm;
> > +   struct process_queue_node *pqn;
> > +   int r = 0;
> > +
> > +   list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
> > +           r = kfd_dbg_set_queue_workaround(pqn->q, enable);
> > +           if (enable && r)
> > +                   goto unwind;
> > +   }
> > +
> > +   return 0;
> > +
> > +unwind:
> > +   list_for_each_entry(pqn, &pqm->queues, process_queue_list)
> > +           kfd_dbg_set_queue_workaround(pqn->q, false);
> > +
> > +   if (enable) {
> > +           target->runtime_info.runtime_state = r == -EBUSY ?
> > +                           DEBUG_RUNTIME_STATE_ENABLED_BUSY :
> > +                           DEBUG_RUNTIME_STATE_ENABLED_ERROR;
> > +   }
>
> Braces are not needed here.
>
>
> > +
> > +   return r;
> > +}
> > +
> > +static int kfd_dbg_set_mes_debug_mode(struct kfd_process_device *pdd)
> > +{
> > +   uint32_t spi_dbg_cntl = pdd->spi_dbg_override | pdd-
> >spi_dbg_launch_mode;
> > +   uint32_t flags = pdd->process->dbg_flags;
> > +
> > +   if (!kfd_dbg_is_per_vmid_supported(pdd->dev))
> > +           return 0;
> > +
> > +   return amdgpu_mes_set_shader_debugger(pdd->dev->adev, pdd-
> >proc_ctx_gpu_addr, spi_dbg_cntl,
> > +                                           pdd->watch_points, flags);
> > +}
> > +
>
> You're adding some unused static functions here. This will cause compile
> warnings until the patch that starts using them. You could avoid this by
> reordering this and the next patch and moving the function calls into
> this patch. That would also make it more obvious where the workaround
> plugs into the debug code.
>
> Regards,
>    Felix
>
>
> >   int kfd_dbg_trap_disable(struct kfd_process *target)
> >   {
> >     if (!target->debug_trap_enabled)
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_debug.h
> b/drivers/gpu/drm/amd/amdkfd/kfd_debug.h
> > index 53c5a3e55bd2..0c09f1729325 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_debug.h
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_debug.h
> > @@ -35,7 +35,8 @@ int kfd_dbg_trap_enable(struct kfd_process *target,
> uint32_t fd,
> >
> >   static inline bool kfd_dbg_is_per_vmid_supported(struct kfd_dev *dev)
> >   {
> > -   return KFD_GC_VERSION(dev) == IP_VERSION(9, 4, 2);
> > +   return KFD_GC_VERSION(dev) == IP_VERSION(9, 4, 2) ||
> > +          KFD_GC_VERSION(dev) >= IP_VERSION(11, 0, 0);
> >   }
> >
> >   void debug_event_write_work_handler(struct work_struct *work);
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> > index 2517716d7cbc..be1985b87ea7 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
> > @@ -214,6 +214,10 @@ static int add_queue_mes(struct
> device_queue_manager *dqm, struct queue *q,
> >     queue_input.paging = false;
> >     queue_input.tba_addr = qpd->tba_addr;
> >     queue_input.tma_addr = qpd->tma_addr;
> > +   queue_input.trap_en = KFD_GC_VERSION(q->device) <
> IP_VERSION(11, 0, 0) ||
> > +                         KFD_GC_VERSION(q->device) >= IP_VERSION(12, 0,
> 0) ||
> > +                         q->properties.is_dbg_wa;
> > +   queue_input.skip_process_ctx_clear = qpd->pqm->process-
> >debug_trap_enabled;
> >
> >     queue_type = convert_to_mes_queue_type(q->properties.type);
> >     if (queue_type < 0) {
> > @@ -1679,6 +1683,9 @@ static int create_queue_cpsch(struct
> device_queue_manager *dqm, struct queue *q,
> >      * updates the is_evicted flag but is a no-op otherwise.
> >      */
> >     q->properties.is_evicted = !!qpd->evicted;
> > +   q->properties.is_dbg_wa = qpd->pqm->process-
> >debug_trap_enabled &&
> > +                   KFD_GC_VERSION(q->device) >= IP_VERSION(11, 0, 0)
> &&
> > +                   KFD_GC_VERSION(q->device) < IP_VERSION(12, 0, 0);
> >
> >     if (qd)
> >             mqd_mgr->restore_mqd(mqd_mgr, &q->mqd, q-
> >mqd_mem_obj, &q->gart_mqd_addr,
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> > index 4889865c725c..c2a7226fc588 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_cik.c
> > @@ -48,8 +48,7 @@ static void update_cu_mask(struct mqd_manager
> *mm, void *mqd,
> >     struct cik_mqd *m;
> >     uint32_t se_mask[4] = {0}; /* 4 is the max # of SEs */
> >
> > -   if (!minfo || (minfo->update_flag != UPDATE_FLAG_CU_MASK) ||
> > -       !minfo->cu_mask.ptr)
> > +   if (!minfo || !minfo->cu_mask.ptr)
> >             return;
> >
> >     mqd_symmetrically_map_cu_mask(mm,
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v10.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v10.c
> > index cb484ace17de..8248e77751e7 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v10.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v10.c
> > @@ -48,8 +48,7 @@ static void update_cu_mask(struct mqd_manager
> *mm, void *mqd,
> >     struct v10_compute_mqd *m;
> >     uint32_t se_mask[4] = {0}; /* 4 is the max # of SEs */
> >
> > -   if (!minfo || (minfo->update_flag != UPDATE_FLAG_CU_MASK) ||
> > -       !minfo->cu_mask.ptr)
> > +   if (!minfo || !minfo->cu_mask.ptr)
> >             return;
> >
> >     mqd_symmetrically_map_cu_mask(mm,
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v11.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v11.c
> > index ac7c8fc83c94..18ab613e787c 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v11.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v11.c
> > @@ -46,15 +46,33 @@ static void update_cu_mask(struct mqd_manager
> *mm, void *mqd,
> >   {
> >     struct v11_compute_mqd *m;
> >     uint32_t se_mask[KFD_MAX_NUM_SE] = {0};
> > +   bool has_wa_flag = minfo && (minfo->update_flag &
> (UPDATE_FLAG_DBG_WA_ENABLE |
> > +                   UPDATE_FLAG_DBG_WA_DISABLE));
> >
> > -   if (!minfo || (minfo->update_flag != UPDATE_FLAG_CU_MASK) ||
> > -       !minfo->cu_mask.ptr)
> > +   if (!minfo || !(has_wa_flag || minfo->cu_mask.ptr))
> >             return;
> >
> > +   m = get_mqd(mqd);
> > +
> > +   if (has_wa_flag) {
> > +           uint32_t wa_mask = minfo->update_flag ==
> UPDATE_FLAG_DBG_WA_ENABLE ?
> > +                                           0xffff : 0xffffffff;
> > +
> > +           m->compute_static_thread_mgmt_se0 = wa_mask;
> > +           m->compute_static_thread_mgmt_se1 = wa_mask;
> > +           m->compute_static_thread_mgmt_se2 = wa_mask;
> > +           m->compute_static_thread_mgmt_se3 = wa_mask;
> > +           m->compute_static_thread_mgmt_se4 = wa_mask;
> > +           m->compute_static_thread_mgmt_se5 = wa_mask;
> > +           m->compute_static_thread_mgmt_se6 = wa_mask;
> > +           m->compute_static_thread_mgmt_se7 = wa_mask;
> > +
> > +           return;
> > +   }
> > +
> >     mqd_symmetrically_map_cu_mask(mm,
> >             minfo->cu_mask.ptr, minfo->cu_mask.count, se_mask);
> >
> > -   m = get_mqd(mqd);
> >     m->compute_static_thread_mgmt_se0 = se_mask[0];
> >     m->compute_static_thread_mgmt_se1 = se_mask[1];
> >     m->compute_static_thread_mgmt_se2 = se_mask[2];
> > @@ -109,6 +127,7 @@ static void init_mqd(struct mqd_manager *mm,
> void **mqd,
> >     uint64_t addr;
> >     struct v11_compute_mqd *m;
> >     int size;
> > +   uint32_t wa_mask = q->is_dbg_wa ? 0xffff : 0xffffffff;
> >
> >     m = (struct v11_compute_mqd *) mqd_mem_obj->cpu_ptr;
> >     addr = mqd_mem_obj->gpu_addr;
> > @@ -122,14 +141,15 @@ static void init_mqd(struct mqd_manager *mm,
> void **mqd,
> >
> >     m->header = 0xC0310800;
> >     m->compute_pipelinestat_enable = 1;
> > -   m->compute_static_thread_mgmt_se0 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se1 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se2 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se3 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se4 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se5 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se6 = 0xFFFFFFFF;
> > -   m->compute_static_thread_mgmt_se7 = 0xFFFFFFFF;
> > +
> > +   m->compute_static_thread_mgmt_se0 = wa_mask;
> > +   m->compute_static_thread_mgmt_se1 = wa_mask;
> > +   m->compute_static_thread_mgmt_se2 = wa_mask;
> > +   m->compute_static_thread_mgmt_se3 = wa_mask;
> > +   m->compute_static_thread_mgmt_se4 = wa_mask;
> > +   m->compute_static_thread_mgmt_se5 = wa_mask;
> > +   m->compute_static_thread_mgmt_se6 = wa_mask;
> > +   m->compute_static_thread_mgmt_se7 = wa_mask;
> >
> >     m->cp_hqd_persistent_state =
> CP_HQD_PERSISTENT_STATE__PRELOAD_REQ_MASK |
> >                     0x55 <<
> CP_HQD_PERSISTENT_STATE__PRELOAD_SIZE__SHIFT;
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
> > index 86f1cf090246..50da16dd4c96 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
> > @@ -49,8 +49,7 @@ static void update_cu_mask(struct mqd_manager
> *mm, void *mqd,
> >     struct v9_mqd *m;
> >     uint32_t se_mask[KFD_MAX_NUM_SE] = {0};
> >
> > -   if (!minfo || (minfo->update_flag != UPDATE_FLAG_CU_MASK) ||
> > -       !minfo->cu_mask.ptr)
> > +   if (!minfo || !minfo->cu_mask.ptr)
> >             return;
> >
> >     mqd_symmetrically_map_cu_mask(mm,
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> > index 530ba6f5b57e..58b40bff3e0c 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_vi.c
> > @@ -51,8 +51,7 @@ static void update_cu_mask(struct mqd_manager
> *mm, void *mqd,
> >     struct vi_mqd *m;
> >     uint32_t se_mask[4] = {0}; /* 4 is the max # of SEs */
> >
> > -   if (!minfo || (minfo->update_flag != UPDATE_FLAG_CU_MASK) ||
> > -       !minfo->cu_mask.ptr)
> > +   if (!minfo || !minfo->cu_mask.ptr)
> >             return;
> >
> >     mqd_symmetrically_map_cu_mask(mm,
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> b/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> > index 8f1e2f9023db..75521d96e937 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_priv.h
> > @@ -479,6 +479,8 @@ struct queue_properties {
> >     bool is_evicted;
> >     bool is_active;
> >     bool is_gws;
> > +   bool is_dbg_wa;
> > +   bool is_user_cu_masked;
> >     /* Not relevant for user mode queues in cp scheduling */
> >     unsigned int vmid;
> >     /* Relevant only for sdma queues*/
> > @@ -501,7 +503,8 @@ struct queue_properties {
> >                         !(q).is_evicted)
> >
> >   enum mqd_update_flag {
> > -   UPDATE_FLAG_CU_MASK = 0,
> > +   UPDATE_FLAG_DBG_WA_ENABLE = 1,
> > +   UPDATE_FLAG_DBG_WA_DISABLE = 2,
> >   };
> >
> >   struct mqd_update_info {
> > diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> b/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> > index 5137476ec18e..d8f032214481 100644
> > --- a/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> > +++ b/drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
> > @@ -498,8 +498,12 @@ int pqm_update_mqd(struct
> process_queue_manager *pqm,
> >             return -EFAULT;
> >     }
> >
> > +   /* CUs are masked for debugger requirements so deny user mask  */
> > +   if (pqn->q->properties.is_dbg_wa && minfo && minfo->cu_mask.ptr)
> > +           return -EBUSY;
> > +
> >     /* ASICs that have WGPs must enforce pairwise enabled mask
> checks. */
> > -   if (minfo && minfo->update_flag == UPDATE_FLAG_CU_MASK &&
> minfo->cu_mask.ptr &&
> > +   if (minfo && minfo->cu_mask.ptr &&
> >                     KFD_GC_VERSION(pqn->q->device) >= IP_VERSION(10,
> 0, 0)) {
> >             int i;
> >
> > @@ -518,6 +522,9 @@ int pqm_update_mqd(struct
> process_queue_manager *pqm,
> >     if (retval != 0)
> >             return retval;
> >
> > +   if (minfo && minfo->cu_mask.ptr)
> > +           pqn->q->properties.is_user_cu_masked = true;
> > +
> >     return 0;
> >   }
> >


More information about the dri-devel mailing list