[PATCH 2/7] drm/amdgpu: Add ring reset callback for JPEG4_0_3
Sundararaju, Sathishkumar
sasundar at amd.com
Fri Jan 31 16:55:34 UTC 2025
On 1/31/2025 10:19 PM, Christian König wrote:
> Am 31.01.25 um 17:40 schrieb Sundararaju, Sathishkumar:
>> Hi Christian,
>>
>>
>> On 1/31/2025 9:56 PM, Christian König wrote:
>>> Am 31.01.25 um 17:23 schrieb Sathishkumar S:
>>>> Add ring reset function callback for JPEG4_0_3 to
>>>> recover from job timeouts without a full gpu reset.
>>>>
>>>> Signed-off-by: Sathishkumar S <sathishkumar.sundararaju at amd.com>
>>>> ---
>>>> drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c | 60
>>>> ++++++++++++++++++++++--
>>>> 1 file changed, 57 insertions(+), 3 deletions(-)
>>>>
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
>>>> b/drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
>>>> index be0b3b4c8690..62d8628dccc5 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
>>>> @@ -204,9 +204,7 @@ static int jpeg_v4_0_3_sw_init(struct
>>>> amdgpu_ip_block *ip_block)
>>>> if (r)
>>>> return r;
>>>> - /* TODO: Add queue reset mask when FW fully supports it */
>>>> - adev->jpeg.supported_reset =
>>>> - amdgpu_get_soft_full_reset_mask(&adev->jpeg.inst[0].ring_dec[0]);
>>>> + adev->jpeg.supported_reset = AMDGPU_RESET_TYPE_PER_PIPE;
>>>> r = amdgpu_jpeg_sysfs_reset_mask_init(adev);
>>>> if (r)
>>>> return r;
>>>> @@ -231,6 +229,7 @@ static int jpeg_v4_0_3_sw_fini(struct
>>>> amdgpu_ip_block *ip_block)
>>>> return r;
>>>> amdgpu_jpeg_sysfs_reset_mask_fini(adev);
>>>> +
>>>> r = amdgpu_jpeg_sw_fini(adev);
>>>> return r;
>>>> @@ -1099,6 +1098,60 @@ static int
>>>> jpeg_v4_0_3_process_interrupt(struct amdgpu_device *adev,
>>>> return 0;
>>>> }
>>>> +static int jpeg_v4_0_3_wait_for_idle_on_inst(struct amdgpu_ring
>>>> *ring)
>>>> +{
>>>> + struct amdgpu_device *adev = ring->adev;
>>>> + struct amdgpu_ring *r;
>>>> + int ret, j;
>>>> +
>>>> + for (j = 0; j < adev->jpeg.num_jpeg_rings; ++j) {
>>>> + r = &adev->jpeg.inst[ring->me].ring_dec[j];
>>>> + r->sched.ready = false;
>>>> + }
>>>> + /* publish update */
>>>> + smp_rmb();
>>>
>>> Using smp_rmb() to publish the update is 100% incorrect.
>>>
>>> What exactly are you trying to do?
>> On JPEG4_0_3, there are possibly multiple contexts submitting jobs to
>> all cores on the affected instance,
>> so I am changing sched_ready to false and using smp_rmb() to publish
>> to other cores which are trying to
>> submit on this instance , for them to read the updated change
>> immediately and stop submission before
>> I wait for idle on this instance, which makes sure other good
>> contexts of jpeg get a chance to be migrated
>> out to any available instance before reset starts, that way good jpeg
>> contexts get a chance to continue
>> without issues while this instance is being reset and made ready to
>> continue decoding after reset.
>
> Ok, completely NAK to this approach!
>
> First of all setting r->sched.ready to false is a completely no-go. I
> can't remember how often I had to remove that nonsense.
>
> With the exception of the KIQ the back-ends should *never* touch that
> stuff!
>
> Then the Linux kernel requires that ever use of smp_rmb() requires to
> document the matching write memory barrier.
>
> Over all please completely remove that code all together. The wait for
> idle function is *only* supposed to wait for the HW to become idle and
> nothing else.
Okay Christian, Thank you for explaining , I will remove this.
Regards,
Sathish
>
> Regards,
> Christian.
>
>>
>> Regards,
>> Sathish
>>>
>>> Regards,
>>> Christian.
>>>
>>>
>>>> + for (j = 0; j < adev->jpeg.num_jpeg_rings; ++j) {
>>>> + r = &adev->jpeg.inst[ring->me].ring_dec[j];
>>>> + if (r->pipe == j)
>>>> + continue;
>>>> + ret = SOC15_WAIT_ON_RREG_OFFSET(JPEG, GET_INST(JPEG,
>>>> ring->me),
>>>> + regUVD_JRBC0_UVD_JRBC_STATUS,
>>>> + jpeg_v4_0_3_core_reg_offset(j),
>>>> + UVD_JRBC0_UVD_JRBC_STATUS__RB_JOB_DONE_MASK,
>>>> + UVD_JRBC0_UVD_JRBC_STATUS__RB_JOB_DONE_MASK);
>>>> + if (ret)
>>>> + return ret;
>>>> + }
>>>> +
>>>> + return 0;
>>>> +}
>>>> +
>>>> +static int jpeg_v4_0_3_ring_reset(struct amdgpu_ring *ring,
>>>> unsigned int vmid)
>>>> +{
>>>> + struct amdgpu_device *adev = ring->adev;
>>>> + struct amdgpu_ring *r;
>>>> + int ret, j;
>>>> +
>>>> + if (amdgpu_sriov_vf(adev))
>>>> + return -EINVAL;
>>>> +
>>>> + jpeg_v4_0_3_wait_for_idle_on_inst(ring);
>>>> + jpeg_v4_0_3_stop_inst(ring->adev, ring->me);
>>>> + jpeg_v4_0_3_start_inst(ring->adev, ring->me);
>>>> + for (j = 0; j < adev->jpeg.num_jpeg_rings; ++j) {
>>>> + r = &adev->jpeg.inst[ring->me].ring_dec[j];
>>>> + jpeg_v4_0_3_start_jrbc(r);
>>>> + ret = amdgpu_ring_test_helper(r);
>>>> + if (ret)
>>>> + return ret;
>>>> + r->sched.ready = true;
>>>> + }
>>>> + /* publish update */
>>>> + smp_rmb();
>>>> + dev_info(adev->dev, "Reset on %s succeeded\n", ring->sched.name);
>>>> + return 0;
>>>> +}
>>>> +
>>>> static const struct amd_ip_funcs jpeg_v4_0_3_ip_funcs = {
>>>> .name = "jpeg_v4_0_3",
>>>> .early_init = jpeg_v4_0_3_early_init,
>>>> @@ -1145,6 +1198,7 @@ static const struct amdgpu_ring_funcs
>>>> jpeg_v4_0_3_dec_ring_vm_funcs = {
>>>> .emit_wreg = jpeg_v4_0_3_dec_ring_emit_wreg,
>>>> .emit_reg_wait = jpeg_v4_0_3_dec_ring_emit_reg_wait,
>>>> .emit_reg_write_reg_wait =
>>>> amdgpu_ring_emit_reg_write_reg_wait_helper,
>>>> + .reset = jpeg_v4_0_3_ring_reset,
>>>> };
>>>> static void jpeg_v4_0_3_set_dec_ring_funcs(struct amdgpu_device
>>>> *adev)
>>>
>>
>
More information about the amd-gfx
mailing list