[v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp in sc7280

Kalyan Thota kalyant at qti.qualcomm.com
Mon Aug 8 10:44:32 UTC 2022



>-----Original Message-----
>From: Dmitry Baryshkov <dmitry.baryshkov at linaro.org>
>Sent: Thursday, August 4, 2022 9:29 PM
>To: Kalyan Thota (QUIC) <quic_kalyant at quicinc.com>
>Cc: dri-devel at lists.freedesktop.org; linux-arm-msm at vger.kernel.org;
>freedreno at lists.freedesktop.org; devicetree at vger.kernel.org; linux-
>kernel at vger.kernel.org; robdclark at gmail.com; dianders at chromium.org;
>swboyd at chromium.org; Vinod Polimera (QUIC) <quic_vpolimer at quicinc.com>;
>Abhinav Kumar (QUIC) <quic_abhinavk at quicinc.com>
>Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp
>in sc7280
>
>WARNING: This email originated from outside of Qualcomm. Please be wary of
>any links or attachments, and do not enable macros.
>
>On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant at quicinc.com> wrote:
>>
>> Flush mechanism for DSPP blocks has changed in sc7280 family, it
>> allows individual sub blocks to be flushed in coordination with master
>> flush control.
>>
>> representation: master_flush && (PCC_flush | IGC_flush .. etc )
>>
>> This change adds necessary support for the above design.
>>
>> Signed-off-by: Kalyan Thota <quic_kalyant at quicinc.com>
>
>I'd like to land at least patches 6-8 from [1] next cycle. They clean up the CTL
>interface. Could you please rebase your patch on top of them?
>

Sure I'll wait for the series to rebase. @Doug can you comment if this is okay and this patch is not needed immediately ?

>[1] https://patchwork.freedesktop.org/series/99909/
>
>> ---
>>  drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c       |  4 +++
>>  drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c |  5 +++-
>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h |  2 ++
>>  drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c     | 40
>+++++++++++++++++++++++++-
>>  drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h     |  3 ++
>>  drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h    |  7 +++++
>>  6 files changed, 59 insertions(+), 2 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>> index 7763558..4eca317 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
>> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct
>drm_crtc *crtc)
>>                 mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl,
>>                         mixer[i].hw_dspp->idx);
>>
>> +               if(ctl->ops.set_dspp_hierarchical_flush)
>> +                       ctl->ops.set_dspp_hierarchical_flush(ctl,
>> +                                               mixer[i].hw_dspp->idx,
>> + DSPP_SUB_PCC);
>> +
>>                 /* stage config flush mask */
>>                 ctl->ops.update_pending_flush(ctl,
>> mixer[i].flush_mask);
>>
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>> index 021eb2f..3b27a87 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c
>> @@ -58,7 +58,10 @@
>>         (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2))
>>
>>  #define CTL_SC7280_MASK \
>> -       (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) |
>BIT(DPU_CTL_VM_CFG))
>> +       (BIT(DPU_CTL_ACTIVE_CFG) | \
>> +        BIT(DPU_CTL_FETCH_ACTIVE) | \
>> +        BIT(DPU_CTL_VM_CFG) | \
>> +        BIT(DPU_CTL_HIERARCHICAL_FLUSH))
>>
>>  #define MERGE_3D_SM8150_MASK (0)
>>
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>> index b85b24b..7922f6c 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
>> @@ -185,6 +185,7 @@ enum {
>>   * @DPU_CTL_SPLIT_DISPLAY:     CTL supports video mode split display
>>   * @DPU_CTL_FETCH_ACTIVE:      Active CTL for fetch HW (SSPPs)
>>   * @DPU_CTL_VM_CFG:            CTL config to support multiple VMs
>> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical
>> + flush
>>   * @DPU_CTL_MAX
>>   */
>>  enum {
>> @@ -192,6 +193,7 @@ enum {
>>         DPU_CTL_ACTIVE_CFG,
>>         DPU_CTL_FETCH_ACTIVE,
>>         DPU_CTL_VM_CFG,
>> +       DPU_CTL_HIERARCHICAL_FLUSH,
>>         DPU_CTL_MAX
>>  };
>>
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>> index 3584f5e..b34fc30 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
>> @@ -28,6 +28,8 @@
>>  #define   CTL_INTF_FLUSH                0x110
>>  #define   CTL_INTF_MASTER               0x134
>>  #define   CTL_FETCH_PIPE_ACTIVE         0x0FC
>> +#define   CTL_DSPP_0_FLUSH             0x13C
>
>Please change to CTL_DSPP_n_FLUSH(n).
>
>> +
>>
>>  #define CTL_MIXER_BORDER_OUT            BIT(24)
>>  #define CTL_FLUSH_MASK_CTL              BIT(17)
>> @@ -292,6 +294,36 @@ static uint32_t dpu_hw_ctl_get_bitmask_dspp(struct
>dpu_hw_ctl *ctx,
>>         return flushbits;
>>  }
>>
>> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx,
>> +       enum dpu_dspp dspp)
>> +{
>> +       return BIT(29);
>> +}
>> +
>> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx,
>> +       enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) {
>> +       uint32_t flushbits = 0, active = 0;
>> +
>> +       switch (dspp_sub_blk) {
>> +       case DSPP_SUB_IGC:
>> +               flushbits = BIT(2);
>> +               break;
>> +       case DSPP_SUB_PCC:
>> +               flushbits = BIT(4);
>> +               break;
>> +       case DSPP_SUB_GC:
>> +               flushbits = BIT(5);
>> +               break;
>> +       default:
>> +               return;
>> +       }
>> +
>> +       active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1)
>> + * 4));
>
>So that this line will be simpler to read.
>
>> +
>> +       DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4),
>> +active | flushbits); }
>> +
>>  static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, u32
>> timeout_us)  {
>>         struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 @@
>> static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops,
>>         ops->setup_blendstage = dpu_hw_ctl_setup_blendstage;
>>         ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp;
>>         ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer;
>> -       ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp;
>> +       if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) {
>> +               ops->get_bitmask_dspp =
>> + dpu_hw_ctl_get_bitmask_dspp_v1;
>
>We have used _v1 for active CTLs. What is the relationship between
>CTL_HIERARCHILCAL_FLUSH and active CTLs?
Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers in grouping the resources such as WB, INTF, pingpong, DSC etc into the data path
DSPP hierarchical flush will gives us a finer control on which post processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. etc )
These blocks are contained in DSPP package.
>
>> +               ops->set_dspp_hierarchical_flush =
>dpu_hw_ctl_set_dspp_hierarchical_flush;
>> +       } else {
>> +               ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp;
>> +       }
>> +
>>         if (cap & BIT(DPU_CTL_FETCH_ACTIVE))
>>                 ops->set_active_pipes =
>> dpu_hw_ctl_set_fetch_pipe_active;  }; diff --git
>> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>> index ac15444..8ecab91 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
>> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops {
>>         uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx,
>>                 enum dpu_dspp blk);
>>
>> +       void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx,
>> +               enum dpu_dspp blk, enum dpu_dspp_sub_blk
>> + dspp_sub_blk);
>
>The word "hierarchical" means particular (internal) implementation.
>Please change to something like set_dspp_block_flush().
>Or with [2] in place, it can be hidden in the
>update_pending_flush_dspp() function. Just pass the subblock to the function and
>let the dpu_hw_ctl care about it.
>
>[2] https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1
>
>
>> +
>>         /**
>>          * Set all blend stages to disabled
>>          * @ctx       : ctl path ctx pointer
>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>> index bb9cead..561e2ab 100644
>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h
>> @@ -166,6 +166,13 @@ enum dpu_dspp {
>>         DSPP_MAX
>>  };
>>
>> +enum dpu_dspp_sub_blk{
>> +       DSPP_SUB_PCC = 1,
>> +       DSPP_SUB_IGC,
>> +       DSPP_SUB_GC,
>> +       DSPP_SUB_MAX
>> +};
>
>I'd prefer if we can use DPU_DSPP_* definitions instead.
>
>> +
>>  enum dpu_ctl {
>>         CTL_0 = 1,
>>         CTL_1,
>
>
>
>--
>With best wishes
>Dmitry


More information about the dri-devel mailing list