Message ID | 1659608930-4370-1-git-send-email-quic_kalyant@quicinc.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp in sc7280 | expand |
On Thu, Aug 4, 2022 at 3:29 AM Kalyan Thota <quic_kalyant@quicinc.com> wrote: > > +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, > + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) > +{ > + uint32_t flushbits = 0, active = 0; nit: don't init to 0 since you just override below. > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h > index ac15444..8ecab91 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h > @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { > uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, > enum dpu_dspp blk); > > + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, > + enum dpu_dspp blk, enum dpu_dspp_sub_blk dspp_sub_blk); > + Given that most of the items in this list have kernel-doc comments explaining them, probably you should have one for your new one too. -Doug
On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@quicinc.com> wrote: > > Flush mechanism for DSPP blocks has changed in sc7280 family, it > allows individual sub blocks to be flushed in coordination with > master flush control. > > representation: master_flush && (PCC_flush | IGC_flush .. etc ) > > This change adds necessary support for the above design. > > Signed-off-by: Kalyan Thota <quic_kalyant@quicinc.com> I'd like to land at least patches 6-8 from [1] next cycle. They clean up the CTL interface. Could you please rebase your patch on top of them? [1] https://patchwork.freedesktop.org/series/99909/ > --- > drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ > drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- > drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ > drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 +++++++++++++++++++++++++- > drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ > drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ > 6 files changed, 59 insertions(+), 2 deletions(-) > > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c > index 7763558..4eca317 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c > @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct drm_crtc *crtc) > mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, > mixer[i].hw_dspp->idx); > > + if(ctl->ops.set_dspp_hierarchical_flush) > + ctl->ops.set_dspp_hierarchical_flush(ctl, > + mixer[i].hw_dspp->idx, DSPP_SUB_PCC); > + > /* stage config flush mask */ > ctl->ops.update_pending_flush(ctl, mixer[i].flush_mask); > > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c > index 021eb2f..3b27a87 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c > @@ -58,7 +58,10 @@ > (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) > > #define CTL_SC7280_MASK \ > - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | BIT(DPU_CTL_VM_CFG)) > + (BIT(DPU_CTL_ACTIVE_CFG) | \ > + BIT(DPU_CTL_FETCH_ACTIVE) | \ > + BIT(DPU_CTL_VM_CFG) | \ > + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) > > #define MERGE_3D_SM8150_MASK (0) > > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h > index b85b24b..7922f6c 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h > @@ -185,6 +185,7 @@ enum { > * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display > * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) > * @DPU_CTL_VM_CFG: CTL config to support multiple VMs > + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical flush > * @DPU_CTL_MAX > */ > enum { > @@ -192,6 +193,7 @@ enum { > DPU_CTL_ACTIVE_CFG, > DPU_CTL_FETCH_ACTIVE, > DPU_CTL_VM_CFG, > + DPU_CTL_HIERARCHICAL_FLUSH, > DPU_CTL_MAX > }; > > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c > index 3584f5e..b34fc30 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c > @@ -28,6 +28,8 @@ > #define CTL_INTF_FLUSH 0x110 > #define CTL_INTF_MASTER 0x134 > #define CTL_FETCH_PIPE_ACTIVE 0x0FC > +#define CTL_DSPP_0_FLUSH 0x13C Please change to CTL_DSPP_n_FLUSH(n). > + > > #define CTL_MIXER_BORDER_OUT BIT(24) > #define CTL_FLUSH_MASK_CTL BIT(17) > @@ -292,6 +294,36 @@ static uint32_t dpu_hw_ctl_get_bitmask_dspp(struct dpu_hw_ctl *ctx, > return flushbits; > } > > +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, > + enum dpu_dspp dspp) > +{ > + return BIT(29); > +} > + > +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, > + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) > +{ > + uint32_t flushbits = 0, active = 0; > + > + switch (dspp_sub_blk) { > + case DSPP_SUB_IGC: > + flushbits = BIT(2); > + break; > + case DSPP_SUB_PCC: > + flushbits = BIT(4); > + break; > + case DSPP_SUB_GC: > + flushbits = BIT(5); > + break; > + default: > + return; > + } > + > + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4)); So that this line will be simpler to read. > + > + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), active | flushbits); > +} > + > static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, u32 timeout_us) > { > struct dpu_hw_blk_reg_map *c = &ctx->hw; > @@ -600,7 +632,13 @@ static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, > ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; > ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; > ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; > - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; > + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { > + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp_v1; We have used _v1 for active CTLs. What is the relationship between CTL_HIERARCHILCAL_FLUSH and active CTLs? > + ops->set_dspp_hierarchical_flush = dpu_hw_ctl_set_dspp_hierarchical_flush; > + } else { > + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; > + } > + > if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) > ops->set_active_pipes = dpu_hw_ctl_set_fetch_pipe_active; > }; > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h > index ac15444..8ecab91 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h > @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { > uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, > enum dpu_dspp blk); > > + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, > + enum dpu_dspp blk, enum dpu_dspp_sub_blk dspp_sub_blk); The word "hierarchical" means particular (internal) implementation. Please change to something like set_dspp_block_flush(). Or with [2] in place, it can be hidden in the update_pending_flush_dspp() function. Just pass the subblock to the function and let the dpu_hw_ctl care about it. [2] https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1 > + > /** > * Set all blend stages to disabled > * @ctx : ctl path ctx pointer > diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h > index bb9cead..561e2ab 100644 > --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h > +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h > @@ -166,6 +166,13 @@ enum dpu_dspp { > DSPP_MAX > }; > > +enum dpu_dspp_sub_blk{ > + DSPP_SUB_PCC = 1, > + DSPP_SUB_IGC, > + DSPP_SUB_GC, > + DSPP_SUB_MAX > +}; I'd prefer if we can use DPU_DSPP_* definitions instead. > + > enum dpu_ctl { > CTL_0 = 1, > CTL_1,
>-----Original Message----- >From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> >Sent: Thursday, August 4, 2022 9:29 PM >To: Kalyan Thota (QUIC) <quic_kalyant@quicinc.com> >Cc: dri-devel@lists.freedesktop.org; linux-arm-msm@vger.kernel.org; >freedreno@lists.freedesktop.org; devicetree@vger.kernel.org; linux- >kernel@vger.kernel.org; robdclark@gmail.com; dianders@chromium.org; >swboyd@chromium.org; Vinod Polimera (QUIC) <quic_vpolimer@quicinc.com>; >Abhinav Kumar (QUIC) <quic_abhinavk@quicinc.com> >Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp >in sc7280 > >WARNING: This email originated from outside of Qualcomm. Please be wary of >any links or attachments, and do not enable macros. > >On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@quicinc.com> wrote: >> >> Flush mechanism for DSPP blocks has changed in sc7280 family, it >> allows individual sub blocks to be flushed in coordination with master >> flush control. >> >> representation: master_flush && (PCC_flush | IGC_flush .. etc ) >> >> This change adds necessary support for the above design. >> >> Signed-off-by: Kalyan Thota <quic_kalyant@quicinc.com> > >I'd like to land at least patches 6-8 from [1] next cycle. They clean up the CTL >interface. Could you please rebase your patch on top of them? > Sure I'll wait for the series to rebase. @Doug can you comment if this is okay and this patch is not needed immediately ? >[1] https://patchwork.freedesktop.org/series/99909/ > >> --- >> drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 >+++++++++++++++++++++++++- >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ >> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ >> 6 files changed, 59 insertions(+), 2 deletions(-) >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> index 7763558..4eca317 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct >drm_crtc *crtc) >> mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, >> mixer[i].hw_dspp->idx); >> >> + if(ctl->ops.set_dspp_hierarchical_flush) >> + ctl->ops.set_dspp_hierarchical_flush(ctl, >> + mixer[i].hw_dspp->idx, >> + DSPP_SUB_PCC); >> + >> /* stage config flush mask */ >> ctl->ops.update_pending_flush(ctl, >> mixer[i].flush_mask); >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> index 021eb2f..3b27a87 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >> @@ -58,7 +58,10 @@ >> (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) >> >> #define CTL_SC7280_MASK \ >> - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | >BIT(DPU_CTL_VM_CFG)) >> + (BIT(DPU_CTL_ACTIVE_CFG) | \ >> + BIT(DPU_CTL_FETCH_ACTIVE) | \ >> + BIT(DPU_CTL_VM_CFG) | \ >> + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) >> >> #define MERGE_3D_SM8150_MASK (0) >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> index b85b24b..7922f6c 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >> @@ -185,6 +185,7 @@ enum { >> * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display >> * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) >> * @DPU_CTL_VM_CFG: CTL config to support multiple VMs >> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical >> + flush >> * @DPU_CTL_MAX >> */ >> enum { >> @@ -192,6 +193,7 @@ enum { >> DPU_CTL_ACTIVE_CFG, >> DPU_CTL_FETCH_ACTIVE, >> DPU_CTL_VM_CFG, >> + DPU_CTL_HIERARCHICAL_FLUSH, >> DPU_CTL_MAX >> }; >> >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> index 3584f5e..b34fc30 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >> @@ -28,6 +28,8 @@ >> #define CTL_INTF_FLUSH 0x110 >> #define CTL_INTF_MASTER 0x134 >> #define CTL_FETCH_PIPE_ACTIVE 0x0FC >> +#define CTL_DSPP_0_FLUSH 0x13C > >Please change to CTL_DSPP_n_FLUSH(n). > >> + >> >> #define CTL_MIXER_BORDER_OUT BIT(24) >> #define CTL_FLUSH_MASK_CTL BIT(17) >> @@ -292,6 +294,36 @@ static uint32_t dpu_hw_ctl_get_bitmask_dspp(struct >dpu_hw_ctl *ctx, >> return flushbits; >> } >> >> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, >> + enum dpu_dspp dspp) >> +{ >> + return BIT(29); >> +} >> + >> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, >> + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) { >> + uint32_t flushbits = 0, active = 0; >> + >> + switch (dspp_sub_blk) { >> + case DSPP_SUB_IGC: >> + flushbits = BIT(2); >> + break; >> + case DSPP_SUB_PCC: >> + flushbits = BIT(4); >> + break; >> + case DSPP_SUB_GC: >> + flushbits = BIT(5); >> + break; >> + default: >> + return; >> + } >> + >> + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) >> + * 4)); > >So that this line will be simpler to read. > >> + >> + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), >> +active | flushbits); } >> + >> static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, u32 >> timeout_us) { >> struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 @@ >> static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, >> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; >> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; >> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; >> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { >> + ops->get_bitmask_dspp = >> + dpu_hw_ctl_get_bitmask_dspp_v1; > >We have used _v1 for active CTLs. What is the relationship between >CTL_HIERARCHILCAL_FLUSH and active CTLs? Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers in grouping the resources such as WB, INTF, pingpong, DSC etc into the data path DSPP hierarchical flush will gives us a finer control on which post processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. etc ) These blocks are contained in DSPP package. > >> + ops->set_dspp_hierarchical_flush = >dpu_hw_ctl_set_dspp_hierarchical_flush; >> + } else { >> + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >> + } >> + >> if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) >> ops->set_active_pipes = >> dpu_hw_ctl_set_fetch_pipe_active; }; diff --git >> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> index ac15444..8ecab91 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { >> uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, >> enum dpu_dspp blk); >> >> + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, >> + enum dpu_dspp blk, enum dpu_dspp_sub_blk >> + dspp_sub_blk); > >The word "hierarchical" means particular (internal) implementation. >Please change to something like set_dspp_block_flush(). >Or with [2] in place, it can be hidden in the >update_pending_flush_dspp() function. Just pass the subblock to the function and >let the dpu_hw_ctl care about it. > >[2] https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1 > > >> + >> /** >> * Set all blend stages to disabled >> * @ctx : ctl path ctx pointer >> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> index bb9cead..561e2ab 100644 >> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >> @@ -166,6 +166,13 @@ enum dpu_dspp { >> DSPP_MAX >> }; >> >> +enum dpu_dspp_sub_blk{ >> + DSPP_SUB_PCC = 1, >> + DSPP_SUB_IGC, >> + DSPP_SUB_GC, >> + DSPP_SUB_MAX >> +}; > >I'd prefer if we can use DPU_DSPP_* definitions instead. > >> + >> enum dpu_ctl { >> CTL_0 = 1, >> CTL_1, > > > >-- >With best wishes >Dmitry
Hi, On Mon, Aug 8, 2022 at 3:44 AM Kalyan Thota <kalyant@qti.qualcomm.com> wrote: > > >I'd like to land at least patches 6-8 from [1] next cycle. They clean up the CTL > >interface. Could you please rebase your patch on top of them? > > > > Sure I'll wait for the series to rebase. @Doug can you comment if this is okay and this patch is not needed immediately ? > > >[1] https://patchwork.freedesktop.org/series/99909/ I don't personally see a problem basing them atop a cleanup. If the patches Dmitry points at are targeted for the next cycle then that seems like a pretty reasonable timeframe to me. -Doug
On 08/08/2022 13:44, Kalyan Thota wrote: > > >> -----Original Message----- >> From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> >> Sent: Thursday, August 4, 2022 9:29 PM >> To: Kalyan Thota (QUIC) <quic_kalyant@quicinc.com> >> Cc: dri-devel@lists.freedesktop.org; linux-arm-msm@vger.kernel.org; >> freedreno@lists.freedesktop.org; devicetree@vger.kernel.org; linux- >> kernel@vger.kernel.org; robdclark@gmail.com; dianders@chromium.org; >> swboyd@chromium.org; Vinod Polimera (QUIC) <quic_vpolimer@quicinc.com>; >> Abhinav Kumar (QUIC) <quic_abhinavk@quicinc.com> >> Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp >> in sc7280 >> >> WARNING: This email originated from outside of Qualcomm. Please be wary of >> any links or attachments, and do not enable macros. >> >> On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@quicinc.com> wrote: >>> >>> Flush mechanism for DSPP blocks has changed in sc7280 family, it >>> allows individual sub blocks to be flushed in coordination with master >>> flush control. >>> >>> representation: master_flush && (PCC_flush | IGC_flush .. etc ) >>> >>> This change adds necessary support for the above design. >>> >>> Signed-off-by: Kalyan Thota <quic_kalyant@quicinc.com> >> >> I'd like to land at least patches 6-8 from [1] next cycle. They clean up the CTL >> interface. Could you please rebase your patch on top of them? >> > > Sure I'll wait for the series to rebase. @Doug can you comment if this is okay and this patch is not needed immediately ? The respective patches have been picked up for 6.1 and were pushed to https://gitlab.freedesktop.org/lumag/msm.git msm-next-lumag . Could you please rebase your patch on top of them? All other comments also needs addressing. > >> [1] https://patchwork.freedesktop.org/series/99909/ >> >>> --- >>> drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ >>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- >>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ >>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 >> +++++++++++++++++++++++++- >>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ >>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ >>> 6 files changed, 59 insertions(+), 2 deletions(-) >>> >>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>> index 7763558..4eca317 100644 >>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct >> drm_crtc *crtc) >>> mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, >>> mixer[i].hw_dspp->idx); >>> >>> + if(ctl->ops.set_dspp_hierarchical_flush) >>> + ctl->ops.set_dspp_hierarchical_flush(ctl, >>> + mixer[i].hw_dspp->idx, >>> + DSPP_SUB_PCC); >>> + >>> /* stage config flush mask */ >>> ctl->ops.update_pending_flush(ctl, >>> mixer[i].flush_mask); >>> >>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>> index 021eb2f..3b27a87 100644 >>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>> @@ -58,7 +58,10 @@ >>> (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) >>> >>> #define CTL_SC7280_MASK \ >>> - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | >> BIT(DPU_CTL_VM_CFG)) >>> + (BIT(DPU_CTL_ACTIVE_CFG) | \ >>> + BIT(DPU_CTL_FETCH_ACTIVE) | \ >>> + BIT(DPU_CTL_VM_CFG) | \ >>> + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) >>> >>> #define MERGE_3D_SM8150_MASK (0) >>> >>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>> index b85b24b..7922f6c 100644 >>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>> @@ -185,6 +185,7 @@ enum { >>> * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display >>> * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) >>> * @DPU_CTL_VM_CFG: CTL config to support multiple VMs >>> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical >>> + flush >>> * @DPU_CTL_MAX >>> */ >>> enum { >>> @@ -192,6 +193,7 @@ enum { >>> DPU_CTL_ACTIVE_CFG, >>> DPU_CTL_FETCH_ACTIVE, >>> DPU_CTL_VM_CFG, >>> + DPU_CTL_HIERARCHICAL_FLUSH, >>> DPU_CTL_MAX >>> }; >>> >>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>> index 3584f5e..b34fc30 100644 >>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>> @@ -28,6 +28,8 @@ >>> #define CTL_INTF_FLUSH 0x110 >>> #define CTL_INTF_MASTER 0x134 >>> #define CTL_FETCH_PIPE_ACTIVE 0x0FC >>> +#define CTL_DSPP_0_FLUSH 0x13C >> >> Please change to CTL_DSPP_n_FLUSH(n). >> >>> + >>> >>> #define CTL_MIXER_BORDER_OUT BIT(24) >>> #define CTL_FLUSH_MASK_CTL BIT(17) >>> @@ -292,6 +294,36 @@ static uint32_t dpu_hw_ctl_get_bitmask_dspp(struct >> dpu_hw_ctl *ctx, >>> return flushbits; >>> } >>> >>> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, >>> + enum dpu_dspp dspp) >>> +{ >>> + return BIT(29); >>> +} >>> + >>> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, >>> + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) { >>> + uint32_t flushbits = 0, active = 0; >>> + >>> + switch (dspp_sub_blk) { >>> + case DSPP_SUB_IGC: >>> + flushbits = BIT(2); >>> + break; >>> + case DSPP_SUB_PCC: >>> + flushbits = BIT(4); >>> + break; >>> + case DSPP_SUB_GC: >>> + flushbits = BIT(5); >>> + break; >>> + default: >>> + return; >>> + } >>> + >>> + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) >>> + * 4)); >> >> So that this line will be simpler to read. >> >>> + >>> + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), >>> +active | flushbits); } >>> + >>> static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, u32 >>> timeout_us) { >>> struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 @@ >>> static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, >>> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; >>> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; >>> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; >>> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >>> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { >>> + ops->get_bitmask_dspp = >>> + dpu_hw_ctl_get_bitmask_dspp_v1; >> >> We have used _v1 for active CTLs. What is the relationship between >> CTL_HIERARCHILCAL_FLUSH and active CTLs? > Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers in grouping the resources such as WB, INTF, pingpong, DSC etc into the data path > DSPP hierarchical flush will gives us a finer control on which post processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. etc ) > These blocks are contained in DSPP package. So, I assume that hierarchical DSPP flush does not exist on non-active CTL SoCs. Which supported SoCs do support the hierarchichal DSPP flush? >> >>> + ops->set_dspp_hierarchical_flush = >> dpu_hw_ctl_set_dspp_hierarchical_flush; >>> + } else { >>> + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >>> + } >>> + >>> if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) >>> ops->set_active_pipes = >>> dpu_hw_ctl_set_fetch_pipe_active; }; diff --git >>> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>> index ac15444..8ecab91 100644 >>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { >>> uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, >>> enum dpu_dspp blk); >>> >>> + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, >>> + enum dpu_dspp blk, enum dpu_dspp_sub_blk >>> + dspp_sub_blk); >> >> The word "hierarchical" means particular (internal) implementation. >> Please change to something like set_dspp_block_flush(). >> Or with [2] in place, it can be hidden in the >> update_pending_flush_dspp() function. Just pass the subblock to the function and >> let the dpu_hw_ctl care about it. >> >> [2] https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1 >> >> >>> + >>> /** >>> * Set all blend stages to disabled >>> * @ctx : ctl path ctx pointer >>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>> index bb9cead..561e2ab 100644 >>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>> @@ -166,6 +166,13 @@ enum dpu_dspp { >>> DSPP_MAX >>> }; >>> >>> +enum dpu_dspp_sub_blk{ >>> + DSPP_SUB_PCC = 1, >>> + DSPP_SUB_IGC, >>> + DSPP_SUB_GC, >>> + DSPP_SUB_MAX >>> +}; >> >> I'd prefer if we can use DPU_DSPP_* definitions instead. >> >>> + >>> enum dpu_ctl { >>> CTL_0 = 1, >>> CTL_1, >> >> >> >> -- >> With best wishes >> Dmitry
>-----Original Message----- >From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> >Sent: Friday, August 26, 2022 5:02 PM >To: Kalyan Thota <kalyant@qti.qualcomm.com>; Kalyan Thota (QUIC) ><quic_kalyant@quicinc.com>; dianders@chromium.org >Cc: dri-devel@lists.freedesktop.org; linux-arm-msm@vger.kernel.org; >freedreno@lists.freedesktop.org; devicetree@vger.kernel.org; linux- >kernel@vger.kernel.org; robdclark@gmail.com; swboyd@chromium.org; Vinod >Polimera (QUIC) <quic_vpolimer@quicinc.com>; Abhinav Kumar (QUIC) ><quic_abhinavk@quicinc.com> >Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp >in sc7280 > >WARNING: This email originated from outside of Qualcomm. Please be wary of >any links or attachments, and do not enable macros. > >On 08/08/2022 13:44, Kalyan Thota wrote: >> >> >>> -----Original Message----- >>> From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> >>> Sent: Thursday, August 4, 2022 9:29 PM >>> To: Kalyan Thota (QUIC) <quic_kalyant@quicinc.com> >>> Cc: dri-devel@lists.freedesktop.org; linux-arm-msm@vger.kernel.org; >>> freedreno@lists.freedesktop.org; devicetree@vger.kernel.org; linux- >>> kernel@vger.kernel.org; robdclark@gmail.com; dianders@chromium.org; >>> swboyd@chromium.org; Vinod Polimera (QUIC) >>> <quic_vpolimer@quicinc.com>; Abhinav Kumar (QUIC) >>> <quic_abhinavk@quicinc.com> >>> Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical >>> flush for dspp in sc7280 >>> >>> WARNING: This email originated from outside of Qualcomm. Please be >>> wary of any links or attachments, and do not enable macros. >>> >>> On Thu, 4 Aug 2022 at 13:29, Kalyan Thota <quic_kalyant@quicinc.com> >wrote: >>>> >>>> Flush mechanism for DSPP blocks has changed in sc7280 family, it >>>> allows individual sub blocks to be flushed in coordination with >>>> master flush control. >>>> >>>> representation: master_flush && (PCC_flush | IGC_flush .. etc ) >>>> >>>> This change adds necessary support for the above design. >>>> >>>> Signed-off-by: Kalyan Thota <quic_kalyant@quicinc.com> >>> >>> I'd like to land at least patches 6-8 from [1] next cycle. They clean >>> up the CTL interface. Could you please rebase your patch on top of them? >>> >> >> Sure I'll wait for the series to rebase. @Doug can you comment if this is okay >and this patch is not needed immediately ? > >The respective patches have been picked up for 6.1 and were pushed to >https://gitlab.freedesktop.org/lumag/msm.git msm-next-lumag . Could you >please rebase your patch on top of them? > >All other comments also needs addressing. > >> >>> [1] https://patchwork.freedesktop.org/series/99909/ >>> >>>> --- >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 >>> +++++++++++++++++++++++++- >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ >>>> drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ >>>> 6 files changed, 59 insertions(+), 2 deletions(-) >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> index 7763558..4eca317 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c >>>> @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct >>> drm_crtc *crtc) >>>> mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, >>>> mixer[i].hw_dspp->idx); >>>> >>>> + if(ctl->ops.set_dspp_hierarchical_flush) >>>> + ctl->ops.set_dspp_hierarchical_flush(ctl, >>>> + >>>> + mixer[i].hw_dspp->idx, DSPP_SUB_PCC); >>>> + >>>> /* stage config flush mask */ >>>> ctl->ops.update_pending_flush(ctl, >>>> mixer[i].flush_mask); >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> index 021eb2f..3b27a87 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c >>>> @@ -58,7 +58,10 @@ >>>> (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) >>>> >>>> #define CTL_SC7280_MASK \ >>>> - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | >>> BIT(DPU_CTL_VM_CFG)) >>>> + (BIT(DPU_CTL_ACTIVE_CFG) | \ >>>> + BIT(DPU_CTL_FETCH_ACTIVE) | \ >>>> + BIT(DPU_CTL_VM_CFG) | \ >>>> + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) >>>> >>>> #define MERGE_3D_SM8150_MASK (0) >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> index b85b24b..7922f6c 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h >>>> @@ -185,6 +185,7 @@ enum { >>>> * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display >>>> * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) >>>> * @DPU_CTL_VM_CFG: CTL config to support multiple VMs >>>> + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical >>>> + flush >>>> * @DPU_CTL_MAX >>>> */ >>>> enum { >>>> @@ -192,6 +193,7 @@ enum { >>>> DPU_CTL_ACTIVE_CFG, >>>> DPU_CTL_FETCH_ACTIVE, >>>> DPU_CTL_VM_CFG, >>>> + DPU_CTL_HIERARCHICAL_FLUSH, >>>> DPU_CTL_MAX >>>> }; >>>> >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> index 3584f5e..b34fc30 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c >>>> @@ -28,6 +28,8 @@ >>>> #define CTL_INTF_FLUSH 0x110 >>>> #define CTL_INTF_MASTER 0x134 >>>> #define CTL_FETCH_PIPE_ACTIVE 0x0FC >>>> +#define CTL_DSPP_0_FLUSH 0x13C >>> >>> Please change to CTL_DSPP_n_FLUSH(n). >>> >>>> + >>>> >>>> #define CTL_MIXER_BORDER_OUT BIT(24) >>>> #define CTL_FLUSH_MASK_CTL BIT(17) >>>> @@ -292,6 +294,36 @@ static uint32_t >>>> dpu_hw_ctl_get_bitmask_dspp(struct >>> dpu_hw_ctl *ctx, >>>> return flushbits; >>>> } >>>> >>>> +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, >>>> + enum dpu_dspp dspp) >>>> +{ >>>> + return BIT(29); >>>> +} >>>> + >>>> +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, >>>> + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) { >>>> + uint32_t flushbits = 0, active = 0; >>>> + >>>> + switch (dspp_sub_blk) { >>>> + case DSPP_SUB_IGC: >>>> + flushbits = BIT(2); >>>> + break; >>>> + case DSPP_SUB_PCC: >>>> + flushbits = BIT(4); >>>> + break; >>>> + case DSPP_SUB_GC: >>>> + flushbits = BIT(5); >>>> + break; >>>> + default: >>>> + return; >>>> + } >>>> + >>>> + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - >>>> + 1) >>>> + * 4)); >>> >>> So that this line will be simpler to read. >>> >>>> + >>>> + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), >>>> +active | flushbits); } >>>> + >>>> static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, >>>> u32 >>>> timeout_us) { >>>> struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 >>>> @@ static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, >>>> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; >>>> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; >>>> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; >>>> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >>>> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { >>>> + ops->get_bitmask_dspp = >>>> + dpu_hw_ctl_get_bitmask_dspp_v1; >>> >>> We have used _v1 for active CTLs. What is the relationship between >>> CTL_HIERARCHILCAL_FLUSH and active CTLs? >> Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers >> in grouping the resources such as WB, INTF, pingpong, DSC etc into the >> data path DSPP hierarchical flush will gives us a finer control on which post >processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. >etc ) These blocks are contained in DSPP package. > >So, I assume that hierarchical DSPP flush does not exist on non-active CTL SoCs. >Which supported SoCs do support the hierarchichal DSPP flush? > Dspp Sub-block flush is supported from the DPU 7-series, the only target in the catalogue is sc7280. >>> >>>> + ops->set_dspp_hierarchical_flush = >>> dpu_hw_ctl_set_dspp_hierarchical_flush; >>>> + } else { >>>> + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; >>>> + } >>>> + >>>> if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) >>>> ops->set_active_pipes = >>>> dpu_hw_ctl_set_fetch_pipe_active; }; diff --git >>>> a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> index ac15444..8ecab91 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h >>>> @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { >>>> uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, >>>> enum dpu_dspp blk); >>>> >>>> + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, >>>> + enum dpu_dspp blk, enum dpu_dspp_sub_blk >>>> + dspp_sub_blk); >>> >>> The word "hierarchical" means particular (internal) implementation. >>> Please change to something like set_dspp_block_flush(). >>> Or with [2] in place, it can be hidden in the >>> update_pending_flush_dspp() function. Just pass the subblock to the >>> function and let the dpu_hw_ctl care about it. >>> >>> [2] >>> https://patchwork.freedesktop.org/patch/473159/?series=99909&rev=1 >>> >>> >>>> + >>>> /** >>>> * Set all blend stages to disabled >>>> * @ctx : ctl path ctx pointer >>>> diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> index bb9cead..561e2ab 100644 >>>> --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h >>>> @@ -166,6 +166,13 @@ enum dpu_dspp { >>>> DSPP_MAX >>>> }; >>>> >>>> +enum dpu_dspp_sub_blk{ >>>> + DSPP_SUB_PCC = 1, >>>> + DSPP_SUB_IGC, >>>> + DSPP_SUB_GC, >>>> + DSPP_SUB_MAX >>>> +}; >>> >>> I'd prefer if we can use DPU_DSPP_* definitions instead. >>> >>>> + >>>> enum dpu_ctl { >>>> CTL_0 = 1, >>>> CTL_1, >>> >>> >>> >>> -- >>> With best wishes >>> Dmitry > >-- >With best wishes >Dmitry
On Fri, 2 Sept 2022 at 12:35, Kalyan Thota <kalyant@qti.qualcomm.com> wrote: > > > > >-----Original Message----- > >From: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> > >Sent: Friday, August 26, 2022 5:02 PM > >To: Kalyan Thota <kalyant@qti.qualcomm.com>; Kalyan Thota (QUIC) > ><quic_kalyant@quicinc.com>; dianders@chromium.org > >Cc: dri-devel@lists.freedesktop.org; linux-arm-msm@vger.kernel.org; > >freedreno@lists.freedesktop.org; devicetree@vger.kernel.org; linux- > >kernel@vger.kernel.org; robdclark@gmail.com; swboyd@chromium.org; Vinod > >Polimera (QUIC) <quic_vpolimer@quicinc.com>; Abhinav Kumar (QUIC) > ><quic_abhinavk@quicinc.com> > >Subject: Re: [v1] drm/msm/disp/dpu1: add support for hierarchical flush for dspp > >in sc7280 Ugh. I'd kindly ask to fix your email client to behave like a normal one. > >>>> @@ static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, > >>>> ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; > >>>> ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; > >>>> ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; > >>>> - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; > >>>> + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { > >>>> + ops->get_bitmask_dspp = > >>>> + dpu_hw_ctl_get_bitmask_dspp_v1; > >>> > >>> We have used _v1 for active CTLs. What is the relationship between > >>> CTL_HIERARCHILCAL_FLUSH and active CTLs? > >> Active CTL design replaces legacy CTL_MEM_SEL, CTL_OUT_SEL registers > >> in grouping the resources such as WB, INTF, pingpong, DSC etc into the > >> data path DSPP hierarchical flush will gives us a finer control on which post > >processing blocks to be flushed as part of the composition ( like IGC, PCC, GC .. > >etc ) These blocks are contained in DSPP package. > > > >So, I assume that hierarchical DSPP flush does not exist on non-active CTL SoCs. > >Which supported SoCs do support the hierarchichal DSPP flush? > > > Dspp Sub-block flush is supported from the DPU 7-series, the only target in the catalogue is sc7280. Ack, thanks.
diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c index 7763558..4eca317 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c @@ -703,6 +703,10 @@ static void _dpu_crtc_setup_cp_blocks(struct drm_crtc *crtc) mixer[i].flush_mask |= ctl->ops.get_bitmask_dspp(ctl, mixer[i].hw_dspp->idx); + if(ctl->ops.set_dspp_hierarchical_flush) + ctl->ops.set_dspp_hierarchical_flush(ctl, + mixer[i].hw_dspp->idx, DSPP_SUB_PCC); + /* stage config flush mask */ ctl->ops.update_pending_flush(ctl, mixer[i].flush_mask); diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c index 021eb2f..3b27a87 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c @@ -58,7 +58,10 @@ (PINGPONG_SDM845_MASK | BIT(DPU_PINGPONG_TE2)) #define CTL_SC7280_MASK \ - (BIT(DPU_CTL_ACTIVE_CFG) | BIT(DPU_CTL_FETCH_ACTIVE) | BIT(DPU_CTL_VM_CFG)) + (BIT(DPU_CTL_ACTIVE_CFG) | \ + BIT(DPU_CTL_FETCH_ACTIVE) | \ + BIT(DPU_CTL_VM_CFG) | \ + BIT(DPU_CTL_HIERARCHICAL_FLUSH)) #define MERGE_3D_SM8150_MASK (0) diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h index b85b24b..7922f6c 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h @@ -185,6 +185,7 @@ enum { * @DPU_CTL_SPLIT_DISPLAY: CTL supports video mode split display * @DPU_CTL_FETCH_ACTIVE: Active CTL for fetch HW (SSPPs) * @DPU_CTL_VM_CFG: CTL config to support multiple VMs + * @DPU_CTL_HIERARCHICAL_FLUSH: CTL config to support hierarchical flush * @DPU_CTL_MAX */ enum { @@ -192,6 +193,7 @@ enum { DPU_CTL_ACTIVE_CFG, DPU_CTL_FETCH_ACTIVE, DPU_CTL_VM_CFG, + DPU_CTL_HIERARCHICAL_FLUSH, DPU_CTL_MAX }; diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c index 3584f5e..b34fc30 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c @@ -28,6 +28,8 @@ #define CTL_INTF_FLUSH 0x110 #define CTL_INTF_MASTER 0x134 #define CTL_FETCH_PIPE_ACTIVE 0x0FC +#define CTL_DSPP_0_FLUSH 0x13C + #define CTL_MIXER_BORDER_OUT BIT(24) #define CTL_FLUSH_MASK_CTL BIT(17) @@ -292,6 +294,36 @@ static uint32_t dpu_hw_ctl_get_bitmask_dspp(struct dpu_hw_ctl *ctx, return flushbits; } +static uint32_t dpu_hw_ctl_get_bitmask_dspp_v1(struct dpu_hw_ctl *ctx, + enum dpu_dspp dspp) +{ + return BIT(29); +} + +static void dpu_hw_ctl_set_dspp_hierarchical_flush(struct dpu_hw_ctl *ctx, + enum dpu_dspp dspp, enum dpu_dspp_sub_blk dspp_sub_blk) +{ + uint32_t flushbits = 0, active = 0; + + switch (dspp_sub_blk) { + case DSPP_SUB_IGC: + flushbits = BIT(2); + break; + case DSPP_SUB_PCC: + flushbits = BIT(4); + break; + case DSPP_SUB_GC: + flushbits = BIT(5); + break; + default: + return; + } + + active = DPU_REG_READ(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4)); + + DPU_REG_WRITE(&ctx->hw, CTL_DSPP_0_FLUSH + ((dspp - 1) * 4), active | flushbits); +} + static u32 dpu_hw_ctl_poll_reset_status(struct dpu_hw_ctl *ctx, u32 timeout_us) { struct dpu_hw_blk_reg_map *c = &ctx->hw; @@ -600,7 +632,13 @@ static void _setup_ctl_ops(struct dpu_hw_ctl_ops *ops, ops->setup_blendstage = dpu_hw_ctl_setup_blendstage; ops->get_bitmask_sspp = dpu_hw_ctl_get_bitmask_sspp; ops->get_bitmask_mixer = dpu_hw_ctl_get_bitmask_mixer; - ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; + if (cap & BIT(DPU_CTL_HIERARCHICAL_FLUSH)) { + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp_v1; + ops->set_dspp_hierarchical_flush = dpu_hw_ctl_set_dspp_hierarchical_flush; + } else { + ops->get_bitmask_dspp = dpu_hw_ctl_get_bitmask_dspp; + } + if (cap & BIT(DPU_CTL_FETCH_ACTIVE)) ops->set_active_pipes = dpu_hw_ctl_set_fetch_pipe_active; }; diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h index ac15444..8ecab91 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h @@ -160,6 +160,9 @@ struct dpu_hw_ctl_ops { uint32_t (*get_bitmask_dspp)(struct dpu_hw_ctl *ctx, enum dpu_dspp blk); + void (*set_dspp_hierarchical_flush)(struct dpu_hw_ctl *ctx, + enum dpu_dspp blk, enum dpu_dspp_sub_blk dspp_sub_blk); + /** * Set all blend stages to disabled * @ctx : ctl path ctx pointer diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h index bb9cead..561e2ab 100644 --- a/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h +++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h @@ -166,6 +166,13 @@ enum dpu_dspp { DSPP_MAX }; +enum dpu_dspp_sub_blk{ + DSPP_SUB_PCC = 1, + DSPP_SUB_IGC, + DSPP_SUB_GC, + DSPP_SUB_MAX +}; + enum dpu_ctl { CTL_0 = 1, CTL_1,
Flush mechanism for DSPP blocks has changed in sc7280 family, it allows individual sub blocks to be flushed in coordination with master flush control. representation: master_flush && (PCC_flush | IGC_flush .. etc ) This change adds necessary support for the above design. Signed-off-by: Kalyan Thota <quic_kalyant@quicinc.com> --- drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c | 4 +++ drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.c | 5 +++- drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h | 2 ++ drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c | 40 +++++++++++++++++++++++++- drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h | 3 ++ drivers/gpu/drm/msm/disp/dpu1/dpu_hw_mdss.h | 7 +++++ 6 files changed, 59 insertions(+), 2 deletions(-)