Message ID | 1584105134-13583-3-git-send-email-akashast@codeaurora.org (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | Add interconnect support to QSPI and QUP drivers | expand |
Hi Akash, On Fri, Mar 13, 2020 at 06:42:08PM +0530, Akash Asthana wrote: > Add necessary macros and structure variables to support ICC BW > voting from individual SE drivers. > > Signed-off-by: Akash Asthana <akashast@codeaurora.org> > --- > Changes in V2: > - As per Bjorn's comment dropped enums for ICC paths, given the three > paths individual members > > include/linux/qcom-geni-se.h | 28 ++++++++++++++++++++++++++++ > 1 file changed, 28 insertions(+) > > diff --git a/include/linux/qcom-geni-se.h b/include/linux/qcom-geni-se.h > index dd46494..eaae16e 100644 > --- a/include/linux/qcom-geni-se.h > +++ b/include/linux/qcom-geni-se.h > @@ -6,6 +6,8 @@ > #ifndef _LINUX_QCOM_GENI_SE > #define _LINUX_QCOM_GENI_SE > > +#include <linux/interconnect.h> > + > /* Transfer mode supported by GENI Serial Engines */ > enum geni_se_xfer_mode { > GENI_SE_INVALID, > @@ -33,6 +35,15 @@ struct clk; > * @clk: Handle to the core serial engine clock > * @num_clk_levels: Number of valid clock levels in clk_perf_tbl > * @clk_perf_tbl: Table of clock frequency input to serial engine clock > + * @icc_path_geni_to_core: ICC path handle for geni to core > + * @icc_path_cpu_to_geni: ICC path handle for cpu to geni > + * @icc_path_geni_to_ddr: ICC path handle for geni to ddr > + * @avg_bw_core: Average bus bandwidth value for QUP core 2x clock > + * @peak_bw_core: Peak bus bandwidth value for QUP core 2x clock > + * @avg_bw_cpu: Average bus bandwidth value for CPU > + * @peak_bw_cpu: Peak bus bandwidth value for CPU > + * @avg_bw_ddr: Average bus bandwidth value for DDR > + * @peak_bw_ddr: Peak bus bandwidth value for DDR > */ > struct geni_se { > void __iomem *base; > @@ -41,6 +52,15 @@ struct geni_se { > struct clk *clk; > unsigned int num_clk_levels; > unsigned long *clk_perf_tbl; > + struct icc_path *icc_path_geni_to_core; > + struct icc_path *icc_path_cpu_to_geni; > + struct icc_path *icc_path_geni_to_ddr; > + unsigned int avg_bw_core; > + unsigned int peak_bw_core; > + unsigned int avg_bw_cpu; > + unsigned int peak_bw_cpu; > + unsigned int avg_bw_ddr; > + unsigned int peak_bw_ddr; Those are a lot of new individual struct members. How about clustering them, e.g.: struct geni_icc_path { struct icc_path *path; unsigned int avg_bw; unsigned int peak_bw; }; struct geni_iccs_paths { struct geni_icc_path to_core; struct geni_icc_path from_cpu; struct geni_icc_path to_ddr; }; And 'struct geni_se' just gets this entry: struct geni_icc_paths icc; or alternatively three 'struct geni_icc_path' entries.
Hi Matthias, On 3/13/2020 10:12 PM, Matthias Kaehlcke wrote: > Hi Akash, > > On Fri, Mar 13, 2020 at 06:42:08PM +0530, Akash Asthana wrote: >> Add necessary macros and structure variables to support ICC BW >> voting from individual SE drivers. >> >> Signed-off-by: Akash Asthana <akashast@codeaurora.org> >> --- >> Changes in V2: >> - As per Bjorn's comment dropped enums for ICC paths, given the three >> paths individual members >> >> include/linux/qcom-geni-se.h | 28 ++++++++++++++++++++++++++++ >> 1 file changed, 28 insertions(+) >> >> diff --git a/include/linux/qcom-geni-se.h b/include/linux/qcom-geni-se.h >> index dd46494..eaae16e 100644 >> --- a/include/linux/qcom-geni-se.h >> +++ b/include/linux/qcom-geni-se.h >> @@ -6,6 +6,8 @@ >> #ifndef _LINUX_QCOM_GENI_SE >> #define _LINUX_QCOM_GENI_SE >> >> +#include <linux/interconnect.h> >> + >> /* Transfer mode supported by GENI Serial Engines */ >> enum geni_se_xfer_mode { >> GENI_SE_INVALID, >> @@ -33,6 +35,15 @@ struct clk; >> * @clk: Handle to the core serial engine clock >> * @num_clk_levels: Number of valid clock levels in clk_perf_tbl >> * @clk_perf_tbl: Table of clock frequency input to serial engine clock >> + * @icc_path_geni_to_core: ICC path handle for geni to core >> + * @icc_path_cpu_to_geni: ICC path handle for cpu to geni >> + * @icc_path_geni_to_ddr: ICC path handle for geni to ddr >> + * @avg_bw_core: Average bus bandwidth value for QUP core 2x clock >> + * @peak_bw_core: Peak bus bandwidth value for QUP core 2x clock >> + * @avg_bw_cpu: Average bus bandwidth value for CPU >> + * @peak_bw_cpu: Peak bus bandwidth value for CPU >> + * @avg_bw_ddr: Average bus bandwidth value for DDR >> + * @peak_bw_ddr: Peak bus bandwidth value for DDR >> */ >> struct geni_se { >> void __iomem *base; >> @@ -41,6 +52,15 @@ struct geni_se { >> struct clk *clk; >> unsigned int num_clk_levels; >> unsigned long *clk_perf_tbl; >> + struct icc_path *icc_path_geni_to_core; >> + struct icc_path *icc_path_cpu_to_geni; >> + struct icc_path *icc_path_geni_to_ddr; >> + unsigned int avg_bw_core; >> + unsigned int peak_bw_core; >> + unsigned int avg_bw_cpu; >> + unsigned int peak_bw_cpu; >> + unsigned int avg_bw_ddr; >> + unsigned int peak_bw_ddr; > Those are a lot of new individual struct members. How about clustering > them, e.g.: > > struct geni_icc_path { > struct icc_path *path; > unsigned int avg_bw; > unsigned int peak_bw; > }; I guess it would be better to add this structure ICC driver as you suggested@https://patchwork.kernel.org/patch/11436905/. > struct geni_iccs_paths { > struct geni_icc_path to_core; > struct geni_icc_path from_cpu; > struct geni_icc_path to_ddr; > }; > > And 'struct geni_se' just gets this entry: > > struct geni_icc_paths icc; > > or alternatively three 'struct geni_icc_path' entries. ok Thanks for reviewing. Regards Akash
On Fri, Mar 13, 2020 at 6:12 AM Akash Asthana <akashast@codeaurora.org> wrote: > > Add necessary macros and structure variables to support ICC BW > voting from individual SE drivers. > > Signed-off-by: Akash Asthana <akashast@codeaurora.org> > --- > Changes in V2: > - As per Bjorn's comment dropped enums for ICC paths, given the three > paths individual members > > include/linux/qcom-geni-se.h | 28 ++++++++++++++++++++++++++++ > 1 file changed, 28 insertions(+) > > diff --git a/include/linux/qcom-geni-se.h b/include/linux/qcom-geni-se.h > index dd46494..eaae16e 100644 > --- a/include/linux/qcom-geni-se.h > +++ b/include/linux/qcom-geni-se.h > @@ -6,6 +6,8 @@ > #ifndef _LINUX_QCOM_GENI_SE > #define _LINUX_QCOM_GENI_SE > > +#include <linux/interconnect.h> > + > /* Transfer mode supported by GENI Serial Engines */ > enum geni_se_xfer_mode { > GENI_SE_INVALID, > @@ -33,6 +35,15 @@ struct clk; > * @clk: Handle to the core serial engine clock > * @num_clk_levels: Number of valid clock levels in clk_perf_tbl > * @clk_perf_tbl: Table of clock frequency input to serial engine clock > + * @icc_path_geni_to_core: ICC path handle for geni to core > + * @icc_path_cpu_to_geni: ICC path handle for cpu to geni > + * @icc_path_geni_to_ddr: ICC path handle for geni to ddr > + * @avg_bw_core: Average bus bandwidth value for QUP core 2x clock > + * @peak_bw_core: Peak bus bandwidth value for QUP core 2x clock > + * @avg_bw_cpu: Average bus bandwidth value for CPU > + * @peak_bw_cpu: Peak bus bandwidth value for CPU > + * @avg_bw_ddr: Average bus bandwidth value for DDR > + * @peak_bw_ddr: Peak bus bandwidth value for DDR > */ > struct geni_se { > void __iomem *base; > @@ -41,6 +52,15 @@ struct geni_se { > struct clk *clk; > unsigned int num_clk_levels; > unsigned long *clk_perf_tbl; > + struct icc_path *icc_path_geni_to_core; > + struct icc_path *icc_path_cpu_to_geni; > + struct icc_path *icc_path_geni_to_ddr; > + unsigned int avg_bw_core; > + unsigned int peak_bw_core; > + unsigned int avg_bw_cpu; > + unsigned int peak_bw_cpu; > + unsigned int avg_bw_ddr; > + unsigned int peak_bw_ddr; > }; > > /* Common SE registers */ > @@ -229,6 +249,14 @@ struct geni_se { > #define GENI_SE_VERSION_MINOR(ver) ((ver & HW_VER_MINOR_MASK) >> HW_VER_MINOR_SHFT) > #define GENI_SE_VERSION_STEP(ver) (ver & HW_VER_STEP_MASK) > > +/* Core 2X clock frequency to BCM threshold mapping */ > +#define CORE_2X_19_2_MHZ 960 > +#define CORE_2X_50_MHZ 2500 > +#define CORE_2X_100_MHZ 5000 > +#define CORE_2X_150_MHZ 7500 > +#define CORE_2X_200_MHZ 10000 > +#define CORE_2X_236_MHZ 16383 These are all just 50 * clock_rate. Can you instead specify that one define of CLK_TO_BW_RATIO 50, and then use clk_get_rate() to get the input clock frequency. That way, if these end up getting clocked at a different rate, the bandwidth also scales appropriately. Also, can you enumerate why 50 is an appropriate ratio? -Evan -Evan
On Fri, Mar 20, 2020 at 4:03 AM Akash Asthana <akashast@codeaurora.org> wrote: > > Hi Evan, > > +/* Core 2X clock frequency to BCM threshold mapping */ > +#define CORE_2X_19_2_MHZ 960 > +#define CORE_2X_50_MHZ 2500 > +#define CORE_2X_100_MHZ 5000 > +#define CORE_2X_150_MHZ 7500 > +#define CORE_2X_200_MHZ 10000 > +#define CORE_2X_236_MHZ 16383 > > These are all just 50 * clock_rate. Can you instead specify that one > define of CLK_TO_BW_RATIO 50, and then use clk_get_rate() to get the > input clock frequency. That way, if these end up getting clocked at a > different rate, the bandwidth also scales appropriately. Also, can you > enumerate why 50 is an appropriate ratio? > -Evan > > -Evan > > Clock rate for Core 2X is controlled by BW voting only, we don't set clock rate for core 2X clock either by DFS or calling clk_set_rate API like we do for SE clocks from individual driver. > > In DT node it's not mentioned as clock. > > As discussed in patch@ https://patchwork.kernel.org/patch/11436897/ We are not scaling Core 2X clock based on dynamic need of driver instead we are putting recommended value from HW team for each driver. Oh I get it. This is pretty opaque, since this table is saying "here are the bandwidth values that happen to work out to a Core2X clock rate of N". But it's not obvious why setting the Core2X clock rate to N is desirable or appropriate. The answer seems to be hardware guys told us these thresholds work well in practice. And if I'm reading into it more, probably they're saying these bandwidths are too low to be worth dynamically managing beyond on/off. At the very least we should explain some of this in the comment above these defines. Something like: /* Define bandwidth thresholds that cause the underlying Core 2X interconnect clock to run at the named frequency. These baseline values are recommended by the hardware team, and are not dynamically scaled with GENI bandwidth beyond basic on/off. */ -Evan
Hi Evan, On 3/20/2020 10:15 PM, Evan Green wrote: > On Fri, Mar 20, 2020 at 4:03 AM Akash Asthana <akashast@codeaurora.org> wrote: >> Hi Evan, >> >> +/* Core 2X clock frequency to BCM threshold mapping */ >> +#define CORE_2X_19_2_MHZ 960 >> +#define CORE_2X_50_MHZ 2500 >> +#define CORE_2X_100_MHZ 5000 >> +#define CORE_2X_150_MHZ 7500 >> +#define CORE_2X_200_MHZ 10000 >> +#define CORE_2X_236_MHZ 16383 >> >> These are all just 50 * clock_rate. Can you instead specify that one >> define of CLK_TO_BW_RATIO 50, and then use clk_get_rate() to get the >> input clock frequency. That way, if these end up getting clocked at a >> different rate, the bandwidth also scales appropriately. Also, can you >> enumerate why 50 is an appropriate ratio? >> -Evan >> >> -Evan >> >> Clock rate for Core 2X is controlled by BW voting only, we don't set clock rate for core 2X clock either by DFS or calling clk_set_rate API like we do for SE clocks from individual driver. >> >> In DT node it's not mentioned as clock. >> >> As discussed in patch@ https://patchwork.kernel.org/patch/11436897/ We are not scaling Core 2X clock based on dynamic need of driver instead we are putting recommended value from HW team for each driver. > Oh I get it. This is pretty opaque, since this table is saying "here > are the bandwidth values that happen to work out to a Core2X clock > rate of N". Hmm, BCM threshold to CORE2X clock rate mapping is exposed to us from clock team. BCM threshold value is internally convert to mentioned clock rate this is something internal to board ICC driver. > But it's not obvious why setting the Core2X clock rate to > N is desirable or appropriate. The answer seems to be hardware guys > told us these thresholds work well in practice. Yes, this is correct as the core clocks behaves different than any other NOC, we rely on the recommendation from VI/HW team. > And if I'm reading > into it more, probably they're saying these bandwidths are too low to > be worth dynamically managing beyond on/off I am not sure whether they intend to say this. > At the very least we should explain some of this in the comment above > these defines. Something like: > /* Define bandwidth thresholds that cause the underlying Core 2X > interconnect clock to run at the named frequency. These baseline > values are recommended by the hardware team, and are not dynamically > scaled with GENI bandwidth beyond basic on/off. */ > -Evan Ok, regards, Akash
diff --git a/include/linux/qcom-geni-se.h b/include/linux/qcom-geni-se.h index dd46494..eaae16e 100644 --- a/include/linux/qcom-geni-se.h +++ b/include/linux/qcom-geni-se.h @@ -6,6 +6,8 @@ #ifndef _LINUX_QCOM_GENI_SE #define _LINUX_QCOM_GENI_SE +#include <linux/interconnect.h> + /* Transfer mode supported by GENI Serial Engines */ enum geni_se_xfer_mode { GENI_SE_INVALID, @@ -33,6 +35,15 @@ struct clk; * @clk: Handle to the core serial engine clock * @num_clk_levels: Number of valid clock levels in clk_perf_tbl * @clk_perf_tbl: Table of clock frequency input to serial engine clock + * @icc_path_geni_to_core: ICC path handle for geni to core + * @icc_path_cpu_to_geni: ICC path handle for cpu to geni + * @icc_path_geni_to_ddr: ICC path handle for geni to ddr + * @avg_bw_core: Average bus bandwidth value for QUP core 2x clock + * @peak_bw_core: Peak bus bandwidth value for QUP core 2x clock + * @avg_bw_cpu: Average bus bandwidth value for CPU + * @peak_bw_cpu: Peak bus bandwidth value for CPU + * @avg_bw_ddr: Average bus bandwidth value for DDR + * @peak_bw_ddr: Peak bus bandwidth value for DDR */ struct geni_se { void __iomem *base; @@ -41,6 +52,15 @@ struct geni_se { struct clk *clk; unsigned int num_clk_levels; unsigned long *clk_perf_tbl; + struct icc_path *icc_path_geni_to_core; + struct icc_path *icc_path_cpu_to_geni; + struct icc_path *icc_path_geni_to_ddr; + unsigned int avg_bw_core; + unsigned int peak_bw_core; + unsigned int avg_bw_cpu; + unsigned int peak_bw_cpu; + unsigned int avg_bw_ddr; + unsigned int peak_bw_ddr; }; /* Common SE registers */ @@ -229,6 +249,14 @@ struct geni_se { #define GENI_SE_VERSION_MINOR(ver) ((ver & HW_VER_MINOR_MASK) >> HW_VER_MINOR_SHFT) #define GENI_SE_VERSION_STEP(ver) (ver & HW_VER_STEP_MASK) +/* Core 2X clock frequency to BCM threshold mapping */ +#define CORE_2X_19_2_MHZ 960 +#define CORE_2X_50_MHZ 2500 +#define CORE_2X_100_MHZ 5000 +#define CORE_2X_150_MHZ 7500 +#define CORE_2X_200_MHZ 10000 +#define CORE_2X_236_MHZ 16383 + #if IS_ENABLED(CONFIG_QCOM_GENI_SE) u32 geni_se_get_qup_hw_version(struct geni_se *se);
Add necessary macros and structure variables to support ICC BW voting from individual SE drivers. Signed-off-by: Akash Asthana <akashast@codeaurora.org> --- Changes in V2: - As per Bjorn's comment dropped enums for ICC paths, given the three paths individual members include/linux/qcom-geni-se.h | 28 ++++++++++++++++++++++++++++ 1 file changed, 28 insertions(+)