Message ID | 6adbec8e4757f3b6c9f47135544a0302f8e7c55c.1565369764.git.robin.murphy@arm.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Arm SMMU refactoring | expand |
Tested-by: Krishna Reddy<vdumpa@nvidia.com> Validated the entire patch set on Tegra194 SOC based platform and confirmed that arm-smmu driver is functional as it has been. -KR -----Original Message----- From: Robin Murphy <robin.murphy@arm.com> Sent: Friday, August 9, 2019 10:08 AM To: will@kernel.org Cc: iommu@lists.linux-foundation.org; linux-arm-kernel@lists.infradead.org; joro@8bytes.org; vivek.gautam@codeaurora.org; bjorn.andersson@linaro.org; Krishna Reddy <vdumpa@nvidia.com>; gregory.clement@bootlin.com; robdclark@gmail.com Subject: [PATCH 15/15] iommu/arm-smmu: Add context init implementation hook Allocating and initialising a context for a domain is another point where certain implementations are known to want special behaviour. Currently the other half of the Cavium workaround comes into play here, so let's finish the job to get the whole thing right out of the way. Signed-off-by: Robin Murphy <robin.murphy@arm.com> --- drivers/iommu/arm-smmu-impl.c | 39 +++++++++++++++++++++++++-- drivers/iommu/arm-smmu.c | 51 +++++++---------------------------- drivers/iommu/arm-smmu.h | 42 +++++++++++++++++++++++++++-- 3 files changed, 86 insertions(+), 46 deletions(-) diff --git a/drivers/iommu/arm-smmu-impl.c b/drivers/iommu/arm-smmu-impl.c index c8904da08354..7a657d47b6ec 100644 --- a/drivers/iommu/arm-smmu-impl.c +++ b/drivers/iommu/arm-smmu-impl.c @@ -48,6 +48,12 @@ const struct arm_smmu_impl calxeda_impl = { }; +struct cavium_smmu { + struct arm_smmu_device smmu; + u32 id_base; +}; +#define to_csmmu(s) container_of(s, struct cavium_smmu, smmu) + static int cavium_cfg_probe(struct arm_smmu_device *smmu) { static atomic_t context_count = ATOMIC_INIT(0); @@ -56,17 +62,46 @@ static int cavium_cfg_probe(struct arm_smmu_device *smmu) * Ensure ASID and VMID allocation is unique across all SMMUs in * the system. */ - smmu->cavium_id_base = atomic_fetch_add(smmu->num_context_banks, + to_csmmu(smmu)->id_base = atomic_fetch_add(smmu->num_context_banks, &context_count); dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n"); return 0; } +int cavium_init_context(struct arm_smmu_domain *smmu_domain) { + u32 id_base = to_csmmu(smmu_domain->smmu)->id_base; + + if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) + smmu_domain->cfg.vmid += id_base; + else + smmu_domain->cfg.asid += id_base; + + return 0; +} + const struct arm_smmu_impl cavium_impl = { .cfg_probe = cavium_cfg_probe, + .init_context = cavium_init_context, }; +struct arm_smmu_device *cavium_smmu_impl_init(struct arm_smmu_device +*smmu) { + struct cavium_smmu *csmmu; + + csmmu = devm_kzalloc(smmu->dev, sizeof(*csmmu), GFP_KERNEL); + if (!csmmu) + return ERR_PTR(-ENOMEM); + + csmmu->smmu = *smmu; + csmmu->smmu.impl = &cavium_impl; + + devm_kfree(smmu->dev, smmu); + + return &csmmu->smmu; +} + #define ARM_MMU500_ACTLR_CPRE (1 << 1) @@ -121,7 +156,7 @@ struct arm_smmu_device *arm_smmu_impl_init(struct arm_smmu_device *smmu) smmu->impl = &calxeda_impl; if (smmu->model == CAVIUM_SMMUV2) - smmu->impl = &cavium_impl; + return cavium_smmu_impl_init(smmu); if (smmu->model == ARM_MMU500) smmu->impl = &arm_mmu500_impl; diff --git a/drivers/iommu/arm-smmu.c b/drivers/iommu/arm-smmu.c index 298ab9e6a6cd..1c1c9ef91d7b 100644 --- a/drivers/iommu/arm-smmu.c +++ b/drivers/iommu/arm-smmu.c @@ -27,7 +27,6 @@ #include <linux/interrupt.h> #include <linux/io.h> #include <linux/io-64-nonatomic-hi-lo.h> -#include <linux/io-pgtable.h> #include <linux/iopoll.h> #include <linux/init.h> #include <linux/moduleparam.h> @@ -111,44 +110,6 @@ struct arm_smmu_master_cfg { #define for_each_cfg_sme(fw, i, idx) \ for (i = 0; idx = fwspec_smendx(fw, i), i < fw->num_ids; ++i) -enum arm_smmu_context_fmt { - ARM_SMMU_CTX_FMT_NONE, - ARM_SMMU_CTX_FMT_AARCH64, - ARM_SMMU_CTX_FMT_AARCH32_L, - ARM_SMMU_CTX_FMT_AARCH32_S, -}; - -struct arm_smmu_cfg { - u8 cbndx; - u8 irptndx; - union { - u16 asid; - u16 vmid; - }; - enum arm_smmu_cbar_type cbar; - enum arm_smmu_context_fmt fmt; -}; -#define INVALID_IRPTNDX 0xff - -enum arm_smmu_domain_stage { - ARM_SMMU_DOMAIN_S1 = 0, - ARM_SMMU_DOMAIN_S2, - ARM_SMMU_DOMAIN_NESTED, - ARM_SMMU_DOMAIN_BYPASS, -}; - -struct arm_smmu_domain { - struct arm_smmu_device *smmu; - struct io_pgtable_ops *pgtbl_ops; - const struct iommu_gather_ops *tlb_ops; - struct arm_smmu_cfg cfg; - enum arm_smmu_domain_stage stage; - bool non_strict; - struct mutex init_mutex; /* Protects smmu pointer */ - spinlock_t cb_lock; /* Serialises ATS1* ops and TLB syncs */ - struct iommu_domain domain; -}; - static bool using_legacy_binding, using_generic_binding; static inline int arm_smmu_rpm_get(struct arm_smmu_device *smmu) @@ -749,9 +710,16 @@ static int arm_smmu_init_domain_context(struct iommu_domain *domain, } if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) - cfg->vmid = cfg->cbndx + 1 + smmu->cavium_id_base; + cfg->vmid = cfg->cbndx + 1; else - cfg->asid = cfg->cbndx + smmu->cavium_id_base; + cfg->asid = cfg->cbndx; + + smmu_domain->smmu = smmu; + if (smmu->impl && smmu->impl->init_context) { + ret = smmu->impl->init_context(smmu_domain); + if (ret) + goto out_unlock; + } pgtbl_cfg = (struct io_pgtable_cfg) { .pgsize_bitmap = smmu->pgsize_bitmap, @@ -765,7 +733,6 @@ static int arm_smmu_init_domain_context(struct iommu_domain *domain, if (smmu_domain->non_strict) pgtbl_cfg.quirks |= IO_PGTABLE_QUIRK_NON_STRICT; - smmu_domain->smmu = smmu; pgtbl_ops = alloc_io_pgtable_ops(fmt, &pgtbl_cfg, smmu_domain); if (!pgtbl_ops) { ret = -ENOMEM; diff --git a/drivers/iommu/arm-smmu.h b/drivers/iommu/arm-smmu.h index 616cc87a05e3..a18b5925b43c 100644 --- a/drivers/iommu/arm-smmu.h +++ b/drivers/iommu/arm-smmu.h @@ -14,6 +14,7 @@ #include <linux/bits.h> #include <linux/clk.h> #include <linux/device.h> +#include <linux/io-pgtable.h> #include <linux/iommu.h> #include <linux/mutex.h> #include <linux/spinlock.h> @@ -270,14 +271,50 @@ struct arm_smmu_device { struct clk_bulk_data *clks; int num_clks; - u32 cavium_id_base; /* Specific to Cavium */ - spinlock_t global_sync_lock; /* IOMMU core code handle */ struct iommu_device iommu; }; +enum arm_smmu_context_fmt { + ARM_SMMU_CTX_FMT_NONE, + ARM_SMMU_CTX_FMT_AARCH64, + ARM_SMMU_CTX_FMT_AARCH32_L, + ARM_SMMU_CTX_FMT_AARCH32_S, +}; + +struct arm_smmu_cfg { + u8 cbndx; + u8 irptndx; + union { + u16 asid; + u16 vmid; + }; + enum arm_smmu_cbar_type cbar; + enum arm_smmu_context_fmt fmt; +}; +#define INVALID_IRPTNDX 0xff + +enum arm_smmu_domain_stage { + ARM_SMMU_DOMAIN_S1 = 0, + ARM_SMMU_DOMAIN_S2, + ARM_SMMU_DOMAIN_NESTED, + ARM_SMMU_DOMAIN_BYPASS, +}; + +struct arm_smmu_domain { + struct arm_smmu_device *smmu; + struct io_pgtable_ops *pgtbl_ops; + const struct iommu_gather_ops *tlb_ops; + struct arm_smmu_cfg cfg; + enum arm_smmu_domain_stage stage; + bool non_strict; + struct mutex init_mutex; /* Protects smmu pointer */ + spinlock_t cb_lock; /* Serialises ATS1* ops and TLB syncs */ + struct iommu_domain domain; +}; + /* Implementation details, yay! */ struct arm_smmu_impl { @@ -289,6 +326,7 @@ struct arm_smmu_impl { u64 val); int (*cfg_probe)(struct arm_smmu_device *smmu); int (*reset)(struct arm_smmu_device *smmu); + int (*init_context)(struct arm_smmu_domain *smmu_domain); }; static inline void __iomem *arm_smmu_page(struct arm_smmu_device *smmu, int n) -- 2.21.0.dirty
On Fri, Aug 09, 2019 at 06:07:52PM +0100, Robin Murphy wrote: > Allocating and initialising a context for a domain is another point > where certain implementations are known to want special behaviour. > Currently the other half of the Cavium workaround comes into play here, > so let's finish the job to get the whole thing right out of the way. > > Signed-off-by: Robin Murphy <robin.murphy@arm.com> > --- > drivers/iommu/arm-smmu-impl.c | 39 +++++++++++++++++++++++++-- > drivers/iommu/arm-smmu.c | 51 +++++++---------------------------- > drivers/iommu/arm-smmu.h | 42 +++++++++++++++++++++++++++-- > 3 files changed, 86 insertions(+), 46 deletions(-) > > diff --git a/drivers/iommu/arm-smmu-impl.c b/drivers/iommu/arm-smmu-impl.c > index c8904da08354..7a657d47b6ec 100644 > --- a/drivers/iommu/arm-smmu-impl.c > +++ b/drivers/iommu/arm-smmu-impl.c > @@ -48,6 +48,12 @@ const struct arm_smmu_impl calxeda_impl = { > }; > > > +struct cavium_smmu { > + struct arm_smmu_device smmu; > + u32 id_base; > +}; > +#define to_csmmu(s) container_of(s, struct cavium_smmu, smmu) To be honest with you, I'd just use container_of directly for the two callsites that need it. "to_csmmu" isn't a great name when we're also got the calxeda thing in here. > static int cavium_cfg_probe(struct arm_smmu_device *smmu) > { > static atomic_t context_count = ATOMIC_INIT(0); > @@ -56,17 +62,46 @@ static int cavium_cfg_probe(struct arm_smmu_device *smmu) > * Ensure ASID and VMID allocation is unique across all SMMUs in > * the system. > */ > - smmu->cavium_id_base = atomic_fetch_add(smmu->num_context_banks, > + to_csmmu(smmu)->id_base = atomic_fetch_add(smmu->num_context_banks, > &context_count); > dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n"); > > return 0; > } > > +int cavium_init_context(struct arm_smmu_domain *smmu_domain) > +{ > + u32 id_base = to_csmmu(smmu_domain->smmu)->id_base; > + > + if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) > + smmu_domain->cfg.vmid += id_base; > + else > + smmu_domain->cfg.asid += id_base; > + > + return 0; > +} > + > const struct arm_smmu_impl cavium_impl = { > .cfg_probe = cavium_cfg_probe, > + .init_context = cavium_init_context, > }; > > +struct arm_smmu_device *cavium_smmu_impl_init(struct arm_smmu_device *smmu) > +{ > + struct cavium_smmu *csmmu; > + > + csmmu = devm_kzalloc(smmu->dev, sizeof(*csmmu), GFP_KERNEL); > + if (!csmmu) > + return ERR_PTR(-ENOMEM); > + > + csmmu->smmu = *smmu; > + csmmu->smmu.impl = &cavium_impl; > + > + devm_kfree(smmu->dev, smmu); > + > + return &csmmu->smmu; > +} > + > > #define ARM_MMU500_ACTLR_CPRE (1 << 1) > > @@ -121,7 +156,7 @@ struct arm_smmu_device *arm_smmu_impl_init(struct arm_smmu_device *smmu) > smmu->impl = &calxeda_impl; > > if (smmu->model == CAVIUM_SMMUV2) > - smmu->impl = &cavium_impl; > + return cavium_smmu_impl_init(smmu); > > if (smmu->model == ARM_MMU500) > smmu->impl = &arm_mmu500_impl; Maybe rework this so we do the calxeda detection first (and return if we match), followed by a switch on smmu->model to make it crystal clear that we match only one? Will
On 15/08/2019 11:56, Will Deacon wrote: > On Fri, Aug 09, 2019 at 06:07:52PM +0100, Robin Murphy wrote: >> Allocating and initialising a context for a domain is another point >> where certain implementations are known to want special behaviour. >> Currently the other half of the Cavium workaround comes into play here, >> so let's finish the job to get the whole thing right out of the way. >> >> Signed-off-by: Robin Murphy <robin.murphy@arm.com> >> --- >> drivers/iommu/arm-smmu-impl.c | 39 +++++++++++++++++++++++++-- >> drivers/iommu/arm-smmu.c | 51 +++++++---------------------------- >> drivers/iommu/arm-smmu.h | 42 +++++++++++++++++++++++++++-- >> 3 files changed, 86 insertions(+), 46 deletions(-) >> >> diff --git a/drivers/iommu/arm-smmu-impl.c b/drivers/iommu/arm-smmu-impl.c >> index c8904da08354..7a657d47b6ec 100644 >> --- a/drivers/iommu/arm-smmu-impl.c >> +++ b/drivers/iommu/arm-smmu-impl.c >> @@ -48,6 +48,12 @@ const struct arm_smmu_impl calxeda_impl = { >> }; >> >> >> +struct cavium_smmu { >> + struct arm_smmu_device smmu; >> + u32 id_base; >> +}; >> +#define to_csmmu(s) container_of(s, struct cavium_smmu, smmu) > > To be honest with you, I'd just use container_of directly for the two > callsites that need it. "to_csmmu" isn't a great name when we're also got > the calxeda thing in here. Sure, by this point I was mostly just going for completeness in terms of sketching out an example for subclassing arm_smmu_device. The Tegra patches will now serve as a more complete example anyway, so indeed we can live without it here. >> static int cavium_cfg_probe(struct arm_smmu_device *smmu) >> { >> static atomic_t context_count = ATOMIC_INIT(0); >> @@ -56,17 +62,46 @@ static int cavium_cfg_probe(struct arm_smmu_device *smmu) >> * Ensure ASID and VMID allocation is unique across all SMMUs in >> * the system. >> */ >> - smmu->cavium_id_base = atomic_fetch_add(smmu->num_context_banks, >> + to_csmmu(smmu)->id_base = atomic_fetch_add(smmu->num_context_banks, >> &context_count); >> dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n"); >> >> return 0; >> } >> >> +int cavium_init_context(struct arm_smmu_domain *smmu_domain) >> +{ >> + u32 id_base = to_csmmu(smmu_domain->smmu)->id_base; >> + >> + if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) >> + smmu_domain->cfg.vmid += id_base; >> + else >> + smmu_domain->cfg.asid += id_base; >> + >> + return 0; >> +} >> + >> const struct arm_smmu_impl cavium_impl = { >> .cfg_probe = cavium_cfg_probe, >> + .init_context = cavium_init_context, >> }; >> >> +struct arm_smmu_device *cavium_smmu_impl_init(struct arm_smmu_device *smmu) >> +{ >> + struct cavium_smmu *csmmu; >> + >> + csmmu = devm_kzalloc(smmu->dev, sizeof(*csmmu), GFP_KERNEL); >> + if (!csmmu) >> + return ERR_PTR(-ENOMEM); >> + >> + csmmu->smmu = *smmu; >> + csmmu->smmu.impl = &cavium_impl; >> + >> + devm_kfree(smmu->dev, smmu); >> + >> + return &csmmu->smmu; >> +} >> + >> >> #define ARM_MMU500_ACTLR_CPRE (1 << 1) >> >> @@ -121,7 +156,7 @@ struct arm_smmu_device *arm_smmu_impl_init(struct arm_smmu_device *smmu) >> smmu->impl = &calxeda_impl; >> >> if (smmu->model == CAVIUM_SMMUV2) >> - smmu->impl = &cavium_impl; >> + return cavium_smmu_impl_init(smmu); >> >> if (smmu->model == ARM_MMU500) >> smmu->impl = &arm_mmu500_impl; > > Maybe rework this so we do the calxeda detection first (and return if we > match), followed by a switch on smmu->model to make it crystal clear that > we match only one? As I see it, "match only one" is really only a short-term thing, though, so I didn't want to get *too* hung up on it. Ultimately we're going to have cases where we need to combine e.g. MMU-500 implementation quirks with platform integration quirks - I've been mostly planning on coming back to think about that (and potentially rework this whole logic) later, but I guess it wouldn't hurt to plan out a bit more structure from the start. I'll have a hack on that (and all the other comments) today and hopefully have a v2 by tomorrow. Robin.
On Thu, Aug 15, 2019 at 01:09:07PM +0100, Robin Murphy wrote: > On 15/08/2019 11:56, Will Deacon wrote: > >On Fri, Aug 09, 2019 at 06:07:52PM +0100, Robin Murphy wrote: > >>Allocating and initialising a context for a domain is another point > >>where certain implementations are known to want special behaviour. > >>Currently the other half of the Cavium workaround comes into play here, > >>so let's finish the job to get the whole thing right out of the way. > >> > >>Signed-off-by: Robin Murphy <robin.murphy@arm.com> > >>--- > >> drivers/iommu/arm-smmu-impl.c | 39 +++++++++++++++++++++++++-- > >> drivers/iommu/arm-smmu.c | 51 +++++++---------------------------- > >> drivers/iommu/arm-smmu.h | 42 +++++++++++++++++++++++++++-- > >> 3 files changed, 86 insertions(+), 46 deletions(-) > >> > >>diff --git a/drivers/iommu/arm-smmu-impl.c b/drivers/iommu/arm-smmu-impl.c > >>index c8904da08354..7a657d47b6ec 100644 > >>--- a/drivers/iommu/arm-smmu-impl.c > >>+++ b/drivers/iommu/arm-smmu-impl.c > >>@@ -48,6 +48,12 @@ const struct arm_smmu_impl calxeda_impl = { > >> }; > >>+struct cavium_smmu { > >>+ struct arm_smmu_device smmu; > >>+ u32 id_base; > >>+}; > >>+#define to_csmmu(s) container_of(s, struct cavium_smmu, smmu) > > > >To be honest with you, I'd just use container_of directly for the two > >callsites that need it. "to_csmmu" isn't a great name when we're also got > >the calxeda thing in here. > > Sure, by this point I was mostly just going for completeness in terms of > sketching out an example for subclassing arm_smmu_device. The Tegra patches > will now serve as a more complete example anyway, so indeed we can live > without it here. > > >> static int cavium_cfg_probe(struct arm_smmu_device *smmu) > >> { > >> static atomic_t context_count = ATOMIC_INIT(0); > >>@@ -56,17 +62,46 @@ static int cavium_cfg_probe(struct arm_smmu_device *smmu) > >> * Ensure ASID and VMID allocation is unique across all SMMUs in > >> * the system. > >> */ > >>- smmu->cavium_id_base = atomic_fetch_add(smmu->num_context_banks, > >>+ to_csmmu(smmu)->id_base = atomic_fetch_add(smmu->num_context_banks, > >> &context_count); > >> dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n"); > >> return 0; > >> } > >>+int cavium_init_context(struct arm_smmu_domain *smmu_domain) > >>+{ > >>+ u32 id_base = to_csmmu(smmu_domain->smmu)->id_base; > >>+ > >>+ if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) > >>+ smmu_domain->cfg.vmid += id_base; > >>+ else > >>+ smmu_domain->cfg.asid += id_base; > >>+ > >>+ return 0; > >>+} > >>+ > >> const struct arm_smmu_impl cavium_impl = { > >> .cfg_probe = cavium_cfg_probe, > >>+ .init_context = cavium_init_context, > >> }; > >>+struct arm_smmu_device *cavium_smmu_impl_init(struct arm_smmu_device *smmu) > >>+{ > >>+ struct cavium_smmu *csmmu; > >>+ > >>+ csmmu = devm_kzalloc(smmu->dev, sizeof(*csmmu), GFP_KERNEL); > >>+ if (!csmmu) > >>+ return ERR_PTR(-ENOMEM); > >>+ > >>+ csmmu->smmu = *smmu; > >>+ csmmu->smmu.impl = &cavium_impl; > >>+ > >>+ devm_kfree(smmu->dev, smmu); > >>+ > >>+ return &csmmu->smmu; > >>+} > >>+ > >> #define ARM_MMU500_ACTLR_CPRE (1 << 1) > >>@@ -121,7 +156,7 @@ struct arm_smmu_device *arm_smmu_impl_init(struct arm_smmu_device *smmu) > >> smmu->impl = &calxeda_impl; > >> if (smmu->model == CAVIUM_SMMUV2) > >>- smmu->impl = &cavium_impl; > >>+ return cavium_smmu_impl_init(smmu); > >> if (smmu->model == ARM_MMU500) > >> smmu->impl = &arm_mmu500_impl; > > > >Maybe rework this so we do the calxeda detection first (and return if we > >match), followed by a switch on smmu->model to make it crystal clear that > >we match only one? > > As I see it, "match only one" is really only a short-term thing, though, so > I didn't want to get *too* hung up on it. Ultimately we're going to have > cases where we need to combine e.g. MMU-500 implementation quirks with > platform integration quirks - I've been mostly planning on coming back to > think about that (and potentially rework this whole logic) later, but I > guess it wouldn't hurt to plan out a bit more structure from the start. I was going to ask something similar. I'm guessing that the intent is that we'll eventually we'll have a couple of arm-smmu-<vendor>.c files and we'll need some sort of centralized place to set up the smmu->impl pointer. I had figured that it would be table based or something, but you make a good point about mixing and matching different workarounds. I don't really have a solution, just something I'm pondering while I'm thinking about how to start merging some of the qcom stuff into this. Jordan
diff --git a/drivers/iommu/arm-smmu-impl.c b/drivers/iommu/arm-smmu-impl.c index c8904da08354..7a657d47b6ec 100644 --- a/drivers/iommu/arm-smmu-impl.c +++ b/drivers/iommu/arm-smmu-impl.c @@ -48,6 +48,12 @@ const struct arm_smmu_impl calxeda_impl = { }; +struct cavium_smmu { + struct arm_smmu_device smmu; + u32 id_base; +}; +#define to_csmmu(s) container_of(s, struct cavium_smmu, smmu) + static int cavium_cfg_probe(struct arm_smmu_device *smmu) { static atomic_t context_count = ATOMIC_INIT(0); @@ -56,17 +62,46 @@ static int cavium_cfg_probe(struct arm_smmu_device *smmu) * Ensure ASID and VMID allocation is unique across all SMMUs in * the system. */ - smmu->cavium_id_base = atomic_fetch_add(smmu->num_context_banks, + to_csmmu(smmu)->id_base = atomic_fetch_add(smmu->num_context_banks, &context_count); dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n"); return 0; } +int cavium_init_context(struct arm_smmu_domain *smmu_domain) +{ + u32 id_base = to_csmmu(smmu_domain->smmu)->id_base; + + if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) + smmu_domain->cfg.vmid += id_base; + else + smmu_domain->cfg.asid += id_base; + + return 0; +} + const struct arm_smmu_impl cavium_impl = { .cfg_probe = cavium_cfg_probe, + .init_context = cavium_init_context, }; +struct arm_smmu_device *cavium_smmu_impl_init(struct arm_smmu_device *smmu) +{ + struct cavium_smmu *csmmu; + + csmmu = devm_kzalloc(smmu->dev, sizeof(*csmmu), GFP_KERNEL); + if (!csmmu) + return ERR_PTR(-ENOMEM); + + csmmu->smmu = *smmu; + csmmu->smmu.impl = &cavium_impl; + + devm_kfree(smmu->dev, smmu); + + return &csmmu->smmu; +} + #define ARM_MMU500_ACTLR_CPRE (1 << 1) @@ -121,7 +156,7 @@ struct arm_smmu_device *arm_smmu_impl_init(struct arm_smmu_device *smmu) smmu->impl = &calxeda_impl; if (smmu->model == CAVIUM_SMMUV2) - smmu->impl = &cavium_impl; + return cavium_smmu_impl_init(smmu); if (smmu->model == ARM_MMU500) smmu->impl = &arm_mmu500_impl; diff --git a/drivers/iommu/arm-smmu.c b/drivers/iommu/arm-smmu.c index 298ab9e6a6cd..1c1c9ef91d7b 100644 --- a/drivers/iommu/arm-smmu.c +++ b/drivers/iommu/arm-smmu.c @@ -27,7 +27,6 @@ #include <linux/interrupt.h> #include <linux/io.h> #include <linux/io-64-nonatomic-hi-lo.h> -#include <linux/io-pgtable.h> #include <linux/iopoll.h> #include <linux/init.h> #include <linux/moduleparam.h> @@ -111,44 +110,6 @@ struct arm_smmu_master_cfg { #define for_each_cfg_sme(fw, i, idx) \ for (i = 0; idx = fwspec_smendx(fw, i), i < fw->num_ids; ++i) -enum arm_smmu_context_fmt { - ARM_SMMU_CTX_FMT_NONE, - ARM_SMMU_CTX_FMT_AARCH64, - ARM_SMMU_CTX_FMT_AARCH32_L, - ARM_SMMU_CTX_FMT_AARCH32_S, -}; - -struct arm_smmu_cfg { - u8 cbndx; - u8 irptndx; - union { - u16 asid; - u16 vmid; - }; - enum arm_smmu_cbar_type cbar; - enum arm_smmu_context_fmt fmt; -}; -#define INVALID_IRPTNDX 0xff - -enum arm_smmu_domain_stage { - ARM_SMMU_DOMAIN_S1 = 0, - ARM_SMMU_DOMAIN_S2, - ARM_SMMU_DOMAIN_NESTED, - ARM_SMMU_DOMAIN_BYPASS, -}; - -struct arm_smmu_domain { - struct arm_smmu_device *smmu; - struct io_pgtable_ops *pgtbl_ops; - const struct iommu_gather_ops *tlb_ops; - struct arm_smmu_cfg cfg; - enum arm_smmu_domain_stage stage; - bool non_strict; - struct mutex init_mutex; /* Protects smmu pointer */ - spinlock_t cb_lock; /* Serialises ATS1* ops and TLB syncs */ - struct iommu_domain domain; -}; - static bool using_legacy_binding, using_generic_binding; static inline int arm_smmu_rpm_get(struct arm_smmu_device *smmu) @@ -749,9 +710,16 @@ static int arm_smmu_init_domain_context(struct iommu_domain *domain, } if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2) - cfg->vmid = cfg->cbndx + 1 + smmu->cavium_id_base; + cfg->vmid = cfg->cbndx + 1; else - cfg->asid = cfg->cbndx + smmu->cavium_id_base; + cfg->asid = cfg->cbndx; + + smmu_domain->smmu = smmu; + if (smmu->impl && smmu->impl->init_context) { + ret = smmu->impl->init_context(smmu_domain); + if (ret) + goto out_unlock; + } pgtbl_cfg = (struct io_pgtable_cfg) { .pgsize_bitmap = smmu->pgsize_bitmap, @@ -765,7 +733,6 @@ static int arm_smmu_init_domain_context(struct iommu_domain *domain, if (smmu_domain->non_strict) pgtbl_cfg.quirks |= IO_PGTABLE_QUIRK_NON_STRICT; - smmu_domain->smmu = smmu; pgtbl_ops = alloc_io_pgtable_ops(fmt, &pgtbl_cfg, smmu_domain); if (!pgtbl_ops) { ret = -ENOMEM; diff --git a/drivers/iommu/arm-smmu.h b/drivers/iommu/arm-smmu.h index 616cc87a05e3..a18b5925b43c 100644 --- a/drivers/iommu/arm-smmu.h +++ b/drivers/iommu/arm-smmu.h @@ -14,6 +14,7 @@ #include <linux/bits.h> #include <linux/clk.h> #include <linux/device.h> +#include <linux/io-pgtable.h> #include <linux/iommu.h> #include <linux/mutex.h> #include <linux/spinlock.h> @@ -270,14 +271,50 @@ struct arm_smmu_device { struct clk_bulk_data *clks; int num_clks; - u32 cavium_id_base; /* Specific to Cavium */ - spinlock_t global_sync_lock; /* IOMMU core code handle */ struct iommu_device iommu; }; +enum arm_smmu_context_fmt { + ARM_SMMU_CTX_FMT_NONE, + ARM_SMMU_CTX_FMT_AARCH64, + ARM_SMMU_CTX_FMT_AARCH32_L, + ARM_SMMU_CTX_FMT_AARCH32_S, +}; + +struct arm_smmu_cfg { + u8 cbndx; + u8 irptndx; + union { + u16 asid; + u16 vmid; + }; + enum arm_smmu_cbar_type cbar; + enum arm_smmu_context_fmt fmt; +}; +#define INVALID_IRPTNDX 0xff + +enum arm_smmu_domain_stage { + ARM_SMMU_DOMAIN_S1 = 0, + ARM_SMMU_DOMAIN_S2, + ARM_SMMU_DOMAIN_NESTED, + ARM_SMMU_DOMAIN_BYPASS, +}; + +struct arm_smmu_domain { + struct arm_smmu_device *smmu; + struct io_pgtable_ops *pgtbl_ops; + const struct iommu_gather_ops *tlb_ops; + struct arm_smmu_cfg cfg; + enum arm_smmu_domain_stage stage; + bool non_strict; + struct mutex init_mutex; /* Protects smmu pointer */ + spinlock_t cb_lock; /* Serialises ATS1* ops and TLB syncs */ + struct iommu_domain domain; +}; + /* Implementation details, yay! */ struct arm_smmu_impl { @@ -289,6 +326,7 @@ struct arm_smmu_impl { u64 val); int (*cfg_probe)(struct arm_smmu_device *smmu); int (*reset)(struct arm_smmu_device *smmu); + int (*init_context)(struct arm_smmu_domain *smmu_domain); }; static inline void __iomem *arm_smmu_page(struct arm_smmu_device *smmu, int n)
Allocating and initialising a context for a domain is another point where certain implementations are known to want special behaviour. Currently the other half of the Cavium workaround comes into play here, so let's finish the job to get the whole thing right out of the way. Signed-off-by: Robin Murphy <robin.murphy@arm.com> --- drivers/iommu/arm-smmu-impl.c | 39 +++++++++++++++++++++++++-- drivers/iommu/arm-smmu.c | 51 +++++++---------------------------- drivers/iommu/arm-smmu.h | 42 +++++++++++++++++++++++++++-- 3 files changed, 86 insertions(+), 46 deletions(-)