@@ -19,7 +19,7 @@ struct arm_smmu_mmu_notifier {
bool cleared;
refcount_t refs;
struct list_head list;
- struct arm_smmu_domain *domain;
+ struct arm_smmu_domain domain;
};
#define mn_to_smmu(mn) container_of(mn, struct arm_smmu_mmu_notifier, mn)
@@ -198,7 +198,7 @@ static void arm_smmu_mm_invalidate_range(struct mmu_notifier *mn,
unsigned long start, unsigned long end)
{
struct arm_smmu_mmu_notifier *smmu_mn = mn_to_smmu(mn);
- struct arm_smmu_domain *smmu_domain = smmu_mn->domain;
+ struct arm_smmu_domain *smmu_domain = &smmu_mn->domain;
size_t size;
/*
@@ -217,7 +217,7 @@ static void arm_smmu_mm_invalidate_range(struct mmu_notifier *mn,
static void arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm)
{
struct arm_smmu_mmu_notifier *smmu_mn = mn_to_smmu(mn);
- struct arm_smmu_domain *smmu_domain = smmu_mn->domain;
+ struct arm_smmu_domain *smmu_domain = &smmu_mn->domain;
struct arm_smmu_master *master;
struct arm_smmu_attached_domain *attached_domain;
unsigned long flags;
@@ -233,15 +233,10 @@ static void arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm)
* but disable translation.
*/
spin_lock_irqsave(&smmu_domain->attached_domains_lock, flags);
- list_for_each_entry(attached_domain, &smmu_domain->attached_domains,
- domain_head) {
+ list_for_each_entry(attached_domain, &smmu_domain->attached_domains, domain_head) {
master = attached_domain->master;
- /*
- * SVA domains piggyback on the attached_domain with SSID 0.
- */
- if (attached_domain->ssid == 0)
- arm_smmu_write_ctx_desc(master->smmu, master->s1_cfg,
- master, mm->pasid, &quiet_cd);
+ arm_smmu_write_ctx_desc(master->smmu, master->s1_cfg, master,
+ attached_domain->ssid, &quiet_cd);
}
spin_unlock_irqrestore(&smmu_domain->attached_domains_lock, flags);
@@ -265,15 +260,13 @@ static const struct mmu_notifier_ops arm_smmu_mmu_notifier_ops = {
/* Allocate or get existing MMU notifier for this {domain, mm} pair */
static struct arm_smmu_mmu_notifier *
-arm_smmu_mmu_notifier_get(struct arm_smmu_domain *smmu_domain,
+arm_smmu_mmu_notifier_get(struct arm_smmu_device *smmu,
+ struct arm_smmu_domain *smmu_domain,
struct mm_struct *mm)
{
int ret;
- unsigned long flags;
struct arm_smmu_ctx_desc *cd;
struct arm_smmu_mmu_notifier *smmu_mn;
- struct arm_smmu_master *master;
- struct arm_smmu_attached_domain *attached_domain;
list_for_each_entry(smmu_mn, &smmu_domain->mmu_notifiers, list) {
if (smmu_mn->mn.mm == mm) {
@@ -294,7 +287,6 @@ arm_smmu_mmu_notifier_get(struct arm_smmu_domain *smmu_domain,
refcount_set(&smmu_mn->refs, 1);
smmu_mn->cd = cd;
- smmu_mn->domain = smmu_domain;
smmu_mn->mn.ops = &arm_smmu_mmu_notifier_ops;
ret = mmu_notifier_register(&smmu_mn->mn, mm);
@@ -302,24 +294,11 @@ arm_smmu_mmu_notifier_get(struct arm_smmu_domain *smmu_domain,
kfree(smmu_mn);
goto err_free_cd;
}
-
- spin_lock_irqsave(&smmu_domain->attached_domains_lock, flags);
- list_for_each_entry(attached_domain, &smmu_domain->attached_domains,
- domain_head) {
- master = attached_domain->master;
- ret = arm_smmu_write_ctx_desc(master->smmu, master->s1_cfg,
- master, mm->pasid, cd);
- }
- spin_unlock_irqrestore(&smmu_domain->attached_domains_lock, flags);
- if (ret)
- goto err_put_notifier;
+ arm_smmu_init_shared_cd_domain(smmu, &smmu_mn->domain, cd);
list_add(&smmu_mn->list, &smmu_domain->mmu_notifiers);
return smmu_mn;
-err_put_notifier:
- /* Frees smmu_mn */
- mmu_notifier_put(&smmu_mn->mn);
err_free_cd:
arm_smmu_free_shared_cd(cd);
return ERR_PTR(ret);
@@ -327,27 +306,15 @@ arm_smmu_mmu_notifier_get(struct arm_smmu_domain *smmu_domain,
static void arm_smmu_mmu_notifier_put(struct arm_smmu_mmu_notifier *smmu_mn)
{
- unsigned long flags;
struct mm_struct *mm = smmu_mn->mn.mm;
struct arm_smmu_ctx_desc *cd = smmu_mn->cd;
- struct arm_smmu_attached_domain *attached_domain;
- struct arm_smmu_master *master;
- struct arm_smmu_domain *smmu_domain = smmu_mn->domain;
+ struct arm_smmu_domain *smmu_domain = &smmu_mn->domain;
if (!refcount_dec_and_test(&smmu_mn->refs))
return;
list_del(&smmu_mn->list);
- spin_lock_irqsave(&smmu_domain->attached_domains_lock, flags);
- list_for_each_entry(attached_domain, &smmu_domain->attached_domains,
- domain_head) {
- master = attached_domain->master;
- arm_smmu_write_ctx_desc(master->smmu, master->s1_cfg, master,
- mm->pasid, NULL);
- }
- spin_unlock_irqrestore(&smmu_domain->attached_domains_lock, flags);
-
/*
* If we went through clear(), we've already invalidated, and no
* new TLB entry can have been formed.
@@ -369,17 +336,26 @@ static int __arm_smmu_sva_bind(struct device *dev,
struct arm_smmu_master *master = dev_iommu_priv_get(dev);
struct iommu_domain *domain = iommu_get_domain_for_dev(dev);
struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
+ int ret;
if (!master || !master->sva_enabled)
return -ENODEV;
- sva_domain->smmu_mn = arm_smmu_mmu_notifier_get(smmu_domain,
+ sva_domain->smmu_mn = arm_smmu_mmu_notifier_get(master->smmu,
+ smmu_domain,
mm);
if (IS_ERR(sva_domain->smmu_mn)) {
sva_domain->smmu_mn = NULL;
return PTR_ERR(sva_domain->smmu_mn);
}
+
master->nr_attached_sva_domains += 1;
+ smmu_domain = &sva_domain->smmu_mn->domain;
+ ret = arm_smmu_domain_set_dev_pasid(dev, master, smmu_domain, mm->pasid);
+ if (ret) {
+ arm_smmu_mmu_notifier_put(sva_domain->smmu_mn);
+ return ret;
+ }
return 0;
}
@@ -544,8 +520,9 @@ void arm_smmu_sva_remove_dev_pasid(struct iommu_domain *domain,
struct arm_smmu_master *master = dev_iommu_priv_get(dev);
mutex_lock(&sva_lock);
- master->nr_attached_sva_domains -= 1;
+ arm_smmu_domain_remove_dev_pasid(dev, &sva_domain->smmu_mn->domain, id);
arm_smmu_mmu_notifier_put(sva_domain->smmu_mn);
+ master->nr_attached_sva_domains -= 1;
mutex_unlock(&sva_lock);
}
Prepare an smmu domain of type S1_SHARED_CD per smmu_mmu_notifier. Attach that domain using the common arm_smmu_domain_set_dev_pasid implementation when attaching an SVA domain. Signed-off-by: Michael Shavit <mshavit@google.com> --- .../iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c | 67 ++++++------------- 1 file changed, 22 insertions(+), 45 deletions(-)