Message ID | 149786364660.14868.13384595731879361399.sendpatchset@little-apple (mailing list archive) |
---|---|
State | Superseded |
Delegated to: | Geert Uytterhoeven |
Headers | show |
On 19/06/17 10:14, Magnus Damm wrote: > From: Magnus Damm <damm+renesas@opensource.se> > > Add root device handling to the IPMMU driver by allowing certain > DT compat strings to enable has_cache_leaf_nodes that in turn will > support both root devices with interrupts and leaf devices that > face the actual IPMMU consumer devices. > > Signed-off-by: Magnus Damm <damm+renesas@opensource.se> > --- > > Changes since V3: > - Reworked root finding code to make it easier to follow, thanks Geert! > > Changes since V2: > - Fixed a bug in ipmmu_find_root() when only leaf devices are present > - Broke out __ipmmu_find_root() to allow ->xlate() check for root devices > > Changes since V1: > - Moved patch to earlier in the series > - Updated code to work with recent changes in: > [PATCH v3 00/06] iommu/ipmmu-vmsa: IPMMU multi-arch update V3 > > drivers/iommu/ipmmu-vmsa.c | 95 ++++++++++++++++++++++++++++++++++++-------- > 1 file changed, 78 insertions(+), 17 deletions(-) > > --- 0015/drivers/iommu/ipmmu-vmsa.c > +++ work/drivers/iommu/ipmmu-vmsa.c 2017-06-19 13:59:41.050607110 +0900 > @@ -36,6 +36,7 @@ > > struct ipmmu_features { > bool use_ns_alias_offset; > + bool has_cache_leaf_nodes; > }; > > struct ipmmu_vmsa_device { > @@ -44,6 +45,7 @@ struct ipmmu_vmsa_device { > struct iommu_device iommu; > struct list_head list; > const struct ipmmu_features *features; > + bool is_leaf; > unsigned int num_utlbs; > spinlock_t lock; /* Protects ctx and domains[] */ > DECLARE_BITMAP(ctx, IPMMU_CTX_MAX); > @@ -54,6 +56,7 @@ struct ipmmu_vmsa_device { > > struct ipmmu_vmsa_domain { > struct ipmmu_vmsa_device *mmu; > + struct ipmmu_vmsa_device *root; Would it not make more sense for this to be a property of the ipmmu_device itself, rather than per ipmmu_domain? I may of course have got the wrong idea of the topology here, but it seems as if mmu->is_leaf could be expressed as mmu->root == mmu vs. mmu->root == some_other_mmu, at which point there's one less thing to worry about in the domain. > struct iommu_domain io_domain; > > struct io_pgtable_cfg cfg; > @@ -203,6 +206,44 @@ static struct ipmmu_vmsa_iommu_priv *to_ > #define IMUASID_ASID0_SHIFT 0 > > /* ----------------------------------------------------------------------------- > + * Root device handling > + */ > + > +static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu) > +{ > + if (mmu->features->has_cache_leaf_nodes) > + return mmu->is_leaf ? false : true> + else > + return true; /* older IPMMU hardware treated as single root */ > +} > + > +static struct ipmmu_vmsa_device *__ipmmu_find_root(void) > +{ > + struct ipmmu_vmsa_device *mmu; > + struct ipmmu_vmsa_device *root = NULL; > + > + spin_lock(&ipmmu_devices_lock); > + > + list_for_each_entry(mmu, &ipmmu_devices, list) { > + if (ipmmu_is_root(mmu)) { > + root = mmu; > + break; > + } > + } > + > + spin_unlock(&ipmmu_devices_lock); > + return root; > +} I wonder if it might be tidier to use driver_for_each_device() for this, and remove the local list at the same time as its previous user. Either way, what happens if things end up hapening in this order: 1: probe leaf IPMMU B 2: probe device X behind IPMMU B 3: create and attach default domain for device X 4: probe root IPMMU A We know X will defer if B isn't ready, but it doesn't seem (at a glance, admittedly) that there's anything to enforce the expected probe ordering between A and B. This seems like another argument for moving the root/leaf association up to the device level, such that B can look up A once in its own probe routine, and defer itself if necessary. > + > +static struct ipmmu_vmsa_device *ipmmu_find_root(struct ipmmu_vmsa_device *leaf) > +{ > + if (ipmmu_is_root(leaf)) > + return leaf; > + else > + return __ipmmu_find_root(); Well actually, looking at that, I think what we have here is a very long-winded way of implementing this: static ipmmu_vmsa_device *root; Yuck. > +} > + > +/* ----------------------------------------------------------------------------- > * Read/Write Access > */ > > @@ -219,13 +260,13 @@ static void ipmmu_write(struct ipmmu_vms > > static u32 ipmmu_ctx_read(struct ipmmu_vmsa_domain *domain, unsigned int reg) > { > - return ipmmu_read(domain->mmu, domain->context_id * IM_CTX_SIZE + reg); > + return ipmmu_read(domain->root, domain->context_id * IM_CTX_SIZE + reg); > } > > static void ipmmu_ctx_write(struct ipmmu_vmsa_domain *domain, unsigned int reg, > u32 data) > { > - ipmmu_write(domain->mmu, domain->context_id * IM_CTX_SIZE + reg, data); > + ipmmu_write(domain->root, domain->context_id * IM_CTX_SIZE + reg, data); > } > > /* ----------------------------------------------------------------------------- > @@ -360,7 +401,7 @@ static int ipmmu_domain_init_context(str > * TODO: Add support for coherent walk through CCI with DVM and remove > * cache handling. For now, delegate it to the io-pgtable code. > */ > - domain->cfg.iommu_dev = domain->mmu->dev; > + domain->cfg.iommu_dev = domain->root->dev; > > domain->iop = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &domain->cfg, > domain); > @@ -370,7 +411,7 @@ static int ipmmu_domain_init_context(str > /* > * Find an unused context. > */ > - ret = ipmmu_domain_allocate_context(domain->mmu, domain); > + ret = ipmmu_domain_allocate_context(domain->root, domain); > if (ret == IPMMU_CTX_MAX) { > free_io_pgtable_ops(domain->iop); > return -EBUSY; > @@ -441,7 +482,7 @@ static void ipmmu_domain_destroy_context > */ > ipmmu_ctx_write(domain, IMCTR, IMCTR_FLUSH); > ipmmu_tlb_sync(domain); > - ipmmu_domain_free_context(domain->mmu, domain->context_id); > + ipmmu_domain_free_context(domain->root, domain->context_id); > } > > /* ----------------------------------------------------------------------------- > @@ -555,7 +596,7 @@ static int ipmmu_attach_device(struct io > { > struct ipmmu_vmsa_iommu_priv *priv = to_priv(dev); > struct iommu_fwspec *fwspec = dev->iommu_fwspec; > - struct ipmmu_vmsa_device *mmu = priv->mmu; > + struct ipmmu_vmsa_device *root, *mmu = priv->mmu; > struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); > unsigned long flags; > unsigned int i; > @@ -566,11 +607,18 @@ static int ipmmu_attach_device(struct io > return -ENXIO; > } > > + root = ipmmu_find_root(priv->mmu); > + if (!root) { > + dev_err(dev, "Unable to locate root IPMMU\n"); > + return -EAGAIN; > + } > + > spin_lock_irqsave(&domain->lock, flags); > > if (!domain->mmu) { > /* The domain hasn't been used yet, initialize it. */ > domain->mmu = mmu; > + domain->root = root; > ret = ipmmu_domain_init_context(domain); > } else if (domain->mmu != mmu) { > /* > @@ -911,6 +959,7 @@ static void ipmmu_device_reset(struct ip > > static const struct ipmmu_features ipmmu_features_default = { > .use_ns_alias_offset = true, > + .has_cache_leaf_nodes = false, > }; > > static const struct of_device_id ipmmu_of_ids[] = { > @@ -965,19 +1014,31 @@ static int ipmmu_probe(struct platform_d > mmu->base += IM_NS_ALIAS_OFFSET; > > irq = platform_get_irq(pdev, 0); > - if (irq < 0) { > - dev_err(&pdev->dev, "no IRQ found\n"); > - return irq; > - } > > - ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0, > - dev_name(&pdev->dev), mmu); > - if (ret < 0) { > - dev_err(&pdev->dev, "failed to request IRQ %d\n", irq); > - return ret; > - } > + /* > + * Determine if this IPMMU instance is a leaf device by checking > + * if the renesas,ipmmu-main property exists or not. > + */ > + if (mmu->features->has_cache_leaf_nodes && > + of_find_property(pdev->dev.of_node, "renesas,ipmmu-main", NULL)) > + mmu->is_leaf = true; Given my previous comments: if (mmu->features->has_cache_leaf_nodes) { root_np = of_parse_phandle(pdev->dev.of_node, "renesas,ipmmu-main", 0) if (root_np) { root_pdev = of_find_device_by_phandle(root_np); mmu->root = platform_get_drvdata(root_pdev); } else { mmu->root = mmu; } } or something along those lines? Robin. > + > + /* Root devices have mandatory IRQs */ > + if (ipmmu_is_root(mmu)) { > + if (irq < 0) { > + dev_err(&pdev->dev, "no IRQ found\n"); > + return irq; > + } > > - ipmmu_device_reset(mmu); > + ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0, > + dev_name(&pdev->dev), mmu); > + if (ret < 0) { > + dev_err(&pdev->dev, "failed to request IRQ %d\n", irq); > + return ret; > + } > + > + ipmmu_device_reset(mmu); > + } > > ret = iommu_device_register(&mmu->iommu); > if (ret) >
Hi Robin, On Tue, Jun 20, 2017 at 2:19 AM, Robin Murphy <robin.murphy@arm.com> wrote: > On 19/06/17 10:14, Magnus Damm wrote: >> From: Magnus Damm <damm+renesas@opensource.se> >> >> Add root device handling to the IPMMU driver by allowing certain >> DT compat strings to enable has_cache_leaf_nodes that in turn will >> support both root devices with interrupts and leaf devices that >> face the actual IPMMU consumer devices. >> >> Signed-off-by: Magnus Damm <damm+renesas@opensource.se> >> --- >> >> Changes since V3: >> - Reworked root finding code to make it easier to follow, thanks Geert! >> >> Changes since V2: >> - Fixed a bug in ipmmu_find_root() when only leaf devices are present >> - Broke out __ipmmu_find_root() to allow ->xlate() check for root devices >> >> Changes since V1: >> - Moved patch to earlier in the series >> - Updated code to work with recent changes in: >> [PATCH v3 00/06] iommu/ipmmu-vmsa: IPMMU multi-arch update V3 >> >> drivers/iommu/ipmmu-vmsa.c | 95 ++++++++++++++++++++++++++++++++++++-------- >> 1 file changed, 78 insertions(+), 17 deletions(-) >> >> --- 0015/drivers/iommu/ipmmu-vmsa.c >> +++ work/drivers/iommu/ipmmu-vmsa.c 2017-06-19 13:59:41.050607110 +0900 >> @@ -36,6 +36,7 @@ >> >> struct ipmmu_features { >> bool use_ns_alias_offset; >> + bool has_cache_leaf_nodes; >> }; >> >> struct ipmmu_vmsa_device { >> @@ -44,6 +45,7 @@ struct ipmmu_vmsa_device { >> struct iommu_device iommu; >> struct list_head list; >> const struct ipmmu_features *features; >> + bool is_leaf; >> unsigned int num_utlbs; >> spinlock_t lock; /* Protects ctx and domains[] */ >> DECLARE_BITMAP(ctx, IPMMU_CTX_MAX); >> @@ -54,6 +56,7 @@ struct ipmmu_vmsa_device { >> >> struct ipmmu_vmsa_domain { >> struct ipmmu_vmsa_device *mmu; >> + struct ipmmu_vmsa_device *root; > > Would it not make more sense for this to be a property of the > ipmmu_device itself, rather than per ipmmu_domain? I may of course have > got the wrong idea of the topology here, but it seems as if mmu->is_leaf > could be expressed as mmu->root == mmu vs. mmu->root == some_other_mmu, > at which point there's one less thing to worry about in the domain. Yes, you are right! Please have a look at patch 2 in V5 that is doing that. >> struct iommu_domain io_domain; >> >> struct io_pgtable_cfg cfg; >> @@ -203,6 +206,44 @@ static struct ipmmu_vmsa_iommu_priv *to_ >> #define IMUASID_ASID0_SHIFT 0 >> >> /* ----------------------------------------------------------------------------- >> + * Root device handling >> + */ >> + >> +static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu) >> +{ >> + if (mmu->features->has_cache_leaf_nodes) >> + return mmu->is_leaf ? false : true> + else >> + return true; /* older IPMMU hardware treated as single root */ >> +} >> + >> +static struct ipmmu_vmsa_device *__ipmmu_find_root(void) >> +{ >> + struct ipmmu_vmsa_device *mmu; >> + struct ipmmu_vmsa_device *root = NULL; >> + >> + spin_lock(&ipmmu_devices_lock); >> + >> + list_for_each_entry(mmu, &ipmmu_devices, list) { >> + if (ipmmu_is_root(mmu)) { >> + root = mmu; >> + break; >> + } >> + } >> + >> + spin_unlock(&ipmmu_devices_lock); >> + return root; >> +} > > I wonder if it might be tidier to use driver_for_each_device() for this, > and remove the local list at the same time as its previous user. Yep, also included in V5. > Either way, what happens if things end up hapening in this order: > > 1: probe leaf IPMMU B > 2: probe device X behind IPMMU B > 3: create and attach default domain for device X > 4: probe root IPMMU A > > We know X will defer if B isn't ready, but it doesn't seem (at a glance, > admittedly) that there's anything to enforce the expected probe ordering > between A and B. This seems like another argument for moving the > root/leaf association up to the device level, such that B can look up A > once in its own probe routine, and defer itself if necessary. There used to be code to check for presence of root device inside xlate() in "[PATCH v4 09/09] iommu/ipmmu-vmsa: Hook up r8a7795 DT matching code" however in V5 I'm deferring probe of non-root devices to handle this. >> + >> +static struct ipmmu_vmsa_device *ipmmu_find_root(struct ipmmu_vmsa_device *leaf) >> +{ >> + if (ipmmu_is_root(leaf)) >> + return leaf; >> + else >> + return __ipmmu_find_root(); > > Well actually, looking at that, I think what we have here is a very > long-winded way of implementing this: > > static ipmmu_vmsa_device *root; > > Yuck. The code has been reworked a bit in V5 but I did not go all the way to single global variable. >> +} >> + >> +/* ----------------------------------------------------------------------------- >> * Read/Write Access >> */ >> >> @@ -219,13 +260,13 @@ static void ipmmu_write(struct ipmmu_vms >> >> static u32 ipmmu_ctx_read(struct ipmmu_vmsa_domain *domain, unsigned int reg) >> { >> - return ipmmu_read(domain->mmu, domain->context_id * IM_CTX_SIZE + reg); >> + return ipmmu_read(domain->root, domain->context_id * IM_CTX_SIZE + reg); >> } >> >> static void ipmmu_ctx_write(struct ipmmu_vmsa_domain *domain, unsigned int reg, >> u32 data) >> { >> - ipmmu_write(domain->mmu, domain->context_id * IM_CTX_SIZE + reg, data); >> + ipmmu_write(domain->root, domain->context_id * IM_CTX_SIZE + reg, data); >> } >> >> /* ----------------------------------------------------------------------------- >> @@ -360,7 +401,7 @@ static int ipmmu_domain_init_context(str >> * TODO: Add support for coherent walk through CCI with DVM and remove >> * cache handling. For now, delegate it to the io-pgtable code. >> */ >> - domain->cfg.iommu_dev = domain->mmu->dev; >> + domain->cfg.iommu_dev = domain->root->dev; >> >> domain->iop = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &domain->cfg, >> domain); >> @@ -370,7 +411,7 @@ static int ipmmu_domain_init_context(str >> /* >> * Find an unused context. >> */ >> - ret = ipmmu_domain_allocate_context(domain->mmu, domain); >> + ret = ipmmu_domain_allocate_context(domain->root, domain); >> if (ret == IPMMU_CTX_MAX) { >> free_io_pgtable_ops(domain->iop); >> return -EBUSY; >> @@ -441,7 +482,7 @@ static void ipmmu_domain_destroy_context >> */ >> ipmmu_ctx_write(domain, IMCTR, IMCTR_FLUSH); >> ipmmu_tlb_sync(domain); >> - ipmmu_domain_free_context(domain->mmu, domain->context_id); >> + ipmmu_domain_free_context(domain->root, domain->context_id); >> } >> >> /* ----------------------------------------------------------------------------- >> @@ -555,7 +596,7 @@ static int ipmmu_attach_device(struct io >> { >> struct ipmmu_vmsa_iommu_priv *priv = to_priv(dev); >> struct iommu_fwspec *fwspec = dev->iommu_fwspec; >> - struct ipmmu_vmsa_device *mmu = priv->mmu; >> + struct ipmmu_vmsa_device *root, *mmu = priv->mmu; >> struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); >> unsigned long flags; >> unsigned int i; >> @@ -566,11 +607,18 @@ static int ipmmu_attach_device(struct io >> return -ENXIO; >> } >> >> + root = ipmmu_find_root(priv->mmu); >> + if (!root) { >> + dev_err(dev, "Unable to locate root IPMMU\n"); >> + return -EAGAIN; >> + } >> + >> spin_lock_irqsave(&domain->lock, flags); >> >> if (!domain->mmu) { >> /* The domain hasn't been used yet, initialize it. */ >> domain->mmu = mmu; >> + domain->root = root; >> ret = ipmmu_domain_init_context(domain); >> } else if (domain->mmu != mmu) { >> /* >> @@ -911,6 +959,7 @@ static void ipmmu_device_reset(struct ip >> >> static const struct ipmmu_features ipmmu_features_default = { >> .use_ns_alias_offset = true, >> + .has_cache_leaf_nodes = false, >> }; >> >> static const struct of_device_id ipmmu_of_ids[] = { >> @@ -965,19 +1014,31 @@ static int ipmmu_probe(struct platform_d >> mmu->base += IM_NS_ALIAS_OFFSET; >> >> irq = platform_get_irq(pdev, 0); >> - if (irq < 0) { >> - dev_err(&pdev->dev, "no IRQ found\n"); >> - return irq; >> - } >> >> - ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0, >> - dev_name(&pdev->dev), mmu); >> - if (ret < 0) { >> - dev_err(&pdev->dev, "failed to request IRQ %d\n", irq); >> - return ret; >> - } >> + /* >> + * Determine if this IPMMU instance is a leaf device by checking >> + * if the renesas,ipmmu-main property exists or not. >> + */ >> + if (mmu->features->has_cache_leaf_nodes && >> + of_find_property(pdev->dev.of_node, "renesas,ipmmu-main", NULL)) >> + mmu->is_leaf = true; > > Given my previous comments: > > if (mmu->features->has_cache_leaf_nodes) { > root_np = of_parse_phandle(pdev->dev.of_node, > "renesas,ipmmu-main", 0) > if (root_np) { > root_pdev = of_find_device_by_phandle(root_np); > mmu->root = platform_get_drvdata(root_pdev); > } else { > mmu->root = mmu; > } > } > > or something along those lines? When ditching the is_leaf variable and using deferred probing I reworked to code above. Getting the phandle like you are pointing out above looks promising. I also need to have a look at your series "[PATCH 0/4] ipmmu-vmsa cleanup", hopefully I can find some time next week in between talks at ELC. Hope to see you there! Cheers, / magnus
--- 0015/drivers/iommu/ipmmu-vmsa.c +++ work/drivers/iommu/ipmmu-vmsa.c 2017-06-19 13:59:41.050607110 +0900 @@ -36,6 +36,7 @@ struct ipmmu_features { bool use_ns_alias_offset; + bool has_cache_leaf_nodes; }; struct ipmmu_vmsa_device { @@ -44,6 +45,7 @@ struct ipmmu_vmsa_device { struct iommu_device iommu; struct list_head list; const struct ipmmu_features *features; + bool is_leaf; unsigned int num_utlbs; spinlock_t lock; /* Protects ctx and domains[] */ DECLARE_BITMAP(ctx, IPMMU_CTX_MAX); @@ -54,6 +56,7 @@ struct ipmmu_vmsa_device { struct ipmmu_vmsa_domain { struct ipmmu_vmsa_device *mmu; + struct ipmmu_vmsa_device *root; struct iommu_domain io_domain; struct io_pgtable_cfg cfg; @@ -203,6 +206,44 @@ static struct ipmmu_vmsa_iommu_priv *to_ #define IMUASID_ASID0_SHIFT 0 /* ----------------------------------------------------------------------------- + * Root device handling + */ + +static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu) +{ + if (mmu->features->has_cache_leaf_nodes) + return mmu->is_leaf ? false : true; + else + return true; /* older IPMMU hardware treated as single root */ +} + +static struct ipmmu_vmsa_device *__ipmmu_find_root(void) +{ + struct ipmmu_vmsa_device *mmu; + struct ipmmu_vmsa_device *root = NULL; + + spin_lock(&ipmmu_devices_lock); + + list_for_each_entry(mmu, &ipmmu_devices, list) { + if (ipmmu_is_root(mmu)) { + root = mmu; + break; + } + } + + spin_unlock(&ipmmu_devices_lock); + return root; +} + +static struct ipmmu_vmsa_device *ipmmu_find_root(struct ipmmu_vmsa_device *leaf) +{ + if (ipmmu_is_root(leaf)) + return leaf; + else + return __ipmmu_find_root(); +} + +/* ----------------------------------------------------------------------------- * Read/Write Access */ @@ -219,13 +260,13 @@ static void ipmmu_write(struct ipmmu_vms static u32 ipmmu_ctx_read(struct ipmmu_vmsa_domain *domain, unsigned int reg) { - return ipmmu_read(domain->mmu, domain->context_id * IM_CTX_SIZE + reg); + return ipmmu_read(domain->root, domain->context_id * IM_CTX_SIZE + reg); } static void ipmmu_ctx_write(struct ipmmu_vmsa_domain *domain, unsigned int reg, u32 data) { - ipmmu_write(domain->mmu, domain->context_id * IM_CTX_SIZE + reg, data); + ipmmu_write(domain->root, domain->context_id * IM_CTX_SIZE + reg, data); } /* ----------------------------------------------------------------------------- @@ -360,7 +401,7 @@ static int ipmmu_domain_init_context(str * TODO: Add support for coherent walk through CCI with DVM and remove * cache handling. For now, delegate it to the io-pgtable code. */ - domain->cfg.iommu_dev = domain->mmu->dev; + domain->cfg.iommu_dev = domain->root->dev; domain->iop = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &domain->cfg, domain); @@ -370,7 +411,7 @@ static int ipmmu_domain_init_context(str /* * Find an unused context. */ - ret = ipmmu_domain_allocate_context(domain->mmu, domain); + ret = ipmmu_domain_allocate_context(domain->root, domain); if (ret == IPMMU_CTX_MAX) { free_io_pgtable_ops(domain->iop); return -EBUSY; @@ -441,7 +482,7 @@ static void ipmmu_domain_destroy_context */ ipmmu_ctx_write(domain, IMCTR, IMCTR_FLUSH); ipmmu_tlb_sync(domain); - ipmmu_domain_free_context(domain->mmu, domain->context_id); + ipmmu_domain_free_context(domain->root, domain->context_id); } /* ----------------------------------------------------------------------------- @@ -555,7 +596,7 @@ static int ipmmu_attach_device(struct io { struct ipmmu_vmsa_iommu_priv *priv = to_priv(dev); struct iommu_fwspec *fwspec = dev->iommu_fwspec; - struct ipmmu_vmsa_device *mmu = priv->mmu; + struct ipmmu_vmsa_device *root, *mmu = priv->mmu; struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain); unsigned long flags; unsigned int i; @@ -566,11 +607,18 @@ static int ipmmu_attach_device(struct io return -ENXIO; } + root = ipmmu_find_root(priv->mmu); + if (!root) { + dev_err(dev, "Unable to locate root IPMMU\n"); + return -EAGAIN; + } + spin_lock_irqsave(&domain->lock, flags); if (!domain->mmu) { /* The domain hasn't been used yet, initialize it. */ domain->mmu = mmu; + domain->root = root; ret = ipmmu_domain_init_context(domain); } else if (domain->mmu != mmu) { /* @@ -911,6 +959,7 @@ static void ipmmu_device_reset(struct ip static const struct ipmmu_features ipmmu_features_default = { .use_ns_alias_offset = true, + .has_cache_leaf_nodes = false, }; static const struct of_device_id ipmmu_of_ids[] = { @@ -965,19 +1014,31 @@ static int ipmmu_probe(struct platform_d mmu->base += IM_NS_ALIAS_OFFSET; irq = platform_get_irq(pdev, 0); - if (irq < 0) { - dev_err(&pdev->dev, "no IRQ found\n"); - return irq; - } - ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0, - dev_name(&pdev->dev), mmu); - if (ret < 0) { - dev_err(&pdev->dev, "failed to request IRQ %d\n", irq); - return ret; - } + /* + * Determine if this IPMMU instance is a leaf device by checking + * if the renesas,ipmmu-main property exists or not. + */ + if (mmu->features->has_cache_leaf_nodes && + of_find_property(pdev->dev.of_node, "renesas,ipmmu-main", NULL)) + mmu->is_leaf = true; + + /* Root devices have mandatory IRQs */ + if (ipmmu_is_root(mmu)) { + if (irq < 0) { + dev_err(&pdev->dev, "no IRQ found\n"); + return irq; + } - ipmmu_device_reset(mmu); + ret = devm_request_irq(&pdev->dev, irq, ipmmu_irq, 0, + dev_name(&pdev->dev), mmu); + if (ret < 0) { + dev_err(&pdev->dev, "failed to request IRQ %d\n", irq); + return ret; + } + + ipmmu_device_reset(mmu); + } ret = iommu_device_register(&mmu->iommu); if (ret)