Message ID | 20-v3-e797f4dc6918+93057-iommu_pages_jgg@nvidia.com (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | iommu: Further abstract iommu-pages | expand |
On 2025/2/26 3:39, Jason Gunthorpe wrote: > Convert most of the places calling get_order() as an argument to the > iommu-pages allocator into order_base_2() or the _sz flavour > instead. These places already have an exact size, there is no particular > reason to use order here. > > Signed-off-by: Jason Gunthorpe<jgg@nvidia.com> > --- > drivers/iommu/amd/init.c | 29 +++++++++++++++-------------- > drivers/iommu/intel/dmar.c | 6 +++--- > drivers/iommu/io-pgtable-arm.c | 3 +-- > drivers/iommu/io-pgtable-dart.c | 12 +++--------- > drivers/iommu/sun50i-iommu.c | 4 ++-- > 5 files changed, 24 insertions(+), 30 deletions(-) For changes in intel iommu driver, Reviewed-by: Lu Baolu <baolu.lu@linux.intel.com>
On Tue, Feb 25, 2025 at 03:39:37PM -0400, Jason Gunthorpe wrote: > Convert most of the places calling get_order() as an argument to the > iommu-pages allocator into order_base_2() or the _sz flavour > instead. These places already have an exact size, there is no particular > reason to use order here. > > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > --- > drivers/iommu/amd/init.c | 29 +++++++++++++++-------------- > drivers/iommu/intel/dmar.c | 6 +++--- > drivers/iommu/io-pgtable-arm.c | 3 +-- > drivers/iommu/io-pgtable-dart.c | 12 +++--------- > drivers/iommu/sun50i-iommu.c | 4 ++-- > 5 files changed, 24 insertions(+), 30 deletions(-) > > diff --git a/drivers/iommu/amd/init.c b/drivers/iommu/amd/init.c > index e3f4283ebbc201..a5720df7b22397 100644 > --- a/drivers/iommu/amd/init.c > +++ b/drivers/iommu/amd/init.c > @@ -635,8 +635,8 @@ static int __init find_last_devid_acpi(struct acpi_table_header *table, u16 pci_ > /* Allocate per PCI segment device table */ > static inline int __init alloc_dev_table(struct amd_iommu_pci_seg *pci_seg) > { > - pci_seg->dev_table = iommu_alloc_pages(GFP_KERNEL | GFP_DMA32, > - get_order(pci_seg->dev_table_size)); > + pci_seg->dev_table = iommu_alloc_pages_sz(GFP_KERNEL | GFP_DMA32, > + pci_seg->dev_table_size); > if (!pci_seg->dev_table) > return -ENOMEM; > > @@ -716,8 +716,7 @@ static void __init free_alias_table(struct amd_iommu_pci_seg *pci_seg) > */ > static int __init alloc_command_buffer(struct amd_iommu *iommu) > { > - iommu->cmd_buf = iommu_alloc_pages(GFP_KERNEL, > - get_order(CMD_BUFFER_SIZE)); > + iommu->cmd_buf = iommu_alloc_pages_sz(GFP_KERNEL, CMD_BUFFER_SIZE); > > return iommu->cmd_buf ? 0 : -ENOMEM; > } > @@ -820,14 +819,16 @@ static void __init free_command_buffer(struct amd_iommu *iommu) > void *__init iommu_alloc_4k_pages(struct amd_iommu *iommu, gfp_t gfp, > size_t size) > { > - int order = get_order(size); > - void *buf = iommu_alloc_pages(gfp, order); > + void *buf; > > - if (buf && > - check_feature(FEATURE_SNP) && > - set_memory_4k((unsigned long)buf, (1 << order))) { > + size = PAGE_ALIGN(size); > + buf = iommu_alloc_pages_sz(gfp, size); > + if (!buf) > + return NULL; > + if (check_feature(FEATURE_SNP) && > + set_memory_4k((unsigned long)buf, size / PAGE_SIZE)) { > iommu_free_pages(buf); > - buf = NULL; > + return NULL; > } > > return buf; > @@ -922,11 +923,11 @@ static int iommu_init_ga_log(struct amd_iommu *iommu) > if (!AMD_IOMMU_GUEST_IR_VAPIC(amd_iommu_guest_ir)) > return 0; > > - iommu->ga_log = iommu_alloc_pages(GFP_KERNEL, get_order(GA_LOG_SIZE)); > + iommu->ga_log = iommu_alloc_pages_sz(GFP_KERNEL, GA_LOG_SIZE); > if (!iommu->ga_log) > goto err_out; > > - iommu->ga_log_tail = iommu_alloc_pages(GFP_KERNEL, get_order(8)); > + iommu->ga_log_tail = iommu_alloc_pages_sz(GFP_KERNEL, 8); > if (!iommu->ga_log_tail) > goto err_out; > > @@ -1021,8 +1022,8 @@ static bool __copy_device_table(struct amd_iommu *iommu) > if (!old_devtb) > return false; > > - pci_seg->old_dev_tbl_cpy = iommu_alloc_pages(GFP_KERNEL | GFP_DMA32, > - get_order(pci_seg->dev_table_size)); > + pci_seg->old_dev_tbl_cpy = iommu_alloc_pages_sz( > + GFP_KERNEL | GFP_DMA32, pci_seg->dev_table_size); > if (pci_seg->old_dev_tbl_cpy == NULL) { > pr_err("Failed to allocate memory for copying old device table!\n"); > memunmap(old_devtb); > diff --git a/drivers/iommu/intel/dmar.c b/drivers/iommu/intel/dmar.c > index c812c83d77da10..4c7ce92acf6976 100644 > --- a/drivers/iommu/intel/dmar.c > +++ b/drivers/iommu/intel/dmar.c > @@ -1681,7 +1681,6 @@ int dmar_enable_qi(struct intel_iommu *iommu) > { > struct q_inval *qi; > void *desc; > - int order; > > if (!ecap_qis(iommu->ecap)) > return -ENOENT; > @@ -1702,8 +1701,9 @@ int dmar_enable_qi(struct intel_iommu *iommu) > * Need two pages to accommodate 256 descriptors of 256 bits each > * if the remapping hardware supports scalable mode translation. > */ > - order = ecap_smts(iommu->ecap) ? 1 : 0; > - desc = iommu_alloc_pages_node(iommu->node, GFP_ATOMIC, order); > + desc = iommu_alloc_pages_node_sz(iommu->node, GFP_ATOMIC, > + ecap_smts(iommu->ecap) ? SZ_8K : > + SZ_4K); > if (!desc) { > kfree(qi); > iommu->qi = NULL; > diff --git a/drivers/iommu/io-pgtable-arm.c b/drivers/iommu/io-pgtable-arm.c > index 08d0f62abe8a09..d13149ec5be77e 100644 > --- a/drivers/iommu/io-pgtable-arm.c > +++ b/drivers/iommu/io-pgtable-arm.c > @@ -263,14 +263,13 @@ static void *__arm_lpae_alloc_pages(size_t size, gfp_t gfp, > void *cookie) > { > struct device *dev = cfg->iommu_dev; > - int order = get_order(size); > dma_addr_t dma; > void *pages; > > if (cfg->alloc) > pages = cfg->alloc(cookie, size, gfp); > else > - pages = iommu_alloc_pages_node(dev_to_node(dev), gfp, order); > + pages = iommu_alloc_pages_node_sz(dev_to_node(dev), gfp, size); Although, the current implementation of iommu_alloc_pages_node_sz() would round the size to order, but this is not correct according to the API definition "The returned allocation is round_up_pow_two(size) big, and is physically aligned to its size." SMMUv3 has special alignment with small number of entries at the start level, according the manual: A 64-byte minimum alignment on starting-level translation table addresses is imposed when TG0 selects 64KB granules and the effective IPS value indicates 52-bit output. In this case bits [5:0] are treated as zero. And according to the Arm Arm (ex D24.2.195 in Version L) - Bits A[(x-1):0] of the stage 1 translation table base address are zero. ... The smallest permitted value of x is 5. Which 32 bytes For a case as (which is valid in Linux) - S1 with IAS 40-bits and 4K, start level has 2 entries (16 bytes) but alignment must be at least 32 bytes. - Similarly with 16K and 48 bits. I'd say we can align the size or use min with 64 bytes before calling the function would be enough (or change the API to state that allocations are rounded to order) Thanks, Mostafa > > if (!pages) > return NULL; > diff --git a/drivers/iommu/io-pgtable-dart.c b/drivers/iommu/io-pgtable-dart.c > index ebf330e67bfa30..a0988669bb951a 100644 > --- a/drivers/iommu/io-pgtable-dart.c > +++ b/drivers/iommu/io-pgtable-dart.c > @@ -107,13 +107,6 @@ static phys_addr_t iopte_to_paddr(dart_iopte pte, > return paddr; > } > > -static void *__dart_alloc_pages(size_t size, gfp_t gfp) > -{ > - int order = get_order(size); > - > - return iommu_alloc_pages(gfp, order); > -} > - > static int dart_init_pte(struct dart_io_pgtable *data, > unsigned long iova, phys_addr_t paddr, > dart_iopte prot, int num_entries, > @@ -255,7 +248,7 @@ static int dart_map_pages(struct io_pgtable_ops *ops, unsigned long iova, > > /* no L2 table present */ > if (!pte) { > - cptep = __dart_alloc_pages(tblsz, gfp); > + cptep = iommu_alloc_pages_sz(gfp, tblsz); > if (!cptep) > return -ENOMEM; > > @@ -412,7 +405,8 @@ apple_dart_alloc_pgtable(struct io_pgtable_cfg *cfg, void *cookie) > cfg->apple_dart_cfg.n_ttbrs = 1 << data->tbl_bits; > > for (i = 0; i < cfg->apple_dart_cfg.n_ttbrs; ++i) { > - data->pgd[i] = __dart_alloc_pages(DART_GRANULE(data), GFP_KERNEL); > + data->pgd[i] = > + iommu_alloc_pages_sz(GFP_KERNEL, DART_GRANULE(data)); > if (!data->pgd[i]) > goto out_free_data; > cfg->apple_dart_cfg.ttbr[i] = virt_to_phys(data->pgd[i]); > diff --git a/drivers/iommu/sun50i-iommu.c b/drivers/iommu/sun50i-iommu.c > index 6385560dbc3fb0..76c9620af4bba8 100644 > --- a/drivers/iommu/sun50i-iommu.c > +++ b/drivers/iommu/sun50i-iommu.c > @@ -690,8 +690,8 @@ sun50i_iommu_domain_alloc_paging(struct device *dev) > if (!sun50i_domain) > return NULL; > > - sun50i_domain->dt = iommu_alloc_pages(GFP_KERNEL | GFP_DMA32, > - get_order(DT_SIZE)); > + sun50i_domain->dt = > + iommu_alloc_pages_sz(GFP_KERNEL | GFP_DMA32, DT_SIZE); > if (!sun50i_domain->dt) > goto err_free_domain; > > -- > 2.43.0 >
diff --git a/drivers/iommu/amd/init.c b/drivers/iommu/amd/init.c index e3f4283ebbc201..a5720df7b22397 100644 --- a/drivers/iommu/amd/init.c +++ b/drivers/iommu/amd/init.c @@ -635,8 +635,8 @@ static int __init find_last_devid_acpi(struct acpi_table_header *table, u16 pci_ /* Allocate per PCI segment device table */ static inline int __init alloc_dev_table(struct amd_iommu_pci_seg *pci_seg) { - pci_seg->dev_table = iommu_alloc_pages(GFP_KERNEL | GFP_DMA32, - get_order(pci_seg->dev_table_size)); + pci_seg->dev_table = iommu_alloc_pages_sz(GFP_KERNEL | GFP_DMA32, + pci_seg->dev_table_size); if (!pci_seg->dev_table) return -ENOMEM; @@ -716,8 +716,7 @@ static void __init free_alias_table(struct amd_iommu_pci_seg *pci_seg) */ static int __init alloc_command_buffer(struct amd_iommu *iommu) { - iommu->cmd_buf = iommu_alloc_pages(GFP_KERNEL, - get_order(CMD_BUFFER_SIZE)); + iommu->cmd_buf = iommu_alloc_pages_sz(GFP_KERNEL, CMD_BUFFER_SIZE); return iommu->cmd_buf ? 0 : -ENOMEM; } @@ -820,14 +819,16 @@ static void __init free_command_buffer(struct amd_iommu *iommu) void *__init iommu_alloc_4k_pages(struct amd_iommu *iommu, gfp_t gfp, size_t size) { - int order = get_order(size); - void *buf = iommu_alloc_pages(gfp, order); + void *buf; - if (buf && - check_feature(FEATURE_SNP) && - set_memory_4k((unsigned long)buf, (1 << order))) { + size = PAGE_ALIGN(size); + buf = iommu_alloc_pages_sz(gfp, size); + if (!buf) + return NULL; + if (check_feature(FEATURE_SNP) && + set_memory_4k((unsigned long)buf, size / PAGE_SIZE)) { iommu_free_pages(buf); - buf = NULL; + return NULL; } return buf; @@ -922,11 +923,11 @@ static int iommu_init_ga_log(struct amd_iommu *iommu) if (!AMD_IOMMU_GUEST_IR_VAPIC(amd_iommu_guest_ir)) return 0; - iommu->ga_log = iommu_alloc_pages(GFP_KERNEL, get_order(GA_LOG_SIZE)); + iommu->ga_log = iommu_alloc_pages_sz(GFP_KERNEL, GA_LOG_SIZE); if (!iommu->ga_log) goto err_out; - iommu->ga_log_tail = iommu_alloc_pages(GFP_KERNEL, get_order(8)); + iommu->ga_log_tail = iommu_alloc_pages_sz(GFP_KERNEL, 8); if (!iommu->ga_log_tail) goto err_out; @@ -1021,8 +1022,8 @@ static bool __copy_device_table(struct amd_iommu *iommu) if (!old_devtb) return false; - pci_seg->old_dev_tbl_cpy = iommu_alloc_pages(GFP_KERNEL | GFP_DMA32, - get_order(pci_seg->dev_table_size)); + pci_seg->old_dev_tbl_cpy = iommu_alloc_pages_sz( + GFP_KERNEL | GFP_DMA32, pci_seg->dev_table_size); if (pci_seg->old_dev_tbl_cpy == NULL) { pr_err("Failed to allocate memory for copying old device table!\n"); memunmap(old_devtb); diff --git a/drivers/iommu/intel/dmar.c b/drivers/iommu/intel/dmar.c index c812c83d77da10..4c7ce92acf6976 100644 --- a/drivers/iommu/intel/dmar.c +++ b/drivers/iommu/intel/dmar.c @@ -1681,7 +1681,6 @@ int dmar_enable_qi(struct intel_iommu *iommu) { struct q_inval *qi; void *desc; - int order; if (!ecap_qis(iommu->ecap)) return -ENOENT; @@ -1702,8 +1701,9 @@ int dmar_enable_qi(struct intel_iommu *iommu) * Need two pages to accommodate 256 descriptors of 256 bits each * if the remapping hardware supports scalable mode translation. */ - order = ecap_smts(iommu->ecap) ? 1 : 0; - desc = iommu_alloc_pages_node(iommu->node, GFP_ATOMIC, order); + desc = iommu_alloc_pages_node_sz(iommu->node, GFP_ATOMIC, + ecap_smts(iommu->ecap) ? SZ_8K : + SZ_4K); if (!desc) { kfree(qi); iommu->qi = NULL; diff --git a/drivers/iommu/io-pgtable-arm.c b/drivers/iommu/io-pgtable-arm.c index 08d0f62abe8a09..d13149ec5be77e 100644 --- a/drivers/iommu/io-pgtable-arm.c +++ b/drivers/iommu/io-pgtable-arm.c @@ -263,14 +263,13 @@ static void *__arm_lpae_alloc_pages(size_t size, gfp_t gfp, void *cookie) { struct device *dev = cfg->iommu_dev; - int order = get_order(size); dma_addr_t dma; void *pages; if (cfg->alloc) pages = cfg->alloc(cookie, size, gfp); else - pages = iommu_alloc_pages_node(dev_to_node(dev), gfp, order); + pages = iommu_alloc_pages_node_sz(dev_to_node(dev), gfp, size); if (!pages) return NULL; diff --git a/drivers/iommu/io-pgtable-dart.c b/drivers/iommu/io-pgtable-dart.c index ebf330e67bfa30..a0988669bb951a 100644 --- a/drivers/iommu/io-pgtable-dart.c +++ b/drivers/iommu/io-pgtable-dart.c @@ -107,13 +107,6 @@ static phys_addr_t iopte_to_paddr(dart_iopte pte, return paddr; } -static void *__dart_alloc_pages(size_t size, gfp_t gfp) -{ - int order = get_order(size); - - return iommu_alloc_pages(gfp, order); -} - static int dart_init_pte(struct dart_io_pgtable *data, unsigned long iova, phys_addr_t paddr, dart_iopte prot, int num_entries, @@ -255,7 +248,7 @@ static int dart_map_pages(struct io_pgtable_ops *ops, unsigned long iova, /* no L2 table present */ if (!pte) { - cptep = __dart_alloc_pages(tblsz, gfp); + cptep = iommu_alloc_pages_sz(gfp, tblsz); if (!cptep) return -ENOMEM; @@ -412,7 +405,8 @@ apple_dart_alloc_pgtable(struct io_pgtable_cfg *cfg, void *cookie) cfg->apple_dart_cfg.n_ttbrs = 1 << data->tbl_bits; for (i = 0; i < cfg->apple_dart_cfg.n_ttbrs; ++i) { - data->pgd[i] = __dart_alloc_pages(DART_GRANULE(data), GFP_KERNEL); + data->pgd[i] = + iommu_alloc_pages_sz(GFP_KERNEL, DART_GRANULE(data)); if (!data->pgd[i]) goto out_free_data; cfg->apple_dart_cfg.ttbr[i] = virt_to_phys(data->pgd[i]); diff --git a/drivers/iommu/sun50i-iommu.c b/drivers/iommu/sun50i-iommu.c index 6385560dbc3fb0..76c9620af4bba8 100644 --- a/drivers/iommu/sun50i-iommu.c +++ b/drivers/iommu/sun50i-iommu.c @@ -690,8 +690,8 @@ sun50i_iommu_domain_alloc_paging(struct device *dev) if (!sun50i_domain) return NULL; - sun50i_domain->dt = iommu_alloc_pages(GFP_KERNEL | GFP_DMA32, - get_order(DT_SIZE)); + sun50i_domain->dt = + iommu_alloc_pages_sz(GFP_KERNEL | GFP_DMA32, DT_SIZE); if (!sun50i_domain->dt) goto err_free_domain;
Convert most of the places calling get_order() as an argument to the iommu-pages allocator into order_base_2() or the _sz flavour instead. These places already have an exact size, there is no particular reason to use order here. Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> --- drivers/iommu/amd/init.c | 29 +++++++++++++++-------------- drivers/iommu/intel/dmar.c | 6 +++--- drivers/iommu/io-pgtable-arm.c | 3 +-- drivers/iommu/io-pgtable-dart.c | 12 +++--------- drivers/iommu/sun50i-iommu.c | 4 ++-- 5 files changed, 24 insertions(+), 30 deletions(-)