Message ID | 20190707210024.26192-2-chris@chris-wilson.co.uk (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [01/11] drm/i915/gtt: Use shallow dma pages for scratch | expand |
Chris Wilson <chris@chris-wilson.co.uk> writes: > We only use the dma pages for scratch, and so do not need to allocate > the extra storage for the shadow page directory. > > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> > Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com> > --- > drivers/gpu/drm/i915/i915_gem_gtt.c | 192 ++++++++++++---------------- > drivers/gpu/drm/i915/i915_gem_gtt.h | 6 +- > 2 files changed, 85 insertions(+), 113 deletions(-) > > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c > index 236c964dd761..937236913e70 100644 > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c > @@ -594,25 +594,17 @@ static void cleanup_page_dma(struct i915_address_space *vm, > > #define kmap_atomic_px(px) kmap_atomic(px_base(px)->page) > > -#define fill_px(vm, px, v) fill_page_dma((vm), px_base(px), (v)) > -#define fill32_px(vm, px, v) fill_page_dma_32((vm), px_base(px), (v)) > +#define fill_px(px, v) fill_page_dma(px_base(px), (v)) > +#define fill32_px(px, v) fill_page_dma_32(px_base(px), (v)) > > -static void fill_page_dma(struct i915_address_space *vm, > - struct i915_page_dma *p, > - const u64 val) > +static void fill_page_dma(struct i915_page_dma *p, const u64 val) > { > - u64 * const vaddr = kmap_atomic(p->page); > - > - memset64(vaddr, val, PAGE_SIZE / sizeof(val)); > - > - kunmap_atomic(vaddr); > + kunmap_atomic(memset64(kmap_atomic(p->page), val, I915_PDES)); Neat. I would go for 512 instead of I915_PDES. There is no magic and there never will be magic as it is as const as carved into stone. > } > > -static void fill_page_dma_32(struct i915_address_space *vm, > - struct i915_page_dma *p, > - const u32 v) > +static void fill_page_dma_32(struct i915_page_dma *p, const u32 v) > { > - fill_page_dma(vm, p, (u64)v << 32 | v); > + fill_page_dma(p, (u64)v << 32 | v); > } > > static int > @@ -687,6 +679,21 @@ static void cleanup_scratch_page(struct i915_address_space *vm) > __free_pages(p->page, order); > } > > +static void free_scratch(struct i915_address_space *vm) > +{ > + if (!vm->scratch_page.daddr) /* set to 0 on clones */ > + return; > + > + if (vm->scratch_pdp.daddr) > + cleanup_page_dma(vm, &vm->scratch_pdp); > + if (vm->scratch_pd.daddr) > + cleanup_page_dma(vm, &vm->scratch_pd); > + if (vm->scratch_pt.daddr) > + cleanup_page_dma(vm, &vm->scratch_pt); > + > + cleanup_scratch_page(vm); > +} > + > static struct i915_page_table *alloc_pt(struct i915_address_space *vm) > { > struct i915_page_table *pt; > @@ -711,18 +718,6 @@ static void free_pt(struct i915_address_space *vm, struct i915_page_table *pt) > kfree(pt); > } > > -static void gen8_initialize_pt(struct i915_address_space *vm, > - struct i915_page_table *pt) > -{ > - fill_px(vm, pt, vm->scratch_pte); > -} > - > -static void gen6_initialize_pt(struct i915_address_space *vm, > - struct i915_page_table *pt) > -{ > - fill32_px(vm, pt, vm->scratch_pte); > -} > - > static struct i915_page_directory *__alloc_pd(void) > { > struct i915_page_directory *pd; > @@ -765,9 +760,11 @@ static void free_pd(struct i915_address_space *vm, > kfree(pd); > } > > -#define init_pd(vm, pd, to) { \ > - fill_px((vm), (pd), gen8_pde_encode(px_dma(to), I915_CACHE_LLC)); \ > - memset_p((pd)->entry, (to), 512); \ > +static void init_pd(struct i915_page_directory *pd, > + struct i915_page_dma *scratch) > +{ > + fill_px(pd, gen8_pde_encode(scratch->daddr, I915_CACHE_LLC)); > + memset_p(pd->entry, scratch, 512); > } > > static inline void > @@ -869,12 +866,11 @@ static void gen8_ppgtt_clear_pd(struct i915_address_space *vm, > u32 pde; > > gen8_for_each_pde(pt, pd, start, length, pde) { > - GEM_BUG_ON(pt == vm->scratch_pt); > + GEM_BUG_ON(px_base(pt) == &vm->scratch_pt); > > atomic_inc(&pt->used); > gen8_ppgtt_clear_pt(vm, pt, start, length); > - if (release_pd_entry(pd, pde, &pt->used, > - px_base(vm->scratch_pt))) > + if (release_pd_entry(pd, pde, &pt->used, &vm->scratch_pt)) > free_pt(vm, pt); > } > } > @@ -890,12 +886,11 @@ static void gen8_ppgtt_clear_pdp(struct i915_address_space *vm, > unsigned int pdpe; > > gen8_for_each_pdpe(pd, pdp, start, length, pdpe) { > - GEM_BUG_ON(pd == vm->scratch_pd); > + GEM_BUG_ON(px_base(pd) == &vm->scratch_pd); Perhaps future will bring pd_points_scratch(pd). Now the intriguing, bordering irritating, question in my mind is that can we fold the scratch_pd and scratch_pdp to be the same thing. Patch lgtm with some dislike towards I915_PDES, Reviewed-by: Mika Kuoppala <mika.kuoppala@linux.intel.com> > > atomic_inc(&pd->used); > gen8_ppgtt_clear_pd(vm, pd, start, length); > - if (release_pd_entry(pdp, pdpe, &pd->used, > - px_base(vm->scratch_pd))) > + if (release_pd_entry(pdp, pdpe, &pd->used, &vm->scratch_pd)) > free_pd(vm, pd); > } > } > @@ -921,12 +916,11 @@ static void gen8_ppgtt_clear_4lvl(struct i915_address_space *vm, > GEM_BUG_ON(!i915_vm_is_4lvl(vm)); > > gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) { > - GEM_BUG_ON(pdp == vm->scratch_pdp); > + GEM_BUG_ON(px_base(pdp) == &vm->scratch_pdp); > > atomic_inc(&pdp->used); > gen8_ppgtt_clear_pdp(vm, pdp, start, length); > - if (release_pd_entry(pml4, pml4e, &pdp->used, > - px_base(vm->scratch_pdp))) > + if (release_pd_entry(pml4, pml4e, &pdp->used, &vm->scratch_pdp)) > free_pd(vm, pdp); > } > } > @@ -1181,7 +1175,7 @@ static void gen8_free_page_tables(struct i915_address_space *vm, > int i; > > for (i = 0; i < I915_PDES; i++) { > - if (pd->entry[i] != vm->scratch_pt) > + if (pd->entry[i] != &vm->scratch_pt) > free_pt(vm, pd->entry[i]); > } > } > @@ -1218,37 +1212,35 @@ static int gen8_init_scratch(struct i915_address_space *vm) > I915_CACHE_LLC, > vm->has_read_only); > > - vm->scratch_pt = alloc_pt(vm); > - if (IS_ERR(vm->scratch_pt)) { > - ret = PTR_ERR(vm->scratch_pt); > + if (unlikely(setup_page_dma(vm, &vm->scratch_pt))) { > + ret = -ENOMEM; > goto free_scratch_page; > } > + fill_page_dma(&vm->scratch_pt, vm->scratch_pte); > > - vm->scratch_pd = alloc_pd(vm); > - if (IS_ERR(vm->scratch_pd)) { > - ret = PTR_ERR(vm->scratch_pd); > + if (unlikely(setup_page_dma(vm, &vm->scratch_pd))) { > + ret = -ENOMEM; > goto free_pt; > } > + fill_page_dma(&vm->scratch_pd, > + gen8_pde_encode(vm->scratch_pd.daddr, I915_CACHE_LLC)); > > if (i915_vm_is_4lvl(vm)) { > - vm->scratch_pdp = alloc_pd(vm); > - if (IS_ERR(vm->scratch_pdp)) { > - ret = PTR_ERR(vm->scratch_pdp); > + if (unlikely(setup_page_dma(vm, &vm->scratch_pdp))) { > + ret = -ENOMEM; > goto free_pd; > } > + fill_page_dma(&vm->scratch_pdp, > + gen8_pde_encode(vm->scratch_pdp.daddr, > + I915_CACHE_LLC)); > } > > - gen8_initialize_pt(vm, vm->scratch_pt); > - init_pd(vm, vm->scratch_pd, vm->scratch_pt); > - if (i915_vm_is_4lvl(vm)) > - init_pd(vm, vm->scratch_pdp, vm->scratch_pd); > - > return 0; > > free_pd: > - free_pd(vm, vm->scratch_pd); > + cleanup_page_dma(vm, &vm->scratch_pd); > free_pt: > - free_pt(vm, vm->scratch_pt); > + cleanup_page_dma(vm, &vm->scratch_pt); > free_scratch_page: > cleanup_scratch_page(vm); > > @@ -1292,18 +1284,6 @@ static int gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create) > return 0; > } > > -static void gen8_free_scratch(struct i915_address_space *vm) > -{ > - if (!vm->scratch_page.daddr) > - return; > - > - if (i915_vm_is_4lvl(vm)) > - free_pd(vm, vm->scratch_pdp); > - free_pd(vm, vm->scratch_pd); > - free_pt(vm, vm->scratch_pt); > - cleanup_scratch_page(vm); > -} > - > static void gen8_ppgtt_cleanup_3lvl(struct i915_address_space *vm, > struct i915_page_directory *pdp) > { > @@ -1311,7 +1291,7 @@ static void gen8_ppgtt_cleanup_3lvl(struct i915_address_space *vm, > int i; > > for (i = 0; i < pdpes; i++) { > - if (pdp->entry[i] == vm->scratch_pd) > + if (pdp->entry[i] == &vm->scratch_pd) > continue; > > gen8_free_page_tables(vm, pdp->entry[i]); > @@ -1329,7 +1309,7 @@ static void gen8_ppgtt_cleanup_4lvl(struct i915_ppgtt *ppgtt) > for (i = 0; i < GEN8_PML4ES_PER_PML4; i++) { > struct i915_page_directory *pdp = i915_pdp_entry(pml4, i); > > - if (pdp == ppgtt->vm.scratch_pdp) > + if (px_base(pdp) == &ppgtt->vm.scratch_pdp) > continue; > > gen8_ppgtt_cleanup_3lvl(&ppgtt->vm, pdp); > @@ -1351,7 +1331,7 @@ static void gen8_ppgtt_cleanup(struct i915_address_space *vm) > else > gen8_ppgtt_cleanup_3lvl(&ppgtt->vm, ppgtt->pd); > > - gen8_free_scratch(vm); > + free_scratch(vm); > } > > static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, > @@ -1367,7 +1347,7 @@ static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, > gen8_for_each_pde(pt, pd, start, length, pde) { > const int count = gen8_pte_count(start, length); > > - if (pt == vm->scratch_pt) { > + if (px_base(pt) == &vm->scratch_pt) { > spin_unlock(&pd->lock); > > pt = fetch_and_zero(&alloc); > @@ -1379,10 +1359,10 @@ static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, > } > > if (count < GEN8_PTES || intel_vgpu_active(vm->i915)) > - gen8_initialize_pt(vm, pt); > + fill_px(pt, vm->scratch_pte); > > spin_lock(&pd->lock); > - if (pd->entry[pde] == vm->scratch_pt) { > + if (pd->entry[pde] == &vm->scratch_pt) { > set_pd_entry(pd, pde, pt); > } else { > alloc = pt; > @@ -1414,7 +1394,7 @@ static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm, > > spin_lock(&pdp->lock); > gen8_for_each_pdpe(pd, pdp, start, length, pdpe) { > - if (pd == vm->scratch_pd) { > + if (px_base(pd) == &vm->scratch_pd) { > spin_unlock(&pdp->lock); > > pd = fetch_and_zero(&alloc); > @@ -1425,10 +1405,10 @@ static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm, > goto unwind; > } > > - init_pd(vm, pd, vm->scratch_pt); > + init_pd(pd, &vm->scratch_pt); > > spin_lock(&pdp->lock); > - if (pdp->entry[pdpe] == vm->scratch_pd) { > + if (pdp->entry[pdpe] == &vm->scratch_pd) { > set_pd_entry(pdp, pdpe, pd); > } else { > alloc = pd; > @@ -1449,7 +1429,7 @@ static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm, > goto out; > > unwind_pd: > - if (release_pd_entry(pdp, pdpe, &pd->used, px_base(vm->scratch_pd))) > + if (release_pd_entry(pdp, pdpe, &pd->used, &vm->scratch_pd)) > free_pd(vm, pd); > unwind: > gen8_ppgtt_clear_pdp(vm, pdp, from, start - from); > @@ -1478,7 +1458,7 @@ static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm, > > spin_lock(&pml4->lock); > gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) { > - if (pdp == vm->scratch_pdp) { > + if (px_base(pdp) == &vm->scratch_pdp) { > spin_unlock(&pml4->lock); > > pdp = fetch_and_zero(&alloc); > @@ -1489,10 +1469,10 @@ static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm, > goto unwind; > } > > - init_pd(vm, pdp, vm->scratch_pd); > + init_pd(pdp, &vm->scratch_pd); > > spin_lock(&pml4->lock); > - if (pml4->entry[pml4e] == vm->scratch_pdp) { > + if (pml4->entry[pml4e] == &vm->scratch_pdp) { > set_pd_entry(pml4, pml4e, pdp); > } else { > alloc = pdp; > @@ -1513,7 +1493,7 @@ static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm, > goto out; > > unwind_pdp: > - if (release_pd_entry(pml4, pml4e, &pdp->used, px_base(vm->scratch_pdp))) > + if (release_pd_entry(pml4, pml4e, &pdp->used, &vm->scratch_pdp)) > free_pd(vm, pdp); > unwind: > gen8_ppgtt_clear_4lvl(vm, from, start - from); > @@ -1537,7 +1517,7 @@ static int gen8_preallocate_top_level_pdp(struct i915_ppgtt *ppgtt) > if (IS_ERR(pd)) > goto unwind; > > - init_pd(vm, pd, vm->scratch_pt); > + init_pd(pd, &vm->scratch_pt); > set_pd_entry(pdp, pdpe, pd); > } > > @@ -1568,10 +1548,10 @@ static void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt) > > static void init_pd_n(struct i915_address_space *vm, > struct i915_page_directory *pd, > - struct i915_page_directory *to, > + struct i915_page_dma *to, > const unsigned int entries) > { > - const u64 daddr = gen8_pde_encode(px_dma(to), I915_CACHE_LLC); > + const u64 daddr = gen8_pde_encode(to->daddr, I915_CACHE_LLC); > u64 * const vaddr = kmap_atomic(pd->base.page); > > memset64(vaddr, daddr, entries); > @@ -1588,7 +1568,7 @@ gen8_alloc_top_pd(struct i915_address_space *vm) > if (i915_vm_is_4lvl(vm)) { > pd = alloc_pd(vm); > if (!IS_ERR(pd)) > - init_pd(vm, pd, vm->scratch_pdp); > + init_pd(pd, &vm->scratch_pdp); > > return pd; > } > @@ -1605,7 +1585,7 @@ gen8_alloc_top_pd(struct i915_address_space *vm) > return ERR_PTR(-ENOMEM); > } > > - init_pd_n(vm, pd, vm->scratch_pd, GEN8_3LVL_PDPES); > + init_pd_n(vm, pd, &vm->scratch_pd, GEN8_3LVL_PDPES); > > return pd; > } > @@ -1678,7 +1658,7 @@ static struct i915_ppgtt *gen8_ppgtt_create(struct drm_i915_private *i915) > err_free_pd: > free_pd(&ppgtt->vm, ppgtt->pd); > err_free_scratch: > - gen8_free_scratch(&ppgtt->vm); > + free_scratch(&ppgtt->vm); > err_free: > kfree(ppgtt); > return ERR_PTR(err); > @@ -1763,7 +1743,7 @@ static void gen6_ppgtt_clear_range(struct i915_address_space *vm, > const unsigned int count = min(num_entries, GEN6_PTES - pte); > gen6_pte_t *vaddr; > > - GEM_BUG_ON(pt == vm->scratch_pt); > + GEM_BUG_ON(px_base(pt) == &vm->scratch_pt); > > num_entries -= count; > > @@ -1800,7 +1780,7 @@ static void gen6_ppgtt_insert_entries(struct i915_address_space *vm, > struct sgt_dma iter = sgt_dma(vma); > gen6_pte_t *vaddr; > > - GEM_BUG_ON(i915_pt_entry(pd, act_pt) == vm->scratch_pt); > + GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch_pt); > > vaddr = kmap_atomic_px(i915_pt_entry(pd, act_pt)); > do { > @@ -1845,7 +1825,7 @@ static int gen6_alloc_va_range(struct i915_address_space *vm, > gen6_for_each_pde(pt, pd, start, length, pde) { > const unsigned int count = gen6_pte_count(start, length); > > - if (pt == vm->scratch_pt) { > + if (px_base(pt) == &vm->scratch_pt) { > spin_unlock(&pd->lock); > > pt = fetch_and_zero(&alloc); > @@ -1856,10 +1836,10 @@ static int gen6_alloc_va_range(struct i915_address_space *vm, > goto unwind_out; > } > > - gen6_initialize_pt(vm, pt); > + fill32_px(pt, vm->scratch_pte); > > spin_lock(&pd->lock); > - if (pd->entry[pde] == vm->scratch_pt) { > + if (pd->entry[pde] == &vm->scratch_pt) { > pd->entry[pde] = pt; > if (i915_vma_is_bound(ppgtt->vma, > I915_VMA_GLOBAL_BIND)) { > @@ -1908,26 +1888,18 @@ static int gen6_ppgtt_init_scratch(struct gen6_ppgtt *ppgtt) > I915_CACHE_NONE, > PTE_READ_ONLY); > > - vm->scratch_pt = alloc_pt(vm); > - if (IS_ERR(vm->scratch_pt)) { > + if (unlikely(setup_page_dma(vm, &vm->scratch_pt))) { > cleanup_scratch_page(vm); > - return PTR_ERR(vm->scratch_pt); > + return -ENOMEM; > } > - > - gen6_initialize_pt(vm, vm->scratch_pt); > + fill_page_dma_32(&vm->scratch_pt, vm->scratch_pte); > > gen6_for_all_pdes(unused, pd, pde) > - pd->entry[pde] = vm->scratch_pt; > + pd->entry[pde] = &vm->scratch_pt; > > return 0; > } > > -static void gen6_ppgtt_free_scratch(struct i915_address_space *vm) > -{ > - free_pt(vm, vm->scratch_pt); > - cleanup_scratch_page(vm); > -} > - > static void gen6_ppgtt_free_pd(struct gen6_ppgtt *ppgtt) > { > struct i915_page_directory * const pd = ppgtt->base.pd; > @@ -1935,7 +1907,7 @@ static void gen6_ppgtt_free_pd(struct gen6_ppgtt *ppgtt) > u32 pde; > > gen6_for_all_pdes(pt, pd, pde) > - if (pt != ppgtt->base.vm.scratch_pt) > + if (px_base(pt) != &ppgtt->base.vm.scratch_pt) > free_pt(&ppgtt->base.vm, pt); > } > > @@ -1950,7 +1922,7 @@ static void gen6_ppgtt_cleanup(struct i915_address_space *vm) > mutex_unlock(&i915->drm.struct_mutex); > > gen6_ppgtt_free_pd(ppgtt); > - gen6_ppgtt_free_scratch(vm); > + free_scratch(vm); > kfree(ppgtt->base.pd); > } > > @@ -1993,7 +1965,7 @@ static void pd_vma_unbind(struct i915_vma *vma) > { > struct gen6_ppgtt *ppgtt = vma->private; > struct i915_page_directory * const pd = ppgtt->base.pd; > - struct i915_page_table * const scratch_pt = ppgtt->base.vm.scratch_pt; > + struct i915_page_dma * const scratch = &ppgtt->base.vm.scratch_pt; > struct i915_page_table *pt; > unsigned int pde; > > @@ -2002,11 +1974,11 @@ static void pd_vma_unbind(struct i915_vma *vma) > > /* Free all no longer used page tables */ > gen6_for_all_pdes(pt, ppgtt->base.pd, pde) { > - if (atomic_read(&pt->used) || pt == scratch_pt) > + if (px_base(pt) == scratch || atomic_read(&pt->used)) > continue; > > free_pt(&ppgtt->base.vm, pt); > - pd->entry[pde] = scratch_pt; > + pd->entry[pde] = scratch; > } > > ppgtt->scan_for_unused_pt = false; > @@ -2148,7 +2120,7 @@ static struct i915_ppgtt *gen6_ppgtt_create(struct drm_i915_private *i915) > return &ppgtt->base; > > err_scratch: > - gen6_ppgtt_free_scratch(&ppgtt->base.vm); > + free_scratch(&ppgtt->base.vm); > err_pd: > kfree(ppgtt->base.pd); > err_free: > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h > index 57a68ef4eda7..860850411a1b 100644 > --- a/drivers/gpu/drm/i915/i915_gem_gtt.h > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h > @@ -304,9 +304,9 @@ struct i915_address_space { > u64 scratch_pte; > int scratch_order; > struct i915_page_dma scratch_page; > - struct i915_page_table *scratch_pt; > - struct i915_page_directory *scratch_pd; > - struct i915_page_directory *scratch_pdp; /* GEN8+ & 48b PPGTT */ > + struct i915_page_dma scratch_pt; > + struct i915_page_dma scratch_pd; > + struct i915_page_dma scratch_pdp; /* GEN8+ & 48b PPGTT */ > > /** > * List of vma currently bound. > -- > 2.20.1
Quoting Mika Kuoppala (2019-07-09 13:24:27) > Chris Wilson <chris@chris-wilson.co.uk> writes: > > > We only use the dma pages for scratch, and so do not need to allocate > > the extra storage for the shadow page directory. > > > > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> > > Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com> > > --- > > drivers/gpu/drm/i915/i915_gem_gtt.c | 192 ++++++++++++---------------- > > drivers/gpu/drm/i915/i915_gem_gtt.h | 6 +- > > 2 files changed, 85 insertions(+), 113 deletions(-) > > > > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c > > index 236c964dd761..937236913e70 100644 > > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c > > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c > > @@ -594,25 +594,17 @@ static void cleanup_page_dma(struct i915_address_space *vm, > > > > #define kmap_atomic_px(px) kmap_atomic(px_base(px)->page) > > > > -#define fill_px(vm, px, v) fill_page_dma((vm), px_base(px), (v)) > > -#define fill32_px(vm, px, v) fill_page_dma_32((vm), px_base(px), (v)) > > +#define fill_px(px, v) fill_page_dma(px_base(px), (v)) > > +#define fill32_px(px, v) fill_page_dma_32(px_base(px), (v)) > > > > -static void fill_page_dma(struct i915_address_space *vm, > > - struct i915_page_dma *p, > > - const u64 val) > > +static void fill_page_dma(struct i915_page_dma *p, const u64 val) > > { > > - u64 * const vaddr = kmap_atomic(p->page); > > - > > - memset64(vaddr, val, PAGE_SIZE / sizeof(val)); > > - > > - kunmap_atomic(vaddr); > > + kunmap_atomic(memset64(kmap_atomic(p->page), val, I915_PDES)); > > Neat. > > I would go for 512 instead of I915_PDES. There is no magic > and there never will be magic as it is as const as carved into stone. I was just going with I915_PDES and I915_PDE_MASK throughout. Later this one becomes count, fwiw. > > > } > > > > -static void fill_page_dma_32(struct i915_address_space *vm, > > - struct i915_page_dma *p, > > - const u32 v) > > +static void fill_page_dma_32(struct i915_page_dma *p, const u32 v) > > { > > - fill_page_dma(vm, p, (u64)v << 32 | v); > > + fill_page_dma(p, (u64)v << 32 | v); > > } > > > > static int > > @@ -687,6 +679,21 @@ static void cleanup_scratch_page(struct i915_address_space *vm) > > __free_pages(p->page, order); > > } > > > > +static void free_scratch(struct i915_address_space *vm) > > +{ > > + if (!vm->scratch_page.daddr) /* set to 0 on clones */ > > + return; > > + > > + if (vm->scratch_pdp.daddr) > > + cleanup_page_dma(vm, &vm->scratch_pdp); > > + if (vm->scratch_pd.daddr) > > + cleanup_page_dma(vm, &vm->scratch_pd); > > + if (vm->scratch_pt.daddr) > > + cleanup_page_dma(vm, &vm->scratch_pt); > > + > > + cleanup_scratch_page(vm); > > +} > > + > > static struct i915_page_table *alloc_pt(struct i915_address_space *vm) > > { > > struct i915_page_table *pt; > > @@ -711,18 +718,6 @@ static void free_pt(struct i915_address_space *vm, struct i915_page_table *pt) > > kfree(pt); > > } > > > > -static void gen8_initialize_pt(struct i915_address_space *vm, > > - struct i915_page_table *pt) > > -{ > > - fill_px(vm, pt, vm->scratch_pte); > > -} > > - > > -static void gen6_initialize_pt(struct i915_address_space *vm, > > - struct i915_page_table *pt) > > -{ > > - fill32_px(vm, pt, vm->scratch_pte); > > -} > > - > > static struct i915_page_directory *__alloc_pd(void) > > { > > struct i915_page_directory *pd; > > @@ -765,9 +760,11 @@ static void free_pd(struct i915_address_space *vm, > > kfree(pd); > > } > > > > -#define init_pd(vm, pd, to) { \ > > - fill_px((vm), (pd), gen8_pde_encode(px_dma(to), I915_CACHE_LLC)); \ > > - memset_p((pd)->entry, (to), 512); \ > > +static void init_pd(struct i915_page_directory *pd, > > + struct i915_page_dma *scratch) > > +{ > > + fill_px(pd, gen8_pde_encode(scratch->daddr, I915_CACHE_LLC)); > > + memset_p(pd->entry, scratch, 512); > > } > > > > static inline void > > @@ -869,12 +866,11 @@ static void gen8_ppgtt_clear_pd(struct i915_address_space *vm, > > u32 pde; > > > > gen8_for_each_pde(pt, pd, start, length, pde) { > > - GEM_BUG_ON(pt == vm->scratch_pt); > > + GEM_BUG_ON(px_base(pt) == &vm->scratch_pt); > > > > atomic_inc(&pt->used); > > gen8_ppgtt_clear_pt(vm, pt, start, length); > > - if (release_pd_entry(pd, pde, &pt->used, > > - px_base(vm->scratch_pt))) > > + if (release_pd_entry(pd, pde, &pt->used, &vm->scratch_pt)) > > free_pt(vm, pt); > > } > > } > > @@ -890,12 +886,11 @@ static void gen8_ppgtt_clear_pdp(struct i915_address_space *vm, > > unsigned int pdpe; > > > > gen8_for_each_pdpe(pd, pdp, start, length, pdpe) { > > - GEM_BUG_ON(pd == vm->scratch_pd); > > + GEM_BUG_ON(px_base(pd) == &vm->scratch_pd); > > Perhaps future will bring pd_points_scratch(pd). > > Now the intriguing, bordering irritating, question in my mind is > that can we fold the scratch_pd and scratch_pdp to be the same thing. No, we can fold the scratch_pd to be the same (dma wise) as they do need to end up at the scratch_pte. And sadly we can't use the scratch_pte as the filler for scratch_pd. > Patch lgtm with some dislike towards I915_PDES, I'm not keen on it tbh. But the mix of alternating between 512/0x1ff does suggest to use some name. -Chris
Chris Wilson <chris@chris-wilson.co.uk> writes: > Quoting Mika Kuoppala (2019-07-09 13:24:27) >> Chris Wilson <chris@chris-wilson.co.uk> writes: >> >> > We only use the dma pages for scratch, and so do not need to allocate >> > the extra storage for the shadow page directory. >> > >> > Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> >> > Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com> >> > --- >> > drivers/gpu/drm/i915/i915_gem_gtt.c | 192 ++++++++++++---------------- >> > drivers/gpu/drm/i915/i915_gem_gtt.h | 6 +- >> > 2 files changed, 85 insertions(+), 113 deletions(-) >> > >> > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c >> > index 236c964dd761..937236913e70 100644 >> > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c >> > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c >> > @@ -594,25 +594,17 @@ static void cleanup_page_dma(struct i915_address_space *vm, >> > >> > #define kmap_atomic_px(px) kmap_atomic(px_base(px)->page) >> > >> > -#define fill_px(vm, px, v) fill_page_dma((vm), px_base(px), (v)) >> > -#define fill32_px(vm, px, v) fill_page_dma_32((vm), px_base(px), (v)) >> > +#define fill_px(px, v) fill_page_dma(px_base(px), (v)) >> > +#define fill32_px(px, v) fill_page_dma_32(px_base(px), (v)) >> > >> > -static void fill_page_dma(struct i915_address_space *vm, >> > - struct i915_page_dma *p, >> > - const u64 val) >> > +static void fill_page_dma(struct i915_page_dma *p, const u64 val) >> > { >> > - u64 * const vaddr = kmap_atomic(p->page); >> > - >> > - memset64(vaddr, val, PAGE_SIZE / sizeof(val)); >> > - >> > - kunmap_atomic(vaddr); >> > + kunmap_atomic(memset64(kmap_atomic(p->page), val, I915_PDES)); >> >> Neat. >> >> I would go for 512 instead of I915_PDES. There is no magic >> and there never will be magic as it is as const as carved into stone. > > I was just going with I915_PDES and I915_PDE_MASK throughout. Later this > one becomes count, fwiw. >> >> > } >> > >> > -static void fill_page_dma_32(struct i915_address_space *vm, >> > - struct i915_page_dma *p, >> > - const u32 v) >> > +static void fill_page_dma_32(struct i915_page_dma *p, const u32 v) >> > { >> > - fill_page_dma(vm, p, (u64)v << 32 | v); >> > + fill_page_dma(p, (u64)v << 32 | v); >> > } >> > >> > static int >> > @@ -687,6 +679,21 @@ static void cleanup_scratch_page(struct i915_address_space *vm) >> > __free_pages(p->page, order); >> > } >> > >> > +static void free_scratch(struct i915_address_space *vm) >> > +{ >> > + if (!vm->scratch_page.daddr) /* set to 0 on clones */ >> > + return; >> > + >> > + if (vm->scratch_pdp.daddr) >> > + cleanup_page_dma(vm, &vm->scratch_pdp); >> > + if (vm->scratch_pd.daddr) >> > + cleanup_page_dma(vm, &vm->scratch_pd); >> > + if (vm->scratch_pt.daddr) >> > + cleanup_page_dma(vm, &vm->scratch_pt); >> > + >> > + cleanup_scratch_page(vm); >> > +} >> > + >> > static struct i915_page_table *alloc_pt(struct i915_address_space *vm) >> > { >> > struct i915_page_table *pt; >> > @@ -711,18 +718,6 @@ static void free_pt(struct i915_address_space *vm, struct i915_page_table *pt) >> > kfree(pt); >> > } >> > >> > -static void gen8_initialize_pt(struct i915_address_space *vm, >> > - struct i915_page_table *pt) >> > -{ >> > - fill_px(vm, pt, vm->scratch_pte); >> > -} >> > - >> > -static void gen6_initialize_pt(struct i915_address_space *vm, >> > - struct i915_page_table *pt) >> > -{ >> > - fill32_px(vm, pt, vm->scratch_pte); >> > -} >> > - >> > static struct i915_page_directory *__alloc_pd(void) >> > { >> > struct i915_page_directory *pd; >> > @@ -765,9 +760,11 @@ static void free_pd(struct i915_address_space *vm, >> > kfree(pd); >> > } >> > >> > -#define init_pd(vm, pd, to) { \ >> > - fill_px((vm), (pd), gen8_pde_encode(px_dma(to), I915_CACHE_LLC)); \ >> > - memset_p((pd)->entry, (to), 512); \ >> > +static void init_pd(struct i915_page_directory *pd, >> > + struct i915_page_dma *scratch) >> > +{ >> > + fill_px(pd, gen8_pde_encode(scratch->daddr, I915_CACHE_LLC)); >> > + memset_p(pd->entry, scratch, 512); >> > } >> > >> > static inline void >> > @@ -869,12 +866,11 @@ static void gen8_ppgtt_clear_pd(struct i915_address_space *vm, >> > u32 pde; >> > >> > gen8_for_each_pde(pt, pd, start, length, pde) { >> > - GEM_BUG_ON(pt == vm->scratch_pt); >> > + GEM_BUG_ON(px_base(pt) == &vm->scratch_pt); >> > >> > atomic_inc(&pt->used); >> > gen8_ppgtt_clear_pt(vm, pt, start, length); >> > - if (release_pd_entry(pd, pde, &pt->used, >> > - px_base(vm->scratch_pt))) >> > + if (release_pd_entry(pd, pde, &pt->used, &vm->scratch_pt)) >> > free_pt(vm, pt); >> > } >> > } >> > @@ -890,12 +886,11 @@ static void gen8_ppgtt_clear_pdp(struct i915_address_space *vm, >> > unsigned int pdpe; >> > >> > gen8_for_each_pdpe(pd, pdp, start, length, pdpe) { >> > - GEM_BUG_ON(pd == vm->scratch_pd); >> > + GEM_BUG_ON(px_base(pd) == &vm->scratch_pd); >> >> Perhaps future will bring pd_points_scratch(pd). >> >> Now the intriguing, bordering irritating, question in my mind is >> that can we fold the scratch_pd and scratch_pdp to be the same thing. > > No, we can fold the scratch_pd to be the same (dma wise) as they do need > to end up at the scratch_pte. And sadly we can't use the scratch_pte as > the filler for scratch_pd. Oh indeed. it needs to be hierarchy even on upper level to break out. *blush* > >> Patch lgtm with some dislike towards I915_PDES, > > I'm not keen on it tbh. But the mix of alternating between 512/0x1ff > does suggest to use some name. 512 everywhere. -Mika
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c index 236c964dd761..937236913e70 100644 --- a/drivers/gpu/drm/i915/i915_gem_gtt.c +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c @@ -594,25 +594,17 @@ static void cleanup_page_dma(struct i915_address_space *vm, #define kmap_atomic_px(px) kmap_atomic(px_base(px)->page) -#define fill_px(vm, px, v) fill_page_dma((vm), px_base(px), (v)) -#define fill32_px(vm, px, v) fill_page_dma_32((vm), px_base(px), (v)) +#define fill_px(px, v) fill_page_dma(px_base(px), (v)) +#define fill32_px(px, v) fill_page_dma_32(px_base(px), (v)) -static void fill_page_dma(struct i915_address_space *vm, - struct i915_page_dma *p, - const u64 val) +static void fill_page_dma(struct i915_page_dma *p, const u64 val) { - u64 * const vaddr = kmap_atomic(p->page); - - memset64(vaddr, val, PAGE_SIZE / sizeof(val)); - - kunmap_atomic(vaddr); + kunmap_atomic(memset64(kmap_atomic(p->page), val, I915_PDES)); } -static void fill_page_dma_32(struct i915_address_space *vm, - struct i915_page_dma *p, - const u32 v) +static void fill_page_dma_32(struct i915_page_dma *p, const u32 v) { - fill_page_dma(vm, p, (u64)v << 32 | v); + fill_page_dma(p, (u64)v << 32 | v); } static int @@ -687,6 +679,21 @@ static void cleanup_scratch_page(struct i915_address_space *vm) __free_pages(p->page, order); } +static void free_scratch(struct i915_address_space *vm) +{ + if (!vm->scratch_page.daddr) /* set to 0 on clones */ + return; + + if (vm->scratch_pdp.daddr) + cleanup_page_dma(vm, &vm->scratch_pdp); + if (vm->scratch_pd.daddr) + cleanup_page_dma(vm, &vm->scratch_pd); + if (vm->scratch_pt.daddr) + cleanup_page_dma(vm, &vm->scratch_pt); + + cleanup_scratch_page(vm); +} + static struct i915_page_table *alloc_pt(struct i915_address_space *vm) { struct i915_page_table *pt; @@ -711,18 +718,6 @@ static void free_pt(struct i915_address_space *vm, struct i915_page_table *pt) kfree(pt); } -static void gen8_initialize_pt(struct i915_address_space *vm, - struct i915_page_table *pt) -{ - fill_px(vm, pt, vm->scratch_pte); -} - -static void gen6_initialize_pt(struct i915_address_space *vm, - struct i915_page_table *pt) -{ - fill32_px(vm, pt, vm->scratch_pte); -} - static struct i915_page_directory *__alloc_pd(void) { struct i915_page_directory *pd; @@ -765,9 +760,11 @@ static void free_pd(struct i915_address_space *vm, kfree(pd); } -#define init_pd(vm, pd, to) { \ - fill_px((vm), (pd), gen8_pde_encode(px_dma(to), I915_CACHE_LLC)); \ - memset_p((pd)->entry, (to), 512); \ +static void init_pd(struct i915_page_directory *pd, + struct i915_page_dma *scratch) +{ + fill_px(pd, gen8_pde_encode(scratch->daddr, I915_CACHE_LLC)); + memset_p(pd->entry, scratch, 512); } static inline void @@ -869,12 +866,11 @@ static void gen8_ppgtt_clear_pd(struct i915_address_space *vm, u32 pde; gen8_for_each_pde(pt, pd, start, length, pde) { - GEM_BUG_ON(pt == vm->scratch_pt); + GEM_BUG_ON(px_base(pt) == &vm->scratch_pt); atomic_inc(&pt->used); gen8_ppgtt_clear_pt(vm, pt, start, length); - if (release_pd_entry(pd, pde, &pt->used, - px_base(vm->scratch_pt))) + if (release_pd_entry(pd, pde, &pt->used, &vm->scratch_pt)) free_pt(vm, pt); } } @@ -890,12 +886,11 @@ static void gen8_ppgtt_clear_pdp(struct i915_address_space *vm, unsigned int pdpe; gen8_for_each_pdpe(pd, pdp, start, length, pdpe) { - GEM_BUG_ON(pd == vm->scratch_pd); + GEM_BUG_ON(px_base(pd) == &vm->scratch_pd); atomic_inc(&pd->used); gen8_ppgtt_clear_pd(vm, pd, start, length); - if (release_pd_entry(pdp, pdpe, &pd->used, - px_base(vm->scratch_pd))) + if (release_pd_entry(pdp, pdpe, &pd->used, &vm->scratch_pd)) free_pd(vm, pd); } } @@ -921,12 +916,11 @@ static void gen8_ppgtt_clear_4lvl(struct i915_address_space *vm, GEM_BUG_ON(!i915_vm_is_4lvl(vm)); gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) { - GEM_BUG_ON(pdp == vm->scratch_pdp); + GEM_BUG_ON(px_base(pdp) == &vm->scratch_pdp); atomic_inc(&pdp->used); gen8_ppgtt_clear_pdp(vm, pdp, start, length); - if (release_pd_entry(pml4, pml4e, &pdp->used, - px_base(vm->scratch_pdp))) + if (release_pd_entry(pml4, pml4e, &pdp->used, &vm->scratch_pdp)) free_pd(vm, pdp); } } @@ -1181,7 +1175,7 @@ static void gen8_free_page_tables(struct i915_address_space *vm, int i; for (i = 0; i < I915_PDES; i++) { - if (pd->entry[i] != vm->scratch_pt) + if (pd->entry[i] != &vm->scratch_pt) free_pt(vm, pd->entry[i]); } } @@ -1218,37 +1212,35 @@ static int gen8_init_scratch(struct i915_address_space *vm) I915_CACHE_LLC, vm->has_read_only); - vm->scratch_pt = alloc_pt(vm); - if (IS_ERR(vm->scratch_pt)) { - ret = PTR_ERR(vm->scratch_pt); + if (unlikely(setup_page_dma(vm, &vm->scratch_pt))) { + ret = -ENOMEM; goto free_scratch_page; } + fill_page_dma(&vm->scratch_pt, vm->scratch_pte); - vm->scratch_pd = alloc_pd(vm); - if (IS_ERR(vm->scratch_pd)) { - ret = PTR_ERR(vm->scratch_pd); + if (unlikely(setup_page_dma(vm, &vm->scratch_pd))) { + ret = -ENOMEM; goto free_pt; } + fill_page_dma(&vm->scratch_pd, + gen8_pde_encode(vm->scratch_pd.daddr, I915_CACHE_LLC)); if (i915_vm_is_4lvl(vm)) { - vm->scratch_pdp = alloc_pd(vm); - if (IS_ERR(vm->scratch_pdp)) { - ret = PTR_ERR(vm->scratch_pdp); + if (unlikely(setup_page_dma(vm, &vm->scratch_pdp))) { + ret = -ENOMEM; goto free_pd; } + fill_page_dma(&vm->scratch_pdp, + gen8_pde_encode(vm->scratch_pdp.daddr, + I915_CACHE_LLC)); } - gen8_initialize_pt(vm, vm->scratch_pt); - init_pd(vm, vm->scratch_pd, vm->scratch_pt); - if (i915_vm_is_4lvl(vm)) - init_pd(vm, vm->scratch_pdp, vm->scratch_pd); - return 0; free_pd: - free_pd(vm, vm->scratch_pd); + cleanup_page_dma(vm, &vm->scratch_pd); free_pt: - free_pt(vm, vm->scratch_pt); + cleanup_page_dma(vm, &vm->scratch_pt); free_scratch_page: cleanup_scratch_page(vm); @@ -1292,18 +1284,6 @@ static int gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create) return 0; } -static void gen8_free_scratch(struct i915_address_space *vm) -{ - if (!vm->scratch_page.daddr) - return; - - if (i915_vm_is_4lvl(vm)) - free_pd(vm, vm->scratch_pdp); - free_pd(vm, vm->scratch_pd); - free_pt(vm, vm->scratch_pt); - cleanup_scratch_page(vm); -} - static void gen8_ppgtt_cleanup_3lvl(struct i915_address_space *vm, struct i915_page_directory *pdp) { @@ -1311,7 +1291,7 @@ static void gen8_ppgtt_cleanup_3lvl(struct i915_address_space *vm, int i; for (i = 0; i < pdpes; i++) { - if (pdp->entry[i] == vm->scratch_pd) + if (pdp->entry[i] == &vm->scratch_pd) continue; gen8_free_page_tables(vm, pdp->entry[i]); @@ -1329,7 +1309,7 @@ static void gen8_ppgtt_cleanup_4lvl(struct i915_ppgtt *ppgtt) for (i = 0; i < GEN8_PML4ES_PER_PML4; i++) { struct i915_page_directory *pdp = i915_pdp_entry(pml4, i); - if (pdp == ppgtt->vm.scratch_pdp) + if (px_base(pdp) == &ppgtt->vm.scratch_pdp) continue; gen8_ppgtt_cleanup_3lvl(&ppgtt->vm, pdp); @@ -1351,7 +1331,7 @@ static void gen8_ppgtt_cleanup(struct i915_address_space *vm) else gen8_ppgtt_cleanup_3lvl(&ppgtt->vm, ppgtt->pd); - gen8_free_scratch(vm); + free_scratch(vm); } static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, @@ -1367,7 +1347,7 @@ static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, gen8_for_each_pde(pt, pd, start, length, pde) { const int count = gen8_pte_count(start, length); - if (pt == vm->scratch_pt) { + if (px_base(pt) == &vm->scratch_pt) { spin_unlock(&pd->lock); pt = fetch_and_zero(&alloc); @@ -1379,10 +1359,10 @@ static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, } if (count < GEN8_PTES || intel_vgpu_active(vm->i915)) - gen8_initialize_pt(vm, pt); + fill_px(pt, vm->scratch_pte); spin_lock(&pd->lock); - if (pd->entry[pde] == vm->scratch_pt) { + if (pd->entry[pde] == &vm->scratch_pt) { set_pd_entry(pd, pde, pt); } else { alloc = pt; @@ -1414,7 +1394,7 @@ static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm, spin_lock(&pdp->lock); gen8_for_each_pdpe(pd, pdp, start, length, pdpe) { - if (pd == vm->scratch_pd) { + if (px_base(pd) == &vm->scratch_pd) { spin_unlock(&pdp->lock); pd = fetch_and_zero(&alloc); @@ -1425,10 +1405,10 @@ static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm, goto unwind; } - init_pd(vm, pd, vm->scratch_pt); + init_pd(pd, &vm->scratch_pt); spin_lock(&pdp->lock); - if (pdp->entry[pdpe] == vm->scratch_pd) { + if (pdp->entry[pdpe] == &vm->scratch_pd) { set_pd_entry(pdp, pdpe, pd); } else { alloc = pd; @@ -1449,7 +1429,7 @@ static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm, goto out; unwind_pd: - if (release_pd_entry(pdp, pdpe, &pd->used, px_base(vm->scratch_pd))) + if (release_pd_entry(pdp, pdpe, &pd->used, &vm->scratch_pd)) free_pd(vm, pd); unwind: gen8_ppgtt_clear_pdp(vm, pdp, from, start - from); @@ -1478,7 +1458,7 @@ static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm, spin_lock(&pml4->lock); gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) { - if (pdp == vm->scratch_pdp) { + if (px_base(pdp) == &vm->scratch_pdp) { spin_unlock(&pml4->lock); pdp = fetch_and_zero(&alloc); @@ -1489,10 +1469,10 @@ static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm, goto unwind; } - init_pd(vm, pdp, vm->scratch_pd); + init_pd(pdp, &vm->scratch_pd); spin_lock(&pml4->lock); - if (pml4->entry[pml4e] == vm->scratch_pdp) { + if (pml4->entry[pml4e] == &vm->scratch_pdp) { set_pd_entry(pml4, pml4e, pdp); } else { alloc = pdp; @@ -1513,7 +1493,7 @@ static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm, goto out; unwind_pdp: - if (release_pd_entry(pml4, pml4e, &pdp->used, px_base(vm->scratch_pdp))) + if (release_pd_entry(pml4, pml4e, &pdp->used, &vm->scratch_pdp)) free_pd(vm, pdp); unwind: gen8_ppgtt_clear_4lvl(vm, from, start - from); @@ -1537,7 +1517,7 @@ static int gen8_preallocate_top_level_pdp(struct i915_ppgtt *ppgtt) if (IS_ERR(pd)) goto unwind; - init_pd(vm, pd, vm->scratch_pt); + init_pd(pd, &vm->scratch_pt); set_pd_entry(pdp, pdpe, pd); } @@ -1568,10 +1548,10 @@ static void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt) static void init_pd_n(struct i915_address_space *vm, struct i915_page_directory *pd, - struct i915_page_directory *to, + struct i915_page_dma *to, const unsigned int entries) { - const u64 daddr = gen8_pde_encode(px_dma(to), I915_CACHE_LLC); + const u64 daddr = gen8_pde_encode(to->daddr, I915_CACHE_LLC); u64 * const vaddr = kmap_atomic(pd->base.page); memset64(vaddr, daddr, entries); @@ -1588,7 +1568,7 @@ gen8_alloc_top_pd(struct i915_address_space *vm) if (i915_vm_is_4lvl(vm)) { pd = alloc_pd(vm); if (!IS_ERR(pd)) - init_pd(vm, pd, vm->scratch_pdp); + init_pd(pd, &vm->scratch_pdp); return pd; } @@ -1605,7 +1585,7 @@ gen8_alloc_top_pd(struct i915_address_space *vm) return ERR_PTR(-ENOMEM); } - init_pd_n(vm, pd, vm->scratch_pd, GEN8_3LVL_PDPES); + init_pd_n(vm, pd, &vm->scratch_pd, GEN8_3LVL_PDPES); return pd; } @@ -1678,7 +1658,7 @@ static struct i915_ppgtt *gen8_ppgtt_create(struct drm_i915_private *i915) err_free_pd: free_pd(&ppgtt->vm, ppgtt->pd); err_free_scratch: - gen8_free_scratch(&ppgtt->vm); + free_scratch(&ppgtt->vm); err_free: kfree(ppgtt); return ERR_PTR(err); @@ -1763,7 +1743,7 @@ static void gen6_ppgtt_clear_range(struct i915_address_space *vm, const unsigned int count = min(num_entries, GEN6_PTES - pte); gen6_pte_t *vaddr; - GEM_BUG_ON(pt == vm->scratch_pt); + GEM_BUG_ON(px_base(pt) == &vm->scratch_pt); num_entries -= count; @@ -1800,7 +1780,7 @@ static void gen6_ppgtt_insert_entries(struct i915_address_space *vm, struct sgt_dma iter = sgt_dma(vma); gen6_pte_t *vaddr; - GEM_BUG_ON(i915_pt_entry(pd, act_pt) == vm->scratch_pt); + GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch_pt); vaddr = kmap_atomic_px(i915_pt_entry(pd, act_pt)); do { @@ -1845,7 +1825,7 @@ static int gen6_alloc_va_range(struct i915_address_space *vm, gen6_for_each_pde(pt, pd, start, length, pde) { const unsigned int count = gen6_pte_count(start, length); - if (pt == vm->scratch_pt) { + if (px_base(pt) == &vm->scratch_pt) { spin_unlock(&pd->lock); pt = fetch_and_zero(&alloc); @@ -1856,10 +1836,10 @@ static int gen6_alloc_va_range(struct i915_address_space *vm, goto unwind_out; } - gen6_initialize_pt(vm, pt); + fill32_px(pt, vm->scratch_pte); spin_lock(&pd->lock); - if (pd->entry[pde] == vm->scratch_pt) { + if (pd->entry[pde] == &vm->scratch_pt) { pd->entry[pde] = pt; if (i915_vma_is_bound(ppgtt->vma, I915_VMA_GLOBAL_BIND)) { @@ -1908,26 +1888,18 @@ static int gen6_ppgtt_init_scratch(struct gen6_ppgtt *ppgtt) I915_CACHE_NONE, PTE_READ_ONLY); - vm->scratch_pt = alloc_pt(vm); - if (IS_ERR(vm->scratch_pt)) { + if (unlikely(setup_page_dma(vm, &vm->scratch_pt))) { cleanup_scratch_page(vm); - return PTR_ERR(vm->scratch_pt); + return -ENOMEM; } - - gen6_initialize_pt(vm, vm->scratch_pt); + fill_page_dma_32(&vm->scratch_pt, vm->scratch_pte); gen6_for_all_pdes(unused, pd, pde) - pd->entry[pde] = vm->scratch_pt; + pd->entry[pde] = &vm->scratch_pt; return 0; } -static void gen6_ppgtt_free_scratch(struct i915_address_space *vm) -{ - free_pt(vm, vm->scratch_pt); - cleanup_scratch_page(vm); -} - static void gen6_ppgtt_free_pd(struct gen6_ppgtt *ppgtt) { struct i915_page_directory * const pd = ppgtt->base.pd; @@ -1935,7 +1907,7 @@ static void gen6_ppgtt_free_pd(struct gen6_ppgtt *ppgtt) u32 pde; gen6_for_all_pdes(pt, pd, pde) - if (pt != ppgtt->base.vm.scratch_pt) + if (px_base(pt) != &ppgtt->base.vm.scratch_pt) free_pt(&ppgtt->base.vm, pt); } @@ -1950,7 +1922,7 @@ static void gen6_ppgtt_cleanup(struct i915_address_space *vm) mutex_unlock(&i915->drm.struct_mutex); gen6_ppgtt_free_pd(ppgtt); - gen6_ppgtt_free_scratch(vm); + free_scratch(vm); kfree(ppgtt->base.pd); } @@ -1993,7 +1965,7 @@ static void pd_vma_unbind(struct i915_vma *vma) { struct gen6_ppgtt *ppgtt = vma->private; struct i915_page_directory * const pd = ppgtt->base.pd; - struct i915_page_table * const scratch_pt = ppgtt->base.vm.scratch_pt; + struct i915_page_dma * const scratch = &ppgtt->base.vm.scratch_pt; struct i915_page_table *pt; unsigned int pde; @@ -2002,11 +1974,11 @@ static void pd_vma_unbind(struct i915_vma *vma) /* Free all no longer used page tables */ gen6_for_all_pdes(pt, ppgtt->base.pd, pde) { - if (atomic_read(&pt->used) || pt == scratch_pt) + if (px_base(pt) == scratch || atomic_read(&pt->used)) continue; free_pt(&ppgtt->base.vm, pt); - pd->entry[pde] = scratch_pt; + pd->entry[pde] = scratch; } ppgtt->scan_for_unused_pt = false; @@ -2148,7 +2120,7 @@ static struct i915_ppgtt *gen6_ppgtt_create(struct drm_i915_private *i915) return &ppgtt->base; err_scratch: - gen6_ppgtt_free_scratch(&ppgtt->base.vm); + free_scratch(&ppgtt->base.vm); err_pd: kfree(ppgtt->base.pd); err_free: diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h index 57a68ef4eda7..860850411a1b 100644 --- a/drivers/gpu/drm/i915/i915_gem_gtt.h +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h @@ -304,9 +304,9 @@ struct i915_address_space { u64 scratch_pte; int scratch_order; struct i915_page_dma scratch_page; - struct i915_page_table *scratch_pt; - struct i915_page_directory *scratch_pd; - struct i915_page_directory *scratch_pdp; /* GEN8+ & 48b PPGTT */ + struct i915_page_dma scratch_pt; + struct i915_page_dma scratch_pd; + struct i915_page_dma scratch_pdp; /* GEN8+ & 48b PPGTT */ /** * List of vma currently bound.
We only use the dma pages for scratch, and so do not need to allocate the extra storage for the shadow page directory. Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com> --- drivers/gpu/drm/i915/i915_gem_gtt.c | 192 ++++++++++++---------------- drivers/gpu/drm/i915/i915_gem_gtt.h | 6 +- 2 files changed, 85 insertions(+), 113 deletions(-)