diff mbox

[09/26] drm/i915: Split out gtt specific header file

Message ID 1395121738-29126-10-git-send-email-benjamin.widawsky@intel.com (mailing list archive)
State New, archived
Headers show

Commit Message

Ben Widawsky March 18, 2014, 5:48 a.m. UTC
TODO: Do header files need a copyright?

Signed-off-by: Ben Widawsky <ben@bwidawsk.net>
---
 drivers/gpu/drm/i915/i915_drv.h     | 162 +-------------------------
 drivers/gpu/drm/i915/i915_gem_gtt.c |  57 ---------
 drivers/gpu/drm/i915/i915_gem_gtt.h | 225 ++++++++++++++++++++++++++++++++++++
 3 files changed, 227 insertions(+), 217 deletions(-)
 create mode 100644 drivers/gpu/drm/i915/i915_gem_gtt.h

Comments

Chris Wilson March 18, 2014, 8:46 a.m. UTC | #1
On Mon, Mar 17, 2014 at 10:48:41PM -0700, Ben Widawsky wrote:
> TODO: Do header files need a copyright?

Short answer: yes.
-Chris
Daniel Vetter March 18, 2014, 9:15 a.m. UTC | #2
On Mon, Mar 17, 2014 at 10:48:41PM -0700, Ben Widawsky wrote:
> TODO: Do header files need a copyright?

Yup ;-)

I like this though, especially since finer-grained files will make
kerneldoc inclusion (well, grouped into sensible chapters at least) much
simpler.
-Daniel

> 
> Signed-off-by: Ben Widawsky <ben@bwidawsk.net>
> ---
>  drivers/gpu/drm/i915/i915_drv.h     | 162 +-------------------------
>  drivers/gpu/drm/i915/i915_gem_gtt.c |  57 ---------
>  drivers/gpu/drm/i915/i915_gem_gtt.h | 225 ++++++++++++++++++++++++++++++++++++
>  3 files changed, 227 insertions(+), 217 deletions(-)
>  create mode 100644 drivers/gpu/drm/i915/i915_gem_gtt.h
> 
> diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> index 084e82f..b19442c 100644
> --- a/drivers/gpu/drm/i915/i915_drv.h
> +++ b/drivers/gpu/drm/i915/i915_drv.h
> @@ -44,6 +44,8 @@
>  #include <linux/kref.h>
>  #include <linux/pm_qos.h>
>  
> +#include "i915_gem_gtt.h"
> +
>  /* General customization:
>   */
>  
> @@ -572,166 +574,6 @@ enum i915_cache_level {
>  	I915_CACHE_WT, /* hsw:gt3e WriteThrough for scanouts */
>  };
>  
> -typedef uint32_t gen6_gtt_pte_t;
> -
> -/**
> - * A VMA represents a GEM BO that is bound into an address space. Therefore, a
> - * VMA's presence cannot be guaranteed before binding, or after unbinding the
> - * object into/from the address space.
> - *
> - * To make things as simple as possible (ie. no refcounting), a VMA's lifetime
> - * will always be <= an objects lifetime. So object refcounting should cover us.
> - */
> -struct i915_vma {
> -	struct drm_mm_node node;
> -	struct drm_i915_gem_object *obj;
> -	struct i915_address_space *vm;
> -
> -	/** This object's place on the active/inactive lists */
> -	struct list_head mm_list;
> -
> -	struct list_head vma_link; /* Link in the object's VMA list */
> -
> -	/** This vma's place in the batchbuffer or on the eviction list */
> -	struct list_head exec_list;
> -
> -	/**
> -	 * Used for performing relocations during execbuffer insertion.
> -	 */
> -	struct hlist_node exec_node;
> -	unsigned long exec_handle;
> -	struct drm_i915_gem_exec_object2 *exec_entry;
> -
> -	/**
> -	 * How many users have pinned this object in GTT space. The following
> -	 * users can each hold at most one reference: pwrite/pread, pin_ioctl
> -	 * (via user_pin_count), execbuffer (objects are not allowed multiple
> -	 * times for the same batchbuffer), and the framebuffer code. When
> -	 * switching/pageflipping, the framebuffer code has at most two buffers
> -	 * pinned per crtc.
> -	 *
> -	 * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3
> -	 * bits with absolutely no headroom. So use 4 bits. */
> -	unsigned int pin_count:4;
> -#define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf
> -
> -	/** Unmap an object from an address space. This usually consists of
> -	 * setting the valid PTE entries to a reserved scratch page. */
> -	void (*unbind_vma)(struct i915_vma *vma);
> -	/* Map an object into an address space with the given cache flags. */
> -#define GLOBAL_BIND (1<<0)
> -	void (*bind_vma)(struct i915_vma *vma,
> -			 enum i915_cache_level cache_level,
> -			 u32 flags);
> -};
> -
> -struct i915_address_space {
> -	struct drm_mm mm;
> -	struct drm_device *dev;
> -	struct list_head global_link;
> -	unsigned long start;		/* Start offset always 0 for dri2 */
> -	size_t total;		/* size addr space maps (ex. 2GB for ggtt) */
> -
> -	struct {
> -		dma_addr_t addr;
> -		struct page *page;
> -	} scratch;
> -
> -	/**
> -	 * List of objects currently involved in rendering.
> -	 *
> -	 * Includes buffers having the contents of their GPU caches
> -	 * flushed, not necessarily primitives.  last_rendering_seqno
> -	 * represents when the rendering involved will be completed.
> -	 *
> -	 * A reference is held on the buffer while on this list.
> -	 */
> -	struct list_head active_list;
> -
> -	/**
> -	 * LRU list of objects which are not in the ringbuffer and
> -	 * are ready to unbind, but are still in the GTT.
> -	 *
> -	 * last_rendering_seqno is 0 while an object is in this list.
> -	 *
> -	 * A reference is not held on the buffer while on this list,
> -	 * as merely being GTT-bound shouldn't prevent its being
> -	 * freed, and we'll pull it off the list in the free path.
> -	 */
> -	struct list_head inactive_list;
> -
> -	/* FIXME: Need a more generic return type */
> -	gen6_gtt_pte_t (*pte_encode)(dma_addr_t addr,
> -				     enum i915_cache_level level,
> -				     bool valid); /* Create a valid PTE */
> -	void (*clear_range)(struct i915_address_space *vm,
> -			    uint64_t start,
> -			    uint64_t length,
> -			    bool use_scratch);
> -	void (*insert_entries)(struct i915_address_space *vm,
> -			       struct sg_table *st,
> -			       uint64_t start,
> -			       enum i915_cache_level cache_level);
> -	void (*cleanup)(struct i915_address_space *vm);
> -};
> -
> -/* The Graphics Translation Table is the way in which GEN hardware translates a
> - * Graphics Virtual Address into a Physical Address. In addition to the normal
> - * collateral associated with any va->pa translations GEN hardware also has a
> - * portion of the GTT which can be mapped by the CPU and remain both coherent
> - * and correct (in cases like swizzling). That region is referred to as GMADR in
> - * the spec.
> - */
> -struct i915_gtt {
> -	struct i915_address_space base;
> -	size_t stolen_size;		/* Total size of stolen memory */
> -
> -	unsigned long mappable_end;	/* End offset that we can CPU map */
> -	struct io_mapping *mappable;	/* Mapping to our CPU mappable region */
> -	phys_addr_t mappable_base;	/* PA of our GMADR */
> -
> -	/** "Graphics Stolen Memory" holds the global PTEs */
> -	void __iomem *gsm;
> -
> -	bool do_idle_maps;
> -
> -	int mtrr;
> -
> -	/* global gtt ops */
> -	int (*gtt_probe)(struct drm_device *dev, size_t *gtt_total,
> -			  size_t *stolen, phys_addr_t *mappable_base,
> -			  unsigned long *mappable_end);
> -};
> -#define gtt_total_entries(gtt) ((gtt).base.total >> PAGE_SHIFT)
> -
> -#define GEN8_LEGACY_PDPS 4
> -struct i915_hw_ppgtt {
> -	struct i915_address_space base;
> -	struct kref ref;
> -	struct drm_mm_node node;
> -	unsigned num_pd_entries;
> -	unsigned num_pd_pages; /* gen8+ */
> -	union {
> -		struct page **pt_pages;
> -		struct page **gen8_pt_pages[GEN8_LEGACY_PDPS];
> -	};
> -	struct page *pd_pages;
> -	union {
> -		uint32_t pd_offset;
> -		dma_addr_t pd_dma_addr[GEN8_LEGACY_PDPS];
> -	};
> -	union {
> -		dma_addr_t *pt_dma_addr;
> -		dma_addr_t *gen8_pt_dma_addr[GEN8_LEGACY_PDPS];
> -	};
> -
> -	int (*enable)(struct i915_hw_ppgtt *ppgtt);
> -	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
> -			 struct intel_ring_buffer *ring,
> -			 bool synchronous);
> -	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
> -};
> -
>  struct i915_ctx_hang_stats {
>  	/* This context had batch pending when hang was declared */
>  	unsigned batch_pending;
> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
> index 5f73284..a239196 100644
> --- a/drivers/gpu/drm/i915/i915_gem_gtt.c
> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
> @@ -53,60 +53,6 @@ bool intel_enable_ppgtt(struct drm_device *dev, bool full)
>  		return HAS_ALIASING_PPGTT(dev);
>  }
>  
> -#define GEN6_PPGTT_PD_ENTRIES 512
> -#define I915_PPGTT_PT_ENTRIES (PAGE_SIZE / sizeof(gen6_gtt_pte_t))
> -typedef uint64_t gen8_gtt_pte_t;
> -typedef gen8_gtt_pte_t gen8_ppgtt_pde_t;
> -
> -/* PPGTT stuff */
> -#define GEN6_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0xff0))
> -#define HSW_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0x7f0))
> -
> -#define GEN6_PDE_VALID			(1 << 0)
> -/* gen6+ has bit 11-4 for physical addr bit 39-32 */
> -#define GEN6_PDE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> -
> -#define GEN6_PTE_VALID			(1 << 0)
> -#define GEN6_PTE_UNCACHED		(1 << 1)
> -#define HSW_PTE_UNCACHED		(0)
> -#define GEN6_PTE_CACHE_LLC		(2 << 1)
> -#define GEN7_PTE_CACHE_L3_LLC		(3 << 1)
> -#define GEN6_PTE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> -#define HSW_PTE_ADDR_ENCODE(addr)	HSW_GTT_ADDR_ENCODE(addr)
> -
> -/* Cacheability Control is a 4-bit value. The low three bits are stored in *
> - * bits 3:1 of the PTE, while the fourth bit is stored in bit 11 of the PTE.
> - */
> -#define HSW_CACHEABILITY_CONTROL(bits)	((((bits) & 0x7) << 1) | \
> -					 (((bits) & 0x8) << (11 - 3)))
> -#define HSW_WB_LLC_AGE3			HSW_CACHEABILITY_CONTROL(0x2)
> -#define HSW_WB_LLC_AGE0			HSW_CACHEABILITY_CONTROL(0x3)
> -#define HSW_WB_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0xb)
> -#define HSW_WB_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x8)
> -#define HSW_WT_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0x6)
> -#define HSW_WT_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x7)
> -
> -#define GEN8_PTES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_gtt_pte_t))
> -#define GEN8_PDES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_ppgtt_pde_t))
> -
> -/* GEN8 legacy style addressis defined as a 3 level page table:
> - * 31:30 | 29:21 | 20:12 |  11:0
> - * PDPE  |  PDE  |  PTE  | offset
> - * The difference as compared to normal x86 3 level page table is the PDPEs are
> - * programmed via register.
> - */
> -#define GEN8_PDPE_SHIFT			30
> -#define GEN8_PDPE_MASK			0x3
> -#define GEN8_PDE_SHIFT			21
> -#define GEN8_PDE_MASK			0x1ff
> -#define GEN8_PTE_SHIFT			12
> -#define GEN8_PTE_MASK			0x1ff
> -
> -#define PPAT_UNCACHED_INDEX		(_PAGE_PWT | _PAGE_PCD)
> -#define PPAT_CACHED_PDE_INDEX		0 /* WB LLC */
> -#define PPAT_CACHED_INDEX		_PAGE_PAT /* WB LLCeLLC */
> -#define PPAT_DISPLAY_ELLC_INDEX		_PAGE_PCD /* WT eLLC */
> -
>  static void ppgtt_bind_vma(struct i915_vma *vma,
>  			   enum i915_cache_level cache_level,
>  			   u32 flags);
> @@ -185,9 +131,6 @@ static gen6_gtt_pte_t ivb_pte_encode(dma_addr_t addr,
>  	return pte;
>  }
>  
> -#define BYT_PTE_WRITEABLE		(1 << 1)
> -#define BYT_PTE_SNOOPED_BY_CPU_CACHES	(1 << 2)
> -
>  static gen6_gtt_pte_t byt_pte_encode(dma_addr_t addr,
>  				     enum i915_cache_level level,
>  				     bool valid)
> diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
> new file mode 100644
> index 0000000..c8d5c77
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
> @@ -0,0 +1,225 @@
> +#ifndef _I915_GEM_GTT_H
> +#define _I915_GEM_GTT_H
> +
> +#define GEN6_PPGTT_PD_ENTRIES 512
> +#define I915_PPGTT_PT_ENTRIES (PAGE_SIZE / sizeof(gen6_gtt_pte_t))
> +typedef uint32_t gen6_gtt_pte_t;
> +typedef uint64_t gen8_gtt_pte_t;
> +typedef gen8_gtt_pte_t gen8_ppgtt_pde_t;
> +
> +/* PPGTT stuff */
> +#define GEN6_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0xff0))
> +#define HSW_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0x7f0))
> +
> +#define GEN6_PDE_VALID			(1 << 0)
> +/* gen6+ has bit 11-4 for physical addr bit 39-32 */
> +#define GEN6_PDE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> +
> +#define GEN6_PTE_VALID			(1 << 0)
> +#define GEN6_PTE_UNCACHED		(1 << 1)
> +#define HSW_PTE_UNCACHED		(0)
> +#define GEN6_PTE_CACHE_LLC		(2 << 1)
> +#define GEN7_PTE_CACHE_L3_LLC		(3 << 1)
> +#define GEN6_PTE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> +#define HSW_PTE_ADDR_ENCODE(addr)	HSW_GTT_ADDR_ENCODE(addr)
> +
> +#define BYT_PTE_WRITEABLE		(1 << 1)
> +#define BYT_PTE_SNOOPED_BY_CPU_CACHES	(1 << 2)
> +
> +/* Cacheability Control is a 4-bit value. The low three bits are stored in *
> + * bits 3:1 of the PTE, while the fourth bit is stored in bit 11 of the PTE.
> + */
> +#define HSW_CACHEABILITY_CONTROL(bits)	((((bits) & 0x7) << 1) | \
> +					 (((bits) & 0x8) << (11 - 3)))
> +#define HSW_WB_LLC_AGE3			HSW_CACHEABILITY_CONTROL(0x2)
> +#define HSW_WB_LLC_AGE0			HSW_CACHEABILITY_CONTROL(0x3)
> +#define HSW_WB_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0xb)
> +#define HSW_WB_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x8)
> +#define HSW_WT_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0x6)
> +#define HSW_WT_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x7)
> +
> +#define PPAT_UNCACHED_INDEX		(_PAGE_PWT | _PAGE_PCD)
> +#define PPAT_CACHED_PDE_INDEX		0 /* WB LLC */
> +#define PPAT_CACHED_INDEX		_PAGE_PAT /* WB LLCeLLC */
> +#define PPAT_DISPLAY_ELLC_INDEX		_PAGE_PCD /* WT eLLC */
> +
> +#define GEN8_LEGACY_PDPS		4
> +#define GEN8_PTES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_gtt_pte_t))
> +#define GEN8_PDES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_ppgtt_pde_t))
> +
> +/* GEN8 legacy style addressis defined as a 3 level page table:
> + * 31:30 | 29:21 | 20:12 |  11:0
> + * PDPE  |  PDE  |  PTE  | offset
> + * The difference as compared to normal x86 3 level page table is the PDPEs are
> + * programmed via register.
> + *
> + * The x86 pagetable code is flexible in its ability to handle varying page
> + * table depths via abstracted PGDIR/PUD/PMD/PTE. I've opted to not do this and
> + * instead replicate the interesting functionality.
> + */
> +#define GEN8_PDPE_SHIFT			30
> +#define GEN8_PDPE_MASK			0x3
> +#define GEN8_PDE_SHIFT			21
> +#define GEN8_PDE_MASK			0x1ff
> +#define GEN8_PTE_SHIFT			12
> +#define GEN8_PTE_MASK			0x1ff
> +
> +enum i915_cache_level;
> +/**
> + * A VMA represents a GEM BO that is bound into an address space. Therefore, a
> + * VMA's presence cannot be guaranteed before binding, or after unbinding the
> + * object into/from the address space.
> + *
> + * To make things as simple as possible (ie. no refcounting), a VMA's lifetime
> + * will always be <= an objects lifetime. So object refcounting should cover us.
> + */
> +struct i915_vma {
> +	struct drm_mm_node node;
> +	struct drm_i915_gem_object *obj;
> +	struct i915_address_space *vm;
> +
> +	/** This object's place on the active/inactive lists */
> +	struct list_head mm_list;
> +
> +	struct list_head vma_link; /* Link in the object's VMA list */
> +
> +	/** This vma's place in the batchbuffer or on the eviction list */
> +	struct list_head exec_list;
> +
> +	/**
> +	 * Used for performing relocations during execbuffer insertion.
> +	 */
> +	struct hlist_node exec_node;
> +	unsigned long exec_handle;
> +	struct drm_i915_gem_exec_object2 *exec_entry;
> +
> +	/**
> +	 * How many users have pinned this object in GTT space. The following
> +	 * users can each hold at most one reference: pwrite/pread, pin_ioctl
> +	 * (via user_pin_count), execbuffer (objects are not allowed multiple
> +	 * times for the same batchbuffer), and the framebuffer code. When
> +	 * switching/pageflipping, the framebuffer code has at most two buffers
> +	 * pinned per crtc.
> +	 *
> +	 * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3
> +	 * bits with absolutely no headroom. So use 4 bits. */
> +	unsigned int pin_count:4;
> +#define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf
> +
> +	/** Unmap an object from an address space. This usually consists of
> +	 * setting the valid PTE entries to a reserved scratch page. */
> +	void (*unbind_vma)(struct i915_vma *vma);
> +	/* Map an object into an address space with the given cache flags. */
> +#define GLOBAL_BIND (1<<0)
> +	void (*bind_vma)(struct i915_vma *vma,
> +			 enum i915_cache_level cache_level,
> +			 u32 flags);
> +};
> +
> +struct i915_address_space {
> +	struct drm_mm mm;
> +	struct drm_device *dev;
> +	struct list_head global_link;
> +	unsigned long start;		/* Start offset always 0 for dri2 */
> +	size_t total;		/* size addr space maps (ex. 2GB for ggtt) */
> +
> +	struct {
> +		dma_addr_t addr;
> +		struct page *page;
> +	} scratch;
> +
> +	/**
> +	 * List of objects currently involved in rendering.
> +	 *
> +	 * Includes buffers having the contents of their GPU caches
> +	 * flushed, not necessarily primitives.  last_rendering_seqno
> +	 * represents when the rendering involved will be completed.
> +	 *
> +	 * A reference is held on the buffer while on this list.
> +	 */
> +	struct list_head active_list;
> +
> +	/**
> +	 * LRU list of objects which are not in the ringbuffer and
> +	 * are ready to unbind, but are still in the GTT.
> +	 *
> +	 * last_rendering_seqno is 0 while an object is in this list.
> +	 *
> +	 * A reference is not held on the buffer while on this list,
> +	 * as merely being GTT-bound shouldn't prevent its being
> +	 * freed, and we'll pull it off the list in the free path.
> +	 */
> +	struct list_head inactive_list;
> +
> +	/* FIXME: Need a more generic return type */
> +	gen6_gtt_pte_t (*pte_encode)(dma_addr_t addr,
> +				     enum i915_cache_level level,
> +				     bool valid); /* Create a valid PTE */
> +	void (*clear_range)(struct i915_address_space *vm,
> +			    uint64_t start,
> +			    uint64_t length,
> +			    bool use_scratch);
> +	void (*insert_entries)(struct i915_address_space *vm,
> +			       struct sg_table *st,
> +			       uint64_t start,
> +			       enum i915_cache_level cache_level);
> +	void (*cleanup)(struct i915_address_space *vm);
> +};
> +
> +/* The Graphics Translation Table is the way in which GEN hardware translates a
> + * Graphics Virtual Address into a Physical Address. In addition to the normal
> + * collateral associated with any va->pa translations GEN hardware also has a
> + * portion of the GTT which can be mapped by the CPU and remain both coherent
> + * and correct (in cases like swizzling). That region is referred to as GMADR in
> + * the spec.
> + */
> +struct i915_gtt {
> +	struct i915_address_space base;
> +	size_t stolen_size;		/* Total size of stolen memory */
> +
> +	unsigned long mappable_end;	/* End offset that we can CPU map */
> +	struct io_mapping *mappable;	/* Mapping to our CPU mappable region */
> +	phys_addr_t mappable_base;	/* PA of our GMADR */
> +
> +	/** "Graphics Stolen Memory" holds the global PTEs */
> +	void __iomem *gsm;
> +
> +	bool do_idle_maps;
> +
> +	int mtrr;
> +
> +	/* global gtt ops */
> +	int (*gtt_probe)(struct drm_device *dev, size_t *gtt_total,
> +			  size_t *stolen, phys_addr_t *mappable_base,
> +			  unsigned long *mappable_end);
> +};
> +#define gtt_total_entries(gtt) ((gtt).base.total >> PAGE_SHIFT)
> +
> +struct i915_hw_ppgtt {
> +	struct i915_address_space base;
> +	struct kref ref;
> +	struct drm_mm_node node;
> +	unsigned num_pd_entries;
> +	unsigned num_pd_pages; /* gen8+ */
> +	union {
> +		struct page **pt_pages;
> +		struct page **gen8_pt_pages[GEN8_LEGACY_PDPS];
> +	};
> +	struct page *pd_pages;
> +	union {
> +		uint32_t pd_offset;
> +		dma_addr_t pd_dma_addr[GEN8_LEGACY_PDPS];
> +	};
> +	union {
> +		dma_addr_t *pt_dma_addr;
> +		dma_addr_t *gen8_pt_dma_addr[GEN8_LEGACY_PDPS];
> +	};
> +
> +	int (*enable)(struct i915_hw_ppgtt *ppgtt);
> +	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
> +			 struct intel_ring_buffer *ring,
> +			 bool synchronous);
> +	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
> +};
> +
> +#endif
> -- 
> 1.9.0
> 
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@lists.freedesktop.org
> http://lists.freedesktop.org/mailman/listinfo/intel-gfx
Ben Widawsky March 22, 2014, 7:44 p.m. UTC | #3
On Tue, Mar 18, 2014 at 10:15:56AM +0100, Daniel Vetter wrote:
> On Mon, Mar 17, 2014 at 10:48:41PM -0700, Ben Widawsky wrote:
> > TODO: Do header files need a copyright?
> 
> Yup ;-)
> 
> I like this though, especially since finer-grained files will make
> kerneldoc inclusion (well, grouped into sensible chapters at least) much
> simpler.
> -Daniel
> 

If I re-submit just this patch (with the copyright), will you merge it?
It will make my life so much easier.

> > 
> > Signed-off-by: Ben Widawsky <ben@bwidawsk.net>
> > ---
> >  drivers/gpu/drm/i915/i915_drv.h     | 162 +-------------------------
> >  drivers/gpu/drm/i915/i915_gem_gtt.c |  57 ---------
> >  drivers/gpu/drm/i915/i915_gem_gtt.h | 225 ++++++++++++++++++++++++++++++++++++
> >  3 files changed, 227 insertions(+), 217 deletions(-)
> >  create mode 100644 drivers/gpu/drm/i915/i915_gem_gtt.h
> > 
> > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
> > index 084e82f..b19442c 100644
> > --- a/drivers/gpu/drm/i915/i915_drv.h
> > +++ b/drivers/gpu/drm/i915/i915_drv.h
> > @@ -44,6 +44,8 @@
> >  #include <linux/kref.h>
> >  #include <linux/pm_qos.h>
> >  
> > +#include "i915_gem_gtt.h"
> > +
> >  /* General customization:
> >   */
> >  
> > @@ -572,166 +574,6 @@ enum i915_cache_level {
> >  	I915_CACHE_WT, /* hsw:gt3e WriteThrough for scanouts */
> >  };
> >  
> > -typedef uint32_t gen6_gtt_pte_t;
> > -
> > -/**
> > - * A VMA represents a GEM BO that is bound into an address space. Therefore, a
> > - * VMA's presence cannot be guaranteed before binding, or after unbinding the
> > - * object into/from the address space.
> > - *
> > - * To make things as simple as possible (ie. no refcounting), a VMA's lifetime
> > - * will always be <= an objects lifetime. So object refcounting should cover us.
> > - */
> > -struct i915_vma {
> > -	struct drm_mm_node node;
> > -	struct drm_i915_gem_object *obj;
> > -	struct i915_address_space *vm;
> > -
> > -	/** This object's place on the active/inactive lists */
> > -	struct list_head mm_list;
> > -
> > -	struct list_head vma_link; /* Link in the object's VMA list */
> > -
> > -	/** This vma's place in the batchbuffer or on the eviction list */
> > -	struct list_head exec_list;
> > -
> > -	/**
> > -	 * Used for performing relocations during execbuffer insertion.
> > -	 */
> > -	struct hlist_node exec_node;
> > -	unsigned long exec_handle;
> > -	struct drm_i915_gem_exec_object2 *exec_entry;
> > -
> > -	/**
> > -	 * How many users have pinned this object in GTT space. The following
> > -	 * users can each hold at most one reference: pwrite/pread, pin_ioctl
> > -	 * (via user_pin_count), execbuffer (objects are not allowed multiple
> > -	 * times for the same batchbuffer), and the framebuffer code. When
> > -	 * switching/pageflipping, the framebuffer code has at most two buffers
> > -	 * pinned per crtc.
> > -	 *
> > -	 * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3
> > -	 * bits with absolutely no headroom. So use 4 bits. */
> > -	unsigned int pin_count:4;
> > -#define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf
> > -
> > -	/** Unmap an object from an address space. This usually consists of
> > -	 * setting the valid PTE entries to a reserved scratch page. */
> > -	void (*unbind_vma)(struct i915_vma *vma);
> > -	/* Map an object into an address space with the given cache flags. */
> > -#define GLOBAL_BIND (1<<0)
> > -	void (*bind_vma)(struct i915_vma *vma,
> > -			 enum i915_cache_level cache_level,
> > -			 u32 flags);
> > -};
> > -
> > -struct i915_address_space {
> > -	struct drm_mm mm;
> > -	struct drm_device *dev;
> > -	struct list_head global_link;
> > -	unsigned long start;		/* Start offset always 0 for dri2 */
> > -	size_t total;		/* size addr space maps (ex. 2GB for ggtt) */
> > -
> > -	struct {
> > -		dma_addr_t addr;
> > -		struct page *page;
> > -	} scratch;
> > -
> > -	/**
> > -	 * List of objects currently involved in rendering.
> > -	 *
> > -	 * Includes buffers having the contents of their GPU caches
> > -	 * flushed, not necessarily primitives.  last_rendering_seqno
> > -	 * represents when the rendering involved will be completed.
> > -	 *
> > -	 * A reference is held on the buffer while on this list.
> > -	 */
> > -	struct list_head active_list;
> > -
> > -	/**
> > -	 * LRU list of objects which are not in the ringbuffer and
> > -	 * are ready to unbind, but are still in the GTT.
> > -	 *
> > -	 * last_rendering_seqno is 0 while an object is in this list.
> > -	 *
> > -	 * A reference is not held on the buffer while on this list,
> > -	 * as merely being GTT-bound shouldn't prevent its being
> > -	 * freed, and we'll pull it off the list in the free path.
> > -	 */
> > -	struct list_head inactive_list;
> > -
> > -	/* FIXME: Need a more generic return type */
> > -	gen6_gtt_pte_t (*pte_encode)(dma_addr_t addr,
> > -				     enum i915_cache_level level,
> > -				     bool valid); /* Create a valid PTE */
> > -	void (*clear_range)(struct i915_address_space *vm,
> > -			    uint64_t start,
> > -			    uint64_t length,
> > -			    bool use_scratch);
> > -	void (*insert_entries)(struct i915_address_space *vm,
> > -			       struct sg_table *st,
> > -			       uint64_t start,
> > -			       enum i915_cache_level cache_level);
> > -	void (*cleanup)(struct i915_address_space *vm);
> > -};
> > -
> > -/* The Graphics Translation Table is the way in which GEN hardware translates a
> > - * Graphics Virtual Address into a Physical Address. In addition to the normal
> > - * collateral associated with any va->pa translations GEN hardware also has a
> > - * portion of the GTT which can be mapped by the CPU and remain both coherent
> > - * and correct (in cases like swizzling). That region is referred to as GMADR in
> > - * the spec.
> > - */
> > -struct i915_gtt {
> > -	struct i915_address_space base;
> > -	size_t stolen_size;		/* Total size of stolen memory */
> > -
> > -	unsigned long mappable_end;	/* End offset that we can CPU map */
> > -	struct io_mapping *mappable;	/* Mapping to our CPU mappable region */
> > -	phys_addr_t mappable_base;	/* PA of our GMADR */
> > -
> > -	/** "Graphics Stolen Memory" holds the global PTEs */
> > -	void __iomem *gsm;
> > -
> > -	bool do_idle_maps;
> > -
> > -	int mtrr;
> > -
> > -	/* global gtt ops */
> > -	int (*gtt_probe)(struct drm_device *dev, size_t *gtt_total,
> > -			  size_t *stolen, phys_addr_t *mappable_base,
> > -			  unsigned long *mappable_end);
> > -};
> > -#define gtt_total_entries(gtt) ((gtt).base.total >> PAGE_SHIFT)
> > -
> > -#define GEN8_LEGACY_PDPS 4
> > -struct i915_hw_ppgtt {
> > -	struct i915_address_space base;
> > -	struct kref ref;
> > -	struct drm_mm_node node;
> > -	unsigned num_pd_entries;
> > -	unsigned num_pd_pages; /* gen8+ */
> > -	union {
> > -		struct page **pt_pages;
> > -		struct page **gen8_pt_pages[GEN8_LEGACY_PDPS];
> > -	};
> > -	struct page *pd_pages;
> > -	union {
> > -		uint32_t pd_offset;
> > -		dma_addr_t pd_dma_addr[GEN8_LEGACY_PDPS];
> > -	};
> > -	union {
> > -		dma_addr_t *pt_dma_addr;
> > -		dma_addr_t *gen8_pt_dma_addr[GEN8_LEGACY_PDPS];
> > -	};
> > -
> > -	int (*enable)(struct i915_hw_ppgtt *ppgtt);
> > -	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
> > -			 struct intel_ring_buffer *ring,
> > -			 bool synchronous);
> > -	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
> > -};
> > -
> >  struct i915_ctx_hang_stats {
> >  	/* This context had batch pending when hang was declared */
> >  	unsigned batch_pending;
> > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
> > index 5f73284..a239196 100644
> > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c
> > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
> > @@ -53,60 +53,6 @@ bool intel_enable_ppgtt(struct drm_device *dev, bool full)
> >  		return HAS_ALIASING_PPGTT(dev);
> >  }
> >  
> > -#define GEN6_PPGTT_PD_ENTRIES 512
> > -#define I915_PPGTT_PT_ENTRIES (PAGE_SIZE / sizeof(gen6_gtt_pte_t))
> > -typedef uint64_t gen8_gtt_pte_t;
> > -typedef gen8_gtt_pte_t gen8_ppgtt_pde_t;
> > -
> > -/* PPGTT stuff */
> > -#define GEN6_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0xff0))
> > -#define HSW_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0x7f0))
> > -
> > -#define GEN6_PDE_VALID			(1 << 0)
> > -/* gen6+ has bit 11-4 for physical addr bit 39-32 */
> > -#define GEN6_PDE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> > -
> > -#define GEN6_PTE_VALID			(1 << 0)
> > -#define GEN6_PTE_UNCACHED		(1 << 1)
> > -#define HSW_PTE_UNCACHED		(0)
> > -#define GEN6_PTE_CACHE_LLC		(2 << 1)
> > -#define GEN7_PTE_CACHE_L3_LLC		(3 << 1)
> > -#define GEN6_PTE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> > -#define HSW_PTE_ADDR_ENCODE(addr)	HSW_GTT_ADDR_ENCODE(addr)
> > -
> > -/* Cacheability Control is a 4-bit value. The low three bits are stored in *
> > - * bits 3:1 of the PTE, while the fourth bit is stored in bit 11 of the PTE.
> > - */
> > -#define HSW_CACHEABILITY_CONTROL(bits)	((((bits) & 0x7) << 1) | \
> > -					 (((bits) & 0x8) << (11 - 3)))
> > -#define HSW_WB_LLC_AGE3			HSW_CACHEABILITY_CONTROL(0x2)
> > -#define HSW_WB_LLC_AGE0			HSW_CACHEABILITY_CONTROL(0x3)
> > -#define HSW_WB_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0xb)
> > -#define HSW_WB_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x8)
> > -#define HSW_WT_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0x6)
> > -#define HSW_WT_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x7)
> > -
> > -#define GEN8_PTES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_gtt_pte_t))
> > -#define GEN8_PDES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_ppgtt_pde_t))
> > -
> > -/* GEN8 legacy style addressis defined as a 3 level page table:
> > - * 31:30 | 29:21 | 20:12 |  11:0
> > - * PDPE  |  PDE  |  PTE  | offset
> > - * The difference as compared to normal x86 3 level page table is the PDPEs are
> > - * programmed via register.
> > - */
> > -#define GEN8_PDPE_SHIFT			30
> > -#define GEN8_PDPE_MASK			0x3
> > -#define GEN8_PDE_SHIFT			21
> > -#define GEN8_PDE_MASK			0x1ff
> > -#define GEN8_PTE_SHIFT			12
> > -#define GEN8_PTE_MASK			0x1ff
> > -
> > -#define PPAT_UNCACHED_INDEX		(_PAGE_PWT | _PAGE_PCD)
> > -#define PPAT_CACHED_PDE_INDEX		0 /* WB LLC */
> > -#define PPAT_CACHED_INDEX		_PAGE_PAT /* WB LLCeLLC */
> > -#define PPAT_DISPLAY_ELLC_INDEX		_PAGE_PCD /* WT eLLC */
> > -
> >  static void ppgtt_bind_vma(struct i915_vma *vma,
> >  			   enum i915_cache_level cache_level,
> >  			   u32 flags);
> > @@ -185,9 +131,6 @@ static gen6_gtt_pte_t ivb_pte_encode(dma_addr_t addr,
> >  	return pte;
> >  }
> >  
> > -#define BYT_PTE_WRITEABLE		(1 << 1)
> > -#define BYT_PTE_SNOOPED_BY_CPU_CACHES	(1 << 2)
> > -
> >  static gen6_gtt_pte_t byt_pte_encode(dma_addr_t addr,
> >  				     enum i915_cache_level level,
> >  				     bool valid)
> > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
> > new file mode 100644
> > index 0000000..c8d5c77
> > --- /dev/null
> > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
> > @@ -0,0 +1,225 @@
> > +#ifndef _I915_GEM_GTT_H
> > +#define _I915_GEM_GTT_H
> > +
> > +#define GEN6_PPGTT_PD_ENTRIES 512
> > +#define I915_PPGTT_PT_ENTRIES (PAGE_SIZE / sizeof(gen6_gtt_pte_t))
> > +typedef uint32_t gen6_gtt_pte_t;
> > +typedef uint64_t gen8_gtt_pte_t;
> > +typedef gen8_gtt_pte_t gen8_ppgtt_pde_t;
> > +
> > +/* PPGTT stuff */
> > +#define GEN6_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0xff0))
> > +#define HSW_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0x7f0))
> > +
> > +#define GEN6_PDE_VALID			(1 << 0)
> > +/* gen6+ has bit 11-4 for physical addr bit 39-32 */
> > +#define GEN6_PDE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> > +
> > +#define GEN6_PTE_VALID			(1 << 0)
> > +#define GEN6_PTE_UNCACHED		(1 << 1)
> > +#define HSW_PTE_UNCACHED		(0)
> > +#define GEN6_PTE_CACHE_LLC		(2 << 1)
> > +#define GEN7_PTE_CACHE_L3_LLC		(3 << 1)
> > +#define GEN6_PTE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
> > +#define HSW_PTE_ADDR_ENCODE(addr)	HSW_GTT_ADDR_ENCODE(addr)
> > +
> > +#define BYT_PTE_WRITEABLE		(1 << 1)
> > +#define BYT_PTE_SNOOPED_BY_CPU_CACHES	(1 << 2)
> > +
> > +/* Cacheability Control is a 4-bit value. The low three bits are stored in *
> > + * bits 3:1 of the PTE, while the fourth bit is stored in bit 11 of the PTE.
> > + */
> > +#define HSW_CACHEABILITY_CONTROL(bits)	((((bits) & 0x7) << 1) | \
> > +					 (((bits) & 0x8) << (11 - 3)))
> > +#define HSW_WB_LLC_AGE3			HSW_CACHEABILITY_CONTROL(0x2)
> > +#define HSW_WB_LLC_AGE0			HSW_CACHEABILITY_CONTROL(0x3)
> > +#define HSW_WB_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0xb)
> > +#define HSW_WB_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x8)
> > +#define HSW_WT_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0x6)
> > +#define HSW_WT_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x7)
> > +
> > +#define PPAT_UNCACHED_INDEX		(_PAGE_PWT | _PAGE_PCD)
> > +#define PPAT_CACHED_PDE_INDEX		0 /* WB LLC */
> > +#define PPAT_CACHED_INDEX		_PAGE_PAT /* WB LLCeLLC */
> > +#define PPAT_DISPLAY_ELLC_INDEX		_PAGE_PCD /* WT eLLC */
> > +
> > +#define GEN8_LEGACY_PDPS		4
> > +#define GEN8_PTES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_gtt_pte_t))
> > +#define GEN8_PDES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_ppgtt_pde_t))
> > +
> > +/* GEN8 legacy style addressis defined as a 3 level page table:
> > + * 31:30 | 29:21 | 20:12 |  11:0
> > + * PDPE  |  PDE  |  PTE  | offset
> > + * The difference as compared to normal x86 3 level page table is the PDPEs are
> > + * programmed via register.
> > + *
> > + * The x86 pagetable code is flexible in its ability to handle varying page
> > + * table depths via abstracted PGDIR/PUD/PMD/PTE. I've opted to not do this and
> > + * instead replicate the interesting functionality.
> > + */
> > +#define GEN8_PDPE_SHIFT			30
> > +#define GEN8_PDPE_MASK			0x3
> > +#define GEN8_PDE_SHIFT			21
> > +#define GEN8_PDE_MASK			0x1ff
> > +#define GEN8_PTE_SHIFT			12
> > +#define GEN8_PTE_MASK			0x1ff
> > +
> > +enum i915_cache_level;
> > +/**
> > + * A VMA represents a GEM BO that is bound into an address space. Therefore, a
> > + * VMA's presence cannot be guaranteed before binding, or after unbinding the
> > + * object into/from the address space.
> > + *
> > + * To make things as simple as possible (ie. no refcounting), a VMA's lifetime
> > + * will always be <= an objects lifetime. So object refcounting should cover us.
> > + */
> > +struct i915_vma {
> > +	struct drm_mm_node node;
> > +	struct drm_i915_gem_object *obj;
> > +	struct i915_address_space *vm;
> > +
> > +	/** This object's place on the active/inactive lists */
> > +	struct list_head mm_list;
> > +
> > +	struct list_head vma_link; /* Link in the object's VMA list */
> > +
> > +	/** This vma's place in the batchbuffer or on the eviction list */
> > +	struct list_head exec_list;
> > +
> > +	/**
> > +	 * Used for performing relocations during execbuffer insertion.
> > +	 */
> > +	struct hlist_node exec_node;
> > +	unsigned long exec_handle;
> > +	struct drm_i915_gem_exec_object2 *exec_entry;
> > +
> > +	/**
> > +	 * How many users have pinned this object in GTT space. The following
> > +	 * users can each hold at most one reference: pwrite/pread, pin_ioctl
> > +	 * (via user_pin_count), execbuffer (objects are not allowed multiple
> > +	 * times for the same batchbuffer), and the framebuffer code. When
> > +	 * switching/pageflipping, the framebuffer code has at most two buffers
> > +	 * pinned per crtc.
> > +	 *
> > +	 * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3
> > +	 * bits with absolutely no headroom. So use 4 bits. */
> > +	unsigned int pin_count:4;
> > +#define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf
> > +
> > +	/** Unmap an object from an address space. This usually consists of
> > +	 * setting the valid PTE entries to a reserved scratch page. */
> > +	void (*unbind_vma)(struct i915_vma *vma);
> > +	/* Map an object into an address space with the given cache flags. */
> > +#define GLOBAL_BIND (1<<0)
> > +	void (*bind_vma)(struct i915_vma *vma,
> > +			 enum i915_cache_level cache_level,
> > +			 u32 flags);
> > +};
> > +
> > +struct i915_address_space {
> > +	struct drm_mm mm;
> > +	struct drm_device *dev;
> > +	struct list_head global_link;
> > +	unsigned long start;		/* Start offset always 0 for dri2 */
> > +	size_t total;		/* size addr space maps (ex. 2GB for ggtt) */
> > +
> > +	struct {
> > +		dma_addr_t addr;
> > +		struct page *page;
> > +	} scratch;
> > +
> > +	/**
> > +	 * List of objects currently involved in rendering.
> > +	 *
> > +	 * Includes buffers having the contents of their GPU caches
> > +	 * flushed, not necessarily primitives.  last_rendering_seqno
> > +	 * represents when the rendering involved will be completed.
> > +	 *
> > +	 * A reference is held on the buffer while on this list.
> > +	 */
> > +	struct list_head active_list;
> > +
> > +	/**
> > +	 * LRU list of objects which are not in the ringbuffer and
> > +	 * are ready to unbind, but are still in the GTT.
> > +	 *
> > +	 * last_rendering_seqno is 0 while an object is in this list.
> > +	 *
> > +	 * A reference is not held on the buffer while on this list,
> > +	 * as merely being GTT-bound shouldn't prevent its being
> > +	 * freed, and we'll pull it off the list in the free path.
> > +	 */
> > +	struct list_head inactive_list;
> > +
> > +	/* FIXME: Need a more generic return type */
> > +	gen6_gtt_pte_t (*pte_encode)(dma_addr_t addr,
> > +				     enum i915_cache_level level,
> > +				     bool valid); /* Create a valid PTE */
> > +	void (*clear_range)(struct i915_address_space *vm,
> > +			    uint64_t start,
> > +			    uint64_t length,
> > +			    bool use_scratch);
> > +	void (*insert_entries)(struct i915_address_space *vm,
> > +			       struct sg_table *st,
> > +			       uint64_t start,
> > +			       enum i915_cache_level cache_level);
> > +	void (*cleanup)(struct i915_address_space *vm);
> > +};
> > +
> > +/* The Graphics Translation Table is the way in which GEN hardware translates a
> > + * Graphics Virtual Address into a Physical Address. In addition to the normal
> > + * collateral associated with any va->pa translations GEN hardware also has a
> > + * portion of the GTT which can be mapped by the CPU and remain both coherent
> > + * and correct (in cases like swizzling). That region is referred to as GMADR in
> > + * the spec.
> > + */
> > +struct i915_gtt {
> > +	struct i915_address_space base;
> > +	size_t stolen_size;		/* Total size of stolen memory */
> > +
> > +	unsigned long mappable_end;	/* End offset that we can CPU map */
> > +	struct io_mapping *mappable;	/* Mapping to our CPU mappable region */
> > +	phys_addr_t mappable_base;	/* PA of our GMADR */
> > +
> > +	/** "Graphics Stolen Memory" holds the global PTEs */
> > +	void __iomem *gsm;
> > +
> > +	bool do_idle_maps;
> > +
> > +	int mtrr;
> > +
> > +	/* global gtt ops */
> > +	int (*gtt_probe)(struct drm_device *dev, size_t *gtt_total,
> > +			  size_t *stolen, phys_addr_t *mappable_base,
> > +			  unsigned long *mappable_end);
> > +};
> > +#define gtt_total_entries(gtt) ((gtt).base.total >> PAGE_SHIFT)
> > +
> > +struct i915_hw_ppgtt {
> > +	struct i915_address_space base;
> > +	struct kref ref;
> > +	struct drm_mm_node node;
> > +	unsigned num_pd_entries;
> > +	unsigned num_pd_pages; /* gen8+ */
> > +	union {
> > +		struct page **pt_pages;
> > +		struct page **gen8_pt_pages[GEN8_LEGACY_PDPS];
> > +	};
> > +	struct page *pd_pages;
> > +	union {
> > +		uint32_t pd_offset;
> > +		dma_addr_t pd_dma_addr[GEN8_LEGACY_PDPS];
> > +	};
> > +	union {
> > +		dma_addr_t *pt_dma_addr;
> > +		dma_addr_t *gen8_pt_dma_addr[GEN8_LEGACY_PDPS];
> > +	};
> > +
> > +	int (*enable)(struct i915_hw_ppgtt *ppgtt);
> > +	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
> > +			 struct intel_ring_buffer *ring,
> > +			 bool synchronous);
> > +	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
> > +};
> > +
> > +#endif
> > -- 
> > 1.9.0
> > 
> > _______________________________________________
> > Intel-gfx mailing list
> > Intel-gfx@lists.freedesktop.org
> > http://lists.freedesktop.org/mailman/listinfo/intel-gfx
> 
> -- 
> Daniel Vetter
> Software Engineer, Intel Corporation
> +41 (0) 79 365 57 48 - http://blog.ffwll.ch
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@lists.freedesktop.org
> http://lists.freedesktop.org/mailman/listinfo/intel-gfx
Daniel Vetter March 23, 2014, 12:46 a.m. UTC | #4
On Sat, Mar 22, 2014 at 8:44 PM, Ben Widawsky <ben@bwidawsk.net> wrote:
> On Tue, Mar 18, 2014 at 10:15:56AM +0100, Daniel Vetter wrote:
>> On Mon, Mar 17, 2014 at 10:48:41PM -0700, Ben Widawsky wrote:
>> > TODO: Do header files need a copyright?
>>
>> Yup ;-)
>>
>> I like this though, especially since finer-grained files will make
>> kerneldoc inclusion (well, grouped into sensible chapters at least) much
>> simpler.
>> -Daniel
>>
>
> If I re-submit just this patch (with the copyright), will you merge it?
> It will make my life so much easier.

Yeah.
-Daniel
diff mbox

Patch

diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
index 084e82f..b19442c 100644
--- a/drivers/gpu/drm/i915/i915_drv.h
+++ b/drivers/gpu/drm/i915/i915_drv.h
@@ -44,6 +44,8 @@ 
 #include <linux/kref.h>
 #include <linux/pm_qos.h>
 
+#include "i915_gem_gtt.h"
+
 /* General customization:
  */
 
@@ -572,166 +574,6 @@  enum i915_cache_level {
 	I915_CACHE_WT, /* hsw:gt3e WriteThrough for scanouts */
 };
 
-typedef uint32_t gen6_gtt_pte_t;
-
-/**
- * A VMA represents a GEM BO that is bound into an address space. Therefore, a
- * VMA's presence cannot be guaranteed before binding, or after unbinding the
- * object into/from the address space.
- *
- * To make things as simple as possible (ie. no refcounting), a VMA's lifetime
- * will always be <= an objects lifetime. So object refcounting should cover us.
- */
-struct i915_vma {
-	struct drm_mm_node node;
-	struct drm_i915_gem_object *obj;
-	struct i915_address_space *vm;
-
-	/** This object's place on the active/inactive lists */
-	struct list_head mm_list;
-
-	struct list_head vma_link; /* Link in the object's VMA list */
-
-	/** This vma's place in the batchbuffer or on the eviction list */
-	struct list_head exec_list;
-
-	/**
-	 * Used for performing relocations during execbuffer insertion.
-	 */
-	struct hlist_node exec_node;
-	unsigned long exec_handle;
-	struct drm_i915_gem_exec_object2 *exec_entry;
-
-	/**
-	 * How many users have pinned this object in GTT space. The following
-	 * users can each hold at most one reference: pwrite/pread, pin_ioctl
-	 * (via user_pin_count), execbuffer (objects are not allowed multiple
-	 * times for the same batchbuffer), and the framebuffer code. When
-	 * switching/pageflipping, the framebuffer code has at most two buffers
-	 * pinned per crtc.
-	 *
-	 * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3
-	 * bits with absolutely no headroom. So use 4 bits. */
-	unsigned int pin_count:4;
-#define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf
-
-	/** Unmap an object from an address space. This usually consists of
-	 * setting the valid PTE entries to a reserved scratch page. */
-	void (*unbind_vma)(struct i915_vma *vma);
-	/* Map an object into an address space with the given cache flags. */
-#define GLOBAL_BIND (1<<0)
-	void (*bind_vma)(struct i915_vma *vma,
-			 enum i915_cache_level cache_level,
-			 u32 flags);
-};
-
-struct i915_address_space {
-	struct drm_mm mm;
-	struct drm_device *dev;
-	struct list_head global_link;
-	unsigned long start;		/* Start offset always 0 for dri2 */
-	size_t total;		/* size addr space maps (ex. 2GB for ggtt) */
-
-	struct {
-		dma_addr_t addr;
-		struct page *page;
-	} scratch;
-
-	/**
-	 * List of objects currently involved in rendering.
-	 *
-	 * Includes buffers having the contents of their GPU caches
-	 * flushed, not necessarily primitives.  last_rendering_seqno
-	 * represents when the rendering involved will be completed.
-	 *
-	 * A reference is held on the buffer while on this list.
-	 */
-	struct list_head active_list;
-
-	/**
-	 * LRU list of objects which are not in the ringbuffer and
-	 * are ready to unbind, but are still in the GTT.
-	 *
-	 * last_rendering_seqno is 0 while an object is in this list.
-	 *
-	 * A reference is not held on the buffer while on this list,
-	 * as merely being GTT-bound shouldn't prevent its being
-	 * freed, and we'll pull it off the list in the free path.
-	 */
-	struct list_head inactive_list;
-
-	/* FIXME: Need a more generic return type */
-	gen6_gtt_pte_t (*pte_encode)(dma_addr_t addr,
-				     enum i915_cache_level level,
-				     bool valid); /* Create a valid PTE */
-	void (*clear_range)(struct i915_address_space *vm,
-			    uint64_t start,
-			    uint64_t length,
-			    bool use_scratch);
-	void (*insert_entries)(struct i915_address_space *vm,
-			       struct sg_table *st,
-			       uint64_t start,
-			       enum i915_cache_level cache_level);
-	void (*cleanup)(struct i915_address_space *vm);
-};
-
-/* The Graphics Translation Table is the way in which GEN hardware translates a
- * Graphics Virtual Address into a Physical Address. In addition to the normal
- * collateral associated with any va->pa translations GEN hardware also has a
- * portion of the GTT which can be mapped by the CPU and remain both coherent
- * and correct (in cases like swizzling). That region is referred to as GMADR in
- * the spec.
- */
-struct i915_gtt {
-	struct i915_address_space base;
-	size_t stolen_size;		/* Total size of stolen memory */
-
-	unsigned long mappable_end;	/* End offset that we can CPU map */
-	struct io_mapping *mappable;	/* Mapping to our CPU mappable region */
-	phys_addr_t mappable_base;	/* PA of our GMADR */
-
-	/** "Graphics Stolen Memory" holds the global PTEs */
-	void __iomem *gsm;
-
-	bool do_idle_maps;
-
-	int mtrr;
-
-	/* global gtt ops */
-	int (*gtt_probe)(struct drm_device *dev, size_t *gtt_total,
-			  size_t *stolen, phys_addr_t *mappable_base,
-			  unsigned long *mappable_end);
-};
-#define gtt_total_entries(gtt) ((gtt).base.total >> PAGE_SHIFT)
-
-#define GEN8_LEGACY_PDPS 4
-struct i915_hw_ppgtt {
-	struct i915_address_space base;
-	struct kref ref;
-	struct drm_mm_node node;
-	unsigned num_pd_entries;
-	unsigned num_pd_pages; /* gen8+ */
-	union {
-		struct page **pt_pages;
-		struct page **gen8_pt_pages[GEN8_LEGACY_PDPS];
-	};
-	struct page *pd_pages;
-	union {
-		uint32_t pd_offset;
-		dma_addr_t pd_dma_addr[GEN8_LEGACY_PDPS];
-	};
-	union {
-		dma_addr_t *pt_dma_addr;
-		dma_addr_t *gen8_pt_dma_addr[GEN8_LEGACY_PDPS];
-	};
-
-	int (*enable)(struct i915_hw_ppgtt *ppgtt);
-	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
-			 struct intel_ring_buffer *ring,
-			 bool synchronous);
-	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
-};
-
 struct i915_ctx_hang_stats {
 	/* This context had batch pending when hang was declared */
 	unsigned batch_pending;
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
index 5f73284..a239196 100644
--- a/drivers/gpu/drm/i915/i915_gem_gtt.c
+++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
@@ -53,60 +53,6 @@  bool intel_enable_ppgtt(struct drm_device *dev, bool full)
 		return HAS_ALIASING_PPGTT(dev);
 }
 
-#define GEN6_PPGTT_PD_ENTRIES 512
-#define I915_PPGTT_PT_ENTRIES (PAGE_SIZE / sizeof(gen6_gtt_pte_t))
-typedef uint64_t gen8_gtt_pte_t;
-typedef gen8_gtt_pte_t gen8_ppgtt_pde_t;
-
-/* PPGTT stuff */
-#define GEN6_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0xff0))
-#define HSW_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0x7f0))
-
-#define GEN6_PDE_VALID			(1 << 0)
-/* gen6+ has bit 11-4 for physical addr bit 39-32 */
-#define GEN6_PDE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
-
-#define GEN6_PTE_VALID			(1 << 0)
-#define GEN6_PTE_UNCACHED		(1 << 1)
-#define HSW_PTE_UNCACHED		(0)
-#define GEN6_PTE_CACHE_LLC		(2 << 1)
-#define GEN7_PTE_CACHE_L3_LLC		(3 << 1)
-#define GEN6_PTE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
-#define HSW_PTE_ADDR_ENCODE(addr)	HSW_GTT_ADDR_ENCODE(addr)
-
-/* Cacheability Control is a 4-bit value. The low three bits are stored in *
- * bits 3:1 of the PTE, while the fourth bit is stored in bit 11 of the PTE.
- */
-#define HSW_CACHEABILITY_CONTROL(bits)	((((bits) & 0x7) << 1) | \
-					 (((bits) & 0x8) << (11 - 3)))
-#define HSW_WB_LLC_AGE3			HSW_CACHEABILITY_CONTROL(0x2)
-#define HSW_WB_LLC_AGE0			HSW_CACHEABILITY_CONTROL(0x3)
-#define HSW_WB_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0xb)
-#define HSW_WB_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x8)
-#define HSW_WT_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0x6)
-#define HSW_WT_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x7)
-
-#define GEN8_PTES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_gtt_pte_t))
-#define GEN8_PDES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_ppgtt_pde_t))
-
-/* GEN8 legacy style addressis defined as a 3 level page table:
- * 31:30 | 29:21 | 20:12 |  11:0
- * PDPE  |  PDE  |  PTE  | offset
- * The difference as compared to normal x86 3 level page table is the PDPEs are
- * programmed via register.
- */
-#define GEN8_PDPE_SHIFT			30
-#define GEN8_PDPE_MASK			0x3
-#define GEN8_PDE_SHIFT			21
-#define GEN8_PDE_MASK			0x1ff
-#define GEN8_PTE_SHIFT			12
-#define GEN8_PTE_MASK			0x1ff
-
-#define PPAT_UNCACHED_INDEX		(_PAGE_PWT | _PAGE_PCD)
-#define PPAT_CACHED_PDE_INDEX		0 /* WB LLC */
-#define PPAT_CACHED_INDEX		_PAGE_PAT /* WB LLCeLLC */
-#define PPAT_DISPLAY_ELLC_INDEX		_PAGE_PCD /* WT eLLC */
-
 static void ppgtt_bind_vma(struct i915_vma *vma,
 			   enum i915_cache_level cache_level,
 			   u32 flags);
@@ -185,9 +131,6 @@  static gen6_gtt_pte_t ivb_pte_encode(dma_addr_t addr,
 	return pte;
 }
 
-#define BYT_PTE_WRITEABLE		(1 << 1)
-#define BYT_PTE_SNOOPED_BY_CPU_CACHES	(1 << 2)
-
 static gen6_gtt_pte_t byt_pte_encode(dma_addr_t addr,
 				     enum i915_cache_level level,
 				     bool valid)
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
new file mode 100644
index 0000000..c8d5c77
--- /dev/null
+++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
@@ -0,0 +1,225 @@ 
+#ifndef _I915_GEM_GTT_H
+#define _I915_GEM_GTT_H
+
+#define GEN6_PPGTT_PD_ENTRIES 512
+#define I915_PPGTT_PT_ENTRIES (PAGE_SIZE / sizeof(gen6_gtt_pte_t))
+typedef uint32_t gen6_gtt_pte_t;
+typedef uint64_t gen8_gtt_pte_t;
+typedef gen8_gtt_pte_t gen8_ppgtt_pde_t;
+
+/* PPGTT stuff */
+#define GEN6_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0xff0))
+#define HSW_GTT_ADDR_ENCODE(addr)	((addr) | (((addr) >> 28) & 0x7f0))
+
+#define GEN6_PDE_VALID			(1 << 0)
+/* gen6+ has bit 11-4 for physical addr bit 39-32 */
+#define GEN6_PDE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
+
+#define GEN6_PTE_VALID			(1 << 0)
+#define GEN6_PTE_UNCACHED		(1 << 1)
+#define HSW_PTE_UNCACHED		(0)
+#define GEN6_PTE_CACHE_LLC		(2 << 1)
+#define GEN7_PTE_CACHE_L3_LLC		(3 << 1)
+#define GEN6_PTE_ADDR_ENCODE(addr)	GEN6_GTT_ADDR_ENCODE(addr)
+#define HSW_PTE_ADDR_ENCODE(addr)	HSW_GTT_ADDR_ENCODE(addr)
+
+#define BYT_PTE_WRITEABLE		(1 << 1)
+#define BYT_PTE_SNOOPED_BY_CPU_CACHES	(1 << 2)
+
+/* Cacheability Control is a 4-bit value. The low three bits are stored in *
+ * bits 3:1 of the PTE, while the fourth bit is stored in bit 11 of the PTE.
+ */
+#define HSW_CACHEABILITY_CONTROL(bits)	((((bits) & 0x7) << 1) | \
+					 (((bits) & 0x8) << (11 - 3)))
+#define HSW_WB_LLC_AGE3			HSW_CACHEABILITY_CONTROL(0x2)
+#define HSW_WB_LLC_AGE0			HSW_CACHEABILITY_CONTROL(0x3)
+#define HSW_WB_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0xb)
+#define HSW_WB_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x8)
+#define HSW_WT_ELLC_LLC_AGE0		HSW_CACHEABILITY_CONTROL(0x6)
+#define HSW_WT_ELLC_LLC_AGE3		HSW_CACHEABILITY_CONTROL(0x7)
+
+#define PPAT_UNCACHED_INDEX		(_PAGE_PWT | _PAGE_PCD)
+#define PPAT_CACHED_PDE_INDEX		0 /* WB LLC */
+#define PPAT_CACHED_INDEX		_PAGE_PAT /* WB LLCeLLC */
+#define PPAT_DISPLAY_ELLC_INDEX		_PAGE_PCD /* WT eLLC */
+
+#define GEN8_LEGACY_PDPS		4
+#define GEN8_PTES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_gtt_pte_t))
+#define GEN8_PDES_PER_PAGE		(PAGE_SIZE / sizeof(gen8_ppgtt_pde_t))
+
+/* GEN8 legacy style addressis defined as a 3 level page table:
+ * 31:30 | 29:21 | 20:12 |  11:0
+ * PDPE  |  PDE  |  PTE  | offset
+ * The difference as compared to normal x86 3 level page table is the PDPEs are
+ * programmed via register.
+ *
+ * The x86 pagetable code is flexible in its ability to handle varying page
+ * table depths via abstracted PGDIR/PUD/PMD/PTE. I've opted to not do this and
+ * instead replicate the interesting functionality.
+ */
+#define GEN8_PDPE_SHIFT			30
+#define GEN8_PDPE_MASK			0x3
+#define GEN8_PDE_SHIFT			21
+#define GEN8_PDE_MASK			0x1ff
+#define GEN8_PTE_SHIFT			12
+#define GEN8_PTE_MASK			0x1ff
+
+enum i915_cache_level;
+/**
+ * A VMA represents a GEM BO that is bound into an address space. Therefore, a
+ * VMA's presence cannot be guaranteed before binding, or after unbinding the
+ * object into/from the address space.
+ *
+ * To make things as simple as possible (ie. no refcounting), a VMA's lifetime
+ * will always be <= an objects lifetime. So object refcounting should cover us.
+ */
+struct i915_vma {
+	struct drm_mm_node node;
+	struct drm_i915_gem_object *obj;
+	struct i915_address_space *vm;
+
+	/** This object's place on the active/inactive lists */
+	struct list_head mm_list;
+
+	struct list_head vma_link; /* Link in the object's VMA list */
+
+	/** This vma's place in the batchbuffer or on the eviction list */
+	struct list_head exec_list;
+
+	/**
+	 * Used for performing relocations during execbuffer insertion.
+	 */
+	struct hlist_node exec_node;
+	unsigned long exec_handle;
+	struct drm_i915_gem_exec_object2 *exec_entry;
+
+	/**
+	 * How many users have pinned this object in GTT space. The following
+	 * users can each hold at most one reference: pwrite/pread, pin_ioctl
+	 * (via user_pin_count), execbuffer (objects are not allowed multiple
+	 * times for the same batchbuffer), and the framebuffer code. When
+	 * switching/pageflipping, the framebuffer code has at most two buffers
+	 * pinned per crtc.
+	 *
+	 * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3
+	 * bits with absolutely no headroom. So use 4 bits. */
+	unsigned int pin_count:4;
+#define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf
+
+	/** Unmap an object from an address space. This usually consists of
+	 * setting the valid PTE entries to a reserved scratch page. */
+	void (*unbind_vma)(struct i915_vma *vma);
+	/* Map an object into an address space with the given cache flags. */
+#define GLOBAL_BIND (1<<0)
+	void (*bind_vma)(struct i915_vma *vma,
+			 enum i915_cache_level cache_level,
+			 u32 flags);
+};
+
+struct i915_address_space {
+	struct drm_mm mm;
+	struct drm_device *dev;
+	struct list_head global_link;
+	unsigned long start;		/* Start offset always 0 for dri2 */
+	size_t total;		/* size addr space maps (ex. 2GB for ggtt) */
+
+	struct {
+		dma_addr_t addr;
+		struct page *page;
+	} scratch;
+
+	/**
+	 * List of objects currently involved in rendering.
+	 *
+	 * Includes buffers having the contents of their GPU caches
+	 * flushed, not necessarily primitives.  last_rendering_seqno
+	 * represents when the rendering involved will be completed.
+	 *
+	 * A reference is held on the buffer while on this list.
+	 */
+	struct list_head active_list;
+
+	/**
+	 * LRU list of objects which are not in the ringbuffer and
+	 * are ready to unbind, but are still in the GTT.
+	 *
+	 * last_rendering_seqno is 0 while an object is in this list.
+	 *
+	 * A reference is not held on the buffer while on this list,
+	 * as merely being GTT-bound shouldn't prevent its being
+	 * freed, and we'll pull it off the list in the free path.
+	 */
+	struct list_head inactive_list;
+
+	/* FIXME: Need a more generic return type */
+	gen6_gtt_pte_t (*pte_encode)(dma_addr_t addr,
+				     enum i915_cache_level level,
+				     bool valid); /* Create a valid PTE */
+	void (*clear_range)(struct i915_address_space *vm,
+			    uint64_t start,
+			    uint64_t length,
+			    bool use_scratch);
+	void (*insert_entries)(struct i915_address_space *vm,
+			       struct sg_table *st,
+			       uint64_t start,
+			       enum i915_cache_level cache_level);
+	void (*cleanup)(struct i915_address_space *vm);
+};
+
+/* The Graphics Translation Table is the way in which GEN hardware translates a
+ * Graphics Virtual Address into a Physical Address. In addition to the normal
+ * collateral associated with any va->pa translations GEN hardware also has a
+ * portion of the GTT which can be mapped by the CPU and remain both coherent
+ * and correct (in cases like swizzling). That region is referred to as GMADR in
+ * the spec.
+ */
+struct i915_gtt {
+	struct i915_address_space base;
+	size_t stolen_size;		/* Total size of stolen memory */
+
+	unsigned long mappable_end;	/* End offset that we can CPU map */
+	struct io_mapping *mappable;	/* Mapping to our CPU mappable region */
+	phys_addr_t mappable_base;	/* PA of our GMADR */
+
+	/** "Graphics Stolen Memory" holds the global PTEs */
+	void __iomem *gsm;
+
+	bool do_idle_maps;
+
+	int mtrr;
+
+	/* global gtt ops */
+	int (*gtt_probe)(struct drm_device *dev, size_t *gtt_total,
+			  size_t *stolen, phys_addr_t *mappable_base,
+			  unsigned long *mappable_end);
+};
+#define gtt_total_entries(gtt) ((gtt).base.total >> PAGE_SHIFT)
+
+struct i915_hw_ppgtt {
+	struct i915_address_space base;
+	struct kref ref;
+	struct drm_mm_node node;
+	unsigned num_pd_entries;
+	unsigned num_pd_pages; /* gen8+ */
+	union {
+		struct page **pt_pages;
+		struct page **gen8_pt_pages[GEN8_LEGACY_PDPS];
+	};
+	struct page *pd_pages;
+	union {
+		uint32_t pd_offset;
+		dma_addr_t pd_dma_addr[GEN8_LEGACY_PDPS];
+	};
+	union {
+		dma_addr_t *pt_dma_addr;
+		dma_addr_t *gen8_pt_dma_addr[GEN8_LEGACY_PDPS];
+	};
+
+	int (*enable)(struct i915_hw_ppgtt *ppgtt);
+	int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
+			 struct intel_ring_buffer *ring,
+			 bool synchronous);
+	void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
+};
+
+#endif