diff mbox series

[v9,31/70] drm/i915: Fix workarounds selftest, part 1

Message ID 20210323155059.628690-32-maarten.lankhorst@linux.intel.com (mailing list archive)
State New, archived
Headers show
Series drm/i915: Remove obj->mm.lock! | expand

Commit Message

Maarten Lankhorst March 23, 2021, 3:50 p.m. UTC
pin_map needs the ww lock, so ensure we pin both before submission.

Signed-off-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com>
Reviewed-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
---
 drivers/gpu/drm/i915/gem/i915_gem_object.h    |  3 +
 drivers/gpu/drm/i915/gem/i915_gem_pages.c     | 12 +++
 .../gpu/drm/i915/gt/selftest_workarounds.c    | 95 +++++++++++++------
 3 files changed, 80 insertions(+), 30 deletions(-)

Comments

Daniel Vetter March 24, 2021, 4:16 p.m. UTC | #1
On Tue, Mar 23, 2021 at 04:50:20PM +0100, Maarten Lankhorst wrote:
> pin_map needs the ww lock, so ensure we pin both before submission.
> 
> Signed-off-by: Maarten Lankhorst <maarten.lankhorst@linux.intel.com>
> Reviewed-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>

Another one where I picked an older version:

https://lore.kernel.org/intel-gfx/20210128162612.927917-32-maarten.lankhorst@linux.intel.com/

Cheers, Daniel

> ---
>  drivers/gpu/drm/i915/gem/i915_gem_object.h    |  3 +
>  drivers/gpu/drm/i915/gem/i915_gem_pages.c     | 12 +++
>  .../gpu/drm/i915/gt/selftest_workarounds.c    | 95 +++++++++++++------
>  3 files changed, 80 insertions(+), 30 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h b/drivers/gpu/drm/i915/gem/i915_gem_object.h
> index 6c3f75adb53c..983f2d4b2a85 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_object.h
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h
> @@ -437,6 +437,9 @@ void i915_gem_object_writeback(struct drm_i915_gem_object *obj);
>  void *__must_check i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
>  					   enum i915_map_type type);
>  
> +void *__must_check i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
> +						    enum i915_map_type type);
> +
>  void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
>  				 unsigned long offset,
>  				 unsigned long size);
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
> index e947d4c0da1f..a24617af3c93 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
> @@ -400,6 +400,18 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
>  	goto out_unlock;
>  }
>  
> +void *i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
> +				       enum i915_map_type type)
> +{
> +	void *ret;
> +
> +	i915_gem_object_lock(obj, NULL);
> +	ret = i915_gem_object_pin_map(obj, type);
> +	i915_gem_object_unlock(obj);
> +
> +	return ret;
> +}
> +
>  void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
>  				 unsigned long offset,
>  				 unsigned long size)
> diff --git a/drivers/gpu/drm/i915/gt/selftest_workarounds.c b/drivers/gpu/drm/i915/gt/selftest_workarounds.c
> index de6136bd10ac..a508614b2fd5 100644
> --- a/drivers/gpu/drm/i915/gt/selftest_workarounds.c
> +++ b/drivers/gpu/drm/i915/gt/selftest_workarounds.c
> @@ -103,15 +103,13 @@ read_nonprivs(struct intel_context *ce, struct i915_vma *result)
>  	int err;
>  	int i;
>  
> -	rq = intel_context_create_request(ce);
> +	rq = i915_request_create(ce);
>  	if (IS_ERR(rq))
>  		return rq;
>  
> -	i915_vma_lock(result);
>  	err = i915_request_await_object(rq, result->obj, true);
>  	if (err == 0)
>  		err = i915_vma_move_to_active(result, rq, EXEC_OBJECT_WRITE);
> -	i915_vma_unlock(result);
>  	if (err)
>  		goto err_rq;
>  
> @@ -176,10 +174,11 @@ static int check_whitelist(struct intel_context *ce)
>  	u32 *vaddr;
>  	int i;
>  
> -	result = __vm_create_scratch_for_read(&engine->gt->ggtt->vm, PAGE_SIZE);
> +	result = __vm_create_scratch_for_read_pinned(&engine->gt->ggtt->vm, PAGE_SIZE);
>  	if (IS_ERR(result))
>  		return PTR_ERR(result);
>  
> +	i915_gem_object_lock(result->obj, NULL);
>  	vaddr = i915_gem_object_pin_map(result->obj, I915_MAP_WB);
>  	if (IS_ERR(vaddr)) {
>  		err = PTR_ERR(vaddr);
> @@ -219,6 +218,8 @@ static int check_whitelist(struct intel_context *ce)
>  out_map:
>  	i915_gem_object_unpin_map(result->obj);
>  out_put:
> +	i915_vma_unpin(result);
> +	i915_gem_object_unlock(result->obj);
>  	i915_vma_put(result);
>  	return err;
>  }
> @@ -279,10 +280,14 @@ static int check_whitelist_across_reset(struct intel_engine_cs *engine,
>  	if (IS_ERR(ce))
>  		return PTR_ERR(ce);
>  
> -	err = igt_spinner_init(&spin, engine->gt);
> +	err = intel_context_pin(ce);
>  	if (err)
>  		goto out_ctx;
>  
> +	err = igt_spinner_init(&spin, engine->gt);
> +	if (err)
> +		goto out_unpin;
> +
>  	err = check_whitelist(ce);
>  	if (err) {
>  		pr_err("Invalid whitelist *before* %s reset!\n", name);
> @@ -315,6 +320,13 @@ static int check_whitelist_across_reset(struct intel_engine_cs *engine,
>  		err = PTR_ERR(tmp);
>  		goto out_spin;
>  	}
> +	err = intel_context_pin(tmp);
> +	if (err) {
> +		intel_context_put(tmp);
> +		goto out_spin;
> +	}
> +
> +	intel_context_unpin(ce);
>  	intel_context_put(ce);
>  	ce = tmp;
>  
> @@ -327,6 +339,8 @@ static int check_whitelist_across_reset(struct intel_engine_cs *engine,
>  
>  out_spin:
>  	igt_spinner_fini(&spin);
> +out_unpin:
> +	intel_context_unpin(ce);
>  out_ctx:
>  	intel_context_put(ce);
>  	return err;
> @@ -475,6 +489,7 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  
>  	for (i = 0; i < engine->whitelist.count; i++) {
>  		u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg);
> +		struct i915_gem_ww_ctx ww;
>  		u64 addr = scratch->node.start;
>  		struct i915_request *rq;
>  		u32 srm, lrm, rsvd;
> @@ -490,6 +505,29 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  
>  		ro_reg = ro_register(reg);
>  
> +		i915_gem_ww_ctx_init(&ww, false);
> +retry:
> +		cs = NULL;
> +		err = i915_gem_object_lock(scratch->obj, &ww);
> +		if (!err)
> +			err = i915_gem_object_lock(batch->obj, &ww);
> +		if (!err)
> +			err = intel_context_pin_ww(ce, &ww);
> +		if (err)
> +			goto out;
> +
> +		cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
> +		if (IS_ERR(cs)) {
> +			err = PTR_ERR(cs);
> +			goto out_ctx;
> +		}
> +
> +		results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
> +		if (IS_ERR(results)) {
> +			err = PTR_ERR(results);
> +			goto out_unmap_batch;
> +		}
> +
>  		/* Clear non priv flags */
>  		reg &= RING_FORCE_TO_NONPRIV_ADDRESS_MASK;
>  
> @@ -501,12 +539,6 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  		pr_debug("%s: Writing garbage to %x\n",
>  			 engine->name, reg);
>  
> -		cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
> -		if (IS_ERR(cs)) {
> -			err = PTR_ERR(cs);
> -			goto out_batch;
> -		}
> -
>  		/* SRM original */
>  		*cs++ = srm;
>  		*cs++ = reg;
> @@ -553,11 +585,12 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  		i915_gem_object_flush_map(batch->obj);
>  		i915_gem_object_unpin_map(batch->obj);
>  		intel_gt_chipset_flush(engine->gt);
> +		cs = NULL;
>  
> -		rq = intel_context_create_request(ce);
> +		rq = i915_request_create(ce);
>  		if (IS_ERR(rq)) {
>  			err = PTR_ERR(rq);
> -			goto out_batch;
> +			goto out_unmap_scratch;
>  		}
>  
>  		if (engine->emit_init_breadcrumb) { /* Be nice if we hang */
> @@ -566,20 +599,16 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  				goto err_request;
>  		}
>  
> -		i915_vma_lock(batch);
>  		err = i915_request_await_object(rq, batch->obj, false);
>  		if (err == 0)
>  			err = i915_vma_move_to_active(batch, rq, 0);
> -		i915_vma_unlock(batch);
>  		if (err)
>  			goto err_request;
>  
> -		i915_vma_lock(scratch);
>  		err = i915_request_await_object(rq, scratch->obj, true);
>  		if (err == 0)
>  			err = i915_vma_move_to_active(scratch, rq,
>  						      EXEC_OBJECT_WRITE);
> -		i915_vma_unlock(scratch);
>  		if (err)
>  			goto err_request;
>  
> @@ -595,13 +624,7 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  			pr_err("%s: Futzing %x timedout; cancelling test\n",
>  			       engine->name, reg);
>  			intel_gt_set_wedged(engine->gt);
> -			goto out_batch;
> -		}
> -
> -		results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
> -		if (IS_ERR(results)) {
> -			err = PTR_ERR(results);
> -			goto out_batch;
> +			goto out_unmap_scratch;
>  		}
>  
>  		GEM_BUG_ON(values[ARRAY_SIZE(values) - 1] != 0xffffffff);
> @@ -612,7 +635,7 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  				pr_err("%s: Unable to write to whitelisted register %x\n",
>  				       engine->name, reg);
>  				err = -EINVAL;
> -				goto out_unpin;
> +				goto out_unmap_scratch;
>  			}
>  		} else {
>  			rsvd = 0;
> @@ -678,15 +701,27 @@ static int check_dirty_whitelist(struct intel_context *ce)
>  
>  			err = -EINVAL;
>  		}
> -out_unpin:
> +out_unmap_scratch:
>  		i915_gem_object_unpin_map(scratch->obj);
> +out_unmap_batch:
> +		if (cs)
> +			i915_gem_object_unpin_map(batch->obj);
> +out_ctx:
> +		intel_context_unpin(ce);
> +out:
> +		if (err == -EDEADLK) {
> +			err = i915_gem_ww_ctx_backoff(&ww);
> +			if (!err)
> +				goto retry;
> +		}
> +		i915_gem_ww_ctx_fini(&ww);
>  		if (err)
>  			break;
>  	}
>  
>  	if (igt_flush_test(engine->i915))
>  		err = -EIO;
> -out_batch:
> +
>  	i915_vma_unpin_and_release(&batch, 0);
>  out_scratch:
>  	i915_vma_unpin_and_release(&scratch, 0);
> @@ -820,7 +855,7 @@ static int scrub_whitelisted_registers(struct intel_context *ce)
>  	if (IS_ERR(batch))
>  		return PTR_ERR(batch);
>  
> -	cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
> +	cs = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
>  	if (IS_ERR(cs)) {
>  		err = PTR_ERR(cs);
>  		goto err_batch;
> @@ -955,11 +990,11 @@ check_whitelisted_registers(struct intel_engine_cs *engine,
>  	u32 *a, *b;
>  	int i, err;
>  
> -	a = i915_gem_object_pin_map(A->obj, I915_MAP_WB);
> +	a = i915_gem_object_pin_map_unlocked(A->obj, I915_MAP_WB);
>  	if (IS_ERR(a))
>  		return PTR_ERR(a);
>  
> -	b = i915_gem_object_pin_map(B->obj, I915_MAP_WB);
> +	b = i915_gem_object_pin_map_unlocked(B->obj, I915_MAP_WB);
>  	if (IS_ERR(b)) {
>  		err = PTR_ERR(b);
>  		goto err_a;
> -- 
> 2.31.0
> 
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx
diff mbox series

Patch

diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h b/drivers/gpu/drm/i915/gem/i915_gem_object.h
index 6c3f75adb53c..983f2d4b2a85 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_object.h
+++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h
@@ -437,6 +437,9 @@  void i915_gem_object_writeback(struct drm_i915_gem_object *obj);
 void *__must_check i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
 					   enum i915_map_type type);
 
+void *__must_check i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
+						    enum i915_map_type type);
+
 void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
 				 unsigned long offset,
 				 unsigned long size);
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
index e947d4c0da1f..a24617af3c93 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c
@@ -400,6 +400,18 @@  void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
 	goto out_unlock;
 }
 
+void *i915_gem_object_pin_map_unlocked(struct drm_i915_gem_object *obj,
+				       enum i915_map_type type)
+{
+	void *ret;
+
+	i915_gem_object_lock(obj, NULL);
+	ret = i915_gem_object_pin_map(obj, type);
+	i915_gem_object_unlock(obj);
+
+	return ret;
+}
+
 void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
 				 unsigned long offset,
 				 unsigned long size)
diff --git a/drivers/gpu/drm/i915/gt/selftest_workarounds.c b/drivers/gpu/drm/i915/gt/selftest_workarounds.c
index de6136bd10ac..a508614b2fd5 100644
--- a/drivers/gpu/drm/i915/gt/selftest_workarounds.c
+++ b/drivers/gpu/drm/i915/gt/selftest_workarounds.c
@@ -103,15 +103,13 @@  read_nonprivs(struct intel_context *ce, struct i915_vma *result)
 	int err;
 	int i;
 
-	rq = intel_context_create_request(ce);
+	rq = i915_request_create(ce);
 	if (IS_ERR(rq))
 		return rq;
 
-	i915_vma_lock(result);
 	err = i915_request_await_object(rq, result->obj, true);
 	if (err == 0)
 		err = i915_vma_move_to_active(result, rq, EXEC_OBJECT_WRITE);
-	i915_vma_unlock(result);
 	if (err)
 		goto err_rq;
 
@@ -176,10 +174,11 @@  static int check_whitelist(struct intel_context *ce)
 	u32 *vaddr;
 	int i;
 
-	result = __vm_create_scratch_for_read(&engine->gt->ggtt->vm, PAGE_SIZE);
+	result = __vm_create_scratch_for_read_pinned(&engine->gt->ggtt->vm, PAGE_SIZE);
 	if (IS_ERR(result))
 		return PTR_ERR(result);
 
+	i915_gem_object_lock(result->obj, NULL);
 	vaddr = i915_gem_object_pin_map(result->obj, I915_MAP_WB);
 	if (IS_ERR(vaddr)) {
 		err = PTR_ERR(vaddr);
@@ -219,6 +218,8 @@  static int check_whitelist(struct intel_context *ce)
 out_map:
 	i915_gem_object_unpin_map(result->obj);
 out_put:
+	i915_vma_unpin(result);
+	i915_gem_object_unlock(result->obj);
 	i915_vma_put(result);
 	return err;
 }
@@ -279,10 +280,14 @@  static int check_whitelist_across_reset(struct intel_engine_cs *engine,
 	if (IS_ERR(ce))
 		return PTR_ERR(ce);
 
-	err = igt_spinner_init(&spin, engine->gt);
+	err = intel_context_pin(ce);
 	if (err)
 		goto out_ctx;
 
+	err = igt_spinner_init(&spin, engine->gt);
+	if (err)
+		goto out_unpin;
+
 	err = check_whitelist(ce);
 	if (err) {
 		pr_err("Invalid whitelist *before* %s reset!\n", name);
@@ -315,6 +320,13 @@  static int check_whitelist_across_reset(struct intel_engine_cs *engine,
 		err = PTR_ERR(tmp);
 		goto out_spin;
 	}
+	err = intel_context_pin(tmp);
+	if (err) {
+		intel_context_put(tmp);
+		goto out_spin;
+	}
+
+	intel_context_unpin(ce);
 	intel_context_put(ce);
 	ce = tmp;
 
@@ -327,6 +339,8 @@  static int check_whitelist_across_reset(struct intel_engine_cs *engine,
 
 out_spin:
 	igt_spinner_fini(&spin);
+out_unpin:
+	intel_context_unpin(ce);
 out_ctx:
 	intel_context_put(ce);
 	return err;
@@ -475,6 +489,7 @@  static int check_dirty_whitelist(struct intel_context *ce)
 
 	for (i = 0; i < engine->whitelist.count; i++) {
 		u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg);
+		struct i915_gem_ww_ctx ww;
 		u64 addr = scratch->node.start;
 		struct i915_request *rq;
 		u32 srm, lrm, rsvd;
@@ -490,6 +505,29 @@  static int check_dirty_whitelist(struct intel_context *ce)
 
 		ro_reg = ro_register(reg);
 
+		i915_gem_ww_ctx_init(&ww, false);
+retry:
+		cs = NULL;
+		err = i915_gem_object_lock(scratch->obj, &ww);
+		if (!err)
+			err = i915_gem_object_lock(batch->obj, &ww);
+		if (!err)
+			err = intel_context_pin_ww(ce, &ww);
+		if (err)
+			goto out;
+
+		cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
+		if (IS_ERR(cs)) {
+			err = PTR_ERR(cs);
+			goto out_ctx;
+		}
+
+		results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
+		if (IS_ERR(results)) {
+			err = PTR_ERR(results);
+			goto out_unmap_batch;
+		}
+
 		/* Clear non priv flags */
 		reg &= RING_FORCE_TO_NONPRIV_ADDRESS_MASK;
 
@@ -501,12 +539,6 @@  static int check_dirty_whitelist(struct intel_context *ce)
 		pr_debug("%s: Writing garbage to %x\n",
 			 engine->name, reg);
 
-		cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
-		if (IS_ERR(cs)) {
-			err = PTR_ERR(cs);
-			goto out_batch;
-		}
-
 		/* SRM original */
 		*cs++ = srm;
 		*cs++ = reg;
@@ -553,11 +585,12 @@  static int check_dirty_whitelist(struct intel_context *ce)
 		i915_gem_object_flush_map(batch->obj);
 		i915_gem_object_unpin_map(batch->obj);
 		intel_gt_chipset_flush(engine->gt);
+		cs = NULL;
 
-		rq = intel_context_create_request(ce);
+		rq = i915_request_create(ce);
 		if (IS_ERR(rq)) {
 			err = PTR_ERR(rq);
-			goto out_batch;
+			goto out_unmap_scratch;
 		}
 
 		if (engine->emit_init_breadcrumb) { /* Be nice if we hang */
@@ -566,20 +599,16 @@  static int check_dirty_whitelist(struct intel_context *ce)
 				goto err_request;
 		}
 
-		i915_vma_lock(batch);
 		err = i915_request_await_object(rq, batch->obj, false);
 		if (err == 0)
 			err = i915_vma_move_to_active(batch, rq, 0);
-		i915_vma_unlock(batch);
 		if (err)
 			goto err_request;
 
-		i915_vma_lock(scratch);
 		err = i915_request_await_object(rq, scratch->obj, true);
 		if (err == 0)
 			err = i915_vma_move_to_active(scratch, rq,
 						      EXEC_OBJECT_WRITE);
-		i915_vma_unlock(scratch);
 		if (err)
 			goto err_request;
 
@@ -595,13 +624,7 @@  static int check_dirty_whitelist(struct intel_context *ce)
 			pr_err("%s: Futzing %x timedout; cancelling test\n",
 			       engine->name, reg);
 			intel_gt_set_wedged(engine->gt);
-			goto out_batch;
-		}
-
-		results = i915_gem_object_pin_map(scratch->obj, I915_MAP_WB);
-		if (IS_ERR(results)) {
-			err = PTR_ERR(results);
-			goto out_batch;
+			goto out_unmap_scratch;
 		}
 
 		GEM_BUG_ON(values[ARRAY_SIZE(values) - 1] != 0xffffffff);
@@ -612,7 +635,7 @@  static int check_dirty_whitelist(struct intel_context *ce)
 				pr_err("%s: Unable to write to whitelisted register %x\n",
 				       engine->name, reg);
 				err = -EINVAL;
-				goto out_unpin;
+				goto out_unmap_scratch;
 			}
 		} else {
 			rsvd = 0;
@@ -678,15 +701,27 @@  static int check_dirty_whitelist(struct intel_context *ce)
 
 			err = -EINVAL;
 		}
-out_unpin:
+out_unmap_scratch:
 		i915_gem_object_unpin_map(scratch->obj);
+out_unmap_batch:
+		if (cs)
+			i915_gem_object_unpin_map(batch->obj);
+out_ctx:
+		intel_context_unpin(ce);
+out:
+		if (err == -EDEADLK) {
+			err = i915_gem_ww_ctx_backoff(&ww);
+			if (!err)
+				goto retry;
+		}
+		i915_gem_ww_ctx_fini(&ww);
 		if (err)
 			break;
 	}
 
 	if (igt_flush_test(engine->i915))
 		err = -EIO;
-out_batch:
+
 	i915_vma_unpin_and_release(&batch, 0);
 out_scratch:
 	i915_vma_unpin_and_release(&scratch, 0);
@@ -820,7 +855,7 @@  static int scrub_whitelisted_registers(struct intel_context *ce)
 	if (IS_ERR(batch))
 		return PTR_ERR(batch);
 
-	cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC);
+	cs = i915_gem_object_pin_map_unlocked(batch->obj, I915_MAP_WC);
 	if (IS_ERR(cs)) {
 		err = PTR_ERR(cs);
 		goto err_batch;
@@ -955,11 +990,11 @@  check_whitelisted_registers(struct intel_engine_cs *engine,
 	u32 *a, *b;
 	int i, err;
 
-	a = i915_gem_object_pin_map(A->obj, I915_MAP_WB);
+	a = i915_gem_object_pin_map_unlocked(A->obj, I915_MAP_WB);
 	if (IS_ERR(a))
 		return PTR_ERR(a);
 
-	b = i915_gem_object_pin_map(B->obj, I915_MAP_WB);
+	b = i915_gem_object_pin_map_unlocked(B->obj, I915_MAP_WB);
 	if (IS_ERR(b)) {
 		err = PTR_ERR(b);
 		goto err_a;