diff mbox series

[2/4] drm/i915/selftests: Include arbitration points in preemption smoketest

Message ID 20181001123204.23982-2-chris@chris-wilson.co.uk (mailing list archive)
State New, archived
Headers show
Series [1/4] drm/i915/selftests: Split preemption smoke test into threads | expand

Commit Message

Chris Wilson Oct. 1, 2018, 12:32 p.m. UTC
Include a batch full of a page of arbitration points in order to provide
a window for inject_preempt_context() in the preemption smoketests.

Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
---
 drivers/gpu/drm/i915/selftests/intel_lrc.c | 134 ++++++++++++++++-----
 1 file changed, 101 insertions(+), 33 deletions(-)

Comments

Tvrtko Ursulin Oct. 1, 2018, 2:01 p.m. UTC | #1
On 01/10/2018 13:32, Chris Wilson wrote:
> Include a batch full of a page of arbitration points in order to provide
> a window for inject_preempt_context() in the preemption smoketests.
> 
> Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> ---
>   drivers/gpu/drm/i915/selftests/intel_lrc.c | 134 ++++++++++++++++-----
>   1 file changed, 101 insertions(+), 33 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/selftests/intel_lrc.c b/drivers/gpu/drm/i915/selftests/intel_lrc.c
> index d68a924c530e..d67fe8335ceb 100644
> --- a/drivers/gpu/drm/i915/selftests/intel_lrc.c
> +++ b/drivers/gpu/drm/i915/selftests/intel_lrc.c
> @@ -588,6 +588,7 @@ struct preempt_smoke {
>   	struct drm_i915_private *i915;
>   	struct i915_gem_context **contexts;
>   	struct intel_engine_cs *engine;
> +	struct drm_i915_gem_object *batch;
>   	unsigned int ncontext;
>   	struct rnd_state prng;
>   	unsigned long count;
> @@ -599,6 +600,49 @@ static struct i915_gem_context *smoke_context(struct preempt_smoke *smoke)
>   							  &smoke->prng)];
>   }
>   
> +static int smoke_submit(struct preempt_smoke *smoke,
> +			struct i915_gem_context *ctx, int prio,
> +			struct drm_i915_gem_object *batch)
> +{
> +	struct i915_request *rq;
> +	struct i915_vma *vma = NULL;
> +	int err = 0;
> +
> +	if (batch) {
> +		vma = i915_vma_instance(batch, &ctx->ppgtt->vm, NULL);
> +		if (IS_ERR(vma))
> +			return PTR_ERR(vma);
> +
> +		err = i915_vma_pin(vma, 0, 0, PIN_USER);
> +		if (err)
> +			return err;
> +	}
> +
> +	ctx->sched.priority = prio;
> +
> +	rq = i915_request_alloc(smoke->engine, ctx);
> +	if (IS_ERR(rq)) {
> +		err = PTR_ERR(rq);
> +		goto unpin;
> +	}
> +
> +	if (vma) {
> +		err = rq->engine->emit_bb_start(rq,
> +						vma->node.start,
> +						PAGE_SIZE, 0);
> +		if (!err)
> +			err = i915_vma_move_to_active(vma, rq, 0);
> +	}
> +
> +	i915_request_add(rq);
> +
> +unpin:
> +	if (vma)
> +		i915_vma_unpin(vma);
> +
> +	return err;
> +}
> +
>   static int smoke_crescendo_thread(void *arg)
>   {
>   	struct preempt_smoke *smoke = arg;
> @@ -608,21 +652,15 @@ static int smoke_crescendo_thread(void *arg)
>   	count = 0;
>   	do {
>   		struct i915_gem_context *ctx = smoke_context(smoke);
> -		struct i915_request *rq;
> +		int err;
>   
>   		mutex_lock(&smoke->i915->drm.struct_mutex);
> -
> -		ctx->sched.priority = count % I915_PRIORITY_MAX;
> -
> -		rq = i915_request_alloc(smoke->engine, ctx);
> -		if (IS_ERR(rq)) {
> -			mutex_unlock(&smoke->i915->drm.struct_mutex);
> -			return PTR_ERR(rq);
> -		}
> -
> -		i915_request_add(rq);
> -
> +		err = smoke_submit(smoke,
> +				   ctx, count % I915_PRIORITY_MAX,
> +				   smoke->batch);
>   		mutex_unlock(&smoke->i915->drm.struct_mutex);
> +		if (err)
> +			return err;
>   
>   		count++;
>   	} while (!__igt_timeout(end_time, NULL));
> @@ -631,7 +669,8 @@ static int smoke_crescendo_thread(void *arg)
>   	return 0;
>   }
>   
> -static int smoke_crescendo(struct preempt_smoke *smoke)
> +static int smoke_crescendo(struct preempt_smoke *smoke, unsigned int flags)
> +#define BATCH BIT(0)
>   {
>   	struct task_struct *tsk[I915_NUM_ENGINES] = {};
>   	struct preempt_smoke arg[I915_NUM_ENGINES];
> @@ -645,6 +684,8 @@ static int smoke_crescendo(struct preempt_smoke *smoke)
>   	for_each_engine(engine, smoke->i915, id) {
>   		arg[id] = *smoke;
>   		arg[id].engine = engine;
> +		if (!(flags & BATCH))
> +			arg[id].batch = NULL;
>   		arg[id].count = 0;
>   
>   		tsk[id] = kthread_run(smoke_crescendo_thread, &arg,
> @@ -671,37 +712,37 @@ static int smoke_crescendo(struct preempt_smoke *smoke)
>   
>   	mutex_lock(&smoke->i915->drm.struct_mutex);
>   
> -	pr_info("Submitted %lu crescendo requests across %d engines and %d contexts\n",
> -		count, INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
> +	pr_info("Submitted %lu crescendo:%x requests across %d engines and %d contexts\n",
> +		count, flags,
> +		INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
>   	return 0;
>   }
>   
> -static int smoke_random(struct preempt_smoke *smoke)
> +static int smoke_random(struct preempt_smoke *smoke, unsigned int flags)
>   {
> -	struct intel_engine_cs *engine;
>   	enum intel_engine_id id;
>   	IGT_TIMEOUT(end_time);
>   	unsigned long count;
>   
>   	count = 0;
>   	do {
> -		for_each_engine(engine, smoke->i915, id) {
> +		for_each_engine(smoke->engine, smoke->i915, id) {
>   			struct i915_gem_context *ctx = smoke_context(smoke);
> -			struct i915_request *rq;
> -
> -			ctx->sched.priority = random_priority(&smoke->prng);
> +			int err;
>   
> -			rq = i915_request_alloc(engine, ctx);
> -			if (IS_ERR(rq))
> -				return PTR_ERR(rq);
> +			err = smoke_submit(smoke,
> +					   ctx, random_priority(&smoke->prng),
> +					   flags & BATCH ? smoke->batch : NULL);
> +			if (err)
> +				return err;
>   
> -			i915_request_add(rq);
>   			count++;
>   		}
>   	} while (!__igt_timeout(end_time, NULL));
>   
> -	pr_info("Submitted %lu random requests across %d engines and %d contexts\n",
> -		count, INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
> +	pr_info("Submitted %lu random:%x requests across %d engines and %d contexts\n",
> +		count, flags,
> +		INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
>   	return 0;
>   }
>   
> @@ -712,7 +753,9 @@ static int live_preempt_smoke(void *arg)
>   		.prng = I915_RND_STATE_INITIALIZER(i915_selftest.random_seed),
>   		.ncontext = 1024,
>   	};
> +	const unsigned int phase[] = { 0, BATCH };
>   	int err = -ENOMEM;
> +	u32 *cs;
>   	int n;
>   
>   	if (!HAS_LOGICAL_RING_PREEMPTION(smoke.i915))
> @@ -727,19 +770,41 @@ static int live_preempt_smoke(void *arg)
>   	mutex_lock(&smoke.i915->drm.struct_mutex);
>   	intel_runtime_pm_get(smoke.i915);
>   
> +	smoke.batch = i915_gem_object_create_internal(smoke.i915, PAGE_SIZE);
> +	if (IS_ERR(smoke.batch)) {
> +		err = PTR_ERR(smoke.batch);
> +		goto err_unlock;
> +	}
> +
> +	cs = i915_gem_object_pin_map(smoke.batch, I915_MAP_WB);
> +	if (IS_ERR(cs)) {
> +		err = PTR_ERR(cs);
> +		goto err_batch;
> +	}
> +	for (n = 0; n < PAGE_SIZE / sizeof(*cs) - 1; n++)
> +		cs[n] = MI_ARB_CHECK;
> +	cs[n] = MI_BATCH_BUFFER_END;
> +	i915_gem_object_unpin_map(smoke.batch);
> +
> +	err = i915_gem_object_set_to_gtt_domain(smoke.batch, false);
> +	if (err)
> +		goto err_batch;
> +
>   	for (n = 0; n < smoke.ncontext; n++) {
>   		smoke.contexts[n] = kernel_context(smoke.i915);
>   		if (!smoke.contexts[n])
>   			goto err_ctx;
>   	}
>   
> -	err = smoke_crescendo(&smoke);
> -	if (err)
> -		goto err_ctx;
> +	for (n = 0; n < ARRAY_SIZE(phase); n++) {
> +		err = smoke_crescendo(&smoke, phase[n]);
> +		if (err)
> +			goto err_ctx;
>   
> -	err = smoke_random(&smoke);
> -	if (err)
> -		goto err_ctx;
> +		err = smoke_random(&smoke, phase[n]);
> +		if (err)
> +			goto err_ctx;
> +	}
>   
>   err_ctx:
>   	if (igt_flush_test(smoke.i915, I915_WAIT_LOCKED))
> @@ -751,6 +816,9 @@ static int live_preempt_smoke(void *arg)
>   		kernel_context_close(smoke.contexts[n]);
>   	}
>   
> +err_batch:
> +	i915_gem_object_put(smoke.batch);
> +err_unlock:
>   	intel_runtime_pm_put(smoke.i915);
>   	mutex_unlock(&smoke.i915->drm.struct_mutex);
>   	kfree(smoke.contexts);
> 

Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>

Regards,

Tvrtko
diff mbox series

Patch

diff --git a/drivers/gpu/drm/i915/selftests/intel_lrc.c b/drivers/gpu/drm/i915/selftests/intel_lrc.c
index d68a924c530e..d67fe8335ceb 100644
--- a/drivers/gpu/drm/i915/selftests/intel_lrc.c
+++ b/drivers/gpu/drm/i915/selftests/intel_lrc.c
@@ -588,6 +588,7 @@  struct preempt_smoke {
 	struct drm_i915_private *i915;
 	struct i915_gem_context **contexts;
 	struct intel_engine_cs *engine;
+	struct drm_i915_gem_object *batch;
 	unsigned int ncontext;
 	struct rnd_state prng;
 	unsigned long count;
@@ -599,6 +600,49 @@  static struct i915_gem_context *smoke_context(struct preempt_smoke *smoke)
 							  &smoke->prng)];
 }
 
+static int smoke_submit(struct preempt_smoke *smoke,
+			struct i915_gem_context *ctx, int prio,
+			struct drm_i915_gem_object *batch)
+{
+	struct i915_request *rq;
+	struct i915_vma *vma = NULL;
+	int err = 0;
+
+	if (batch) {
+		vma = i915_vma_instance(batch, &ctx->ppgtt->vm, NULL);
+		if (IS_ERR(vma))
+			return PTR_ERR(vma);
+
+		err = i915_vma_pin(vma, 0, 0, PIN_USER);
+		if (err)
+			return err;
+	}
+
+	ctx->sched.priority = prio;
+
+	rq = i915_request_alloc(smoke->engine, ctx);
+	if (IS_ERR(rq)) {
+		err = PTR_ERR(rq);
+		goto unpin;
+	}
+
+	if (vma) {
+		err = rq->engine->emit_bb_start(rq,
+						vma->node.start,
+						PAGE_SIZE, 0);
+		if (!err)
+			err = i915_vma_move_to_active(vma, rq, 0);
+	}
+
+	i915_request_add(rq);
+
+unpin:
+	if (vma)
+		i915_vma_unpin(vma);
+
+	return err;
+}
+
 static int smoke_crescendo_thread(void *arg)
 {
 	struct preempt_smoke *smoke = arg;
@@ -608,21 +652,15 @@  static int smoke_crescendo_thread(void *arg)
 	count = 0;
 	do {
 		struct i915_gem_context *ctx = smoke_context(smoke);
-		struct i915_request *rq;
+		int err;
 
 		mutex_lock(&smoke->i915->drm.struct_mutex);
-
-		ctx->sched.priority = count % I915_PRIORITY_MAX;
-
-		rq = i915_request_alloc(smoke->engine, ctx);
-		if (IS_ERR(rq)) {
-			mutex_unlock(&smoke->i915->drm.struct_mutex);
-			return PTR_ERR(rq);
-		}
-
-		i915_request_add(rq);
-
+		err = smoke_submit(smoke,
+				   ctx, count % I915_PRIORITY_MAX,
+				   smoke->batch);
 		mutex_unlock(&smoke->i915->drm.struct_mutex);
+		if (err)
+			return err;
 
 		count++;
 	} while (!__igt_timeout(end_time, NULL));
@@ -631,7 +669,8 @@  static int smoke_crescendo_thread(void *arg)
 	return 0;
 }
 
-static int smoke_crescendo(struct preempt_smoke *smoke)
+static int smoke_crescendo(struct preempt_smoke *smoke, unsigned int flags)
+#define BATCH BIT(0)
 {
 	struct task_struct *tsk[I915_NUM_ENGINES] = {};
 	struct preempt_smoke arg[I915_NUM_ENGINES];
@@ -645,6 +684,8 @@  static int smoke_crescendo(struct preempt_smoke *smoke)
 	for_each_engine(engine, smoke->i915, id) {
 		arg[id] = *smoke;
 		arg[id].engine = engine;
+		if (!(flags & BATCH))
+			arg[id].batch = NULL;
 		arg[id].count = 0;
 
 		tsk[id] = kthread_run(smoke_crescendo_thread, &arg,
@@ -671,37 +712,37 @@  static int smoke_crescendo(struct preempt_smoke *smoke)
 
 	mutex_lock(&smoke->i915->drm.struct_mutex);
 
-	pr_info("Submitted %lu crescendo requests across %d engines and %d contexts\n",
-		count, INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
+	pr_info("Submitted %lu crescendo:%x requests across %d engines and %d contexts\n",
+		count, flags,
+		INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
 	return 0;
 }
 
-static int smoke_random(struct preempt_smoke *smoke)
+static int smoke_random(struct preempt_smoke *smoke, unsigned int flags)
 {
-	struct intel_engine_cs *engine;
 	enum intel_engine_id id;
 	IGT_TIMEOUT(end_time);
 	unsigned long count;
 
 	count = 0;
 	do {
-		for_each_engine(engine, smoke->i915, id) {
+		for_each_engine(smoke->engine, smoke->i915, id) {
 			struct i915_gem_context *ctx = smoke_context(smoke);
-			struct i915_request *rq;
-
-			ctx->sched.priority = random_priority(&smoke->prng);
+			int err;
 
-			rq = i915_request_alloc(engine, ctx);
-			if (IS_ERR(rq))
-				return PTR_ERR(rq);
+			err = smoke_submit(smoke,
+					   ctx, random_priority(&smoke->prng),
+					   flags & BATCH ? smoke->batch : NULL);
+			if (err)
+				return err;
 
-			i915_request_add(rq);
 			count++;
 		}
 	} while (!__igt_timeout(end_time, NULL));
 
-	pr_info("Submitted %lu random requests across %d engines and %d contexts\n",
-		count, INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
+	pr_info("Submitted %lu random:%x requests across %d engines and %d contexts\n",
+		count, flags,
+		INTEL_INFO(smoke->i915)->num_rings, smoke->ncontext);
 	return 0;
 }
 
@@ -712,7 +753,9 @@  static int live_preempt_smoke(void *arg)
 		.prng = I915_RND_STATE_INITIALIZER(i915_selftest.random_seed),
 		.ncontext = 1024,
 	};
+	const unsigned int phase[] = { 0, BATCH };
 	int err = -ENOMEM;
+	u32 *cs;
 	int n;
 
 	if (!HAS_LOGICAL_RING_PREEMPTION(smoke.i915))
@@ -727,19 +770,41 @@  static int live_preempt_smoke(void *arg)
 	mutex_lock(&smoke.i915->drm.struct_mutex);
 	intel_runtime_pm_get(smoke.i915);
 
+	smoke.batch = i915_gem_object_create_internal(smoke.i915, PAGE_SIZE);
+	if (IS_ERR(smoke.batch)) {
+		err = PTR_ERR(smoke.batch);
+		goto err_unlock;
+	}
+
+	cs = i915_gem_object_pin_map(smoke.batch, I915_MAP_WB);
+	if (IS_ERR(cs)) {
+		err = PTR_ERR(cs);
+		goto err_batch;
+	}
+	for (n = 0; n < PAGE_SIZE / sizeof(*cs) - 1; n++)
+		cs[n] = MI_ARB_CHECK;
+	cs[n] = MI_BATCH_BUFFER_END;
+	i915_gem_object_unpin_map(smoke.batch);
+
+	err = i915_gem_object_set_to_gtt_domain(smoke.batch, false);
+	if (err)
+		goto err_batch;
+
 	for (n = 0; n < smoke.ncontext; n++) {
 		smoke.contexts[n] = kernel_context(smoke.i915);
 		if (!smoke.contexts[n])
 			goto err_ctx;
 	}
 
-	err = smoke_crescendo(&smoke);
-	if (err)
-		goto err_ctx;
+	for (n = 0; n < ARRAY_SIZE(phase); n++) {
+		err = smoke_crescendo(&smoke, phase[n]);
+		if (err)
+			goto err_ctx;
 
-	err = smoke_random(&smoke);
-	if (err)
-		goto err_ctx;
+		err = smoke_random(&smoke, phase[n]);
+		if (err)
+			goto err_ctx;
+	}
 
 err_ctx:
 	if (igt_flush_test(smoke.i915, I915_WAIT_LOCKED))
@@ -751,6 +816,9 @@  static int live_preempt_smoke(void *arg)
 		kernel_context_close(smoke.contexts[n]);
 	}
 
+err_batch:
+	i915_gem_object_put(smoke.batch);
+err_unlock:
 	intel_runtime_pm_put(smoke.i915);
 	mutex_unlock(&smoke.i915->drm.struct_mutex);
 	kfree(smoke.contexts);