@@ -1245,6 +1245,7 @@ i915_gem_ringbuffer_submission(struct i915_execbuffer_params *params,
int i915_gem_ringbuffer_submission_final(struct i915_execbuffer_params *params)
{
struct drm_i915_private *dev_priv = params->dev->dev_private;
+ struct drm_i915_gem_request *req = params->request;
struct intel_engine_cs *ring = params->ring;
u64 exec_start, exec_len;
int ret;
@@ -1258,12 +1259,12 @@ int i915_gem_ringbuffer_submission_final(struct i915_execbuffer_params *params)
* Unconditionally invalidate gpu caches and ensure that we do flush
* any residual writes from the previous batch.
*/
- ret = intel_ring_invalidate_all_caches(params->request);
+ ret = intel_ring_invalidate_all_caches(req);
if (ret)
goto error;
/* Switch to the correct context for the batch */
- ret = i915_switch_context(params->request);
+ ret = i915_switch_context(req);
if (ret)
goto error;
@@ -1272,7 +1273,7 @@ int i915_gem_ringbuffer_submission_final(struct i915_execbuffer_params *params)
if (ring == &dev_priv->ring[RCS] &&
params->instp_mode != dev_priv->relative_constants_mode) {
- ret = intel_ring_begin(params->request, 4);
+ ret = intel_ring_begin(req, 4);
if (ret)
goto error;
@@ -1286,7 +1287,7 @@ int i915_gem_ringbuffer_submission_final(struct i915_execbuffer_params *params)
}
if (params->args_flags & I915_EXEC_GEN7_SOL_RESET) {
- ret = i915_reset_gen7_sol_offsets(params->dev, params->request);
+ ret = i915_reset_gen7_sol_offsets(params->dev, req);
if (ret)
goto error;
}
@@ -1295,13 +1296,13 @@ int i915_gem_ringbuffer_submission_final(struct i915_execbuffer_params *params)
exec_start = params->batch_obj_vm_offset +
params->args_batch_start_offset;
- ret = ring->dispatch_execbuffer(params->request,
+ ret = ring->dispatch_execbuffer(req,
exec_start, exec_len,
params->dispatch_flags);
if (ret)
goto error;
- trace_i915_gem_ring_dispatch(params->request, params->dispatch_flags);
+ trace_i915_gem_ring_dispatch(req, params->dispatch_flags);
i915_gem_execbuffer_retire_commands(params);
@@ -929,7 +929,8 @@ int intel_execlists_submission(struct i915_execbuffer_params *params,
int intel_execlists_submission_final(struct i915_execbuffer_params *params)
{
struct drm_i915_private *dev_priv = params->dev->dev_private;
- struct intel_ringbuffer *ringbuf = params->request->ringbuf;
+ struct drm_i915_gem_request *req = params->request;
+ struct intel_ringbuffer *ringbuf = req->ringbuf;
struct intel_engine_cs *ring = params->ring;
u64 exec_start;
int ret;
@@ -941,13 +942,13 @@ int intel_execlists_submission_final(struct i915_execbuffer_params *params)
* Unconditionally invalidate gpu caches and ensure that we do flush
* any residual writes from the previous batch.
*/
- ret = logical_ring_invalidate_all_caches(params->request);
+ ret = logical_ring_invalidate_all_caches(req);
if (ret)
return ret;
if (ring == &dev_priv->ring[RCS] &&
params->instp_mode != dev_priv->relative_constants_mode) {
- ret = intel_logical_ring_begin(params->request, 4);
+ ret = intel_logical_ring_begin(req, 4);
if (ret)
return ret;
@@ -963,11 +964,11 @@ int intel_execlists_submission_final(struct i915_execbuffer_params *params)
exec_start = params->batch_obj_vm_offset +
params->args_batch_start_offset;
- ret = ring->emit_bb_start(params->request, exec_start, params->dispatch_flags);
+ ret = ring->emit_bb_start(req, exec_start, params->dispatch_flags);
if (ret)
return ret;
- trace_i915_gem_ring_dispatch(params->request, params->dispatch_flags);
+ trace_i915_gem_ring_dispatch(req, params->dispatch_flags);
i915_gem_execbuffer_retire_commands(params);