@@ -56,7 +56,8 @@
intel_ring_emit(LP_RING(dev_priv), x)
#define ADVANCE_LP_RING() \
- __intel_ring_advance(LP_RING(dev_priv), LP_RING(dev_priv)->default_context)
+ intel_ringbuffer_advance_and_submit(LP_RING(dev_priv), \
+ LP_RING(dev_priv)->default_context)
/**
* Lock test for when it's just for synchronization of ring access.
@@ -8609,7 +8609,7 @@ static int intel_gen2_queue_flip(struct drm_device *dev,
intel_ring_emit(ring, 0); /* aux display base address, unused */
intel_mark_page_flip_active(intel_crtc);
- __intel_ring_advance(ring, ring->default_context);
+ intel_ringbuffer_advance_and_submit(ring, ring->default_context);
return 0;
err_unpin:
@@ -8651,7 +8651,7 @@ static int intel_gen3_queue_flip(struct drm_device *dev,
intel_ring_emit(ring, MI_NOOP);
intel_mark_page_flip_active(intel_crtc);
- __intel_ring_advance(ring, ring->default_context);
+ intel_ringbuffer_advance_and_submit(ring, ring->default_context);
return 0;
err_unpin:
@@ -8700,7 +8700,7 @@ static int intel_gen4_queue_flip(struct drm_device *dev,
intel_ring_emit(ring, pf | pipesrc);
intel_mark_page_flip_active(intel_crtc);
- __intel_ring_advance(ring, ring->default_context);
+ intel_ringbuffer_advance_and_submit(ring, ring->default_context);
return 0;
err_unpin:
@@ -8745,7 +8745,7 @@ static int intel_gen6_queue_flip(struct drm_device *dev,
intel_ring_emit(ring, pf | pipesrc);
intel_mark_page_flip_active(intel_crtc);
- __intel_ring_advance(ring, ring->default_context);
+ intel_ringbuffer_advance_and_submit(ring, ring->default_context);
return 0;
err_unpin:
@@ -8855,7 +8855,7 @@ static int intel_gen7_queue_flip(struct drm_device *dev,
intel_ring_emit(ring, (MI_NOOP));
intel_mark_page_flip_active(intel_crtc);
- __intel_ring_advance(ring, ring->default_context);
+ intel_ringbuffer_advance_and_submit(ring, ring->default_context);
return 0;
err_unpin:
@@ -56,7 +56,7 @@ static bool intel_ring_stopped(struct intel_engine *ring)
return dev_priv->gpu_error.stop_rings & intel_ring_flag(ring);
}
-void __intel_ring_advance(struct intel_engine *ring,
+void intel_ringbuffer_advance_and_submit(struct intel_engine *ring,
struct i915_hw_context *ctx)
{
struct intel_ringbuffer *ringbuf = __get_ringbuf(ring);
@@ -738,7 +738,7 @@ gen6_add_request(struct intel_engine *ring,
intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
intel_ring_emit(ring, ring->outstanding_lazy_seqno);
intel_ring_emit(ring, MI_USER_INTERRUPT);
- __intel_ring_advance(ring, ctx);
+ intel_ringbuffer_advance_and_submit(ring, ctx);
return 0;
}
@@ -850,7 +850,7 @@ pc_render_add_request(struct intel_engine *ring,
intel_ring_emit(ring, ring->scratch.gtt_offset | PIPE_CONTROL_GLOBAL_GTT);
intel_ring_emit(ring, ring->outstanding_lazy_seqno);
intel_ring_emit(ring, 0);
- __intel_ring_advance(ring, ctx);
+ intel_ringbuffer_advance_and_submit(ring, ctx);
return 0;
}
@@ -1092,7 +1092,7 @@ i9xx_add_request(struct intel_engine *ring,
intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
intel_ring_emit(ring, ring->outstanding_lazy_seqno);
intel_ring_emit(ring, MI_USER_INTERRUPT);
- __intel_ring_advance(ring, ctx);
+ intel_ringbuffer_advance_and_submit(ring, ctx);
return 0;
}
@@ -1577,7 +1577,7 @@ static int ring_wait_for_space(struct intel_engine *ring,
return ret;
/* force the tail write in case we have been skipping them */
- __intel_ring_advance(ring, ctx);
+ intel_ringbuffer_advance_and_submit(ring, ctx);
trace_i915_ring_wait_begin(ring);
/* With GEM the hangcheck timer should kick us out of the loop,
@@ -313,8 +313,8 @@ static inline void intel_ring_advance(struct intel_engine *ring)
ringbuf->tail &= ringbuf->size - 1;
}
-void __intel_ring_advance(struct intel_engine *ring,
- struct i915_hw_context *ctx);
+void intel_ringbuffer_advance_and_submit(struct intel_engine *ring,
+ struct i915_hw_context *ctx);
int __must_check intel_ring_idle(struct intel_engine *ring);
void intel_ring_init_seqno(struct intel_engine *ring, u32 seqno);