Message ID | 20231006182051.1599923-4-jonathan.cavitt@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | drm/i915: Define and use GuC and CTB TLB invalidation routines | expand |
On 06/10/2023 19:20, Jonathan Cavitt wrote: > From: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> > > The GuC firmware had defined the interface for Translation Look-Aside > Buffer (TLB) invalidation. We should use this interface when > invalidating the engine and GuC TLBs. > Add additional functionality to intel_gt_invalidate_tlb, invalidating > the GuC TLBs and falling back to GT invalidation when the GuC is > disabled. > The invalidation is done by sending a request directly to the GuC > tlb_lookup that invalidates the table. The invalidation is submitted as > a wait request and is performed in the CT event handler. This means we > cannot perform this TLB invalidation path if the CT is not enabled. > If the request isn't fulfilled in two seconds, this would constitute > an error in the invalidation as that would constitute either a lost > request or a severe GuC overload. > > With this new invalidation routine, we can perform GuC-based GGTT > invalidations. GuC-based GGTT invalidation is incompatible with > MMIO invalidation so we should not perform MMIO invalidation when > GuC-based GGTT invalidation is expected. > > Purpose of xarray: > The tlb_lookup table is allocated as an xarray because the set of > pending TLB invalidations may have no upper bound. The consequence of > this is that all actions interfacing with this table need to use the > xarray functions, such as xa_alloc_cyclic_irq for array insertion. > > Purpose of must_wait_woken: > Our wait for the G2H ack for the completion of a TLB invalidation is > mandatory; we must wait for the HW to confirm that the physical > addresses are no longer accessible before we return those to the system. > > On switching to using the wait_woken() convenience routine, we > introduced ourselves to an issue where wait_woken() may complete early > under a kthread that is stopped. Since we send a TLB invalidation when > we try to release pages from the shrinker, we can be called from any > process; including kthreads. > > Using wait_woken() from any process context causes another issue. The > use of is_kthread_should_stop() assumes that any task with PF_KTHREAD > set was made by kthread_create() and has called set_kthread_struct(). > This is not true for the raw kernel_thread(): This explanation misses the main point of my ask - which is to explain why a simpler scheme isn't sufficient. Simpler scheme aka not needed the xarray or any flavour of wait_token(). In other words it is obvious we have to wait for the invalidation ack, but not obvious why we need a complicated scheme. > BUG: kernel NULL pointer dereference, address: 0000000000000000 > [ 3089.759660] Call Trace: > [ 3089.762110] wait_woken+0x4f/0x80 > [ 3089.765496] guc_send_invalidate_tlb+0x1fe/0x310 [i915] > [ 3089.770725] ? syscall_return_via_sysret+0xf/0x7f > [ 3089.775426] ? do_wait_intr_irq+0xb0/0xb0 > [ 3089.779430] ? __switch_to_asm+0x40/0x70 > [ 3089.783349] ? __switch_to_asm+0x34/0x70 > [ 3089.787273] ? __switch_to+0x7a/0x3e0 > [ 3089.790930] ? __switch_to_asm+0x34/0x70 > [ 3089.794883] intel_guc_invalidate_tlb_full+0x92/0xa0 [i915] > [ 3089.800487] intel_invalidate_tlb_full+0x94/0x190 [i915] > [ 3089.805824] intel_invalidate_tlb_full_sync+0x1b/0x30 [i915] > [ 3089.811508] __i915_gem_object_unset_pages+0x138/0x150 [i915] > [ 3089.817279] __i915_gem_object_put_pages+0x25/0x90 [i915] > [ 3089.822706] i915_gem_shrink+0x532/0x7e0 [i915] > [ 3089.827264] i915_gem_shrinker_scan+0x3d/0xd0 [i915] > [ 3089.832230] do_shrink_slab+0x12c/0x2a0 > [ 3089.836065] shrink_slab+0xad/0x2b0 > [ 3089.839550] shrink_node+0xcc/0x410 > [ 3089.843035] do_try_to_free_pages+0xc6/0x380 > [ 3089.847306] try_to_free_pages+0xec/0x1c0 > [ 3089.851312] __alloc_pages_slowpath+0x3ad/0xd10 > [ 3089.855845] ? update_sd_lb_stats+0x636/0x710 > [ 3089.860204] __alloc_pages_nodemask+0x2d5/0x310 > [ 3089.864737] new_slab+0x265/0xa80 > [ 3089.868053] ___slab_alloc+0y_to_free_pages+0xec/0x1c0 > [ 3089.871798] ? copy_process+0x1e5/0x1a00 > [ 3089.875717] ? load_balance+0x165/0xb20 > [ 3089.879555] __slab_alloc+0x1c/0x30 > [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 > [ 3089.887397] ? copy_process+0x1e5/0x1a00 > [ 3089.891314] copy_process+0x1e5/0x1a00 > [ 3089.895058] ? __switch_to_asm+0x40/0x70 > [ 3089.879555] __slab_alloc+0x1c/0x30 > [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 > [ 3089.887397] ? copy_process+0x1e5/0x1a00 > [ 3089.891314] copy_process+0x1e5/0x1a00 > [ 3089.895058] ? __switch_to_asm+0x40/0x70 > [ 3089.898977] ? __switch_to_asm+0x34/0x70 > [ 3089.902903] ? __switch_to_asm+0x40/0x70 > [ 3089.906828] ? __switch_to_asm+0x34/0x70 > [ 3089.910745] _do_fork+0x83/0x350 > [ 3089.913969] ? __switch_to+0x7a/0x3e0 > [ 3089.917626] ? __switch_to_asm+0x34/0x70 > [ 3089.921545] kernel_thread+0x58/0x80 > [ 3089.925124] ? kthread_park+0x80/0x80 > [ 3089.928788] kthreadd+0x162/0x1b0 > [ 3089.932098] ? kthread_create_on_cpu+0xa0/0xa0 > [ 3089.936538] ret_from_fork+0x1f/0x40 > > Signed-off-by: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> > Signed-off-by: Bruce Chang <yu.bruce.chang@intel.com> > Signed-off-by: Chris Wilson <chris.p.wilson@intel.com> > Signed-off-by: Umesh Nerlige Ramappa <umesh.nerlige.ramappa@intel.com> > Signed-off-by: Jonathan Cavitt <jonathan.cavitt@intel.com> > Signed-off-by: Aravind Iddamsetty <aravind.iddamsetty@intel.com> > Signed-off-by: Fei Yang <fei.yang@intel.com> > CC: Andi Shyti <andi.shyti@linux.intel.com> > --- > drivers/gpu/drm/i915/gt/intel_ggtt.c | 34 ++- > drivers/gpu/drm/i915/gt/intel_tlb.c | 15 +- > .../gpu/drm/i915/gt/uc/abi/guc_actions_abi.h | 33 +++ > drivers/gpu/drm/i915/gt/uc/intel_guc.h | 22 ++ > drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 4 + > drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h | 1 + > .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 200 +++++++++++++++++- > 7 files changed, 297 insertions(+), 12 deletions(-) > > diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c > index 4d7d88b92632b..a1f7bdc602996 100644 > --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c > +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c > @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt) > intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN); > } > > +static void guc_ggtt_ct_invalidate(struct intel_gt *gt) > +{ > + struct intel_uncore *uncore = gt->uncore; > + intel_wakeref_t wakeref; > + > + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) { > + struct intel_guc *guc = >->uc.guc; > + > + intel_guc_invalidate_tlb_guc(guc); > + } > +} > + > static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) > { > struct drm_i915_private *i915 = ggtt->vm.i915; > + struct intel_gt *gt; > > - gen8_ggtt_invalidate(ggtt); > - > - if (GRAPHICS_VER(i915) >= 12) { > - struct intel_gt *gt; > + if (!HAS_GUC_TLB_INVALIDATION(i915)) > + gen8_ggtt_invalidate(ggtt); > > - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) > + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) { > + if (HAS_GUC_TLB_INVALIDATION(i915) && > + intel_guc_is_ready(>->uc.guc)) { > + guc_ggtt_ct_invalidate(gt); > + } else if (GRAPHICS_VER(i915) >= 12) { > intel_uncore_write_fw(gt->uncore, > GEN12_GUC_TLB_INV_CR, > GEN12_GUC_TLB_INV_CR_INVALIDATE); > - } else { > - intel_uncore_write_fw(ggtt->vm.gt->uncore, > - GEN8_GTCR, GEN8_GTCR_INVALIDATE); > + } else { > + intel_uncore_write_fw(gt->uncore, > + GEN8_GTCR, GEN8_GTCR_INVALIDATE); > + } > } > } > > @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) > ggtt->vm.raw_insert_page = gen8_ggtt_insert_page; > } > > - if (intel_uc_wants_guc(&ggtt->vm.gt->uc)) > + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc)) > ggtt->invalidate = guc_ggtt_invalidate; > else > ggtt->invalidate = gen8_ggtt_invalidate; > diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c b/drivers/gpu/drm/i915/gt/intel_tlb.c > index 139608c30d978..1278622b9554d 100644 > --- a/drivers/gpu/drm/i915/gt/intel_tlb.c > +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c > @@ -12,6 +12,7 @@ > #include "intel_gt_print.h" > #include "intel_gt_regs.h" > #include "intel_tlb.h" > +#include "uc/intel_guc.h" > > /* > * HW architecture suggest typical invalidation time at 40us, > @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) > return; > > with_intel_gt_pm_if_awake(gt, wakeref) { > + struct intel_guc *guc = >->uc.guc; > + > mutex_lock(>->tlb.invalidate_lock); > if (tlb_seqno_passed(gt, seqno)) > goto unlock; > > - mmio_invalidate_full(gt); > + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { > + /* > + * Only perform GuC TLB invalidation if GuC is ready. > + * If GuC is not ready, then there are no TLBs to > + * invalidate. Ergo, skip invalidation. > + */ > + if (intel_guc_is_ready(guc)) > + intel_guc_invalidate_tlb_engines(guc); What was the answer to John's question on why it is okay to just skip and not maybe fall back to mmio? > + } else { > + mmio_invalidate_full(gt); > + } > > write_seqcount_invalidate(>->tlb.seqno); > unlock: > diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > index f359bef046e0b..33f253410d0c8 100644 > --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > @@ -138,6 +138,8 @@ enum intel_guc_action { > INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601, > INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507, > INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A, > + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000, > + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001, > INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002, > INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003, > INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004, > @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status { > > #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF > > +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0) > +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8) > +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31) > + > +enum intel_guc_tlb_invalidation_type { > + INTEL_GUC_TLB_INVAL_ENGINES = 0x0, > + INTEL_GUC_TLB_INVAL_GUC = 0x3, > +}; > + > +/* > + * 0: Heavy mode of Invalidation: > + * The pipeline of the engine(s) for which the invalidation is targeted to is > + * blocked, and all the in-flight transactions are guaranteed to be Globally > + * Observed before completing the TLB invalidation > + * 1: Lite mode of Invalidation: > + * TLBs of the targeted engine(s) are immediately invalidated. > + * In-flight transactions are NOT guaranteed to be Globally Observed before > + * completing TLB invalidation. > + * Light Invalidation Mode is to be used only when > + * it can be guaranteed (by SW) that the address translations remain invariant > + * for the in-flight transactions across the TLB invalidation. In other words, > + * this mode can be used when the TLB invalidation is intended to clear out the > + * stale cached translations that are no longer in use. Light Invalidation Mode > + * is much faster than the Heavy Invalidation Mode, as it does not wait for the > + * in-flight transactions to be GOd. > + */ > +enum intel_guc_tlb_inval_mode { > + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0, > + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1, > +}; > + > #endif /* _ABI_GUC_ACTIONS_ABI_H */ > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h > index 6c392bad29c19..88deb43bbdc48 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h > @@ -79,6 +79,18 @@ struct intel_guc { > */ > atomic_t outstanding_submission_g2h; > > + /** @tlb_lookup: xarray to store all pending TLB invalidation requests */ > + struct xarray tlb_lookup; > + > + /** > + * @serial_slot: id to the initial waiter created in tlb_lookup, > + * which is used only when failed to allocate new waiter. > + */ > + u32 serial_slot; > + > + /** @next_seqno: the next id (sequence no.) to allocate. */ > + u32 next_seqno; > + > /** @interrupts: pointers to GuC interrupt-managing functions. */ > struct { > bool enabled; > @@ -296,6 +308,11 @@ struct intel_guc { > #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, (ver).minor, (ver).patch) > #define GUC_SUBMIT_VER(guc) MAKE_GUC_VER_STRUCT((guc)->submission_version) > > +struct intel_guc_tlb_wait { > + struct wait_queue_head wq; > + bool busy; > +}; > + > static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) > { > return container_of(log, struct intel_guc, log); > @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct intel_guc *guc) > return intel_uc_fw_is_supported(&guc->fw); > } > > +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc); > +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc); > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, > + u32 len, u32 seqno); > + > static inline bool intel_guc_is_wanted(struct intel_guc *guc) > { > return intel_uc_fw_is_enabled(&guc->fw); > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > index 6e22af31513a5..e7ac8402245b7 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > @@ -1115,6 +1115,9 @@ static int ct_process_request(struct intel_guc_ct *ct, struct ct_incoming_msg *r > case INTEL_GUC_ACTION_NOTIFY_EXCEPTION: > ret = intel_guc_crash_process_msg(guc, action); > break; > + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: > + ret = intel_guc_tlb_invalidation_done(guc, request->size, len, payload[0]); > + break; > default: > ret = -EOPNOTSUPP; > break; > @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct *ct, struct ct_incoming_msg *requ > switch (action) { > case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE: > case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE: > + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: > g2h_release_space(ct, request->size); > } > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > index b4d56eccfb1f0..a7c9874e122a3 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > @@ -22,6 +22,7 @@ > /* Payload length only i.e. don't include G2H header length */ > #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2 > #define G2H_LEN_DW_DEREGISTER_CONTEXT 1 > +#define G2H_LEN_DW_INVALIDATE_TLB 1 > > #define GUC_CONTEXT_DISABLE 0 > #define GUC_CONTEXT_ENABLE 1 > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > index 2cce5ec1ff00d..0c5ccda1b3e87 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct intel_context *ce, intel_engine_mask_t st > > void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) > { > + struct intel_guc_tlb_wait *wait; > struct intel_context *ce; > unsigned long index; > unsigned long flags; > + unsigned long i; > > if (unlikely(!guc_submission_initialized(guc))) { > /* Reset called during driver load? GuC not yet initialised! */ > @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stall > > /* GuC is blown away, drop all references to contexts */ > xa_destroy(&guc->context_lookup); > + > + /* > + * The full GT reset will have cleared the TLB caches and flushed the > + * G2H message queue; we can release all the blocked waiters. > + * > + * This is safe to do unlocked because the xarray is not dependent > + * on the GT reset, and there's a separate execution path for TLB > + * invalidations on GT reset, and there's a large window of time > + * between the GT reset and GuC becoming available. > + */ > + xa_for_each(&guc->tlb_lookup, i, wait) > + wake_up(&wait->wq); If you are confident there can be no failures to wake up someone, who maybe just added themselves to the xarray (via put pages for instance), while reset in ongoing. Or even removed themselves after say timing out the wait and so freed their entry... > } > > static void guc_cancel_context_requests(struct intel_context *ce) > @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) > static void destroyed_worker_func(struct work_struct *w); > static void reset_fail_worker_func(struct work_struct *w); > > +static int init_tlb_lookup(struct intel_guc *guc) > +{ > + struct intel_guc_tlb_wait *wait; > + int err; > + > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > + return 0; > + > + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); > + > + wait = kzalloc(sizeof(*wait), GFP_KERNEL); > + if (!wait) > + return -ENOMEM; > + > + init_waitqueue_head(&wait->wq); > + > + /* Preallocate a shared id for use under memory pressure. */ > + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, > + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); > + /* Only error if out of memory, not when busy (list full)*/ > + if (err == -ENOMEM) { > + kfree(wait); > + return err; > + } I agreed with John here that only looking at ENOMEM reads odd and I did not see that answered. Did I miss it? Otherwise, I _know_ it is not likely to get any other error having *just* created a new xarray, but still, why not simply catch all? It is not like the slot fallback code at runtime would handle guc->serial_slot being empty?! It appears it would just explode in guc_send_invalidate_tlb if it would hit it.. > + > + return 0; > +} > + > +static void fini_tlb_lookup(struct intel_guc *guc) > +{ > + struct intel_guc_tlb_wait *wait; > + > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > + return; > + > + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); > + if (wait) { > + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); Hm wait, why is this unexpected when init_tlb_lookup() pre-allocated that entry? Who frees it? guc_send_invalidate_tlb() does not appear to. > + kfree(wait); > + } > + > + xa_destroy(&guc->tlb_lookup); > +} > + > /* > * Set up the memory resources to be shared with the GuC (via the GGTT) > * at firmware loading time. > @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct intel_guc *guc) > return ret; > } > > + ret = init_tlb_lookup(guc); > + if (ret) > + goto destroy_pool; > + > guc->submission_state.guc_ids_bitmap = > bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL); > if (!guc->submission_state.guc_ids_bitmap) { > ret = -ENOMEM; > - goto destroy_pool; > + goto destroy_tlb; > } > > guc->timestamp.ping_delay = (POLL_TIME_CLKS / gt->clock_frequency + 1) * HZ; > @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc *guc) > > return 0; > > +destroy_tlb: > + fini_tlb_lookup(guc); > destroy_pool: > guc_lrc_desc_pool_destroy_v69(guc); > - > return ret; > } > > @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct intel_guc *guc) > guc_lrc_desc_pool_destroy_v69(guc); > i915_sched_engine_put(guc->sched_engine); > bitmap_free(guc->submission_state.guc_ids_bitmap); > + fini_tlb_lookup(guc); > guc->submission_initialized = false; > } > > @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) > return ce; > } > > +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) > +{ > + struct intel_guc_tlb_wait *wait; > + unsigned long flags; > + > + xa_lock_irqsave(&guc->tlb_lookup, flags); > + wait = xa_load(&guc->tlb_lookup, seqno); > + > + if (wait) > + wake_up(&wait->wq); > + else > + guc_dbg(guc, > + "Stale TLB invalidation response with seqno %d\n", seqno); > + > + xa_unlock_irqrestore(&guc->tlb_lookup, flags); > +} > + > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) > +{ > + /* Check for underflow */ > + if (unlikely(len < 1 || len > size)) > + return -EPROTO; These check are not valid for any message/action type ct_process_request() can receive? > + > + wait_wake_outstanding_tlb_g2h(guc, seqno); > + return 0; > +} > + > +static long must_wait_woken(struct wait_queue_entry *wq_entry, long timeout) > +{ > + /* > + * This is equivalent to wait_woken() with the exception that > + * we do not wake up early if the kthread task has been completed. > + * As we are called from page reclaim in any task context, > + * we may be invoked from stopped kthreads, but we *must* > + * complete the wait from the HW . > + * > + * A second problem is that since we are called under reclaim > + * and wait_woken() inspected the thread state, it makes an invalid > + * assumption that all PF_KTHREAD tasks have set_kthread_struct() > + * called upon them, and will trigger a GPF in is_kthread_should_stop(). > + */ > + do { > + set_current_state(TASK_UNINTERRUPTIBLE); > + if (wq_entry->flags & WQ_FLAG_WOKEN) > + break; > + > + timeout = schedule_timeout(timeout); > + } while (timeout); > + __set_current_state(TASK_RUNNING); > + > + /* See wait_woken() and woken_wake_function() */ > + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); > + > + return timeout; > +} > + > +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_inval_mode type) > +{ > + struct intel_guc_tlb_wait _wq, *wq = &_wq; > + DEFINE_WAIT_FUNC(wait, woken_wake_function); > + int err; > + u32 seqno; > + u32 action[] = { > + INTEL_GUC_ACTION_TLB_INVALIDATION, > + 0, > + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | > + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, > + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | > + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, > + }; > + u32 size = ARRAY_SIZE(action); > + > + init_waitqueue_head(&_wq.wq); > + > + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, > + xa_limit_32b, &guc->next_seqno, > + GFP_ATOMIC | __GFP_NOWARN) < 0) { > + /* Under severe memory pressure? Serialise TLB allocations */ > + xa_lock_irq(&guc->tlb_lookup); > + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); > + wait_event_lock_irq(wq->wq, > + !READ_ONCE(wq->busy), > + guc->tlb_lookup.xa_lock); > + /* > + * Update wq->busy under lock to ensure only one waiter can > + * issue the TLB invalidation command using the serial slot at a > + * time. The condition is set to true before releasing the lock > + * so that other caller continue to wait until woken up again. > + */ > + wq->busy = true; > + xa_unlock_irq(&guc->tlb_lookup); > + > + seqno = guc->serial_slot; > + } > + > + action[1] = seqno; > + > + add_wait_queue(&wq->wq, &wait); > + > + /* > + * This is a critical reclaim path and thus we must loop here: > + * We cannot block for anything that is on the GPU. > + */ > + err = intel_guc_send_busy_loop(guc, action, size, G2H_LEN_DW_INVALIDATE_TLB, true); > + if (err) > + goto out; > + > + if (!must_wait_woken(&wait, intel_guc_ct_expected_delay(&guc->ct))) { > + guc_err(guc, > + "TLB invalidation response timed out for seqno %u\n", seqno); > + err = -ETIME; > + } > +out: > + remove_wait_queue(&wq->wq, &wait); > + if (seqno != guc->serial_slot) > + xa_erase_irq(&guc->tlb_lookup, seqno); > + > + return err; > +} > + > +/* Full TLB invalidation */ > +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) > +{ > + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); > +} > + > +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ > +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) > +{ > + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); > +} > + > int intel_guc_deregister_done_process_msg(struct intel_guc *guc, > const u32 *msg, > u32 len) Regards, Tvrtko
Hi, ... > > @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) > > return; > > with_intel_gt_pm_if_awake(gt, wakeref) { > > + struct intel_guc *guc = >->uc.guc; > > + > > mutex_lock(>->tlb.invalidate_lock); > > if (tlb_seqno_passed(gt, seqno)) > > goto unlock; > > - mmio_invalidate_full(gt); > > + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { > > + /* > > + * Only perform GuC TLB invalidation if GuC is ready. > > + * If GuC is not ready, then there are no TLBs to > > + * invalidate. Ergo, skip invalidation. > > + */ > > + if (intel_guc_is_ready(guc)) > > + intel_guc_invalidate_tlb_engines(guc); > > What was the answer to John's question on why it is okay to just skip and > not maybe fall back to mmio? this maybe can be written as: if (HAS_GUC_TLB_INVALIDATION(gt->i915) && intel_guc_is_ready(guc)) intel_guc_invalidate_tlb_engines(guc); else mmio_invalidate_full(gt); > > + } else { > > + mmio_invalidate_full(gt); > > + } > > write_seqcount_invalidate(>->tlb.seqno); > > unlock: ... > > + /* > > + * The full GT reset will have cleared the TLB caches and flushed the > > + * G2H message queue; we can release all the blocked waiters. > > + * > > + * This is safe to do unlocked because the xarray is not dependent > > + * on the GT reset, and there's a separate execution path for TLB > > + * invalidations on GT reset, and there's a large window of time > > + * between the GT reset and GuC becoming available. > > + */ > > + xa_for_each(&guc->tlb_lookup, i, wait) > > + wake_up(&wait->wq); > > If you are confident there can be no failures to wake up someone, who maybe > just added themselves to the xarray (via put pages for instance), while > reset in ongoing. Or even removed themselves after say timing out the wait > and so freed their entry... I guess yuo are suggesting here to spinlock around this. The reset is protected by the uncore->spinlock, but I don't really see it colliding with reset, to be honest. > > } > > static void guc_cancel_context_requests(struct intel_context *ce) > > @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) > > static void destroyed_worker_func(struct work_struct *w); > > static void reset_fail_worker_func(struct work_struct *w); > > +static int init_tlb_lookup(struct intel_guc *guc) > > +{ > > + struct intel_guc_tlb_wait *wait; > > + int err; > > + > > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > > + return 0; > > + > > + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); > > + > > + wait = kzalloc(sizeof(*wait), GFP_KERNEL); > > + if (!wait) > > + return -ENOMEM; > > + > > + init_waitqueue_head(&wait->wq); > > + > > + /* Preallocate a shared id for use under memory pressure. */ > > + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, > > + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); > > + /* Only error if out of memory, not when busy (list full)*/ > > + if (err == -ENOMEM) { > > + kfree(wait); > > + return err; > > + } > > I agreed with John here that only looking at ENOMEM reads odd and I did not > see that answered. Did I miss it? xa_alloc_cyclic_irq() can also fail with -EBUSY... so that I think this is a matter... > Otherwise, I _know_ it is not likely to get any other error having *just* > created a new xarray, but still, why not simply catch all? It is not like > the slot fallback code at runtime would handle guc->serial_slot being > empty?! It appears it would just explode in guc_send_invalidate_tlb if it > would hit it.. ... of what we want to do for such errors. E.g. Jonathan decided here not to fail, but ignore the error. Should we fail for every error? > > + > > + return 0; > > +} > > + > > +static void fini_tlb_lookup(struct intel_guc *guc) > > +{ > > + struct intel_guc_tlb_wait *wait; > > + > > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > > + return; > > + > > + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); > > + if (wait) { > > + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); > > Hm wait, why is this unexpected when init_tlb_lookup() pre-allocated that > entry? Who frees it? guc_send_invalidate_tlb() does not appear to. I think this links with my answer above, right? With th refactoring of the if's for not skipping tlb invalidation. > > + kfree(wait); > > + } > > + > > + xa_destroy(&guc->tlb_lookup); > > +} > > + > > /* > > * Set up the memory resources to be shared with the GuC (via the GGTT) > > * at firmware loading time. ... > > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) > > +{ > > + /* Check for underflow */ > > + if (unlikely(len < 1 || len > size)) > > + return -EPROTO; > > These check are not valid for any message/action type ct_process_request() > can receive? You mean discriminating by payload? Jonathan... you konw the details here?
On 10/6/2023 8:20 PM, Jonathan Cavitt wrote: > From: Prathap Kumar Valsan<prathap.kumar.valsan@intel.com> > > The GuC firmware had defined the interface for Translation Look-Aside > Buffer (TLB) invalidation. We should use this interface when > invalidating the engine and GuC TLBs. > Add additional functionality to intel_gt_invalidate_tlb, invalidating > the GuC TLBs and falling back to GT invalidation when the GuC is > disabled. > The invalidation is done by sending a request directly to the GuC > tlb_lookup that invalidates the table. The invalidation is submitted as > a wait request and is performed in the CT event handler. This means we > cannot perform this TLB invalidation path if the CT is not enabled. > If the request isn't fulfilled in two seconds, this would constitute > an error in the invalidation as that would constitute either a lost > request or a severe GuC overload. > > With this new invalidation routine, we can perform GuC-based GGTT > invalidations. GuC-based GGTT invalidation is incompatible with > MMIO invalidation so we should not perform MMIO invalidation when > GuC-based GGTT invalidation is expected. > > Purpose of xarray: > The tlb_lookup table is allocated as an xarray because the set of > pending TLB invalidations may have no upper bound. The consequence of > this is that all actions interfacing with this table need to use the > xarray functions, such as xa_alloc_cyclic_irq for array insertion. > > Purpose of must_wait_woken: > Our wait for the G2H ack for the completion of a TLB invalidation is > mandatory; we must wait for the HW to confirm that the physical > addresses are no longer accessible before we return those to the system. > > On switching to using the wait_woken() convenience routine, we > introduced ourselves to an issue where wait_woken() may complete early > under a kthread that is stopped. Since we send a TLB invalidation when > we try to release pages from the shrinker, we can be called from any > process; including kthreads. > > Using wait_woken() from any process context causes another issue. The > use of is_kthread_should_stop() assumes that any task with PF_KTHREAD > set was made by kthread_create() and has called set_kthread_struct(). > This is not true for the raw kernel_thread(): > > BUG: kernel NULL pointer dereference, address: 0000000000000000 > [ 3089.759660] Call Trace: > [ 3089.762110] wait_woken+0x4f/0x80 > [ 3089.765496] guc_send_invalidate_tlb+0x1fe/0x310 [i915] > [ 3089.770725] ? syscall_return_via_sysret+0xf/0x7f > [ 3089.775426] ? do_wait_intr_irq+0xb0/0xb0 > [ 3089.779430] ? __switch_to_asm+0x40/0x70 > [ 3089.783349] ? __switch_to_asm+0x34/0x70 > [ 3089.787273] ? __switch_to+0x7a/0x3e0 > [ 3089.790930] ? __switch_to_asm+0x34/0x70 > [ 3089.794883] intel_guc_invalidate_tlb_full+0x92/0xa0 [i915] > [ 3089.800487] intel_invalidate_tlb_full+0x94/0x190 [i915] > [ 3089.805824] intel_invalidate_tlb_full_sync+0x1b/0x30 [i915] > [ 3089.811508] __i915_gem_object_unset_pages+0x138/0x150 [i915] > [ 3089.817279] __i915_gem_object_put_pages+0x25/0x90 [i915] > [ 3089.822706] i915_gem_shrink+0x532/0x7e0 [i915] > [ 3089.827264] i915_gem_shrinker_scan+0x3d/0xd0 [i915] > [ 3089.832230] do_shrink_slab+0x12c/0x2a0 > [ 3089.836065] shrink_slab+0xad/0x2b0 > [ 3089.839550] shrink_node+0xcc/0x410 > [ 3089.843035] do_try_to_free_pages+0xc6/0x380 > [ 3089.847306] try_to_free_pages+0xec/0x1c0 > [ 3089.851312] __alloc_pages_slowpath+0x3ad/0xd10 > [ 3089.855845] ? update_sd_lb_stats+0x636/0x710 > [ 3089.860204] __alloc_pages_nodemask+0x2d5/0x310 > [ 3089.864737] new_slab+0x265/0xa80 > [ 3089.868053] ___slab_alloc+0y_to_free_pages+0xec/0x1c0 > [ 3089.871798] ? copy_process+0x1e5/0x1a00 > [ 3089.875717] ? load_balance+0x165/0xb20 > [ 3089.879555] __slab_alloc+0x1c/0x30 > [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 > [ 3089.887397] ? copy_process+0x1e5/0x1a00 > [ 3089.891314] copy_process+0x1e5/0x1a00 > [ 3089.895058] ? __switch_to_asm+0x40/0x70 > [ 3089.879555] __slab_alloc+0x1c/0x30 > [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 > [ 3089.887397] ? copy_process+0x1e5/0x1a00 > [ 3089.891314] copy_process+0x1e5/0x1a00 > [ 3089.895058] ? __switch_to_asm+0x40/0x70 > [ 3089.898977] ? __switch_to_asm+0x34/0x70 > [ 3089.902903] ? __switch_to_asm+0x40/0x70 > [ 3089.906828] ? __switch_to_asm+0x34/0x70 > [ 3089.910745] _do_fork+0x83/0x350 > [ 3089.913969] ? __switch_to+0x7a/0x3e0 > [ 3089.917626] ? __switch_to_asm+0x34/0x70 > [ 3089.921545] kernel_thread+0x58/0x80 > [ 3089.925124] ? kthread_park+0x80/0x80 > [ 3089.928788] kthreadd+0x162/0x1b0 > [ 3089.932098] ? kthread_create_on_cpu+0xa0/0xa0 > [ 3089.936538] ret_from_fork+0x1f/0x40 > > Signed-off-by: Prathap Kumar Valsan<prathap.kumar.valsan@intel.com> > Signed-off-by: Bruce Chang<yu.bruce.chang@intel.com> > Signed-off-by: Chris Wilson<chris.p.wilson@intel.com> > Signed-off-by: Umesh Nerlige Ramappa<umesh.nerlige.ramappa@intel.com> > Signed-off-by: Jonathan Cavitt<jonathan.cavitt@intel.com> > Signed-off-by: Aravind Iddamsetty<aravind.iddamsetty@intel.com> > Signed-off-by: Fei Yang<fei.yang@intel.com> > CC: Andi Shyti<andi.shyti@linux.intel.com> > --- > drivers/gpu/drm/i915/gt/intel_ggtt.c | 34 ++- > drivers/gpu/drm/i915/gt/intel_tlb.c | 15 +- > .../gpu/drm/i915/gt/uc/abi/guc_actions_abi.h | 33 +++ > drivers/gpu/drm/i915/gt/uc/intel_guc.h | 22 ++ > drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 4 + > drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h | 1 + > .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 200 +++++++++++++++++- > 7 files changed, 297 insertions(+), 12 deletions(-) > > diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c > index 4d7d88b92632b..a1f7bdc602996 100644 > --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c > +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c > @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt) > intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN); > } > > +static void guc_ggtt_ct_invalidate(struct intel_gt *gt) > +{ > + struct intel_uncore *uncore = gt->uncore; > + intel_wakeref_t wakeref; > + > + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) { > + struct intel_guc *guc = >->uc.guc; > + > + intel_guc_invalidate_tlb_guc(guc); > + } > +} > + > static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) > { > struct drm_i915_private *i915 = ggtt->vm.i915; > + struct intel_gt *gt; > > - gen8_ggtt_invalidate(ggtt); > - > - if (GRAPHICS_VER(i915) >= 12) { > - struct intel_gt *gt; > + if (!HAS_GUC_TLB_INVALIDATION(i915)) > + gen8_ggtt_invalidate(ggtt); > > - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) > + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) { > + if (HAS_GUC_TLB_INVALIDATION(i915) && > + intel_guc_is_ready(>->uc.guc)) { > + guc_ggtt_ct_invalidate(gt); > + } else if (GRAPHICS_VER(i915) >= 12) { > intel_uncore_write_fw(gt->uncore, > GEN12_GUC_TLB_INV_CR, > GEN12_GUC_TLB_INV_CR_INVALIDATE); > - } else { > - intel_uncore_write_fw(ggtt->vm.gt->uncore, > - GEN8_GTCR, GEN8_GTCR_INVALIDATE); > + } else { > + intel_uncore_write_fw(gt->uncore, > + GEN8_GTCR, GEN8_GTCR_INVALIDATE); > + } > } > } > > @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) > ggtt->vm.raw_insert_page = gen8_ggtt_insert_page; > } > > - if (intel_uc_wants_guc(&ggtt->vm.gt->uc)) > + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc)) > ggtt->invalidate = guc_ggtt_invalidate; > else > ggtt->invalidate = gen8_ggtt_invalidate; > diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c b/drivers/gpu/drm/i915/gt/intel_tlb.c > index 139608c30d978..1278622b9554d 100644 > --- a/drivers/gpu/drm/i915/gt/intel_tlb.c > +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c > @@ -12,6 +12,7 @@ > #include "intel_gt_print.h" > #include "intel_gt_regs.h" > #include "intel_tlb.h" > +#include "uc/intel_guc.h" > > /* > * HW architecture suggest typical invalidation time at 40us, > @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) > return; > > with_intel_gt_pm_if_awake(gt, wakeref) { > + struct intel_guc *guc = >->uc.guc; > + > mutex_lock(>->tlb.invalidate_lock); > if (tlb_seqno_passed(gt, seqno)) > goto unlock; > > - mmio_invalidate_full(gt); > + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { > + /* > + * Only perform GuC TLB invalidation if GuC is ready. > + * If GuC is not ready, then there are no TLBs to > + * invalidate. Ergo, skip invalidation. > + */ > + if (intel_guc_is_ready(guc)) > + intel_guc_invalidate_tlb_engines(guc); > + } else { > + mmio_invalidate_full(gt); > + } > > write_seqcount_invalidate(>->tlb.seqno); > unlock: > diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > index f359bef046e0b..33f253410d0c8 100644 > --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > @@ -138,6 +138,8 @@ enum intel_guc_action { > INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601, > INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507, > INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A, > + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000, > + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001, > INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002, > INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003, > INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004, > @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status { > > #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF > > +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0) > +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8) > +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31) > + > +enum intel_guc_tlb_invalidation_type { > + INTEL_GUC_TLB_INVAL_ENGINES = 0x0, > + INTEL_GUC_TLB_INVAL_GUC = 0x3, > +}; > + > +/* > + * 0: Heavy mode of Invalidation: > + * The pipeline of the engine(s) for which the invalidation is targeted to is > + * blocked, and all the in-flight transactions are guaranteed to be Globally > + * Observed before completing the TLB invalidation > + * 1: Lite mode of Invalidation: > + * TLBs of the targeted engine(s) are immediately invalidated. > + * In-flight transactions are NOT guaranteed to be Globally Observed before > + * completing TLB invalidation. > + * Light Invalidation Mode is to be used only when > + * it can be guaranteed (by SW) that the address translations remain invariant > + * for the in-flight transactions across the TLB invalidation. In other words, > + * this mode can be used when the TLB invalidation is intended to clear out the > + * stale cached translations that are no longer in use. Light Invalidation Mode > + * is much faster than the Heavy Invalidation Mode, as it does not wait for the > + * in-flight transactions to be GOd. > + */ > +enum intel_guc_tlb_inval_mode { > + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0, > + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1, > +}; > + > #endif /* _ABI_GUC_ACTIONS_ABI_H */ > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h > index 6c392bad29c19..88deb43bbdc48 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h > @@ -79,6 +79,18 @@ struct intel_guc { > */ > atomic_t outstanding_submission_g2h; > > + /** @tlb_lookup: xarray to store all pending TLB invalidation requests */ > + struct xarray tlb_lookup; > + > + /** > + * @serial_slot: id to the initial waiter created in tlb_lookup, > + * which is used only when failed to allocate new waiter. > + */ > + u32 serial_slot; > + > + /** @next_seqno: the next id (sequence no.) to allocate. */ > + u32 next_seqno; > + > /** @interrupts: pointers to GuC interrupt-managing functions. */ > struct { > bool enabled; > @@ -296,6 +308,11 @@ struct intel_guc { > #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, (ver).minor, (ver).patch) > #define GUC_SUBMIT_VER(guc) MAKE_GUC_VER_STRUCT((guc)->submission_version) > > +struct intel_guc_tlb_wait { > + struct wait_queue_head wq; > + bool busy; > +}; > + > static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) > { > return container_of(log, struct intel_guc, log); > @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct intel_guc *guc) > return intel_uc_fw_is_supported(&guc->fw); > } > > +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc); > +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc); > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, > + u32 len, u32 seqno); > + > static inline bool intel_guc_is_wanted(struct intel_guc *guc) > { > return intel_uc_fw_is_enabled(&guc->fw); > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > index 6e22af31513a5..e7ac8402245b7 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > @@ -1115,6 +1115,9 @@ static int ct_process_request(struct intel_guc_ct *ct, struct ct_incoming_msg *r > case INTEL_GUC_ACTION_NOTIFY_EXCEPTION: > ret = intel_guc_crash_process_msg(guc, action); > break; > + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: > + ret = intel_guc_tlb_invalidation_done(guc, request->size, len, payload[0]); > + break; > default: > ret = -EOPNOTSUPP; > break; > @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct *ct, struct ct_incoming_msg *requ > switch (action) { > case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE: > case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE: > + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: > g2h_release_space(ct, request->size); > } > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > index b4d56eccfb1f0..a7c9874e122a3 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > @@ -22,6 +22,7 @@ > /* Payload length only i.e. don't include G2H header length */ > #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2 > #define G2H_LEN_DW_DEREGISTER_CONTEXT 1 > +#define G2H_LEN_DW_INVALIDATE_TLB 1 > > #define GUC_CONTEXT_DISABLE 0 > #define GUC_CONTEXT_ENABLE 1 > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > index 2cce5ec1ff00d..0c5ccda1b3e87 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct intel_context *ce, intel_engine_mask_t st > > void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) > { > + struct intel_guc_tlb_wait *wait; > struct intel_context *ce; > unsigned long index; > unsigned long flags; > + unsigned long i; > > if (unlikely(!guc_submission_initialized(guc))) { > /* Reset called during driver load? GuC not yet initialised! */ > @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stall > > /* GuC is blown away, drop all references to contexts */ > xa_destroy(&guc->context_lookup); > + > + /* > + * The full GT reset will have cleared the TLB caches and flushed the > + * G2H message queue; we can release all the blocked waiters. > + * > + * This is safe to do unlocked because the xarray is not dependent > + * on the GT reset, and there's a separate execution path for TLB > + * invalidations on GT reset, and there's a large window of time > + * between the GT reset and GuC becoming available. > + */ > + xa_for_each(&guc->tlb_lookup, i, wait) > + wake_up(&wait->wq); > } > > static void guc_cancel_context_requests(struct intel_context *ce) > @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) > static void destroyed_worker_func(struct work_struct *w); > static void reset_fail_worker_func(struct work_struct *w); > > +static int init_tlb_lookup(struct intel_guc *guc) > +{ > + struct intel_guc_tlb_wait *wait; > + int err; > + > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > + return 0; > + > + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); > + > + wait = kzalloc(sizeof(*wait), GFP_KERNEL); > + if (!wait) > + return -ENOMEM; > + > + init_waitqueue_head(&wait->wq); > + > + /* Preallocate a shared id for use under memory pressure. */ > + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, > + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); > + /* Only error if out of memory, not when busy (list full)*/ > + if (err == -ENOMEM) { > + kfree(wait); > + return err; > + } > + > + return 0; > +} > + > +static void fini_tlb_lookup(struct intel_guc *guc) > +{ > + struct intel_guc_tlb_wait *wait; > + > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > + return; > + > + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); > + if (wait) { > + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); > + kfree(wait); > + } > + > + xa_destroy(&guc->tlb_lookup); > +} > + > /* > * Set up the memory resources to be shared with the GuC (via the GGTT) > * at firmware loading time. > @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct intel_guc *guc) > return ret; > } > > + ret = init_tlb_lookup(guc); > + if (ret) > + goto destroy_pool; > + > guc->submission_state.guc_ids_bitmap = > bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL); > if (!guc->submission_state.guc_ids_bitmap) { > ret = -ENOMEM; > - goto destroy_pool; > + goto destroy_tlb; > } > > guc->timestamp.ping_delay = (POLL_TIME_CLKS / gt->clock_frequency + 1) * HZ; > @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc *guc) > > return 0; > > +destroy_tlb: > + fini_tlb_lookup(guc); > destroy_pool: > guc_lrc_desc_pool_destroy_v69(guc); > - > return ret; > } > > @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct intel_guc *guc) > guc_lrc_desc_pool_destroy_v69(guc); > i915_sched_engine_put(guc->sched_engine); > bitmap_free(guc->submission_state.guc_ids_bitmap); > + fini_tlb_lookup(guc); > guc->submission_initialized = false; > } > > @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) > return ce; > } > > +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) > +{ > + struct intel_guc_tlb_wait *wait; > + unsigned long flags; > + > + xa_lock_irqsave(&guc->tlb_lookup, flags); > + wait = xa_load(&guc->tlb_lookup, seqno); > + > + if (wait) > + wake_up(&wait->wq); > + else > + guc_dbg(guc, > + "Stale TLB invalidation response with seqno %d\n", seqno); > + > + xa_unlock_irqrestore(&guc->tlb_lookup, flags); > +} > + > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) > +{ > + /* Check for underflow */ > + if (unlikely(len < 1 || len > size)) > + return -EPROTO; > + > + wait_wake_outstanding_tlb_g2h(guc, seqno); > + return 0; > +} > + > +static long must_wait_woken(struct wait_queue_entry *wq_entry, long timeout) > +{ > + /* > + * This is equivalent to wait_woken() with the exception that > + * we do not wake up early if the kthread task has been completed. > + * As we are called from page reclaim in any task context, > + * we may be invoked from stopped kthreads, but we *must* > + * complete the wait from the HW . > + * > + * A second problem is that since we are called under reclaim > + * and wait_woken() inspected the thread state, it makes an invalid > + * assumption that all PF_KTHREAD tasks have set_kthread_struct() > + * called upon them, and will trigger a GPF As discussed internally, the GPF issue is resolved with https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/ <https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/> > in is_kthread_should_stop(). > + */ > + do { > + set_current_state(TASK_UNINTERRUPTIBLE); > + if (wq_entry->flags & WQ_FLAG_WOKEN) > + break; > + > + timeout = schedule_timeout(timeout); > + } while (timeout); > + __set_current_state(TASK_RUNNING); > + > + /* See wait_woken() and woken_wake_function() */ > + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); > + > + return timeout; > +} > + > +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_inval_mode type) 2nd param should be intel_guc_tlb_invalidation_type not intel_guc_tlb_inval_mod. Not sure why didn't CI complained. Regards, Nirmoy > +{ > + struct intel_guc_tlb_wait _wq, *wq = &_wq; > + DEFINE_WAIT_FUNC(wait, woken_wake_function); > + int err; > + u32 seqno; > + u32 action[] = { > + INTEL_GUC_ACTION_TLB_INVALIDATION, > + 0, > + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | > + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, > + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | > + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, > + }; > + u32 size = ARRAY_SIZE(action); > + > + init_waitqueue_head(&_wq.wq); > + > + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, > + xa_limit_32b, &guc->next_seqno, > + GFP_ATOMIC | __GFP_NOWARN) < 0) { > + /* Under severe memory pressure? Serialise TLB allocations */ > + xa_lock_irq(&guc->tlb_lookup); > + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); > + wait_event_lock_irq(wq->wq, > + !READ_ONCE(wq->busy), > + guc->tlb_lookup.xa_lock); > + /* > + * Update wq->busy under lock to ensure only one waiter can > + * issue the TLB invalidation command using the serial slot at a > + * time. The condition is set to true before releasing the lock > + * so that other caller continue to wait until woken up again. > + */ > + wq->busy = true; > + xa_unlock_irq(&guc->tlb_lookup); > + > + seqno = guc->serial_slot; > + } > + > + action[1] = seqno; > + > + add_wait_queue(&wq->wq, &wait); > + > + /* > + * This is a critical reclaim path and thus we must loop here: > + * We cannot block for anything that is on the GPU. > + */ > + err = intel_guc_send_busy_loop(guc, action, size, G2H_LEN_DW_INVALIDATE_TLB, true); > + if (err) > + goto out; > + > + if (!must_wait_woken(&wait, intel_guc_ct_expected_delay(&guc->ct))) { > + guc_err(guc, > + "TLB invalidation response timed out for seqno %u\n", seqno); > + err = -ETIME; > + } > +out: > + remove_wait_queue(&wq->wq, &wait); > + if (seqno != guc->serial_slot) > + xa_erase_irq(&guc->tlb_lookup, seqno); > + > + return err; > +} > + > +/* Full TLB invalidation */ > +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) > +{ > + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); > +} > + > +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ > +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) > +{ > + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); > +} > + > int intel_guc_deregister_done_process_msg(struct intel_guc *guc, > const u32 *msg, > u32 len)
On 09/10/2023 12:40, Andi Shyti wrote: > Hi, > > ... > >>> @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) >>> return; >>> with_intel_gt_pm_if_awake(gt, wakeref) { >>> + struct intel_guc *guc = >->uc.guc; >>> + >>> mutex_lock(>->tlb.invalidate_lock); >>> if (tlb_seqno_passed(gt, seqno)) >>> goto unlock; >>> - mmio_invalidate_full(gt); >>> + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { >>> + /* >>> + * Only perform GuC TLB invalidation if GuC is ready. >>> + * If GuC is not ready, then there are no TLBs to >>> + * invalidate. Ergo, skip invalidation. >>> + */ >>> + if (intel_guc_is_ready(guc)) >>> + intel_guc_invalidate_tlb_engines(guc); >> >> What was the answer to John's question on why it is okay to just skip and >> not maybe fall back to mmio? > > this maybe can be written as: > > if (HAS_GUC_TLB_INVALIDATION(gt->i915) && > intel_guc_is_ready(guc)) > intel_guc_invalidate_tlb_engines(guc); > else > mmio_invalidate_full(gt); > >>> + } else { >>> + mmio_invalidate_full(gt); >>> + } >>> write_seqcount_invalidate(>->tlb.seqno); >>> unlock: > > ... > >>> + /* >>> + * The full GT reset will have cleared the TLB caches and flushed the >>> + * G2H message queue; we can release all the blocked waiters. >>> + * >>> + * This is safe to do unlocked because the xarray is not dependent >>> + * on the GT reset, and there's a separate execution path for TLB >>> + * invalidations on GT reset, and there's a large window of time >>> + * between the GT reset and GuC becoming available. >>> + */ >>> + xa_for_each(&guc->tlb_lookup, i, wait) >>> + wake_up(&wait->wq); >> >> If you are confident there can be no failures to wake up someone, who maybe >> just added themselves to the xarray (via put pages for instance), while >> reset in ongoing. Or even removed themselves after say timing out the wait >> and so freed their entry... > > I guess yuo are suggesting here to spinlock around this. The > reset is protected by the uncore->spinlock, but I don't really > see it colliding with reset, to be honest. I am not suggesting since I don't know the flows well enough, just asking. If you are confident what I wrote is impossible then okay. >>> } >>> static void guc_cancel_context_requests(struct intel_context *ce) >>> @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) >>> static void destroyed_worker_func(struct work_struct *w); >>> static void reset_fail_worker_func(struct work_struct *w); >>> +static int init_tlb_lookup(struct intel_guc *guc) >>> +{ >>> + struct intel_guc_tlb_wait *wait; >>> + int err; >>> + >>> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) >>> + return 0; >>> + >>> + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); >>> + >>> + wait = kzalloc(sizeof(*wait), GFP_KERNEL); >>> + if (!wait) >>> + return -ENOMEM; >>> + >>> + init_waitqueue_head(&wait->wq); >>> + >>> + /* Preallocate a shared id for use under memory pressure. */ >>> + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, >>> + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); >>> + /* Only error if out of memory, not when busy (list full)*/ >>> + if (err == -ENOMEM) { >>> + kfree(wait); >>> + return err; >>> + } >> >> I agreed with John here that only looking at ENOMEM reads odd and I did not >> see that answered. Did I miss it? > > xa_alloc_cyclic_irq() can also fail with -EBUSY... so that I > think this is a matter... > >> Otherwise, I _know_ it is not likely to get any other error having *just* >> created a new xarray, but still, why not simply catch all? It is not like >> the slot fallback code at runtime would handle guc->serial_slot being >> empty?! It appears it would just explode in guc_send_invalidate_tlb if it >> would hit it.. > > ... of what we want to do for such errors. E.g. Jonathan decided > here not to fail, but ignore the error. Should we fail for every > error? Why not fail on *any* error? This is a fresh and empty xarray. If there is any error at this point how can the driver continue operating? As I wrote, AFAICT it would explode in guc_send_invalidate_tlb() where it assumes xa_load on the guc_serial slot always returns a valid pointer. > >>> + >>> + return 0; >>> +} >>> + >>> +static void fini_tlb_lookup(struct intel_guc *guc) >>> +{ >>> + struct intel_guc_tlb_wait *wait; >>> + >>> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) >>> + return; >>> + >>> + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); >>> + if (wait) { >>> + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); >> >> Hm wait, why is this unexpected when init_tlb_lookup() pre-allocated that >> entry? Who frees it? guc_send_invalidate_tlb() does not appear to. > > I think this links with my answer above, right? With th > refactoring of the if's for not skipping tlb invalidation. I don't think so, init/fini is driver load/unload, no? I went and looked at CI results and they agree with me: <7> [156.380021] i915 0000:00:02.0: [drm:fini_tlb_lookup.part.0 [i915]] GT0: GUC: fini_tlb_lookup: Unexpected item in tlb_lookup <7> [156.395370] i915 0000:00:02.0: [drm:fini_tlb_lookup.part.0 [i915]] GT1: GUC: fini_tlb_lookup: Unexpected item in tlb_lookup So yeah.. code is a bit confused here. >>> + kfree(wait); >>> + } >>> + >>> + xa_destroy(&guc->tlb_lookup); >>> +} >>> + >>> /* >>> * Set up the memory resources to be shared with the GuC (via the GGTT) >>> * at firmware loading time. > > ... > >>> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) >>> +{ >>> + /* Check for underflow */ >>> + if (unlikely(len < 1 || len > size)) >>> + return -EPROTO; >> >> These check are not valid for any message/action type ct_process_request() >> can receive? > > You mean discriminating by payload? Jonathan... you konw the > details here? I mean that ct_process_request() already does: hxg_len = request->size - GUC_CTB_MSG_MIN_LEN; payload = &hxg[GUC_HXG_MSG_MIN_LEN]; action = FIELD_GET(GUC_HXG_EVENT_MSG_0_ACTION, hxg[0]); len = hxg_len - GUC_HXG_MSG_MIN_LEN; So I guess at least the len > size check looks dubious since it appears it is only doubting what ct_process_request() should already sanitize. len < 1 after a look at other handlers looks justified since it is message specific. Also all other handlers in there have the signature of (guc, payload, len). So maybe this one should do the same for consistency? Regards, Tvrtko
On 09/10/2023 13:12, Nirmoy Das wrote: > > On 10/6/2023 8:20 PM, Jonathan Cavitt wrote: >> From: Prathap Kumar Valsan<prathap.kumar.valsan@intel.com> >> >> The GuC firmware had defined the interface for Translation Look-Aside >> Buffer (TLB) invalidation. We should use this interface when >> invalidating the engine and GuC TLBs. >> Add additional functionality to intel_gt_invalidate_tlb, invalidating >> the GuC TLBs and falling back to GT invalidation when the GuC is >> disabled. >> The invalidation is done by sending a request directly to the GuC >> tlb_lookup that invalidates the table. The invalidation is submitted as >> a wait request and is performed in the CT event handler. This means we >> cannot perform this TLB invalidation path if the CT is not enabled. >> If the request isn't fulfilled in two seconds, this would constitute >> an error in the invalidation as that would constitute either a lost >> request or a severe GuC overload. >> >> With this new invalidation routine, we can perform GuC-based GGTT >> invalidations. GuC-based GGTT invalidation is incompatible with >> MMIO invalidation so we should not perform MMIO invalidation when >> GuC-based GGTT invalidation is expected. >> >> Purpose of xarray: >> The tlb_lookup table is allocated as an xarray because the set of >> pending TLB invalidations may have no upper bound. The consequence of >> this is that all actions interfacing with this table need to use the >> xarray functions, such as xa_alloc_cyclic_irq for array insertion. >> >> Purpose of must_wait_woken: >> Our wait for the G2H ack for the completion of a TLB invalidation is >> mandatory; we must wait for the HW to confirm that the physical >> addresses are no longer accessible before we return those to the system. >> >> On switching to using the wait_woken() convenience routine, we >> introduced ourselves to an issue where wait_woken() may complete early >> under a kthread that is stopped. Since we send a TLB invalidation when >> we try to release pages from the shrinker, we can be called from any >> process; including kthreads. >> >> Using wait_woken() from any process context causes another issue. The >> use of is_kthread_should_stop() assumes that any task with PF_KTHREAD >> set was made by kthread_create() and has called set_kthread_struct(). >> This is not true for the raw kernel_thread(): >> >> BUG: kernel NULL pointer dereference, address: 0000000000000000 >> [ 3089.759660] Call Trace: >> [ 3089.762110] wait_woken+0x4f/0x80 >> [ 3089.765496] guc_send_invalidate_tlb+0x1fe/0x310 [i915] >> [ 3089.770725] ? syscall_return_via_sysret+0xf/0x7f >> [ 3089.775426] ? do_wait_intr_irq+0xb0/0xb0 >> [ 3089.779430] ? __switch_to_asm+0x40/0x70 >> [ 3089.783349] ? __switch_to_asm+0x34/0x70 >> [ 3089.787273] ? __switch_to+0x7a/0x3e0 >> [ 3089.790930] ? __switch_to_asm+0x34/0x70 >> [ 3089.794883] intel_guc_invalidate_tlb_full+0x92/0xa0 [i915] >> [ 3089.800487] intel_invalidate_tlb_full+0x94/0x190 [i915] >> [ 3089.805824] intel_invalidate_tlb_full_sync+0x1b/0x30 [i915] >> [ 3089.811508] __i915_gem_object_unset_pages+0x138/0x150 [i915] >> [ 3089.817279] __i915_gem_object_put_pages+0x25/0x90 [i915] >> [ 3089.822706] i915_gem_shrink+0x532/0x7e0 [i915] >> [ 3089.827264] i915_gem_shrinker_scan+0x3d/0xd0 [i915] >> [ 3089.832230] do_shrink_slab+0x12c/0x2a0 >> [ 3089.836065] shrink_slab+0xad/0x2b0 >> [ 3089.839550] shrink_node+0xcc/0x410 >> [ 3089.843035] do_try_to_free_pages+0xc6/0x380 >> [ 3089.847306] try_to_free_pages+0xec/0x1c0 >> [ 3089.851312] __alloc_pages_slowpath+0x3ad/0xd10 >> [ 3089.855845] ? update_sd_lb_stats+0x636/0x710 >> [ 3089.860204] __alloc_pages_nodemask+0x2d5/0x310 >> [ 3089.864737] new_slab+0x265/0xa80 >> [ 3089.868053] ___slab_alloc+0y_to_free_pages+0xec/0x1c0 >> [ 3089.871798] ? copy_process+0x1e5/0x1a00 >> [ 3089.875717] ? load_balance+0x165/0xb20 >> [ 3089.879555] __slab_alloc+0x1c/0x30 >> [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 >> [ 3089.887397] ? copy_process+0x1e5/0x1a00 >> [ 3089.891314] copy_process+0x1e5/0x1a00 >> [ 3089.895058] ? __switch_to_asm+0x40/0x70 >> [ 3089.879555] __slab_alloc+0x1c/0x30 >> [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 >> [ 3089.887397] ? copy_process+0x1e5/0x1a00 >> [ 3089.891314] copy_process+0x1e5/0x1a00 >> [ 3089.895058] ? __switch_to_asm+0x40/0x70 >> [ 3089.898977] ? __switch_to_asm+0x34/0x70 >> [ 3089.902903] ? __switch_to_asm+0x40/0x70 >> [ 3089.906828] ? __switch_to_asm+0x34/0x70 >> [ 3089.910745] _do_fork+0x83/0x350 >> [ 3089.913969] ? __switch_to+0x7a/0x3e0 >> [ 3089.917626] ? __switch_to_asm+0x34/0x70 >> [ 3089.921545] kernel_thread+0x58/0x80 >> [ 3089.925124] ? kthread_park+0x80/0x80 >> [ 3089.928788] kthreadd+0x162/0x1b0 >> [ 3089.932098] ? kthread_create_on_cpu+0xa0/0xa0 >> [ 3089.936538] ret_from_fork+0x1f/0x40 >> >> Signed-off-by: Prathap Kumar Valsan<prathap.kumar.valsan@intel.com> >> Signed-off-by: Bruce Chang<yu.bruce.chang@intel.com> >> Signed-off-by: Chris Wilson<chris.p.wilson@intel.com> >> Signed-off-by: Umesh Nerlige Ramappa<umesh.nerlige.ramappa@intel.com> >> Signed-off-by: Jonathan Cavitt<jonathan.cavitt@intel.com> >> Signed-off-by: Aravind Iddamsetty<aravind.iddamsetty@intel.com> >> Signed-off-by: Fei Yang<fei.yang@intel.com> >> CC: Andi Shyti<andi.shyti@linux.intel.com> >> --- >> drivers/gpu/drm/i915/gt/intel_ggtt.c | 34 ++- >> drivers/gpu/drm/i915/gt/intel_tlb.c | 15 +- >> .../gpu/drm/i915/gt/uc/abi/guc_actions_abi.h | 33 +++ >> drivers/gpu/drm/i915/gt/uc/intel_guc.h | 22 ++ >> drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 4 + >> drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h | 1 + >> .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 200 +++++++++++++++++- >> 7 files changed, 297 insertions(+), 12 deletions(-) >> >> diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c >> index 4d7d88b92632b..a1f7bdc602996 100644 >> --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c >> +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c >> @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt) >> intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN); >> } >> >> +static void guc_ggtt_ct_invalidate(struct intel_gt *gt) >> +{ >> + struct intel_uncore *uncore = gt->uncore; >> + intel_wakeref_t wakeref; >> + >> + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) { >> + struct intel_guc *guc = >->uc.guc; >> + >> + intel_guc_invalidate_tlb_guc(guc); >> + } >> +} >> + >> static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) >> { >> struct drm_i915_private *i915 = ggtt->vm.i915; >> + struct intel_gt *gt; >> >> - gen8_ggtt_invalidate(ggtt); >> - >> - if (GRAPHICS_VER(i915) >= 12) { >> - struct intel_gt *gt; >> + if (!HAS_GUC_TLB_INVALIDATION(i915)) >> + gen8_ggtt_invalidate(ggtt); >> >> - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) >> + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) { >> + if (HAS_GUC_TLB_INVALIDATION(i915) && >> + intel_guc_is_ready(>->uc.guc)) { >> + guc_ggtt_ct_invalidate(gt); >> + } else if (GRAPHICS_VER(i915) >= 12) { >> intel_uncore_write_fw(gt->uncore, >> GEN12_GUC_TLB_INV_CR, >> GEN12_GUC_TLB_INV_CR_INVALIDATE); >> - } else { >> - intel_uncore_write_fw(ggtt->vm.gt->uncore, >> - GEN8_GTCR, GEN8_GTCR_INVALIDATE); >> + } else { >> + intel_uncore_write_fw(gt->uncore, >> + GEN8_GTCR, GEN8_GTCR_INVALIDATE); >> + } >> } >> } >> >> @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) >> ggtt->vm.raw_insert_page = gen8_ggtt_insert_page; >> } >> >> - if (intel_uc_wants_guc(&ggtt->vm.gt->uc)) >> + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc)) >> ggtt->invalidate = guc_ggtt_invalidate; >> else >> ggtt->invalidate = gen8_ggtt_invalidate; >> diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c b/drivers/gpu/drm/i915/gt/intel_tlb.c >> index 139608c30d978..1278622b9554d 100644 >> --- a/drivers/gpu/drm/i915/gt/intel_tlb.c >> +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c >> @@ -12,6 +12,7 @@ >> #include "intel_gt_print.h" >> #include "intel_gt_regs.h" >> #include "intel_tlb.h" >> +#include "uc/intel_guc.h" >> >> /* >> * HW architecture suggest typical invalidation time at 40us, >> @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) >> return; >> >> with_intel_gt_pm_if_awake(gt, wakeref) { >> + struct intel_guc *guc = >->uc.guc; >> + >> mutex_lock(>->tlb.invalidate_lock); >> if (tlb_seqno_passed(gt, seqno)) >> goto unlock; >> >> - mmio_invalidate_full(gt); >> + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { >> + /* >> + * Only perform GuC TLB invalidation if GuC is ready. >> + * If GuC is not ready, then there are no TLBs to >> + * invalidate. Ergo, skip invalidation. >> + */ >> + if (intel_guc_is_ready(guc)) >> + intel_guc_invalidate_tlb_engines(guc); >> + } else { >> + mmio_invalidate_full(gt); >> + } >> >> write_seqcount_invalidate(>->tlb.seqno); >> unlock: >> diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> index f359bef046e0b..33f253410d0c8 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> @@ -138,6 +138,8 @@ enum intel_guc_action { >> INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601, >> INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507, >> INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A, >> + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000, >> + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001, >> INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002, >> INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003, >> INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004, >> @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status { >> >> #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF >> >> +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0) >> +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8) >> +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31) >> + >> +enum intel_guc_tlb_invalidation_type { >> + INTEL_GUC_TLB_INVAL_ENGINES = 0x0, >> + INTEL_GUC_TLB_INVAL_GUC = 0x3, >> +}; >> + >> +/* >> + * 0: Heavy mode of Invalidation: >> + * The pipeline of the engine(s) for which the invalidation is targeted to is >> + * blocked, and all the in-flight transactions are guaranteed to be Globally >> + * Observed before completing the TLB invalidation >> + * 1: Lite mode of Invalidation: >> + * TLBs of the targeted engine(s) are immediately invalidated. >> + * In-flight transactions are NOT guaranteed to be Globally Observed before >> + * completing TLB invalidation. >> + * Light Invalidation Mode is to be used only when >> + * it can be guaranteed (by SW) that the address translations remain invariant >> + * for the in-flight transactions across the TLB invalidation. In other words, >> + * this mode can be used when the TLB invalidation is intended to clear out the >> + * stale cached translations that are no longer in use. Light Invalidation Mode >> + * is much faster than the Heavy Invalidation Mode, as it does not wait for the >> + * in-flight transactions to be GOd. >> + */ >> +enum intel_guc_tlb_inval_mode { >> + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0, >> + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1, >> +}; >> + >> #endif /* _ABI_GUC_ACTIONS_ABI_H */ >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> index 6c392bad29c19..88deb43bbdc48 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> @@ -79,6 +79,18 @@ struct intel_guc { >> */ >> atomic_t outstanding_submission_g2h; >> >> + /** @tlb_lookup: xarray to store all pending TLB invalidation requests */ >> + struct xarray tlb_lookup; >> + >> + /** >> + * @serial_slot: id to the initial waiter created in tlb_lookup, >> + * which is used only when failed to allocate new waiter. >> + */ >> + u32 serial_slot; >> + >> + /** @next_seqno: the next id (sequence no.) to allocate. */ >> + u32 next_seqno; >> + >> /** @interrupts: pointers to GuC interrupt-managing functions. */ >> struct { >> bool enabled; >> @@ -296,6 +308,11 @@ struct intel_guc { >> #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, (ver).minor, (ver).patch) >> #define GUC_SUBMIT_VER(guc) MAKE_GUC_VER_STRUCT((guc)->submission_version) >> >> +struct intel_guc_tlb_wait { >> + struct wait_queue_head wq; >> + bool busy; >> +}; >> + >> static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) >> { >> return container_of(log, struct intel_guc, log); >> @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct intel_guc *guc) >> return intel_uc_fw_is_supported(&guc->fw); >> } >> >> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc); >> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc); >> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, >> + u32 len, u32 seqno); >> + >> static inline bool intel_guc_is_wanted(struct intel_guc *guc) >> { >> return intel_uc_fw_is_enabled(&guc->fw); >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> index 6e22af31513a5..e7ac8402245b7 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> @@ -1115,6 +1115,9 @@ static int ct_process_request(struct intel_guc_ct *ct, struct ct_incoming_msg *r >> case INTEL_GUC_ACTION_NOTIFY_EXCEPTION: >> ret = intel_guc_crash_process_msg(guc, action); >> break; >> + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: >> + ret = intel_guc_tlb_invalidation_done(guc, request->size, len, payload[0]); >> + break; >> default: >> ret = -EOPNOTSUPP; >> break; >> @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct *ct, struct ct_incoming_msg *requ >> switch (action) { >> case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE: >> case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE: >> + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: >> g2h_release_space(ct, request->size); >> } >> >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> index b4d56eccfb1f0..a7c9874e122a3 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> @@ -22,6 +22,7 @@ >> /* Payload length only i.e. don't include G2H header length */ >> #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2 >> #define G2H_LEN_DW_DEREGISTER_CONTEXT 1 >> +#define G2H_LEN_DW_INVALIDATE_TLB 1 >> >> #define GUC_CONTEXT_DISABLE 0 >> #define GUC_CONTEXT_ENABLE 1 >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> index 2cce5ec1ff00d..0c5ccda1b3e87 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct intel_context *ce, intel_engine_mask_t st >> >> void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) >> { >> + struct intel_guc_tlb_wait *wait; >> struct intel_context *ce; >> unsigned long index; >> unsigned long flags; >> + unsigned long i; >> >> if (unlikely(!guc_submission_initialized(guc))) { >> /* Reset called during driver load? GuC not yet initialised! */ >> @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stall >> >> /* GuC is blown away, drop all references to contexts */ >> xa_destroy(&guc->context_lookup); >> + >> + /* >> + * The full GT reset will have cleared the TLB caches and flushed the >> + * G2H message queue; we can release all the blocked waiters. >> + * >> + * This is safe to do unlocked because the xarray is not dependent >> + * on the GT reset, and there's a separate execution path for TLB >> + * invalidations on GT reset, and there's a large window of time >> + * between the GT reset and GuC becoming available. >> + */ >> + xa_for_each(&guc->tlb_lookup, i, wait) >> + wake_up(&wait->wq); >> } >> >> static void guc_cancel_context_requests(struct intel_context *ce) >> @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) >> static void destroyed_worker_func(struct work_struct *w); >> static void reset_fail_worker_func(struct work_struct *w); >> >> +static int init_tlb_lookup(struct intel_guc *guc) >> +{ >> + struct intel_guc_tlb_wait *wait; >> + int err; >> + >> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) >> + return 0; >> + >> + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); >> + >> + wait = kzalloc(sizeof(*wait), GFP_KERNEL); >> + if (!wait) >> + return -ENOMEM; >> + >> + init_waitqueue_head(&wait->wq); >> + >> + /* Preallocate a shared id for use under memory pressure. */ >> + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, >> + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); >> + /* Only error if out of memory, not when busy (list full)*/ >> + if (err == -ENOMEM) { >> + kfree(wait); >> + return err; >> + } >> + >> + return 0; >> +} >> + >> +static void fini_tlb_lookup(struct intel_guc *guc) >> +{ >> + struct intel_guc_tlb_wait *wait; >> + >> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) >> + return; >> + >> + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); >> + if (wait) { >> + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); >> + kfree(wait); >> + } >> + >> + xa_destroy(&guc->tlb_lookup); >> +} >> + >> /* >> * Set up the memory resources to be shared with the GuC (via the GGTT) >> * at firmware loading time. >> @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct intel_guc *guc) >> return ret; >> } >> >> + ret = init_tlb_lookup(guc); >> + if (ret) >> + goto destroy_pool; >> + >> guc->submission_state.guc_ids_bitmap = >> bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL); >> if (!guc->submission_state.guc_ids_bitmap) { >> ret = -ENOMEM; >> - goto destroy_pool; >> + goto destroy_tlb; >> } >> >> guc->timestamp.ping_delay = (POLL_TIME_CLKS / gt->clock_frequency + 1) * HZ; >> @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc *guc) >> >> return 0; >> >> +destroy_tlb: >> + fini_tlb_lookup(guc); >> destroy_pool: >> guc_lrc_desc_pool_destroy_v69(guc); >> - >> return ret; >> } >> >> @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct intel_guc *guc) >> guc_lrc_desc_pool_destroy_v69(guc); >> i915_sched_engine_put(guc->sched_engine); >> bitmap_free(guc->submission_state.guc_ids_bitmap); >> + fini_tlb_lookup(guc); >> guc->submission_initialized = false; >> } >> >> @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) >> return ce; >> } >> >> +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) >> +{ >> + struct intel_guc_tlb_wait *wait; >> + unsigned long flags; >> + >> + xa_lock_irqsave(&guc->tlb_lookup, flags); >> + wait = xa_load(&guc->tlb_lookup, seqno); >> + >> + if (wait) >> + wake_up(&wait->wq); >> + else >> + guc_dbg(guc, >> + "Stale TLB invalidation response with seqno %d\n", seqno); >> + >> + xa_unlock_irqrestore(&guc->tlb_lookup, flags); >> +} >> + >> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) >> +{ >> + /* Check for underflow */ >> + if (unlikely(len < 1 || len > size)) >> + return -EPROTO; >> + >> + wait_wake_outstanding_tlb_g2h(guc, seqno); >> + return 0; >> +} >> + >> +static long must_wait_woken(struct wait_queue_entry *wq_entry, long timeout) >> +{ >> + /* >> + * This is equivalent to wait_woken() with the exception that >> + * we do not wake up early if the kthread task has been completed. >> + * As we are called from page reclaim in any task context, >> + * we may be invoked from stopped kthreads, but we *must* >> + * complete the wait from the HW . >> + * >> + * A second problem is that since we are called under reclaim >> + * and wait_woken() inspected the thread state, it makes an invalid >> + * assumption that all PF_KTHREAD tasks have set_kthread_struct() >> + * called upon them, and will trigger a GPF > > As discussed internally, the GPF issue is resolved with > https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/ > <https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/> If it means no open coding a slighlt different wait_token() that would be a lot better ineed. Although the question of why a single wait queue is not good enough still remains. As I wrote before - give each invalidation a seqno, upon receiving the done h2g store the latest seqno in the driver and wake up all sleepers. They check the seqno and if their has passed exit, otherwise go back to sleep. No xarray needed. Put_pages() invalidations are already serialized so no ill-effect and GGTT invalidations are, or are not, performance sensitive for some reason? Regards, Tvrtko > > >> in is_kthread_should_stop(). >> + */ >> + do { >> + set_current_state(TASK_UNINTERRUPTIBLE); >> + if (wq_entry->flags & WQ_FLAG_WOKEN) >> + break; >> + >> + timeout = schedule_timeout(timeout); >> + } while (timeout); >> + __set_current_state(TASK_RUNNING); >> + >> + /* See wait_woken() and woken_wake_function() */ >> + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); >> + >> + return timeout; >> +} >> + >> +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_inval_mode type) > > > 2nd param should be intel_guc_tlb_invalidation_type not > intel_guc_tlb_inval_mod. > > Not sure why didn't CI complained. > > > Regards, > > Nirmoy > >> +{ >> + struct intel_guc_tlb_wait _wq, *wq = &_wq; >> + DEFINE_WAIT_FUNC(wait, woken_wake_function); >> + int err; >> + u32 seqno; >> + u32 action[] = { >> + INTEL_GUC_ACTION_TLB_INVALIDATION, >> + 0, >> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | >> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, >> + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | >> + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, >> + }; >> + u32 size = ARRAY_SIZE(action); >> + >> + init_waitqueue_head(&_wq.wq); >> + >> + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, >> + xa_limit_32b, &guc->next_seqno, >> + GFP_ATOMIC | __GFP_NOWARN) < 0) { >> + /* Under severe memory pressure? Serialise TLB allocations */ >> + xa_lock_irq(&guc->tlb_lookup); >> + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); >> + wait_event_lock_irq(wq->wq, >> + !READ_ONCE(wq->busy), >> + guc->tlb_lookup.xa_lock); >> + /* >> + * Update wq->busy under lock to ensure only one waiter can >> + * issue the TLB invalidation command using the serial slot at a >> + * time. The condition is set to true before releasing the lock >> + * so that other caller continue to wait until woken up again. >> + */ >> + wq->busy = true; >> + xa_unlock_irq(&guc->tlb_lookup); >> + >> + seqno = guc->serial_slot; >> + } >> + >> + action[1] = seqno; >> + >> + add_wait_queue(&wq->wq, &wait); >> + >> + /* >> + * This is a critical reclaim path and thus we must loop here: >> + * We cannot block for anything that is on the GPU. >> + */ >> + err = intel_guc_send_busy_loop(guc, action, size, G2H_LEN_DW_INVALIDATE_TLB, true); >> + if (err) >> + goto out; >> + >> + if (!must_wait_woken(&wait, intel_guc_ct_expected_delay(&guc->ct))) { >> + guc_err(guc, >> + "TLB invalidation response timed out for seqno %u\n", seqno); >> + err = -ETIME; >> + } >> +out: >> + remove_wait_queue(&wq->wq, &wait); >> + if (seqno != guc->serial_slot) >> + xa_erase_irq(&guc->tlb_lookup, seqno); >> + >> + return err; >> +} >> + >> +/* Full TLB invalidation */ >> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) >> +{ >> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); >> +} >> + >> +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ >> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) >> +{ >> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); >> +} >> + >> int intel_guc_deregister_done_process_msg(struct intel_guc *guc, >> const u32 *msg, >> u32 len)
Hi Tvrtko, On 10/9/2023 2:54 PM, Tvrtko Ursulin wrote: <snip> > >>> +static long must_wait_woken(struct wait_queue_entry *wq_entry, long >>> timeout) >>> +{ >>> + /* >>> + * This is equivalent to wait_woken() with the exception that >>> + * we do not wake up early if the kthread task has been completed. >>> + * As we are called from page reclaim in any task context, >>> + * we may be invoked from stopped kthreads, but we *must* >>> + * complete the wait from the HW . >>> + * >>> + * A second problem is that since we are called under reclaim >>> + * and wait_woken() inspected the thread state, it makes an >>> invalid >>> + * assumption that all PF_KTHREAD tasks have set_kthread_struct() >>> + * called upon them, and will trigger a GPF >> >> As discussed internally, the GPF issue is resolved with >> https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/ >> <https://lore.kernel.org/all/20230602212350.535358-1-jstultz@google.com/> >> > > If it means no open coding a slighlt different wait_token() that would > be a lot better ineed. > > Although the question of why a single wait queue is not good enough > still remains. As I wrote before - give each invalidation a seqno, > upon receiving the done h2g store the latest seqno in the driver and > wake up all sleepers. They check the seqno and if their has passed > exit, otherwise go back to sleep. No xarray needed. Put_pages() > invalidations are already serialized so no ill-effect and GGTT > invalidations are, or are not, performance sensitive for some reason? I think your proposed solution should work as per my understanding because we are doing TLB invalidation of all engines and it is serialized with gt->tlb.invalidate_lock. We might need this when we want to make it more fine grain and do vma based ranged invalidation for better performance. So I think we should try with a single wait queue now. I don't have an concrete answer yet, may be someone else? Regards, Nirmoy > > Regards, > > Tvrtko > >> >> >>> in is_kthread_should_stop(). >>> + */ >>> + do { >>> + set_current_state(TASK_UNINTERRUPTIBLE); >>> + if (wq_entry->flags & WQ_FLAG_WOKEN) >>> + break; >>> + >>> + timeout = schedule_timeout(timeout); >>> + } while (timeout); >>> + __set_current_state(TASK_RUNNING); >>> + >>> + /* See wait_woken() and woken_wake_function() */ >>> + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); >>> + >>> + return timeout; >>> +} >>> + >>> +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum >>> intel_guc_tlb_inval_mode type) >> >> >> 2nd param should be intel_guc_tlb_invalidation_type not >> intel_guc_tlb_inval_mod. >> >> Not sure why didn't CI complained. >> >> >> Regards, >> >> Nirmoy >> >>> +{ >>> + struct intel_guc_tlb_wait _wq, *wq = &_wq; >>> + DEFINE_WAIT_FUNC(wait, woken_wake_function); >>> + int err; >>> + u32 seqno; >>> + u32 action[] = { >>> + INTEL_GUC_ACTION_TLB_INVALIDATION, >>> + 0, >>> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | >>> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, >>> + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | >>> + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, >>> + }; >>> + u32 size = ARRAY_SIZE(action); >>> + >>> + init_waitqueue_head(&_wq.wq); >>> + >>> + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, >>> + xa_limit_32b, &guc->next_seqno, >>> + GFP_ATOMIC | __GFP_NOWARN) < 0) { >>> + /* Under severe memory pressure? Serialise TLB allocations */ >>> + xa_lock_irq(&guc->tlb_lookup); >>> + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); >>> + wait_event_lock_irq(wq->wq, >>> + !READ_ONCE(wq->busy), >>> + guc->tlb_lookup.xa_lock); >>> + /* >>> + * Update wq->busy under lock to ensure only one waiter can >>> + * issue the TLB invalidation command using the serial slot >>> at a >>> + * time. The condition is set to true before releasing the >>> lock >>> + * so that other caller continue to wait until woken up again. >>> + */ >>> + wq->busy = true; >>> + xa_unlock_irq(&guc->tlb_lookup); >>> + >>> + seqno = guc->serial_slot; >>> + } >>> + >>> + action[1] = seqno; >>> + >>> + add_wait_queue(&wq->wq, &wait); >>> + >>> + /* >>> + * This is a critical reclaim path and thus we must loop here: >>> + * We cannot block for anything that is on the GPU. >>> + */ >>> + err = intel_guc_send_busy_loop(guc, action, size, >>> G2H_LEN_DW_INVALIDATE_TLB, true); >>> + if (err) >>> + goto out; >>> + >>> + if (!must_wait_woken(&wait, >>> intel_guc_ct_expected_delay(&guc->ct))) { >>> + guc_err(guc, >>> + "TLB invalidation response timed out for seqno %u\n", >>> seqno); >>> + err = -ETIME; >>> + } >>> +out: >>> + remove_wait_queue(&wq->wq, &wait); >>> + if (seqno != guc->serial_slot) >>> + xa_erase_irq(&guc->tlb_lookup, seqno); >>> + >>> + return err; >>> +} >>> + >>> +/* Full TLB invalidation */ >>> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) >>> +{ >>> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); >>> +} >>> + >>> +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ >>> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) >>> +{ >>> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); >>> +} >>> + >>> int intel_guc_deregister_done_process_msg(struct intel_guc *guc, >>> const u32 *msg, >>> u32 len)
-----Original Message----- From: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com> Sent: Monday, October 9, 2023 1:57 AM To: Cavitt, Jonathan <jonathan.cavitt@intel.com>; intel-gfx@lists.freedesktop.org Cc: Dutt, Sudeep <sudeep.dutt@intel.com>; Chang, Yu bruce <yu.bruce.chang@intel.com>; chris.p.wilson@linux.intel.com; Iddamsetty, Aravind <aravind.iddamsetty@intel.com>; Yang, Fei <fei.yang@intel.com>; Shyti, Andi <andi.shyti@intel.com>; Harrison, John C <john.c.harrison@intel.com>; Das, Nirmoy <nirmoy.das@intel.com>; Krzysztofik, Janusz <janusz.krzysztofik@intel.com>; Roper, Matthew D <matthew.d.roper@intel.com>; jani.nikula@linux.intel.com Subject: Re: [PATCH v8 3/7] drm/i915: Define and use GuC and CTB TLB invalidation routines > > > On 06/10/2023 19:20, Jonathan Cavitt wrote: > > From: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> > > > > The GuC firmware had defined the interface for Translation Look-Aside > > Buffer (TLB) invalidation. We should use this interface when > > invalidating the engine and GuC TLBs. > > Add additional functionality to intel_gt_invalidate_tlb, invalidating > > the GuC TLBs and falling back to GT invalidation when the GuC is > > disabled. > > The invalidation is done by sending a request directly to the GuC > > tlb_lookup that invalidates the table. The invalidation is submitted as > > a wait request and is performed in the CT event handler. This means we > > cannot perform this TLB invalidation path if the CT is not enabled. > > If the request isn't fulfilled in two seconds, this would constitute > > an error in the invalidation as that would constitute either a lost > > request or a severe GuC overload. > > > > With this new invalidation routine, we can perform GuC-based GGTT > > invalidations. GuC-based GGTT invalidation is incompatible with > > MMIO invalidation so we should not perform MMIO invalidation when > > GuC-based GGTT invalidation is expected. > > > > Purpose of xarray: > > The tlb_lookup table is allocated as an xarray because the set of > > pending TLB invalidations may have no upper bound. The consequence of > > this is that all actions interfacing with this table need to use the > > xarray functions, such as xa_alloc_cyclic_irq for array insertion. > > > > Purpose of must_wait_woken: > > Our wait for the G2H ack for the completion of a TLB invalidation is > > mandatory; we must wait for the HW to confirm that the physical > > addresses are no longer accessible before we return those to the system. > > > > On switching to using the wait_woken() convenience routine, we > > introduced ourselves to an issue where wait_woken() may complete early > > under a kthread that is stopped. Since we send a TLB invalidation when > > we try to release pages from the shrinker, we can be called from any > > process; including kthreads. > > > > Using wait_woken() from any process context causes another issue. The > > use of is_kthread_should_stop() assumes that any task with PF_KTHREAD > > set was made by kthread_create() and has called set_kthread_struct(). > > This is not true for the raw kernel_thread(): > > This explanation misses the main point of my ask - which is to explain > why a simpler scheme isn't sufficient. Simpler scheme aka not needed the > xarray or any flavour of wait_token(). > > In other words it is obvious we have to wait for the invalidation ack, > but not obvious why we need a complicated scheme. Okay. I'll remove these chunks and explain that the complexity is required for range-based TLB invalidaitons, which will land upstream eventually. > > > BUG: kernel NULL pointer dereference, address: 0000000000000000 > > [ 3089.759660] Call Trace: > > [ 3089.762110] wait_woken+0x4f/0x80 > > [ 3089.765496] guc_send_invalidate_tlb+0x1fe/0x310 [i915] > > [ 3089.770725] ? syscall_return_via_sysret+0xf/0x7f > > [ 3089.775426] ? do_wait_intr_irq+0xb0/0xb0 > > [ 3089.779430] ? __switch_to_asm+0x40/0x70 > > [ 3089.783349] ? __switch_to_asm+0x34/0x70 > > [ 3089.787273] ? __switch_to+0x7a/0x3e0 > > [ 3089.790930] ? __switch_to_asm+0x34/0x70 > > [ 3089.794883] intel_guc_invalidate_tlb_full+0x92/0xa0 [i915] > > [ 3089.800487] intel_invalidate_tlb_full+0x94/0x190 [i915] > > [ 3089.805824] intel_invalidate_tlb_full_sync+0x1b/0x30 [i915] > > [ 3089.811508] __i915_gem_object_unset_pages+0x138/0x150 [i915] > > [ 3089.817279] __i915_gem_object_put_pages+0x25/0x90 [i915] > > [ 3089.822706] i915_gem_shrink+0x532/0x7e0 [i915] > > [ 3089.827264] i915_gem_shrinker_scan+0x3d/0xd0 [i915] > > [ 3089.832230] do_shrink_slab+0x12c/0x2a0 > > [ 3089.836065] shrink_slab+0xad/0x2b0 > > [ 3089.839550] shrink_node+0xcc/0x410 > > [ 3089.843035] do_try_to_free_pages+0xc6/0x380 > > [ 3089.847306] try_to_free_pages+0xec/0x1c0 > > [ 3089.851312] __alloc_pages_slowpath+0x3ad/0xd10 > > [ 3089.855845] ? update_sd_lb_stats+0x636/0x710 > > [ 3089.860204] __alloc_pages_nodemask+0x2d5/0x310 > > [ 3089.864737] new_slab+0x265/0xa80 > > [ 3089.868053] ___slab_alloc+0y_to_free_pages+0xec/0x1c0 > > [ 3089.871798] ? copy_process+0x1e5/0x1a00 > > [ 3089.875717] ? load_balance+0x165/0xb20 > > [ 3089.879555] __slab_alloc+0x1c/0x30 > > [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 > > [ 3089.887397] ? copy_process+0x1e5/0x1a00 > > [ 3089.891314] copy_process+0x1e5/0x1a00 > > [ 3089.895058] ? __switch_to_asm+0x40/0x70 > > [ 3089.879555] __slab_alloc+0x1c/0x30 > > [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 > > [ 3089.887397] ? copy_process+0x1e5/0x1a00 > > [ 3089.891314] copy_process+0x1e5/0x1a00 > > [ 3089.895058] ? __switch_to_asm+0x40/0x70 > > [ 3089.898977] ? __switch_to_asm+0x34/0x70 > > [ 3089.902903] ? __switch_to_asm+0x40/0x70 > > [ 3089.906828] ? __switch_to_asm+0x34/0x70 > > [ 3089.910745] _do_fork+0x83/0x350 > > [ 3089.913969] ? __switch_to+0x7a/0x3e0 > > [ 3089.917626] ? __switch_to_asm+0x34/0x70 > > [ 3089.921545] kernel_thread+0x58/0x80 > > [ 3089.925124] ? kthread_park+0x80/0x80 > > [ 3089.928788] kthreadd+0x162/0x1b0 > > [ 3089.932098] ? kthread_create_on_cpu+0xa0/0xa0 > > [ 3089.936538] ret_from_fork+0x1f/0x40 > > > > Signed-off-by: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> > > Signed-off-by: Bruce Chang <yu.bruce.chang@intel.com> > > Signed-off-by: Chris Wilson <chris.p.wilson@intel.com> > > Signed-off-by: Umesh Nerlige Ramappa <umesh.nerlige.ramappa@intel.com> > > Signed-off-by: Jonathan Cavitt <jonathan.cavitt@intel.com> > > Signed-off-by: Aravind Iddamsetty <aravind.iddamsetty@intel.com> > > Signed-off-by: Fei Yang <fei.yang@intel.com> > > CC: Andi Shyti <andi.shyti@linux.intel.com> > > --- > > drivers/gpu/drm/i915/gt/intel_ggtt.c | 34 ++- > > drivers/gpu/drm/i915/gt/intel_tlb.c | 15 +- > > .../gpu/drm/i915/gt/uc/abi/guc_actions_abi.h | 33 +++ > > drivers/gpu/drm/i915/gt/uc/intel_guc.h | 22 ++ > > drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 4 + > > drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h | 1 + > > .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 200 +++++++++++++++++- > > 7 files changed, 297 insertions(+), 12 deletions(-) > > > > diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c > > index 4d7d88b92632b..a1f7bdc602996 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c > > +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c > > @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt) > > intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN); > > } > > > > +static void guc_ggtt_ct_invalidate(struct intel_gt *gt) > > +{ > > + struct intel_uncore *uncore = gt->uncore; > > + intel_wakeref_t wakeref; > > + > > + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) { > > + struct intel_guc *guc = >->uc.guc; > > + > > + intel_guc_invalidate_tlb_guc(guc); > > + } > > +} > > + > > static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) > > { > > struct drm_i915_private *i915 = ggtt->vm.i915; > > + struct intel_gt *gt; > > > > - gen8_ggtt_invalidate(ggtt); > > - > > - if (GRAPHICS_VER(i915) >= 12) { > > - struct intel_gt *gt; > > + if (!HAS_GUC_TLB_INVALIDATION(i915)) > > + gen8_ggtt_invalidate(ggtt); > > > > - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) > > + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) { > > + if (HAS_GUC_TLB_INVALIDATION(i915) && > > + intel_guc_is_ready(>->uc.guc)) { > > + guc_ggtt_ct_invalidate(gt); > > + } else if (GRAPHICS_VER(i915) >= 12) { > > intel_uncore_write_fw(gt->uncore, > > GEN12_GUC_TLB_INV_CR, > > GEN12_GUC_TLB_INV_CR_INVALIDATE); > > - } else { > > - intel_uncore_write_fw(ggtt->vm.gt->uncore, > > - GEN8_GTCR, GEN8_GTCR_INVALIDATE); > > + } else { > > + intel_uncore_write_fw(gt->uncore, > > + GEN8_GTCR, GEN8_GTCR_INVALIDATE); > > + } > > } > > } > > > > @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) > > ggtt->vm.raw_insert_page = gen8_ggtt_insert_page; > > } > > > > - if (intel_uc_wants_guc(&ggtt->vm.gt->uc)) > > + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc)) > > ggtt->invalidate = guc_ggtt_invalidate; > > else > > ggtt->invalidate = gen8_ggtt_invalidate; > > diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c b/drivers/gpu/drm/i915/gt/intel_tlb.c > > index 139608c30d978..1278622b9554d 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_tlb.c > > +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c > > @@ -12,6 +12,7 @@ > > #include "intel_gt_print.h" > > #include "intel_gt_regs.h" > > #include "intel_tlb.h" > > +#include "uc/intel_guc.h" > > > > /* > > * HW architecture suggest typical invalidation time at 40us, > > @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) > > return; > > > > with_intel_gt_pm_if_awake(gt, wakeref) { > > + struct intel_guc *guc = >->uc.guc; > > + > > mutex_lock(>->tlb.invalidate_lock); > > if (tlb_seqno_passed(gt, seqno)) > > goto unlock; > > > > - mmio_invalidate_full(gt); > > + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { > > + /* > > + * Only perform GuC TLB invalidation if GuC is ready. > > + * If GuC is not ready, then there are no TLBs to > > + * invalidate. Ergo, skip invalidation. > > + */ > > + if (intel_guc_is_ready(guc)) > > + intel_guc_invalidate_tlb_engines(guc); > > What was the answer to John's question on why it is okay to just skip > and not maybe fall back to mmio? Intel_guc_is_ready is only disabled on a GT reset, which will clobber all the TLBs anyways, so it is safe to skip invalidations in this case. I'll add this as a comment. > > > + } else { > > + mmio_invalidate_full(gt); > > + } > > > > write_seqcount_invalidate(>->tlb.seqno); > > unlock: > > diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > > index f359bef046e0b..33f253410d0c8 100644 > > --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > > +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h > > @@ -138,6 +138,8 @@ enum intel_guc_action { > > INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601, > > INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507, > > INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A, > > + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000, > > + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001, > > INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002, > > INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003, > > INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004, > > @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status { > > > > #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF > > > > +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0) > > +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8) > > +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31) > > + > > +enum intel_guc_tlb_invalidation_type { > > + INTEL_GUC_TLB_INVAL_ENGINES = 0x0, > > + INTEL_GUC_TLB_INVAL_GUC = 0x3, > > +}; > > + > > +/* > > + * 0: Heavy mode of Invalidation: > > + * The pipeline of the engine(s) for which the invalidation is targeted to is > > + * blocked, and all the in-flight transactions are guaranteed to be Globally > > + * Observed before completing the TLB invalidation > > + * 1: Lite mode of Invalidation: > > + * TLBs of the targeted engine(s) are immediately invalidated. > > + * In-flight transactions are NOT guaranteed to be Globally Observed before > > + * completing TLB invalidation. > > + * Light Invalidation Mode is to be used only when > > + * it can be guaranteed (by SW) that the address translations remain invariant > > + * for the in-flight transactions across the TLB invalidation. In other words, > > + * this mode can be used when the TLB invalidation is intended to clear out the > > + * stale cached translations that are no longer in use. Light Invalidation Mode > > + * is much faster than the Heavy Invalidation Mode, as it does not wait for the > > + * in-flight transactions to be GOd. > > + */ > > +enum intel_guc_tlb_inval_mode { > > + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0, > > + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1, > > +}; > > + > > #endif /* _ABI_GUC_ACTIONS_ABI_H */ > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h > > index 6c392bad29c19..88deb43bbdc48 100644 > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h > > @@ -79,6 +79,18 @@ struct intel_guc { > > */ > > atomic_t outstanding_submission_g2h; > > > > + /** @tlb_lookup: xarray to store all pending TLB invalidation requests */ > > + struct xarray tlb_lookup; > > + > > + /** > > + * @serial_slot: id to the initial waiter created in tlb_lookup, > > + * which is used only when failed to allocate new waiter. > > + */ > > + u32 serial_slot; > > + > > + /** @next_seqno: the next id (sequence no.) to allocate. */ > > + u32 next_seqno; > > + > > /** @interrupts: pointers to GuC interrupt-managing functions. */ > > struct { > > bool enabled; > > @@ -296,6 +308,11 @@ struct intel_guc { > > #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, (ver).minor, (ver).patch) > > #define GUC_SUBMIT_VER(guc) MAKE_GUC_VER_STRUCT((guc)->submission_version) > > > > +struct intel_guc_tlb_wait { > > + struct wait_queue_head wq; > > + bool busy; > > +}; > > + > > static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) > > { > > return container_of(log, struct intel_guc, log); > > @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct intel_guc *guc) > > return intel_uc_fw_is_supported(&guc->fw); > > } > > > > +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc); > > +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc); > > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, > > + u32 len, u32 seqno); > > + > > static inline bool intel_guc_is_wanted(struct intel_guc *guc) > > { > > return intel_uc_fw_is_enabled(&guc->fw); > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > > index 6e22af31513a5..e7ac8402245b7 100644 > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c > > @@ -1115,6 +1115,9 @@ static int ct_process_request(struct intel_guc_ct *ct, struct ct_incoming_msg *r > > case INTEL_GUC_ACTION_NOTIFY_EXCEPTION: > > ret = intel_guc_crash_process_msg(guc, action); > > break; > > + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: > > + ret = intel_guc_tlb_invalidation_done(guc, request->size, len, payload[0]); > > + break; > > default: > > ret = -EOPNOTSUPP; > > break; > > @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct *ct, struct ct_incoming_msg *requ > > switch (action) { > > case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE: > > case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE: > > + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: > > g2h_release_space(ct, request->size); > > } > > > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > > index b4d56eccfb1f0..a7c9874e122a3 100644 > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h > > @@ -22,6 +22,7 @@ > > /* Payload length only i.e. don't include G2H header length */ > > #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2 > > #define G2H_LEN_DW_DEREGISTER_CONTEXT 1 > > +#define G2H_LEN_DW_INVALIDATE_TLB 1 > > > > #define GUC_CONTEXT_DISABLE 0 > > #define GUC_CONTEXT_ENABLE 1 > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > > index 2cce5ec1ff00d..0c5ccda1b3e87 100644 > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c > > @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct intel_context *ce, intel_engine_mask_t st > > > > void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) > > { > > + struct intel_guc_tlb_wait *wait; > > struct intel_context *ce; > > unsigned long index; > > unsigned long flags; > > + unsigned long i; > > > > if (unlikely(!guc_submission_initialized(guc))) { > > /* Reset called during driver load? GuC not yet initialised! */ > > @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stall > > > > /* GuC is blown away, drop all references to contexts */ > > xa_destroy(&guc->context_lookup); > > + > > + /* > > + * The full GT reset will have cleared the TLB caches and flushed the > > + * G2H message queue; we can release all the blocked waiters. > > + * > > + * This is safe to do unlocked because the xarray is not dependent > > + * on the GT reset, and there's a separate execution path for TLB > > + * invalidations on GT reset, and there's a large window of time > > + * between the GT reset and GuC becoming available. > > + */ > > + xa_for_each(&guc->tlb_lookup, i, wait) > > + wake_up(&wait->wq); > > If you are confident there can be no failures to wake up someone, who > maybe just added themselves to the xarray (via put pages for instance), > while reset in ongoing. Or even removed themselves after say timing out > the wait and so freed their entry... I'll reduce the comment complexity to just "This is safe to do unlocked because there are no races with waiters entering the array." > > > } > > > > static void guc_cancel_context_requests(struct intel_context *ce) > > @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) > > static void destroyed_worker_func(struct work_struct *w); > > static void reset_fail_worker_func(struct work_struct *w); > > > > +static int init_tlb_lookup(struct intel_guc *guc) > > +{ > > + struct intel_guc_tlb_wait *wait; > > + int err; > > + > > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > > + return 0; > > + > > + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); > > + > > + wait = kzalloc(sizeof(*wait), GFP_KERNEL); > > + if (!wait) > > + return -ENOMEM; > > + > > + init_waitqueue_head(&wait->wq); > > + > > + /* Preallocate a shared id for use under memory pressure. */ > > + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, > > + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); > > + /* Only error if out of memory, not when busy (list full)*/ > > + if (err == -ENOMEM) { > > + kfree(wait); > > + return err; > > + } > > I agreed with John here that only looking at ENOMEM reads odd and I did > not see that answered. Did I miss it? > > Otherwise, I _know_ it is not likely to get any other error having > *just* created a new xarray, but still, why not simply catch all? It is > not like the slot fallback code at runtime would handle guc->serial_slot > being empty?! It appears it would just explode in > guc_send_invalidate_tlb if it would hit it.. I'll error on all err < 0, then. > > > + > > + return 0; > > +} > > + > > +static void fini_tlb_lookup(struct intel_guc *guc) > > +{ > > + struct intel_guc_tlb_wait *wait; > > + > > + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) > > + return; > > + > > + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); > > + if (wait) { > > + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); > > Hm wait, why is this unexpected when init_tlb_lookup() pre-allocated > that entry? Who frees it? guc_send_invalidate_tlb() does not appear to. > It seems I accidentally dropped the if (wait->busy) check, though it seems a debug message here is not needed anyways. I'll drop it. > > + kfree(wait); > > + } > > + > > + xa_destroy(&guc->tlb_lookup); > > +} > > + > > /* > > * Set up the memory resources to be shared with the GuC (via the GGTT) > > * at firmware loading time. > > @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct intel_guc *guc) > > return ret; > > } > > > > + ret = init_tlb_lookup(guc); > > + if (ret) > > + goto destroy_pool; > > + > > guc->submission_state.guc_ids_bitmap = > > bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL); > > if (!guc->submission_state.guc_ids_bitmap) { > > ret = -ENOMEM; > > - goto destroy_pool; > > + goto destroy_tlb; > > } > > > > guc->timestamp.ping_delay = (POLL_TIME_CLKS / gt->clock_frequency + 1) * HZ; > > @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc *guc) > > > > return 0; > > > > +destroy_tlb: > > + fini_tlb_lookup(guc); > > destroy_pool: > > guc_lrc_desc_pool_destroy_v69(guc); > > - > > return ret; > > } > > > > @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct intel_guc *guc) > > guc_lrc_desc_pool_destroy_v69(guc); > > i915_sched_engine_put(guc->sched_engine); > > bitmap_free(guc->submission_state.guc_ids_bitmap); > > + fini_tlb_lookup(guc); > > guc->submission_initialized = false; > > } > > > > @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) > > return ce; > > } > > > > +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) > > +{ > > + struct intel_guc_tlb_wait *wait; > > + unsigned long flags; > > + > > + xa_lock_irqsave(&guc->tlb_lookup, flags); > > + wait = xa_load(&guc->tlb_lookup, seqno); > > + > > + if (wait) > > + wake_up(&wait->wq); > > + else > > + guc_dbg(guc, > > + "Stale TLB invalidation response with seqno %d\n", seqno); > > + > > + xa_unlock_irqrestore(&guc->tlb_lookup, flags); > > +} > > + > > +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) > > +{ > > + /* Check for underflow */ > > + if (unlikely(len < 1 || len > size)) > > + return -EPROTO; > > These check are not valid for any message/action type > ct_process_request() can receive? > Okay, I'll drop this check and reorient intel_guc_tlb_invalidation_done to use (guc, payload, len) format. > > + > > + wait_wake_outstanding_tlb_g2h(guc, seqno); > > + return 0; > > +} > > + > > +static long must_wait_woken(struct wait_queue_entry *wq_entry, long timeout) > > +{ > > + /* > > + * This is equivalent to wait_woken() with the exception that > > + * we do not wake up early if the kthread task has been completed. > > + * As we are called from page reclaim in any task context, > > + * we may be invoked from stopped kthreads, but we *must* > > + * complete the wait from the HW . > > + * > > + * A second problem is that since we are called under reclaim > > + * and wait_woken() inspected the thread state, it makes an invalid > > + * assumption that all PF_KTHREAD tasks have set_kthread_struct() > > + * called upon them, and will trigger a GPF in is_kthread_should_stop(). > > + */ It seems this GPF issue will be resolved soon. We still need the additional complexity because we need an uninterruptible wait, which is not currently provided by the kernel. > > + do { > > + set_current_state(TASK_UNINTERRUPTIBLE); > > + if (wq_entry->flags & WQ_FLAG_WOKEN) > > + break; > > + > > + timeout = schedule_timeout(timeout); > > + } while (timeout); > > + __set_current_state(TASK_RUNNING); > > + > > + /* See wait_woken() and woken_wake_function() */ > > + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); > > + > > + return timeout; > > +} > > + > > +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_inval_mode type) This turned out to be the wrong enum. I'll fix it. -Jonathan Cavitt > > +{ > > + struct intel_guc_tlb_wait _wq, *wq = &_wq; > > + DEFINE_WAIT_FUNC(wait, woken_wake_function); > > + int err; > > + u32 seqno; > > + u32 action[] = { > > + INTEL_GUC_ACTION_TLB_INVALIDATION, > > + 0, > > + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | > > + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, > > + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | > > + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, > > + }; > > + u32 size = ARRAY_SIZE(action); > > + > > + init_waitqueue_head(&_wq.wq); > > + > > + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, > > + xa_limit_32b, &guc->next_seqno, > > + GFP_ATOMIC | __GFP_NOWARN) < 0) { > > + /* Under severe memory pressure? Serialise TLB allocations */ > > + xa_lock_irq(&guc->tlb_lookup); > > + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); > > + wait_event_lock_irq(wq->wq, > > + !READ_ONCE(wq->busy), > > + guc->tlb_lookup.xa_lock); > > + /* > > + * Update wq->busy under lock to ensure only one waiter can > > + * issue the TLB invalidation command using the serial slot at a > > + * time. The condition is set to true before releasing the lock > > + * so that other caller continue to wait until woken up again. > > + */ > > + wq->busy = true; > > + xa_unlock_irq(&guc->tlb_lookup); > > + > > + seqno = guc->serial_slot; > > + } > > + > > + action[1] = seqno; > > + > > + add_wait_queue(&wq->wq, &wait); > > + > > + /* > > + * This is a critical reclaim path and thus we must loop here: > > + * We cannot block for anything that is on the GPU. > > + */ > > + err = intel_guc_send_busy_loop(guc, action, size, G2H_LEN_DW_INVALIDATE_TLB, true); > > + if (err) > > + goto out; > > + > > + if (!must_wait_woken(&wait, intel_guc_ct_expected_delay(&guc->ct))) { > > + guc_err(guc, > > + "TLB invalidation response timed out for seqno %u\n", seqno); > > + err = -ETIME; > > + } > > +out: > > + remove_wait_queue(&wq->wq, &wait); > > + if (seqno != guc->serial_slot) > > + xa_erase_irq(&guc->tlb_lookup, seqno); > > + > > + return err; > > +} > > + > > +/* Full TLB invalidation */ > > +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) > > +{ > > + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); > > +} > > + > > +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ > > +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) > > +{ > > + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); > > +} > > + > > int intel_guc_deregister_done_process_msg(struct intel_guc *guc, > > const u32 *msg, > > u32 len) > > Regards, > > Tvrtko >
On 10/9/2023 01:56, Tvrtko Ursulin wrote: > On 06/10/2023 19:20, Jonathan Cavitt wrote: >> From: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> >> >> The GuC firmware had defined the interface for Translation Look-Aside >> Buffer (TLB) invalidation. We should use this interface when >> invalidating the engine and GuC TLBs. >> Add additional functionality to intel_gt_invalidate_tlb, invalidating >> the GuC TLBs and falling back to GT invalidation when the GuC is >> disabled. >> The invalidation is done by sending a request directly to the GuC >> tlb_lookup that invalidates the table. The invalidation is submitted as >> a wait request and is performed in the CT event handler. This means we >> cannot perform this TLB invalidation path if the CT is not enabled. >> If the request isn't fulfilled in two seconds, this would constitute >> an error in the invalidation as that would constitute either a lost >> request or a severe GuC overload. >> >> With this new invalidation routine, we can perform GuC-based GGTT >> invalidations. GuC-based GGTT invalidation is incompatible with >> MMIO invalidation so we should not perform MMIO invalidation when >> GuC-based GGTT invalidation is expected. >> >> Purpose of xarray: >> The tlb_lookup table is allocated as an xarray because the set of >> pending TLB invalidations may have no upper bound. The consequence of >> this is that all actions interfacing with this table need to use the >> xarray functions, such as xa_alloc_cyclic_irq for array insertion. >> >> Purpose of must_wait_woken: >> Our wait for the G2H ack for the completion of a TLB invalidation is >> mandatory; we must wait for the HW to confirm that the physical >> addresses are no longer accessible before we return those to the system. >> >> On switching to using the wait_woken() convenience routine, we >> introduced ourselves to an issue where wait_woken() may complete early >> under a kthread that is stopped. Since we send a TLB invalidation when >> we try to release pages from the shrinker, we can be called from any >> process; including kthreads. >> >> Using wait_woken() from any process context causes another issue. The >> use of is_kthread_should_stop() assumes that any task with PF_KTHREAD >> set was made by kthread_create() and has called set_kthread_struct(). >> This is not true for the raw kernel_thread(): > > This explanation misses the main point of my ask - which is to explain > why a simpler scheme isn't sufficient. Simpler scheme aka not needed > the xarray or any flavour of wait_token(). > > In other words it is obvious we have to wait for the invalidation ack, > but not obvious why we need a complicated scheme. The alternative being to simply serialise all TLB invalidation requests? Thus, no complex tracking required as there is only one in flight at a time? That seems inefficient and a potential performance impact if a bunch of invalidations are required back to back. But given that the current scheme is global invalidation only (no support for ranges / per page invalidation yet), is it possible to get multiple back-to-back requests? > >> BUG: kernel NULL pointer dereference, address: 0000000000000000 >> [ 3089.759660] Call Trace: >> [ 3089.762110] wait_woken+0x4f/0x80 >> [ 3089.765496] guc_send_invalidate_tlb+0x1fe/0x310 [i915] >> [ 3089.770725] ? syscall_return_via_sysret+0xf/0x7f >> [ 3089.775426] ? do_wait_intr_irq+0xb0/0xb0 >> [ 3089.779430] ? __switch_to_asm+0x40/0x70 >> [ 3089.783349] ? __switch_to_asm+0x34/0x70 >> [ 3089.787273] ? __switch_to+0x7a/0x3e0 >> [ 3089.790930] ? __switch_to_asm+0x34/0x70 >> [ 3089.794883] intel_guc_invalidate_tlb_full+0x92/0xa0 [i915] >> [ 3089.800487] intel_invalidate_tlb_full+0x94/0x190 [i915] >> [ 3089.805824] intel_invalidate_tlb_full_sync+0x1b/0x30 [i915] >> [ 3089.811508] __i915_gem_object_unset_pages+0x138/0x150 [i915] >> [ 3089.817279] __i915_gem_object_put_pages+0x25/0x90 [i915] >> [ 3089.822706] i915_gem_shrink+0x532/0x7e0 [i915] >> [ 3089.827264] i915_gem_shrinker_scan+0x3d/0xd0 [i915] >> [ 3089.832230] do_shrink_slab+0x12c/0x2a0 >> [ 3089.836065] shrink_slab+0xad/0x2b0 >> [ 3089.839550] shrink_node+0xcc/0x410 >> [ 3089.843035] do_try_to_free_pages+0xc6/0x380 >> [ 3089.847306] try_to_free_pages+0xec/0x1c0 >> [ 3089.851312] __alloc_pages_slowpath+0x3ad/0xd10 >> [ 3089.855845] ? update_sd_lb_stats+0x636/0x710 >> [ 3089.860204] __alloc_pages_nodemask+0x2d5/0x310 >> [ 3089.864737] new_slab+0x265/0xa80 >> [ 3089.868053] ___slab_alloc+0y_to_free_pages+0xec/0x1c0 >> [ 3089.871798] ? copy_process+0x1e5/0x1a00 >> [ 3089.875717] ? load_balance+0x165/0xb20 >> [ 3089.879555] __slab_alloc+0x1c/0x30 >> [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 >> [ 3089.887397] ? copy_process+0x1e5/0x1a00 >> [ 3089.891314] copy_process+0x1e5/0x1a00 >> [ 3089.895058] ? __switch_to_asm+0x40/0x70 >> [ 3089.879555] __slab_alloc+0x1c/0x30 >> [ 3089.883047] kmem_cache_alloc_node+0x9f/0x240 >> [ 3089.887397] ? copy_process+0x1e5/0x1a00 >> [ 3089.891314] copy_process+0x1e5/0x1a00 >> [ 3089.895058] ? __switch_to_asm+0x40/0x70 >> [ 3089.898977] ? __switch_to_asm+0x34/0x70 >> [ 3089.902903] ? __switch_to_asm+0x40/0x70 >> [ 3089.906828] ? __switch_to_asm+0x34/0x70 >> [ 3089.910745] _do_fork+0x83/0x350 >> [ 3089.913969] ? __switch_to+0x7a/0x3e0 >> [ 3089.917626] ? __switch_to_asm+0x34/0x70 >> [ 3089.921545] kernel_thread+0x58/0x80 >> [ 3089.925124] ? kthread_park+0x80/0x80 >> [ 3089.928788] kthreadd+0x162/0x1b0 >> [ 3089.932098] ? kthread_create_on_cpu+0xa0/0xa0 >> [ 3089.936538] ret_from_fork+0x1f/0x40 >> >> Signed-off-by: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> >> Signed-off-by: Bruce Chang <yu.bruce.chang@intel.com> >> Signed-off-by: Chris Wilson <chris.p.wilson@intel.com> >> Signed-off-by: Umesh Nerlige Ramappa <umesh.nerlige.ramappa@intel.com> >> Signed-off-by: Jonathan Cavitt <jonathan.cavitt@intel.com> >> Signed-off-by: Aravind Iddamsetty <aravind.iddamsetty@intel.com> >> Signed-off-by: Fei Yang <fei.yang@intel.com> >> CC: Andi Shyti <andi.shyti@linux.intel.com> >> --- >> drivers/gpu/drm/i915/gt/intel_ggtt.c | 34 ++- >> drivers/gpu/drm/i915/gt/intel_tlb.c | 15 +- >> .../gpu/drm/i915/gt/uc/abi/guc_actions_abi.h | 33 +++ >> drivers/gpu/drm/i915/gt/uc/intel_guc.h | 22 ++ >> drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c | 4 + >> drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h | 1 + >> .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 200 +++++++++++++++++- >> 7 files changed, 297 insertions(+), 12 deletions(-) >> >> diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c >> b/drivers/gpu/drm/i915/gt/intel_ggtt.c >> index 4d7d88b92632b..a1f7bdc602996 100644 >> --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c >> +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c >> @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct >> i915_ggtt *ggtt) >> intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, >> GFX_FLSH_CNTL_EN); >> } >> +static void guc_ggtt_ct_invalidate(struct intel_gt *gt) >> +{ >> + struct intel_uncore *uncore = gt->uncore; >> + intel_wakeref_t wakeref; >> + >> + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) { >> + struct intel_guc *guc = >->uc.guc; >> + >> + intel_guc_invalidate_tlb_guc(guc); >> + } >> +} >> + >> static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) >> { >> struct drm_i915_private *i915 = ggtt->vm.i915; >> + struct intel_gt *gt; >> - gen8_ggtt_invalidate(ggtt); >> - >> - if (GRAPHICS_VER(i915) >= 12) { >> - struct intel_gt *gt; >> + if (!HAS_GUC_TLB_INVALIDATION(i915)) >> + gen8_ggtt_invalidate(ggtt); >> - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) >> + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) { >> + if (HAS_GUC_TLB_INVALIDATION(i915) && >> + intel_guc_is_ready(>->uc.guc)) { >> + guc_ggtt_ct_invalidate(gt); >> + } else if (GRAPHICS_VER(i915) >= 12) { >> intel_uncore_write_fw(gt->uncore, >> GEN12_GUC_TLB_INV_CR, >> GEN12_GUC_TLB_INV_CR_INVALIDATE); >> - } else { >> - intel_uncore_write_fw(ggtt->vm.gt->uncore, >> - GEN8_GTCR, GEN8_GTCR_INVALIDATE); >> + } else { >> + intel_uncore_write_fw(gt->uncore, >> + GEN8_GTCR, GEN8_GTCR_INVALIDATE); >> + } >> } >> } >> @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt >> *ggtt) >> ggtt->vm.raw_insert_page = gen8_ggtt_insert_page; >> } >> - if (intel_uc_wants_guc(&ggtt->vm.gt->uc)) >> + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc)) >> ggtt->invalidate = guc_ggtt_invalidate; >> else >> ggtt->invalidate = gen8_ggtt_invalidate; >> diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c >> b/drivers/gpu/drm/i915/gt/intel_tlb.c >> index 139608c30d978..1278622b9554d 100644 >> --- a/drivers/gpu/drm/i915/gt/intel_tlb.c >> +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c >> @@ -12,6 +12,7 @@ >> #include "intel_gt_print.h" >> #include "intel_gt_regs.h" >> #include "intel_tlb.h" >> +#include "uc/intel_guc.h" >> /* >> * HW architecture suggest typical invalidation time at 40us, >> @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct >> intel_gt *gt, u32 seqno) >> return; >> with_intel_gt_pm_if_awake(gt, wakeref) { >> + struct intel_guc *guc = >->uc.guc; >> + >> mutex_lock(>->tlb.invalidate_lock); >> if (tlb_seqno_passed(gt, seqno)) >> goto unlock; >> - mmio_invalidate_full(gt); >> + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { >> + /* >> + * Only perform GuC TLB invalidation if GuC is ready. >> + * If GuC is not ready, then there are no TLBs to >> + * invalidate. Ergo, skip invalidation. >> + */ >> + if (intel_guc_is_ready(guc)) >> + intel_guc_invalidate_tlb_engines(guc); > > What was the answer to John's question on why it is okay to just skip > and not maybe fall back to mmio? > >> + } else { >> + mmio_invalidate_full(gt); >> + } >> write_seqcount_invalidate(>->tlb.seqno); >> unlock: >> diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> index f359bef046e0b..33f253410d0c8 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h >> @@ -138,6 +138,8 @@ enum intel_guc_action { >> INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601, >> INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507, >> INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A, >> + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000, >> + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001, >> INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002, >> INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003, >> INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004, >> @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status { >> #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF >> +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0) >> +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8) >> +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31) >> + >> +enum intel_guc_tlb_invalidation_type { >> + INTEL_GUC_TLB_INVAL_ENGINES = 0x0, >> + INTEL_GUC_TLB_INVAL_GUC = 0x3, >> +}; >> + >> +/* >> + * 0: Heavy mode of Invalidation: >> + * The pipeline of the engine(s) for which the invalidation is >> targeted to is >> + * blocked, and all the in-flight transactions are guaranteed to be >> Globally >> + * Observed before completing the TLB invalidation >> + * 1: Lite mode of Invalidation: >> + * TLBs of the targeted engine(s) are immediately invalidated. >> + * In-flight transactions are NOT guaranteed to be Globally Observed >> before >> + * completing TLB invalidation. >> + * Light Invalidation Mode is to be used only when >> + * it can be guaranteed (by SW) that the address translations remain >> invariant >> + * for the in-flight transactions across the TLB invalidation. In >> other words, >> + * this mode can be used when the TLB invalidation is intended to >> clear out the >> + * stale cached translations that are no longer in use. Light >> Invalidation Mode >> + * is much faster than the Heavy Invalidation Mode, as it does not >> wait for the >> + * in-flight transactions to be GOd. >> + */ >> +enum intel_guc_tlb_inval_mode { >> + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0, >> + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1, >> +}; >> + >> #endif /* _ABI_GUC_ACTIONS_ABI_H */ >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> b/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> index 6c392bad29c19..88deb43bbdc48 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h >> @@ -79,6 +79,18 @@ struct intel_guc { >> */ >> atomic_t outstanding_submission_g2h; >> + /** @tlb_lookup: xarray to store all pending TLB invalidation >> requests */ >> + struct xarray tlb_lookup; >> + >> + /** >> + * @serial_slot: id to the initial waiter created in tlb_lookup, >> + * which is used only when failed to allocate new waiter. >> + */ >> + u32 serial_slot; >> + >> + /** @next_seqno: the next id (sequence no.) to allocate. */ >> + u32 next_seqno; >> + >> /** @interrupts: pointers to GuC interrupt-managing functions. */ >> struct { >> bool enabled; >> @@ -296,6 +308,11 @@ struct intel_guc { >> #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, >> (ver).minor, (ver).patch) >> #define GUC_SUBMIT_VER(guc) >> MAKE_GUC_VER_STRUCT((guc)->submission_version) >> +struct intel_guc_tlb_wait { >> + struct wait_queue_head wq; >> + bool busy; >> +}; >> + >> static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) >> { >> return container_of(log, struct intel_guc, log); >> @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct >> intel_guc *guc) >> return intel_uc_fw_is_supported(&guc->fw); >> } >> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc); >> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc); >> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, >> + u32 len, u32 seqno); >> + >> static inline bool intel_guc_is_wanted(struct intel_guc *guc) >> { >> return intel_uc_fw_is_enabled(&guc->fw); >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> index 6e22af31513a5..e7ac8402245b7 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c >> @@ -1115,6 +1115,9 @@ static int ct_process_request(struct >> intel_guc_ct *ct, struct ct_incoming_msg *r >> case INTEL_GUC_ACTION_NOTIFY_EXCEPTION: >> ret = intel_guc_crash_process_msg(guc, action); >> break; >> + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: >> + ret = intel_guc_tlb_invalidation_done(guc, request->size, >> len, payload[0]); This is still non-standard with respect to all other notification handlers. The convention is to pass '(guc, payload, len)'. The only exception is the two exception notifications above which have no payload and share a common handler for two different actions. This is just a normal single action, with payload handler. So should match the convention of the others. >> + break; >> default: >> ret = -EOPNOTSUPP; >> break; >> @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct >> *ct, struct ct_incoming_msg *requ >> switch (action) { >> case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE: >> case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE: >> + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: >> g2h_release_space(ct, request->size); >> } >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> index b4d56eccfb1f0..a7c9874e122a3 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h >> @@ -22,6 +22,7 @@ >> /* Payload length only i.e. don't include G2H header length */ >> #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2 >> #define G2H_LEN_DW_DEREGISTER_CONTEXT 1 >> +#define G2H_LEN_DW_INVALIDATE_TLB 1 >> #define GUC_CONTEXT_DISABLE 0 >> #define GUC_CONTEXT_ENABLE 1 >> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> index 2cce5ec1ff00d..0c5ccda1b3e87 100644 >> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c >> @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct >> intel_context *ce, intel_engine_mask_t st >> void intel_guc_submission_reset(struct intel_guc *guc, >> intel_engine_mask_t stalled) >> { >> + struct intel_guc_tlb_wait *wait; >> struct intel_context *ce; >> unsigned long index; >> unsigned long flags; >> + unsigned long i; >> if (unlikely(!guc_submission_initialized(guc))) { >> /* Reset called during driver load? GuC not yet >> initialised! */ >> @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct >> intel_guc *guc, intel_engine_mask_t stall >> /* GuC is blown away, drop all references to contexts */ >> xa_destroy(&guc->context_lookup); >> + >> + /* >> + * The full GT reset will have cleared the TLB caches and >> flushed the >> + * G2H message queue; we can release all the blocked waiters. >> + * >> + * This is safe to do unlocked because the xarray is not dependent >> + * on the GT reset, and there's a separate execution path for TLB >> + * invalidations on GT reset, and there's a large window of time >> + * between the GT reset and GuC becoming available. >> + */ >> + xa_for_each(&guc->tlb_lookup, i, wait) >> + wake_up(&wait->wq); > > If you are confident there can be no failures to wake up someone, who > maybe just added themselves to the xarray (via put pages for > instance), while reset in ongoing. Or even removed themselves after > say timing out the wait and so freed their entry... > >> } >> static void guc_cancel_context_requests(struct intel_context *ce) >> @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct >> intel_guc *guc) >> static void destroyed_worker_func(struct work_struct *w); >> static void reset_fail_worker_func(struct work_struct *w); >> +static int init_tlb_lookup(struct intel_guc *guc) >> +{ >> + struct intel_guc_tlb_wait *wait; >> + int err; >> + >> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) >> + return 0; >> + >> + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); >> + >> + wait = kzalloc(sizeof(*wait), GFP_KERNEL); >> + if (!wait) >> + return -ENOMEM; >> + >> + init_waitqueue_head(&wait->wq); >> + >> + /* Preallocate a shared id for use under memory pressure. */ >> + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, >> wait, >> + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); >> + /* Only error if out of memory, not when busy (list full)*/ >> + if (err == -ENOMEM) { >> + kfree(wait); >> + return err; >> + } > > I agreed with John here that only looking at ENOMEM reads odd and I > did not see that answered. Did I miss it? > > Otherwise, I _know_ it is not likely to get any other error having > *just* created a new xarray, but still, why not simply catch all? It > is not like the slot fallback code at runtime would handle > guc->serial_slot being empty?! It appears it would just explode in > guc_send_invalidate_tlb if it would hit it.. To be clear, if you are going to write an error handler that checks for one specific error and only that one error, i.e. it completely ignores any other error, then it needs a big clear comment as to exactly why it is safe and valid to ignore any and all other errors. Saying 'do not error out when busy' makes no sense for an an allocation function. This is not acquiring a resource. This is allocating a tracking structure. How can it be busy? How can it not be a fatal problem if a tracking structure failed to allocate due to being 'busy'? And even if 'busy' is safe to ignore, the code is not explicitly checking for EBUSY and letting that through. What if the call were to return EINVAL, EINTR, EANYTHINGELSEATALL? If you have a known safe error case then you must explicitly check for that and fail on anything else. Not fail on one specific case and pass through anything else. > >> + >> + return 0; >> +} >> + >> +static void fini_tlb_lookup(struct intel_guc *guc) >> +{ >> + struct intel_guc_tlb_wait *wait; >> + >> + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) >> + return; >> + >> + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); >> + if (wait) { >> + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in >> tlb_lookup\n"); > > Hm wait, why is this unexpected when init_tlb_lookup() pre-allocated > that entry? Who frees it? guc_send_invalidate_tlb() does not appear to. The original version of this code was 'BUG_ON(wait->busy)';. Should the dbg print be conditional on wait->busy being set? And should it be an error not a debug print? John. > >> + kfree(wait); >> + } >> + >> + xa_destroy(&guc->tlb_lookup); >> +} >> + >> /* >> * Set up the memory resources to be shared with the GuC (via the >> GGTT) >> * at firmware loading time. >> @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct >> intel_guc *guc) >> return ret; >> } >> + ret = init_tlb_lookup(guc); >> + if (ret) >> + goto destroy_pool; >> + >> guc->submission_state.guc_ids_bitmap = >> bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL); >> if (!guc->submission_state.guc_ids_bitmap) { >> ret = -ENOMEM; >> - goto destroy_pool; >> + goto destroy_tlb; >> } >> guc->timestamp.ping_delay = (POLL_TIME_CLKS / >> gt->clock_frequency + 1) * HZ; >> @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc >> *guc) >> return 0; >> +destroy_tlb: >> + fini_tlb_lookup(guc); >> destroy_pool: >> guc_lrc_desc_pool_destroy_v69(guc); >> - >> return ret; >> } >> @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct >> intel_guc *guc) >> guc_lrc_desc_pool_destroy_v69(guc); >> i915_sched_engine_put(guc->sched_engine); >> bitmap_free(guc->submission_state.guc_ids_bitmap); >> + fini_tlb_lookup(guc); >> guc->submission_initialized = false; >> } >> @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, >> u32 ctx_id) >> return ce; >> } >> +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, >> u32 seqno) >> +{ >> + struct intel_guc_tlb_wait *wait; >> + unsigned long flags; >> + >> + xa_lock_irqsave(&guc->tlb_lookup, flags); >> + wait = xa_load(&guc->tlb_lookup, seqno); >> + >> + if (wait) >> + wake_up(&wait->wq); >> + else >> + guc_dbg(guc, >> + "Stale TLB invalidation response with seqno %d\n", seqno); >> + >> + xa_unlock_irqrestore(&guc->tlb_lookup, flags); >> +} >> + >> +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, >> u32 len, u32 seqno) >> +{ >> + /* Check for underflow */ >> + if (unlikely(len < 1 || len > size)) >> + return -EPROTO; > > These check are not valid for any message/action type > ct_process_request() can receive? > >> + >> + wait_wake_outstanding_tlb_g2h(guc, seqno); >> + return 0; >> +} >> + >> +static long must_wait_woken(struct wait_queue_entry *wq_entry, long >> timeout) >> +{ >> + /* >> + * This is equivalent to wait_woken() with the exception that >> + * we do not wake up early if the kthread task has been completed. >> + * As we are called from page reclaim in any task context, >> + * we may be invoked from stopped kthreads, but we *must* >> + * complete the wait from the HW . >> + * >> + * A second problem is that since we are called under reclaim >> + * and wait_woken() inspected the thread state, it makes an invalid >> + * assumption that all PF_KTHREAD tasks have set_kthread_struct() >> + * called upon them, and will trigger a GPF in >> is_kthread_should_stop(). >> + */ >> + do { >> + set_current_state(TASK_UNINTERRUPTIBLE); >> + if (wq_entry->flags & WQ_FLAG_WOKEN) >> + break; >> + >> + timeout = schedule_timeout(timeout); >> + } while (timeout); >> + __set_current_state(TASK_RUNNING); >> + >> + /* See wait_woken() and woken_wake_function() */ >> + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); >> + >> + return timeout; >> +} >> + >> +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum >> intel_guc_tlb_inval_mode type) >> +{ >> + struct intel_guc_tlb_wait _wq, *wq = &_wq; >> + DEFINE_WAIT_FUNC(wait, woken_wake_function); >> + int err; >> + u32 seqno; >> + u32 action[] = { >> + INTEL_GUC_ACTION_TLB_INVALIDATION, >> + 0, >> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | >> + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, >> + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | >> + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, >> + }; >> + u32 size = ARRAY_SIZE(action); >> + >> + init_waitqueue_head(&_wq.wq); >> + >> + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, >> + xa_limit_32b, &guc->next_seqno, >> + GFP_ATOMIC | __GFP_NOWARN) < 0) { >> + /* Under severe memory pressure? Serialise TLB allocations */ >> + xa_lock_irq(&guc->tlb_lookup); >> + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); >> + wait_event_lock_irq(wq->wq, >> + !READ_ONCE(wq->busy), >> + guc->tlb_lookup.xa_lock); >> + /* >> + * Update wq->busy under lock to ensure only one waiter can >> + * issue the TLB invalidation command using the serial slot >> at a >> + * time. The condition is set to true before releasing the lock >> + * so that other caller continue to wait until woken up again. >> + */ >> + wq->busy = true; >> + xa_unlock_irq(&guc->tlb_lookup); >> + >> + seqno = guc->serial_slot; >> + } >> + >> + action[1] = seqno; >> + >> + add_wait_queue(&wq->wq, &wait); >> + >> + /* >> + * This is a critical reclaim path and thus we must loop here: >> + * We cannot block for anything that is on the GPU. >> + */ >> + err = intel_guc_send_busy_loop(guc, action, size, >> G2H_LEN_DW_INVALIDATE_TLB, true); >> + if (err) >> + goto out; >> + >> + if (!must_wait_woken(&wait, >> intel_guc_ct_expected_delay(&guc->ct))) { >> + guc_err(guc, >> + "TLB invalidation response timed out for seqno %u\n", >> seqno); >> + err = -ETIME; >> + } >> +out: >> + remove_wait_queue(&wq->wq, &wait); >> + if (seqno != guc->serial_slot) >> + xa_erase_irq(&guc->tlb_lookup, seqno); >> + >> + return err; >> +} >> + >> +/* Full TLB invalidation */ >> +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) >> +{ >> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); >> +} >> + >> +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ >> +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) >> +{ >> + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); >> +} >> + >> int intel_guc_deregister_done_process_msg(struct intel_guc *guc, >> const u32 *msg, >> u32 len) > > Regards, > > Tvrtko
On 09/10/2023 20:14, John Harrison wrote: > On 10/9/2023 01:56, Tvrtko Ursulin wrote: >> On 06/10/2023 19:20, Jonathan Cavitt wrote: >>> From: Prathap Kumar Valsan <prathap.kumar.valsan@intel.com> >>> >>> The GuC firmware had defined the interface for Translation Look-Aside >>> Buffer (TLB) invalidation. We should use this interface when >>> invalidating the engine and GuC TLBs. >>> Add additional functionality to intel_gt_invalidate_tlb, invalidating >>> the GuC TLBs and falling back to GT invalidation when the GuC is >>> disabled. >>> The invalidation is done by sending a request directly to the GuC >>> tlb_lookup that invalidates the table. The invalidation is submitted as >>> a wait request and is performed in the CT event handler. This means we >>> cannot perform this TLB invalidation path if the CT is not enabled. >>> If the request isn't fulfilled in two seconds, this would constitute >>> an error in the invalidation as that would constitute either a lost >>> request or a severe GuC overload. >>> >>> With this new invalidation routine, we can perform GuC-based GGTT >>> invalidations. GuC-based GGTT invalidation is incompatible with >>> MMIO invalidation so we should not perform MMIO invalidation when >>> GuC-based GGTT invalidation is expected. >>> >>> Purpose of xarray: >>> The tlb_lookup table is allocated as an xarray because the set of >>> pending TLB invalidations may have no upper bound. The consequence of >>> this is that all actions interfacing with this table need to use the >>> xarray functions, such as xa_alloc_cyclic_irq for array insertion. >>> >>> Purpose of must_wait_woken: >>> Our wait for the G2H ack for the completion of a TLB invalidation is >>> mandatory; we must wait for the HW to confirm that the physical >>> addresses are no longer accessible before we return those to the system. >>> >>> On switching to using the wait_woken() convenience routine, we >>> introduced ourselves to an issue where wait_woken() may complete early >>> under a kthread that is stopped. Since we send a TLB invalidation when >>> we try to release pages from the shrinker, we can be called from any >>> process; including kthreads. >>> >>> Using wait_woken() from any process context causes another issue. The >>> use of is_kthread_should_stop() assumes that any task with PF_KTHREAD >>> set was made by kthread_create() and has called set_kthread_struct(). >>> This is not true for the raw kernel_thread(): >> >> This explanation misses the main point of my ask - which is to explain >> why a simpler scheme isn't sufficient. Simpler scheme aka not needed >> the xarray or any flavour of wait_token(). >> >> In other words it is obvious we have to wait for the invalidation ack, >> but not obvious why we need a complicated scheme. > The alternative being to simply serialise all TLB invalidation requests? > Thus, no complex tracking required as there is only one in flight at a > time? That seems inefficient and a potential performance impact if a > bunch of invalidations are required back to back. But given that the > current scheme is global invalidation only (no support for ranges / per > page invalidation yet), is it possible to get multiple back-to-back > requests? It is possible to get a stream of invalidation requests but all that come with put_pages() are serialized on the gt->tlb.invalidate_lock anyway. So for them only benefit with the complicated approach versus the dumb single wait queue is avoiding wake up storms. Then the second source of TLB invalidations is ggtt->invalidate(). I am not sure if those are frequent enough to warrant parallelism. Definitely shouldn't be for things like context images and ringbuffers. So I was asking if maybe framebuffers but don't know. Regards, Tvrtko
diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c index 4d7d88b92632b..a1f7bdc602996 100644 --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c @@ -206,22 +206,38 @@ static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt) intel_uncore_write_fw(uncore, GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN); } +static void guc_ggtt_ct_invalidate(struct intel_gt *gt) +{ + struct intel_uncore *uncore = gt->uncore; + intel_wakeref_t wakeref; + + with_intel_runtime_pm_if_active(uncore->rpm, wakeref) { + struct intel_guc *guc = >->uc.guc; + + intel_guc_invalidate_tlb_guc(guc); + } +} + static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) { struct drm_i915_private *i915 = ggtt->vm.i915; + struct intel_gt *gt; - gen8_ggtt_invalidate(ggtt); - - if (GRAPHICS_VER(i915) >= 12) { - struct intel_gt *gt; + if (!HAS_GUC_TLB_INVALIDATION(i915)) + gen8_ggtt_invalidate(ggtt); - list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) + list_for_each_entry(gt, &ggtt->gt_list, ggtt_link) { + if (HAS_GUC_TLB_INVALIDATION(i915) && + intel_guc_is_ready(>->uc.guc)) { + guc_ggtt_ct_invalidate(gt); + } else if (GRAPHICS_VER(i915) >= 12) { intel_uncore_write_fw(gt->uncore, GEN12_GUC_TLB_INV_CR, GEN12_GUC_TLB_INV_CR_INVALIDATE); - } else { - intel_uncore_write_fw(ggtt->vm.gt->uncore, - GEN8_GTCR, GEN8_GTCR_INVALIDATE); + } else { + intel_uncore_write_fw(gt->uncore, + GEN8_GTCR, GEN8_GTCR_INVALIDATE); + } } } @@ -1243,7 +1259,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) ggtt->vm.raw_insert_page = gen8_ggtt_insert_page; } - if (intel_uc_wants_guc(&ggtt->vm.gt->uc)) + if (intel_uc_wants_guc_submission(&ggtt->vm.gt->uc)) ggtt->invalidate = guc_ggtt_invalidate; else ggtt->invalidate = gen8_ggtt_invalidate; diff --git a/drivers/gpu/drm/i915/gt/intel_tlb.c b/drivers/gpu/drm/i915/gt/intel_tlb.c index 139608c30d978..1278622b9554d 100644 --- a/drivers/gpu/drm/i915/gt/intel_tlb.c +++ b/drivers/gpu/drm/i915/gt/intel_tlb.c @@ -12,6 +12,7 @@ #include "intel_gt_print.h" #include "intel_gt_regs.h" #include "intel_tlb.h" +#include "uc/intel_guc.h" /* * HW architecture suggest typical invalidation time at 40us, @@ -131,11 +132,23 @@ void intel_gt_invalidate_tlb_full(struct intel_gt *gt, u32 seqno) return; with_intel_gt_pm_if_awake(gt, wakeref) { + struct intel_guc *guc = >->uc.guc; + mutex_lock(>->tlb.invalidate_lock); if (tlb_seqno_passed(gt, seqno)) goto unlock; - mmio_invalidate_full(gt); + if (HAS_GUC_TLB_INVALIDATION(gt->i915)) { + /* + * Only perform GuC TLB invalidation if GuC is ready. + * If GuC is not ready, then there are no TLBs to + * invalidate. Ergo, skip invalidation. + */ + if (intel_guc_is_ready(guc)) + intel_guc_invalidate_tlb_engines(guc); + } else { + mmio_invalidate_full(gt); + } write_seqcount_invalidate(>->tlb.seqno); unlock: diff --git a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h index f359bef046e0b..33f253410d0c8 100644 --- a/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h +++ b/drivers/gpu/drm/i915/gt/uc/abi/guc_actions_abi.h @@ -138,6 +138,8 @@ enum intel_guc_action { INTEL_GUC_ACTION_REGISTER_CONTEXT_MULTI_LRC = 0x4601, INTEL_GUC_ACTION_CLIENT_SOFT_RESET = 0x5507, INTEL_GUC_ACTION_SET_ENG_UTIL_BUFF = 0x550A, + INTEL_GUC_ACTION_TLB_INVALIDATION = 0x7000, + INTEL_GUC_ACTION_TLB_INVALIDATION_DONE = 0x7001, INTEL_GUC_ACTION_STATE_CAPTURE_NOTIFICATION = 0x8002, INTEL_GUC_ACTION_NOTIFY_FLUSH_LOG_BUFFER_TO_FILE = 0x8003, INTEL_GUC_ACTION_NOTIFY_CRASH_DUMP_POSTED = 0x8004, @@ -181,4 +183,35 @@ enum intel_guc_state_capture_event_status { #define INTEL_GUC_STATE_CAPTURE_EVENT_STATUS_MASK 0x000000FF +#define INTEL_GUC_TLB_INVAL_TYPE_MASK REG_GENMASK(7, 0) +#define INTEL_GUC_TLB_INVAL_MODE_MASK REG_GENMASK(11, 8) +#define INTEL_GUC_TLB_INVAL_FLUSH_CACHE REG_BIT(31) + +enum intel_guc_tlb_invalidation_type { + INTEL_GUC_TLB_INVAL_ENGINES = 0x0, + INTEL_GUC_TLB_INVAL_GUC = 0x3, +}; + +/* + * 0: Heavy mode of Invalidation: + * The pipeline of the engine(s) for which the invalidation is targeted to is + * blocked, and all the in-flight transactions are guaranteed to be Globally + * Observed before completing the TLB invalidation + * 1: Lite mode of Invalidation: + * TLBs of the targeted engine(s) are immediately invalidated. + * In-flight transactions are NOT guaranteed to be Globally Observed before + * completing TLB invalidation. + * Light Invalidation Mode is to be used only when + * it can be guaranteed (by SW) that the address translations remain invariant + * for the in-flight transactions across the TLB invalidation. In other words, + * this mode can be used when the TLB invalidation is intended to clear out the + * stale cached translations that are no longer in use. Light Invalidation Mode + * is much faster than the Heavy Invalidation Mode, as it does not wait for the + * in-flight transactions to be GOd. + */ +enum intel_guc_tlb_inval_mode { + INTEL_GUC_TLB_INVAL_MODE_HEAVY = 0x0, + INTEL_GUC_TLB_INVAL_MODE_LITE = 0x1, +}; + #endif /* _ABI_GUC_ACTIONS_ABI_H */ diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h index 6c392bad29c19..88deb43bbdc48 100644 --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h @@ -79,6 +79,18 @@ struct intel_guc { */ atomic_t outstanding_submission_g2h; + /** @tlb_lookup: xarray to store all pending TLB invalidation requests */ + struct xarray tlb_lookup; + + /** + * @serial_slot: id to the initial waiter created in tlb_lookup, + * which is used only when failed to allocate new waiter. + */ + u32 serial_slot; + + /** @next_seqno: the next id (sequence no.) to allocate. */ + u32 next_seqno; + /** @interrupts: pointers to GuC interrupt-managing functions. */ struct { bool enabled; @@ -296,6 +308,11 @@ struct intel_guc { #define MAKE_GUC_VER_STRUCT(ver) MAKE_GUC_VER((ver).major, (ver).minor, (ver).patch) #define GUC_SUBMIT_VER(guc) MAKE_GUC_VER_STRUCT((guc)->submission_version) +struct intel_guc_tlb_wait { + struct wait_queue_head wq; + bool busy; +}; + static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) { return container_of(log, struct intel_guc, log); @@ -418,6 +435,11 @@ static inline bool intel_guc_is_supported(struct intel_guc *guc) return intel_uc_fw_is_supported(&guc->fw); } +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc); +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc); +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, + u32 len, u32 seqno); + static inline bool intel_guc_is_wanted(struct intel_guc *guc) { return intel_uc_fw_is_enabled(&guc->fw); diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c index 6e22af31513a5..e7ac8402245b7 100644 --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c @@ -1115,6 +1115,9 @@ static int ct_process_request(struct intel_guc_ct *ct, struct ct_incoming_msg *r case INTEL_GUC_ACTION_NOTIFY_EXCEPTION: ret = intel_guc_crash_process_msg(guc, action); break; + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: + ret = intel_guc_tlb_invalidation_done(guc, request->size, len, payload[0]); + break; default: ret = -EOPNOTSUPP; break; @@ -1186,6 +1189,7 @@ static int ct_handle_event(struct intel_guc_ct *ct, struct ct_incoming_msg *requ switch (action) { case INTEL_GUC_ACTION_SCHED_CONTEXT_MODE_DONE: case INTEL_GUC_ACTION_DEREGISTER_CONTEXT_DONE: + case INTEL_GUC_ACTION_TLB_INVALIDATION_DONE: g2h_release_space(ct, request->size); } diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h index b4d56eccfb1f0..a7c9874e122a3 100644 --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_fwif.h @@ -22,6 +22,7 @@ /* Payload length only i.e. don't include G2H header length */ #define G2H_LEN_DW_SCHED_CONTEXT_MODE_SET 2 #define G2H_LEN_DW_DEREGISTER_CONTEXT 1 +#define G2H_LEN_DW_INVALIDATE_TLB 1 #define GUC_CONTEXT_DISABLE 0 #define GUC_CONTEXT_ENABLE 1 diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c index 2cce5ec1ff00d..0c5ccda1b3e87 100644 --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c @@ -1798,9 +1798,11 @@ static void __guc_reset_context(struct intel_context *ce, intel_engine_mask_t st void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stalled) { + struct intel_guc_tlb_wait *wait; struct intel_context *ce; unsigned long index; unsigned long flags; + unsigned long i; if (unlikely(!guc_submission_initialized(guc))) { /* Reset called during driver load? GuC not yet initialised! */ @@ -1826,6 +1828,18 @@ void intel_guc_submission_reset(struct intel_guc *guc, intel_engine_mask_t stall /* GuC is blown away, drop all references to contexts */ xa_destroy(&guc->context_lookup); + + /* + * The full GT reset will have cleared the TLB caches and flushed the + * G2H message queue; we can release all the blocked waiters. + * + * This is safe to do unlocked because the xarray is not dependent + * on the GT reset, and there's a separate execution path for TLB + * invalidations on GT reset, and there's a large window of time + * between the GT reset and GuC becoming available. + */ + xa_for_each(&guc->tlb_lookup, i, wait) + wake_up(&wait->wq); } static void guc_cancel_context_requests(struct intel_context *ce) @@ -1948,6 +1962,50 @@ void intel_guc_submission_reset_finish(struct intel_guc *guc) static void destroyed_worker_func(struct work_struct *w); static void reset_fail_worker_func(struct work_struct *w); +static int init_tlb_lookup(struct intel_guc *guc) +{ + struct intel_guc_tlb_wait *wait; + int err; + + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) + return 0; + + xa_init_flags(&guc->tlb_lookup, XA_FLAGS_ALLOC); + + wait = kzalloc(sizeof(*wait), GFP_KERNEL); + if (!wait) + return -ENOMEM; + + init_waitqueue_head(&wait->wq); + + /* Preallocate a shared id for use under memory pressure. */ + err = xa_alloc_cyclic_irq(&guc->tlb_lookup, &guc->serial_slot, wait, + xa_limit_32b, &guc->next_seqno, GFP_KERNEL); + /* Only error if out of memory, not when busy (list full)*/ + if (err == -ENOMEM) { + kfree(wait); + return err; + } + + return 0; +} + +static void fini_tlb_lookup(struct intel_guc *guc) +{ + struct intel_guc_tlb_wait *wait; + + if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915)) + return; + + wait = xa_load(&guc->tlb_lookup, guc->serial_slot); + if (wait) { + guc_dbg(guc, "fini_tlb_lookup: Unexpected item in tlb_lookup\n"); + kfree(wait); + } + + xa_destroy(&guc->tlb_lookup); +} + /* * Set up the memory resources to be shared with the GuC (via the GGTT) * at firmware loading time. @@ -1966,11 +2024,15 @@ int intel_guc_submission_init(struct intel_guc *guc) return ret; } + ret = init_tlb_lookup(guc); + if (ret) + goto destroy_pool; + guc->submission_state.guc_ids_bitmap = bitmap_zalloc(NUMBER_MULTI_LRC_GUC_ID(guc), GFP_KERNEL); if (!guc->submission_state.guc_ids_bitmap) { ret = -ENOMEM; - goto destroy_pool; + goto destroy_tlb; } guc->timestamp.ping_delay = (POLL_TIME_CLKS / gt->clock_frequency + 1) * HZ; @@ -1979,9 +2041,10 @@ int intel_guc_submission_init(struct intel_guc *guc) return 0; +destroy_tlb: + fini_tlb_lookup(guc); destroy_pool: guc_lrc_desc_pool_destroy_v69(guc); - return ret; } @@ -1994,6 +2057,7 @@ void intel_guc_submission_fini(struct intel_guc *guc) guc_lrc_desc_pool_destroy_v69(guc); i915_sched_engine_put(guc->sched_engine); bitmap_free(guc->submission_state.guc_ids_bitmap); + fini_tlb_lookup(guc); guc->submission_initialized = false; } @@ -4624,6 +4688,138 @@ g2h_context_lookup(struct intel_guc *guc, u32 ctx_id) return ce; } +static void wait_wake_outstanding_tlb_g2h(struct intel_guc *guc, u32 seqno) +{ + struct intel_guc_tlb_wait *wait; + unsigned long flags; + + xa_lock_irqsave(&guc->tlb_lookup, flags); + wait = xa_load(&guc->tlb_lookup, seqno); + + if (wait) + wake_up(&wait->wq); + else + guc_dbg(guc, + "Stale TLB invalidation response with seqno %d\n", seqno); + + xa_unlock_irqrestore(&guc->tlb_lookup, flags); +} + +int intel_guc_tlb_invalidation_done(struct intel_guc *guc, u32 size, u32 len, u32 seqno) +{ + /* Check for underflow */ + if (unlikely(len < 1 || len > size)) + return -EPROTO; + + wait_wake_outstanding_tlb_g2h(guc, seqno); + return 0; +} + +static long must_wait_woken(struct wait_queue_entry *wq_entry, long timeout) +{ + /* + * This is equivalent to wait_woken() with the exception that + * we do not wake up early if the kthread task has been completed. + * As we are called from page reclaim in any task context, + * we may be invoked from stopped kthreads, but we *must* + * complete the wait from the HW . + * + * A second problem is that since we are called under reclaim + * and wait_woken() inspected the thread state, it makes an invalid + * assumption that all PF_KTHREAD tasks have set_kthread_struct() + * called upon them, and will trigger a GPF in is_kthread_should_stop(). + */ + do { + set_current_state(TASK_UNINTERRUPTIBLE); + if (wq_entry->flags & WQ_FLAG_WOKEN) + break; + + timeout = schedule_timeout(timeout); + } while (timeout); + __set_current_state(TASK_RUNNING); + + /* See wait_woken() and woken_wake_function() */ + smp_store_mb(wq_entry->flags, wq_entry->flags & ~WQ_FLAG_WOKEN); + + return timeout; +} + +static int guc_send_invalidate_tlb(struct intel_guc *guc, enum intel_guc_tlb_inval_mode type) +{ + struct intel_guc_tlb_wait _wq, *wq = &_wq; + DEFINE_WAIT_FUNC(wait, woken_wake_function); + int err; + u32 seqno; + u32 action[] = { + INTEL_GUC_ACTION_TLB_INVALIDATION, + 0, + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_TYPE_MASK, type) | + REG_FIELD_PREP(INTEL_GUC_TLB_INVAL_MODE_MASK, + INTEL_GUC_TLB_INVAL_MODE_HEAVY) | + INTEL_GUC_TLB_INVAL_FLUSH_CACHE, + }; + u32 size = ARRAY_SIZE(action); + + init_waitqueue_head(&_wq.wq); + + if (xa_alloc_cyclic_irq(&guc->tlb_lookup, &seqno, wq, + xa_limit_32b, &guc->next_seqno, + GFP_ATOMIC | __GFP_NOWARN) < 0) { + /* Under severe memory pressure? Serialise TLB allocations */ + xa_lock_irq(&guc->tlb_lookup); + wq = xa_load(&guc->tlb_lookup, guc->serial_slot); + wait_event_lock_irq(wq->wq, + !READ_ONCE(wq->busy), + guc->tlb_lookup.xa_lock); + /* + * Update wq->busy under lock to ensure only one waiter can + * issue the TLB invalidation command using the serial slot at a + * time. The condition is set to true before releasing the lock + * so that other caller continue to wait until woken up again. + */ + wq->busy = true; + xa_unlock_irq(&guc->tlb_lookup); + + seqno = guc->serial_slot; + } + + action[1] = seqno; + + add_wait_queue(&wq->wq, &wait); + + /* + * This is a critical reclaim path and thus we must loop here: + * We cannot block for anything that is on the GPU. + */ + err = intel_guc_send_busy_loop(guc, action, size, G2H_LEN_DW_INVALIDATE_TLB, true); + if (err) + goto out; + + if (!must_wait_woken(&wait, intel_guc_ct_expected_delay(&guc->ct))) { + guc_err(guc, + "TLB invalidation response timed out for seqno %u\n", seqno); + err = -ETIME; + } +out: + remove_wait_queue(&wq->wq, &wait); + if (seqno != guc->serial_slot) + xa_erase_irq(&guc->tlb_lookup, seqno); + + return err; +} + +/* Full TLB invalidation */ +int intel_guc_invalidate_tlb_engines(struct intel_guc *guc) +{ + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_ENGINES); +} + +/* GuC TLB Invalidation: Invalidate the TLB's of GuC itself. */ +int intel_guc_invalidate_tlb_guc(struct intel_guc *guc) +{ + return guc_send_invalidate_tlb(guc, INTEL_GUC_TLB_INVAL_GUC); +} + int intel_guc_deregister_done_process_msg(struct intel_guc *guc, const u32 *msg, u32 len)