diff mbox series

[05/27] drm/i915: Add GT PM unpark worker

Message ID 20210820224446.30620-6-matthew.brost@intel.com (mailing list archive)
State New, archived
Headers show
Series Parallel submission aka multi-bb execbuf | expand

Commit Message

Matthew Brost Aug. 20, 2021, 10:44 p.m. UTC
Sometimes it is desirable to queue work up for later if the GT PM isn't
held and run that work on next GT PM unpark.

Implemented with a list in the GT of all pending work, workqueues in
the list, a callback to add a workqueue to the list, and finally a
wakeref post_get callback that iterates / drains the list + queues the
workqueues.

First user of this is deregistration of GuC contexts.

Signed-off-by: Matthew Brost <matthew.brost@intel.com>
---
 drivers/gpu/drm/i915/Makefile                 |  1 +
 drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
 drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
 .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
 .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
 drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
 drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
 .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
 drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
 drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
 10 files changed, 119 insertions(+), 7 deletions(-)
 create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
 create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h

Comments

John Harrison Sept. 9, 2021, 10:36 p.m. UTC | #1
On 8/20/2021 15:44, Matthew Brost wrote:
> Sometimes it is desirable to queue work up for later if the GT PM isn't
> held and run that work on next GT PM unpark.
What is the reason for doing this? Why is it important? Why not just 
take the GT PM at the time the work is requested?

>
> Implemented with a list in the GT of all pending work, workqueues in
> the list, a callback to add a workqueue to the list, and finally a
> wakeref post_get callback that iterates / drains the list + queues the
> workqueues.
>
> First user of this is deregistration of GuC contexts.
This statement should be in the first paragraph but needs to be more 
detailed - why is it necessary to add all this extra complexity rather 
than just taking the GT PM?


>
> Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> ---
>   drivers/gpu/drm/i915/Makefile                 |  1 +
>   drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
>   drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
>   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
>   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
>   drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
>   drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
>   .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
>   drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
>   drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
>   10 files changed, 119 insertions(+), 7 deletions(-)
>   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
>   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
>
> diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> index 642a5b5a1b81..579bdc069f25 100644
> --- a/drivers/gpu/drm/i915/Makefile
> +++ b/drivers/gpu/drm/i915/Makefile
> @@ -103,6 +103,7 @@ gt-y += \
>   	gt/intel_gt_clock_utils.o \
>   	gt/intel_gt_irq.o \
>   	gt/intel_gt_pm.o \
> +	gt/intel_gt_pm_unpark_work.o \
>   	gt/intel_gt_pm_irq.o \
>   	gt/intel_gt_requests.o \
>   	gt/intel_gtt.o \
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
> index 62d40c986642..7e690e74baa2 100644
> --- a/drivers/gpu/drm/i915/gt/intel_gt.c
> +++ b/drivers/gpu/drm/i915/gt/intel_gt.c
> @@ -29,6 +29,9 @@ void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
>   
>   	spin_lock_init(&gt->irq_lock);
>   
> +	spin_lock_init(&gt->pm_unpark_work_lock);
> +	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
> +
>   	INIT_LIST_HEAD(&gt->closed_vma);
>   	spin_lock_init(&gt->closed_lock);
>   
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> index dea8e2479897..564c11a3748b 100644
> --- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> @@ -90,6 +90,13 @@ static int __gt_unpark(struct intel_wakeref *wf)
>   	return 0;
>   }
>   
> +static void __gt_unpark_work_queue(struct intel_wakeref *wf)
> +{
> +	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> +
> +	intel_gt_pm_unpark_work_queue(gt);
> +}
> +
>   static int __gt_park(struct intel_wakeref *wf)
>   {
>   	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> @@ -118,6 +125,7 @@ static int __gt_park(struct intel_wakeref *wf)
>   
>   static const struct intel_wakeref_ops wf_ops = {
>   	.get = __gt_unpark,
> +	.post_get = __gt_unpark_work_queue,
>   	.put = __gt_park,
>   };
>   
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> new file mode 100644
> index 000000000000..23162dbd0c35
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> @@ -0,0 +1,35 @@
> +// SPDX-License-Identifier: MIT
> +/*
> + * Copyright © 2021 Intel Corporation
> + */
> +
> +#include "i915_drv.h"
> +#include "intel_runtime_pm.h"
> +#include "intel_gt_pm.h"
> +
> +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
> +{
> +	struct intel_gt_pm_unpark_work *work, *next;
> +	unsigned long flags;
> +
> +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> +	list_for_each_entry_safe(work, next,
> +				 &gt->pm_unpark_work_list, link) {
> +		list_del_init(&work->link);
> +		queue_work(system_unbound_wq, &work->worker);
> +	}
> +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> +}
> +
> +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> +				 struct intel_gt_pm_unpark_work *work)
> +{
> +	unsigned long flags;
> +
> +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> +	if (intel_gt_pm_is_awake(gt))
> +		queue_work(system_unbound_wq, &work->worker);
> +	else if (list_empty(&work->link))
> +		list_add_tail(&work->link, &gt->pm_unpark_work_list);
> +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> +}
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> new file mode 100644
> index 000000000000..eaf1dc313aa2
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> @@ -0,0 +1,40 @@
> +/* SPDX-License-Identifier: MIT */
> +/*
> + * Copyright © 2021 Intel Corporation
> + */
> +
> +#ifndef INTEL_GT_PM_UNPARK_WORK_H
> +#define INTEL_GT_PM_UNPARK_WORK_H
> +
> +#include <linux/list.h>
> +#include <linux/workqueue.h>
> +
> +struct intel_gt;
> +
> +/**
> + * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
> + */
> +struct intel_gt_pm_unpark_work {
> +	/**
> +	 * @link: link into gt->pm_unpark_work_list of workers that need to be
> +	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
> +	 */
> +	struct list_head link;
> +	/** @worker: will be scheduled when GT unparked */
> +	struct work_struct worker;
> +};
> +
> +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
> +
> +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> +				 struct intel_gt_pm_unpark_work *work);
> +
> +static inline void
> +intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
> +			     work_func_t fn)
> +{
> +	INIT_LIST_HEAD(&work->link);
> +	INIT_WORK(&work->worker, fn);
> +}
> +
> +#endif /* INTEL_GT_PM_UNPARK_WORK_H */
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> index a81e21bf1bd1..4480312f0add 100644
> --- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> @@ -96,6 +96,16 @@ struct intel_gt {
>   	struct intel_wakeref wakeref;
>   	atomic_t user_wakeref;
>   
> +	/**
> +	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
> +	 * unparked, protected by pm_unpark_work_lock
> +	 */
> +	struct list_head pm_unpark_work_list;
> +	/**
> +	 * @pm_unpark_work_lock: protects pm_unpark_work_list
> +	 */
> +	spinlock_t pm_unpark_work_lock;
> +
>   	struct list_head closed_vma;
>   	spinlock_t closed_lock; /* guards the list of closed_vma */
>   
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> index 7358883f1540..023953e77553 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> @@ -19,6 +19,7 @@
>   #include "intel_uc_fw.h"
>   #include "i915_utils.h"
>   #include "i915_vma.h"
> +#include "gt/intel_gt_pm_unpark_work.h"
>   
>   struct __guc_ads_blob;
>   
> @@ -78,11 +79,12 @@ struct intel_guc {
>   		 */
>   		struct list_head destroyed_contexts;
>   		/**
> -		 * @destroyed_worker: worker to deregister contexts, need as we
> +		 * @destroyed_worker: Worker to deregister contexts, need as we
>   		 * need to take a GT PM reference and can't from destroy
> -		 * function as it might be in an atomic context (no sleeping)
> +		 * function as it might be in an atomic context (no sleeping).
These corrections should be squashed into the previous patch that added 
these comments in the first place.

John.


> +		 * Worker only issues deregister when GT is unparked.
>   		 */
> -		struct work_struct destroyed_worker;
> +		struct intel_gt_pm_unpark_work destroyed_worker;
>   	} submission_state;
>   
>   	bool submission_supported;
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> index f835e06e5f9f..dbf919801de2 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> @@ -1135,7 +1135,8 @@ int intel_guc_submission_init(struct intel_guc *guc)
>   	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
>   	ida_init(&guc->submission_state.guc_ids);
>   	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
> -	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
> +	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
> +				     destroyed_worker_func);
>   
>   	return 0;
>   }
> @@ -1942,13 +1943,18 @@ static void deregister_destroyed_contexts(struct intel_guc *guc)
>   
>   static void destroyed_worker_func(struct work_struct *w)
>   {
> -	struct intel_guc *guc = container_of(w, struct intel_guc,
> +	struct intel_gt_pm_unpark_work *destroyed_worker =
> +		container_of(w, struct intel_gt_pm_unpark_work, worker);
> +	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
>   					     submission_state.destroyed_worker);
>   	struct intel_gt *gt = guc_to_gt(guc);
>   	int tmp;
>   
> -	with_intel_gt_pm(gt, tmp)
> +	with_intel_gt_pm_if_awake(gt, tmp)
>   		deregister_destroyed_contexts(guc);
> +
> +	if (!list_empty(&guc->submission_state.destroyed_contexts))
> +		intel_gt_pm_unpark_work_add(gt, destroyed_worker);
>   }
>   
>   static void guc_context_destroy(struct kref *kref)
> @@ -1985,7 +1991,8 @@ static void guc_context_destroy(struct kref *kref)
>   	 * take the GT PM for the first time which isn't allowed from an atomic
>   	 * context.
>   	 */
> -	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
> +	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
> +				    &guc->submission_state.destroyed_worker);
>   }
>   
>   static int guc_context_alloc(struct intel_context *ce)
> diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
> index dfd87d082218..282fc4f312e3 100644
> --- a/drivers/gpu/drm/i915/intel_wakeref.c
> +++ b/drivers/gpu/drm/i915/intel_wakeref.c
> @@ -24,6 +24,8 @@ static void rpm_put(struct intel_wakeref *wf)
>   
>   int __intel_wakeref_get_first(struct intel_wakeref *wf)
>   {
> +	bool do_post = false;
> +
>   	/*
>   	 * Treat get/put as different subclasses, as we may need to run
>   	 * the put callback from under the shrinker and do not want to
> @@ -44,8 +46,11 @@ int __intel_wakeref_get_first(struct intel_wakeref *wf)
>   		}
>   
>   		smp_mb__before_atomic(); /* release wf->count */
> +		do_post = true;
>   	}
>   	atomic_inc(&wf->count);
> +	if (do_post && wf->ops->post_get)
> +		wf->ops->post_get(wf);
>   	mutex_unlock(&wf->mutex);
>   
>   	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
> diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
> index 545c8f277c46..ef7e6a698e8a 100644
> --- a/drivers/gpu/drm/i915/intel_wakeref.h
> +++ b/drivers/gpu/drm/i915/intel_wakeref.h
> @@ -30,6 +30,7 @@ typedef depot_stack_handle_t intel_wakeref_t;
>   
>   struct intel_wakeref_ops {
>   	int (*get)(struct intel_wakeref *wf);
> +	void (*post_get)(struct intel_wakeref *wf);
>   	int (*put)(struct intel_wakeref *wf);
>   };
>
Matthew Brost Sept. 10, 2021, 12:34 a.m. UTC | #2
On Thu, Sep 09, 2021 at 03:36:44PM -0700, John Harrison wrote:
> On 8/20/2021 15:44, Matthew Brost wrote:
> > Sometimes it is desirable to queue work up for later if the GT PM isn't
> > held and run that work on next GT PM unpark.
> What is the reason for doing this? Why is it important? Why not just take
> the GT PM at the time the work is requested?
>

This was suggestion from a long time back, don't take GT PM (waking the
GPU) if is is idle. I believe Daniele suggested this a couple years ago
/ we still have this FIXME comment in DII.
 
> > 
> > Implemented with a list in the GT of all pending work, workqueues in
> > the list, a callback to add a workqueue to the list, and finally a
> > wakeref post_get callback that iterates / drains the list + queues the
> > workqueues.
> > 
> > First user of this is deregistration of GuC contexts.
> This statement should be in the first paragraph but needs to be more
> detailed - why is it necessary to add all this extra complexity rather than
> just taking the GT PM?
> 
>

This is just an optimization - don't take the GT PM when deregistering a
context if the GT is idle. If you think it is too complex we can can
delete this but IMO it really isn't any more complex than the previous
patch as either way we need a list + worker. Also once we have this
framework in place we might be able to find other users of this.
 
> > 
> > Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> > ---
> >   drivers/gpu/drm/i915/Makefile                 |  1 +
> >   drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
> >   drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
> >   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
> >   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
> >   drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
> >   drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
> >   .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
> >   drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
> >   drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
> >   10 files changed, 119 insertions(+), 7 deletions(-)
> >   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> >   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > 
> > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> > index 642a5b5a1b81..579bdc069f25 100644
> > --- a/drivers/gpu/drm/i915/Makefile
> > +++ b/drivers/gpu/drm/i915/Makefile
> > @@ -103,6 +103,7 @@ gt-y += \
> >   	gt/intel_gt_clock_utils.o \
> >   	gt/intel_gt_irq.o \
> >   	gt/intel_gt_pm.o \
> > +	gt/intel_gt_pm_unpark_work.o \
> >   	gt/intel_gt_pm_irq.o \
> >   	gt/intel_gt_requests.o \
> >   	gt/intel_gtt.o \
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
> > index 62d40c986642..7e690e74baa2 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gt.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt.c
> > @@ -29,6 +29,9 @@ void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
> >   	spin_lock_init(&gt->irq_lock);
> > +	spin_lock_init(&gt->pm_unpark_work_lock);
> > +	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
> > +
> >   	INIT_LIST_HEAD(&gt->closed_vma);
> >   	spin_lock_init(&gt->closed_lock);
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > index dea8e2479897..564c11a3748b 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > @@ -90,6 +90,13 @@ static int __gt_unpark(struct intel_wakeref *wf)
> >   	return 0;
> >   }
> > +static void __gt_unpark_work_queue(struct intel_wakeref *wf)
> > +{
> > +	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> > +
> > +	intel_gt_pm_unpark_work_queue(gt);
> > +}
> > +
> >   static int __gt_park(struct intel_wakeref *wf)
> >   {
> >   	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> > @@ -118,6 +125,7 @@ static int __gt_park(struct intel_wakeref *wf)
> >   static const struct intel_wakeref_ops wf_ops = {
> >   	.get = __gt_unpark,
> > +	.post_get = __gt_unpark_work_queue,
> >   	.put = __gt_park,
> >   };
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > new file mode 100644
> > index 000000000000..23162dbd0c35
> > --- /dev/null
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > @@ -0,0 +1,35 @@
> > +// SPDX-License-Identifier: MIT
> > +/*
> > + * Copyright © 2021 Intel Corporation
> > + */
> > +
> > +#include "i915_drv.h"
> > +#include "intel_runtime_pm.h"
> > +#include "intel_gt_pm.h"
> > +
> > +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
> > +{
> > +	struct intel_gt_pm_unpark_work *work, *next;
> > +	unsigned long flags;
> > +
> > +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> > +	list_for_each_entry_safe(work, next,
> > +				 &gt->pm_unpark_work_list, link) {
> > +		list_del_init(&work->link);
> > +		queue_work(system_unbound_wq, &work->worker);
> > +	}
> > +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> > +}
> > +
> > +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> > +				 struct intel_gt_pm_unpark_work *work)
> > +{
> > +	unsigned long flags;
> > +
> > +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> > +	if (intel_gt_pm_is_awake(gt))
> > +		queue_work(system_unbound_wq, &work->worker);
> > +	else if (list_empty(&work->link))
> > +		list_add_tail(&work->link, &gt->pm_unpark_work_list);
> > +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> > +}
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > new file mode 100644
> > index 000000000000..eaf1dc313aa2
> > --- /dev/null
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > @@ -0,0 +1,40 @@
> > +/* SPDX-License-Identifier: MIT */
> > +/*
> > + * Copyright © 2021 Intel Corporation
> > + */
> > +
> > +#ifndef INTEL_GT_PM_UNPARK_WORK_H
> > +#define INTEL_GT_PM_UNPARK_WORK_H
> > +
> > +#include <linux/list.h>
> > +#include <linux/workqueue.h>
> > +
> > +struct intel_gt;
> > +
> > +/**
> > + * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
> > + */
> > +struct intel_gt_pm_unpark_work {
> > +	/**
> > +	 * @link: link into gt->pm_unpark_work_list of workers that need to be
> > +	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
> > +	 */
> > +	struct list_head link;
> > +	/** @worker: will be scheduled when GT unparked */
> > +	struct work_struct worker;
> > +};
> > +
> > +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
> > +
> > +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> > +				 struct intel_gt_pm_unpark_work *work);
> > +
> > +static inline void
> > +intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
> > +			     work_func_t fn)
> > +{
> > +	INIT_LIST_HEAD(&work->link);
> > +	INIT_WORK(&work->worker, fn);
> > +}
> > +
> > +#endif /* INTEL_GT_PM_UNPARK_WORK_H */
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > index a81e21bf1bd1..4480312f0add 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > @@ -96,6 +96,16 @@ struct intel_gt {
> >   	struct intel_wakeref wakeref;
> >   	atomic_t user_wakeref;
> > +	/**
> > +	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
> > +	 * unparked, protected by pm_unpark_work_lock
> > +	 */
> > +	struct list_head pm_unpark_work_list;
> > +	/**
> > +	 * @pm_unpark_work_lock: protects pm_unpark_work_list
> > +	 */
> > +	spinlock_t pm_unpark_work_lock;
> > +
> >   	struct list_head closed_vma;
> >   	spinlock_t closed_lock; /* guards the list of closed_vma */
> > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > index 7358883f1540..023953e77553 100644
> > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > @@ -19,6 +19,7 @@
> >   #include "intel_uc_fw.h"
> >   #include "i915_utils.h"
> >   #include "i915_vma.h"
> > +#include "gt/intel_gt_pm_unpark_work.h"
> >   struct __guc_ads_blob;
> > @@ -78,11 +79,12 @@ struct intel_guc {
> >   		 */
> >   		struct list_head destroyed_contexts;
> >   		/**
> > -		 * @destroyed_worker: worker to deregister contexts, need as we
> > +		 * @destroyed_worker: Worker to deregister contexts, need as we
> >   		 * need to take a GT PM reference and can't from destroy
> > -		 * function as it might be in an atomic context (no sleeping)
> > +		 * function as it might be in an atomic context (no sleeping).
> These corrections should be squashed into the previous patch that added
> these comments in the first place.
> 

Yep.

Matt

> John.
> 
> 
> > +		 * Worker only issues deregister when GT is unparked.
> >   		 */
> > -		struct work_struct destroyed_worker;
> > +		struct intel_gt_pm_unpark_work destroyed_worker;
> >   	} submission_state;
> >   	bool submission_supported;
> > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > index f835e06e5f9f..dbf919801de2 100644
> > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > @@ -1135,7 +1135,8 @@ int intel_guc_submission_init(struct intel_guc *guc)
> >   	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
> >   	ida_init(&guc->submission_state.guc_ids);
> >   	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
> > -	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
> > +	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
> > +				     destroyed_worker_func);
> >   	return 0;
> >   }
> > @@ -1942,13 +1943,18 @@ static void deregister_destroyed_contexts(struct intel_guc *guc)
> >   static void destroyed_worker_func(struct work_struct *w)
> >   {
> > -	struct intel_guc *guc = container_of(w, struct intel_guc,
> > +	struct intel_gt_pm_unpark_work *destroyed_worker =
> > +		container_of(w, struct intel_gt_pm_unpark_work, worker);
> > +	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
> >   					     submission_state.destroyed_worker);
> >   	struct intel_gt *gt = guc_to_gt(guc);
> >   	int tmp;
> > -	with_intel_gt_pm(gt, tmp)
> > +	with_intel_gt_pm_if_awake(gt, tmp)
> >   		deregister_destroyed_contexts(guc);
> > +
> > +	if (!list_empty(&guc->submission_state.destroyed_contexts))
> > +		intel_gt_pm_unpark_work_add(gt, destroyed_worker);
> >   }
> >   static void guc_context_destroy(struct kref *kref)
> > @@ -1985,7 +1991,8 @@ static void guc_context_destroy(struct kref *kref)
> >   	 * take the GT PM for the first time which isn't allowed from an atomic
> >   	 * context.
> >   	 */
> > -	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
> > +	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
> > +				    &guc->submission_state.destroyed_worker);
> >   }
> >   static int guc_context_alloc(struct intel_context *ce)
> > diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
> > index dfd87d082218..282fc4f312e3 100644
> > --- a/drivers/gpu/drm/i915/intel_wakeref.c
> > +++ b/drivers/gpu/drm/i915/intel_wakeref.c
> > @@ -24,6 +24,8 @@ static void rpm_put(struct intel_wakeref *wf)
> >   int __intel_wakeref_get_first(struct intel_wakeref *wf)
> >   {
> > +	bool do_post = false;
> > +
> >   	/*
> >   	 * Treat get/put as different subclasses, as we may need to run
> >   	 * the put callback from under the shrinker and do not want to
> > @@ -44,8 +46,11 @@ int __intel_wakeref_get_first(struct intel_wakeref *wf)
> >   		}
> >   		smp_mb__before_atomic(); /* release wf->count */
> > +		do_post = true;
> >   	}
> >   	atomic_inc(&wf->count);
> > +	if (do_post && wf->ops->post_get)
> > +		wf->ops->post_get(wf);
> >   	mutex_unlock(&wf->mutex);
> >   	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
> > diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
> > index 545c8f277c46..ef7e6a698e8a 100644
> > --- a/drivers/gpu/drm/i915/intel_wakeref.h
> > +++ b/drivers/gpu/drm/i915/intel_wakeref.h
> > @@ -30,6 +30,7 @@ typedef depot_stack_handle_t intel_wakeref_t;
> >   struct intel_wakeref_ops {
> >   	int (*get)(struct intel_wakeref *wf);
> > +	void (*post_get)(struct intel_wakeref *wf);
> >   	int (*put)(struct intel_wakeref *wf);
> >   };
>
Tvrtko Ursulin Sept. 10, 2021, 8:36 a.m. UTC | #3
On 20/08/2021 23:44, Matthew Brost wrote:
> Sometimes it is desirable to queue work up for later if the GT PM isn't
> held and run that work on next GT PM unpark.

Sounds maybe plausible, but it depends how much work can happen on 
unpark and whether it can have too much of a negative impact on latency 
for interactive loads? Or from a reverse angle, why the work wouldn't be 
done on parking?

Also what kind of mechanism for dealing with too much stuff being put on 
this list you have? Can there be pressure which triggers (or would need 
to trigger) these deregistrations to happen at runtime (no park/unpark 
transitions)?

> Implemented with a list in the GT of all pending work, workqueues in
> the list, a callback to add a workqueue to the list, and finally a
> wakeref post_get callback that iterates / drains the list + queues the
> workqueues.
> 
> First user of this is deregistration of GuC contexts.

Does first imply there are more incoming?

> Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> ---
>   drivers/gpu/drm/i915/Makefile                 |  1 +
>   drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
>   drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
>   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
>   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
>   drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
>   drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
>   .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
>   drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
>   drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
>   10 files changed, 119 insertions(+), 7 deletions(-)
>   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
>   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> 
> diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> index 642a5b5a1b81..579bdc069f25 100644
> --- a/drivers/gpu/drm/i915/Makefile
> +++ b/drivers/gpu/drm/i915/Makefile
> @@ -103,6 +103,7 @@ gt-y += \
>   	gt/intel_gt_clock_utils.o \
>   	gt/intel_gt_irq.o \
>   	gt/intel_gt_pm.o \
> +	gt/intel_gt_pm_unpark_work.o \
>   	gt/intel_gt_pm_irq.o \
>   	gt/intel_gt_requests.o \
>   	gt/intel_gtt.o \
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
> index 62d40c986642..7e690e74baa2 100644
> --- a/drivers/gpu/drm/i915/gt/intel_gt.c
> +++ b/drivers/gpu/drm/i915/gt/intel_gt.c
> @@ -29,6 +29,9 @@ void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
>   
>   	spin_lock_init(&gt->irq_lock);
>   
> +	spin_lock_init(&gt->pm_unpark_work_lock);
> +	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
> +
>   	INIT_LIST_HEAD(&gt->closed_vma);
>   	spin_lock_init(&gt->closed_lock);
>   
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> index dea8e2479897..564c11a3748b 100644
> --- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> @@ -90,6 +90,13 @@ static int __gt_unpark(struct intel_wakeref *wf)
>   	return 0;
>   }
>   
> +static void __gt_unpark_work_queue(struct intel_wakeref *wf)
> +{
> +	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> +
> +	intel_gt_pm_unpark_work_queue(gt);
> +}
> +
>   static int __gt_park(struct intel_wakeref *wf)
>   {
>   	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> @@ -118,6 +125,7 @@ static int __gt_park(struct intel_wakeref *wf)
>   
>   static const struct intel_wakeref_ops wf_ops = {
>   	.get = __gt_unpark,
> +	.post_get = __gt_unpark_work_queue,
>   	.put = __gt_park,
>   };
>   
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> new file mode 100644
> index 000000000000..23162dbd0c35
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> @@ -0,0 +1,35 @@
> +// SPDX-License-Identifier: MIT
> +/*
> + * Copyright © 2021 Intel Corporation
> + */
> +
> +#include "i915_drv.h"
> +#include "intel_runtime_pm.h"
> +#include "intel_gt_pm.h"
> +
> +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
> +{
> +	struct intel_gt_pm_unpark_work *work, *next;
> +	unsigned long flags;
> +
> +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> +	list_for_each_entry_safe(work, next,
> +				 &gt->pm_unpark_work_list, link) {
> +		list_del_init(&work->link);
> +		queue_work(system_unbound_wq, &work->worker);
> +	}
> +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> +}
> +
> +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> +				 struct intel_gt_pm_unpark_work *work)
> +{
> +	unsigned long flags;
> +
> +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> +	if (intel_gt_pm_is_awake(gt))
> +		queue_work(system_unbound_wq, &work->worker);
> +	else if (list_empty(&work->link))

What's the list_empty check for, something can race by design?

> +		list_add_tail(&work->link, &gt->pm_unpark_work_list);
> +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> +}
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> new file mode 100644
> index 000000000000..eaf1dc313aa2
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> @@ -0,0 +1,40 @@
> +/* SPDX-License-Identifier: MIT */
> +/*
> + * Copyright © 2021 Intel Corporation
> + */
> +
> +#ifndef INTEL_GT_PM_UNPARK_WORK_H
> +#define INTEL_GT_PM_UNPARK_WORK_H
> +
> +#include <linux/list.h>
> +#include <linux/workqueue.h>
> +
> +struct intel_gt;
> +
> +/**
> + * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
> + */
> +struct intel_gt_pm_unpark_work {
> +	/**
> +	 * @link: link into gt->pm_unpark_work_list of workers that need to be
> +	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
> +	 */
> +	struct list_head link;
> +	/** @worker: will be scheduled when GT unparked */
> +	struct work_struct worker;
> +};
> +
> +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
> +
> +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> +				 struct intel_gt_pm_unpark_work *work);
> +
> +static inline void
> +intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
> +			     work_func_t fn)
> +{
> +	INIT_LIST_HEAD(&work->link);
> +	INIT_WORK(&work->worker, fn);
> +}
> +
> +#endif /* INTEL_GT_PM_UNPARK_WORK_H */
> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> index a81e21bf1bd1..4480312f0add 100644
> --- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
> +++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> @@ -96,6 +96,16 @@ struct intel_gt {
>   	struct intel_wakeref wakeref;
>   	atomic_t user_wakeref;
>   
> +	/**
> +	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
> +	 * unparked, protected by pm_unpark_work_lock
> +	 */
> +	struct list_head pm_unpark_work_list;
> +	/**
> +	 * @pm_unpark_work_lock: protects pm_unpark_work_list
> +	 */
> +	spinlock_t pm_unpark_work_lock;
> +
>   	struct list_head closed_vma;
>   	spinlock_t closed_lock; /* guards the list of closed_vma */
>   
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> index 7358883f1540..023953e77553 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> @@ -19,6 +19,7 @@
>   #include "intel_uc_fw.h"
>   #include "i915_utils.h"
>   #include "i915_vma.h"
> +#include "gt/intel_gt_pm_unpark_work.h"
>   
>   struct __guc_ads_blob;
>   
> @@ -78,11 +79,12 @@ struct intel_guc {
>   		 */
>   		struct list_head destroyed_contexts;
>   		/**
> -		 * @destroyed_worker: worker to deregister contexts, need as we
> +		 * @destroyed_worker: Worker to deregister contexts, need as we
>   		 * need to take a GT PM reference and can't from destroy
> -		 * function as it might be in an atomic context (no sleeping)
> +		 * function as it might be in an atomic context (no sleeping).
> +		 * Worker only issues deregister when GT is unparked.
>   		 */
> -		struct work_struct destroyed_worker;
> +		struct intel_gt_pm_unpark_work destroyed_worker;
>   	} submission_state;
>   
>   	bool submission_supported;
> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> index f835e06e5f9f..dbf919801de2 100644
> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> @@ -1135,7 +1135,8 @@ int intel_guc_submission_init(struct intel_guc *guc)
>   	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
>   	ida_init(&guc->submission_state.guc_ids);
>   	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
> -	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
> +	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
> +				     destroyed_worker_func);
>   
>   	return 0;
>   }
> @@ -1942,13 +1943,18 @@ static void deregister_destroyed_contexts(struct intel_guc *guc)
>   
>   static void destroyed_worker_func(struct work_struct *w)
>   {
> -	struct intel_guc *guc = container_of(w, struct intel_guc,
> +	struct intel_gt_pm_unpark_work *destroyed_worker =
> +		container_of(w, struct intel_gt_pm_unpark_work, worker);
> +	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
>   					     submission_state.destroyed_worker);
>   	struct intel_gt *gt = guc_to_gt(guc);
>   	int tmp;
>   
> -	with_intel_gt_pm(gt, tmp)
> +	with_intel_gt_pm_if_awake(gt, tmp)
>   		deregister_destroyed_contexts(guc);
> +
> +	if (!list_empty(&guc->submission_state.destroyed_contexts))
> +		intel_gt_pm_unpark_work_add(gt, destroyed_worker);

This is the worker itself, right?

There's a "if awake" here followed by another "if awake" inside 
intel_gt_pm_unpark_work_add which raises questions.

Second question is what's the list_empty for - why is the state of the 
list itself relevant to a single worker deciding whether to re-add 
itself to it or not? And is there a lock protecting this list?

On the overall it feels questionable to have unpark work which 
apparently can race with subsequent parking. Presumably you cannot have 
it run sync on unpark due execution context issues?

>   }
>   
>   static void guc_context_destroy(struct kref *kref)
> @@ -1985,7 +1991,8 @@ static void guc_context_destroy(struct kref *kref)
>   	 * take the GT PM for the first time which isn't allowed from an atomic
>   	 * context.
>   	 */
> -	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
> +	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
> +				    &guc->submission_state.destroyed_worker);
>   }
>   
>   static int guc_context_alloc(struct intel_context *ce)
> diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
> index dfd87d082218..282fc4f312e3 100644
> --- a/drivers/gpu/drm/i915/intel_wakeref.c
> +++ b/drivers/gpu/drm/i915/intel_wakeref.c
> @@ -24,6 +24,8 @@ static void rpm_put(struct intel_wakeref *wf)
>   
>   int __intel_wakeref_get_first(struct intel_wakeref *wf)
>   {
> +	bool do_post = false;
> +
>   	/*
>   	 * Treat get/put as different subclasses, as we may need to run
>   	 * the put callback from under the shrinker and do not want to
> @@ -44,8 +46,11 @@ int __intel_wakeref_get_first(struct intel_wakeref *wf)
>   		}
>   
>   		smp_mb__before_atomic(); /* release wf->count */
> +		do_post = true;
>   	}
>   	atomic_inc(&wf->count);
> +	if (do_post && wf->ops->post_get)
> +		wf->ops->post_get(wf);

You want this hook under the wf->mutex and why?

Regards,

Tvrtko

>   	mutex_unlock(&wf->mutex);
>   
>   	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
> diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
> index 545c8f277c46..ef7e6a698e8a 100644
> --- a/drivers/gpu/drm/i915/intel_wakeref.h
> +++ b/drivers/gpu/drm/i915/intel_wakeref.h
> @@ -30,6 +30,7 @@ typedef depot_stack_handle_t intel_wakeref_t;
>   
>   struct intel_wakeref_ops {
>   	int (*get)(struct intel_wakeref *wf);
> +	void (*post_get)(struct intel_wakeref *wf);
>   	int (*put)(struct intel_wakeref *wf);
>   };
>   
>
Matthew Brost Sept. 10, 2021, 8:09 p.m. UTC | #4
On Fri, Sep 10, 2021 at 09:36:17AM +0100, Tvrtko Ursulin wrote:
> 
> On 20/08/2021 23:44, Matthew Brost wrote:
> > Sometimes it is desirable to queue work up for later if the GT PM isn't
> > held and run that work on next GT PM unpark.
> 
> Sounds maybe plausible, but it depends how much work can happen on unpark
> and whether it can have too much of a negative impact on latency for
> interactive loads? Or from a reverse angle, why the work wouldn't be done on

All it is does is add an interface to kick a work queue on unpark. i.e.
All the actually work is done async in the work queue so it shouldn't
add any latency.

> parking?
> 
> Also what kind of mechanism for dealing with too much stuff being put on
> this list you have? Can there be pressure which triggers (or would need to

No limits on pressure. See above, I don't think this is a concern.

> trigger) these deregistrations to happen at runtime (no park/unpark
> transitions)?
>
> > Implemented with a list in the GT of all pending work, workqueues in
> > the list, a callback to add a workqueue to the list, and finally a
> > wakeref post_get callback that iterates / drains the list + queues the
> > workqueues.
> > 
> > First user of this is deregistration of GuC contexts.
> 
> Does first imply there are more incoming?
>

Haven't found another user yet but this is generic mechanism so we can
add more in the future if other use cases arrise.
 
> > Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> > ---
> >   drivers/gpu/drm/i915/Makefile                 |  1 +
> >   drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
> >   drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
> >   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
> >   .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
> >   drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
> >   drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
> >   .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
> >   drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
> >   drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
> >   10 files changed, 119 insertions(+), 7 deletions(-)
> >   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> >   create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > 
> > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> > index 642a5b5a1b81..579bdc069f25 100644
> > --- a/drivers/gpu/drm/i915/Makefile
> > +++ b/drivers/gpu/drm/i915/Makefile
> > @@ -103,6 +103,7 @@ gt-y += \
> >   	gt/intel_gt_clock_utils.o \
> >   	gt/intel_gt_irq.o \
> >   	gt/intel_gt_pm.o \
> > +	gt/intel_gt_pm_unpark_work.o \
> >   	gt/intel_gt_pm_irq.o \
> >   	gt/intel_gt_requests.o \
> >   	gt/intel_gtt.o \
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
> > index 62d40c986642..7e690e74baa2 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gt.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt.c
> > @@ -29,6 +29,9 @@ void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
> >   	spin_lock_init(&gt->irq_lock);
> > +	spin_lock_init(&gt->pm_unpark_work_lock);
> > +	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
> > +
> >   	INIT_LIST_HEAD(&gt->closed_vma);
> >   	spin_lock_init(&gt->closed_lock);
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > index dea8e2479897..564c11a3748b 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > @@ -90,6 +90,13 @@ static int __gt_unpark(struct intel_wakeref *wf)
> >   	return 0;
> >   }
> > +static void __gt_unpark_work_queue(struct intel_wakeref *wf)
> > +{
> > +	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> > +
> > +	intel_gt_pm_unpark_work_queue(gt);
> > +}
> > +
> >   static int __gt_park(struct intel_wakeref *wf)
> >   {
> >   	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> > @@ -118,6 +125,7 @@ static int __gt_park(struct intel_wakeref *wf)
> >   static const struct intel_wakeref_ops wf_ops = {
> >   	.get = __gt_unpark,
> > +	.post_get = __gt_unpark_work_queue,
> >   	.put = __gt_park,
> >   };
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > new file mode 100644
> > index 000000000000..23162dbd0c35
> > --- /dev/null
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > @@ -0,0 +1,35 @@
> > +// SPDX-License-Identifier: MIT
> > +/*
> > + * Copyright © 2021 Intel Corporation
> > + */
> > +
> > +#include "i915_drv.h"
> > +#include "intel_runtime_pm.h"
> > +#include "intel_gt_pm.h"
> > +
> > +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
> > +{
> > +	struct intel_gt_pm_unpark_work *work, *next;
> > +	unsigned long flags;
> > +
> > +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> > +	list_for_each_entry_safe(work, next,
> > +				 &gt->pm_unpark_work_list, link) {
> > +		list_del_init(&work->link);
> > +		queue_work(system_unbound_wq, &work->worker);
> > +	}
> > +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> > +}
> > +
> > +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> > +				 struct intel_gt_pm_unpark_work *work)
> > +{
> > +	unsigned long flags;
> > +
> > +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> > +	if (intel_gt_pm_is_awake(gt))
> > +		queue_work(system_unbound_wq, &work->worker);
> > +	else if (list_empty(&work->link))
> 
> What's the list_empty check for, something can race by design?
> 

This function is allowed to be called twice, e.g. Two contexts can be
tried to be deregistered while the GT PM is ideal but only first context
results in the worker being added to the list of workers to be kicked on
unpark.

> > +		list_add_tail(&work->link, &gt->pm_unpark_work_list);
> > +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> > +}
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > new file mode 100644
> > index 000000000000..eaf1dc313aa2
> > --- /dev/null
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > @@ -0,0 +1,40 @@
> > +/* SPDX-License-Identifier: MIT */
> > +/*
> > + * Copyright © 2021 Intel Corporation
> > + */
> > +
> > +#ifndef INTEL_GT_PM_UNPARK_WORK_H
> > +#define INTEL_GT_PM_UNPARK_WORK_H
> > +
> > +#include <linux/list.h>
> > +#include <linux/workqueue.h>
> > +
> > +struct intel_gt;
> > +
> > +/**
> > + * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
> > + */
> > +struct intel_gt_pm_unpark_work {
> > +	/**
> > +	 * @link: link into gt->pm_unpark_work_list of workers that need to be
> > +	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
> > +	 */
> > +	struct list_head link;
> > +	/** @worker: will be scheduled when GT unparked */
> > +	struct work_struct worker;
> > +};
> > +
> > +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
> > +
> > +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> > +				 struct intel_gt_pm_unpark_work *work);
> > +
> > +static inline void
> > +intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
> > +			     work_func_t fn)
> > +{
> > +	INIT_LIST_HEAD(&work->link);
> > +	INIT_WORK(&work->worker, fn);
> > +}
> > +
> > +#endif /* INTEL_GT_PM_UNPARK_WORK_H */
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > index a81e21bf1bd1..4480312f0add 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > +++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > @@ -96,6 +96,16 @@ struct intel_gt {
> >   	struct intel_wakeref wakeref;
> >   	atomic_t user_wakeref;
> > +	/**
> > +	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
> > +	 * unparked, protected by pm_unpark_work_lock
> > +	 */
> > +	struct list_head pm_unpark_work_list;
> > +	/**
> > +	 * @pm_unpark_work_lock: protects pm_unpark_work_list
> > +	 */
> > +	spinlock_t pm_unpark_work_lock;
> > +
> >   	struct list_head closed_vma;
> >   	spinlock_t closed_lock; /* guards the list of closed_vma */
> > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > index 7358883f1540..023953e77553 100644
> > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > @@ -19,6 +19,7 @@
> >   #include "intel_uc_fw.h"
> >   #include "i915_utils.h"
> >   #include "i915_vma.h"
> > +#include "gt/intel_gt_pm_unpark_work.h"
> >   struct __guc_ads_blob;
> > @@ -78,11 +79,12 @@ struct intel_guc {
> >   		 */
> >   		struct list_head destroyed_contexts;
> >   		/**
> > -		 * @destroyed_worker: worker to deregister contexts, need as we
> > +		 * @destroyed_worker: Worker to deregister contexts, need as we
> >   		 * need to take a GT PM reference and can't from destroy
> > -		 * function as it might be in an atomic context (no sleeping)
> > +		 * function as it might be in an atomic context (no sleeping).
> > +		 * Worker only issues deregister when GT is unparked.
> >   		 */
> > -		struct work_struct destroyed_worker;
> > +		struct intel_gt_pm_unpark_work destroyed_worker;
> >   	} submission_state;
> >   	bool submission_supported;
> > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > index f835e06e5f9f..dbf919801de2 100644
> > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > @@ -1135,7 +1135,8 @@ int intel_guc_submission_init(struct intel_guc *guc)
> >   	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
> >   	ida_init(&guc->submission_state.guc_ids);
> >   	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
> > -	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
> > +	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
> > +				     destroyed_worker_func);
> >   	return 0;
> >   }
> > @@ -1942,13 +1943,18 @@ static void deregister_destroyed_contexts(struct intel_guc *guc)
> >   static void destroyed_worker_func(struct work_struct *w)
> >   {
> > -	struct intel_guc *guc = container_of(w, struct intel_guc,
> > +	struct intel_gt_pm_unpark_work *destroyed_worker =
> > +		container_of(w, struct intel_gt_pm_unpark_work, worker);
> > +	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
> >   					     submission_state.destroyed_worker);
> >   	struct intel_gt *gt = guc_to_gt(guc);
> >   	int tmp;
> > -	with_intel_gt_pm(gt, tmp)
> > +	with_intel_gt_pm_if_awake(gt, tmp)
> >   		deregister_destroyed_contexts(guc);
> > +
> > +	if (!list_empty(&guc->submission_state.destroyed_contexts))
> > +		intel_gt_pm_unpark_work_add(gt, destroyed_worker);
> 
> This is the worker itself, right?
> 

Yes.

> There's a "if awake" here followed by another "if awake" inside
> intel_gt_pm_unpark_work_add which raises questions.
>

Even in the worker we only deregister the context if the GT is awake.
 
> Second question is what's the list_empty for - why is the state of the list
> itself relevant to a single worker deciding whether to re-add itself to it
> or not? And is there a lock protecting this list?
>

Yes we have locking - pm_unpark_work_lock. It is basically all 2
threads to call intel_gt_pm_unpark_work_add while the GT is idle - in
that case only the first call adds the worker to the list of workers to
kick when unparked (same explaination as above).
 
> On the overall it feels questionable to have unpark work which apparently
> can race with subsequent parking. Presumably you cannot have it run sync on
> unpark due execution context issues?
> 

No race, just allowed to be called twice.

> >   }
> >   static void guc_context_destroy(struct kref *kref)
> > @@ -1985,7 +1991,8 @@ static void guc_context_destroy(struct kref *kref)
> >   	 * take the GT PM for the first time which isn't allowed from an atomic
> >   	 * context.
> >   	 */
> > -	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
> > +	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
> > +				    &guc->submission_state.destroyed_worker);
> >   }
> >   static int guc_context_alloc(struct intel_context *ce)
> > diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
> > index dfd87d082218..282fc4f312e3 100644
> > --- a/drivers/gpu/drm/i915/intel_wakeref.c
> > +++ b/drivers/gpu/drm/i915/intel_wakeref.c
> > @@ -24,6 +24,8 @@ static void rpm_put(struct intel_wakeref *wf)
> >   int __intel_wakeref_get_first(struct intel_wakeref *wf)
> >   {
> > +	bool do_post = false;
> > +
> >   	/*
> >   	 * Treat get/put as different subclasses, as we may need to run
> >   	 * the put callback from under the shrinker and do not want to
> > @@ -44,8 +46,11 @@ int __intel_wakeref_get_first(struct intel_wakeref *wf)
> >   		}
> >   		smp_mb__before_atomic(); /* release wf->count */
> > +		do_post = true;
> >   	}
> >   	atomic_inc(&wf->count);
> > +	if (do_post && wf->ops->post_get)
> > +		wf->ops->post_get(wf);
> 
> You want this hook under the wf->mutex and why?
> 

I didn't really think about this but everything else in under the mutex
so I included this under the mutex too. In this case this post_get op
could likely be outside the mutex but I think it harmless to keep it
under the mutex. For future safety, I think it should stay under the
mutex.

Matt

> Regards,
> 
> Tvrtko
> 
> >   	mutex_unlock(&wf->mutex);
> >   	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
> > diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
> > index 545c8f277c46..ef7e6a698e8a 100644
> > --- a/drivers/gpu/drm/i915/intel_wakeref.h
> > +++ b/drivers/gpu/drm/i915/intel_wakeref.h
> > @@ -30,6 +30,7 @@ typedef depot_stack_handle_t intel_wakeref_t;
> >   struct intel_wakeref_ops {
> >   	int (*get)(struct intel_wakeref *wf);
> > +	void (*post_get)(struct intel_wakeref *wf);
> >   	int (*put)(struct intel_wakeref *wf);
> >   };
> >
Tvrtko Ursulin Sept. 13, 2021, 10:33 a.m. UTC | #5
On 10/09/2021 21:09, Matthew Brost wrote:
> On Fri, Sep 10, 2021 at 09:36:17AM +0100, Tvrtko Ursulin wrote:
>>
>> On 20/08/2021 23:44, Matthew Brost wrote:
>>> Sometimes it is desirable to queue work up for later if the GT PM isn't
>>> held and run that work on next GT PM unpark.
>>
>> Sounds maybe plausible, but it depends how much work can happen on unpark
>> and whether it can have too much of a negative impact on latency for
>> interactive loads? Or from a reverse angle, why the work wouldn't be done on
> 
> All it is does is add an interface to kick a work queue on unpark. i.e.
> All the actually work is done async in the work queue so it shouldn't
> add any latency.
> 
>> parking?
>>
>> Also what kind of mechanism for dealing with too much stuff being put on
>> this list you have? Can there be pressure which triggers (or would need to
> 
> No limits on pressure. See above, I don't think this is a concern.

On unpark it has the potential to send an unbound amount of actions for 
the GuC to process. Which will be competing, in GuC internal processing 
power, with the user action which caused the unpark. That logically does 
feel like can have effect on initial latency. Why you think it cannot?

Why the work wouldn't be done on parking?

With this scheme couldn't we end up with a situation that the worker 
keeps missing the GT unparked state and so keeps piling items on the 
deregistration list? Can you run of some ids like that (which is related 
to my question of how is pressure handled here).

Unpark
Register context
Submit work
Retire
Schedule context deregister
Park

Worker runs
GT parked
Work put on a list

Unpark
Schedule deregistration worker
Register new context
Submit work
Retire
Schedule contect deregister
Park

Worker runs (lets say there was CPU pressure)
GT already parked
  -> deregistration queue now has two contexts on it

... repeat until disaster ...

Unless I have misunderstood the logic.

>> trigger) these deregistrations to happen at runtime (no park/unpark
>> transitions)?
>>
>>> Implemented with a list in the GT of all pending work, workqueues in
>>> the list, a callback to add a workqueue to the list, and finally a
>>> wakeref post_get callback that iterates / drains the list + queues the
>>> workqueues.
>>>
>>> First user of this is deregistration of GuC contexts.
>>
>> Does first imply there are more incoming?
>>
> 
> Haven't found another user yet but this is generic mechanism so we can
> add more in the future if other use cases arrise.

My feeling is it would be best to leave it for later.

>   
>>> Signed-off-by: Matthew Brost <matthew.brost@intel.com>
>>> ---
>>>    drivers/gpu/drm/i915/Makefile                 |  1 +
>>>    drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
>>>    drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
>>>    .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
>>>    .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
>>>    drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
>>>    drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
>>>    .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
>>>    drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
>>>    drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
>>>    10 files changed, 119 insertions(+), 7 deletions(-)
>>>    create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
>>>    create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
>>>
>>> diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
>>> index 642a5b5a1b81..579bdc069f25 100644
>>> --- a/drivers/gpu/drm/i915/Makefile
>>> +++ b/drivers/gpu/drm/i915/Makefile
>>> @@ -103,6 +103,7 @@ gt-y += \
>>>    	gt/intel_gt_clock_utils.o \
>>>    	gt/intel_gt_irq.o \
>>>    	gt/intel_gt_pm.o \
>>> +	gt/intel_gt_pm_unpark_work.o \
>>>    	gt/intel_gt_pm_irq.o \
>>>    	gt/intel_gt_requests.o \
>>>    	gt/intel_gtt.o \
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
>>> index 62d40c986642..7e690e74baa2 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_gt.c
>>> +++ b/drivers/gpu/drm/i915/gt/intel_gt.c
>>> @@ -29,6 +29,9 @@ void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
>>>    	spin_lock_init(&gt->irq_lock);
>>> +	spin_lock_init(&gt->pm_unpark_work_lock);
>>> +	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
>>> +
>>>    	INIT_LIST_HEAD(&gt->closed_vma);
>>>    	spin_lock_init(&gt->closed_lock);
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
>>> index dea8e2479897..564c11a3748b 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
>>> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
>>> @@ -90,6 +90,13 @@ static int __gt_unpark(struct intel_wakeref *wf)
>>>    	return 0;
>>>    }
>>> +static void __gt_unpark_work_queue(struct intel_wakeref *wf)
>>> +{
>>> +	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
>>> +
>>> +	intel_gt_pm_unpark_work_queue(gt);
>>> +}
>>> +
>>>    static int __gt_park(struct intel_wakeref *wf)
>>>    {
>>>    	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
>>> @@ -118,6 +125,7 @@ static int __gt_park(struct intel_wakeref *wf)
>>>    static const struct intel_wakeref_ops wf_ops = {
>>>    	.get = __gt_unpark,
>>> +	.post_get = __gt_unpark_work_queue,
>>>    	.put = __gt_park,
>>>    };
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
>>> new file mode 100644
>>> index 000000000000..23162dbd0c35
>>> --- /dev/null
>>> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
>>> @@ -0,0 +1,35 @@
>>> +// SPDX-License-Identifier: MIT
>>> +/*
>>> + * Copyright © 2021 Intel Corporation
>>> + */
>>> +
>>> +#include "i915_drv.h"
>>> +#include "intel_runtime_pm.h"
>>> +#include "intel_gt_pm.h"
>>> +
>>> +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
>>> +{
>>> +	struct intel_gt_pm_unpark_work *work, *next;
>>> +	unsigned long flags;
>>> +
>>> +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
>>> +	list_for_each_entry_safe(work, next,
>>> +				 &gt->pm_unpark_work_list, link) {
>>> +		list_del_init(&work->link);
>>> +		queue_work(system_unbound_wq, &work->worker);
>>> +	}
>>> +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
>>> +}
>>> +
>>> +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
>>> +				 struct intel_gt_pm_unpark_work *work)
>>> +{
>>> +	unsigned long flags;
>>> +
>>> +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
>>> +	if (intel_gt_pm_is_awake(gt))
>>> +		queue_work(system_unbound_wq, &work->worker);
>>> +	else if (list_empty(&work->link))
>>
>> What's the list_empty check for, something can race by design?
>>
> 
> This function is allowed to be called twice, e.g. Two contexts can be
> tried to be deregistered while the GT PM is ideal but only first context
> results in the worker being added to the list of workers to be kicked on
> unpark.

Hmm okay.

> 
>>> +		list_add_tail(&work->link, &gt->pm_unpark_work_list);
>>> +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
>>> +}
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
>>> new file mode 100644
>>> index 000000000000..eaf1dc313aa2
>>> --- /dev/null
>>> +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
>>> @@ -0,0 +1,40 @@
>>> +/* SPDX-License-Identifier: MIT */
>>> +/*
>>> + * Copyright © 2021 Intel Corporation
>>> + */
>>> +
>>> +#ifndef INTEL_GT_PM_UNPARK_WORK_H
>>> +#define INTEL_GT_PM_UNPARK_WORK_H
>>> +
>>> +#include <linux/list.h>
>>> +#include <linux/workqueue.h>
>>> +
>>> +struct intel_gt;
>>> +
>>> +/**
>>> + * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
>>> + */
>>> +struct intel_gt_pm_unpark_work {
>>> +	/**
>>> +	 * @link: link into gt->pm_unpark_work_list of workers that need to be
>>> +	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
>>> +	 */
>>> +	struct list_head link;
>>> +	/** @worker: will be scheduled when GT unparked */
>>> +	struct work_struct worker;
>>> +};
>>> +
>>> +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
>>> +
>>> +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
>>> +				 struct intel_gt_pm_unpark_work *work);
>>> +
>>> +static inline void
>>> +intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
>>> +			     work_func_t fn)
>>> +{
>>> +	INIT_LIST_HEAD(&work->link);
>>> +	INIT_WORK(&work->worker, fn);
>>> +}
>>> +
>>> +#endif /* INTEL_GT_PM_UNPARK_WORK_H */
>>> diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
>>> index a81e21bf1bd1..4480312f0add 100644
>>> --- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
>>> +++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
>>> @@ -96,6 +96,16 @@ struct intel_gt {
>>>    	struct intel_wakeref wakeref;
>>>    	atomic_t user_wakeref;
>>> +	/**
>>> +	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
>>> +	 * unparked, protected by pm_unpark_work_lock
>>> +	 */
>>> +	struct list_head pm_unpark_work_list;
>>> +	/**
>>> +	 * @pm_unpark_work_lock: protects pm_unpark_work_list
>>> +	 */
>>> +	spinlock_t pm_unpark_work_lock;
>>> +
>>>    	struct list_head closed_vma;
>>>    	spinlock_t closed_lock; /* guards the list of closed_vma */
>>> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
>>> index 7358883f1540..023953e77553 100644
>>> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
>>> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
>>> @@ -19,6 +19,7 @@
>>>    #include "intel_uc_fw.h"
>>>    #include "i915_utils.h"
>>>    #include "i915_vma.h"
>>> +#include "gt/intel_gt_pm_unpark_work.h"
>>>    struct __guc_ads_blob;
>>> @@ -78,11 +79,12 @@ struct intel_guc {
>>>    		 */
>>>    		struct list_head destroyed_contexts;
>>>    		/**
>>> -		 * @destroyed_worker: worker to deregister contexts, need as we
>>> +		 * @destroyed_worker: Worker to deregister contexts, need as we
>>>    		 * need to take a GT PM reference and can't from destroy
>>> -		 * function as it might be in an atomic context (no sleeping)
>>> +		 * function as it might be in an atomic context (no sleeping).
>>> +		 * Worker only issues deregister when GT is unparked.
>>>    		 */
>>> -		struct work_struct destroyed_worker;
>>> +		struct intel_gt_pm_unpark_work destroyed_worker;
>>>    	} submission_state;
>>>    	bool submission_supported;
>>> diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
>>> index f835e06e5f9f..dbf919801de2 100644
>>> --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
>>> +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
>>> @@ -1135,7 +1135,8 @@ int intel_guc_submission_init(struct intel_guc *guc)
>>>    	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
>>>    	ida_init(&guc->submission_state.guc_ids);
>>>    	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
>>> -	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
>>> +	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
>>> +				     destroyed_worker_func);
>>>    	return 0;
>>>    }
>>> @@ -1942,13 +1943,18 @@ static void deregister_destroyed_contexts(struct intel_guc *guc)
>>>    static void destroyed_worker_func(struct work_struct *w)
>>>    {
>>> -	struct intel_guc *guc = container_of(w, struct intel_guc,
>>> +	struct intel_gt_pm_unpark_work *destroyed_worker =
>>> +		container_of(w, struct intel_gt_pm_unpark_work, worker);
>>> +	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
>>>    					     submission_state.destroyed_worker);
>>>    	struct intel_gt *gt = guc_to_gt(guc);
>>>    	int tmp;
>>> -	with_intel_gt_pm(gt, tmp)
>>> +	with_intel_gt_pm_if_awake(gt, tmp)
>>>    		deregister_destroyed_contexts(guc);
>>> +
>>> +	if (!list_empty(&guc->submission_state.destroyed_contexts))
>>> +		intel_gt_pm_unpark_work_add(gt, destroyed_worker);
>>
>> This is the worker itself, right?
>>
> 
> Yes.
> 
>> There's a "if awake" here followed by another "if awake" inside
>> intel_gt_pm_unpark_work_add which raises questions.
>>
> 
> Even in the worker we only deregister the context if the GT is awake.

Yeah but the two "if awake" checks followed by one another is the 
confusing part since it is not atomic. Coupled with the list empty check 
I mean. So two external async condition can influence the flow, one of 
which is even checked twice.

The worker can decide not to deregister, but by the time 
intel_gt_pm_unpark_work_add runs GT may have became unparked so worker 
gets queued. When it runs GT is parked again and so it can bounce in a 
loop forever.

>   
>> Second question is what's the list_empty for - why is the state of the list
>> itself relevant to a single worker deciding whether to re-add itself to it
>> or not? And is there a lock protecting this list?
>>
> 
> Yes we have locking - pm_unpark_work_lock. It is basically all 2
> threads to call intel_gt_pm_unpark_work_add while the GT is idle - in
> that case only the first call adds the worker to the list of workers to
> kick when unparked (same explaination as above).
>   
>> On the overall it feels questionable to have unpark work which apparently
>> can race with subsequent parking. Presumably you cannot have it run sync on
>> unpark due execution context issues?
>>
> 
> No race, just allowed to be called twice.

But list_empty check is not under any locks so I don't follow.

> 
>>>    }
>>>    static void guc_context_destroy(struct kref *kref)
>>> @@ -1985,7 +1991,8 @@ static void guc_context_destroy(struct kref *kref)
>>>    	 * take the GT PM for the first time which isn't allowed from an atomic
>>>    	 * context.
>>>    	 */
>>> -	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
>>> +	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
>>> +				    &guc->submission_state.destroyed_worker);
>>>    }
>>>    static int guc_context_alloc(struct intel_context *ce)
>>> diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
>>> index dfd87d082218..282fc4f312e3 100644
>>> --- a/drivers/gpu/drm/i915/intel_wakeref.c
>>> +++ b/drivers/gpu/drm/i915/intel_wakeref.c
>>> @@ -24,6 +24,8 @@ static void rpm_put(struct intel_wakeref *wf)
>>>    int __intel_wakeref_get_first(struct intel_wakeref *wf)
>>>    {
>>> +	bool do_post = false;
>>> +
>>>    	/*
>>>    	 * Treat get/put as different subclasses, as we may need to run
>>>    	 * the put callback from under the shrinker and do not want to
>>> @@ -44,8 +46,11 @@ int __intel_wakeref_get_first(struct intel_wakeref *wf)
>>>    		}
>>>    		smp_mb__before_atomic(); /* release wf->count */
>>> +		do_post = true;
>>>    	}
>>>    	atomic_inc(&wf->count);
>>> +	if (do_post && wf->ops->post_get)
>>> +		wf->ops->post_get(wf);
>>
>> You want this hook under the wf->mutex and why?
>>
> 
> I didn't really think about this but everything else in under the mutex
> so I included this under the mutex too. In this case this post_get op
> could likely be outside the mutex but I think it harmless to keep it
> under the mutex. For future safety, I think it should stay under the
> mutex.

If it is under the mutex then what is the point of two hooks? The only 
difference is get sees wf->count == 0, while post_get sees it as one.

No wait.. you have put post_get outside the 0->1 transition check so 
potentially called more than once. You probably do not want this.. But 
could you just do this from the existing hook is the main question.

Regards,

Tvrtko

> 
> Matt
> 
>> Regards,
>>
>> Tvrtko
>>
>>>    	mutex_unlock(&wf->mutex);
>>>    	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
>>> diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
>>> index 545c8f277c46..ef7e6a698e8a 100644
>>> --- a/drivers/gpu/drm/i915/intel_wakeref.h
>>> +++ b/drivers/gpu/drm/i915/intel_wakeref.h
>>> @@ -30,6 +30,7 @@ typedef depot_stack_handle_t intel_wakeref_t;
>>>    struct intel_wakeref_ops {
>>>    	int (*get)(struct intel_wakeref *wf);
>>> +	void (*post_get)(struct intel_wakeref *wf);
>>>    	int (*put)(struct intel_wakeref *wf);
>>>    };
>>>
Matthew Brost Sept. 13, 2021, 5:20 p.m. UTC | #6
On Mon, Sep 13, 2021 at 11:33:46AM +0100, Tvrtko Ursulin wrote:
> 
> On 10/09/2021 21:09, Matthew Brost wrote:
> > On Fri, Sep 10, 2021 at 09:36:17AM +0100, Tvrtko Ursulin wrote:
> > > 
> > > On 20/08/2021 23:44, Matthew Brost wrote:
> > > > Sometimes it is desirable to queue work up for later if the GT PM isn't
> > > > held and run that work on next GT PM unpark.
> > > 
> > > Sounds maybe plausible, but it depends how much work can happen on unpark
> > > and whether it can have too much of a negative impact on latency for
> > > interactive loads? Or from a reverse angle, why the work wouldn't be done on
> > 
> > All it is does is add an interface to kick a work queue on unpark. i.e.
> > All the actually work is done async in the work queue so it shouldn't
> > add any latency.
> > 
> > > parking?
> > > 
> > > Also what kind of mechanism for dealing with too much stuff being put on
> > > this list you have? Can there be pressure which triggers (or would need to
> > 
> > No limits on pressure. See above, I don't think this is a concern.
> 
> On unpark it has the potential to send an unbound amount of actions for the
> GuC to process. Which will be competing, in GuC internal processing power,
> with the user action which caused the unpark. That logically does feel like
> can have effect on initial latency. Why you think it cannot?
> 

Ah, I see what you mean. Yes, bunch of deregisters could be sent before
a submission adding latency. Maybe I just drop this whole idea / patch
for now. Not going to respond to individual comments because this will
be dropped.

Matt

> Why the work wouldn't be done on parking?
> 
> With this scheme couldn't we end up with a situation that the worker keeps
> missing the GT unparked state and so keeps piling items on the
> deregistration list? Can you run of some ids like that (which is related to
> my question of how is pressure handled here).
> 
> Unpark
> Register context
> Submit work
> Retire
> Schedule context deregister
> Park
> 
> Worker runs
> GT parked
> Work put on a list
> 
> Unpark
> Schedule deregistration worker
> Register new context
> Submit work
> Retire
> Schedule contect deregister
> Park
> 
> Worker runs (lets say there was CPU pressure)
> GT already parked
>  -> deregistration queue now has two contexts on it
> 
> ... repeat until disaster ...
> 
> Unless I have misunderstood the logic.
> 
> > > trigger) these deregistrations to happen at runtime (no park/unpark
> > > transitions)?
> > > 
> > > > Implemented with a list in the GT of all pending work, workqueues in
> > > > the list, a callback to add a workqueue to the list, and finally a
> > > > wakeref post_get callback that iterates / drains the list + queues the
> > > > workqueues.
> > > > 
> > > > First user of this is deregistration of GuC contexts.
> > > 
> > > Does first imply there are more incoming?
> > > 
> > 
> > Haven't found another user yet but this is generic mechanism so we can
> > add more in the future if other use cases arrise.
> 
> My feeling is it would be best to leave it for later.
> 
> > > > Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> > > > ---
> > > >    drivers/gpu/drm/i915/Makefile                 |  1 +
> > > >    drivers/gpu/drm/i915/gt/intel_gt.c            |  3 ++
> > > >    drivers/gpu/drm/i915/gt/intel_gt_pm.c         |  8 ++++
> > > >    .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.c | 35 ++++++++++++++++
> > > >    .../gpu/drm/i915/gt/intel_gt_pm_unpark_work.h | 40 +++++++++++++++++++
> > > >    drivers/gpu/drm/i915/gt/intel_gt_types.h      | 10 +++++
> > > >    drivers/gpu/drm/i915/gt/uc/intel_guc.h        |  8 ++--
> > > >    .../gpu/drm/i915/gt/uc/intel_guc_submission.c | 15 +++++--
> > > >    drivers/gpu/drm/i915/intel_wakeref.c          |  5 +++
> > > >    drivers/gpu/drm/i915/intel_wakeref.h          |  1 +
> > > >    10 files changed, 119 insertions(+), 7 deletions(-)
> > > >    create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > > >    create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > > > 
> > > > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> > > > index 642a5b5a1b81..579bdc069f25 100644
> > > > --- a/drivers/gpu/drm/i915/Makefile
> > > > +++ b/drivers/gpu/drm/i915/Makefile
> > > > @@ -103,6 +103,7 @@ gt-y += \
> > > >    	gt/intel_gt_clock_utils.o \
> > > >    	gt/intel_gt_irq.o \
> > > >    	gt/intel_gt_pm.o \
> > > > +	gt/intel_gt_pm_unpark_work.o \
> > > >    	gt/intel_gt_pm_irq.o \
> > > >    	gt/intel_gt_requests.o \
> > > >    	gt/intel_gtt.o \
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
> > > > index 62d40c986642..7e690e74baa2 100644
> > > > --- a/drivers/gpu/drm/i915/gt/intel_gt.c
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_gt.c
> > > > @@ -29,6 +29,9 @@ void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
> > > >    	spin_lock_init(&gt->irq_lock);
> > > > +	spin_lock_init(&gt->pm_unpark_work_lock);
> > > > +	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
> > > > +
> > > >    	INIT_LIST_HEAD(&gt->closed_vma);
> > > >    	spin_lock_init(&gt->closed_lock);
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > > > index dea8e2479897..564c11a3748b 100644
> > > > --- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
> > > > @@ -90,6 +90,13 @@ static int __gt_unpark(struct intel_wakeref *wf)
> > > >    	return 0;
> > > >    }
> > > > +static void __gt_unpark_work_queue(struct intel_wakeref *wf)
> > > > +{
> > > > +	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> > > > +
> > > > +	intel_gt_pm_unpark_work_queue(gt);
> > > > +}
> > > > +
> > > >    static int __gt_park(struct intel_wakeref *wf)
> > > >    {
> > > >    	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
> > > > @@ -118,6 +125,7 @@ static int __gt_park(struct intel_wakeref *wf)
> > > >    static const struct intel_wakeref_ops wf_ops = {
> > > >    	.get = __gt_unpark,
> > > > +	.post_get = __gt_unpark_work_queue,
> > > >    	.put = __gt_park,
> > > >    };
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > > > new file mode 100644
> > > > index 000000000000..23162dbd0c35
> > > > --- /dev/null
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
> > > > @@ -0,0 +1,35 @@
> > > > +// SPDX-License-Identifier: MIT
> > > > +/*
> > > > + * Copyright © 2021 Intel Corporation
> > > > + */
> > > > +
> > > > +#include "i915_drv.h"
> > > > +#include "intel_runtime_pm.h"
> > > > +#include "intel_gt_pm.h"
> > > > +
> > > > +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
> > > > +{
> > > > +	struct intel_gt_pm_unpark_work *work, *next;
> > > > +	unsigned long flags;
> > > > +
> > > > +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> > > > +	list_for_each_entry_safe(work, next,
> > > > +				 &gt->pm_unpark_work_list, link) {
> > > > +		list_del_init(&work->link);
> > > > +		queue_work(system_unbound_wq, &work->worker);
> > > > +	}
> > > > +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> > > > +}
> > > > +
> > > > +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> > > > +				 struct intel_gt_pm_unpark_work *work)
> > > > +{
> > > > +	unsigned long flags;
> > > > +
> > > > +	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
> > > > +	if (intel_gt_pm_is_awake(gt))
> > > > +		queue_work(system_unbound_wq, &work->worker);
> > > > +	else if (list_empty(&work->link))
> > > 
> > > What's the list_empty check for, something can race by design?
> > > 
> > 
> > This function is allowed to be called twice, e.g. Two contexts can be
> > tried to be deregistered while the GT PM is ideal but only first context
> > results in the worker being added to the list of workers to be kicked on
> > unpark.
> 
> Hmm okay.
> 
> > 
> > > > +		list_add_tail(&work->link, &gt->pm_unpark_work_list);
> > > > +	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
> > > > +}
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > > > new file mode 100644
> > > > index 000000000000..eaf1dc313aa2
> > > > --- /dev/null
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
> > > > @@ -0,0 +1,40 @@
> > > > +/* SPDX-License-Identifier: MIT */
> > > > +/*
> > > > + * Copyright © 2021 Intel Corporation
> > > > + */
> > > > +
> > > > +#ifndef INTEL_GT_PM_UNPARK_WORK_H
> > > > +#define INTEL_GT_PM_UNPARK_WORK_H
> > > > +
> > > > +#include <linux/list.h>
> > > > +#include <linux/workqueue.h>
> > > > +
> > > > +struct intel_gt;
> > > > +
> > > > +/**
> > > > + * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
> > > > + */
> > > > +struct intel_gt_pm_unpark_work {
> > > > +	/**
> > > > +	 * @link: link into gt->pm_unpark_work_list of workers that need to be
> > > > +	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
> > > > +	 */
> > > > +	struct list_head link;
> > > > +	/** @worker: will be scheduled when GT unparked */
> > > > +	struct work_struct worker;
> > > > +};
> > > > +
> > > > +void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
> > > > +
> > > > +void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
> > > > +				 struct intel_gt_pm_unpark_work *work);
> > > > +
> > > > +static inline void
> > > > +intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
> > > > +			     work_func_t fn)
> > > > +{
> > > > +	INIT_LIST_HEAD(&work->link);
> > > > +	INIT_WORK(&work->worker, fn);
> > > > +}
> > > > +
> > > > +#endif /* INTEL_GT_PM_UNPARK_WORK_H */
> > > > diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > > > index a81e21bf1bd1..4480312f0add 100644
> > > > --- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > > > +++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
> > > > @@ -96,6 +96,16 @@ struct intel_gt {
> > > >    	struct intel_wakeref wakeref;
> > > >    	atomic_t user_wakeref;
> > > > +	/**
> > > > +	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
> > > > +	 * unparked, protected by pm_unpark_work_lock
> > > > +	 */
> > > > +	struct list_head pm_unpark_work_list;
> > > > +	/**
> > > > +	 * @pm_unpark_work_lock: protects pm_unpark_work_list
> > > > +	 */
> > > > +	spinlock_t pm_unpark_work_lock;
> > > > +
> > > >    	struct list_head closed_vma;
> > > >    	spinlock_t closed_lock; /* guards the list of closed_vma */
> > > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > > > index 7358883f1540..023953e77553 100644
> > > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
> > > > @@ -19,6 +19,7 @@
> > > >    #include "intel_uc_fw.h"
> > > >    #include "i915_utils.h"
> > > >    #include "i915_vma.h"
> > > > +#include "gt/intel_gt_pm_unpark_work.h"
> > > >    struct __guc_ads_blob;
> > > > @@ -78,11 +79,12 @@ struct intel_guc {
> > > >    		 */
> > > >    		struct list_head destroyed_contexts;
> > > >    		/**
> > > > -		 * @destroyed_worker: worker to deregister contexts, need as we
> > > > +		 * @destroyed_worker: Worker to deregister contexts, need as we
> > > >    		 * need to take a GT PM reference and can't from destroy
> > > > -		 * function as it might be in an atomic context (no sleeping)
> > > > +		 * function as it might be in an atomic context (no sleeping).
> > > > +		 * Worker only issues deregister when GT is unparked.
> > > >    		 */
> > > > -		struct work_struct destroyed_worker;
> > > > +		struct intel_gt_pm_unpark_work destroyed_worker;
> > > >    	} submission_state;
> > > >    	bool submission_supported;
> > > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > index f835e06e5f9f..dbf919801de2 100644
> > > > --- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
> > > > @@ -1135,7 +1135,8 @@ int intel_guc_submission_init(struct intel_guc *guc)
> > > >    	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
> > > >    	ida_init(&guc->submission_state.guc_ids);
> > > >    	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
> > > > -	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
> > > > +	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
> > > > +				     destroyed_worker_func);
> > > >    	return 0;
> > > >    }
> > > > @@ -1942,13 +1943,18 @@ static void deregister_destroyed_contexts(struct intel_guc *guc)
> > > >    static void destroyed_worker_func(struct work_struct *w)
> > > >    {
> > > > -	struct intel_guc *guc = container_of(w, struct intel_guc,
> > > > +	struct intel_gt_pm_unpark_work *destroyed_worker =
> > > > +		container_of(w, struct intel_gt_pm_unpark_work, worker);
> > > > +	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
> > > >    					     submission_state.destroyed_worker);
> > > >    	struct intel_gt *gt = guc_to_gt(guc);
> > > >    	int tmp;
> > > > -	with_intel_gt_pm(gt, tmp)
> > > > +	with_intel_gt_pm_if_awake(gt, tmp)
> > > >    		deregister_destroyed_contexts(guc);
> > > > +
> > > > +	if (!list_empty(&guc->submission_state.destroyed_contexts))
> > > > +		intel_gt_pm_unpark_work_add(gt, destroyed_worker);
> > > 
> > > This is the worker itself, right?
> > > 
> > 
> > Yes.
> > 
> > > There's a "if awake" here followed by another "if awake" inside
> > > intel_gt_pm_unpark_work_add which raises questions.
> > > 
> > 
> > Even in the worker we only deregister the context if the GT is awake.
> 
> Yeah but the two "if awake" checks followed by one another is the confusing
> part since it is not atomic. Coupled with the list empty check I mean. So
> two external async condition can influence the flow, one of which is even
> checked twice.
> 
> The worker can decide not to deregister, but by the time
> intel_gt_pm_unpark_work_add runs GT may have became unparked so worker gets
> queued. When it runs GT is parked again and so it can bounce in a loop
> forever.
> 
> > > Second question is what's the list_empty for - why is the state of the list
> > > itself relevant to a single worker deciding whether to re-add itself to it
> > > or not? And is there a lock protecting this list?
> > > 
> > 
> > Yes we have locking - pm_unpark_work_lock. It is basically all 2
> > threads to call intel_gt_pm_unpark_work_add while the GT is idle - in
> > that case only the first call adds the worker to the list of workers to
> > kick when unparked (same explaination as above).
> > > On the overall it feels questionable to have unpark work which apparently
> > > can race with subsequent parking. Presumably you cannot have it run sync on
> > > unpark due execution context issues?
> > > 
> > 
> > No race, just allowed to be called twice.
> 
> But list_empty check is not under any locks so I don't follow.
> 
> > 
> > > >    }
> > > >    static void guc_context_destroy(struct kref *kref)
> > > > @@ -1985,7 +1991,8 @@ static void guc_context_destroy(struct kref *kref)
> > > >    	 * take the GT PM for the first time which isn't allowed from an atomic
> > > >    	 * context.
> > > >    	 */
> > > > -	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
> > > > +	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
> > > > +				    &guc->submission_state.destroyed_worker);
> > > >    }
> > > >    static int guc_context_alloc(struct intel_context *ce)
> > > > diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
> > > > index dfd87d082218..282fc4f312e3 100644
> > > > --- a/drivers/gpu/drm/i915/intel_wakeref.c
> > > > +++ b/drivers/gpu/drm/i915/intel_wakeref.c
> > > > @@ -24,6 +24,8 @@ static void rpm_put(struct intel_wakeref *wf)
> > > >    int __intel_wakeref_get_first(struct intel_wakeref *wf)
> > > >    {
> > > > +	bool do_post = false;
> > > > +
> > > >    	/*
> > > >    	 * Treat get/put as different subclasses, as we may need to run
> > > >    	 * the put callback from under the shrinker and do not want to
> > > > @@ -44,8 +46,11 @@ int __intel_wakeref_get_first(struct intel_wakeref *wf)
> > > >    		}
> > > >    		smp_mb__before_atomic(); /* release wf->count */
> > > > +		do_post = true;
> > > >    	}
> > > >    	atomic_inc(&wf->count);
> > > > +	if (do_post && wf->ops->post_get)
> > > > +		wf->ops->post_get(wf);
> > > 
> > > You want this hook under the wf->mutex and why?
> > > 
> > 
> > I didn't really think about this but everything else in under the mutex
> > so I included this under the mutex too. In this case this post_get op
> > could likely be outside the mutex but I think it harmless to keep it
> > under the mutex. For future safety, I think it should stay under the
> > mutex.
> 
> If it is under the mutex then what is the point of two hooks? The only
> difference is get sees wf->count == 0, while post_get sees it as one.
> 
> No wait.. you have put post_get outside the 0->1 transition check so
> potentially called more than once. You probably do not want this.. But could
> you just do this from the existing hook is the main question.
> 
> Regards,
> 
> Tvrtko
> 
> > 
> > Matt
> > 
> > > Regards,
> > > 
> > > Tvrtko
> > > 
> > > >    	mutex_unlock(&wf->mutex);
> > > >    	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
> > > > diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
> > > > index 545c8f277c46..ef7e6a698e8a 100644
> > > > --- a/drivers/gpu/drm/i915/intel_wakeref.h
> > > > +++ b/drivers/gpu/drm/i915/intel_wakeref.h
> > > > @@ -30,6 +30,7 @@ typedef depot_stack_handle_t intel_wakeref_t;
> > > >    struct intel_wakeref_ops {
> > > >    	int (*get)(struct intel_wakeref *wf);
> > > > +	void (*post_get)(struct intel_wakeref *wf);
> > > >    	int (*put)(struct intel_wakeref *wf);
> > > >    };
> > > >
diff mbox series

Patch

diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
index 642a5b5a1b81..579bdc069f25 100644
--- a/drivers/gpu/drm/i915/Makefile
+++ b/drivers/gpu/drm/i915/Makefile
@@ -103,6 +103,7 @@  gt-y += \
 	gt/intel_gt_clock_utils.o \
 	gt/intel_gt_irq.o \
 	gt/intel_gt_pm.o \
+	gt/intel_gt_pm_unpark_work.o \
 	gt/intel_gt_pm_irq.o \
 	gt/intel_gt_requests.o \
 	gt/intel_gtt.o \
diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c
index 62d40c986642..7e690e74baa2 100644
--- a/drivers/gpu/drm/i915/gt/intel_gt.c
+++ b/drivers/gpu/drm/i915/gt/intel_gt.c
@@ -29,6 +29,9 @@  void intel_gt_init_early(struct intel_gt *gt, struct drm_i915_private *i915)
 
 	spin_lock_init(&gt->irq_lock);
 
+	spin_lock_init(&gt->pm_unpark_work_lock);
+	INIT_LIST_HEAD(&gt->pm_unpark_work_list);
+
 	INIT_LIST_HEAD(&gt->closed_vma);
 	spin_lock_init(&gt->closed_lock);
 
diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm.c b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
index dea8e2479897..564c11a3748b 100644
--- a/drivers/gpu/drm/i915/gt/intel_gt_pm.c
+++ b/drivers/gpu/drm/i915/gt/intel_gt_pm.c
@@ -90,6 +90,13 @@  static int __gt_unpark(struct intel_wakeref *wf)
 	return 0;
 }
 
+static void __gt_unpark_work_queue(struct intel_wakeref *wf)
+{
+	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
+
+	intel_gt_pm_unpark_work_queue(gt);
+}
+
 static int __gt_park(struct intel_wakeref *wf)
 {
 	struct intel_gt *gt = container_of(wf, typeof(*gt), wakeref);
@@ -118,6 +125,7 @@  static int __gt_park(struct intel_wakeref *wf)
 
 static const struct intel_wakeref_ops wf_ops = {
 	.get = __gt_unpark,
+	.post_get = __gt_unpark_work_queue,
 	.put = __gt_park,
 };
 
diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
new file mode 100644
index 000000000000..23162dbd0c35
--- /dev/null
+++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.c
@@ -0,0 +1,35 @@ 
+// SPDX-License-Identifier: MIT
+/*
+ * Copyright © 2021 Intel Corporation
+ */
+
+#include "i915_drv.h"
+#include "intel_runtime_pm.h"
+#include "intel_gt_pm.h"
+
+void intel_gt_pm_unpark_work_queue(struct intel_gt *gt)
+{
+	struct intel_gt_pm_unpark_work *work, *next;
+	unsigned long flags;
+
+	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
+	list_for_each_entry_safe(work, next,
+				 &gt->pm_unpark_work_list, link) {
+		list_del_init(&work->link);
+		queue_work(system_unbound_wq, &work->worker);
+	}
+	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
+}
+
+void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
+				 struct intel_gt_pm_unpark_work *work)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&gt->pm_unpark_work_lock, flags);
+	if (intel_gt_pm_is_awake(gt))
+		queue_work(system_unbound_wq, &work->worker);
+	else if (list_empty(&work->link))
+		list_add_tail(&work->link, &gt->pm_unpark_work_list);
+	spin_unlock_irqrestore(&gt->pm_unpark_work_lock, flags);
+}
diff --git a/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
new file mode 100644
index 000000000000..eaf1dc313aa2
--- /dev/null
+++ b/drivers/gpu/drm/i915/gt/intel_gt_pm_unpark_work.h
@@ -0,0 +1,40 @@ 
+/* SPDX-License-Identifier: MIT */
+/*
+ * Copyright © 2021 Intel Corporation
+ */
+
+#ifndef INTEL_GT_PM_UNPARK_WORK_H
+#define INTEL_GT_PM_UNPARK_WORK_H
+
+#include <linux/list.h>
+#include <linux/workqueue.h>
+
+struct intel_gt;
+
+/**
+ * struct intel_gt_pm_unpark_work - work to be scheduled when GT unparked
+ */
+struct intel_gt_pm_unpark_work {
+	/**
+	 * @link: link into gt->pm_unpark_work_list of workers that need to be
+	 * scheduled when GT is unpark, protected by gt->pm_unpark_work_lock
+	 */
+	struct list_head link;
+	/** @worker: will be scheduled when GT unparked */
+	struct work_struct worker;
+};
+
+void intel_gt_pm_unpark_work_queue(struct intel_gt *gt);
+
+void intel_gt_pm_unpark_work_add(struct intel_gt *gt,
+				 struct intel_gt_pm_unpark_work *work);
+
+static inline void
+intel_gt_pm_unpark_work_init(struct intel_gt_pm_unpark_work *work,
+			     work_func_t fn)
+{
+	INIT_LIST_HEAD(&work->link);
+	INIT_WORK(&work->worker, fn);
+}
+
+#endif /* INTEL_GT_PM_UNPARK_WORK_H */
diff --git a/drivers/gpu/drm/i915/gt/intel_gt_types.h b/drivers/gpu/drm/i915/gt/intel_gt_types.h
index a81e21bf1bd1..4480312f0add 100644
--- a/drivers/gpu/drm/i915/gt/intel_gt_types.h
+++ b/drivers/gpu/drm/i915/gt/intel_gt_types.h
@@ -96,6 +96,16 @@  struct intel_gt {
 	struct intel_wakeref wakeref;
 	atomic_t user_wakeref;
 
+	/**
+	 * @pm_unpark_work_list: list of delayed work to scheduled which GT is
+	 * unparked, protected by pm_unpark_work_lock
+	 */
+	struct list_head pm_unpark_work_list;
+	/**
+	 * @pm_unpark_work_lock: protects pm_unpark_work_list
+	 */
+	spinlock_t pm_unpark_work_lock;
+
 	struct list_head closed_vma;
 	spinlock_t closed_lock; /* guards the list of closed_vma */
 
diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc.h b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
index 7358883f1540..023953e77553 100644
--- a/drivers/gpu/drm/i915/gt/uc/intel_guc.h
+++ b/drivers/gpu/drm/i915/gt/uc/intel_guc.h
@@ -19,6 +19,7 @@ 
 #include "intel_uc_fw.h"
 #include "i915_utils.h"
 #include "i915_vma.h"
+#include "gt/intel_gt_pm_unpark_work.h"
 
 struct __guc_ads_blob;
 
@@ -78,11 +79,12 @@  struct intel_guc {
 		 */
 		struct list_head destroyed_contexts;
 		/**
-		 * @destroyed_worker: worker to deregister contexts, need as we
+		 * @destroyed_worker: Worker to deregister contexts, need as we
 		 * need to take a GT PM reference and can't from destroy
-		 * function as it might be in an atomic context (no sleeping)
+		 * function as it might be in an atomic context (no sleeping).
+		 * Worker only issues deregister when GT is unparked.
 		 */
-		struct work_struct destroyed_worker;
+		struct intel_gt_pm_unpark_work destroyed_worker;
 	} submission_state;
 
 	bool submission_supported;
diff --git a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
index f835e06e5f9f..dbf919801de2 100644
--- a/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
+++ b/drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
@@ -1135,7 +1135,8 @@  int intel_guc_submission_init(struct intel_guc *guc)
 	INIT_LIST_HEAD(&guc->submission_state.guc_id_list);
 	ida_init(&guc->submission_state.guc_ids);
 	INIT_LIST_HEAD(&guc->submission_state.destroyed_contexts);
-	INIT_WORK(&guc->submission_state.destroyed_worker, destroyed_worker_func);
+	intel_gt_pm_unpark_work_init(&guc->submission_state.destroyed_worker,
+				     destroyed_worker_func);
 
 	return 0;
 }
@@ -1942,13 +1943,18 @@  static void deregister_destroyed_contexts(struct intel_guc *guc)
 
 static void destroyed_worker_func(struct work_struct *w)
 {
-	struct intel_guc *guc = container_of(w, struct intel_guc,
+	struct intel_gt_pm_unpark_work *destroyed_worker =
+		container_of(w, struct intel_gt_pm_unpark_work, worker);
+	struct intel_guc *guc = container_of(destroyed_worker, struct intel_guc,
 					     submission_state.destroyed_worker);
 	struct intel_gt *gt = guc_to_gt(guc);
 	int tmp;
 
-	with_intel_gt_pm(gt, tmp)
+	with_intel_gt_pm_if_awake(gt, tmp)
 		deregister_destroyed_contexts(guc);
+
+	if (!list_empty(&guc->submission_state.destroyed_contexts))
+		intel_gt_pm_unpark_work_add(gt, destroyed_worker);
 }
 
 static void guc_context_destroy(struct kref *kref)
@@ -1985,7 +1991,8 @@  static void guc_context_destroy(struct kref *kref)
 	 * take the GT PM for the first time which isn't allowed from an atomic
 	 * context.
 	 */
-	queue_work(system_unbound_wq, &guc->submission_state.destroyed_worker);
+	intel_gt_pm_unpark_work_add(guc_to_gt(guc),
+				    &guc->submission_state.destroyed_worker);
 }
 
 static int guc_context_alloc(struct intel_context *ce)
diff --git a/drivers/gpu/drm/i915/intel_wakeref.c b/drivers/gpu/drm/i915/intel_wakeref.c
index dfd87d082218..282fc4f312e3 100644
--- a/drivers/gpu/drm/i915/intel_wakeref.c
+++ b/drivers/gpu/drm/i915/intel_wakeref.c
@@ -24,6 +24,8 @@  static void rpm_put(struct intel_wakeref *wf)
 
 int __intel_wakeref_get_first(struct intel_wakeref *wf)
 {
+	bool do_post = false;
+
 	/*
 	 * Treat get/put as different subclasses, as we may need to run
 	 * the put callback from under the shrinker and do not want to
@@ -44,8 +46,11 @@  int __intel_wakeref_get_first(struct intel_wakeref *wf)
 		}
 
 		smp_mb__before_atomic(); /* release wf->count */
+		do_post = true;
 	}
 	atomic_inc(&wf->count);
+	if (do_post && wf->ops->post_get)
+		wf->ops->post_get(wf);
 	mutex_unlock(&wf->mutex);
 
 	INTEL_WAKEREF_BUG_ON(atomic_read(&wf->count) <= 0);
diff --git a/drivers/gpu/drm/i915/intel_wakeref.h b/drivers/gpu/drm/i915/intel_wakeref.h
index 545c8f277c46..ef7e6a698e8a 100644
--- a/drivers/gpu/drm/i915/intel_wakeref.h
+++ b/drivers/gpu/drm/i915/intel_wakeref.h
@@ -30,6 +30,7 @@  typedef depot_stack_handle_t intel_wakeref_t;
 
 struct intel_wakeref_ops {
 	int (*get)(struct intel_wakeref *wf);
+	void (*post_get)(struct intel_wakeref *wf);
 	int (*put)(struct intel_wakeref *wf);
 };