From patchwork Tue May 28 10:32:37 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: =?utf-8?b?SsO8cmdlbiBHcm/Dnw==?= X-Patchwork-Id: 10964587 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 9E54892A for ; Tue, 28 May 2019 10:34:31 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 8F60B1FE82 for ; Tue, 28 May 2019 10:34:31 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 83C2D28815; Tue, 28 May 2019 10:34:31 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.2 required=2.0 tests=BAYES_00,MAILING_LIST_MULTI, RCVD_IN_DNSWL_MED autolearn=ham version=3.3.1 Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) (using TLSv1.2 with cipher AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.wl.linuxfoundation.org (Postfix) with ESMTPS id 44ADB26E4F for ; Tue, 28 May 2019 10:34:30 +0000 (UTC) Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.89) (envelope-from ) id 1hVZQE-0004sa-3e; Tue, 28 May 2019 10:33:34 +0000 Received: from us1-rack-dfw2.inumbo.com ([104.130.134.6]) by lists.xenproject.org with esmtp (Exim 4.89) (envelope-from ) id 1hVZQ6-0004dr-OU for xen-devel@lists.xenproject.org; Tue, 28 May 2019 10:33:26 +0000 X-Inumbo-ID: 0528ab92-8134-11e9-8980-bc764e045a96 Received: from mx1.suse.de (unknown [195.135.220.15]) by us1-rack-dfw2.inumbo.com (Halon) with ESMTPS id 0528ab92-8134-11e9-8980-bc764e045a96; Tue, 28 May 2019 10:33:24 +0000 (UTC) X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id 9EE50B040; Tue, 28 May 2019 10:33:22 +0000 (UTC) From: Juergen Gross To: xen-devel@lists.xenproject.org Date: Tue, 28 May 2019 12:32:37 +0200 Message-Id: <20190528103313.1343-25-jgross@suse.com> X-Mailer: git-send-email 2.16.4 In-Reply-To: <20190528103313.1343-1-jgross@suse.com> References: <20190528103313.1343-1-jgross@suse.com> Subject: [Xen-devel] [PATCH 24/60] xen/sched: make arinc653 scheduler vcpu agnostic. X-BeenThere: xen-devel@lists.xenproject.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Cc: Juergen Gross , George Dunlap , Josh Whitehead , Robert VanVossen , Dario Faggioli MIME-Version: 1.0 Errors-To: xen-devel-bounces@lists.xenproject.org Sender: "Xen-devel" X-Virus-Scanned: ClamAV using ClamSMTP Switch arinc653 scheduler completely from vcpu to sched_unit usage. Signed-off-by: Juergen Gross --- xen/common/sched_arinc653.c | 208 +++++++++++++++++++++----------------------- 1 file changed, 101 insertions(+), 107 deletions(-) diff --git a/xen/common/sched_arinc653.c b/xen/common/sched_arinc653.c index 0c75440bd0..213bc960ef 100644 --- a/xen/common/sched_arinc653.c +++ b/xen/common/sched_arinc653.c @@ -45,15 +45,15 @@ #define DEFAULT_TIMESLICE MILLISECS(10) /** - * Retrieve the idle VCPU for a given physical CPU + * Retrieve the idle UNIT for a given physical CPU */ -#define IDLETASK(cpu) (idle_vcpu[cpu]) +#define IDLETASK(cpu) (sched_idle_unit(cpu)) /** * Return a pointer to the ARINC 653-specific scheduler data information - * associated with the given VCPU (vc) + * associated with the given UNIT (unit) */ -#define AVCPU(vc) ((arinc653_vcpu_t *)(vc)->sched_unit->priv) +#define AUNIT(unit) ((arinc653_unit_t *)(unit)->priv) /** * Return the global scheduler private data given the scheduler ops pointer @@ -65,20 +65,20 @@ **************************************************************************/ /** - * The arinc653_vcpu_t structure holds ARINC 653-scheduler-specific - * information for all non-idle VCPUs + * The arinc653_unit_t structure holds ARINC 653-scheduler-specific + * information for all non-idle UNITs */ -typedef struct arinc653_vcpu_s +typedef struct arinc653_unit_s { - /* vc points to Xen's struct vcpu so we can get to it from an - * arinc653_vcpu_t pointer. */ - struct vcpu * vc; - /* awake holds whether the VCPU has been woken with vcpu_wake() */ + /* unit points to Xen's struct sched_unit so we can get to it from an + * arinc653_unit_t pointer. */ + struct sched_unit * unit; + /* awake holds whether the UNIT has been woken with vcpu_wake() */ bool_t awake; - /* list holds the linked list information for the list this VCPU + /* list holds the linked list information for the list this UNIT * is stored in */ struct list_head list; -} arinc653_vcpu_t; +} arinc653_unit_t; /** * The sched_entry_t structure holds a single entry of the @@ -89,14 +89,14 @@ typedef struct sched_entry_s /* dom_handle holds the handle ("UUID") for the domain that this * schedule entry refers to. */ xen_domain_handle_t dom_handle; - /* vcpu_id holds the VCPU number for the VCPU that this schedule + /* unit_id holds the UNIT number for the UNIT that this schedule * entry refers to. */ - int vcpu_id; - /* runtime holds the number of nanoseconds that the VCPU for this + int unit_id; + /* runtime holds the number of nanoseconds that the UNIT for this * schedule entry should be allowed to run per major frame. */ s_time_t runtime; - /* vc holds a pointer to the Xen VCPU structure */ - struct vcpu * vc; + /* unit holds a pointer to the Xen sched_unit structure */ + struct sched_unit * unit; } sched_entry_t; /** @@ -110,9 +110,9 @@ typedef struct a653sched_priv_s /** * This array holds the active ARINC 653 schedule. * - * When the system tries to start a new VCPU, this schedule is scanned - * to look for a matching (handle, VCPU #) pair. If both the handle (UUID) - * and VCPU number match, then the VCPU is allowed to run. Its run time + * When the system tries to start a new UNIT, this schedule is scanned + * to look for a matching (handle, UNIT #) pair. If both the handle (UUID) + * and UNIT number match, then the UNIT is allowed to run. Its run time * (per major frame) is given in the third entry of the schedule. */ sched_entry_t schedule[ARINC653_MAX_DOMAINS_PER_SCHEDULE]; @@ -123,8 +123,8 @@ typedef struct a653sched_priv_s * * This is not necessarily the same as the number of domains in the * schedule. A domain could be listed multiple times within the schedule, - * or a domain with multiple VCPUs could have a different - * schedule entry for each VCPU. + * or a domain with multiple UNITs could have a different + * schedule entry for each UNIT. */ unsigned int num_schedule_entries; @@ -139,9 +139,9 @@ typedef struct a653sched_priv_s s_time_t next_major_frame; /** - * pointers to all Xen VCPU structures for iterating through + * pointers to all Xen UNIT structures for iterating through */ - struct list_head vcpu_list; + struct list_head unit_list; } a653sched_priv_t; /************************************************************************** @@ -167,50 +167,50 @@ static int dom_handle_cmp(const xen_domain_handle_t h1, } /** - * This function searches the vcpu list to find a VCPU that matches - * the domain handle and VCPU ID specified. + * This function searches the unit list to find a UNIT that matches + * the domain handle and UNIT ID specified. * * @param ops Pointer to this instance of the scheduler structure * @param handle Pointer to handler - * @param vcpu_id VCPU ID + * @param unit_id UNIT ID * * @return
    - *
  • Pointer to the matching VCPU if one is found + *
  • Pointer to the matching UNIT if one is found *
  • NULL otherwise *
*/ -static struct vcpu *find_vcpu( +static struct sched_unit *find_unit( const struct scheduler *ops, xen_domain_handle_t handle, - int vcpu_id) + int unit_id) { - arinc653_vcpu_t *avcpu; + arinc653_unit_t *aunit; - /* loop through the vcpu_list looking for the specified VCPU */ - list_for_each_entry ( avcpu, &SCHED_PRIV(ops)->vcpu_list, list ) - if ( (dom_handle_cmp(avcpu->vc->domain->handle, handle) == 0) - && (vcpu_id == avcpu->vc->vcpu_id) ) - return avcpu->vc; + /* loop through the unit_list looking for the specified UNIT */ + list_for_each_entry ( aunit, &SCHED_PRIV(ops)->unit_list, list ) + if ( (dom_handle_cmp(aunit->unit->domain->handle, handle) == 0) + && (unit_id == aunit->unit->unit_id) ) + return aunit->unit; return NULL; } /** - * This function updates the pointer to the Xen VCPU structure for each entry + * This function updates the pointer to the Xen UNIT structure for each entry * in the ARINC 653 schedule. * * @param ops Pointer to this instance of the scheduler structure * @return */ -static void update_schedule_vcpus(const struct scheduler *ops) +static void update_schedule_units(const struct scheduler *ops) { unsigned int i, n_entries = SCHED_PRIV(ops)->num_schedule_entries; for ( i = 0; i < n_entries; i++ ) - SCHED_PRIV(ops)->schedule[i].vc = - find_vcpu(ops, + SCHED_PRIV(ops)->schedule[i].unit = + find_unit(ops, SCHED_PRIV(ops)->schedule[i].dom_handle, - SCHED_PRIV(ops)->schedule[i].vcpu_id); + SCHED_PRIV(ops)->schedule[i].unit_id); } /** @@ -268,12 +268,12 @@ arinc653_sched_set( memcpy(sched_priv->schedule[i].dom_handle, schedule->sched_entries[i].dom_handle, sizeof(sched_priv->schedule[i].dom_handle)); - sched_priv->schedule[i].vcpu_id = + sched_priv->schedule[i].unit_id = schedule->sched_entries[i].vcpu_id; sched_priv->schedule[i].runtime = schedule->sched_entries[i].runtime; } - update_schedule_vcpus(ops); + update_schedule_units(ops); /* * The newly-installed schedule takes effect immediately. We do not even @@ -319,7 +319,7 @@ arinc653_sched_get( memcpy(schedule->sched_entries[i].dom_handle, sched_priv->schedule[i].dom_handle, sizeof(sched_priv->schedule[i].dom_handle)); - schedule->sched_entries[i].vcpu_id = sched_priv->schedule[i].vcpu_id; + schedule->sched_entries[i].vcpu_id = sched_priv->schedule[i].unit_id; schedule->sched_entries[i].runtime = sched_priv->schedule[i].runtime; } @@ -355,7 +355,7 @@ a653sched_init(struct scheduler *ops) prv->next_major_frame = 0; spin_lock_init(&prv->lock); - INIT_LIST_HEAD(&prv->vcpu_list); + INIT_LIST_HEAD(&prv->unit_list); return 0; } @@ -373,7 +373,7 @@ a653sched_deinit(struct scheduler *ops) } /** - * This function allocates scheduler-specific data for a VCPU + * This function allocates scheduler-specific data for a UNIT * * @param ops Pointer to this instance of the scheduler structure * @param unit Pointer to struct sched_unit @@ -385,35 +385,34 @@ a653sched_alloc_vdata(const struct scheduler *ops, struct sched_unit *unit, void *dd) { a653sched_priv_t *sched_priv = SCHED_PRIV(ops); - struct vcpu *vc = unit->vcpu; - arinc653_vcpu_t *svc; + arinc653_unit_t *svc; unsigned int entry; unsigned long flags; /* * Allocate memory for the ARINC 653-specific scheduler data information - * associated with the given VCPU (vc). + * associated with the given UNIT (unit). */ - svc = xmalloc(arinc653_vcpu_t); + svc = xmalloc(arinc653_unit_t); if ( svc == NULL ) return NULL; spin_lock_irqsave(&sched_priv->lock, flags); - /* - * Add every one of dom0's vcpus to the schedule, as long as there are + /* + * Add every one of dom0's units to the schedule, as long as there are * slots available. */ - if ( vc->domain->domain_id == 0 ) + if ( unit->domain->domain_id == 0 ) { entry = sched_priv->num_schedule_entries; if ( entry < ARINC653_MAX_DOMAINS_PER_SCHEDULE ) { sched_priv->schedule[entry].dom_handle[0] = '\0'; - sched_priv->schedule[entry].vcpu_id = vc->vcpu_id; + sched_priv->schedule[entry].unit_id = unit->unit_id; sched_priv->schedule[entry].runtime = DEFAULT_TIMESLICE; - sched_priv->schedule[entry].vc = vc; + sched_priv->schedule[entry].unit = unit; sched_priv->major_frame += DEFAULT_TIMESLICE; ++sched_priv->num_schedule_entries; @@ -421,16 +420,16 @@ a653sched_alloc_vdata(const struct scheduler *ops, struct sched_unit *unit, } /* - * Initialize our ARINC 653 scheduler-specific information for the VCPU. - * The VCPU starts "asleep." When Xen is ready for the VCPU to run, it + * Initialize our ARINC 653 scheduler-specific information for the UNIT. + * The UNIT starts "asleep." When Xen is ready for the UNIT to run, it * will call the vcpu_wake scheduler callback function and our scheduler - * will mark the VCPU awake. + * will mark the UNIT awake. */ - svc->vc = vc; + svc->unit = unit; svc->awake = 0; - if ( !is_idle_vcpu(vc) ) - list_add(&svc->list, &SCHED_PRIV(ops)->vcpu_list); - update_schedule_vcpus(ops); + if ( !is_idle_unit(unit) ) + list_add(&svc->list, &SCHED_PRIV(ops)->unit_list); + update_schedule_units(ops); spin_unlock_irqrestore(&sched_priv->lock, flags); @@ -438,27 +437,27 @@ a653sched_alloc_vdata(const struct scheduler *ops, struct sched_unit *unit, } /** - * This function frees scheduler-specific VCPU data + * This function frees scheduler-specific UNIT data * * @param ops Pointer to this instance of the scheduler structure */ static void a653sched_free_vdata(const struct scheduler *ops, void *priv) { - arinc653_vcpu_t *av = priv; + arinc653_unit_t *av = priv; if (av == NULL) return; - if ( !is_idle_vcpu(av->vc) ) + if ( !is_idle_unit(av->unit) ) list_del(&av->list); xfree(av); - update_schedule_vcpus(ops); + update_schedule_units(ops); } /** - * Xen scheduler callback function to sleep a VCPU + * Xen scheduler callback function to sleep a UNIT * * @param ops Pointer to this instance of the scheduler structure * @param unit Pointer to struct sched_unit @@ -466,21 +465,19 @@ a653sched_free_vdata(const struct scheduler *ops, void *priv) static void a653sched_unit_sleep(const struct scheduler *ops, struct sched_unit *unit) { - struct vcpu *vc = unit->vcpu; - - if ( AVCPU(vc) != NULL ) - AVCPU(vc)->awake = 0; + if ( AUNIT(unit) != NULL ) + AUNIT(unit)->awake = 0; /* - * If the VCPU being put to sleep is the same one that is currently + * If the UNIT being put to sleep is the same one that is currently * running, raise a softirq to invoke the scheduler to switch domains. */ - if ( get_sched_res(vc->processor)->curr == unit ) - cpu_raise_softirq(vc->processor, SCHEDULE_SOFTIRQ); + if ( get_sched_res(sched_unit_cpu(unit))->curr == unit ) + cpu_raise_softirq(sched_unit_cpu(unit), SCHEDULE_SOFTIRQ); } /** - * Xen scheduler callback function to wake up a VCPU + * Xen scheduler callback function to wake up a UNIT * * @param ops Pointer to this instance of the scheduler structure * @param unit Pointer to struct sched_unit @@ -488,24 +485,22 @@ a653sched_unit_sleep(const struct scheduler *ops, struct sched_unit *unit) static void a653sched_unit_wake(const struct scheduler *ops, struct sched_unit *unit) { - struct vcpu *vc = unit->vcpu; + if ( AUNIT(unit) != NULL ) + AUNIT(unit)->awake = 1; - if ( AVCPU(vc) != NULL ) - AVCPU(vc)->awake = 1; - - cpu_raise_softirq(vc->processor, SCHEDULE_SOFTIRQ); + cpu_raise_softirq(sched_unit_cpu(unit), SCHEDULE_SOFTIRQ); } /** - * Xen scheduler callback function to select a VCPU to run. + * Xen scheduler callback function to select a UNIT to run. * This is the main scheduler routine. * * @param ops Pointer to this instance of the scheduler structure * @param now Current time * - * @return Address of the VCPU structure scheduled to be run next - * Amount of time to execute the returned VCPU - * Flag for whether the VCPU was migrated + * @return Address of the UNIT structure scheduled to be run next + * Amount of time to execute the returned UNIT + * Flag for whether the UNIT was migrated */ static struct task_slice a653sched_do_schedule( @@ -514,7 +509,7 @@ a653sched_do_schedule( bool_t tasklet_work_scheduled) { struct task_slice ret; /* hold the chosen domain */ - struct vcpu * new_task = NULL; + struct sched_unit *new_task = NULL; static unsigned int sched_index = 0; static s_time_t next_switch_time; a653sched_priv_t *sched_priv = SCHED_PRIV(ops); @@ -559,14 +554,14 @@ a653sched_do_schedule( * sched_unit structure. */ new_task = (sched_index < sched_priv->num_schedule_entries) - ? sched_priv->schedule[sched_index].vc + ? sched_priv->schedule[sched_index].unit : IDLETASK(cpu); /* Check to see if the new task can be run (awake & runnable). */ if ( !((new_task != NULL) - && (AVCPU(new_task) != NULL) - && AVCPU(new_task)->awake - && vcpu_runnable(new_task)) ) + && (AUNIT(new_task) != NULL) + && AUNIT(new_task)->awake + && unit_runnable(new_task)) ) new_task = IDLETASK(cpu); BUG_ON(new_task == NULL); @@ -578,21 +573,21 @@ a653sched_do_schedule( spin_unlock_irqrestore(&sched_priv->lock, flags); - /* Tasklet work (which runs in idle VCPU context) overrides all else. */ + /* Tasklet work (which runs in idle UNIT context) overrides all else. */ if ( tasklet_work_scheduled ) new_task = IDLETASK(cpu); /* Running this task would result in a migration */ - if ( !is_idle_vcpu(new_task) - && (new_task->processor != cpu) ) + if ( !is_idle_unit(new_task) + && (sched_unit_cpu(new_task) != cpu) ) new_task = IDLETASK(cpu); /* * Return the amount of time the next domain has to run and the address - * of the selected task's VCPU structure. + * of the selected task's UNIT structure. */ ret.time = next_switch_time - now; - ret.task = new_task->sched_unit; + ret.task = new_task; ret.migrated = 0; BUG_ON(ret.time <= 0); @@ -601,7 +596,7 @@ a653sched_do_schedule( } /** - * Xen scheduler callback function to select a resource for the VCPU to run on + * Xen scheduler callback function to select a resource for the UNIT to run on * * @param ops Pointer to this instance of the scheduler structure * @param unit Pointer to struct sched_unit @@ -611,21 +606,20 @@ a653sched_do_schedule( static struct sched_resource * a653sched_pick_resource(const struct scheduler *ops, struct sched_unit *unit) { - struct vcpu *vc = unit->vcpu; cpumask_t *online; unsigned int cpu; - /* - * If present, prefer vc's current processor, else - * just find the first valid vcpu . + /* + * If present, prefer unit's current processor, else + * just find the first valid unit. */ - online = cpupool_domain_cpumask(vc->domain); + online = cpupool_domain_cpumask(unit->domain); cpu = cpumask_first(online); - if ( cpumask_test_cpu(vc->processor, online) + if ( cpumask_test_cpu(sched_unit_cpu(unit), online) || (cpu >= nr_cpu_ids) ) - cpu = vc->processor; + cpu = sched_unit_cpu(unit); return get_sched_res(cpu); } @@ -636,18 +630,18 @@ a653sched_pick_resource(const struct scheduler *ops, struct sched_unit *unit) * @param new_ops Pointer to this instance of the scheduler structure * @param cpu The cpu that is changing scheduler * @param pdata scheduler specific PCPU data (we don't have any) - * @param vdata scheduler specific VCPU data of the idle vcpu + * @param vdata scheduler specific UNIT data of the idle unit */ static spinlock_t * a653_switch_sched(struct scheduler *new_ops, unsigned int cpu, void *pdata, void *vdata) { struct sched_resource *sd = get_sched_res(cpu); - arinc653_vcpu_t *svc = vdata; + arinc653_unit_t *svc = vdata; - ASSERT(!pdata && svc && is_idle_vcpu(svc->vc)); + ASSERT(!pdata && svc && is_idle_unit(svc->unit)); - idle_vcpu[cpu]->sched_unit->priv = vdata; + sched_idle_unit(cpu)->priv = vdata; return &sd->_lock; }