diff mbox series

[v1.1] evtchn: add early-out to evtchn_move_pirqs()

Message ID acd0dfae-b045-8505-3f6c-30ce72653660@suse.com (mailing list archive)
State New, archived
Headers show
Series [v1.1] evtchn: add early-out to evtchn_move_pirqs() | expand

Commit Message

Jan Beulich April 26, 2022, 10:33 a.m. UTC
See the code comment. The higher the rate of vCPU-s migrating across
pCPU-s, the less useful this attempted optimization actually is. With
credit2 the migration rate looks to be unduly high even on mostly idle
systems, and hence on large systems lock contention here isn't very
difficult to observe (as was the case for a failed 4.12 osstest flight).

Signed-off-by: Jan Beulich <jbeulich@suse.com>
Tested-by: Luca Fancellu <luca.fancellu@arm.com>

Comments

Julien Grall April 26, 2022, 10:44 a.m. UTC | #1
Hi Jan,

On 26/04/2022 11:33, Jan Beulich wrote:
> See the code comment. The higher the rate of vCPU-s migrating across
> pCPU-s, the less useful this attempted optimization actually is. With
> credit2 the migration rate looks to be unduly high even on mostly idle
> systems, and hence on large systems lock contention here isn't very
> difficult to observe (as was the case for a failed 4.12 osstest flight).
> 
> Signed-off-by: Jan Beulich <jbeulich@suse.com>
> Tested-by: Luca Fancellu <luca.fancellu@arm.com>
> 
> --- a/xen/common/event_channel.c
> +++ b/xen/common/event_channel.c
> @@ -1559,6 +1559,16 @@ void evtchn_move_pirqs(struct vcpu *v)
>       unsigned int port;
>       struct evtchn *chn;
>   
> +    /*
> +     * The work done below is an attempt to keep pIRQ-s on the pCPU-s that the
> +     * vCPU-s they're to be delivered to run on. In order to limit lock
> +     * contention, check for an empty list prior to acquiring the lock. In the
> +     * worst case a pIRQ just bound to this vCPU will be delivered elsewhere
> +     * until the vCPU is migrated (again) to another pCPU.
> +     */
> +    if ( !v->pirq_evtchn_head )
> +        return;

I was hoping Andrew would give some insight (hence why I haven't replied 
to your previous answer).

I am still not really convinced about this optimization. Aside what I 
wrote about the IRQ raised on the "wrong" pCPU, the lock contention 
would still be present if an OS is deciding to spread the PIRQs across 
all the vCPUs.

So it seems to me switching to a rwlock would help to address the 
contention on all the cases.

> +
>       spin_lock(&d->event_lock);
>       for ( port = v->pirq_evtchn_head; port; port = chn->u.pirq.next_port )
>       {
> 

Cheers,
Jan Beulich July 5, 2022, 4:10 p.m. UTC | #2
On 26.04.2022 12:44, Julien Grall wrote:
> On 26/04/2022 11:33, Jan Beulich wrote:
>> See the code comment. The higher the rate of vCPU-s migrating across
>> pCPU-s, the less useful this attempted optimization actually is. With
>> credit2 the migration rate looks to be unduly high even on mostly idle
>> systems, and hence on large systems lock contention here isn't very
>> difficult to observe (as was the case for a failed 4.12 osstest flight).
>>
>> Signed-off-by: Jan Beulich <jbeulich@suse.com>
>> Tested-by: Luca Fancellu <luca.fancellu@arm.com>
>>
>> --- a/xen/common/event_channel.c
>> +++ b/xen/common/event_channel.c
>> @@ -1559,6 +1559,16 @@ void evtchn_move_pirqs(struct vcpu *v)
>>       unsigned int port;
>>       struct evtchn *chn;
>>   
>> +    /*
>> +     * The work done below is an attempt to keep pIRQ-s on the pCPU-s that the
>> +     * vCPU-s they're to be delivered to run on. In order to limit lock
>> +     * contention, check for an empty list prior to acquiring the lock. In the
>> +     * worst case a pIRQ just bound to this vCPU will be delivered elsewhere
>> +     * until the vCPU is migrated (again) to another pCPU.
>> +     */
>> +    if ( !v->pirq_evtchn_head )
>> +        return;
> 
> I was hoping Andrew would give some insight (hence why I haven't replied 
> to your previous answer).
> 
> I am still not really convinced about this optimization. Aside what I 
> wrote about the IRQ raised on the "wrong" pCPU, the lock contention 
> would still be present if an OS is deciding to spread the PIRQs across 
> all the vCPUs.
> 
> So it seems to me switching to a rwlock would help to address the 
> contention on all the cases.

But that patch of mine didn't get any ack either, and hence at some
point I've shelved it. The same is going to happen to this patch,
and sooner or later we'll re-observe the issue osstest had hit (at
least) once.

Jan
Julien Grall July 5, 2022, 9:43 p.m. UTC | #3
Hi Jan,

On 05/07/2022 17:10, Jan Beulich wrote:
> On 26.04.2022 12:44, Julien Grall wrote:
>> On 26/04/2022 11:33, Jan Beulich wrote:
>>> See the code comment. The higher the rate of vCPU-s migrating across
>>> pCPU-s, the less useful this attempted optimization actually is. With
>>> credit2 the migration rate looks to be unduly high even on mostly idle
>>> systems, and hence on large systems lock contention here isn't very
>>> difficult to observe (as was the case for a failed 4.12 osstest flight).
>>>
>>> Signed-off-by: Jan Beulich <jbeulich@suse.com>
>>> Tested-by: Luca Fancellu <luca.fancellu@arm.com>
>>>
>>> --- a/xen/common/event_channel.c
>>> +++ b/xen/common/event_channel.c
>>> @@ -1559,6 +1559,16 @@ void evtchn_move_pirqs(struct vcpu *v)
>>>        unsigned int port;
>>>        struct evtchn *chn;
>>>    
>>> +    /*
>>> +     * The work done below is an attempt to keep pIRQ-s on the pCPU-s that the
>>> +     * vCPU-s they're to be delivered to run on. In order to limit lock
>>> +     * contention, check for an empty list prior to acquiring the lock. In the
>>> +     * worst case a pIRQ just bound to this vCPU will be delivered elsewhere
>>> +     * until the vCPU is migrated (again) to another pCPU.
>>> +     */
>>> +    if ( !v->pirq_evtchn_head )
>>> +        return;
>>
>> I was hoping Andrew would give some insight (hence why I haven't replied
>> to your previous answer).
>>
>> I am still not really convinced about this optimization. Aside what I
>> wrote about the IRQ raised on the "wrong" pCPU, the lock contention
>> would still be present if an OS is deciding to spread the PIRQs across
>> all the vCPUs.
>>
>> So it seems to me switching to a rwlock would help to address the
>> contention on all the cases.
> 
> But that patch of mine didn't get any ack either, and hence at some
> point I've shelved it.

Looking at v5, your latest answer suggested you were going to drop the 
patch. So I didn't bother to review/ack the latest version.

Now, in the context of this discussion, I think this is better than this 
approach. I will review the other patch.

Cheers,
diff mbox series

Patch

--- a/xen/common/event_channel.c
+++ b/xen/common/event_channel.c
@@ -1559,6 +1559,16 @@  void evtchn_move_pirqs(struct vcpu *v)
     unsigned int port;
     struct evtchn *chn;
 
+    /*
+     * The work done below is an attempt to keep pIRQ-s on the pCPU-s that the
+     * vCPU-s they're to be delivered to run on. In order to limit lock
+     * contention, check for an empty list prior to acquiring the lock. In the
+     * worst case a pIRQ just bound to this vCPU will be delivered elsewhere
+     * until the vCPU is migrated (again) to another pCPU.
+     */
+    if ( !v->pirq_evtchn_head )
+        return;
+
     spin_lock(&d->event_lock);
     for ( port = v->pirq_evtchn_head; port; port = chn->u.pirq.next_port )
     {