diff mbox

[5/7] xen/9pfs: send requests to the backend

Message ID 1488830488-18506-5-git-send-email-sstabellini@kernel.org (mailing list archive)
State New, archived
Headers show

Commit Message

Stefano Stabellini March 6, 2017, 8:01 p.m. UTC
Implement struct p9_trans_module create and close functions by looking
at the available Xen 9pfs frontend-backend connections. We don't expect
many frontend-backend connections, thus walking a list is OK.

Send requests to the backend by copying each request to one of the
available rings (each frontend-backend connection comes with multiple
rings). Handle the ring and notifications following the 9pfs
specification. If there are not enough free bytes on the ring for the
request, wait on the wait_queue: the backend will send a notification
after consuming more requests.

Signed-off-by: Stefano Stabellini <stefano@aporeto.com>
CC: boris.ostrovsky@oracle.com
CC: jgross@suse.com
CC: Eric Van Hensbergen <ericvh@gmail.com>
CC: Ron Minnich <rminnich@sandia.gov>
CC: Latchesar Ionkov <lucho@ionkov.net>
CC: v9fs-developer@lists.sourceforge.net
---
 net/9p/trans_xen.c | 83 +++++++++++++++++++++++++++++++++++++++++++++++++++++-
 1 file changed, 82 insertions(+), 1 deletion(-)

Comments

Boris Ostrovsky March 7, 2017, 3:27 p.m. UTC | #1
On 03/06/2017 03:01 PM, Stefano Stabellini wrote:
> Implement struct p9_trans_module create and close functions by looking
> at the available Xen 9pfs frontend-backend connections. We don't expect
> many frontend-backend connections, thus walking a list is OK.
> 
> Send requests to the backend by copying each request to one of the
> available rings (each frontend-backend connection comes with multiple
> rings). Handle the ring and notifications following the 9pfs
> specification. If there are not enough free bytes on the ring for the
> request, wait on the wait_queue: the backend will send a notification
> after consuming more requests.
> 
> Signed-off-by: Stefano Stabellini <stefano@aporeto.com>
> CC: boris.ostrovsky@oracle.com
> CC: jgross@suse.com
> CC: Eric Van Hensbergen <ericvh@gmail.com>
> CC: Ron Minnich <rminnich@sandia.gov>
> CC: Latchesar Ionkov <lucho@ionkov.net>
> CC: v9fs-developer@lists.sourceforge.net
> ---
>  net/9p/trans_xen.c | 83 +++++++++++++++++++++++++++++++++++++++++++++++++++++-
>  1 file changed, 82 insertions(+), 1 deletion(-)
> 
> diff --git a/net/9p/trans_xen.c b/net/9p/trans_xen.c
> index 9f6cf8d..4e26556 100644
> --- a/net/9p/trans_xen.c
> +++ b/net/9p/trans_xen.c
> @@ -47,22 +47,103 @@ struct xen_9pfs_front_priv {
>  };
>  static LIST_HEAD(xen_9pfs_devs);
>  
> +/* We don't currently allow canceling of requests */
>  static int p9_xen_cancel(struct p9_client *client, struct p9_req_t *req)
>  {
> -	return 0;
> +	return 1;
>  }
>  
>  static int p9_xen_create(struct p9_client *client, const char *addr, char *args)
>  {
> +	struct xen_9pfs_front_priv *priv = NULL;
> +
> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> +		if (!strcmp(priv->tag, addr))
> +			break;
> +	}


You could simplify this (and p9_xen_close()) but assigning client and
returning from inside the 'if' statement.

I am also not sure you need to initialize priv.


> +	if (!priv || strcmp(priv->tag, addr))
> +		return -EINVAL;
> +
> +	priv->client = client; 
>  	return 0;
>  }
>  
>  static void p9_xen_close(struct p9_client *client)
>  {
> +	struct xen_9pfs_front_priv *priv = NULL;
> +
> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> +		if (priv->client == client)
> +			break;
> +	}
> +	if (!priv || priv->client != client)
> +		return;
> +
> +	priv->client = NULL; 
> +	return;
> +}
> +
> +static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
> +{
> +	RING_IDX cons, prod;
> +
> +	cons = ring->intf->out_cons;
> +	prod = ring->intf->out_prod;
> +	mb();
> +
> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
> +		return 1;
> +	else
> +		return 0;
>  }
>  
>  static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
>  {
> +	struct xen_9pfs_front_priv *priv = NULL;
> +	RING_IDX cons, prod, masked_cons, masked_prod;
> +	unsigned long flags;
> +	uint32_t size = p9_req->tc->size;
> +	struct xen_9pfs_dataring *ring;
> +	int num;
> +
> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> +		if (priv->client == client)
> +			break;
> +	}
> +	if (priv == NULL || priv->client != client)
> +		return -EINVAL;
> +
> +	num = p9_req->tc->tag % priv->num_rings;
> +	ring = &priv->rings[num];
> +
> +again:
> +	while (wait_event_interruptible(ring->wq,
> +				p9_xen_write_todo(ring, size) > 0) != 0);
> +
> +	spin_lock_irqsave(&ring->lock, flags);
> +	cons = ring->intf->out_cons;
> +	prod = ring->intf->out_prod;
> +	mb();
> +
> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {


This looks like p9_xen_write_todo(). BTW, where is xen_9pfs_queued()
defined? I couldn't find it. Same for xen_9pfs_mask() and
xen_9pfs_write_packet().

-boris


> +		spin_unlock_irqrestore(&ring->lock, flags);
> +		goto again;
> +	}
> +
> +	masked_prod = xen_9pfs_mask(prod, XEN_9PFS_RING_SIZE);
> +	masked_cons = xen_9pfs_mask(cons, XEN_9PFS_RING_SIZE);
> +
> +	xen_9pfs_write_packet(ring->ring.out,
> +				&masked_prod, masked_cons,
> +				XEN_9PFS_RING_SIZE, p9_req->tc->sdata, size);
> +
> +	p9_req->status = REQ_STATUS_SENT;
> +	wmb();			/* write ring before updating pointer */
> +	prod += size;
> +	ring->intf->out_prod = prod;
> +	spin_unlock_irqrestore(&ring->lock, flags);
> +	notify_remote_via_irq(ring->irq);
> +
>  	return 0;
>  }
>  
>
Stefano Stabellini March 8, 2017, 12:55 a.m. UTC | #2
On Tue, 7 Mar 2017, Boris Ostrovsky wrote:
> On 03/06/2017 03:01 PM, Stefano Stabellini wrote:
> > Implement struct p9_trans_module create and close functions by looking
> > at the available Xen 9pfs frontend-backend connections. We don't expect
> > many frontend-backend connections, thus walking a list is OK.
> > 
> > Send requests to the backend by copying each request to one of the
> > available rings (each frontend-backend connection comes with multiple
> > rings). Handle the ring and notifications following the 9pfs
> > specification. If there are not enough free bytes on the ring for the
> > request, wait on the wait_queue: the backend will send a notification
> > after consuming more requests.
> > 
> > Signed-off-by: Stefano Stabellini <stefano@aporeto.com>
> > CC: boris.ostrovsky@oracle.com
> > CC: jgross@suse.com
> > CC: Eric Van Hensbergen <ericvh@gmail.com>
> > CC: Ron Minnich <rminnich@sandia.gov>
> > CC: Latchesar Ionkov <lucho@ionkov.net>
> > CC: v9fs-developer@lists.sourceforge.net
> > ---
> >  net/9p/trans_xen.c | 83 +++++++++++++++++++++++++++++++++++++++++++++++++++++-
> >  1 file changed, 82 insertions(+), 1 deletion(-)
> > 
> > diff --git a/net/9p/trans_xen.c b/net/9p/trans_xen.c
> > index 9f6cf8d..4e26556 100644
> > --- a/net/9p/trans_xen.c
> > +++ b/net/9p/trans_xen.c
> > @@ -47,22 +47,103 @@ struct xen_9pfs_front_priv {
> >  };
> >  static LIST_HEAD(xen_9pfs_devs);
> >  
> > +/* We don't currently allow canceling of requests */
> >  static int p9_xen_cancel(struct p9_client *client, struct p9_req_t *req)
> >  {
> > -	return 0;
> > +	return 1;
> >  }
> >  
> >  static int p9_xen_create(struct p9_client *client, const char *addr, char *args)
> >  {
> > +	struct xen_9pfs_front_priv *priv = NULL;
> > +
> > +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> > +		if (!strcmp(priv->tag, addr))
> > +			break;
> > +	}
> 
> 
> You could simplify this (and p9_xen_close()) but assigning client and
> returning from inside the 'if' statement.

I'll do that.


> I am also not sure you need to initialize priv.
 
With the new changes, I won't need to.

 
> > +	if (!priv || strcmp(priv->tag, addr))
> > +		return -EINVAL;
> > +
> > +	priv->client = client; 
> >  	return 0;
> >  }
> >  
> >  static void p9_xen_close(struct p9_client *client)
> >  {
> > +	struct xen_9pfs_front_priv *priv = NULL;
> > +
> > +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> > +		if (priv->client == client)
> > +			break;
> > +	}
> > +	if (!priv || priv->client != client)
> > +		return;
> > +
> > +	priv->client = NULL; 
> > +	return;
> > +}
> > +
> > +static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
> > +{
> > +	RING_IDX cons, prod;
> > +
> > +	cons = ring->intf->out_cons;
> > +	prod = ring->intf->out_prod;
> > +	mb();
> > +
> > +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
> > +		return 1;
> > +	else
> > +		return 0;
> >  }
> >  
> >  static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
> >  {
> > +	struct xen_9pfs_front_priv *priv = NULL;
> > +	RING_IDX cons, prod, masked_cons, masked_prod;
> > +	unsigned long flags;
> > +	uint32_t size = p9_req->tc->size;
> > +	struct xen_9pfs_dataring *ring;
> > +	int num;
> > +
> > +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> > +		if (priv->client == client)
> > +			break;
> > +	}
> > +	if (priv == NULL || priv->client != client)
> > +		return -EINVAL;
> > +
> > +	num = p9_req->tc->tag % priv->num_rings;
> > +	ring = &priv->rings[num];
> > +
> > +again:
> > +	while (wait_event_interruptible(ring->wq,
> > +				p9_xen_write_todo(ring, size) > 0) != 0);
> > +
> > +	spin_lock_irqsave(&ring->lock, flags);
> > +	cons = ring->intf->out_cons;
> > +	prod = ring->intf->out_prod;
> > +	mb();
> > +
> > +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {
> 
> 
> This looks like p9_xen_write_todo().

p9_xen_write_todo is just a wrapper around xen_9pfs_queued to provide
a return value that works well with wait_event_interruptible.

I would prefer not to call p9_xen_write_todo here, because it's simpler
if we don't read prod and cons twice.


> BTW, where is xen_9pfs_queued()
> defined? I couldn't find it. Same for xen_9pfs_mask() and
> xen_9pfs_write_packet().

They are provided by the new ring macros, see
include/xen/interface/io/ring.h (the first patch).


> > +		spin_unlock_irqrestore(&ring->lock, flags);
> > +		goto again;
> > +	}
> > +
> > +	masked_prod = xen_9pfs_mask(prod, XEN_9PFS_RING_SIZE);
> > +	masked_cons = xen_9pfs_mask(cons, XEN_9PFS_RING_SIZE);
> > +
> > +	xen_9pfs_write_packet(ring->ring.out,
> > +				&masked_prod, masked_cons,
> > +				XEN_9PFS_RING_SIZE, p9_req->tc->sdata, size);
> > +
> > +	p9_req->status = REQ_STATUS_SENT;
> > +	wmb();			/* write ring before updating pointer */
> > +	prod += size;
> > +	ring->intf->out_prod = prod;
> > +	spin_unlock_irqrestore(&ring->lock, flags);
> > +	notify_remote_via_irq(ring->irq);
> > +
> >  	return 0;
> >  }
> >  
> > 
>
Boris Ostrovsky March 8, 2017, 1:58 p.m. UTC | #3
>>> +}
>>> +
>>> +static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
>>> +{
>>> +	RING_IDX cons, prod;
>>> +
>>> +	cons = ring->intf->out_cons;
>>> +	prod = ring->intf->out_prod;
>>> +	mb();
>>> +
>>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
>>> +		return 1;
>>> +	else
>>> +		return 0;
>>>  }
>>>  
>>>  static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
>>>  {
>>> +	struct xen_9pfs_front_priv *priv = NULL;
>>> +	RING_IDX cons, prod, masked_cons, masked_prod;
>>> +	unsigned long flags;
>>> +	uint32_t size = p9_req->tc->size;
>>> +	struct xen_9pfs_dataring *ring;
>>> +	int num;
>>> +
>>> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
>>> +		if (priv->client == client)
>>> +			break;
>>> +	}
>>> +	if (priv == NULL || priv->client != client)
>>> +		return -EINVAL;
>>> +
>>> +	num = p9_req->tc->tag % priv->num_rings;
>>> +	ring = &priv->rings[num];
>>> +
>>> +again:
>>> +	while (wait_event_interruptible(ring->wq,
>>> +				p9_xen_write_todo(ring, size) > 0) != 0);
>>> +
>>> +	spin_lock_irqsave(&ring->lock, flags);
>>> +	cons = ring->intf->out_cons;
>>> +	prod = ring->intf->out_prod;
>>> +	mb();
>>> +
>>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {
>>
>> This looks like p9_xen_write_todo().
> p9_xen_write_todo is just a wrapper around xen_9pfs_queued to provide
> a return value that works well with wait_event_interruptible.
>
> I would prefer not to call p9_xen_write_todo here, because it's simpler
> if we don't read prod and cons twice.

I was referring to the whole code fragment after spin_lock_irqsave(),
not just the last line. Isn't it exactly !p9_xen_write_todo()?


>
>
>> BTW, where is xen_9pfs_queued()
>> defined? I couldn't find it. Same for xen_9pfs_mask() and
>> xen_9pfs_write_packet().
> They are provided by the new ring macros, see
> include/xen/interface/io/ring.h (the first patch).

Oh, right. I was searching for the string literally.

-boris
Stefano Stabellini March 8, 2017, 7:33 p.m. UTC | #4
On Wed, 8 Mar 2017, Boris Ostrovsky wrote:
> >>> +}
> >>> +
> >>> +static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
> >>> +{
> >>> +	RING_IDX cons, prod;
> >>> +
> >>> +	cons = ring->intf->out_cons;
> >>> +	prod = ring->intf->out_prod;
> >>> +	mb();
> >>> +
> >>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
> >>> +		return 1;
> >>> +	else
> >>> +		return 0;
> >>>  }
> >>>  
> >>>  static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
> >>>  {
> >>> +	struct xen_9pfs_front_priv *priv = NULL;
> >>> +	RING_IDX cons, prod, masked_cons, masked_prod;
> >>> +	unsigned long flags;
> >>> +	uint32_t size = p9_req->tc->size;
> >>> +	struct xen_9pfs_dataring *ring;
> >>> +	int num;
> >>> +
> >>> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> >>> +		if (priv->client == client)
> >>> +			break;
> >>> +	}
> >>> +	if (priv == NULL || priv->client != client)
> >>> +		return -EINVAL;
> >>> +
> >>> +	num = p9_req->tc->tag % priv->num_rings;
> >>> +	ring = &priv->rings[num];
> >>> +
> >>> +again:
> >>> +	while (wait_event_interruptible(ring->wq,
> >>> +				p9_xen_write_todo(ring, size) > 0) != 0);
> >>> +
> >>> +	spin_lock_irqsave(&ring->lock, flags);
> >>> +	cons = ring->intf->out_cons;
> >>> +	prod = ring->intf->out_prod;
> >>> +	mb();
> >>> +
> >>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {
> >>
> >> This looks like p9_xen_write_todo().
> > p9_xen_write_todo is just a wrapper around xen_9pfs_queued to provide
> > a return value that works well with wait_event_interruptible.
> >
> > I would prefer not to call p9_xen_write_todo here, because it's simpler
> > if we don't read prod and cons twice.
> 
> I was referring to the whole code fragment after spin_lock_irqsave(),
> not just the last line. Isn't it exactly !p9_xen_write_todo()?

Yes, it is true they are almost the same. The difference, and the reason
for p9_xen_write_todo to exist, is that p9_xen_write_todo is called in
the wait_event_interruptible loop, as such it needs to read prod and
cons every time. On the other end, here we want to read them once. Does
it make sense?
Boris Ostrovsky March 8, 2017, 8:02 p.m. UTC | #5
On 03/08/2017 02:33 PM, Stefano Stabellini wrote:
> On Wed, 8 Mar 2017, Boris Ostrovsky wrote:
>>>>> +}
>>>>> +
>>>>> +static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
>>>>> +{
>>>>> +	RING_IDX cons, prod;
>>>>> +
>>>>> +	cons = ring->intf->out_cons;
>>>>> +	prod = ring->intf->out_prod;
>>>>> +	mb();
>>>>> +
>>>>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
>>>>> +		return 1;
>>>>> +	else
>>>>> +		return 0;
>>>>>  }
>>>>>  
>>>>>  static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
>>>>>  {
>>>>> +	struct xen_9pfs_front_priv *priv = NULL;
>>>>> +	RING_IDX cons, prod, masked_cons, masked_prod;
>>>>> +	unsigned long flags;
>>>>> +	uint32_t size = p9_req->tc->size;
>>>>> +	struct xen_9pfs_dataring *ring;
>>>>> +	int num;
>>>>> +
>>>>> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
>>>>> +		if (priv->client == client)
>>>>> +			break;
>>>>> +	}
>>>>> +	if (priv == NULL || priv->client != client)
>>>>> +		return -EINVAL;
>>>>> +
>>>>> +	num = p9_req->tc->tag % priv->num_rings;
>>>>> +	ring = &priv->rings[num];
>>>>> +
>>>>> +again:
>>>>> +	while (wait_event_interruptible(ring->wq,
>>>>> +				p9_xen_write_todo(ring, size) > 0) != 0);
>>>>> +
>>>>> +	spin_lock_irqsave(&ring->lock, flags);
>>>>> +	cons = ring->intf->out_cons;
>>>>> +	prod = ring->intf->out_prod;
>>>>> +	mb();
>>>>> +
>>>>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {
>>>> This looks like p9_xen_write_todo().
>>> p9_xen_write_todo is just a wrapper around xen_9pfs_queued to provide
>>> a return value that works well with wait_event_interruptible.
>>>
>>> I would prefer not to call p9_xen_write_todo here, because it's simpler
>>> if we don't read prod and cons twice.
>> I was referring to the whole code fragment after spin_lock_irqsave(),
>> not just the last line. Isn't it exactly !p9_xen_write_todo()?
> Yes, it is true they are almost the same. The difference, and the reason
> for p9_xen_write_todo to exist, is that p9_xen_write_todo is called in
> the wait_event_interruptible loop, as such it needs to read prod and
> cons every time. On the other end, here we want to read them once. Does
> it make sense?


I am clearly being particularly dense here but what I was thinking was:

again:
	while (wait_event_interruptible(ring->wq,
				p9_xen_write_todo(ring, size) > 0) != 0);

	spin_lock_irqsave(&ring->lock, flags);
	if (!p9_xen_write_todo(ring, size)) {
		spin_unlock_irqrestore(&ring->lock, flags);
		goto again;
	}

There is no extra read of prod/cons.

-boris
Stefano Stabellini March 8, 2017, 8:56 p.m. UTC | #6
On Wed, 8 Mar 2017, Boris Ostrovsky wrote:
> On 03/08/2017 02:33 PM, Stefano Stabellini wrote:
> > On Wed, 8 Mar 2017, Boris Ostrovsky wrote:
> >>>>> +}
> >>>>> +
> >>>>> +static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
> >>>>> +{
> >>>>> +	RING_IDX cons, prod;
> >>>>> +
> >>>>> +	cons = ring->intf->out_cons;
> >>>>> +	prod = ring->intf->out_prod;
> >>>>> +	mb();
> >>>>> +
> >>>>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
> >>>>> +		return 1;
> >>>>> +	else
> >>>>> +		return 0;
> >>>>>  }
> >>>>>  
> >>>>>  static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
> >>>>>  {
> >>>>> +	struct xen_9pfs_front_priv *priv = NULL;
> >>>>> +	RING_IDX cons, prod, masked_cons, masked_prod;
> >>>>> +	unsigned long flags;
> >>>>> +	uint32_t size = p9_req->tc->size;
> >>>>> +	struct xen_9pfs_dataring *ring;
> >>>>> +	int num;
> >>>>> +
> >>>>> +	list_for_each_entry(priv, &xen_9pfs_devs, list) {
> >>>>> +		if (priv->client == client)
> >>>>> +			break;
> >>>>> +	}
> >>>>> +	if (priv == NULL || priv->client != client)
> >>>>> +		return -EINVAL;
> >>>>> +
> >>>>> +	num = p9_req->tc->tag % priv->num_rings;
> >>>>> +	ring = &priv->rings[num];
> >>>>> +
> >>>>> +again:
> >>>>> +	while (wait_event_interruptible(ring->wq,
> >>>>> +				p9_xen_write_todo(ring, size) > 0) != 0);
> >>>>> +
> >>>>> +	spin_lock_irqsave(&ring->lock, flags);
> >>>>> +	cons = ring->intf->out_cons;
> >>>>> +	prod = ring->intf->out_prod;
> >>>>> +	mb();
> >>>>> +
> >>>>> +	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {
> >>>> This looks like p9_xen_write_todo().
> >>> p9_xen_write_todo is just a wrapper around xen_9pfs_queued to provide
> >>> a return value that works well with wait_event_interruptible.
> >>>
> >>> I would prefer not to call p9_xen_write_todo here, because it's simpler
> >>> if we don't read prod and cons twice.
> >> I was referring to the whole code fragment after spin_lock_irqsave(),
> >> not just the last line. Isn't it exactly !p9_xen_write_todo()?
> > Yes, it is true they are almost the same. The difference, and the reason
> > for p9_xen_write_todo to exist, is that p9_xen_write_todo is called in
> > the wait_event_interruptible loop, as such it needs to read prod and
> > cons every time. On the other end, here we want to read them once. Does
> > it make sense?
> 
> 
> I am clearly being particularly dense here but what I was thinking was:
> 
> again:
> 	while (wait_event_interruptible(ring->wq,
> 				p9_xen_write_todo(ring, size) > 0) != 0);
> 
> 	spin_lock_irqsave(&ring->lock, flags);
> 	if (!p9_xen_write_todo(ring, size)) {
> 		spin_unlock_irqrestore(&ring->lock, flags);
> 		goto again;
> 	}
> 
> There is no extra read of prod/cons.

Yes, there are: just after this if statement we would have to read them
again to calculate masked_prod and masked_cons.
Boris Ostrovsky March 8, 2017, 9:01 p.m. UTC | #7
>> There is no extra read of prod/cons.
> Yes, there are: just after this if statement we would have to read them
> again to calculate masked_prod and masked_cons.

Ah, of course. Thanks.

-boris
diff mbox

Patch

diff --git a/net/9p/trans_xen.c b/net/9p/trans_xen.c
index 9f6cf8d..4e26556 100644
--- a/net/9p/trans_xen.c
+++ b/net/9p/trans_xen.c
@@ -47,22 +47,103 @@  struct xen_9pfs_front_priv {
 };
 static LIST_HEAD(xen_9pfs_devs);
 
+/* We don't currently allow canceling of requests */
 static int p9_xen_cancel(struct p9_client *client, struct p9_req_t *req)
 {
-	return 0;
+	return 1;
 }
 
 static int p9_xen_create(struct p9_client *client, const char *addr, char *args)
 {
+	struct xen_9pfs_front_priv *priv = NULL;
+
+	list_for_each_entry(priv, &xen_9pfs_devs, list) {
+		if (!strcmp(priv->tag, addr))
+			break;
+	}
+	if (!priv || strcmp(priv->tag, addr))
+		return -EINVAL;
+
+	priv->client = client; 
 	return 0;
 }
 
 static void p9_xen_close(struct p9_client *client)
 {
+	struct xen_9pfs_front_priv *priv = NULL;
+
+	list_for_each_entry(priv, &xen_9pfs_devs, list) {
+		if (priv->client == client)
+			break;
+	}
+	if (!priv || priv->client != client)
+		return;
+
+	priv->client = NULL; 
+	return;
+}
+
+static int p9_xen_write_todo(struct xen_9pfs_dataring *ring, RING_IDX size)
+{
+	RING_IDX cons, prod;
+
+	cons = ring->intf->out_cons;
+	prod = ring->intf->out_prod;
+	mb();
+
+	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) >= size)
+		return 1;
+	else
+		return 0;
 }
 
 static int p9_xen_request(struct p9_client *client, struct p9_req_t *p9_req)
 {
+	struct xen_9pfs_front_priv *priv = NULL;
+	RING_IDX cons, prod, masked_cons, masked_prod;
+	unsigned long flags;
+	uint32_t size = p9_req->tc->size;
+	struct xen_9pfs_dataring *ring;
+	int num;
+
+	list_for_each_entry(priv, &xen_9pfs_devs, list) {
+		if (priv->client == client)
+			break;
+	}
+	if (priv == NULL || priv->client != client)
+		return -EINVAL;
+
+	num = p9_req->tc->tag % priv->num_rings;
+	ring = &priv->rings[num];
+
+again:
+	while (wait_event_interruptible(ring->wq,
+				p9_xen_write_todo(ring, size) > 0) != 0);
+
+	spin_lock_irqsave(&ring->lock, flags);
+	cons = ring->intf->out_cons;
+	prod = ring->intf->out_prod;
+	mb();
+
+	if (XEN_9PFS_RING_SIZE - xen_9pfs_queued(prod, cons, XEN_9PFS_RING_SIZE) < size) {
+		spin_unlock_irqrestore(&ring->lock, flags);
+		goto again;
+	}
+
+	masked_prod = xen_9pfs_mask(prod, XEN_9PFS_RING_SIZE);
+	masked_cons = xen_9pfs_mask(cons, XEN_9PFS_RING_SIZE);
+
+	xen_9pfs_write_packet(ring->ring.out,
+				&masked_prod, masked_cons,
+				XEN_9PFS_RING_SIZE, p9_req->tc->sdata, size);
+
+	p9_req->status = REQ_STATUS_SENT;
+	wmb();			/* write ring before updating pointer */
+	prod += size;
+	ring->intf->out_prod = prod;
+	spin_unlock_irqrestore(&ring->lock, flags);
+	notify_remote_via_irq(ring->irq);
+
 	return 0;
 }