diff mbox series

[v2,1/2] remoteproc: fall back to using parent memory pool if no dedicated available

Message ID 20200319162321.20632-2-s-anna@ti.com (mailing list archive)
State Superseded
Headers show
Series Misc. rproc fixes around fixed memory region support | expand

Commit Message

Suman Anna March 19, 2020, 4:23 p.m. UTC
From: Tero Kristo <t-kristo@ti.com>

In some cases, like with OMAP remoteproc, we are not creating dedicated
memory pool for the virtio device. Instead, we use the same memory pool
for all shared memories. The current virtio memory pool handling forces
a split between these two, as a separate device is created for it,
causing memory to be allocated from bad location if the dedicated pool
is not available. Fix this by falling back to using the parent device
memory pool if dedicated is not available.

Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
Signed-off-by: Tero Kristo <t-kristo@ti.com>
Signed-off-by: Suman Anna <s-anna@ti.com>
---
v2:
 - Address Arnaud's concerns about hard-coded memory-region index 0
 - Update the comment around the new code addition
v1: https://patchwork.kernel.org/patch/11422721/

 drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
 include/linux/remoteproc.h             |  2 ++
 2 files changed, 17 insertions(+)

Comments

Arnaud POULIQUEN March 20, 2020, 8:39 a.m. UTC | #1
Hi Suman,

On 3/19/20 5:23 PM, Suman Anna wrote:
> From: Tero Kristo <t-kristo@ti.com>
> 
> In some cases, like with OMAP remoteproc, we are not creating dedicated
> memory pool for the virtio device. Instead, we use the same memory pool
> for all shared memories. The current virtio memory pool handling forces
> a split between these two, as a separate device is created for it,
> causing memory to be allocated from bad location if the dedicated pool
> is not available. Fix this by falling back to using the parent device
> memory pool if dedicated is not available.
> 
> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
> Signed-off-by: Tero Kristo <t-kristo@ti.com>
> Signed-off-by: Suman Anna <s-anna@ti.com>
> ---
> v2:
>  - Address Arnaud's concerns about hard-coded memory-region index 0
>  - Update the comment around the new code addition
> v1: https://patchwork.kernel.org/patch/11422721/
> 
>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>  include/linux/remoteproc.h             |  2 ++
>  2 files changed, 17 insertions(+)
> 
> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
> index eb817132bc5f..b687715cdf4b 100644
> --- a/drivers/remoteproc/remoteproc_virtio.c
> +++ b/drivers/remoteproc/remoteproc_virtio.c
> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>  				goto out;
>  			}
>  		}
> +	} else {
> +		struct device_node *np = rproc->dev.parent->of_node;
> +
> +		/*
> +		 * If we don't have dedicated buffer, just attempt to re-assign
> +		 * the reserved memory from our parent. A default memory-region
> +		 * at index 0 from the parent's memory-regions is assigned for
> +		 * the rvdev dev to allocate from, and this can be customized
> +		 * by updating the vdevbuf_mem_id in platform drivers if
> +		 * desired. Failure is non-critical and the allocations will
> +		 * fall back to global pools, so don't check return value
> +		 * either.
> +		 */
> +		of_reserved_mem_device_init_by_idx(dev, np,
> +						   rproc->vdevbuf_mem_id);
>  	}
>  
>  	/* Allocate virtio device */
> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
> index ed127b2d35ca..07bd73a6d72a 100644
> --- a/include/linux/remoteproc.h
> +++ b/include/linux/remoteproc.h
> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>   * @auto_boot: flag to indicate if remote processor should be auto-started
>   * @dump_segments: list of segments in the firmware
>   * @nb_vdev: number of vdev currently handled by rproc
> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>   */
>  struct rproc {
>  	struct list_head node;
> @@ -514,6 +515,7 @@ struct rproc {
>  	bool auto_boot;
>  	struct list_head dump_segments;
>  	int nb_vdev;
> +	u8 vdevbuf_mem_id;
>  	u8 elf_class;
>  };
>

LGTM, just a minor remark concerning the name. As this memory region could be used for
different allocations (e.g rpmsg device allocation, vrings...) i would suggest you to rename
it with a more generic name such as shared_mem_id or default_shmem_id.
Anyway, first let's wait for further feedback on your solution. 

Thanks,
Arnaud
Mathieu Poirier March 25, 2020, 8:38 p.m. UTC | #2
On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
> From: Tero Kristo <t-kristo@ti.com>
> 
> In some cases, like with OMAP remoteproc, we are not creating dedicated
> memory pool for the virtio device. Instead, we use the same memory pool
> for all shared memories. The current virtio memory pool handling forces
> a split between these two, as a separate device is created for it,
> causing memory to be allocated from bad location if the dedicated pool
> is not available. Fix this by falling back to using the parent device
> memory pool if dedicated is not available.
> 
> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
> Signed-off-by: Tero Kristo <t-kristo@ti.com>
> Signed-off-by: Suman Anna <s-anna@ti.com>
> ---
> v2:
>  - Address Arnaud's concerns about hard-coded memory-region index 0
>  - Update the comment around the new code addition
> v1: https://patchwork.kernel.org/patch/11422721/
> 
>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>  include/linux/remoteproc.h             |  2 ++
>  2 files changed, 17 insertions(+)
> 
> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
> index eb817132bc5f..b687715cdf4b 100644
> --- a/drivers/remoteproc/remoteproc_virtio.c
> +++ b/drivers/remoteproc/remoteproc_virtio.c
> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>  				goto out;
>  			}
>  		}
> +	} else {
> +		struct device_node *np = rproc->dev.parent->of_node;
> +
> +		/*
> +		 * If we don't have dedicated buffer, just attempt to re-assign
> +		 * the reserved memory from our parent. A default memory-region
> +		 * at index 0 from the parent's memory-regions is assigned for
> +		 * the rvdev dev to allocate from, and this can be customized
> +		 * by updating the vdevbuf_mem_id in platform drivers if
> +		 * desired. Failure is non-critical and the allocations will
> +		 * fall back to global pools, so don't check return value
> +		 * either.

I'm perplex...  In the changelog it is indicated that if a memory pool is
not dedicated allocation happens from a bad location but here failure of
getting a hold of a dedicated memory pool is not critical. 

> +		 */
> +		of_reserved_mem_device_init_by_idx(dev, np,
> +						   rproc->vdevbuf_mem_id);

I wonder if using an index setup by platform code is really the best way
forward when we already have the carveout mechanic available to us.  I see the
platform code adding a carveout that would have the same name as rproc->name.
From there in rproc_add_virtio_dev() we could have something like:

        mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);


That would be very flexible, the location of the reserved memory withing the
memory-region could change without fear of breaking things and no need to add to
struct rproc.

Let me know what you think.

Thanks,
Mathieu

>  	}
>  
>  	/* Allocate virtio device */
> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
> index ed127b2d35ca..07bd73a6d72a 100644
> --- a/include/linux/remoteproc.h
> +++ b/include/linux/remoteproc.h
> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>   * @auto_boot: flag to indicate if remote processor should be auto-started
>   * @dump_segments: list of segments in the firmware
>   * @nb_vdev: number of vdev currently handled by rproc
> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>   */
>  struct rproc {
>  	struct list_head node;
> @@ -514,6 +515,7 @@ struct rproc {
>  	bool auto_boot;
>  	struct list_head dump_segments;
>  	int nb_vdev;
> +	u8 vdevbuf_mem_id;
>  	u8 elf_class;
>  };
>  
> -- 
> 2.23.0
>
Suman Anna March 25, 2020, 11:39 p.m. UTC | #3
Hi Mathieu,

On 3/25/20 3:38 PM, Mathieu Poirier wrote:
> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>> From: Tero Kristo <t-kristo@ti.com>
>>
>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>> memory pool for the virtio device. Instead, we use the same memory pool
>> for all shared memories. The current virtio memory pool handling forces
>> a split between these two, as a separate device is created for it,
>> causing memory to be allocated from bad location if the dedicated pool
>> is not available. Fix this by falling back to using the parent device
>> memory pool if dedicated is not available.
>>
>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>> Signed-off-by: Suman Anna <s-anna@ti.com>
>> ---
>> v2:
>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>  - Update the comment around the new code addition
>> v1: https://patchwork.kernel.org/patch/11422721/
>>
>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>  include/linux/remoteproc.h             |  2 ++
>>  2 files changed, 17 insertions(+)
>>
>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>> index eb817132bc5f..b687715cdf4b 100644
>> --- a/drivers/remoteproc/remoteproc_virtio.c
>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>  				goto out;
>>  			}
>>  		}
>> +	} else {
>> +		struct device_node *np = rproc->dev.parent->of_node;
>> +
>> +		/*
>> +		 * If we don't have dedicated buffer, just attempt to re-assign
>> +		 * the reserved memory from our parent. A default memory-region
>> +		 * at index 0 from the parent's memory-regions is assigned for
>> +		 * the rvdev dev to allocate from, and this can be customized
>> +		 * by updating the vdevbuf_mem_id in platform drivers if
>> +		 * desired. Failure is non-critical and the allocations will
>> +		 * fall back to global pools, so don't check return value
>> +		 * either.
> 
> I'm perplex...  In the changelog it is indicated that if a memory pool is
> not dedicated allocation happens from a bad location but here failure of
> getting a hold of a dedicated memory pool is not critical.

So, the comment here is a generic one while the bad location part in the
commit description is actually from OMAP remoteproc usage perspective
(if you remember the dev_warn messages we added to the memory-region
parse logic in the driver).

Before the fixed-memory carveout support, all the DMA allocations in
remoteproc core were made from the rproc platform device's DMA pool (
which can be NULL). That is lost after the fixed-memory support, and
they were always allocated from global DMA pools if no dedicated pools
are used. After this patch, that continues to be case for drivers that
still do not use any dedicated pools, while it does restore the usage of
the platform device's DMA pool if a driver uses one (OMAP remoteproc
falls into the latter).

> 
>> +		 */
>> +		of_reserved_mem_device_init_by_idx(dev, np,
>> +						   rproc->vdevbuf_mem_id);
> 
> I wonder if using an index setup by platform code is really the best way
> forward when we already have the carveout mechanic available to us.  I see the
> platform code adding a carveout that would have the same name as rproc->name.
> From there in rproc_add_virtio_dev() we could have something like:
> 
>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
> 
> 
> That would be very flexible, the location of the reserved memory withing the
> memory-region could change without fear of breaking things and no need to add to
> struct rproc.
> 
> Let me know what you think.

I think that can work as well but I feel it is lot more cumbersome. It
does require every platform driver to add code adding/registering that
carveout, and parse the reserved memory region etc. End of the day, we
rely on DMA API and we just have to assign the region to the newly
created device. The DMA pool assignment for devices using
reserved-memory nodes has simply been the of_reserved_mem_device_init()
function.

regards
Suman

> 
> Thanks,
> Mathieu
> 
>>  	}
>>  
>>  	/* Allocate virtio device */
>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>> index ed127b2d35ca..07bd73a6d72a 100644
>> --- a/include/linux/remoteproc.h
>> +++ b/include/linux/remoteproc.h
>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>   * @dump_segments: list of segments in the firmware
>>   * @nb_vdev: number of vdev currently handled by rproc
>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>   */
>>  struct rproc {
>>  	struct list_head node;
>> @@ -514,6 +515,7 @@ struct rproc {
>>  	bool auto_boot;
>>  	struct list_head dump_segments;
>>  	int nb_vdev;
>> +	u8 vdevbuf_mem_id;
>>  	u8 elf_class;
>>  };
>>  
>> -- 
>> 2.23.0
>>
Mathieu Poirier March 27, 2020, 9:09 p.m. UTC | #4
On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>
> Hi Mathieu,
>
> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
> > On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
> >> From: Tero Kristo <t-kristo@ti.com>
> >>
> >> In some cases, like with OMAP remoteproc, we are not creating dedicated
> >> memory pool for the virtio device. Instead, we use the same memory pool
> >> for all shared memories. The current virtio memory pool handling forces
> >> a split between these two, as a separate device is created for it,
> >> causing memory to be allocated from bad location if the dedicated pool
> >> is not available. Fix this by falling back to using the parent device
> >> memory pool if dedicated is not available.
> >>
> >> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
> >> Signed-off-by: Tero Kristo <t-kristo@ti.com>
> >> Signed-off-by: Suman Anna <s-anna@ti.com>
> >> ---
> >> v2:
> >>  - Address Arnaud's concerns about hard-coded memory-region index 0
> >>  - Update the comment around the new code addition
> >> v1: https://patchwork.kernel.org/patch/11422721/
> >>
> >>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
> >>  include/linux/remoteproc.h             |  2 ++
> >>  2 files changed, 17 insertions(+)
> >>
> >> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
> >> index eb817132bc5f..b687715cdf4b 100644
> >> --- a/drivers/remoteproc/remoteproc_virtio.c
> >> +++ b/drivers/remoteproc/remoteproc_virtio.c
> >> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
> >>                              goto out;
> >>                      }
> >>              }
> >> +    } else {
> >> +            struct device_node *np = rproc->dev.parent->of_node;
> >> +
> >> +            /*
> >> +             * If we don't have dedicated buffer, just attempt to re-assign
> >> +             * the reserved memory from our parent. A default memory-region
> >> +             * at index 0 from the parent's memory-regions is assigned for
> >> +             * the rvdev dev to allocate from, and this can be customized
> >> +             * by updating the vdevbuf_mem_id in platform drivers if
> >> +             * desired. Failure is non-critical and the allocations will
> >> +             * fall back to global pools, so don't check return value
> >> +             * either.
> >
> > I'm perplex...  In the changelog it is indicated that if a memory pool is
> > not dedicated allocation happens from a bad location but here failure of
> > getting a hold of a dedicated memory pool is not critical.
>
> So, the comment here is a generic one while the bad location part in the
> commit description is actually from OMAP remoteproc usage perspective
> (if you remember the dev_warn messages we added to the memory-region
> parse logic in the driver).

I can't tell... Are you referring to the comment lines after
of_reserved_mem_device_init() in omap_rproc_probe()?

>
> Before the fixed-memory carveout support, all the DMA allocations in
> remoteproc core were made from the rproc platform device's DMA pool (
> which can be NULL). That is lost after the fixed-memory support, and
> they were always allocated from global DMA pools if no dedicated pools
> are used. After this patch, that continues to be case for drivers that
> still do not use any dedicated pools, while it does restore the usage of
> the platform device's DMA pool if a driver uses one (OMAP remoteproc
> falls into the latter).
>
> >
> >> +             */
> >> +            of_reserved_mem_device_init_by_idx(dev, np,
> >> +                                               rproc->vdevbuf_mem_id);
> >
> > I wonder if using an index setup by platform code is really the best way
> > forward when we already have the carveout mechanic available to us.  I see the
> > platform code adding a carveout that would have the same name as rproc->name.
> > From there in rproc_add_virtio_dev() we could have something like:
> >
> >         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
> >
> >
> > That would be very flexible, the location of the reserved memory withing the
> > memory-region could change without fear of breaking things and no need to add to
> > struct rproc.
> >
> > Let me know what you think.
>
> I think that can work as well but I feel it is lot more cumbersome. It
> does require every platform driver to add code adding/registering that
> carveout, and parse the reserved memory region etc. End of the day, we
> rely on DMA API and we just have to assign the region to the newly
> created device. The DMA pool assignment for devices using
> reserved-memory nodes has simply been the of_reserved_mem_device_init()
> function.

Given all the things happening in the platform drivers adding and
registering a single carveout doesn't seem that onerous to me.   I
also expect setting rproc->vdevbuf_mem_id would involve some form of
parsing.  Lastly if a couple of platforms end up doing the same thing
might as well bring the code in the core, hence choosing a generic
name such as rproc->name for the memory region.

At the very least I would use of_reserved_mem_device_init_by_idx(dev,
np, 0).  I agree it is not flexible but I'll take that over adding a
new field to structure rproc.

Thanks,
Mathieu

>
> regards
> Suman
>
> >
> > Thanks,
> > Mathieu
> >
> >>      }
> >>
> >>      /* Allocate virtio device */
> >> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
> >> index ed127b2d35ca..07bd73a6d72a 100644
> >> --- a/include/linux/remoteproc.h
> >> +++ b/include/linux/remoteproc.h
> >> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
> >>   * @auto_boot: flag to indicate if remote processor should be auto-started
> >>   * @dump_segments: list of segments in the firmware
> >>   * @nb_vdev: number of vdev currently handled by rproc
> >> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
> >>   */
> >>  struct rproc {
> >>      struct list_head node;
> >> @@ -514,6 +515,7 @@ struct rproc {
> >>      bool auto_boot;
> >>      struct list_head dump_segments;
> >>      int nb_vdev;
> >> +    u8 vdevbuf_mem_id;
> >>      u8 elf_class;
> >>  };
> >>
> >> --
> >> 2.23.0
> >>
>
Arnaud POULIQUEN March 30, 2020, 12:29 p.m. UTC | #5
On 3/27/20 10:09 PM, Mathieu Poirier wrote:
> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>
>> Hi Mathieu,
>>
>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>
>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>> for all shared memories. The current virtio memory pool handling forces
>>>> a split between these two, as a separate device is created for it,
>>>> causing memory to be allocated from bad location if the dedicated pool
>>>> is not available. Fix this by falling back to using the parent device
>>>> memory pool if dedicated is not available.
>>>>
>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>> ---
>>>> v2:
>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>  - Update the comment around the new code addition
>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>
>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>  include/linux/remoteproc.h             |  2 ++
>>>>  2 files changed, 17 insertions(+)
>>>>
>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>> index eb817132bc5f..b687715cdf4b 100644
>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>                              goto out;
>>>>                      }
>>>>              }
>>>> +    } else {
>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>> +
>>>> +            /*
>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>> +             * the reserved memory from our parent. A default memory-region
>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>> +             * desired. Failure is non-critical and the allocations will
>>>> +             * fall back to global pools, so don't check return value
>>>> +             * either.
>>>
>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>> not dedicated allocation happens from a bad location but here failure of
>>> getting a hold of a dedicated memory pool is not critical.
>>
>> So, the comment here is a generic one while the bad location part in the
>> commit description is actually from OMAP remoteproc usage perspective
>> (if you remember the dev_warn messages we added to the memory-region
>> parse logic in the driver).
> 
> I can't tell... Are you referring to the comment lines after
> of_reserved_mem_device_init() in omap_rproc_probe()?
> 
>>
>> Before the fixed-memory carveout support, all the DMA allocations in
>> remoteproc core were made from the rproc platform device's DMA pool (
>> which can be NULL). That is lost after the fixed-memory support, and
>> they were always allocated from global DMA pools if no dedicated pools
>> are used. After this patch, that continues to be case for drivers that
>> still do not use any dedicated pools, while it does restore the usage of
>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>> falls into the latter).
>>
>>>
>>>> +             */
>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>> +                                               rproc->vdevbuf_mem_id);
>>>
>>> I wonder if using an index setup by platform code is really the best way
>>> forward when we already have the carveout mechanic available to us.  I see the
>>> platform code adding a carveout that would have the same name as rproc->name.
>>> From there in rproc_add_virtio_dev() we could have something like:
>>>
>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>
>>>
>>> That would be very flexible, the location of the reserved memory withing the
>>> memory-region could change without fear of breaking things and no need to add to
>>> struct rproc.
>>>
>>> Let me know what you think.
>>
>> I think that can work as well but I feel it is lot more cumbersome. It
>> does require every platform driver to add code adding/registering that
>> carveout, and parse the reserved memory region etc. End of the day, we
>> rely on DMA API and we just have to assign the region to the newly
>> created device. The DMA pool assignment for devices using
>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>> function.
> 
> Given all the things happening in the platform drivers adding and
> registering a single carveout doesn't seem that onerous to me.   I
> also expect setting rproc->vdevbuf_mem_id would involve some form of
> parsing.  Lastly if a couple of platforms end up doing the same thing
> might as well bring the code in the core, hence choosing a generic
> name such as rproc->name for the memory region.
> 
> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
> np, 0).  I agree it is not flexible but I'll take that over adding a
> new field to structure rproc.

I wonder whether this would not introduce side effect for some legacy
drivers. Some rproc platforms can have a memory region defined but not
used for the virtio buffers which is allocated in the Linux default
memory pool.

If the aim is to minimize impact in the core part, the solution i proposed
in V1 using rproc_of_resm_mem_entry_init seems a good default candidate too.
The constraint would be that the platform driver has to pre-register
vdev<X>buffer reserved memory associated for a max number of vdev.
This max would limit the number of vdev that a remote firmware can request.
Also not very flexible more that the index 0, as managed at platform level.

Having a default name or a default index seems to me a good compromise...
One advantage of the default name (in this case not rproc->name) is the 
ability to define the memory region in a resource table carveout, instead
of a static definition in DT.

Regards,
Arnaud

> 
> Thanks,
> Mathieu
> 
>>
>> regards
>> Suman
>>
>>>
>>> Thanks,
>>> Mathieu
>>>
>>>>      }
>>>>
>>>>      /* Allocate virtio device */
>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>> --- a/include/linux/remoteproc.h
>>>> +++ b/include/linux/remoteproc.h
>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>   * @dump_segments: list of segments in the firmware
>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>   */
>>>>  struct rproc {
>>>>      struct list_head node;
>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>      bool auto_boot;
>>>>      struct list_head dump_segments;
>>>>      int nb_vdev;
>>>> +    u8 vdevbuf_mem_id;
>>>>      u8 elf_class;
>>>>  };
>>>>
>>>> --
>>>> 2.23.0
>>>>
>>
Suman Anna April 7, 2020, 7:36 p.m. UTC | #6
Hi Mathieu,

On 3/27/20 4:09 PM, Mathieu Poirier wrote:
> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>
>> Hi Mathieu,
>>
>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>
>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>> for all shared memories. The current virtio memory pool handling forces
>>>> a split between these two, as a separate device is created for it,
>>>> causing memory to be allocated from bad location if the dedicated pool
>>>> is not available. Fix this by falling back to using the parent device
>>>> memory pool if dedicated is not available.
>>>>
>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>> ---
>>>> v2:
>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>  - Update the comment around the new code addition
>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>
>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>  include/linux/remoteproc.h             |  2 ++
>>>>  2 files changed, 17 insertions(+)
>>>>
>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>> index eb817132bc5f..b687715cdf4b 100644
>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>                              goto out;
>>>>                      }
>>>>              }
>>>> +    } else {
>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>> +
>>>> +            /*
>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>> +             * the reserved memory from our parent. A default memory-region
>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>> +             * desired. Failure is non-critical and the allocations will
>>>> +             * fall back to global pools, so don't check return value
>>>> +             * either.
>>>
>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>> not dedicated allocation happens from a bad location but here failure of
>>> getting a hold of a dedicated memory pool is not critical.
>>
>> So, the comment here is a generic one while the bad location part in the
>> commit description is actually from OMAP remoteproc usage perspective
>> (if you remember the dev_warn messages we added to the memory-region
>> parse logic in the driver).
> 
> I can't tell... Are you referring to the comment lines after
> of_reserved_mem_device_init() in omap_rproc_probe()?

Yes indeed, the dev_warn traces after of_reserved_mem_device_init().

> 
>>
>> Before the fixed-memory carveout support, all the DMA allocations in
>> remoteproc core were made from the rproc platform device's DMA pool (
>> which can be NULL). That is lost after the fixed-memory support, and
>> they were always allocated from global DMA pools if no dedicated pools
>> are used. After this patch, that continues to be case for drivers that
>> still do not use any dedicated pools, while it does restore the usage of
>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>> falls into the latter).
>>
>>>
>>>> +             */
>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>> +                                               rproc->vdevbuf_mem_id);
>>>
>>> I wonder if using an index setup by platform code is really the best way
>>> forward when we already have the carveout mechanic available to us.  I see the
>>> platform code adding a carveout that would have the same name as rproc->name.
>>> From there in rproc_add_virtio_dev() we could have something like:
>>>
>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>
>>>
>>> That would be very flexible, the location of the reserved memory withing the
>>> memory-region could change without fear of breaking things and no need to add to
>>> struct rproc.
>>>
>>> Let me know what you think.
>>
>> I think that can work as well but I feel it is lot more cumbersome. It
>> does require every platform driver to add code adding/registering that
>> carveout, and parse the reserved memory region etc. End of the day, we
>> rely on DMA API and we just have to assign the region to the newly
>> created device. The DMA pool assignment for devices using
>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>> function.
> 
> Given all the things happening in the platform drivers adding and
> registering a single carveout doesn't seem that onerous to me.   I
> also expect setting rproc->vdevbuf_mem_id would involve some form of
> parsing.  

So, no additional parsing other than to know which id if you have
multiple regions. A device can only be assigned one default DMA/CMA pool
to use with the DMA API. One would need to add the assignment statement
only if region 0 is not being used as the device DMA-API pool.

Lastly if a couple of platforms end up doing the same thing
> might as well bring the code in the core, hence choosing a generic
> name such as rproc->name for the memory region.

That is actually lot more code than the current code. First you would
need to lookup and parse the reserved mem to get the address and size to
initialize the rproc mem structure, and then use the filled in values to
declare the DMA pool.

> 
> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
> np, 0).  I agree it is not flexible but I'll take that over adding a
> new field to structure rproc.

Yep, I started out indeed with exactly that code in v1, and only
introduced the new field to address Arnaud's comments. Even the new
field is intrinsically initialized to 0, so the code is equivalent, and
supports the cases in case you need to use a different reserved-memory
region than at index 0.

regards
Suman


> 
> Thanks,
> Mathieu
> 
>>
>> regards
>> Suman
>>
>>>
>>> Thanks,
>>> Mathieu
>>>
>>>>      }
>>>>
>>>>      /* Allocate virtio device */
>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>> --- a/include/linux/remoteproc.h
>>>> +++ b/include/linux/remoteproc.h
>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>   * @dump_segments: list of segments in the firmware
>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>   */
>>>>  struct rproc {
>>>>      struct list_head node;
>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>      bool auto_boot;
>>>>      struct list_head dump_segments;
>>>>      int nb_vdev;
>>>> +    u8 vdevbuf_mem_id;
>>>>      u8 elf_class;
>>>>  };
>>>>
>>>> --
>>>> 2.23.0
>>>>
>>
Suman Anna April 7, 2020, 7:47 p.m. UTC | #7
Hi Arnaud,

On 3/30/20 7:29 AM, Arnaud POULIQUEN wrote:
> 
> 
> On 3/27/20 10:09 PM, Mathieu Poirier wrote:
>> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>>
>>> Hi Mathieu,
>>>
>>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>>
>>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>>> for all shared memories. The current virtio memory pool handling forces
>>>>> a split between these two, as a separate device is created for it,
>>>>> causing memory to be allocated from bad location if the dedicated pool
>>>>> is not available. Fix this by falling back to using the parent device
>>>>> memory pool if dedicated is not available.
>>>>>
>>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>>> ---
>>>>> v2:
>>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>>  - Update the comment around the new code addition
>>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>>
>>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>>  include/linux/remoteproc.h             |  2 ++
>>>>>  2 files changed, 17 insertions(+)
>>>>>
>>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>>> index eb817132bc5f..b687715cdf4b 100644
>>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>>                              goto out;
>>>>>                      }
>>>>>              }
>>>>> +    } else {
>>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>>> +
>>>>> +            /*
>>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>>> +             * the reserved memory from our parent. A default memory-region
>>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>>> +             * desired. Failure is non-critical and the allocations will
>>>>> +             * fall back to global pools, so don't check return value
>>>>> +             * either.
>>>>
>>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>>> not dedicated allocation happens from a bad location but here failure of
>>>> getting a hold of a dedicated memory pool is not critical.
>>>
>>> So, the comment here is a generic one while the bad location part in the
>>> commit description is actually from OMAP remoteproc usage perspective
>>> (if you remember the dev_warn messages we added to the memory-region
>>> parse logic in the driver).
>>
>> I can't tell... Are you referring to the comment lines after
>> of_reserved_mem_device_init() in omap_rproc_probe()?
>>
>>>
>>> Before the fixed-memory carveout support, all the DMA allocations in
>>> remoteproc core were made from the rproc platform device's DMA pool (
>>> which can be NULL). That is lost after the fixed-memory support, and
>>> they were always allocated from global DMA pools if no dedicated pools
>>> are used. After this patch, that continues to be case for drivers that
>>> still do not use any dedicated pools, while it does restore the usage of
>>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>>> falls into the latter).
>>>
>>>>
>>>>> +             */
>>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>>> +                                               rproc->vdevbuf_mem_id);
>>>>
>>>> I wonder if using an index setup by platform code is really the best way
>>>> forward when we already have the carveout mechanic available to us.  I see the
>>>> platform code adding a carveout that would have the same name as rproc->name.
>>>> From there in rproc_add_virtio_dev() we could have something like:
>>>>
>>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>>
>>>>
>>>> That would be very flexible, the location of the reserved memory withing the
>>>> memory-region could change without fear of breaking things and no need to add to
>>>> struct rproc.
>>>>
>>>> Let me know what you think.
>>>
>>> I think that can work as well but I feel it is lot more cumbersome. It
>>> does require every platform driver to add code adding/registering that
>>> carveout, and parse the reserved memory region etc. End of the day, we
>>> rely on DMA API and we just have to assign the region to the newly
>>> created device. The DMA pool assignment for devices using
>>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>>> function.
>>
>> Given all the things happening in the platform drivers adding and
>> registering a single carveout doesn't seem that onerous to me.   I
>> also expect setting rproc->vdevbuf_mem_id would involve some form of
>> parsing.  Lastly if a couple of platforms end up doing the same thing
>> might as well bring the code in the core, hence choosing a generic
>> name such as rproc->name for the memory region.
>>
>> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
>> np, 0).  I agree it is not flexible but I'll take that over adding a
>> new field to structure rproc.
> 
> I wonder whether this would not introduce side effect for some legacy
> drivers. Some rproc platforms can have a memory region defined but not
> used for the virtio buffers which is allocated in the Linux default
> memory pool.

I am actually trying to restore the behavior before the separate vdev
device creation. All allocations were coming from the rproc platform
device's DMA pool (if assigned, otherwise will fall back to global pool).

You have a valid point though, but are there any drivers relying on this
today? All the usage I saw were only using of_reserved_mem_device_init()
which is 0-indexed region by default. If this is really an issue, we can
use -1 as a means to conditionally follow this path.

> If the aim is to minimize impact in the core part, the solution i proposed
> in V1 using rproc_of_resm_mem_entry_init seems a good default candidate too.
> The constraint would be that the platform driver has to pre-register
> vdev<X>buffer reserved memory associated for a max number of vdev.
> This max would limit the number of vdev that a remote firmware can request.
> Also not very flexible more that the index 0, as managed at platform level.
> 
> Having a default name or a default index seems to me a good compromise...
> One advantage of the default name (in this case not rproc->name) is the 
> ability to define the memory region in a resource table carveout, instead
> of a static definition in DT.

Hmm, regarding the default name, vdev rings and vdev buffers are already
looking for specific names, right? So, what resource type are you
envisioning for this - RSC_CARVEOUT? Wouldn't that become a separate
allocation by itself?

regards
Suman

>>>>
>>>>>      }
>>>>>
>>>>>      /* Allocate virtio device */
>>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>>> --- a/include/linux/remoteproc.h
>>>>> +++ b/include/linux/remoteproc.h
>>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>>   * @dump_segments: list of segments in the firmware
>>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>>   */
>>>>>  struct rproc {
>>>>>      struct list_head node;
>>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>>      bool auto_boot;
>>>>>      struct list_head dump_segments;
>>>>>      int nb_vdev;
>>>>> +    u8 vdevbuf_mem_id;
>>>>>      u8 elf_class;
>>>>>  };
>>>>>
>>>>> --
>>>>> 2.23.0
>>>>>
>>>
Arnaud POULIQUEN April 8, 2020, 9:42 a.m. UTC | #8
Hi Suman,

On 4/7/20 9:47 PM, Suman Anna wrote:
> Hi Arnaud,
> 
> On 3/30/20 7:29 AM, Arnaud POULIQUEN wrote:
>>
>>
>> On 3/27/20 10:09 PM, Mathieu Poirier wrote:
>>> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>>>
>>>> Hi Mathieu,
>>>>
>>>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>>>
>>>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>>>> for all shared memories. The current virtio memory pool handling forces
>>>>>> a split between these two, as a separate device is created for it,
>>>>>> causing memory to be allocated from bad location if the dedicated pool
>>>>>> is not available. Fix this by falling back to using the parent device
>>>>>> memory pool if dedicated is not available.
>>>>>>
>>>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>>>> ---
>>>>>> v2:
>>>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>>>  - Update the comment around the new code addition
>>>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>>>
>>>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>>>  include/linux/remoteproc.h             |  2 ++
>>>>>>  2 files changed, 17 insertions(+)
>>>>>>
>>>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>>>> index eb817132bc5f..b687715cdf4b 100644
>>>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>>>                              goto out;
>>>>>>                      }
>>>>>>              }
>>>>>> +    } else {
>>>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>>>> +
>>>>>> +            /*
>>>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>>>> +             * the reserved memory from our parent. A default memory-region
>>>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>>>> +             * desired. Failure is non-critical and the allocations will
>>>>>> +             * fall back to global pools, so don't check return value
>>>>>> +             * either.
>>>>>
>>>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>>>> not dedicated allocation happens from a bad location but here failure of
>>>>> getting a hold of a dedicated memory pool is not critical.
>>>>
>>>> So, the comment here is a generic one while the bad location part in the
>>>> commit description is actually from OMAP remoteproc usage perspective
>>>> (if you remember the dev_warn messages we added to the memory-region
>>>> parse logic in the driver).
>>>
>>> I can't tell... Are you referring to the comment lines after
>>> of_reserved_mem_device_init() in omap_rproc_probe()?
>>>
>>>>
>>>> Before the fixed-memory carveout support, all the DMA allocations in
>>>> remoteproc core were made from the rproc platform device's DMA pool (
>>>> which can be NULL). That is lost after the fixed-memory support, and
>>>> they were always allocated from global DMA pools if no dedicated pools
>>>> are used. After this patch, that continues to be case for drivers that
>>>> still do not use any dedicated pools, while it does restore the usage of
>>>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>>>> falls into the latter).
>>>>
>>>>>
>>>>>> +             */
>>>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>>>> +                                               rproc->vdevbuf_mem_id);
>>>>>
>>>>> I wonder if using an index setup by platform code is really the best way
>>>>> forward when we already have the carveout mechanic available to us.  I see the
>>>>> platform code adding a carveout that would have the same name as rproc->name.
>>>>> From there in rproc_add_virtio_dev() we could have something like:
>>>>>
>>>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>>>
>>>>>
>>>>> That would be very flexible, the location of the reserved memory withing the
>>>>> memory-region could change without fear of breaking things and no need to add to
>>>>> struct rproc.
>>>>>
>>>>> Let me know what you think.
>>>>
>>>> I think that can work as well but I feel it is lot more cumbersome. It
>>>> does require every platform driver to add code adding/registering that
>>>> carveout, and parse the reserved memory region etc. End of the day, we
>>>> rely on DMA API and we just have to assign the region to the newly
>>>> created device. The DMA pool assignment for devices using
>>>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>>>> function.
>>>
>>> Given all the things happening in the platform drivers adding and
>>> registering a single carveout doesn't seem that onerous to me.   I
>>> also expect setting rproc->vdevbuf_mem_id would involve some form of
>>> parsing.  Lastly if a couple of platforms end up doing the same thing
>>> might as well bring the code in the core, hence choosing a generic
>>> name such as rproc->name for the memory region.
>>>
>>> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
>>> np, 0).  I agree it is not flexible but I'll take that over adding a
>>> new field to structure rproc.
>>
>> I wonder whether this would not introduce side effect for some legacy
>> drivers. Some rproc platforms can have a memory region defined but not
>> used for the virtio buffers which is allocated in the Linux default
>> memory pool.
> 
> I am actually trying to restore the behavior before the separate vdev
> device creation. All allocations were coming from the rproc platform
> device's DMA pool (if assigned, otherwise will fall back to global pool).

Yes deprecated dma_declare_coherent_memory was used to declare the rproc
platform pool for vdev if i well remember).

> 
> You have a valid point though, but are there any drivers relying on this
> today? All the usage I saw were only using of_reserved_mem_device_init()
> which is 0-indexed region by default. If this is really an issue, we can
> use -1 as a means to conditionally follow this path.

Up-streamed driver seems using nothing but address 0 for time being. 
I don't know for non-up-streamed drivers...
But if you have a look to the stm32 platform driver, we also define the
memory regions for the code. 
Even if we are also interesting by the possibility of having a default pool,
it is no simple to implement for our driver using index 0.
We will have to declare it first in DT. But we would also have to determine
if the first memory pool declaration is a default pool or not vs the other
memory region declarations.

> 
>> If the aim is to minimize impact in the core part, the solution i proposed
>> in V1 using rproc_of_resm_mem_entry_init seems a good default candidate too.
>> The constraint would be that the platform driver has to pre-register
>> vdev<X>buffer reserved memory associated for a max number of vdev.
>> This max would limit the number of vdev that a remote firmware can request.
>> Also not very flexible more that the index 0, as managed at platform level.
>>
>> Having a default name or a default index seems to me a good compromise...
>> One advantage of the default name (in this case not rproc->name) is the 
>> ability to define the memory region in a resource table carveout, instead
>> of a static definition in DT.
> 
> Hmm, regarding the default name, vdev rings and vdev buffers are already
> looking for specific names, right? So, what resource type are you
> envisioning for this - RSC_CARVEOUT? Wouldn't that become a separate
> allocation by itself?

After checking, I think my point on the carveout is not valid. As we should no
longer use dma_declare_coherent_memory which seems obsolete since the
introduction of the devicetree. Thus, even if a carveout is defined, we will
have to declare an associated memory region in DT.

Anyway as a name is used today for other memory regions, i would be in flavor
of using the same method to declare a default memory pool in remote proc.
relying on the rproc_of_resm_mem_entry_init function.

If an index is preferred in core, platform drivers that declare more that
one memory region will probably have to implement code based on a default
memory region name to retrieve the associated index.

Regards,
Arnaud 


> 
> regards
> Suman
> 
>>>>>
>>>>>>      }
>>>>>>
>>>>>>      /* Allocate virtio device */
>>>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>>>> --- a/include/linux/remoteproc.h
>>>>>> +++ b/include/linux/remoteproc.h
>>>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>>>   * @dump_segments: list of segments in the firmware
>>>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>>>   */
>>>>>>  struct rproc {
>>>>>>      struct list_head node;
>>>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>>>      bool auto_boot;
>>>>>>      struct list_head dump_segments;
>>>>>>      int nb_vdev;
>>>>>> +    u8 vdevbuf_mem_id;
>>>>>>      u8 elf_class;
>>>>>>  };
>>>>>>
>>>>>> --
>>>>>> 2.23.0
>>>>>>
>>>>
>
Suman Anna April 8, 2020, 11:36 p.m. UTC | #9
Hi Arnaud,

>>
>> On 3/30/20 7:29 AM, Arnaud POULIQUEN wrote:
>>>
>>>
>>> On 3/27/20 10:09 PM, Mathieu Poirier wrote:
>>>> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>>>>
>>>>> Hi Mathieu,
>>>>>
>>>>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>>>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>>>>
>>>>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>>>>> for all shared memories. The current virtio memory pool handling forces
>>>>>>> a split between these two, as a separate device is created for it,
>>>>>>> causing memory to be allocated from bad location if the dedicated pool
>>>>>>> is not available. Fix this by falling back to using the parent device
>>>>>>> memory pool if dedicated is not available.
>>>>>>>
>>>>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>>>>> ---
>>>>>>> v2:
>>>>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>>>>  - Update the comment around the new code addition
>>>>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>>>>
>>>>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>>>>  include/linux/remoteproc.h             |  2 ++
>>>>>>>  2 files changed, 17 insertions(+)
>>>>>>>
>>>>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>>>>> index eb817132bc5f..b687715cdf4b 100644
>>>>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>>>>                              goto out;
>>>>>>>                      }
>>>>>>>              }
>>>>>>> +    } else {
>>>>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>>>>> +
>>>>>>> +            /*
>>>>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>>>>> +             * the reserved memory from our parent. A default memory-region
>>>>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>>>>> +             * desired. Failure is non-critical and the allocations will
>>>>>>> +             * fall back to global pools, so don't check return value
>>>>>>> +             * either.
>>>>>>
>>>>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>>>>> not dedicated allocation happens from a bad location but here failure of
>>>>>> getting a hold of a dedicated memory pool is not critical.
>>>>>
>>>>> So, the comment here is a generic one while the bad location part in the
>>>>> commit description is actually from OMAP remoteproc usage perspective
>>>>> (if you remember the dev_warn messages we added to the memory-region
>>>>> parse logic in the driver).
>>>>
>>>> I can't tell... Are you referring to the comment lines after
>>>> of_reserved_mem_device_init() in omap_rproc_probe()?
>>>>
>>>>>
>>>>> Before the fixed-memory carveout support, all the DMA allocations in
>>>>> remoteproc core were made from the rproc platform device's DMA pool (
>>>>> which can be NULL). That is lost after the fixed-memory support, and
>>>>> they were always allocated from global DMA pools if no dedicated pools
>>>>> are used. After this patch, that continues to be case for drivers that
>>>>> still do not use any dedicated pools, while it does restore the usage of
>>>>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>>>>> falls into the latter).
>>>>>
>>>>>>
>>>>>>> +             */
>>>>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>>>>> +                                               rproc->vdevbuf_mem_id);
>>>>>>
>>>>>> I wonder if using an index setup by platform code is really the best way
>>>>>> forward when we already have the carveout mechanic available to us.  I see the
>>>>>> platform code adding a carveout that would have the same name as rproc->name.
>>>>>> From there in rproc_add_virtio_dev() we could have something like:
>>>>>>
>>>>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>>>>
>>>>>>
>>>>>> That would be very flexible, the location of the reserved memory withing the
>>>>>> memory-region could change without fear of breaking things and no need to add to
>>>>>> struct rproc.
>>>>>>
>>>>>> Let me know what you think.
>>>>>
>>>>> I think that can work as well but I feel it is lot more cumbersome. It
>>>>> does require every platform driver to add code adding/registering that
>>>>> carveout, and parse the reserved memory region etc. End of the day, we
>>>>> rely on DMA API and we just have to assign the region to the newly
>>>>> created device. The DMA pool assignment for devices using
>>>>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>>>>> function.
>>>>
>>>> Given all the things happening in the platform drivers adding and
>>>> registering a single carveout doesn't seem that onerous to me.   I
>>>> also expect setting rproc->vdevbuf_mem_id would involve some form of
>>>> parsing.  Lastly if a couple of platforms end up doing the same thing
>>>> might as well bring the code in the core, hence choosing a generic
>>>> name such as rproc->name for the memory region.
>>>>
>>>> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
>>>> np, 0).  I agree it is not flexible but I'll take that over adding a
>>>> new field to structure rproc.
>>>
>>> I wonder whether this would not introduce side effect for some legacy
>>> drivers. Some rproc platforms can have a memory region defined but not
>>> used for the virtio buffers which is allocated in the Linux default
>>> memory pool.
>>
>> I am actually trying to restore the behavior before the separate vdev
>> device creation. All allocations were coming from the rproc platform
>> device's DMA pool (if assigned, otherwise will fall back to global pool).
> 
> Yes deprecated dma_declare_coherent_memory was used to declare the rproc
> platform pool for vdev if i well remember).

Right, the dma_declare_coherent_memory() usage was introduced only in
v5.1 as part of vdev subdevice creation. Prior to that, the allocations
used the rproc dev DMA pool which would have been initialized in the
platform drivers.

> 
>>
>> You have a valid point though, but are there any drivers relying on this
>> today? All the usage I saw were only using of_reserved_mem_device_init()
>> which is 0-indexed region by default. If this is really an issue, we can
>> use -1 as a means to conditionally follow this path.
> 
> Up-streamed driver seems using nothing but address 0 for time being. 
> I don't know for non-up-streamed drivers...
> But if you have a look to the stm32 platform driver, we also define the
> memory regions for the code. 

ST drivers are already using the "vdev0buffer" on latest kernels, so
they will not be entering the fallback path anyway right. Looking at
4.19 kernel, I see that you guys were also using a single
reserved-memory region (didn't have DT nodes though), so all allocations
would have come from that region. The patch is for fixing the
allocations for drivers that were defining a DMA rproc pool, and still
using them (so not having to force the drivers to add code to fix the
issue).

> Even if we are also interesting by the possibility of having a default pool,
> it is no simple to implement for our driver using index 0.
> We will have to declare it first in DT. But we would also have to determine
> if the first memory pool declaration is a default pool or not vs the other
> memory region declarations.

Yeah, if we used the hard-coded 0 and you were adapting to use a default
pool. Otherwise, you could technically update or define the new generic
name (which has to be added to remoteproc core as well).

> 
>>
>>> If the aim is to minimize impact in the core part, the solution i proposed
>>> in V1 using rproc_of_resm_mem_entry_init seems a good default candidate too.
>>> The constraint would be that the platform driver has to pre-register
>>> vdev<X>buffer reserved memory associated for a max number of vdev.
>>> This max would limit the number of vdev that a remote firmware can request.
>>> Also not very flexible more that the index 0, as managed at platform level.
>>>
>>> Having a default name or a default index seems to me a good compromise...
>>> One advantage of the default name (in this case not rproc->name) is the 
>>> ability to define the memory region in a resource table carveout, instead
>>> of a static definition in DT.
>>
>> Hmm, regarding the default name, vdev rings and vdev buffers are already
>> looking for specific names, right? So, what resource type are you
>> envisioning for this - RSC_CARVEOUT? Wouldn't that become a separate
>> allocation by itself?
> 
> After checking, I think my point on the carveout is not valid. As we should no
> longer use dma_declare_coherent_memory which seems obsolete since the
> introduction of the devicetree. Thus, even if a carveout is defined, we will
> have to declare an associated memory region in DT.
> 
> Anyway as a name is used today for other memory regions, i would be in flavor
> of using the same method to declare a default memory pool in remote proc.
> relying on the rproc_of_resm_mem_entry_init function.

This could definitely be an enhancement and conversion of existing
drivers, but at this point, the idea is to restore the allocations to
how they were for drivers defining a DMA pool and without having to
modify them.

> If an index is preferred in core, platform drivers that declare more that
> one memory region will probably have to implement code based on a default
> memory region name to retrieve the associated index.

So, if we were to get rid of the dma_declare_coherent_memory() usage
from remoteproc core, then I actually don't see need for the other args
in rproc_of_resm_mem_entry_init(), just the name and id. And even that
you would only need if you are going to have to deal with multiple
devices to assign the DMA pool for, for which you would need the name.

regards
Suman

> 
> Regards,
> Arnaud 
> 
> 
>>
>> regards
>> Suman
>>
>>>>>>
>>>>>>>      }
>>>>>>>
>>>>>>>      /* Allocate virtio device */
>>>>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>>>>> --- a/include/linux/remoteproc.h
>>>>>>> +++ b/include/linux/remoteproc.h
>>>>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>>>>   * @dump_segments: list of segments in the firmware
>>>>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>>>>   */
>>>>>>>  struct rproc {
>>>>>>>      struct list_head node;
>>>>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>>>>      bool auto_boot;
>>>>>>>      struct list_head dump_segments;
>>>>>>>      int nb_vdev;
>>>>>>> +    u8 vdevbuf_mem_id;
>>>>>>>      u8 elf_class;
>>>>>>>  };
>>>>>>>
>>>>>>> --
>>>>>>> 2.23.0
>>>>>>>
>>>>>
>>
Arnaud POULIQUEN April 9, 2020, 9:58 a.m. UTC | #10
On 4/9/20 1:36 AM, Suman Anna wrote:
> Hi Arnaud,
> 
>>>
>>> On 3/30/20 7:29 AM, Arnaud POULIQUEN wrote:
>>>>
>>>>
>>>> On 3/27/20 10:09 PM, Mathieu Poirier wrote:
>>>>> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>>>>>
>>>>>> Hi Mathieu,
>>>>>>
>>>>>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>>>>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>>>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>>>>>
>>>>>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>>>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>>>>>> for all shared memories. The current virtio memory pool handling forces
>>>>>>>> a split between these two, as a separate device is created for it,
>>>>>>>> causing memory to be allocated from bad location if the dedicated pool
>>>>>>>> is not available. Fix this by falling back to using the parent device
>>>>>>>> memory pool if dedicated is not available.
>>>>>>>>
>>>>>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>>>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>>>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>>>>>> ---
>>>>>>>> v2:
>>>>>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>>>>>  - Update the comment around the new code addition
>>>>>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>>>>>
>>>>>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>>>>>  include/linux/remoteproc.h             |  2 ++
>>>>>>>>  2 files changed, 17 insertions(+)
>>>>>>>>
>>>>>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>>>>>> index eb817132bc5f..b687715cdf4b 100644
>>>>>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>>>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>>>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>>>>>                              goto out;
>>>>>>>>                      }
>>>>>>>>              }
>>>>>>>> +    } else {
>>>>>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>>>>>> +
>>>>>>>> +            /*
>>>>>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>>>>>> +             * the reserved memory from our parent. A default memory-region
>>>>>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>>>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>>>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>>>>>> +             * desired. Failure is non-critical and the allocations will
>>>>>>>> +             * fall back to global pools, so don't check return value
>>>>>>>> +             * either.
>>>>>>>
>>>>>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>>>>>> not dedicated allocation happens from a bad location but here failure of
>>>>>>> getting a hold of a dedicated memory pool is not critical.
>>>>>>
>>>>>> So, the comment here is a generic one while the bad location part in the
>>>>>> commit description is actually from OMAP remoteproc usage perspective
>>>>>> (if you remember the dev_warn messages we added to the memory-region
>>>>>> parse logic in the driver).
>>>>>
>>>>> I can't tell... Are you referring to the comment lines after
>>>>> of_reserved_mem_device_init() in omap_rproc_probe()?
>>>>>
>>>>>>
>>>>>> Before the fixed-memory carveout support, all the DMA allocations in
>>>>>> remoteproc core were made from the rproc platform device's DMA pool (
>>>>>> which can be NULL). That is lost after the fixed-memory support, and
>>>>>> they were always allocated from global DMA pools if no dedicated pools
>>>>>> are used. After this patch, that continues to be case for drivers that
>>>>>> still do not use any dedicated pools, while it does restore the usage of
>>>>>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>>>>>> falls into the latter).
>>>>>>
>>>>>>>
>>>>>>>> +             */
>>>>>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>>>>>> +                                               rproc->vdevbuf_mem_id);
>>>>>>>
>>>>>>> I wonder if using an index setup by platform code is really the best way
>>>>>>> forward when we already have the carveout mechanic available to us.  I see the
>>>>>>> platform code adding a carveout that would have the same name as rproc->name.
>>>>>>> From there in rproc_add_virtio_dev() we could have something like:
>>>>>>>
>>>>>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>>>>>
>>>>>>>
>>>>>>> That would be very flexible, the location of the reserved memory withing the
>>>>>>> memory-region could change without fear of breaking things and no need to add to
>>>>>>> struct rproc.
>>>>>>>
>>>>>>> Let me know what you think.
>>>>>>
>>>>>> I think that can work as well but I feel it is lot more cumbersome. It
>>>>>> does require every platform driver to add code adding/registering that
>>>>>> carveout, and parse the reserved memory region etc. End of the day, we
>>>>>> rely on DMA API and we just have to assign the region to the newly
>>>>>> created device. The DMA pool assignment for devices using
>>>>>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>>>>>> function.
>>>>>
>>>>> Given all the things happening in the platform drivers adding and
>>>>> registering a single carveout doesn't seem that onerous to me.   I
>>>>> also expect setting rproc->vdevbuf_mem_id would involve some form of
>>>>> parsing.  Lastly if a couple of platforms end up doing the same thing
>>>>> might as well bring the code in the core, hence choosing a generic
>>>>> name such as rproc->name for the memory region.
>>>>>
>>>>> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
>>>>> np, 0).  I agree it is not flexible but I'll take that over adding a
>>>>> new field to structure rproc.
>>>>
>>>> I wonder whether this would not introduce side effect for some legacy
>>>> drivers. Some rproc platforms can have a memory region defined but not
>>>> used for the virtio buffers which is allocated in the Linux default
>>>> memory pool.
>>>
>>> I am actually trying to restore the behavior before the separate vdev
>>> device creation. All allocations were coming from the rproc platform
>>> device's DMA pool (if assigned, otherwise will fall back to global pool).
>>
>> Yes deprecated dma_declare_coherent_memory was used to declare the rproc
>> platform pool for vdev if i well remember).
> 
> Right, the dma_declare_coherent_memory() usage was introduced only in
> v5.1 as part of vdev subdevice creation. Prior to that, the allocations
> used the rproc dev DMA pool which would have been initialized in the
> platform drivers.
> 
>>
>>>
>>> You have a valid point though, but are there any drivers relying on this
>>> today? All the usage I saw were only using of_reserved_mem_device_init()
>>> which is 0-indexed region by default. If this is really an issue, we can
>>> use -1 as a means to conditionally follow this path.
>>
>> Up-streamed driver seems using nothing but address 0 for time being. 
>> I don't know for non-up-streamed drivers...
>> But if you have a look to the stm32 platform driver, we also define the
>> memory regions for the code. 
> 
> ST drivers are already using the "vdev0buffer" on latest kernels, so
> they will not be entering the fallback path anyway right. Looking at
> 4.19 kernel, I see that you guys were also using a single
> reserved-memory region (didn't have DT nodes though), so all allocations
> would have come from that region. The patch is for fixing the
> allocations for drivers that were defining a DMA rproc pool, and still
> using them (so not having to force the drivers to add code to fix the
> issue).

Right, but today the limitation with our implementation is the support of
additional shared buffers(big data). The possibility to add a single DMA pool
would help to solve this issue. Furthermore we have a DT generator tools that
should support the shared memory regions tuning. Having an unique shared memory
region could simplify this tool usage.
For these reasons i would be interested in finding a more scalable solution.  

> 
>> Even if we are also interesting by the possibility of having a default pool,
>> it is no simple to implement for our driver using index 0.
>> We will have to declare it first in DT. But we would also have to determine
>> if the first memory pool declaration is a default pool or not vs the other
>> memory region declarations.
> 
> Yeah, if we used the hard-coded 0 and you were adapting to use a default
> pool. Otherwise, you could technically update or define the new generic
> name (which has to be added to remoteproc core as well).
> 
>>
>>>
>>>> If the aim is to minimize impact in the core part, the solution i proposed
>>>> in V1 using rproc_of_resm_mem_entry_init seems a good default candidate too.
>>>> The constraint would be that the platform driver has to pre-register
>>>> vdev<X>buffer reserved memory associated for a max number of vdev.
>>>> This max would limit the number of vdev that a remote firmware can request.
>>>> Also not very flexible more that the index 0, as managed at platform level.
>>>>
>>>> Having a default name or a default index seems to me a good compromise...
>>>> One advantage of the default name (in this case not rproc->name) is the 
>>>> ability to define the memory region in a resource table carveout, instead
>>>> of a static definition in DT.
>>>
>>> Hmm, regarding the default name, vdev rings and vdev buffers are already
>>> looking for specific names, right? So, what resource type are you
>>> envisioning for this - RSC_CARVEOUT? Wouldn't that become a separate
>>> allocation by itself?
>>
>> After checking, I think my point on the carveout is not valid. As we should no
>> longer use dma_declare_coherent_memory which seems obsolete since the
>> introduction of the devicetree. Thus, even if a carveout is defined, we will
>> have to declare an associated memory region in DT.
>>
>> Anyway as a name is used today for other memory regions, i would be in flavor
>> of using the same method to declare a default memory pool in remote proc.
>> relying on the rproc_of_resm_mem_entry_init function.
> 
> This could definitely be an enhancement and conversion of existing
> drivers, but at this point, the idea is to restore the allocations to
> how they were for drivers defining a DMA pool and without having to
> modify them.
ok, in this case the fix would be your V1 patch and the question is do we take
the opportunity to implement something more scalable as existing is already broken?
or only first fix the issue with the associated limitations?
 
> 
>> If an index is preferred in core, platform drivers that declare more that
>> one memory region will probably have to implement code based on a default
>> memory region name to retrieve the associated index.
> 
> So, if we were to get rid of the dma_declare_coherent_memory() usage
> from remoteproc core, then I actually don't see need for the other args
> in rproc_of_resm_mem_entry_init(), just the name and id. 

yes probably.

>And even that
> you would only need if you are going to have to deal with multiple
> devices to assign the DMA pool for, for which you would need the name.

Here i was speaking about you V2 implementation. As consequence i would
implement a function in the stm32_rproc platform that would set the vdevbuf_mem_id,
depending on a default memory region name, i would define in stm32 bindings.

Regards
Arnaud

> 
> regards
> Suman
> 
>>
>> Regards,
>> Arnaud 
>>
>>
>>>
>>> regards
>>> Suman
>>>
>>>>>>>
>>>>>>>>      }
>>>>>>>>
>>>>>>>>      /* Allocate virtio device */
>>>>>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>>>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>>>>>> --- a/include/linux/remoteproc.h
>>>>>>>> +++ b/include/linux/remoteproc.h
>>>>>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>>>>>   * @dump_segments: list of segments in the firmware
>>>>>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>>>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>>>>>   */
>>>>>>>>  struct rproc {
>>>>>>>>      struct list_head node;
>>>>>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>>>>>      bool auto_boot;
>>>>>>>>      struct list_head dump_segments;
>>>>>>>>      int nb_vdev;
>>>>>>>> +    u8 vdevbuf_mem_id;
>>>>>>>>      u8 elf_class;
>>>>>>>>  };
>>>>>>>>
>>>>>>>> --
>>>>>>>> 2.23.0
>>>>>>>>
>>>>>>
>>>
>
Suman Anna April 9, 2020, 1:20 p.m. UTC | #11
On 4/9/20 4:58 AM, Arnaud POULIQUEN wrote:
> 
> On 4/9/20 1:36 AM, Suman Anna wrote:
>> Hi Arnaud,
>>
>>>>
>>>> On 3/30/20 7:29 AM, Arnaud POULIQUEN wrote:
>>>>>
>>>>>
>>>>> On 3/27/20 10:09 PM, Mathieu Poirier wrote:
>>>>>> On Wed, 25 Mar 2020 at 17:39, Suman Anna <s-anna@ti.com> wrote:
>>>>>>>
>>>>>>> Hi Mathieu,
>>>>>>>
>>>>>>> On 3/25/20 3:38 PM, Mathieu Poirier wrote:
>>>>>>>> On Thu, Mar 19, 2020 at 11:23:20AM -0500, Suman Anna wrote:
>>>>>>>>> From: Tero Kristo <t-kristo@ti.com>
>>>>>>>>>
>>>>>>>>> In some cases, like with OMAP remoteproc, we are not creating dedicated
>>>>>>>>> memory pool for the virtio device. Instead, we use the same memory pool
>>>>>>>>> for all shared memories. The current virtio memory pool handling forces
>>>>>>>>> a split between these two, as a separate device is created for it,
>>>>>>>>> causing memory to be allocated from bad location if the dedicated pool
>>>>>>>>> is not available. Fix this by falling back to using the parent device
>>>>>>>>> memory pool if dedicated is not available.
>>>>>>>>>
>>>>>>>>> Fixes: 086d08725d34 ("remoteproc: create vdev subdevice with specific dma memory pool")
>>>>>>>>> Signed-off-by: Tero Kristo <t-kristo@ti.com>
>>>>>>>>> Signed-off-by: Suman Anna <s-anna@ti.com>
>>>>>>>>> ---
>>>>>>>>> v2:
>>>>>>>>>  - Address Arnaud's concerns about hard-coded memory-region index 0
>>>>>>>>>  - Update the comment around the new code addition
>>>>>>>>> v1: https://patchwork.kernel.org/patch/11422721/
>>>>>>>>>
>>>>>>>>>  drivers/remoteproc/remoteproc_virtio.c | 15 +++++++++++++++
>>>>>>>>>  include/linux/remoteproc.h             |  2 ++
>>>>>>>>>  2 files changed, 17 insertions(+)
>>>>>>>>>
>>>>>>>>> diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
>>>>>>>>> index eb817132bc5f..b687715cdf4b 100644
>>>>>>>>> --- a/drivers/remoteproc/remoteproc_virtio.c
>>>>>>>>> +++ b/drivers/remoteproc/remoteproc_virtio.c
>>>>>>>>> @@ -369,6 +369,21 @@ int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
>>>>>>>>>                              goto out;
>>>>>>>>>                      }
>>>>>>>>>              }
>>>>>>>>> +    } else {
>>>>>>>>> +            struct device_node *np = rproc->dev.parent->of_node;
>>>>>>>>> +
>>>>>>>>> +            /*
>>>>>>>>> +             * If we don't have dedicated buffer, just attempt to re-assign
>>>>>>>>> +             * the reserved memory from our parent. A default memory-region
>>>>>>>>> +             * at index 0 from the parent's memory-regions is assigned for
>>>>>>>>> +             * the rvdev dev to allocate from, and this can be customized
>>>>>>>>> +             * by updating the vdevbuf_mem_id in platform drivers if
>>>>>>>>> +             * desired. Failure is non-critical and the allocations will
>>>>>>>>> +             * fall back to global pools, so don't check return value
>>>>>>>>> +             * either.
>>>>>>>>
>>>>>>>> I'm perplex...  In the changelog it is indicated that if a memory pool is
>>>>>>>> not dedicated allocation happens from a bad location but here failure of
>>>>>>>> getting a hold of a dedicated memory pool is not critical.
>>>>>>>
>>>>>>> So, the comment here is a generic one while the bad location part in the
>>>>>>> commit description is actually from OMAP remoteproc usage perspective
>>>>>>> (if you remember the dev_warn messages we added to the memory-region
>>>>>>> parse logic in the driver).
>>>>>>
>>>>>> I can't tell... Are you referring to the comment lines after
>>>>>> of_reserved_mem_device_init() in omap_rproc_probe()?
>>>>>>
>>>>>>>
>>>>>>> Before the fixed-memory carveout support, all the DMA allocations in
>>>>>>> remoteproc core were made from the rproc platform device's DMA pool (
>>>>>>> which can be NULL). That is lost after the fixed-memory support, and
>>>>>>> they were always allocated from global DMA pools if no dedicated pools
>>>>>>> are used. After this patch, that continues to be case for drivers that
>>>>>>> still do not use any dedicated pools, while it does restore the usage of
>>>>>>> the platform device's DMA pool if a driver uses one (OMAP remoteproc
>>>>>>> falls into the latter).
>>>>>>>
>>>>>>>>
>>>>>>>>> +             */
>>>>>>>>> +            of_reserved_mem_device_init_by_idx(dev, np,
>>>>>>>>> +                                               rproc->vdevbuf_mem_id);
>>>>>>>>
>>>>>>>> I wonder if using an index setup by platform code is really the best way
>>>>>>>> forward when we already have the carveout mechanic available to us.  I see the
>>>>>>>> platform code adding a carveout that would have the same name as rproc->name.
>>>>>>>> From there in rproc_add_virtio_dev() we could have something like:
>>>>>>>>
>>>>>>>>         mem = rproc_find_carveout_by_name(rproc, "%s", rproc->name);
>>>>>>>>
>>>>>>>>
>>>>>>>> That would be very flexible, the location of the reserved memory withing the
>>>>>>>> memory-region could change without fear of breaking things and no need to add to
>>>>>>>> struct rproc.
>>>>>>>>
>>>>>>>> Let me know what you think.
>>>>>>>
>>>>>>> I think that can work as well but I feel it is lot more cumbersome. It
>>>>>>> does require every platform driver to add code adding/registering that
>>>>>>> carveout, and parse the reserved memory region etc. End of the day, we
>>>>>>> rely on DMA API and we just have to assign the region to the newly
>>>>>>> created device. The DMA pool assignment for devices using
>>>>>>> reserved-memory nodes has simply been the of_reserved_mem_device_init()
>>>>>>> function.
>>>>>>
>>>>>> Given all the things happening in the platform drivers adding and
>>>>>> registering a single carveout doesn't seem that onerous to me.   I
>>>>>> also expect setting rproc->vdevbuf_mem_id would involve some form of
>>>>>> parsing.  Lastly if a couple of platforms end up doing the same thing
>>>>>> might as well bring the code in the core, hence choosing a generic
>>>>>> name such as rproc->name for the memory region.
>>>>>>
>>>>>> At the very least I would use of_reserved_mem_device_init_by_idx(dev,
>>>>>> np, 0).  I agree it is not flexible but I'll take that over adding a
>>>>>> new field to structure rproc.
>>>>>
>>>>> I wonder whether this would not introduce side effect for some legacy
>>>>> drivers. Some rproc platforms can have a memory region defined but not
>>>>> used for the virtio buffers which is allocated in the Linux default
>>>>> memory pool.
>>>>
>>>> I am actually trying to restore the behavior before the separate vdev
>>>> device creation. All allocations were coming from the rproc platform
>>>> device's DMA pool (if assigned, otherwise will fall back to global pool).
>>>
>>> Yes deprecated dma_declare_coherent_memory was used to declare the rproc
>>> platform pool for vdev if i well remember).
>>
>> Right, the dma_declare_coherent_memory() usage was introduced only in
>> v5.1 as part of vdev subdevice creation. Prior to that, the allocations
>> used the rproc dev DMA pool which would have been initialized in the
>> platform drivers.
>>
>>>
>>>>
>>>> You have a valid point though, but are there any drivers relying on this
>>>> today? All the usage I saw were only using of_reserved_mem_device_init()
>>>> which is 0-indexed region by default. If this is really an issue, we can
>>>> use -1 as a means to conditionally follow this path.
>>>
>>> Up-streamed driver seems using nothing but address 0 for time being. 
>>> I don't know for non-up-streamed drivers...
>>> But if you have a look to the stm32 platform driver, we also define the
>>> memory regions for the code. 
>>
>> ST drivers are already using the "vdev0buffer" on latest kernels, so
>> they will not be entering the fallback path anyway right. Looking at
>> 4.19 kernel, I see that you guys were also using a single
>> reserved-memory region (didn't have DT nodes though), so all allocations
>> would have come from that region. The patch is for fixing the
>> allocations for drivers that were defining a DMA rproc pool, and still
>> using them (so not having to force the drivers to add code to fix the
>> issue).
> 
> Right, but today the limitation with our implementation is the support of
> additional shared buffers(big data). The possibility to add a single DMA pool
> would help to solve this issue. Furthermore we have a DT generator tools that
> should support the shared memory regions tuning. Having an unique shared memory
> region could simplify this tool usage.

big data is always interesting, and you probably need some more tooling
around it than just the DMA pool in rproc especially if you were to
share it with other kernel subsystems. You would want to keep that
separate from the firmware regions. Our solution has been to use the
in-kernel buffer allocators/dma-buf exporters. We had our own
out-of-tree solution before, and now converging on to use the recently
added dma-buf heaps (drivers/dma-buf/heaps). You would want to know the
pool especially on non-MMU rprocs given that you would have to configure
certain Cache settings for the regions.

> For these reasons i would be interested in finding a more scalable solution.  
> 
>>
>>> Even if we are also interesting by the possibility of having a default pool,
>>> it is no simple to implement for our driver using index 0.
>>> We will have to declare it first in DT. But we would also have to determine
>>> if the first memory pool declaration is a default pool or not vs the other
>>> memory region declarations.
>>
>> Yeah, if we used the hard-coded 0 and you were adapting to use a default
>> pool. Otherwise, you could technically update or define the new generic
>> name (which has to be added to remoteproc core as well).
>>
>>>
>>>>
>>>>> If the aim is to minimize impact in the core part, the solution i proposed
>>>>> in V1 using rproc_of_resm_mem_entry_init seems a good default candidate too.
>>>>> The constraint would be that the platform driver has to pre-register
>>>>> vdev<X>buffer reserved memory associated for a max number of vdev.
>>>>> This max would limit the number of vdev that a remote firmware can request.
>>>>> Also not very flexible more that the index 0, as managed at platform level.
>>>>>
>>>>> Having a default name or a default index seems to me a good compromise...
>>>>> One advantage of the default name (in this case not rproc->name) is the 
>>>>> ability to define the memory region in a resource table carveout, instead
>>>>> of a static definition in DT.
>>>>
>>>> Hmm, regarding the default name, vdev rings and vdev buffers are already
>>>> looking for specific names, right? So, what resource type are you
>>>> envisioning for this - RSC_CARVEOUT? Wouldn't that become a separate
>>>> allocation by itself?
>>>
>>> After checking, I think my point on the carveout is not valid. As we should no
>>> longer use dma_declare_coherent_memory which seems obsolete since the
>>> introduction of the devicetree. Thus, even if a carveout is defined, we will
>>> have to declare an associated memory region in DT.
>>>
>>> Anyway as a name is used today for other memory regions, i would be in flavor
>>> of using the same method to declare a default memory pool in remote proc.
>>> relying on the rproc_of_resm_mem_entry_init function.
>>
>> This could definitely be an enhancement and conversion of existing
>> drivers, but at this point, the idea is to restore the allocations to
>> how they were for drivers defining a DMA pool and without having to
>> modify them.

> ok, in this case the fix would be your V1 patch and the question is do we take
> the opportunity to implement something more scalable as existing is already broken?
> or only first fix the issue with the associated limitations?

I am fine with going back to v1 (and ok with either of the versions :)).
I can post a v3 going back if we all agree that is what we want to merge.

Bjorn,
Your thoughts on this patch, and what's your preference?

The improvements can be follow-up patches. The first thing we should get
to is to get rid of the dma_declare_coherent_memory() before anything
else. We do not want any users of rproc_of_resm_mem_entry_init() to
exercise that path.

>  
>>
>>> If an index is preferred in core, platform drivers that declare more that
>>> one memory region will probably have to implement code based on a default
>>> memory region name to retrieve the associated index.
>>
>> So, if we were to get rid of the dma_declare_coherent_memory() usage
>> from remoteproc core, then I actually don't see need for the other args
>> in rproc_of_resm_mem_entry_init(), just the name and id. 
> 
> yes probably.
> 
>> And even that
>> you would only need if you are going to have to deal with multiple
>> devices to assign the DMA pool for, for which you would need the name.
> 
> Here i was speaking about you V2 implementation. As consequence i would
> implement a function in the stm32_rproc platform that would set the vdevbuf_mem_id,
> depending on a default memory region name, i would define in stm32 bindings.

Yep, that was the idea. It would have fixed the existing drivers, and if
one wanted to customize, the platform drivers can choose the right id.
The name + id registration would also entail the platform drivers
knowing the id anyways. And today, the only other device where we are
registering a DMA pool is the vdev subdevice.

regards
Suman

> 
> Regards
> Arnaud
> 
>>
>> regards
>> Suman
>>
>>>
>>> Regards,
>>> Arnaud 
>>>
>>>
>>>>
>>>> regards
>>>> Suman
>>>>
>>>>>>>>
>>>>>>>>>      }
>>>>>>>>>
>>>>>>>>>      /* Allocate virtio device */
>>>>>>>>> diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
>>>>>>>>> index ed127b2d35ca..07bd73a6d72a 100644
>>>>>>>>> --- a/include/linux/remoteproc.h
>>>>>>>>> +++ b/include/linux/remoteproc.h
>>>>>>>>> @@ -481,6 +481,7 @@ struct rproc_dump_segment {
>>>>>>>>>   * @auto_boot: flag to indicate if remote processor should be auto-started
>>>>>>>>>   * @dump_segments: list of segments in the firmware
>>>>>>>>>   * @nb_vdev: number of vdev currently handled by rproc
>>>>>>>>> + * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
>>>>>>>>>   */
>>>>>>>>>  struct rproc {
>>>>>>>>>      struct list_head node;
>>>>>>>>> @@ -514,6 +515,7 @@ struct rproc {
>>>>>>>>>      bool auto_boot;
>>>>>>>>>      struct list_head dump_segments;
>>>>>>>>>      int nb_vdev;
>>>>>>>>> +    u8 vdevbuf_mem_id;
>>>>>>>>>      u8 elf_class;
>>>>>>>>>  };
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> 2.23.0
>>>>>>>>>
>>>>>>>
>>>>
>>
diff mbox series

Patch

diff --git a/drivers/remoteproc/remoteproc_virtio.c b/drivers/remoteproc/remoteproc_virtio.c
index eb817132bc5f..b687715cdf4b 100644
--- a/drivers/remoteproc/remoteproc_virtio.c
+++ b/drivers/remoteproc/remoteproc_virtio.c
@@ -369,6 +369,21 @@  int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
 				goto out;
 			}
 		}
+	} else {
+		struct device_node *np = rproc->dev.parent->of_node;
+
+		/*
+		 * If we don't have dedicated buffer, just attempt to re-assign
+		 * the reserved memory from our parent. A default memory-region
+		 * at index 0 from the parent's memory-regions is assigned for
+		 * the rvdev dev to allocate from, and this can be customized
+		 * by updating the vdevbuf_mem_id in platform drivers if
+		 * desired. Failure is non-critical and the allocations will
+		 * fall back to global pools, so don't check return value
+		 * either.
+		 */
+		of_reserved_mem_device_init_by_idx(dev, np,
+						   rproc->vdevbuf_mem_id);
 	}
 
 	/* Allocate virtio device */
diff --git a/include/linux/remoteproc.h b/include/linux/remoteproc.h
index ed127b2d35ca..07bd73a6d72a 100644
--- a/include/linux/remoteproc.h
+++ b/include/linux/remoteproc.h
@@ -481,6 +481,7 @@  struct rproc_dump_segment {
  * @auto_boot: flag to indicate if remote processor should be auto-started
  * @dump_segments: list of segments in the firmware
  * @nb_vdev: number of vdev currently handled by rproc
+ * @vdevbuf_mem_id: default memory-region index for allocating vdev buffers
  */
 struct rproc {
 	struct list_head node;
@@ -514,6 +515,7 @@  struct rproc {
 	bool auto_boot;
 	struct list_head dump_segments;
 	int nb_vdev;
+	u8 vdevbuf_mem_id;
 	u8 elf_class;
 };