Message ID | 20240206172759.421737-8-sui.jingfeng@linux.dev (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | drm/etnaviv: Add driver wrapper for vivante GPUs attached on PCI(e) device | expand |
On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: > The component helper functions are the glue, which is used to bind multiple > GPU cores to a virtual master platform device. Which is fine and works well > for the SoCs who contains multiple GPU cores. > > The problem is that usperspace programs (such as X server and Mesa) will > search the PCIe device to use if it is exist. In other words, usperspace > programs open the PCIe device with higher priority. Creating a virtual > master platform device for PCI(e) GPUs is unnecessary, as the PCI device > has been created by the time drm/etnaviv is loaded. > > we create virtual platform devices as a representation for the vivante GPU > ip core. As all of subcomponent are attached via the PCIe master device, > we reflect this hardware layout by binding all of the virtual child to the > the real master. > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> Uh so my understanding is that drivers really shouldn't create platform devices of their own. For this case here I think the aux-bus framework is the right thing to use. Alternatively would be some infrastructure where you feed a DT tree to driver core or pci subsystem and it instantiates it all for you correctly, and especially with hotunplug all done right since this is pci now, not actually part of the soc that cannot be hotunplugged. I think I've seen some other pci devices from arm soc designs that would benefit from this too, so lifting this logic into a pci function would make sense imo. Cheers, Sima > --- > drivers/gpu/drm/etnaviv/Kconfig | 8 + > drivers/gpu/drm/etnaviv/Makefile | 2 + > drivers/gpu/drm/etnaviv/etnaviv_drv.c | 12 +- > drivers/gpu/drm/etnaviv/etnaviv_gpu.c | 75 ++++++-- > drivers/gpu/drm/etnaviv/etnaviv_gpu.h | 4 + > drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c | 224 ++++++++++++++++++++++ > drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h | 18 ++ > 7 files changed, 327 insertions(+), 16 deletions(-) > create mode 100644 drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c > create mode 100644 drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h > > diff --git a/drivers/gpu/drm/etnaviv/Kconfig b/drivers/gpu/drm/etnaviv/Kconfig > index faa7fc68b009..38c251585ec1 100644 > --- a/drivers/gpu/drm/etnaviv/Kconfig > +++ b/drivers/gpu/drm/etnaviv/Kconfig > @@ -15,6 +15,14 @@ config DRM_ETNAVIV > help > DRM driver for Vivante GPUs. > > +config DRM_ETNAVIV_PCI_DRIVER > + bool "enable ETNAVIV PCI driver support" > + depends on DRM_ETNAVIV > + depends on PCI > + help > + Compile in support for Vivante GPUs attached via PCI(e). > + Say Y if you have such hardwares. > + > config DRM_ETNAVIV_THERMAL > bool "enable ETNAVIV thermal throttling" > depends on DRM_ETNAVIV > diff --git a/drivers/gpu/drm/etnaviv/Makefile b/drivers/gpu/drm/etnaviv/Makefile > index 46e5ffad69a6..6829e1ebf2db 100644 > --- a/drivers/gpu/drm/etnaviv/Makefile > +++ b/drivers/gpu/drm/etnaviv/Makefile > @@ -16,4 +16,6 @@ etnaviv-y := \ > etnaviv_perfmon.o \ > etnaviv_sched.o > > +etnaviv-$(CONFIG_DRM_ETNAVIV_PCI_DRIVER) += etnaviv_pci_drv.o > + > obj-$(CONFIG_DRM_ETNAVIV) += etnaviv.o > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_drv.c b/drivers/gpu/drm/etnaviv/etnaviv_drv.c > index 5f65f2dead44..48e2402adbe3 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_drv.c > +++ b/drivers/gpu/drm/etnaviv/etnaviv_drv.c > @@ -24,6 +24,7 @@ > #include "etnaviv_gpu.h" > #include "etnaviv_gem.h" > #include "etnaviv_mmu.h" > +#include "etnaviv_pci_drv.h" > #include "etnaviv_perfmon.h" > > /* > @@ -566,6 +567,10 @@ static int etnaviv_bind(struct device *dev) > if (ret < 0) > goto out_free_priv; > > + ret = etnaviv_register_irq_handler(dev, priv); > + if (ret) > + goto out_unbind; > + > load_gpu(drm); > > ret = drm_dev_register(drm, 0); > @@ -594,7 +599,7 @@ static void etnaviv_unbind(struct device *dev) > etnaviv_private_fini(priv); > } > > -static const struct component_master_ops etnaviv_master_ops = { > +const struct component_master_ops etnaviv_master_ops = { > .bind = etnaviv_bind, > .unbind = etnaviv_unbind, > }; > @@ -740,6 +745,10 @@ static int __init etnaviv_init(void) > if (ret != 0) > goto unregister_gpu_driver; > > + ret = etnaviv_register_pci_driver(); > + if (ret != 0) > + goto unregister_platform_driver; > + > /* > * If the DT contains at least one available GPU device, instantiate > * the DRM platform device. > @@ -769,6 +778,7 @@ module_init(etnaviv_init); > static void __exit etnaviv_exit(void) > { > etnaviv_destroy_platform_device(&etnaviv_drm); > + etnaviv_unregister_pci_driver(); > platform_driver_unregister(&etnaviv_platform_driver); > platform_driver_unregister(&etnaviv_gpu_driver); > } > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c > index 3fd637c17797..221020e98900 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c > +++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c > @@ -10,6 +10,7 @@ > #include <linux/dma-mapping.h> > #include <linux/mod_devicetable.h> > #include <linux/module.h> > +#include <linux/pci.h> > #include <linux/platform_device.h> > #include <linux/pm_runtime.h> > #include <linux/regulator/consumer.h> > @@ -29,6 +30,7 @@ > > static const struct platform_device_id gpu_ids[] = { > { .name = "etnaviv-gpu,2d" }, > + { .name = "etnaviv-gpu,3d" }, > { }, > }; > > @@ -1555,14 +1557,22 @@ static void dump_mmu_fault(struct etnaviv_gpu *gpu) > > static irqreturn_t irq_handler(int irq, void *data) > { > - struct etnaviv_gpu *gpu = data; > + struct etnaviv_drm_private *priv = data; > irqreturn_t ret = IRQ_NONE; > + int i; > > - u32 intr = gpu_read(gpu, VIVS_HI_INTR_ACKNOWLEDGE); > - > - if (intr != 0) { > + for (i = 0; i < priv->num_gpus; i++) { > + struct etnaviv_gpu *gpu = priv->gpu[i]; > + u32 intr; > int event; > > + if (!gpu) > + continue; > + > + intr = gpu_read(gpu, VIVS_HI_INTR_ACKNOWLEDGE); > + if (!intr) > + continue; > + > pm_runtime_mark_last_busy(gpu->dev); > > dev_dbg(gpu->dev, "intr 0x%08x\n", intr); > @@ -1893,10 +1903,44 @@ static const struct of_device_id etnaviv_gpu_match[] = { > }; > MODULE_DEVICE_TABLE(of, etnaviv_gpu_match); > > +/* > + * dev point to the master. For platform device, it is virtual. > + * For PCI(e) device, it is real. > + */ > +int etnaviv_register_irq_handler(struct device *dev, > + struct etnaviv_drm_private *priv) > +{ > + bool is_pci = dev_is_pci(dev); > + int ret = 0; > + > + if (is_pci) { > + struct pci_dev *pdev = to_pci_dev(dev); > + > + ret = request_irq(pdev->irq, irq_handler, IRQF_SHARED, > + dev_name(dev), priv); > + } else { > + int i; > + > + for (i = 0; i < priv->num_gpus; i++) { > + struct etnaviv_gpu *gpu = priv->gpu[i]; > + > + ret = devm_request_irq(gpu->dev, gpu->irq, irq_handler, > + 0, dev_name(dev), priv); > + if (ret) { > + dev_err(dev, "failed to request IRQ handler: %d\n", ret); > + break; > + } > + } > + } > + > + return ret; > +} > + > static int etnaviv_gpu_platform_probe(struct platform_device *pdev) > { > struct device *dev = &pdev->dev; > struct etnaviv_gpu *gpu; > + bool is_pci; > int err; > > gpu = devm_kzalloc(dev, sizeof(*gpu), GFP_KERNEL); > @@ -1912,22 +1956,23 @@ static int etnaviv_gpu_platform_probe(struct platform_device *pdev) > if (IS_ERR(gpu->mmio)) > return PTR_ERR(gpu->mmio); > > + is_pci = dev->parent ? dev_is_pci(dev->parent) : false; > + > /* Get Interrupt: */ > - gpu->irq = platform_get_irq(pdev, 0); > + if (is_pci) > + gpu->irq = to_pci_dev(dev->parent)->irq; > + else > + gpu->irq = platform_get_irq(pdev, 0); > + > if (gpu->irq < 0) > return gpu->irq; > > - err = devm_request_irq(dev, gpu->irq, irq_handler, 0, > - dev_name(dev), gpu); > - if (err) { > - dev_err(dev, "failed to request IRQ%u: %d\n", gpu->irq, err); > - return err; > - } > - > /* Get Clocks: */ > - err = etnaviv_gpu_clk_get(gpu); > - if (err) > - return err; > + if (!is_pci) { > + err = etnaviv_gpu_clk_get(gpu); > + if (err) > + return err; > + } > > /* TODO: figure out max mapped size */ > dev_set_drvdata(dev, gpu); > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h > index 7d5e9158e13c..d354e096652c 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h > +++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h > @@ -208,6 +208,10 @@ static inline u32 gpu_read_power(struct etnaviv_gpu *gpu, u32 reg) > int etnaviv_gpu_get_param(struct etnaviv_gpu *gpu, u32 param, u64 *value); > > int etnaviv_gpu_init(struct etnaviv_gpu *gpu); > + > +int etnaviv_register_irq_handler(struct device *dev, > + struct etnaviv_drm_private *priv); > + > bool etnaviv_fill_identity_from_hwdb(struct etnaviv_gpu *gpu); > > #ifdef CONFIG_DEBUG_FS > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c > new file mode 100644 > index 000000000000..ec5039866eda > --- /dev/null > +++ b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c > @@ -0,0 +1,224 @@ > +// SPDX-License-Identifier: GPL-2.0 > + > +#include <linux/component.h> > +#include <linux/pci.h> > + > +#include "etnaviv_drv.h" > +#include "etnaviv_pci_drv.h" > + > +enum etnaviv_pci_gpu_chip_id { > + GC_CORE_UNKNOWN = 0, > + JM9100 = 1, > + JD9230P = 2, > + GP102 = 3, > + GC1000_IN_LS7A1000 = 4, > + GC1000_IN_LS2K1000 = 5, > + GC_CORE_PCI_LAST, > +}; > + > +struct etnaviv_pci_gpu_data { > + enum etnaviv_pci_gpu_chip_id chip_id; > + u32 num_core; > + u32 num_vram; > + u32 vram_bars[2]; > + u32 mmio_bar; > + struct { > + u32 id; > + u32 offset; > + u32 size; > + char compatible[20]; > + } cores[ETNA_MAX_PIPES]; > + > + bool has_dedicated_vram; > + char market_name[24]; > +}; > + > +static const struct etnaviv_pci_gpu_data > +gc_core_plaform_data[GC_CORE_PCI_LAST] = { > + { > + .chip_id = GC_CORE_UNKNOWN, > + }, > + { > + .chip_id = JM9100, > + .num_core = 1, > + .num_vram = 2, > + .vram_bars = {0, 2}, > + .mmio_bar = 1, > + .cores = {{0, 0x00900000, 0x00010000, "etnaviv-gpu,3d"},}, > + .has_dedicated_vram = true, > + .market_name = "JingJia Micro JM9100", > + }, > + { > + .chip_id = JD9230P, > + .num_core = 2, > + .num_vram = 2, > + .vram_bars = {0, 2}, > + .mmio_bar = 1, > + .cores = {{0, 0x00900000, 0x00010000, "etnaviv-gpu,3d"}, > + {1, 0x00910000, 0x00010000, "etnaviv-gpu,3d"},}, > + .has_dedicated_vram = true, > + .market_name = "JingJia Micro JD9230P", > + }, > + { > + .chip_id = GP102, > + .num_core = 2, > + .num_vram = 1, > + .vram_bars = {0,}, > + .mmio_bar = 2, > + .cores = {{0, 0x00040000, 0x00010000, "etnaviv-gpu,3d"}, > + {0, 0x000C0000, 0x00010000, "etnaviv-gpu,2d"},}, > + .has_dedicated_vram = true, > + .market_name = "LingJiu GP102", > + }, > + { > + .chip_id = GC1000_IN_LS7A1000, > + .num_core = 1, > + .num_vram = 1, > + .vram_bars = {2,}, > + .mmio_bar = 0, > + .cores = {{0, 0, 0x00010000, "etnaviv-gpu,3d"}, {}, {}, {}}, > + .has_dedicated_vram = true, > + .market_name = "GC1000 in LS7A1000", > + }, > + { > + .chip_id = GC1000_IN_LS2K1000, > + .num_core = 1, > + .num_vram = 0, > + .mmio_bar = 0, > + .cores = {{0, 0, 0x00010000, "etnaviv-gpu,3d"}, {}, {}, {}}, > + .has_dedicated_vram = false, > + .market_name = "GC1000 in LS2K1000", > + }, > +}; > + > +static const struct etnaviv_pci_gpu_data * > +etnaviv_pci_get_platform_data(const struct pci_device_id *entity) > +{ > + enum etnaviv_pci_gpu_chip_id chip_id = entity->driver_data; > + static const struct etnaviv_pci_gpu_data *pdata; > + > + pdata = &gc_core_plaform_data[chip_id]; > + if (!pdata || pdata->chip_id == GC_CORE_UNKNOWN) > + return NULL; > + > + return pdata; > +} > + > +extern const struct component_master_ops etnaviv_master_ops; > + > +static int etnaviv_pci_probe(struct pci_dev *pdev, > + const struct pci_device_id *ent) > +{ > + const struct etnaviv_pci_gpu_data *pdata; > + struct device *dev = &pdev->dev; > + struct component_match *matches = NULL; > + unsigned int i; > + unsigned int num_core; > + int ret; > + > + ret = pcim_enable_device(pdev); > + if (ret) { > + dev_err(dev, "failed to enable\n"); > + return ret; > + } > + > + pci_set_master(pdev); > + > + ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); > + if (ret) > + return ret; > + > + pdata = etnaviv_pci_get_platform_data(ent); > + if (!pdata) > + return -ENODEV; > + > + num_core = pdata->num_core; > + > + dev_info(dev, "%s has %u GPU cores\n", pdata->market_name, num_core); > + > + /* > + * Create a virtual platform device for the sub-component, > + * a sub-component is refer to a single vivante GPU core. > + * But it can also be extended to stand for a display controller > + * or any other IP core attached via the same PCIe master. > + */ > + for (i = 0; i < num_core; i++) { > + struct platform_device *virtual_child; > + resource_size_t start, offset, size; > + struct resource res; > + > + start = pci_resource_start(pdev, pdata->mmio_bar); > + offset = pdata->cores[i].offset; > + size = pdata->cores[i].size; > + > + memset(&res, 0, sizeof(res)); > + res.flags = IORESOURCE_MEM; > + res.name = "reg"; > + res.start = start + offset; > + res.end = start + offset + size - 1; > + > + ret = etnaviv_create_platform_device(dev, > + pdata->cores[i].compatible, > + pdata->cores[i].id, > + &res, > + (void *)pdata, > + &virtual_child); > + if (ret) > + return ret; > + > + component_match_add(dev, &matches, component_compare_dev, > + &virtual_child->dev); > + } > + > + ret = component_master_add_with_match(dev, &etnaviv_master_ops, matches); > + > + return ret; > +} > + > +static int platform_device_remove_callback(struct device *dev, void *data) > +{ > + struct platform_device *pdev = to_platform_device(dev); > + > + etnaviv_destroy_platform_device(&pdev); > + > + return 0; > +} > + > +static void etnaviv_pci_remove(struct pci_dev *pdev) > +{ > + struct device *dev = &pdev->dev; > + > + component_master_del(dev, &etnaviv_master_ops); > + > + device_for_each_child(dev, NULL, platform_device_remove_callback); > + > + pci_clear_master(pdev); > +} > + > +static const struct pci_device_id etnaviv_pci_id_lists[] = { > + {0x0731, 0x9100, PCI_ANY_ID, PCI_ANY_ID, 0, 0, JM9100}, > + {0x0731, 0x9230, PCI_ANY_ID, PCI_ANY_ID, 0, 0, JD9230P}, > + {0x0709, 0x0001, PCI_ANY_ID, PCI_ANY_ID, 0, 0, GP102}, > + {0x0014, 0x7A15, PCI_ANY_ID, PCI_ANY_ID, 0, 0, GC1000_IN_LS7A1000}, > + {0x0014, 0x7A05, PCI_ANY_ID, PCI_ANY_ID, 0, 0, GC1000_IN_LS2K1000}, > + { } > +}; > + > +static struct pci_driver etnaviv_pci_driver = { > + .name = "etnaviv", > + .id_table = etnaviv_pci_id_lists, > + .probe = etnaviv_pci_probe, > + .remove = etnaviv_pci_remove, > +}; > + > +int etnaviv_register_pci_driver(void) > +{ > + return pci_register_driver(&etnaviv_pci_driver); > +} > + > +void etnaviv_unregister_pci_driver(void) > +{ > + pci_unregister_driver(&etnaviv_pci_driver); > +} > + > +MODULE_DEVICE_TABLE(pci, etnaviv_pci_id_lists); > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h > new file mode 100644 > index 000000000000..1db559ee5e9b > --- /dev/null > +++ b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h > @@ -0,0 +1,18 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > + > +#ifndef __ETNAVIV_PCI_DRV_H__ > +#define __ETNAVIV_PCI_DRV_H__ > + > +#ifdef CONFIG_DRM_ETNAVIV_PCI_DRIVER > + > +int etnaviv_register_pci_driver(void); > +void etnaviv_unregister_pci_driver(void); > + > +#else > + > +static inline int etnaviv_register_pci_driver(void) { return 0; } > +static inline void etnaviv_unregister_pci_driver(void) { } > + > +#endif > + > +#endif > -- > 2.34.1 >
Hi, On 2024/2/7 17:35, Daniel Vetter wrote: > On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: >> The component helper functions are the glue, which is used to bind multiple >> GPU cores to a virtual master platform device. Which is fine and works well >> for the SoCs who contains multiple GPU cores. >> >> The problem is that usperspace programs (such as X server and Mesa) will >> search the PCIe device to use if it is exist. In other words, usperspace >> programs open the PCIe device with higher priority. Creating a virtual >> master platform device for PCI(e) GPUs is unnecessary, as the PCI device >> has been created by the time drm/etnaviv is loaded. >> >> we create virtual platform devices as a representation for the vivante GPU >> ip core. As all of subcomponent are attached via the PCIe master device, >> we reflect this hardware layout by binding all of the virtual child to the >> the real master. >> >> Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > Uh so my understanding is that drivers really shouldn't create platform > devices of their own. > Yes, At least for DT-based systems, this driver can be modified to let the core to create the virtual master for us. We don't have to create platform devices by our own(refer to the drm/etnaviv driver). I means that we could put the following example device node into the .dts file. gpu_2d: gpu@A0000 { compatible = "vivante,gc"; reg = <0xA0000 0x4000>; }; gpu_3d: gpu@90000 { compatible = "vivante,gc"; reg = <0x90000 0x4000>; }; gpu@0 { compatible = "etnaviv"; cores = <&gpu_2d &gpu_3d>; dma-coherent; dma-mask = <0xffffffff> virtual_master; }; But now, I'm afraid it's too late. Because the DTS/DTB may already have been burned into board's BIOS for years. I guess, nowadays, modifying(changes) this driver have to take the backward compatibility constraint into consideration. Since we only have one chance to form the spec, that happens when this driver was initially merged. Apparently, we miss it.
On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: > On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: > > The component helper functions are the glue, which is used to bind multiple > > GPU cores to a virtual master platform device. Which is fine and works well > > for the SoCs who contains multiple GPU cores. > > > > The problem is that usperspace programs (such as X server and Mesa) will > > search the PCIe device to use if it is exist. In other words, usperspace > > programs open the PCIe device with higher priority. Creating a virtual > > master platform device for PCI(e) GPUs is unnecessary, as the PCI device > > has been created by the time drm/etnaviv is loaded. > > > > we create virtual platform devices as a representation for the vivante GPU > > ip core. As all of subcomponent are attached via the PCIe master device, > > we reflect this hardware layout by binding all of the virtual child to the > > the real master. > > > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > > Uh so my understanding is that drivers really shouldn't create platform > devices of their own. For this case here I think the aux-bus framework is > the right thing to use. Alternatively would be some infrastructure where > you feed a DT tree to driver core or pci subsystem and it instantiates it > all for you correctly, and especially with hotunplug all done right since > this is pci now, not actually part of the soc that cannot be hotunplugged. I don't think we need intermediate platform devices at all. We just need to register our GPU against the PCI device and that's it. We don't need a platform device, we don't need the component framework. > I think I've seen some other pci devices from arm soc designs that would > benefit from this too, so lifting this logic into a pci function would > make sense imo. Nouveau supports both iirc. Maxime
On Thu, Feb 08, 2024 at 04:27:02PM +0100, Maxime Ripard wrote: > On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: > > On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: > > > The component helper functions are the glue, which is used to bind multiple > > > GPU cores to a virtual master platform device. Which is fine and works well > > > for the SoCs who contains multiple GPU cores. > > > > > > The problem is that usperspace programs (such as X server and Mesa) will > > > search the PCIe device to use if it is exist. In other words, usperspace > > > programs open the PCIe device with higher priority. Creating a virtual > > > master platform device for PCI(e) GPUs is unnecessary, as the PCI device > > > has been created by the time drm/etnaviv is loaded. > > > > > > we create virtual platform devices as a representation for the vivante GPU > > > ip core. As all of subcomponent are attached via the PCIe master device, > > > we reflect this hardware layout by binding all of the virtual child to the > > > the real master. > > > > > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > > > > Uh so my understanding is that drivers really shouldn't create platform > > devices of their own. For this case here I think the aux-bus framework is > > the right thing to use. Alternatively would be some infrastructure where > > you feed a DT tree to driver core or pci subsystem and it instantiates it > > all for you correctly, and especially with hotunplug all done right since > > this is pci now, not actually part of the soc that cannot be hotunplugged. > > I don't think we need intermediate platform devices at all. We just need > to register our GPU against the PCI device and that's it. We don't need > a platform device, we don't need the component framework. Afaik that's what this series does. The component stuff is for the internal structure of the gpu ip, so that the same modular approach that works for arm-soc also works for pci chips. Otherwise we end up with each driver hand-rolling that stuff, which is defacto what both nouveau and amdgpu do (intel hw is too much a mess for that component-driver based approach to actually work reasonably well). Cheers, Sima > > I think I've seen some other pci devices from arm soc designs that would > > benefit from this too, so lifting this logic into a pci function would > > make sense imo. > > Nouveau supports both iirc. > > Maxime
On Fri, Feb 09, 2024 at 12:02:48PM +0100, Daniel Vetter wrote: > On Thu, Feb 08, 2024 at 04:27:02PM +0100, Maxime Ripard wrote: > > On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: > > > On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: > > > > The component helper functions are the glue, which is used to bind multiple > > > > GPU cores to a virtual master platform device. Which is fine and works well > > > > for the SoCs who contains multiple GPU cores. > > > > > > > > The problem is that usperspace programs (such as X server and Mesa) will > > > > search the PCIe device to use if it is exist. In other words, usperspace > > > > programs open the PCIe device with higher priority. Creating a virtual > > > > master platform device for PCI(e) GPUs is unnecessary, as the PCI device > > > > has been created by the time drm/etnaviv is loaded. > > > > > > > > we create virtual platform devices as a representation for the vivante GPU > > > > ip core. As all of subcomponent are attached via the PCIe master device, > > > > we reflect this hardware layout by binding all of the virtual child to the > > > > the real master. > > > > > > > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > > > > > > Uh so my understanding is that drivers really shouldn't create platform > > > devices of their own. For this case here I think the aux-bus framework is > > > the right thing to use. Alternatively would be some infrastructure where > > > you feed a DT tree to driver core or pci subsystem and it instantiates it > > > all for you correctly, and especially with hotunplug all done right since > > > this is pci now, not actually part of the soc that cannot be hotunplugged. > > > > I don't think we need intermediate platform devices at all. We just need > > to register our GPU against the PCI device and that's it. We don't need > > a platform device, we don't need the component framework. > > Afaik that's what this series does. The component stuff is for the > internal structure of the gpu ip, so that the same modular approach that > works for arm-soc also works for pci chips. But there should be a single PCI device, while we have multiple "DT" devices, right? Or is there several PCI devices too on that PCI card? Maxime
Hi, On 2024/2/9 23:15, Maxime Ripard wrote: > On Fri, Feb 09, 2024 at 12:02:48PM +0100, Daniel Vetter wrote: >> On Thu, Feb 08, 2024 at 04:27:02PM +0100, Maxime Ripard wrote: >>> On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: >>>> On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: >>>>> The component helper functions are the glue, which is used to bind multiple >>>>> GPU cores to a virtual master platform device. Which is fine and works well >>>>> for the SoCs who contains multiple GPU cores. >>>>> >>>>> The problem is that usperspace programs (such as X server and Mesa) will >>>>> search the PCIe device to use if it is exist. In other words, usperspace >>>>> programs open the PCIe device with higher priority. Creating a virtual >>>>> master platform device for PCI(e) GPUs is unnecessary, as the PCI device >>>>> has been created by the time drm/etnaviv is loaded. >>>>> >>>>> we create virtual platform devices as a representation for the vivante GPU >>>>> ip core. As all of subcomponent are attached via the PCIe master device, >>>>> we reflect this hardware layout by binding all of the virtual child to the >>>>> the real master. >>>>> >>>>> Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> >>>> Uh so my understanding is that drivers really shouldn't create platform >>>> devices of their own. For this case here I think the aux-bus framework is >>>> the right thing to use. Alternatively would be some infrastructure where >>>> you feed a DT tree to driver core or pci subsystem and it instantiates it >>>> all for you correctly, and especially with hotunplug all done right since >>>> this is pci now, not actually part of the soc that cannot be hotunplugged. >>> I don't think we need intermediate platform devices at all. We just need >>> to register our GPU against the PCI device and that's it. We don't need >>> a platform device, we don't need the component framework. >> Afaik that's what this series does. The component stuff is for the >> internal structure of the gpu ip, so that the same modular approach that >> works for arm-soc also works for pci chips. > But there should be a single PCI device, while we have multiple "DT" > devices, right? Or is there several PCI devices too on that PCI card? There is only a single PCI(e) device on that PCI(e) card, this single PCI(e) device is selected as the component master. All other Hardware IP blocks are shipped by the single PCI(e) master. It may includes Display controllers, GPUs, video decoders, HDMI display bridges hardware unit etc. But all of those Hardware IP share the same MMIO registers PCI BAR, this PCI BAR is a kind of PCI(e) MEM resource. It is a relative *big* chunk, as large as 32MB in address ranges for the JingJia Macro dGPU. Therefore, I break the whole registers memory(MMIO) resource into smaller pieces by creating platform device manually, manually created platform device is called as virtual child in this series. In short, we cut the whole into smaller piece, each smaller piece is a single hardware IP block, thus deserve a single device driver. We will have multiple platform devices if the dGPU contains multiple hardware IP block. On the driver side, we bind all of the scattered driver module with component. Bind with another PCI(e) device is also possible, if it is going to be used under the same driver. It is just that this will not need us to create a platform device for it manually. We won't set its parent, they are siblings then. > Maxime
On Sat, Feb 10, 2024 at 12:25:33AM +0800, Sui Jingfeng wrote: > On 2024/2/9 23:15, Maxime Ripard wrote: > > On Fri, Feb 09, 2024 at 12:02:48PM +0100, Daniel Vetter wrote: > > > On Thu, Feb 08, 2024 at 04:27:02PM +0100, Maxime Ripard wrote: > > > > On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: > > > > > On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: > > > > > > The component helper functions are the glue, which is used to bind multiple > > > > > > GPU cores to a virtual master platform device. Which is fine and works well > > > > > > for the SoCs who contains multiple GPU cores. > > > > > > > > > > > > The problem is that usperspace programs (such as X server and Mesa) will > > > > > > search the PCIe device to use if it is exist. In other words, usperspace > > > > > > programs open the PCIe device with higher priority. Creating a virtual > > > > > > master platform device for PCI(e) GPUs is unnecessary, as the PCI device > > > > > > has been created by the time drm/etnaviv is loaded. > > > > > > > > > > > > we create virtual platform devices as a representation for the vivante GPU > > > > > > ip core. As all of subcomponent are attached via the PCIe master device, > > > > > > we reflect this hardware layout by binding all of the virtual child to the > > > > > > the real master. > > > > > > > > > > > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > > > > > Uh so my understanding is that drivers really shouldn't create platform > > > > > devices of their own. For this case here I think the aux-bus framework is > > > > > the right thing to use. Alternatively would be some infrastructure where > > > > > you feed a DT tree to driver core or pci subsystem and it instantiates it > > > > > all for you correctly, and especially with hotunplug all done right since > > > > > this is pci now, not actually part of the soc that cannot be hotunplugged. > > > > I don't think we need intermediate platform devices at all. We just need > > > > to register our GPU against the PCI device and that's it. We don't need > > > > a platform device, we don't need the component framework. > > > Afaik that's what this series does. The component stuff is for the > > > internal structure of the gpu ip, so that the same modular approach that > > > works for arm-soc also works for pci chips. > > But there should be a single PCI device, while we have multiple "DT" > > devices, right? Or is there several PCI devices too on that PCI card? > > > There is only a single PCI(e) device on that PCI(e) card, this single > PCI(e) device is selected as the component master. All other Hardware IP > blocks are shipped by the single PCI(e) master. It may includes Display > controllers, GPUs, video decoders, HDMI display bridges hardware unit etc. > > But all of those Hardware IP share the same MMIO registers PCI BAR, this > PCI BAR is a kind of PCI(e) MEM resource. It is a relative *big* chunk, > as large as 32MB in address ranges for the JingJia Macro dGPU. Therefore, > I break the whole registers memory(MMIO) resource into smaller pieces by > creating platform device manually, manually created platform device is > called as virtual child in this series. > > In short, we cut the whole into smaller piece, each smaller piece is a > single hardware IP block, thus deserve a single device driver. We will > have multiple platform devices if the dGPU contains multiple hardware > IP block. On the driver side, we bind all of the scattered driver module > with component. That's kind of my point then. If there's a single device, there's no need to create intermediate devices and use the component framework to tie them all together. You can have a simpler approach where you create a function that takes the memory area it operates on (and whatever additional resource it needs: interrupt, clocks, etc.) and call that directly from the PCIe device probe, and the MMIO device bind. Maxime
Hi, On 2024/2/13 22:38, Maxime Ripard wrote: > On Sat, Feb 10, 2024 at 12:25:33AM +0800, Sui Jingfeng wrote: >> On 2024/2/9 23:15, Maxime Ripard wrote: >>> On Fri, Feb 09, 2024 at 12:02:48PM +0100, Daniel Vetter wrote: >>>> On Thu, Feb 08, 2024 at 04:27:02PM +0100, Maxime Ripard wrote: >>>>> On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: >>>>>> On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: >>>>>>> The component helper functions are the glue, which is used to bind multiple >>>>>>> GPU cores to a virtual master platform device. Which is fine and works well >>>>>>> for the SoCs who contains multiple GPU cores. >>>>>>> >>>>>>> The problem is that usperspace programs (such as X server and Mesa) will >>>>>>> search the PCIe device to use if it is exist. In other words, usperspace >>>>>>> programs open the PCIe device with higher priority. Creating a virtual >>>>>>> master platform device for PCI(e) GPUs is unnecessary, as the PCI device >>>>>>> has been created by the time drm/etnaviv is loaded. >>>>>>> >>>>>>> we create virtual platform devices as a representation for the vivante GPU >>>>>>> ip core. As all of subcomponent are attached via the PCIe master device, >>>>>>> we reflect this hardware layout by binding all of the virtual child to the >>>>>>> the real master. >>>>>>> >>>>>>> Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> >>>>>> Uh so my understanding is that drivers really shouldn't create platform >>>>>> devices of their own. For this case here I think the aux-bus framework is >>>>>> the right thing to use. Alternatively would be some infrastructure where >>>>>> you feed a DT tree to driver core or pci subsystem and it instantiates it >>>>>> all for you correctly, and especially with hotunplug all done right since >>>>>> this is pci now, not actually part of the soc that cannot be hotunplugged. >>>>> I don't think we need intermediate platform devices at all. We just need >>>>> to register our GPU against the PCI device and that's it. We don't need >>>>> a platform device, we don't need the component framework. >>>> Afaik that's what this series does. The component stuff is for the >>>> internal structure of the gpu ip, so that the same modular approach that >>>> works for arm-soc also works for pci chips. >>> But there should be a single PCI device, while we have multiple "DT" >>> devices, right? Or is there several PCI devices too on that PCI card? >> >> There is only a single PCI(e) device on that PCI(e) card, this single >> PCI(e) device is selected as the component master. All other Hardware IP >> blocks are shipped by the single PCI(e) master. It may includes Display >> controllers, GPUs, video decoders, HDMI display bridges hardware unit etc. >> >> But all of those Hardware IP share the same MMIO registers PCI BAR, this >> PCI BAR is a kind of PCI(e) MEM resource. It is a relative *big* chunk, >> as large as 32MB in address ranges for the JingJia Macro dGPU. Therefore, >> I break the whole registers memory(MMIO) resource into smaller pieces by >> creating platform device manually, manually created platform device is >> called as virtual child in this series. >> >> In short, we cut the whole into smaller piece, each smaller piece is a >> single hardware IP block, thus deserve a single device driver. We will >> have multiple platform devices if the dGPU contains multiple hardware >> IP block. On the driver side, we bind all of the scattered driver module >> with component. > That's kind of my point then. If there's a single device, there's no > need to create intermediate devices and use the component framework to > tie them all together. You can have a simpler approach where you create > a function that takes the memory area it operates on (and whatever > additional resource it needs: interrupt, clocks, etc.) and call that > directly from the PCIe device probe, and the MMIO device bind. Yes, you are right. I have implemented the method just as you told me at V12 of this series (see 0004 patch at [1]). But at V13, I suddenly realized that the component code path plus(+) non-component code path yield a "side-by-side" implement. The old non-component approach can not support binding multiple sub-core, it can only support one Vivante GPU IP core case. But there are dGPU which shipping two identical core. While, the component-based approach implemented at this version is the most universal and the most flexible and extensible. We could bind a virtual display driver and/or a real display driver to the real master (refer to the PCI(e) device). The PCI(e) device is responsible for present the DRM service to userspace, like a leader or agent. All other sub hardware and software are hiding behind of the master. Besides, Lucas asked me implement the driver like this way at V10 (see [2]) Lucas said: "My favorite option would be to just always use the component path even when the GPU is on a PCI device to keep both paths mostly aligned. One could easily image both a 3D and a 2D core being made available though the same PCI device." Lucas, are you watching? How about this version? Can you help to review please? [1] https://patchwork.freedesktop.org/series/127084/ [2] https://lore.kernel.org/dri-devel/0f1095ef333da7ea103486a1121ca9038815e57c.camel@pengutronix.de/ > Maxime
Hi, On 2024/2/9 19:02, Daniel Vetter wrote: > On Thu, Feb 08, 2024 at 04:27:02PM +0100, Maxime Ripard wrote: >> On Wed, Feb 07, 2024 at 10:35:49AM +0100, Daniel Vetter wrote: >>> On Wed, Feb 07, 2024 at 01:27:59AM +0800, Sui Jingfeng wrote: >>>> The component helper functions are the glue, which is used to bind multiple >>>> GPU cores to a virtual master platform device. Which is fine and works well >>>> for the SoCs who contains multiple GPU cores. >>>> >>>> The problem is that usperspace programs (such as X server and Mesa) will >>>> search the PCIe device to use if it is exist. In other words, usperspace >>>> programs open the PCIe device with higher priority. Creating a virtual >>>> master platform device for PCI(e) GPUs is unnecessary, as the PCI device >>>> has been created by the time drm/etnaviv is loaded. >>>> >>>> we create virtual platform devices as a representation for the vivante GPU >>>> ip core. As all of subcomponent are attached via the PCIe master device, >>>> we reflect this hardware layout by binding all of the virtual child to the >>>> the real master. >>>> >>>> Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> >>> Uh so my understanding is that drivers really shouldn't create platform >>> devices of their own. For this case here I think the aux-bus framework is >>> the right thing to use. Alternatively would be some infrastructure where >>> you feed a DT tree to driver core or pci subsystem and it instantiates it >>> all for you correctly, and especially with hotunplug all done right since >>> this is pci now, not actually part of the soc that cannot be hotunplugged. >> I don't think we need intermediate platform devices at all. We just need >> to register our GPU against the PCI device and that's it. We don't need >> a platform device, we don't need the component framework. > Afaik that's what this series does. The component stuff is for the > internal structure of the gpu ip, so that the same modular approach that > works for arm-soc also works for pci chips. > > Otherwise we end up with each driver hand-rolling that stuff, which is > defacto what both nouveau and amdgpu do (intel hw is too much a mess for > that component-driver based approach to actually work reasonably well). Emmm, I spend years to achieve this, and only to find that you have fully understand my patch within two days. > Cheers, Sima > >>> I think I've seen some other pci devices from arm soc designs that would >>> benefit from this too, so lifting this logic into a pci function would >>> make sense imo. Yes, as you said, we are trying to avoid the hand-rolling stuff. I guess, extremely advanced drivers(like i915, amdgpu, and nouveau) won't need this. So I'm not sure if lifting this logic into PCI core would benefit to other people. While investigating the aux-bus framework a few days, I find it is not as concise as this one. It introduce a lot of new structure, I fear that it may cause namespace pollution if adopt it. So, I thinks I should choose the alternative way. While taking a lot from contribution, we are really want to do some feedback(pay-back). We are happy if there are other users(or new drivers) would like to adopt this idea, I think, the idea itself has already been conveyed. Which probably can be seen as a trivial contribution. Other programmer are free to copy and modify. But as a initial commit, I minimized the mount of changes as required by Locus. meanwhile, I'm willing to following the expectation in the long term. If there are other users or other problem need to solve, I would like help to improve and to cooperate to testing in the future.
diff --git a/drivers/gpu/drm/etnaviv/Kconfig b/drivers/gpu/drm/etnaviv/Kconfig index faa7fc68b009..38c251585ec1 100644 --- a/drivers/gpu/drm/etnaviv/Kconfig +++ b/drivers/gpu/drm/etnaviv/Kconfig @@ -15,6 +15,14 @@ config DRM_ETNAVIV help DRM driver for Vivante GPUs. +config DRM_ETNAVIV_PCI_DRIVER + bool "enable ETNAVIV PCI driver support" + depends on DRM_ETNAVIV + depends on PCI + help + Compile in support for Vivante GPUs attached via PCI(e). + Say Y if you have such hardwares. + config DRM_ETNAVIV_THERMAL bool "enable ETNAVIV thermal throttling" depends on DRM_ETNAVIV diff --git a/drivers/gpu/drm/etnaviv/Makefile b/drivers/gpu/drm/etnaviv/Makefile index 46e5ffad69a6..6829e1ebf2db 100644 --- a/drivers/gpu/drm/etnaviv/Makefile +++ b/drivers/gpu/drm/etnaviv/Makefile @@ -16,4 +16,6 @@ etnaviv-y := \ etnaviv_perfmon.o \ etnaviv_sched.o +etnaviv-$(CONFIG_DRM_ETNAVIV_PCI_DRIVER) += etnaviv_pci_drv.o + obj-$(CONFIG_DRM_ETNAVIV) += etnaviv.o diff --git a/drivers/gpu/drm/etnaviv/etnaviv_drv.c b/drivers/gpu/drm/etnaviv/etnaviv_drv.c index 5f65f2dead44..48e2402adbe3 100644 --- a/drivers/gpu/drm/etnaviv/etnaviv_drv.c +++ b/drivers/gpu/drm/etnaviv/etnaviv_drv.c @@ -24,6 +24,7 @@ #include "etnaviv_gpu.h" #include "etnaviv_gem.h" #include "etnaviv_mmu.h" +#include "etnaviv_pci_drv.h" #include "etnaviv_perfmon.h" /* @@ -566,6 +567,10 @@ static int etnaviv_bind(struct device *dev) if (ret < 0) goto out_free_priv; + ret = etnaviv_register_irq_handler(dev, priv); + if (ret) + goto out_unbind; + load_gpu(drm); ret = drm_dev_register(drm, 0); @@ -594,7 +599,7 @@ static void etnaviv_unbind(struct device *dev) etnaviv_private_fini(priv); } -static const struct component_master_ops etnaviv_master_ops = { +const struct component_master_ops etnaviv_master_ops = { .bind = etnaviv_bind, .unbind = etnaviv_unbind, }; @@ -740,6 +745,10 @@ static int __init etnaviv_init(void) if (ret != 0) goto unregister_gpu_driver; + ret = etnaviv_register_pci_driver(); + if (ret != 0) + goto unregister_platform_driver; + /* * If the DT contains at least one available GPU device, instantiate * the DRM platform device. @@ -769,6 +778,7 @@ module_init(etnaviv_init); static void __exit etnaviv_exit(void) { etnaviv_destroy_platform_device(&etnaviv_drm); + etnaviv_unregister_pci_driver(); platform_driver_unregister(&etnaviv_platform_driver); platform_driver_unregister(&etnaviv_gpu_driver); } diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c index 3fd637c17797..221020e98900 100644 --- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c +++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c @@ -10,6 +10,7 @@ #include <linux/dma-mapping.h> #include <linux/mod_devicetable.h> #include <linux/module.h> +#include <linux/pci.h> #include <linux/platform_device.h> #include <linux/pm_runtime.h> #include <linux/regulator/consumer.h> @@ -29,6 +30,7 @@ static const struct platform_device_id gpu_ids[] = { { .name = "etnaviv-gpu,2d" }, + { .name = "etnaviv-gpu,3d" }, { }, }; @@ -1555,14 +1557,22 @@ static void dump_mmu_fault(struct etnaviv_gpu *gpu) static irqreturn_t irq_handler(int irq, void *data) { - struct etnaviv_gpu *gpu = data; + struct etnaviv_drm_private *priv = data; irqreturn_t ret = IRQ_NONE; + int i; - u32 intr = gpu_read(gpu, VIVS_HI_INTR_ACKNOWLEDGE); - - if (intr != 0) { + for (i = 0; i < priv->num_gpus; i++) { + struct etnaviv_gpu *gpu = priv->gpu[i]; + u32 intr; int event; + if (!gpu) + continue; + + intr = gpu_read(gpu, VIVS_HI_INTR_ACKNOWLEDGE); + if (!intr) + continue; + pm_runtime_mark_last_busy(gpu->dev); dev_dbg(gpu->dev, "intr 0x%08x\n", intr); @@ -1893,10 +1903,44 @@ static const struct of_device_id etnaviv_gpu_match[] = { }; MODULE_DEVICE_TABLE(of, etnaviv_gpu_match); +/* + * dev point to the master. For platform device, it is virtual. + * For PCI(e) device, it is real. + */ +int etnaviv_register_irq_handler(struct device *dev, + struct etnaviv_drm_private *priv) +{ + bool is_pci = dev_is_pci(dev); + int ret = 0; + + if (is_pci) { + struct pci_dev *pdev = to_pci_dev(dev); + + ret = request_irq(pdev->irq, irq_handler, IRQF_SHARED, + dev_name(dev), priv); + } else { + int i; + + for (i = 0; i < priv->num_gpus; i++) { + struct etnaviv_gpu *gpu = priv->gpu[i]; + + ret = devm_request_irq(gpu->dev, gpu->irq, irq_handler, + 0, dev_name(dev), priv); + if (ret) { + dev_err(dev, "failed to request IRQ handler: %d\n", ret); + break; + } + } + } + + return ret; +} + static int etnaviv_gpu_platform_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; struct etnaviv_gpu *gpu; + bool is_pci; int err; gpu = devm_kzalloc(dev, sizeof(*gpu), GFP_KERNEL); @@ -1912,22 +1956,23 @@ static int etnaviv_gpu_platform_probe(struct platform_device *pdev) if (IS_ERR(gpu->mmio)) return PTR_ERR(gpu->mmio); + is_pci = dev->parent ? dev_is_pci(dev->parent) : false; + /* Get Interrupt: */ - gpu->irq = platform_get_irq(pdev, 0); + if (is_pci) + gpu->irq = to_pci_dev(dev->parent)->irq; + else + gpu->irq = platform_get_irq(pdev, 0); + if (gpu->irq < 0) return gpu->irq; - err = devm_request_irq(dev, gpu->irq, irq_handler, 0, - dev_name(dev), gpu); - if (err) { - dev_err(dev, "failed to request IRQ%u: %d\n", gpu->irq, err); - return err; - } - /* Get Clocks: */ - err = etnaviv_gpu_clk_get(gpu); - if (err) - return err; + if (!is_pci) { + err = etnaviv_gpu_clk_get(gpu); + if (err) + return err; + } /* TODO: figure out max mapped size */ dev_set_drvdata(dev, gpu); diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h index 7d5e9158e13c..d354e096652c 100644 --- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h +++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h @@ -208,6 +208,10 @@ static inline u32 gpu_read_power(struct etnaviv_gpu *gpu, u32 reg) int etnaviv_gpu_get_param(struct etnaviv_gpu *gpu, u32 param, u64 *value); int etnaviv_gpu_init(struct etnaviv_gpu *gpu); + +int etnaviv_register_irq_handler(struct device *dev, + struct etnaviv_drm_private *priv); + bool etnaviv_fill_identity_from_hwdb(struct etnaviv_gpu *gpu); #ifdef CONFIG_DEBUG_FS diff --git a/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c new file mode 100644 index 000000000000..ec5039866eda --- /dev/null +++ b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c @@ -0,0 +1,224 @@ +// SPDX-License-Identifier: GPL-2.0 + +#include <linux/component.h> +#include <linux/pci.h> + +#include "etnaviv_drv.h" +#include "etnaviv_pci_drv.h" + +enum etnaviv_pci_gpu_chip_id { + GC_CORE_UNKNOWN = 0, + JM9100 = 1, + JD9230P = 2, + GP102 = 3, + GC1000_IN_LS7A1000 = 4, + GC1000_IN_LS2K1000 = 5, + GC_CORE_PCI_LAST, +}; + +struct etnaviv_pci_gpu_data { + enum etnaviv_pci_gpu_chip_id chip_id; + u32 num_core; + u32 num_vram; + u32 vram_bars[2]; + u32 mmio_bar; + struct { + u32 id; + u32 offset; + u32 size; + char compatible[20]; + } cores[ETNA_MAX_PIPES]; + + bool has_dedicated_vram; + char market_name[24]; +}; + +static const struct etnaviv_pci_gpu_data +gc_core_plaform_data[GC_CORE_PCI_LAST] = { + { + .chip_id = GC_CORE_UNKNOWN, + }, + { + .chip_id = JM9100, + .num_core = 1, + .num_vram = 2, + .vram_bars = {0, 2}, + .mmio_bar = 1, + .cores = {{0, 0x00900000, 0x00010000, "etnaviv-gpu,3d"},}, + .has_dedicated_vram = true, + .market_name = "JingJia Micro JM9100", + }, + { + .chip_id = JD9230P, + .num_core = 2, + .num_vram = 2, + .vram_bars = {0, 2}, + .mmio_bar = 1, + .cores = {{0, 0x00900000, 0x00010000, "etnaviv-gpu,3d"}, + {1, 0x00910000, 0x00010000, "etnaviv-gpu,3d"},}, + .has_dedicated_vram = true, + .market_name = "JingJia Micro JD9230P", + }, + { + .chip_id = GP102, + .num_core = 2, + .num_vram = 1, + .vram_bars = {0,}, + .mmio_bar = 2, + .cores = {{0, 0x00040000, 0x00010000, "etnaviv-gpu,3d"}, + {0, 0x000C0000, 0x00010000, "etnaviv-gpu,2d"},}, + .has_dedicated_vram = true, + .market_name = "LingJiu GP102", + }, + { + .chip_id = GC1000_IN_LS7A1000, + .num_core = 1, + .num_vram = 1, + .vram_bars = {2,}, + .mmio_bar = 0, + .cores = {{0, 0, 0x00010000, "etnaviv-gpu,3d"}, {}, {}, {}}, + .has_dedicated_vram = true, + .market_name = "GC1000 in LS7A1000", + }, + { + .chip_id = GC1000_IN_LS2K1000, + .num_core = 1, + .num_vram = 0, + .mmio_bar = 0, + .cores = {{0, 0, 0x00010000, "etnaviv-gpu,3d"}, {}, {}, {}}, + .has_dedicated_vram = false, + .market_name = "GC1000 in LS2K1000", + }, +}; + +static const struct etnaviv_pci_gpu_data * +etnaviv_pci_get_platform_data(const struct pci_device_id *entity) +{ + enum etnaviv_pci_gpu_chip_id chip_id = entity->driver_data; + static const struct etnaviv_pci_gpu_data *pdata; + + pdata = &gc_core_plaform_data[chip_id]; + if (!pdata || pdata->chip_id == GC_CORE_UNKNOWN) + return NULL; + + return pdata; +} + +extern const struct component_master_ops etnaviv_master_ops; + +static int etnaviv_pci_probe(struct pci_dev *pdev, + const struct pci_device_id *ent) +{ + const struct etnaviv_pci_gpu_data *pdata; + struct device *dev = &pdev->dev; + struct component_match *matches = NULL; + unsigned int i; + unsigned int num_core; + int ret; + + ret = pcim_enable_device(pdev); + if (ret) { + dev_err(dev, "failed to enable\n"); + return ret; + } + + pci_set_master(pdev); + + ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); + if (ret) + return ret; + + pdata = etnaviv_pci_get_platform_data(ent); + if (!pdata) + return -ENODEV; + + num_core = pdata->num_core; + + dev_info(dev, "%s has %u GPU cores\n", pdata->market_name, num_core); + + /* + * Create a virtual platform device for the sub-component, + * a sub-component is refer to a single vivante GPU core. + * But it can also be extended to stand for a display controller + * or any other IP core attached via the same PCIe master. + */ + for (i = 0; i < num_core; i++) { + struct platform_device *virtual_child; + resource_size_t start, offset, size; + struct resource res; + + start = pci_resource_start(pdev, pdata->mmio_bar); + offset = pdata->cores[i].offset; + size = pdata->cores[i].size; + + memset(&res, 0, sizeof(res)); + res.flags = IORESOURCE_MEM; + res.name = "reg"; + res.start = start + offset; + res.end = start + offset + size - 1; + + ret = etnaviv_create_platform_device(dev, + pdata->cores[i].compatible, + pdata->cores[i].id, + &res, + (void *)pdata, + &virtual_child); + if (ret) + return ret; + + component_match_add(dev, &matches, component_compare_dev, + &virtual_child->dev); + } + + ret = component_master_add_with_match(dev, &etnaviv_master_ops, matches); + + return ret; +} + +static int platform_device_remove_callback(struct device *dev, void *data) +{ + struct platform_device *pdev = to_platform_device(dev); + + etnaviv_destroy_platform_device(&pdev); + + return 0; +} + +static void etnaviv_pci_remove(struct pci_dev *pdev) +{ + struct device *dev = &pdev->dev; + + component_master_del(dev, &etnaviv_master_ops); + + device_for_each_child(dev, NULL, platform_device_remove_callback); + + pci_clear_master(pdev); +} + +static const struct pci_device_id etnaviv_pci_id_lists[] = { + {0x0731, 0x9100, PCI_ANY_ID, PCI_ANY_ID, 0, 0, JM9100}, + {0x0731, 0x9230, PCI_ANY_ID, PCI_ANY_ID, 0, 0, JD9230P}, + {0x0709, 0x0001, PCI_ANY_ID, PCI_ANY_ID, 0, 0, GP102}, + {0x0014, 0x7A15, PCI_ANY_ID, PCI_ANY_ID, 0, 0, GC1000_IN_LS7A1000}, + {0x0014, 0x7A05, PCI_ANY_ID, PCI_ANY_ID, 0, 0, GC1000_IN_LS2K1000}, + { } +}; + +static struct pci_driver etnaviv_pci_driver = { + .name = "etnaviv", + .id_table = etnaviv_pci_id_lists, + .probe = etnaviv_pci_probe, + .remove = etnaviv_pci_remove, +}; + +int etnaviv_register_pci_driver(void) +{ + return pci_register_driver(&etnaviv_pci_driver); +} + +void etnaviv_unregister_pci_driver(void) +{ + pci_unregister_driver(&etnaviv_pci_driver); +} + +MODULE_DEVICE_TABLE(pci, etnaviv_pci_id_lists); diff --git a/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h new file mode 100644 index 000000000000..1db559ee5e9b --- /dev/null +++ b/drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h @@ -0,0 +1,18 @@ +/* SPDX-License-Identifier: GPL-2.0 */ + +#ifndef __ETNAVIV_PCI_DRV_H__ +#define __ETNAVIV_PCI_DRV_H__ + +#ifdef CONFIG_DRM_ETNAVIV_PCI_DRIVER + +int etnaviv_register_pci_driver(void); +void etnaviv_unregister_pci_driver(void); + +#else + +static inline int etnaviv_register_pci_driver(void) { return 0; } +static inline void etnaviv_unregister_pci_driver(void) { } + +#endif + +#endif
The component helper functions are the glue, which is used to bind multiple GPU cores to a virtual master platform device. Which is fine and works well for the SoCs who contains multiple GPU cores. The problem is that usperspace programs (such as X server and Mesa) will search the PCIe device to use if it is exist. In other words, usperspace programs open the PCIe device with higher priority. Creating a virtual master platform device for PCI(e) GPUs is unnecessary, as the PCI device has been created by the time drm/etnaviv is loaded. we create virtual platform devices as a representation for the vivante GPU ip core. As all of subcomponent are attached via the PCIe master device, we reflect this hardware layout by binding all of the virtual child to the the real master. Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> --- drivers/gpu/drm/etnaviv/Kconfig | 8 + drivers/gpu/drm/etnaviv/Makefile | 2 + drivers/gpu/drm/etnaviv/etnaviv_drv.c | 12 +- drivers/gpu/drm/etnaviv/etnaviv_gpu.c | 75 ++++++-- drivers/gpu/drm/etnaviv/etnaviv_gpu.h | 4 + drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c | 224 ++++++++++++++++++++++ drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h | 18 ++ 7 files changed, 327 insertions(+), 16 deletions(-) create mode 100644 drivers/gpu/drm/etnaviv/etnaviv_pci_drv.c create mode 100644 drivers/gpu/drm/etnaviv/etnaviv_pci_drv.h