Message ID | 20220215053844.7119-4-Frank.Li@nxp.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | NTB function for PCIe RC to EP connection | expand |
[+cc Jon, Dave, Allen, linux-ntb] This patch only touches drivers/pci; the subject line should reflect that to help people know which things to look at. Maybe something like: PCI: endpoint: Support NTB transfer between RC and EP On Mon, Feb 14, 2022 at 11:38:43PM -0600, Frank Li wrote: > Add NTB function driver and virtual PCI BUS and Virtual NTB driver > to implement communication between PCIe RC and PCIe EP devices s/PCI BUS/PCI Bus/ s/PCIe RC/PCIe Root Port/ (I think; the RC itself doesn't have a software representation) > ┌────────────┐ ┌─────────────────────────────────────┐ > │ │ │ │ > ├────────────┤ │ ┌──────────────┤ > │ NTB │ │ │ NTB │ > │ NetDev │ │ │ NetDev │ > ├────────────┤ │ ├──────────────┤ > │ NTB │ │ │ NTB │ > │ Transfer │ │ │ Transfer │ > ├────────────┤ │ ├──────────────┤ > │ │ │ │ │ > │ PCI NTB │ │ │ │ > │ EPF │ │ │ │ > │ Driver │ │ │ PCI Virtual │ > │ │ ├───────────────┐ │ NTB Driver │ > │ │ │ PCI EP NTB │◄────►│ │ > │ │ │ FN Driver │ │ │ > ├────────────┤ ├───────────────┤ ├──────────────┤ > │ │ │ │ │ │ > │ PCI BUS │ ◄─────► │ PCI EP BUS │ │ Virtual PCI │ > │ │ PCI │ │ │ BUS │ > └────────────┘ └───────────────┴──────┴──────────────┘ s/PCI BUS/PCI Bus/ > PCI RC PCI EP s/PCI RC/PCIe Root Port/ ? > This driver include 3 part: > 1 PCI EP NTB function driver > 2 Virtual PCI bus > 3 PCI virutal NTB driver, which is loaded only by above virtual pci bus s/include 3 part/includes 3 parts/ s/virutal/virtual/ s/pci bus/PCI bus/ > Signed-off-by: Frank Li <Frank.Li@nxp.com> > --- > drivers/pci/endpoint/functions/Kconfig | 11 + > drivers/pci/endpoint/functions/Makefile | 1 + > drivers/pci/endpoint/functions/pci-epf-vntb.c | 1425 +++++++++++++++++ > 3 files changed, 1437 insertions(+) > create mode 100644 drivers/pci/endpoint/functions/pci-epf-vntb.c > > diff --git a/drivers/pci/endpoint/functions/Kconfig b/drivers/pci/endpoint/functions/Kconfig > index 5f1242ca2f4e4..362555b024e8f 100644 > --- a/drivers/pci/endpoint/functions/Kconfig > +++ b/drivers/pci/endpoint/functions/Kconfig > @@ -25,3 +25,14 @@ config PCI_EPF_NTB > device tree. > > If in doubt, say "N" to disable Endpoint NTB driver. > + > +config PCI_EPF_VNTB > + tristate "PCI Endpoint NTB driver" > + depends on PCI_ENDPOINT > + select CONFIGFS_FS > + help > + Select this configuration option to enable the Non-Transparent > + Bridge (NTB) driver for PCI Endpoint. NTB driver implements NTB > + between PCI host and PCIe Endpoint. "PCI Endpoint" vs "PCIe Endpoint". Pick one. It seems like you require an upstream bridge, i.e., a Root Port, so maybe "PCIe Root Port" and "PCIe Endpoint" is what you want? > + If in doubt, say "N" to disable Endpoint NTB driver. > diff --git a/drivers/pci/endpoint/functions/Makefile b/drivers/pci/endpoint/functions/Makefile > index 96ab932a537a2..5c13001deaba1 100644 > --- a/drivers/pci/endpoint/functions/Makefile > +++ b/drivers/pci/endpoint/functions/Makefile > @@ -5,3 +5,4 @@ > > obj-$(CONFIG_PCI_EPF_TEST) += pci-epf-test.o > obj-$(CONFIG_PCI_EPF_NTB) += pci-epf-ntb.o > +obj-$(CONFIG_PCI_EPF_VNTB) += pci-epf-vntb.o > diff --git a/drivers/pci/endpoint/functions/pci-epf-vntb.c b/drivers/pci/endpoint/functions/pci-epf-vntb.c > new file mode 100644 > index 0000000000000..ebf7e243eefa4 > --- /dev/null > +++ b/drivers/pci/endpoint/functions/pci-epf-vntb.c > @@ -0,0 +1,1425 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Endpoint Function Driver to implement Non-Transparent Bridge functionality > + * Between PCI RC and EP > + * > + * Copyright (C) 2020 Texas Instruments > + * Copyright (C) 2022 NXP > + * > + * Based on pci-epf-ntb.c > + * Author: Frank Li <Frank.Li@nxp.com> > + * Author: Kishon Vijay Abraham I <kishon@ti.com> > + */ > + > +/** > + * +------------+ +---------------------------------------+ > + * | | | | > + * +------------+ | +--------------+ > + * | NTB | | | NTB | > + * | NetDev | | | NetDev | > + * +------------+ | +--------------+ > + * | NTB | | | NTB | > + * | Transfer | | | Transfer | > + * +------------+ | +--------------+ > + * | | | | | > + * | PCI NTB | | | | > + * | EPF | | | | > + * | Driver | | | PCI Virtual | > + * | | +---------------+ | NTB Driver | > + * | | | PCI EP NTB |<------>| | > + * | | | FN Driver | | | > + * +------------+ +---------------+ +--------------+ > + * | | | | | | > + * | PCI BUS | <-----> | PCI EP BUS | | Virtual PCI | > + * | | PCI | | | BUS | > + * +------------+ +---------------+--------+--------------+ > + * PCI RC PCI EP s/PCI RC/PCIe Root Port/ ? > + */ > + > +#include <linux/delay.h> > +#include <linux/io.h> > +#include <linux/module.h> > +#include <linux/slab.h> > + > +#include <linux/pci-epc.h> > +#include <linux/pci-epf.h> > +#include <linux/ntb.h> > + > +static struct workqueue_struct *kpcintb_workqueue; > + > +#define COMMAND_CONFIGURE_DOORBELL 1 > +#define COMMAND_TEARDOWN_DOORBELL 2 > +#define COMMAND_CONFIGURE_MW 3 > +#define COMMAND_TEARDOWN_MW 4 > +#define COMMAND_LINK_UP 5 > +#define COMMAND_LINK_DOWN 6 > + > +#define COMMAND_STATUS_OK 1 > +#define COMMAND_STATUS_ERROR 2 > + > +#define LINK_STATUS_UP BIT(0) > + > +#define SPAD_COUNT 64 > +#define DB_COUNT 4 > +#define NTB_MW_OFFSET 2 > +#define DB_COUNT_MASK GENMASK(15, 0) > +#define MSIX_ENABLE BIT(16) > +#define MAX_DB_COUNT 32 > +#define MAX_MW 4 > + > +#define VNTB_VID 0x1957 This looks like PCI_VENDOR_ID_NXP (or PCI_VENDOR_ID_FREESCALE). How did you choose that, why is it safe to use, and why don't use use PCI_VENDOR_ID_NXP instead? > +#define VNTB_PID 0x080A > + > +enum epf_ntb_bar { > + BAR_CONFIG, > + BAR_DB, > + BAR_MW0, > + BAR_MW1, > + BAR_MW2, > +}; > + > +/* > + * +--------------------------------------------------+ Base > + * | | > + * | | > + * | | > + * | Common Control Register | > + * | | > + * | | > + * | | > + * +-----------------------+--------------------------+ Base+span_offset > + * | | | > + * | Peer Span Space | Span Space | > + * | | | > + * | | | > + * +-----------------------+--------------------------+ Base+span_offset > + * | | | +span_count * 4 > + * | | | > + * | Span Space | Peer Span Space | > + * | | | > + * +-----------------------+--------------------------+ > + * Virtual PCI Pcie Endpoint > + * NTB Driver NTB Driver s/Pcie/PCIe/ > + */ > +struct epf_ntb_ctrl { > + u32 command; > + u32 argument; > + u16 command_status; > + u16 link_status; > + u32 topology; > + u64 addr; > + u64 size; > + u32 num_mws; > + u32 reserved; > + u32 spad_offset; > + u32 spad_count; > + u32 db_entry_size; > + u32 db_data[MAX_DB_COUNT]; > + u32 db_offset[MAX_DB_COUNT]; > +} __packed; > + > +struct epf_ntb { > + struct ntb_dev ntb; > + struct pci_epf *epf; > + struct config_group group; > + > + u32 num_mws; > + u32 db_count; > + u32 spad_count; > + u64 mws_size[MAX_MW]; > + u64 db; > + u32 vbus_number; > + > + bool linkup; > + u32 spad_size; > + > + enum pci_barno epf_ntb_bar[6]; > + > + struct epf_ntb_ctrl *reg; > + > + phys_addr_t epf_db_phy; > + void __iomem *epf_db; > + > + phys_addr_t vpci_mw_phy[MAX_MW]; > + void __iomem *vpci_mw_addr[MAX_MW]; > + > + struct delayed_work cmd_handler; > +}; > + > +#define to_epf_ntb(epf_group) container_of((epf_group), struct epf_ntb, group) > +#define ntb_ndev(__ntb) container_of(__ntb, struct epf_ntb, ntb) > + > +static struct pci_epf_header epf_ntb_header = { > + .vendorid = PCI_ANY_ID, > + .deviceid = PCI_ANY_ID, > + .baseclass_code = PCI_BASE_CLASS_MEMORY, > + .interrupt_pin = PCI_INTERRUPT_INTA, > +}; > + > +/** > + * epf_ntb_link_up() - Raise link_up interrupt to Virtual Host > + * @ntb: NTB device that facilitates communication between HOST and VHOST > + * @link_up: true or false indicating Link is UP or Down > + * > + * Once NTB function in HOST invoke ntb_link_enable(), > + * this NTB function driver will trigger a link event to vhost. > + */ > +static int epf_ntb_link_up(struct epf_ntb *ntb, bool link_up) > +{ > + if (link_up) > + ntb->reg->link_status |= LINK_STATUS_UP; > + else > + ntb->reg->link_status &= ~LINK_STATUS_UP; > + > + ntb_link_event(&ntb->ntb); > + return 0; > +} > + > +/** > + * epf_ntb_configure_mw() - Configure the Outbound Address Space for vhost > + * to access the memory window of host > + * @ntb: NTB device that facilitates communication between host and vhost > + * @mw: Index of the memory window (either 0, 1, 2 or 3) > + * > + * EP Outbound Window > + * +--------+ +-----------+ > + * | | | | > + * | | | | > + * | | | | > + * | | | | > + * | | +-----------+ > + * | Virtual| | Memory Win| > + * | NTB | -----------> | | > + * | Driver | | | > + * | | +-----------+ > + * | | | | > + * | | | | > + * +--------+ +-----------+ > + * VHost PCI EP > + */ > +static int epf_ntb_configure_mw(struct epf_ntb *ntb, u32 mw) > +{ > + phys_addr_t phys_addr; > + u8 func_no, vfunc_no; > + u64 addr, size; > + int ret = 0; > + > + phys_addr = ntb->vpci_mw_phy[mw]; > + addr = ntb->reg->addr; > + size = ntb->reg->size; > + > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + ret = pci_epc_map_addr(ntb->epf->epc, func_no, vfunc_no, phys_addr, addr, size); > + if (ret) > + dev_err(&ntb->epf->epc->dev, > + "intf: Failed to map memory window %d address\n", mw); Lots of these messages start with "intf:". What is that telling us? Is it useful? > + return ret; > +} > + > +/** > + * epf_ntb_teardown_mw() - Teardown the configured OB ATU > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * @mw: Index of the memory window (either 0, 1, 2 or 3) > + * > + * Teardown the configured OB ATU configured in epf_ntb_configure_mw() using > + * pci_epc_unmap_addr() > + */ > +static void epf_ntb_teardown_mw(struct epf_ntb *ntb, u32 mw) > +{ > + pci_epc_unmap_addr(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + ntb->vpci_mw_phy[mw]); > +} > + > +/** > + * epf_ntb_cmd_handler() - Handle commands provided by the NTB Host > + * @work: work_struct for the epf_ntb_epc > + * > + * Workqueue function that gets invoked for the two epf_ntb_epc > + * periodically (once every 5ms) to see if it has received any commands > + * from NTB host. The host can send commands to configure doorbell or > + * configure memory window or to update link status. > + */ > +static void epf_ntb_cmd_handler(struct work_struct *work) > +{ > + struct epf_ntb_ctrl *ctrl; > + u32 command, argument; > + struct epf_ntb *ntb; > + struct device *dev; > + int ret; > + int i; > + > + ntb = container_of(work, struct epf_ntb, cmd_handler.work); > + > + for (i = 1; i < ntb->db_count; i++) { > + if (readl(ntb->epf_db + i * 4)) { > + if (readl(ntb->epf_db + i * 4)) > + ntb->db |= 1 << (i - 1); > + > + ntb_db_event(&ntb->ntb, i); > + writel(0, ntb->epf_db + i * 4); > + } > + } > + > + ctrl = ntb->reg; > + command = ctrl->command; > + if (!command) > + goto reset_handler; > + argument = ctrl->argument; > + > + ctrl->command = 0; > + ctrl->argument = 0; > + > + ctrl = ntb->reg; > + dev = &ntb->epf->dev; > + > + switch (command) { > + case COMMAND_CONFIGURE_DOORBELL: > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_TEARDOWN_DOORBELL: > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_CONFIGURE_MW: > + ret = epf_ntb_configure_mw(ntb, argument); > + if (ret < 0) > + ctrl->command_status = COMMAND_STATUS_ERROR; > + else > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_TEARDOWN_MW: > + epf_ntb_teardown_mw(ntb, argument); > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_LINK_UP: > + ntb->linkup = true; > + ret = epf_ntb_link_up(ntb, true); > + if (ret < 0) > + ctrl->command_status = COMMAND_STATUS_ERROR; > + else > + ctrl->command_status = COMMAND_STATUS_OK; > + goto reset_handler; > + case COMMAND_LINK_DOWN: > + ntb->linkup = false; > + ret = epf_ntb_link_up(ntb, false); > + if (ret < 0) > + ctrl->command_status = COMMAND_STATUS_ERROR; > + else > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + default: > + dev_err(dev, "intf UNKNOWN command: %d\n", command); > + break; > + } > + > +reset_handler: > + queue_delayed_work(kpcintb_workqueue, &ntb->cmd_handler, > + msecs_to_jiffies(5)); > +} > + > +/** > + * epf_ntb_config_sspad_bar_clear() - Clear Config + Self scratchpad BAR > + * @ntb_epc: EPC associated with one of the HOST which holds peer's outbound > + * address. > + * > + * Clear BAR0 of EP CONTROLLER 1 which contains the HOST1's config and > + * self scratchpad region (removes inbound ATU configuration). While BAR0 is > + * the default self scratchpad BAR, an NTB could have other BARs for self > + * scratchpad (because of reserved BARs). This function can get the exact BAR > + * used for self scratchpad from epf_ntb_bar[BAR_CONFIG]. > + * > + * Please note the self scratchpad region and config region is combined to > + * a single region and mapped using the same BAR. Also note HOST2's peer > + * scratchpad is HOST1's self scratchpad. > + */ > +static void epf_ntb_config_sspad_bar_clear(struct epf_ntb *ntb) > +{ > + struct pci_epf_bar *epf_bar; > + enum pci_barno barno; > + > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + epf_bar = &ntb->epf->bar[barno]; > + > + pci_epc_clear_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); > +} > + > +/** > + * epf_ntb_config_sspad_bar_set() - Set Config + Self scratchpad BAR > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + * Map BAR0 of EP CONTROLLER 1 which contains the HOST1's config and > + * self scratchpad region. > + * > + * Please note the self scratchpad region and config region is combined to > + * a single region and mapped using the same BAR. > + */ > +static int epf_ntb_config_sspad_bar_set(struct epf_ntb *ntb) > +{ > + struct pci_epf_bar *epf_bar; > + enum pci_barno barno; > + u8 func_no, vfunc_no; > + struct device *dev; > + int ret; > + > + dev = &ntb->epf->dev; > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + epf_bar = &ntb->epf->bar[barno]; > + > + ret = pci_epc_set_bar(ntb->epf->epc, func_no, vfunc_no, epf_bar); > + if (ret) { > + dev_err(dev, "inft: Config/Status/SPAD BAR set failed\n"); > + return ret; > + } > + return 0; > +} > + > +/** > + * epf_ntb_config_spad_bar_free() - Free the physical memory associated with > + * config + scratchpad region > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + */ > +static void epf_ntb_config_spad_bar_free(struct epf_ntb *ntb) > +{ > + enum pci_barno barno; > + > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + pci_epf_free_space(ntb->epf, ntb->reg, barno, 0); > +} > + > +/** > + * epf_ntb_config_spad_bar_alloc() - Allocate memory for config + scratchpad > + * region > + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 > + * > + * Allocate the Local Memory mentioned in the above diagram. The size of > + * CONFIG REGION is sizeof(struct epf_ntb_ctrl) and size of SCRATCHPAD REGION > + * is obtained from "spad-count" configfs entry. > + */ > +static int epf_ntb_config_spad_bar_alloc(struct epf_ntb *ntb) > +{ > + size_t align; > + enum pci_barno barno; > + struct epf_ntb_ctrl *ctrl; > + u32 spad_size, ctrl_size; > + u64 size; > + struct pci_epf *epf = ntb->epf; > + struct device *dev = &epf->dev; > + u32 spad_count; > + void *base; > + int i; > + const struct pci_epc_features *epc_features = pci_epc_get_features(epf->epc, > + epf->func_no, > + epf->vfunc_no); > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + size = epc_features->bar_fixed_size[barno]; > + align = epc_features->align; > + > + if ((!IS_ALIGNED(size, align))) > + return -EINVAL; > + > + spad_count = ntb->spad_count; > + > + ctrl_size = sizeof(struct epf_ntb_ctrl); > + spad_size = 2 * spad_count * 4; > + > + if (!align) { > + ctrl_size = roundup_pow_of_two(ctrl_size); > + spad_size = roundup_pow_of_two(spad_size); > + } else { > + ctrl_size = ALIGN(ctrl_size, align); > + spad_size = ALIGN(spad_size, align); > + } > + > + if (!size) > + size = ctrl_size + spad_size; > + else if (size < ctrl_size + spad_size) > + return -EINVAL; > + > + base = pci_epf_alloc_space(epf, size, barno, align, 0); > + if (!base) { > + dev_err(dev, "intf: Config/Status/SPAD alloc region fail\n"); > + return -ENOMEM; > + } > + > + ntb->reg = base; > + > + ctrl = ntb->reg; > + ctrl->spad_offset = ctrl_size; > + > + ctrl->spad_count = spad_count; > + ctrl->num_mws = ntb->num_mws; > + ntb->spad_size = spad_size; > + > + ctrl->db_entry_size = 4; > + > + for (i = 0; i < ntb->db_count; i++) { > + ntb->reg->db_data[i] = 1 + i; > + ntb->reg->db_offset[i] = 0; > + } > + > + return 0; > +} > + > +/** > + * epf_ntb_configure_interrupt() - Configure MSI/MSI-X capaiblity > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + * Configure MSI/MSI-X capability for each interface with number of > + * interrupts equal to "db_count" configfs entry. > + */ > +static int epf_ntb_configure_interrupt(struct epf_ntb *ntb) > +{ > + const struct pci_epc_features *epc_features; > + bool msix_capable, msi_capable; > + u8 func_no, vfunc_no; > + struct device *dev; > + u32 db_count; > + int ret; > + > + dev = &ntb->epf->dev; > + > + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); > + msix_capable = epc_features->msix_capable; > + msi_capable = epc_features->msi_capable; > + > + if (!(msix_capable || msi_capable)) { I don't think the "msix_capable" and "msi_capable" local variables really add any readability since they're only used once. Not sure about "func_no", "vfunc_no", either. If you keep them, you could at least assign them earlier and use them in the pci_epc_get_features() call. > + dev_err(dev, "MSI or MSI-X is required for doorbell\n"); > + return -EINVAL; > + } > + > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + db_count = ntb->db_count; > + if (db_count > MAX_DB_COUNT) { > + dev_err(dev, "DB count cannot be more than %d\n", MAX_DB_COUNT); > + return -EINVAL; > + } > + > + ntb->db_count = db_count; > + > + if (msi_capable) { > + ret = pci_epc_set_msi(ntb->epf->epc, func_no, vfunc_no, 16); > + if (ret) { > + dev_err(dev, "intf: MSI configuration failed\n"); > + return ret; > + } > + } > + > + return 0; > +} > + > +/** > + * epf_ntb_db_bar_init() - Configure Doorbell window BARs > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * Spurious blank line. > + */ > +static int epf_ntb_db_bar_init(struct epf_ntb *ntb) > +{ > + const struct pci_epc_features *epc_features; > + u32 align; > + struct device *dev = &ntb->epf->dev; > + int ret; > + struct pci_epf_bar *epf_bar; > + void __iomem *mw_addr; > + enum pci_barno barno; > + size_t size = 4 * ntb->db_count; > + > + epc_features = pci_epc_get_features(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no); > + align = epc_features->align; > + > + if (size < 128) > + size = 128; > + > + if (align) > + size = ALIGN(size, align); > + else > + size = roundup_pow_of_two(size); > + > + barno = ntb->epf_ntb_bar[BAR_DB]; > + > + mw_addr = pci_epf_alloc_space(ntb->epf, size, barno, align, 0); > + if (!mw_addr) { > + dev_err(dev, "intf: Failed to allocate OB address\n"); > + return -ENOMEM; > + } > + > + ntb->epf_db = mw_addr; > + > + epf_bar = &ntb->epf->bar[barno]; > + > + ret = pci_epc_set_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); > + if (ret) { > + dev_err(dev, "intf: DoorBell BAR set failed\n"); s/DoorBell/Doorbell/ > + goto err_alloc_peer_mem; > + } > + return ret; > + > +err_alloc_peer_mem: > + pci_epc_mem_free_addr(ntb->epf->epc, epf_bar->phys_addr, mw_addr, epf_bar->size); > + return -1; > +} > + > +/** > + * epf_ntb_db_bar_clear() - Clear doorbell BAR and free memory > + * allocated in peers outbound address space s/peers/peer's/ > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + */ > +static void epf_ntb_db_bar_clear(struct epf_ntb *ntb) > +{ > + enum pci_barno barno; > + > + barno = ntb->epf_ntb_bar[BAR_DB]; > + pci_epf_free_space(ntb->epf, ntb->epf_db, barno, 0); > + pci_epc_clear_bar(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + &ntb->epf->bar[barno]); > +} > + > +/** > + * epf_ntb_mw_bar_init() - Configure Memory window BARs > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + */ > +static int epf_ntb_mw_bar_init(struct epf_ntb *ntb) > +{ > + int ret = 0; > + int i; > + u64 size; > + enum pci_barno barno; > + struct device *dev = &ntb->epf->dev; > + > + for (i = 0; i < ntb->num_mws; i++) { > + Spurious blank line. > + size = ntb->mws_size[i]; > + > + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; > + > + ntb->epf->bar[barno].barno = barno; > + ntb->epf->bar[barno].size = size; > + ntb->epf->bar[barno].addr = 0; > + ntb->epf->bar[barno].phys_addr = 0; > + ntb->epf->bar[barno].flags |= upper_32_bits(size) ? > + PCI_BASE_ADDRESS_MEM_TYPE_64 : > + PCI_BASE_ADDRESS_MEM_TYPE_32; > + > + ret = pci_epc_set_bar(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + &ntb->epf->bar[barno]); > + if (ret) { > + dev_err(dev, "intf: MW set failed\n"); > + goto err_alloc_mem; > + } > + > + /* allocate epc outbound memory windows to vpci vntb device */ s/allocate/Allocate/ to match other comments s/epc/EPC/ > + ntb->vpci_mw_addr[i] = pci_epc_mem_alloc_addr(ntb->epf->epc, > + &ntb->vpci_mw_phy[i], > + size); > + if (!ntb->vpci_mw_addr[i]) { > + dev_err(dev, "Failed to allocate source address\n"); > + goto err_alloc_mem; > + } > + } > + > + return ret; > +err_alloc_mem: > + return ret; > +} > + > +/** > + * epf_ntb_mw_bar_clear() - Clear Memory window BARs > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * Spurious blank line. > + */ > +static void epf_ntb_mw_bar_clear(struct epf_ntb *ntb) > +{ > + enum pci_barno barno; > + int i; > + > + for (i = 0; i < ntb->num_mws; i++) { > + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; > + pci_epc_clear_bar(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + &ntb->epf->bar[barno]); > + > + pci_epc_mem_free_addr(ntb->epf->epc, > + ntb->vpci_mw_phy[i], > + ntb->vpci_mw_addr[i], > + ntb->mws_size[i]); > + } > +} > + > +/** > + * epf_ntb_epc_destroy() - Cleanup NTB EPC interface > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + * Wrapper for epf_ntb_epc_destroy_interface() to cleanup all the NTB interfaces > + */ > +static void epf_ntb_epc_destroy(struct epf_ntb *ntb) > +{ > + pci_epc_remove_epf(ntb->epf->epc, ntb->epf, 0); > + pci_epc_put(ntb->epf->epc); > +} > + > +/** > + * epf_ntb_init_epc_bar() - Identify BARs to be used for each of the NTB > + * constructs (scratchpad region, doorbell, memorywindow) > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * Spurious blank line. > + */ > +static int epf_ntb_init_epc_bar(struct epf_ntb *ntb) > +{ > + const struct pci_epc_features *epc_features; > + enum pci_barno barno; > + enum epf_ntb_bar bar; > + struct device *dev; > + u32 num_mws; > + int i; > + > + barno = BAR_0; > + num_mws = ntb->num_mws; > + dev = &ntb->epf->dev; > + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); > + > + /* These are required BARs which are mandatory for NTB functionality */ > + for (bar = BAR_CONFIG; bar <= BAR_MW0; bar++, barno++) { > + barno = pci_epc_get_next_free_bar(epc_features, barno); > + if (barno < 0) { > + dev_err(dev, "intf: Fail to get NTB function BAR\n"); > + return barno; > + } > + ntb->epf_ntb_bar[bar] = barno; > + } > + > + /* These are optional BARs which don't impact NTB functionality */ > + for (bar = BAR_MW1, i = 1; i < num_mws; bar++, barno++, i++) { > + barno = pci_epc_get_next_free_bar(epc_features, barno); > + if (barno < 0) { > + ntb->num_mws = i; > + dev_dbg(dev, "BAR not available for > MW%d\n", i + 1); > + } > + ntb->epf_ntb_bar[bar] = barno; > + } > + > + return 0; > +} > + > +/** > + * epf_ntb_epc_init() - Initialize NTB interface > + * @ntb: NTB device that facilitates communication between HOST and vHOST2 > + * > + * Wrapper to initialize a particular EPC interface and start the workqueue > + * to check for commands from host. This function will write to the > + * EP controller HW for configuring it. > + */ > +static int epf_ntb_epc_init(struct epf_ntb *ntb) > +{ > + u8 func_no, vfunc_no; > + struct pci_epc *epc; > + struct pci_epf *epf; > + struct device *dev; > + int ret; > + > + epf = ntb->epf; > + dev = &epf->dev; > + epc = epf->epc; > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + ret = epf_ntb_config_sspad_bar_set(ntb); > + if (ret) { > + dev_err(dev, "intf: Config/self SPAD BAR init failed"); > + return ret; > + } > + > + ret = epf_ntb_configure_interrupt(ntb); > + if (ret) { > + dev_err(dev, "intf: Interrupt configuration failed\n"); > + goto err_config_interrupt; > + } > + > + ret = epf_ntb_db_bar_init(ntb); > + if (ret) { > + dev_err(dev, "intf: DB BAR init failed\n"); > + goto err_db_bar_init; > + } > + > + ret = epf_ntb_mw_bar_init(ntb); > + if (ret) { > + dev_err(dev, "intf: MW BAR init failed\n"); > + goto err_mw_bar_init; > + } > + > + if (vfunc_no <= 1) { > + ret = pci_epc_write_header(epc, func_no, vfunc_no, epf->header); > + if (ret) { > + dev_err(dev, "intf: Configuration header write failed\n"); > + goto err_write_header; > + } > + } > + > + INIT_DELAYED_WORK(&ntb->cmd_handler, epf_ntb_cmd_handler); > + queue_work(kpcintb_workqueue, &ntb->cmd_handler.work); > + > + return 0; > + > +err_write_header: > + epf_ntb_mw_bar_clear(ntb); > +err_mw_bar_init: > + epf_ntb_db_bar_clear(ntb); > +err_db_bar_init: > +err_config_interrupt: > + epf_ntb_config_sspad_bar_clear(ntb); > + > + return ret; > +} > + > + > +/** > + * epf_ntb_epc_cleanup() - Cleanup all NTB interfaces > + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 > + * > + * Wrapper to cleanup all NTB interfaces. > + */ > +static void epf_ntb_epc_cleanup(struct epf_ntb *ntb) > +{ > + epf_ntb_db_bar_clear(ntb); > + epf_ntb_mw_bar_clear(ntb); > +} > + > +#define EPF_NTB_R(_name) \ > +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ > + char *page) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + \ > + return sprintf(page, "%d\n", ntb->_name); \ > +} > + > +#define EPF_NTB_W(_name) \ > +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ > + const char *page, size_t len) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + u32 val; \ > + int ret; \ > + \ > + ret = kstrtou32(page, 0, &val); \ > + if (ret) \ > + return ret; \ > + \ > + ntb->_name = val; \ > + \ > + return len; \ > +} > + > +#define EPF_NTB_MW_R(_name) \ > +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ > + char *page) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + int win_no; \ > + \ > + sscanf(#_name, "mw%d", &win_no); \ > + \ > + return sprintf(page, "%lld\n", ntb->mws_size[win_no - 1]); \ > +} > + > +#define EPF_NTB_MW_W(_name) \ > +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ > + const char *page, size_t len) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + struct device *dev = &ntb->epf->dev; \ > + int win_no; \ > + u64 val; \ > + int ret; \ > + \ > + ret = kstrtou64(page, 0, &val); \ > + if (ret) \ > + return ret; \ > + \ > + if (sscanf(#_name, "mw%d", &win_no) != 1) \ > + return -EINVAL; \ > + \ > + if (ntb->num_mws < win_no) { \ > + dev_err(dev, "Invalid num_nws: %d value\n", ntb->num_mws); \ > + return -EINVAL; \ > + } \ > + \ > + ntb->mws_size[win_no - 1] = val; \ > + \ > + return len; \ > +} > + > +static ssize_t epf_ntb_num_mws_store(struct config_item *item, > + const char *page, size_t len) > +{ > + struct config_group *group = to_config_group(item); > + struct epf_ntb *ntb = to_epf_ntb(group); > + u32 val; > + int ret; > + > + ret = kstrtou32(page, 0, &val); > + if (ret) > + return ret; > + > + if (val > MAX_MW) > + return -EINVAL; > + > + ntb->num_mws = val; > + > + return len; > +} > + > +EPF_NTB_R(spad_count) > +EPF_NTB_W(spad_count) > +EPF_NTB_R(db_count) > +EPF_NTB_W(db_count) > +EPF_NTB_R(num_mws) > +EPF_NTB_R(vbus_number) > +EPF_NTB_W(vbus_number) > +EPF_NTB_MW_R(mw1) > +EPF_NTB_MW_W(mw1) > +EPF_NTB_MW_R(mw2) > +EPF_NTB_MW_W(mw2) > +EPF_NTB_MW_R(mw3) > +EPF_NTB_MW_W(mw3) > +EPF_NTB_MW_R(mw4) > +EPF_NTB_MW_W(mw4) > + > +CONFIGFS_ATTR(epf_ntb_, spad_count); > +CONFIGFS_ATTR(epf_ntb_, db_count); > +CONFIGFS_ATTR(epf_ntb_, num_mws); > +CONFIGFS_ATTR(epf_ntb_, mw1); > +CONFIGFS_ATTR(epf_ntb_, mw2); > +CONFIGFS_ATTR(epf_ntb_, mw3); > +CONFIGFS_ATTR(epf_ntb_, mw4); > +CONFIGFS_ATTR(epf_ntb_, vbus_number); > + > +static struct configfs_attribute *epf_ntb_attrs[] = { > + &epf_ntb_attr_spad_count, > + &epf_ntb_attr_db_count, > + &epf_ntb_attr_num_mws, > + &epf_ntb_attr_mw1, > + &epf_ntb_attr_mw2, > + &epf_ntb_attr_mw3, > + &epf_ntb_attr_mw4, > + &epf_ntb_attr_vbus_number, > + NULL, > +}; > + > +static const struct config_item_type ntb_group_type = { > + .ct_attrs = epf_ntb_attrs, > + .ct_owner = THIS_MODULE, > +}; > + > +/** > + * epf_ntb_add_cfs() - Add configfs directory specific to NTB > + * @epf: NTB endpoint function device > + * @group: A pointer to the config_group structure referencing a group of > + * config_items of a specific type that belong to a specific sub-system. > + * > + * Add configfs directory specific to NTB. This directory will hold > + * NTB specific properties like db_count, spad_count, num_mws etc., > + */ > +static struct config_group *epf_ntb_add_cfs(struct pci_epf *epf, > + struct config_group *group) > +{ > + struct epf_ntb *ntb = epf_get_drvdata(epf); > + struct config_group *ntb_group = &ntb->group; > + struct device *dev = &epf->dev; > + > + config_group_init_type_name(ntb_group, dev_name(dev), &ntb_group_type); > + > + return ntb_group; > +} > + > +/*==== virtual PCI bus driver, which only load virutal ntb pci driver ====*/ s/virutal ntb pci/virtual NTB PCI/ > +#define VPCI_BUS_NUM 0x10 How did you pick this number and how do we know it is available? > +uint32_t pci_space[] = { "static u32"? > + (VNTB_VID | (VNTB_PID << 16)), //DeviceID, Vendor ID > + 0, // status, Command > + 0xffffffff, // Class code, subclass, prog if, revision id > + 0x40, //bist, header type, latency Timer, cache line size > + 0, //bar 0 > + 0, //bar 1 > + 0, //bar 2 > + 0, //bar 3 > + 0, //bar 4 > + 0, //bar 5 s/bar/BAR/ > + 0, //cardbus cis point > + 0, //Subsystem ID Subystem vendor id > + 0, //ROM Base Address > + 0, //Reserved, Cap. Point > + 0, //Reserved, > + 0, //Max Lat, Min Gnt, interrupt pin, interrupt line Use /* */ instead of //, like the rest of drivers/pci/ > +}; > + > +int pci_read(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 *val) > +{ > + if (devfn == 0) { > + memcpy(val, ((uint8_t *)pci_space) + where, size); "u8 *"? > + return 0; > + } > + return -1; These should return PCIBIOS_SUCCESSFUL or PCIBIOS_DEVICE_NOT_FOUND. > +} > + > +int pci_write(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 val) > +{ > + return 0; > +} > + > +struct pci_ops vpci_ops = { > + .read = pci_read, > + .write = pci_write, > +}; > + > +static int vpci_bus(void *sysdata) The function name should say something about what it does. Maybe vpci_scan_bus() or something? > +{ > + struct pci_bus *vpci_bus; > + > + vpci_bus = pci_scan_bus(VPCI_BUS_NUM, &vpci_ops, sysdata); > + if (vpci_bus) > + pr_err("create pci bus\n"); > + > + pci_bus_add_devices(vpci_bus); > + > + return 0; > +} > + > +/*==================== Virtual PCIe NTB driver ==========================*/ > + > +static int vntb_epf_mw_count(struct ntb_dev *ntb, int pidx) > +{ > + struct epf_ntb *ndev = ntb_ndev(ntb); > + > + return ndev->num_mws; > +} > + > +static int vntb_epf_spad_count(struct ntb_dev *ntb) > +{ > + return ntb_ndev(ntb)->spad_count; > +} > + > +static int vntb_epf_peer_mw_count(struct ntb_dev *ntb) > +{ > + return ntb_ndev(ntb)->num_mws; > +} > + > +static u64 vntb_epf_db_valid_mask(struct ntb_dev *ntb) > +{ > + return BIT_ULL(ntb_ndev(ntb)->db_count) - 1; > +} > + > +static int vntb_epf_db_set_mask(struct ntb_dev *ntb, u64 db_bits) > +{ > + return 0; > +} > + > +static int vntb_epf_mw_set_trans(struct ntb_dev *ndev, int pidx, int idx, > + dma_addr_t addr, resource_size_t size) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct pci_epf_bar *epf_bar; > + enum pci_barno barno; > + int ret; > + struct device *dev; > + > + dev = &ntb->ntb.dev; > + barno = ntb->epf_ntb_bar[BAR_MW0 + idx]; > + epf_bar = &ntb->epf->bar[barno]; > + epf_bar->phys_addr = addr; > + epf_bar->barno = barno; > + epf_bar->size = size; > + > + ret = pci_epc_set_bar(ntb->epf->epc, 0, 0, epf_bar); > + if (ret) { > + dev_err(dev, "failure set mw trans\n"); > + return ret; > + } > + return 0; > +} > + > +static int vntb_epf_mw_clear_trans(struct ntb_dev *ntb, int pidx, int idx) > +{ > + return 0; > +} > + > +static int vntb_epf_peer_mw_get_addr(struct ntb_dev *ndev, int idx, > + phys_addr_t *base, resource_size_t *size) > +{ > + > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + if (base) > + *base = ntb->vpci_mw_phy[idx]; > + > + if (size) > + *size = ntb->mws_size[idx]; > + > + return 0; > +} > + > +static int vntb_epf_link_enable(struct ntb_dev *ntb, > + enum ntb_speed max_speed, > + enum ntb_width max_width) > +{ > + return 0; > +} > + > +static u32 vntb_epf_spad_read(struct ntb_dev *ndev, int idx) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + int off = ntb->reg->spad_offset, ct = ntb->reg->spad_count * 4; > + u32 val; > + void __iomem *base = ntb->reg; > + > + val = readl(base + off + ct + idx * 4); > + return val; > +} > + > +static int vntb_epf_spad_write(struct ntb_dev *ndev, int idx, u32 val) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct epf_ntb_ctrl *ctrl = ntb->reg; > + int off = ctrl->spad_offset, ct = ctrl->spad_count * 4; > + void __iomem *base = ntb->reg; > + > + writel(val, base + off + ct + idx * 4); > + return 0; > +} > + > +static u32 vntb_epf_peer_spad_read(struct ntb_dev *ndev, int pidx, int idx) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct epf_ntb_ctrl *ctrl = ntb->reg; > + int off = ctrl->spad_offset; > + void __iomem *base = ntb->reg; > + u32 val; > + > + val = readl(base + off + idx * 4); > + return val; > +} > + > +static int vntb_epf_peer_spad_write(struct ntb_dev *ndev, int pidx, int idx, u32 val) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct epf_ntb_ctrl *ctrl = ntb->reg; > + int off = ctrl->spad_offset; > + void __iomem *base = ntb->reg; > + > + writel(val, base + off + idx * 4); > + return 0; > +} > + > +static int vntb_epf_peer_db_set(struct ntb_dev *ndev, u64 db_bits) > +{ > + u32 interrupt_num = ffs(db_bits) + 1; > + struct epf_ntb *ntb = ntb_ndev(ndev); > + u8 func_no, vfunc_no; > + int ret; > + > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + ret = pci_epc_raise_irq(ntb->epf->epc, > + func_no, > + vfunc_no, > + PCI_EPC_IRQ_MSI, > + interrupt_num + 1); > + if (ret) { > + dev_err(&ntb->ntb.dev, "intf: Failed to raise IRQ\n"); > + return ret; > + } > + > + return 0; Equivalent to: if (ret) dev_err(...); return ret; > +} > + > +static u64 vntb_epf_db_read(struct ntb_dev *ndev) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + return ntb->db; > +} > + > +static int vntb_epf_mw_get_align(struct ntb_dev *ndev, int pidx, int idx, > + resource_size_t *addr_align, > + resource_size_t *size_align, > + resource_size_t *size_max) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + if (addr_align) > + *addr_align = SZ_4K; > + > + if (size_align) > + *size_align = 1; > + > + if (size_max) > + *size_max = ntb->mws_size[idx]; > + > + return 0; > +} > + > +static u64 vntb_epf_link_is_up(struct ntb_dev *ndev, > + enum ntb_speed *speed, > + enum ntb_width *width) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + return ntb->reg->link_status; > +} > + > +static int vntb_epf_db_clear_mask(struct ntb_dev *ndev, u64 db_bits) > +{ > + return 0; > +} > + > +static int vntb_epf_db_clear(struct ntb_dev *ndev, u64 db_bits) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + ntb->db &= ~db_bits; > + return 0; > +} > + > +static int vntb_epf_link_disable(struct ntb_dev *ntb) > +{ > + return 0; > +} > + > +static const struct ntb_dev_ops vntb_epf_ops = { > + .mw_count = vntb_epf_mw_count, > + .spad_count = vntb_epf_spad_count, > + .peer_mw_count = vntb_epf_peer_mw_count, > + .db_valid_mask = vntb_epf_db_valid_mask, > + .db_set_mask = vntb_epf_db_set_mask, > + .mw_set_trans = vntb_epf_mw_set_trans, > + .mw_clear_trans = vntb_epf_mw_clear_trans, > + .peer_mw_get_addr = vntb_epf_peer_mw_get_addr, > + .link_enable = vntb_epf_link_enable, > + .spad_read = vntb_epf_spad_read, > + .spad_write = vntb_epf_spad_write, > + .peer_spad_read = vntb_epf_peer_spad_read, > + .peer_spad_write = vntb_epf_peer_spad_write, > + .peer_db_set = vntb_epf_peer_db_set, > + .db_read = vntb_epf_db_read, > + .mw_get_align = vntb_epf_mw_get_align, > + .link_is_up = vntb_epf_link_is_up, > + .db_clear_mask = vntb_epf_db_clear_mask, > + .db_clear = vntb_epf_db_clear, > + .link_disable = vntb_epf_link_disable, > +}; > + > +static int pci_vntb_probe(struct pci_dev *pdev, const struct pci_device_id *id) > +{ > + int ret; > + struct epf_ntb *ndev = (struct epf_ntb *)pdev->sysdata; > + struct device *dev = &pdev->dev; > + > + ndev->ntb.pdev = pdev; > + ndev->ntb.topo = NTB_TOPO_NONE; > + ndev->ntb.ops = &vntb_epf_ops; > + > + ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); > + if (ret) { > + dev_err(dev, "Cannot set DMA mask\n"); > + return -1; return -EINVAL; > + } > + > + ret = ntb_register_device(&ndev->ntb); > + if (ret) { > + dev_err(dev, "Failed to register NTB device\n"); > + goto err_register_dev; > + } > + > + dev_info(dev, "PCI Virtual NTB driver loaded\n"); > + return 0; > + > +err_register_dev: > + return -1; return -EINVAL; > +} > + > +static const struct pci_device_id pci_vntb_table[] = { > + { > + PCI_DEVICE(VNTB_VID, VNTB_PID), > + }, > + {}, > +}; > + > +static struct pci_driver vntb_pci_driver = { > + .name = "pci-vntb", > + .id_table = pci_vntb_table, > + .probe = pci_vntb_probe, > +}; > + > +/* ============ PCIe EPF Driver Bind ====================*/ > + > +/** > + * epf_ntb_bind() - Initialize endpoint controller to provide NTB functionality > + * @epf: NTB endpoint function device > + * > + * Initialize both the endpoint controllers associated with NTB function device. > + * Invoked when a primary interface or secondary interface is bound to EPC > + * device. This function will succeed only when EPC is bound to both the > + * interfaces. > + */ > +static int epf_ntb_bind(struct pci_epf *epf) > +{ > + struct epf_ntb *ntb = epf_get_drvdata(epf); > + struct device *dev = &epf->dev; > + int ret; > + > + if (!epf->epc) { > + dev_dbg(dev, "PRIMARY EPC interface not yet bound\n"); > + return 0; > + } > + > + ret = epf_ntb_init_epc_bar(ntb); > + if (ret) { > + dev_err(dev, "Failed to create NTB EPC\n"); > + goto err_bar_init; > + } > + > + ret = epf_ntb_config_spad_bar_alloc(ntb); > + if (ret) { > + dev_err(dev, "Failed to allocate BAR memory\n"); > + goto err_bar_alloc; > + } > + > + ret = epf_ntb_epc_init(ntb); > + if (ret) { > + dev_err(dev, "Failed to initialize EPC\n"); > + goto err_bar_alloc; > + } > + > + epf_set_drvdata(epf, ntb); > + > + if (pci_register_driver(&vntb_pci_driver)) { > + dev_err(dev, "failure register vntb pci driver\n"); > + goto err_bar_alloc; > + } > + > + vpci_bus(ntb); > + > + return 0; > + > +err_bar_alloc: > + epf_ntb_config_spad_bar_free(ntb); > + > +err_bar_init: > + epf_ntb_epc_destroy(ntb); > + > + return ret; > +} > + > +/** > + * epf_ntb_unbind() - Cleanup the initialization from epf_ntb_bind() > + * @epf: NTB endpoint function device > + * > + * Cleanup the initialization from epf_ntb_bind() > + */ > +static void epf_ntb_unbind(struct pci_epf *epf) > +{ > + struct epf_ntb *ntb = epf_get_drvdata(epf); > + > + epf_ntb_epc_cleanup(ntb); > + epf_ntb_config_spad_bar_free(ntb); > + epf_ntb_epc_destroy(ntb); > + > + pci_unregister_driver(&vntb_pci_driver); > +} > + > +// EPF driver probe > +static struct pci_epf_ops epf_ntb_ops = { > + .bind = epf_ntb_bind, > + .unbind = epf_ntb_unbind, > + .add_cfs = epf_ntb_add_cfs, > +}; > + > +/** > + * epf_ntb_probe() - Probe NTB function driver > + * @epf: NTB endpoint function device > + * > + * Probe NTB function driver when endpoint function bus detects a NTB > + * endpoint function. > + */ > +static int epf_ntb_probe(struct pci_epf *epf) > +{ > + struct epf_ntb *ntb; > + struct device *dev; > + > + dev = &epf->dev; > + > + ntb = devm_kzalloc(dev, sizeof(*ntb), GFP_KERNEL); > + if (!ntb) > + return -ENOMEM; > + > + epf->header = &epf_ntb_header; > + ntb->epf = epf; > + epf_set_drvdata(epf, ntb); > + > + dev_info(dev, "pci-ep epf driver loaded\n"); I think most drivers don't announce when they're loaded because it really doesn't help the user. > + return 0; > +} > + > +static const struct pci_epf_device_id epf_ntb_ids[] = { > + { > + .name = "pci_epf_vntb", > + }, > + {}, > +}; > + > +static struct pci_epf_driver epf_ntb_driver = { > + .driver.name = "pci_epf_vntb", > + .probe = epf_ntb_probe, > + .id_table = epf_ntb_ids, > + .ops = &epf_ntb_ops, > + .owner = THIS_MODULE, > +}; > + > + Spurious blank line. > +static int __init epf_ntb_init(void) > +{ > + int ret; > + > + kpcintb_workqueue = alloc_workqueue("kpcintb", WQ_MEM_RECLAIM | > + WQ_HIGHPRI, 0); > + ret = pci_epf_register_driver(&epf_ntb_driver); > + if (ret) { > + destroy_workqueue(kpcintb_workqueue); > + pr_err("Failed to register pci epf ntb driver --> %d\n", ret); > + return ret; > + } > + > + return 0; > +} > +module_init(epf_ntb_init); > + > +static void __exit epf_ntb_exit(void) > +{ > + pci_epf_unregister_driver(&epf_ntb_driver); > + destroy_workqueue(kpcintb_workqueue); > +} > +module_exit(epf_ntb_exit); > + > +MODULE_DESCRIPTION("PCI EPF NTB DRIVER"); > +MODULE_AUTHOR("Frank Li <Frank.li@nxp.com>"); > +MODULE_LICENSE("GPL v2"); > -- > 2.24.0.rc1 >
On Thu, Feb 17, 2022 at 4:38 PM Bjorn Helgaas <helgaas@kernel.org> wrote: > > [+cc Jon, Dave, Allen, linux-ntb] > > This patch only touches drivers/pci; the subject line should reflect > that to help people know which things to look at. > > Maybe something like: > > PCI: endpoint: Support NTB transfer between RC and EP > > On Mon, Feb 14, 2022 at 11:38:43PM -0600, Frank Li wrote: > > Add NTB function driver and virtual PCI BUS and Virtual NTB driver > > to implement communication between PCIe RC and PCIe EP devices > > s/PCI BUS/PCI Bus/ > s/PCIe RC/PCIe Root Port/ (I think; the RC itself doesn't have a > software representation) > > > ┌────────────┐ ┌─────────────────────────────────────┐ > > │ │ │ │ > > ├────────────┤ │ ┌──────────────┤ > > │ NTB │ │ │ NTB │ > > │ NetDev │ │ │ NetDev │ > > ├────────────┤ │ ├──────────────┤ > > │ NTB │ │ │ NTB │ > > │ Transfer │ │ │ Transfer │ > > ├────────────┤ │ ├──────────────┤ > > │ │ │ │ │ > > │ PCI NTB │ │ │ │ > > │ EPF │ │ │ │ > > │ Driver │ │ │ PCI Virtual │ > > │ │ ├───────────────┐ │ NTB Driver │ > > │ │ │ PCI EP NTB │◄────►│ │ > > │ │ │ FN Driver │ │ │ > > ├────────────┤ ├───────────────┤ ├──────────────┤ > > │ │ │ │ │ │ > > │ PCI BUS │ ◄─────► │ PCI EP BUS │ │ Virtual PCI │ > > │ │ PCI │ │ │ BUS │ > > └────────────┘ └───────────────┴──────┴──────────────┘ > > s/PCI BUS/PCI Bus/ > > > PCI RC PCI EP > > s/PCI RC/PCIe Root Port/ ? > > > This driver include 3 part: > > 1 PCI EP NTB function driver > > 2 Virtual PCI bus > > 3 PCI virutal NTB driver, which is loaded only by above virtual pci bus > > s/include 3 part/includes 3 parts/ > s/virutal/virtual/ > s/pci bus/PCI bus/ > > > Signed-off-by: Frank Li <Frank.Li@nxp.com> > > --- > > drivers/pci/endpoint/functions/Kconfig | 11 + > > drivers/pci/endpoint/functions/Makefile | 1 + > > drivers/pci/endpoint/functions/pci-epf-vntb.c | 1425 +++++++++++++++++ > > 3 files changed, 1437 insertions(+) > > create mode 100644 drivers/pci/endpoint/functions/pci-epf-vntb.c > > > > diff --git a/drivers/pci/endpoint/functions/Kconfig b/drivers/pci/endpoint/functions/Kconfig > > index 5f1242ca2f4e4..362555b024e8f 100644 > > --- a/drivers/pci/endpoint/functions/Kconfig > > +++ b/drivers/pci/endpoint/functions/Kconfig > > @@ -25,3 +25,14 @@ config PCI_EPF_NTB > > device tree. > > > > If in doubt, say "N" to disable Endpoint NTB driver. > > + > > +config PCI_EPF_VNTB > > + tristate "PCI Endpoint NTB driver" > > + depends on PCI_ENDPOINT > > + select CONFIGFS_FS > > + help > > + Select this configuration option to enable the Non-Transparent > > + Bridge (NTB) driver for PCI Endpoint. NTB driver implements NTB > > + between PCI host and PCIe Endpoint. > > "PCI Endpoint" vs "PCIe Endpoint". Pick one. It seems like you > require an upstream bridge, i.e., a Root Port, so maybe "PCIe Root > Port" and "PCIe Endpoint" is what you want? > > > + If in doubt, say "N" to disable Endpoint NTB driver. > > diff --git a/drivers/pci/endpoint/functions/Makefile b/drivers/pci/endpoint/functions/Makefile > > index 96ab932a537a2..5c13001deaba1 100644 > > --- a/drivers/pci/endpoint/functions/Makefile > > +++ b/drivers/pci/endpoint/functions/Makefile > > @@ -5,3 +5,4 @@ > > > > obj-$(CONFIG_PCI_EPF_TEST) += pci-epf-test.o > > obj-$(CONFIG_PCI_EPF_NTB) += pci-epf-ntb.o > > +obj-$(CONFIG_PCI_EPF_VNTB) += pci-epf-vntb.o > > diff --git a/drivers/pci/endpoint/functions/pci-epf-vntb.c b/drivers/pci/endpoint/functions/pci-epf-vntb.c > > new file mode 100644 > > index 0000000000000..ebf7e243eefa4 > > --- /dev/null > > +++ b/drivers/pci/endpoint/functions/pci-epf-vntb.c > > @@ -0,0 +1,1425 @@ > > +// SPDX-License-Identifier: GPL-2.0 > > +/* > > + * Endpoint Function Driver to implement Non-Transparent Bridge functionality > > + * Between PCI RC and EP > > + * > > + * Copyright (C) 2020 Texas Instruments > > + * Copyright (C) 2022 NXP > > + * > > + * Based on pci-epf-ntb.c > > + * Author: Frank Li <Frank.Li@nxp.com> > > + * Author: Kishon Vijay Abraham I <kishon@ti.com> > > + */ > > + > > +/** > > + * +------------+ +---------------------------------------+ > > + * | | | | > > + * +------------+ | +--------------+ > > + * | NTB | | | NTB | > > + * | NetDev | | | NetDev | > > + * +------------+ | +--------------+ > > + * | NTB | | | NTB | > > + * | Transfer | | | Transfer | > > + * +------------+ | +--------------+ > > + * | | | | | > > + * | PCI NTB | | | | > > + * | EPF | | | | > > + * | Driver | | | PCI Virtual | > > + * | | +---------------+ | NTB Driver | > > + * | | | PCI EP NTB |<------>| | > > + * | | | FN Driver | | | > > + * +------------+ +---------------+ +--------------+ > > + * | | | | | | > > + * | PCI BUS | <-----> | PCI EP BUS | | Virtual PCI | > > + * | | PCI | | | BUS | > > + * +------------+ +---------------+--------+--------------+ > > + * PCI RC PCI EP > > s/PCI RC/PCIe Root Port/ ? > > > + */ > > + > > +#include <linux/delay.h> > > +#include <linux/io.h> > > +#include <linux/module.h> > > +#include <linux/slab.h> > > + > > +#include <linux/pci-epc.h> > > +#include <linux/pci-epf.h> > > +#include <linux/ntb.h> > > + > > +static struct workqueue_struct *kpcintb_workqueue; > > + > > +#define COMMAND_CONFIGURE_DOORBELL 1 > > +#define COMMAND_TEARDOWN_DOORBELL 2 > > +#define COMMAND_CONFIGURE_MW 3 > > +#define COMMAND_TEARDOWN_MW 4 > > +#define COMMAND_LINK_UP 5 > > +#define COMMAND_LINK_DOWN 6 > > + > > +#define COMMAND_STATUS_OK 1 > > +#define COMMAND_STATUS_ERROR 2 > > + > > +#define LINK_STATUS_UP BIT(0) > > + > > +#define SPAD_COUNT 64 > > +#define DB_COUNT 4 > > +#define NTB_MW_OFFSET 2 > > +#define DB_COUNT_MASK GENMASK(15, 0) > > +#define MSIX_ENABLE BIT(16) > > +#define MAX_DB_COUNT 32 > > +#define MAX_MW 4 > > + > > +#define VNTB_VID 0x1957 > > This looks like PCI_VENDOR_ID_NXP (or PCI_VENDOR_ID_FREESCALE). > How did you choose that, why is it safe to use, and why don't use use > PCI_VENDOR_ID_NXP instead? This is a problem. I am working on NXP. But I have not found an apartment, who managed the PCIe PID list yet. I can put it in the configfs, let the user change it. Or someone can donate one PID/VID. > > > +#define VNTB_PID 0x080A > > + > > +enum epf_ntb_bar { > > + BAR_CONFIG, > > + BAR_DB, > > + BAR_MW0, > > + BAR_MW1, > > + BAR_MW2, > > +}; > > + > > +/* > > + * +--------------------------------------------------+ Base > > + * | | > > + * | | > > + * | | > > + * | Common Control Register | > > + * | | > > + * | | > > + * | | > > + * +-----------------------+--------------------------+ Base+span_offset > > + * | | | > > + * | Peer Span Space | Span Space | > > + * | | | > > + * | | | > > + * +-----------------------+--------------------------+ Base+span_offset > > + * | | | +span_count * 4 > > + * | | | > > + * | Span Space | Peer Span Space | > > + * | | | > > + * +-----------------------+--------------------------+ > > + * Virtual PCI Pcie Endpoint > > + * NTB Driver NTB Driver > > s/Pcie/PCIe/ > > > + */ > > +struct epf_ntb_ctrl { > > + u32 command; > > + u32 argument; > > + u16 command_status; > > + u16 link_status; > > + u32 topology; > > + u64 addr; > > + u64 size; > > + u32 num_mws; > > + u32 reserved; > > + u32 spad_offset; > > + u32 spad_count; > > + u32 db_entry_size; > > + u32 db_data[MAX_DB_COUNT]; > > + u32 db_offset[MAX_DB_COUNT]; > > +} __packed; > > + > > +struct epf_ntb { > > + struct ntb_dev ntb; > > + struct pci_epf *epf; > > + struct config_group group; > > + > > + u32 num_mws; > > + u32 db_count; > > + u32 spad_count; > > + u64 mws_size[MAX_MW]; > > + u64 db; > > + u32 vbus_number; > > + > > + bool linkup; > > + u32 spad_size; > > + > > + enum pci_barno epf_ntb_bar[6]; > > + > > + struct epf_ntb_ctrl *reg; > > + > > + phys_addr_t epf_db_phy; > > + void __iomem *epf_db; > > + > > + phys_addr_t vpci_mw_phy[MAX_MW]; > > + void __iomem *vpci_mw_addr[MAX_MW]; > > + > > + struct delayed_work cmd_handler; > > +}; > > + > > +#define to_epf_ntb(epf_group) container_of((epf_group), struct epf_ntb, group) > > +#define ntb_ndev(__ntb) container_of(__ntb, struct epf_ntb, ntb) > > + > > +static struct pci_epf_header epf_ntb_header = { > > + .vendorid = PCI_ANY_ID, > > + .deviceid = PCI_ANY_ID, > > + .baseclass_code = PCI_BASE_CLASS_MEMORY, > > + .interrupt_pin = PCI_INTERRUPT_INTA, > > +}; > > + > > +/** > > + * epf_ntb_link_up() - Raise link_up interrupt to Virtual Host > > + * @ntb: NTB device that facilitates communication between HOST and VHOST > > + * @link_up: true or false indicating Link is UP or Down > > + * > > + * Once NTB function in HOST invoke ntb_link_enable(), > > + * this NTB function driver will trigger a link event to vhost. > > + */ > > +static int epf_ntb_link_up(struct epf_ntb *ntb, bool link_up) > > +{ > > + if (link_up) > > + ntb->reg->link_status |= LINK_STATUS_UP; > > + else > > + ntb->reg->link_status &= ~LINK_STATUS_UP; > > + > > + ntb_link_event(&ntb->ntb); > > + return 0; > > +} > > + > > +/** > > + * epf_ntb_configure_mw() - Configure the Outbound Address Space for vhost > > + * to access the memory window of host > > + * @ntb: NTB device that facilitates communication between host and vhost > > + * @mw: Index of the memory window (either 0, 1, 2 or 3) > > + * > > + * EP Outbound Window > > + * +--------+ +-----------+ > > + * | | | | > > + * | | | | > > + * | | | | > > + * | | | | > > + * | | +-----------+ > > + * | Virtual| | Memory Win| > > + * | NTB | -----------> | | > > + * | Driver | | | > > + * | | +-----------+ > > + * | | | | > > + * | | | | > > + * +--------+ +-----------+ > > + * VHost PCI EP > > + */ > > +static int epf_ntb_configure_mw(struct epf_ntb *ntb, u32 mw) > > +{ > > + phys_addr_t phys_addr; > > + u8 func_no, vfunc_no; > > + u64 addr, size; > > + int ret = 0; > > + > > + phys_addr = ntb->vpci_mw_phy[mw]; > > + addr = ntb->reg->addr; > > + size = ntb->reg->size; > > + > > + func_no = ntb->epf->func_no; > > + vfunc_no = ntb->epf->vfunc_no; > > + > > + ret = pci_epc_map_addr(ntb->epf->epc, func_no, vfunc_no, phys_addr, addr, size); > > + if (ret) > > + dev_err(&ntb->epf->epc->dev, > > + "intf: Failed to map memory window %d address\n", mw); > > Lots of these messages start with "intf:". What is that telling us? > Is it useful? I will clear it. > > > + return ret; > > +} > > + > > +/** > > + * epf_ntb_teardown_mw() - Teardown the configured OB ATU > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * @mw: Index of the memory window (either 0, 1, 2 or 3) > > + * > > + * Teardown the configured OB ATU configured in epf_ntb_configure_mw() using > > + * pci_epc_unmap_addr() > > + */ > > +static void epf_ntb_teardown_mw(struct epf_ntb *ntb, u32 mw) > > +{ > > + pci_epc_unmap_addr(ntb->epf->epc, > > + ntb->epf->func_no, > > + ntb->epf->vfunc_no, > > + ntb->vpci_mw_phy[mw]); > > +} > > + > > +/** > > + * epf_ntb_cmd_handler() - Handle commands provided by the NTB Host > > + * @work: work_struct for the epf_ntb_epc > > + * > > + * Workqueue function that gets invoked for the two epf_ntb_epc > > + * periodically (once every 5ms) to see if it has received any commands > > + * from NTB host. The host can send commands to configure doorbell or > > + * configure memory window or to update link status. > > + */ > > +static void epf_ntb_cmd_handler(struct work_struct *work) > > +{ > > + struct epf_ntb_ctrl *ctrl; > > + u32 command, argument; > > + struct epf_ntb *ntb; > > + struct device *dev; > > + int ret; > > + int i; > > + > > + ntb = container_of(work, struct epf_ntb, cmd_handler.work); > > + > > + for (i = 1; i < ntb->db_count; i++) { > > + if (readl(ntb->epf_db + i * 4)) { > > + if (readl(ntb->epf_db + i * 4)) > > + ntb->db |= 1 << (i - 1); > > + > > + ntb_db_event(&ntb->ntb, i); > > + writel(0, ntb->epf_db + i * 4); > > + } > > + } > > + > > + ctrl = ntb->reg; > > + command = ctrl->command; > > + if (!command) > > + goto reset_handler; > > + argument = ctrl->argument; > > + > > + ctrl->command = 0; > > + ctrl->argument = 0; > > + > > + ctrl = ntb->reg; > > + dev = &ntb->epf->dev; > > + > > + switch (command) { > > + case COMMAND_CONFIGURE_DOORBELL: > > + ctrl->command_status = COMMAND_STATUS_OK; > > + break; > > + case COMMAND_TEARDOWN_DOORBELL: > > + ctrl->command_status = COMMAND_STATUS_OK; > > + break; > > + case COMMAND_CONFIGURE_MW: > > + ret = epf_ntb_configure_mw(ntb, argument); > > + if (ret < 0) > > + ctrl->command_status = COMMAND_STATUS_ERROR; > > + else > > + ctrl->command_status = COMMAND_STATUS_OK; > > + break; > > + case COMMAND_TEARDOWN_MW: > > + epf_ntb_teardown_mw(ntb, argument); > > + ctrl->command_status = COMMAND_STATUS_OK; > > + break; > > + case COMMAND_LINK_UP: > > + ntb->linkup = true; > > + ret = epf_ntb_link_up(ntb, true); > > + if (ret < 0) > > + ctrl->command_status = COMMAND_STATUS_ERROR; > > + else > > + ctrl->command_status = COMMAND_STATUS_OK; > > + goto reset_handler; > > + case COMMAND_LINK_DOWN: > > + ntb->linkup = false; > > + ret = epf_ntb_link_up(ntb, false); > > + if (ret < 0) > > + ctrl->command_status = COMMAND_STATUS_ERROR; > > + else > > + ctrl->command_status = COMMAND_STATUS_OK; > > + break; > > + default: > > + dev_err(dev, "intf UNKNOWN command: %d\n", command); > > + break; > > + } > > + > > +reset_handler: > > + queue_delayed_work(kpcintb_workqueue, &ntb->cmd_handler, > > + msecs_to_jiffies(5)); > > +} > > + > > +/** > > + * epf_ntb_config_sspad_bar_clear() - Clear Config + Self scratchpad BAR > > + * @ntb_epc: EPC associated with one of the HOST which holds peer's outbound > > + * address. > > + * > > + * Clear BAR0 of EP CONTROLLER 1 which contains the HOST1's config and > > + * self scratchpad region (removes inbound ATU configuration). While BAR0 is > > + * the default self scratchpad BAR, an NTB could have other BARs for self > > + * scratchpad (because of reserved BARs). This function can get the exact BAR > > + * used for self scratchpad from epf_ntb_bar[BAR_CONFIG]. > > + * > > + * Please note the self scratchpad region and config region is combined to > > + * a single region and mapped using the same BAR. Also note HOST2's peer > > + * scratchpad is HOST1's self scratchpad. > > + */ > > +static void epf_ntb_config_sspad_bar_clear(struct epf_ntb *ntb) > > +{ > > + struct pci_epf_bar *epf_bar; > > + enum pci_barno barno; > > + > > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > > + epf_bar = &ntb->epf->bar[barno]; > > + > > + pci_epc_clear_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); > > +} > > + > > +/** > > + * epf_ntb_config_sspad_bar_set() - Set Config + Self scratchpad BAR > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > + * Map BAR0 of EP CONTROLLER 1 which contains the HOST1's config and > > + * self scratchpad region. > > + * > > + * Please note the self scratchpad region and config region is combined to > > + * a single region and mapped using the same BAR. > > + */ > > +static int epf_ntb_config_sspad_bar_set(struct epf_ntb *ntb) > > +{ > > + struct pci_epf_bar *epf_bar; > > + enum pci_barno barno; > > + u8 func_no, vfunc_no; > > + struct device *dev; > > + int ret; > > + > > + dev = &ntb->epf->dev; > > + func_no = ntb->epf->func_no; > > + vfunc_no = ntb->epf->vfunc_no; > > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > > + epf_bar = &ntb->epf->bar[barno]; > > + > > + ret = pci_epc_set_bar(ntb->epf->epc, func_no, vfunc_no, epf_bar); > > + if (ret) { > > + dev_err(dev, "inft: Config/Status/SPAD BAR set failed\n"); > > + return ret; > > + } > > + return 0; > > +} > > + > > +/** > > + * epf_ntb_config_spad_bar_free() - Free the physical memory associated with > > + * config + scratchpad region > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + */ > > +static void epf_ntb_config_spad_bar_free(struct epf_ntb *ntb) > > +{ > > + enum pci_barno barno; > > + > > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > > + pci_epf_free_space(ntb->epf, ntb->reg, barno, 0); > > +} > > + > > +/** > > + * epf_ntb_config_spad_bar_alloc() - Allocate memory for config + scratchpad > > + * region > > + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 > > + * > > + * Allocate the Local Memory mentioned in the above diagram. The size of > > + * CONFIG REGION is sizeof(struct epf_ntb_ctrl) and size of SCRATCHPAD REGION > > + * is obtained from "spad-count" configfs entry. > > + */ > > +static int epf_ntb_config_spad_bar_alloc(struct epf_ntb *ntb) > > +{ > > + size_t align; > > + enum pci_barno barno; > > + struct epf_ntb_ctrl *ctrl; > > + u32 spad_size, ctrl_size; > > + u64 size; > > + struct pci_epf *epf = ntb->epf; > > + struct device *dev = &epf->dev; > > + u32 spad_count; > > + void *base; > > + int i; > > + const struct pci_epc_features *epc_features = pci_epc_get_features(epf->epc, > > + epf->func_no, > > + epf->vfunc_no); > > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > > + size = epc_features->bar_fixed_size[barno]; > > + align = epc_features->align; > > + > > + if ((!IS_ALIGNED(size, align))) > > + return -EINVAL; > > + > > + spad_count = ntb->spad_count; > > + > > + ctrl_size = sizeof(struct epf_ntb_ctrl); > > + spad_size = 2 * spad_count * 4; > > + > > + if (!align) { > > + ctrl_size = roundup_pow_of_two(ctrl_size); > > + spad_size = roundup_pow_of_two(spad_size); > > + } else { > > + ctrl_size = ALIGN(ctrl_size, align); > > + spad_size = ALIGN(spad_size, align); > > + } > > + > > + if (!size) > > + size = ctrl_size + spad_size; > > + else if (size < ctrl_size + spad_size) > > + return -EINVAL; > > + > > + base = pci_epf_alloc_space(epf, size, barno, align, 0); > > + if (!base) { > > + dev_err(dev, "intf: Config/Status/SPAD alloc region fail\n"); > > + return -ENOMEM; > > + } > > + > > + ntb->reg = base; > > + > > + ctrl = ntb->reg; > > + ctrl->spad_offset = ctrl_size; > > + > > + ctrl->spad_count = spad_count; > > + ctrl->num_mws = ntb->num_mws; > > + ntb->spad_size = spad_size; > > + > > + ctrl->db_entry_size = 4; > > + > > + for (i = 0; i < ntb->db_count; i++) { > > + ntb->reg->db_data[i] = 1 + i; > > + ntb->reg->db_offset[i] = 0; > > + } > > + > > + return 0; > > +} > > + > > +/** > > + * epf_ntb_configure_interrupt() - Configure MSI/MSI-X capaiblity > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > + * Configure MSI/MSI-X capability for each interface with number of > > + * interrupts equal to "db_count" configfs entry. > > + */ > > +static int epf_ntb_configure_interrupt(struct epf_ntb *ntb) > > +{ > > + const struct pci_epc_features *epc_features; > > + bool msix_capable, msi_capable; > > + u8 func_no, vfunc_no; > > + struct device *dev; > > + u32 db_count; > > + int ret; > > + > > + dev = &ntb->epf->dev; > > + > > + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); > > + msix_capable = epc_features->msix_capable; > > + msi_capable = epc_features->msi_capable; > > + > > + if (!(msix_capable || msi_capable)) { > > I don't think the "msix_capable" and "msi_capable" local variables > really add any readability since they're only used once. > > Not sure about "func_no", "vfunc_no", either. If you keep them, you > could at least assign them earlier and use them in the > pci_epc_get_features() call. > > > + dev_err(dev, "MSI or MSI-X is required for doorbell\n"); > > + return -EINVAL; > > + } > > + > > + func_no = ntb->epf->func_no; > > + vfunc_no = ntb->epf->vfunc_no; > > + > > + db_count = ntb->db_count; > > + if (db_count > MAX_DB_COUNT) { > > + dev_err(dev, "DB count cannot be more than %d\n", MAX_DB_COUNT); > > + return -EINVAL; > > + } > > + > > + ntb->db_count = db_count; > > + > > + if (msi_capable) { > > + ret = pci_epc_set_msi(ntb->epf->epc, func_no, vfunc_no, 16); > > + if (ret) { > > + dev_err(dev, "intf: MSI configuration failed\n"); > > + return ret; > > + } > > + } > > + > > + return 0; > > +} > > + > > +/** > > + * epf_ntb_db_bar_init() - Configure Doorbell window BARs > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > Spurious blank line. > > > + */ > > +static int epf_ntb_db_bar_init(struct epf_ntb *ntb) > > +{ > > + const struct pci_epc_features *epc_features; > > + u32 align; > > + struct device *dev = &ntb->epf->dev; > > + int ret; > > + struct pci_epf_bar *epf_bar; > > + void __iomem *mw_addr; > > + enum pci_barno barno; > > + size_t size = 4 * ntb->db_count; > > + > > + epc_features = pci_epc_get_features(ntb->epf->epc, > > + ntb->epf->func_no, > > + ntb->epf->vfunc_no); > > + align = epc_features->align; > > + > > + if (size < 128) > > + size = 128; > > + > > + if (align) > > + size = ALIGN(size, align); > > + else > > + size = roundup_pow_of_two(size); > > + > > + barno = ntb->epf_ntb_bar[BAR_DB]; > > + > > + mw_addr = pci_epf_alloc_space(ntb->epf, size, barno, align, 0); > > + if (!mw_addr) { > > + dev_err(dev, "intf: Failed to allocate OB address\n"); > > + return -ENOMEM; > > + } > > + > > + ntb->epf_db = mw_addr; > > + > > + epf_bar = &ntb->epf->bar[barno]; > > + > > + ret = pci_epc_set_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); > > + if (ret) { > > + dev_err(dev, "intf: DoorBell BAR set failed\n"); > > s/DoorBell/Doorbell/ > > > + goto err_alloc_peer_mem; > > + } > > + return ret; > > + > > +err_alloc_peer_mem: > > + pci_epc_mem_free_addr(ntb->epf->epc, epf_bar->phys_addr, mw_addr, epf_bar->size); > > + return -1; > > +} > > + > > +/** > > + * epf_ntb_db_bar_clear() - Clear doorbell BAR and free memory > > + * allocated in peers outbound address space > > s/peers/peer's/ > > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + */ > > +static void epf_ntb_db_bar_clear(struct epf_ntb *ntb) > > +{ > > + enum pci_barno barno; > > + > > + barno = ntb->epf_ntb_bar[BAR_DB]; > > + pci_epf_free_space(ntb->epf, ntb->epf_db, barno, 0); > > + pci_epc_clear_bar(ntb->epf->epc, > > + ntb->epf->func_no, > > + ntb->epf->vfunc_no, > > + &ntb->epf->bar[barno]); > > +} > > + > > +/** > > + * epf_ntb_mw_bar_init() - Configure Memory window BARs > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > + */ > > +static int epf_ntb_mw_bar_init(struct epf_ntb *ntb) > > +{ > > + int ret = 0; > > + int i; > > + u64 size; > > + enum pci_barno barno; > > + struct device *dev = &ntb->epf->dev; > > + > > + for (i = 0; i < ntb->num_mws; i++) { > > + > > Spurious blank line. > > > + size = ntb->mws_size[i]; > > + > > + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; > > + > > + ntb->epf->bar[barno].barno = barno; > > + ntb->epf->bar[barno].size = size; > > + ntb->epf->bar[barno].addr = 0; > > + ntb->epf->bar[barno].phys_addr = 0; > > + ntb->epf->bar[barno].flags |= upper_32_bits(size) ? > > + PCI_BASE_ADDRESS_MEM_TYPE_64 : > > + PCI_BASE_ADDRESS_MEM_TYPE_32; > > + > > + ret = pci_epc_set_bar(ntb->epf->epc, > > + ntb->epf->func_no, > > + ntb->epf->vfunc_no, > > + &ntb->epf->bar[barno]); > > + if (ret) { > > + dev_err(dev, "intf: MW set failed\n"); > > + goto err_alloc_mem; > > + } > > + > > + /* allocate epc outbound memory windows to vpci vntb device */ > > s/allocate/Allocate/ to match other comments > s/epc/EPC/ > > > + ntb->vpci_mw_addr[i] = pci_epc_mem_alloc_addr(ntb->epf->epc, > > + &ntb->vpci_mw_phy[i], > > + size); > > + if (!ntb->vpci_mw_addr[i]) { > > + dev_err(dev, "Failed to allocate source address\n"); > > + goto err_alloc_mem; > > + } > > + } > > + > > + return ret; > > +err_alloc_mem: > > + return ret; > > +} > > + > > +/** > > + * epf_ntb_mw_bar_clear() - Clear Memory window BARs > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > Spurious blank line. > > > + */ > > +static void epf_ntb_mw_bar_clear(struct epf_ntb *ntb) > > +{ > > + enum pci_barno barno; > > + int i; > > + > > + for (i = 0; i < ntb->num_mws; i++) { > > + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; > > + pci_epc_clear_bar(ntb->epf->epc, > > + ntb->epf->func_no, > > + ntb->epf->vfunc_no, > > + &ntb->epf->bar[barno]); > > + > > + pci_epc_mem_free_addr(ntb->epf->epc, > > + ntb->vpci_mw_phy[i], > > + ntb->vpci_mw_addr[i], > > + ntb->mws_size[i]); > > + } > > +} > > + > > +/** > > + * epf_ntb_epc_destroy() - Cleanup NTB EPC interface > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > + * Wrapper for epf_ntb_epc_destroy_interface() to cleanup all the NTB interfaces > > + */ > > +static void epf_ntb_epc_destroy(struct epf_ntb *ntb) > > +{ > > + pci_epc_remove_epf(ntb->epf->epc, ntb->epf, 0); > > + pci_epc_put(ntb->epf->epc); > > +} > > + > > +/** > > + * epf_ntb_init_epc_bar() - Identify BARs to be used for each of the NTB > > + * constructs (scratchpad region, doorbell, memorywindow) > > + * @ntb: NTB device that facilitates communication between HOST and vHOST > > + * > > Spurious blank line. > > > + */ > > +static int epf_ntb_init_epc_bar(struct epf_ntb *ntb) > > +{ > > + const struct pci_epc_features *epc_features; > > + enum pci_barno barno; > > + enum epf_ntb_bar bar; > > + struct device *dev; > > + u32 num_mws; > > + int i; > > + > > + barno = BAR_0; > > + num_mws = ntb->num_mws; > > + dev = &ntb->epf->dev; > > + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); > > + > > + /* These are required BARs which are mandatory for NTB functionality */ > > + for (bar = BAR_CONFIG; bar <= BAR_MW0; bar++, barno++) { > > + barno = pci_epc_get_next_free_bar(epc_features, barno); > > + if (barno < 0) { > > + dev_err(dev, "intf: Fail to get NTB function BAR\n"); > > + return barno; > > + } > > + ntb->epf_ntb_bar[bar] = barno; > > + } > > + > > + /* These are optional BARs which don't impact NTB functionality */ > > + for (bar = BAR_MW1, i = 1; i < num_mws; bar++, barno++, i++) { > > + barno = pci_epc_get_next_free_bar(epc_features, barno); > > + if (barno < 0) { > > + ntb->num_mws = i; > > + dev_dbg(dev, "BAR not available for > MW%d\n", i + 1); > > + } > > + ntb->epf_ntb_bar[bar] = barno; > > + } > > + > > + return 0; > > +} > > + > > +/** > > + * epf_ntb_epc_init() - Initialize NTB interface > > + * @ntb: NTB device that facilitates communication between HOST and vHOST2 > > + * > > + * Wrapper to initialize a particular EPC interface and start the workqueue > > + * to check for commands from host. This function will write to the > > + * EP controller HW for configuring it. > > + */ > > +static int epf_ntb_epc_init(struct epf_ntb *ntb) > > +{ > > + u8 func_no, vfunc_no; > > + struct pci_epc *epc; > > + struct pci_epf *epf; > > + struct device *dev; > > + int ret; > > + > > + epf = ntb->epf; > > + dev = &epf->dev; > > + epc = epf->epc; > > + func_no = ntb->epf->func_no; > > + vfunc_no = ntb->epf->vfunc_no; > > + > > + ret = epf_ntb_config_sspad_bar_set(ntb); > > + if (ret) { > > + dev_err(dev, "intf: Config/self SPAD BAR init failed"); > > + return ret; > > + } > > + > > + ret = epf_ntb_configure_interrupt(ntb); > > + if (ret) { > > + dev_err(dev, "intf: Interrupt configuration failed\n"); > > + goto err_config_interrupt; > > + } > > + > > + ret = epf_ntb_db_bar_init(ntb); > > + if (ret) { > > + dev_err(dev, "intf: DB BAR init failed\n"); > > + goto err_db_bar_init; > > + } > > + > > + ret = epf_ntb_mw_bar_init(ntb); > > + if (ret) { > > + dev_err(dev, "intf: MW BAR init failed\n"); > > + goto err_mw_bar_init; > > + } > > + > > + if (vfunc_no <= 1) { > > + ret = pci_epc_write_header(epc, func_no, vfunc_no, epf->header); > > + if (ret) { > > + dev_err(dev, "intf: Configuration header write failed\n"); > > + goto err_write_header; > > + } > > + } > > + > > + INIT_DELAYED_WORK(&ntb->cmd_handler, epf_ntb_cmd_handler); > > + queue_work(kpcintb_workqueue, &ntb->cmd_handler.work); > > + > > + return 0; > > + > > +err_write_header: > > + epf_ntb_mw_bar_clear(ntb); > > +err_mw_bar_init: > > + epf_ntb_db_bar_clear(ntb); > > +err_db_bar_init: > > +err_config_interrupt: > > + epf_ntb_config_sspad_bar_clear(ntb); > > + > > + return ret; > > +} > > + > > + > > +/** > > + * epf_ntb_epc_cleanup() - Cleanup all NTB interfaces > > + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 > > + * > > + * Wrapper to cleanup all NTB interfaces. > > + */ > > +static void epf_ntb_epc_cleanup(struct epf_ntb *ntb) > > +{ > > + epf_ntb_db_bar_clear(ntb); > > + epf_ntb_mw_bar_clear(ntb); > > +} > > + > > +#define EPF_NTB_R(_name) \ > > +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ > > + char *page) \ > > +{ \ > > + struct config_group *group = to_config_group(item); \ > > + struct epf_ntb *ntb = to_epf_ntb(group); \ > > + \ > > + return sprintf(page, "%d\n", ntb->_name); \ > > +} > > + > > +#define EPF_NTB_W(_name) \ > > +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ > > + const char *page, size_t len) \ > > +{ \ > > + struct config_group *group = to_config_group(item); \ > > + struct epf_ntb *ntb = to_epf_ntb(group); \ > > + u32 val; \ > > + int ret; \ > > + \ > > + ret = kstrtou32(page, 0, &val); \ > > + if (ret) \ > > + return ret; \ > > + \ > > + ntb->_name = val; \ > > + \ > > + return len; \ > > +} > > + > > +#define EPF_NTB_MW_R(_name) \ > > +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ > > + char *page) \ > > +{ \ > > + struct config_group *group = to_config_group(item); \ > > + struct epf_ntb *ntb = to_epf_ntb(group); \ > > + int win_no; \ > > + \ > > + sscanf(#_name, "mw%d", &win_no); \ > > + \ > > + return sprintf(page, "%lld\n", ntb->mws_size[win_no - 1]); \ > > +} > > + > > +#define EPF_NTB_MW_W(_name) \ > > +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ > > + const char *page, size_t len) \ > > +{ \ > > + struct config_group *group = to_config_group(item); \ > > + struct epf_ntb *ntb = to_epf_ntb(group); \ > > + struct device *dev = &ntb->epf->dev; \ > > + int win_no; \ > > + u64 val; \ > > + int ret; \ > > + \ > > + ret = kstrtou64(page, 0, &val); \ > > + if (ret) \ > > + return ret; \ > > + \ > > + if (sscanf(#_name, "mw%d", &win_no) != 1) \ > > + return -EINVAL; \ > > + \ > > + if (ntb->num_mws < win_no) { \ > > + dev_err(dev, "Invalid num_nws: %d value\n", ntb->num_mws); \ > > + return -EINVAL; \ > > + } \ > > + \ > > + ntb->mws_size[win_no - 1] = val; \ > > + \ > > + return len; \ > > +} > > + > > +static ssize_t epf_ntb_num_mws_store(struct config_item *item, > > + const char *page, size_t len) > > +{ > > + struct config_group *group = to_config_group(item); > > + struct epf_ntb *ntb = to_epf_ntb(group); > > + u32 val; > > + int ret; > > + > > + ret = kstrtou32(page, 0, &val); > > + if (ret) > > + return ret; > > + > > + if (val > MAX_MW) > > + return -EINVAL; > > + > > + ntb->num_mws = val; > > + > > + return len; > > +} > > + > > +EPF_NTB_R(spad_count) > > +EPF_NTB_W(spad_count) > > +EPF_NTB_R(db_count) > > +EPF_NTB_W(db_count) > > +EPF_NTB_R(num_mws) > > +EPF_NTB_R(vbus_number) > > +EPF_NTB_W(vbus_number) > > +EPF_NTB_MW_R(mw1) > > +EPF_NTB_MW_W(mw1) > > +EPF_NTB_MW_R(mw2) > > +EPF_NTB_MW_W(mw2) > > +EPF_NTB_MW_R(mw3) > > +EPF_NTB_MW_W(mw3) > > +EPF_NTB_MW_R(mw4) > > +EPF_NTB_MW_W(mw4) > > + > > +CONFIGFS_ATTR(epf_ntb_, spad_count); > > +CONFIGFS_ATTR(epf_ntb_, db_count); > > +CONFIGFS_ATTR(epf_ntb_, num_mws); > > +CONFIGFS_ATTR(epf_ntb_, mw1); > > +CONFIGFS_ATTR(epf_ntb_, mw2); > > +CONFIGFS_ATTR(epf_ntb_, mw3); > > +CONFIGFS_ATTR(epf_ntb_, mw4); > > +CONFIGFS_ATTR(epf_ntb_, vbus_number); > > + > > +static struct configfs_attribute *epf_ntb_attrs[] = { > > + &epf_ntb_attr_spad_count, > > + &epf_ntb_attr_db_count, > > + &epf_ntb_attr_num_mws, > > + &epf_ntb_attr_mw1, > > + &epf_ntb_attr_mw2, > > + &epf_ntb_attr_mw3, > > + &epf_ntb_attr_mw4, > > + &epf_ntb_attr_vbus_number, > > + NULL, > > +}; > > + > > +static const struct config_item_type ntb_group_type = { > > + .ct_attrs = epf_ntb_attrs, > > + .ct_owner = THIS_MODULE, > > +}; > > + > > +/** > > + * epf_ntb_add_cfs() - Add configfs directory specific to NTB > > + * @epf: NTB endpoint function device > > + * @group: A pointer to the config_group structure referencing a group of > > + * config_items of a specific type that belong to a specific sub-system. > > + * > > + * Add configfs directory specific to NTB. This directory will hold > > + * NTB specific properties like db_count, spad_count, num_mws etc., > > + */ > > +static struct config_group *epf_ntb_add_cfs(struct pci_epf *epf, > > + struct config_group *group) > > +{ > > + struct epf_ntb *ntb = epf_get_drvdata(epf); > > + struct config_group *ntb_group = &ntb->group; > > + struct device *dev = &epf->dev; > > + > > + config_group_init_type_name(ntb_group, dev_name(dev), &ntb_group_type); > > + > > + return ntb_group; > > +} > > + > > +/*==== virtual PCI bus driver, which only load virutal ntb pci driver ====*/ > > s/virutal ntb pci/virtual NTB PCI/ > > > +#define VPCI_BUS_NUM 0x10 > > How did you pick this number and how do we know it is available? This is problem. Do you know auto detect bus number method? or can I put it to configfs, let user to config it. > > > +uint32_t pci_space[] = { > > "static u32"? > > > + (VNTB_VID | (VNTB_PID << 16)), //DeviceID, Vendor ID > > + 0, // status, Command > > + 0xffffffff, // Class code, subclass, prog if, revision id > > + 0x40, //bist, header type, latency Timer, cache line size > > + 0, //bar 0 > > + 0, //bar 1 > > + 0, //bar 2 > > + 0, //bar 3 > > + 0, //bar 4 > > + 0, //bar 5 > > s/bar/BAR/ > > > + 0, //cardbus cis point > > + 0, //Subsystem ID Subystem vendor id > > + 0, //ROM Base Address > > + 0, //Reserved, Cap. Point > > + 0, //Reserved, > > + 0, //Max Lat, Min Gnt, interrupt pin, interrupt line > > Use /* */ instead of //, like the rest of drivers/pci/ > > > +}; > > + > > +int pci_read(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 *val) > > +{ > > + if (devfn == 0) { > > + memcpy(val, ((uint8_t *)pci_space) + where, size); > > "u8 *"? > > > + return 0; > > + } > > + return -1; > > These should return PCIBIOS_SUCCESSFUL or PCIBIOS_DEVICE_NOT_FOUND. > > > +} > > + > > +int pci_write(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 val) > > +{ > > + return 0; > > +} > > + > > +struct pci_ops vpci_ops = { > > + .read = pci_read, > > + .write = pci_write, > > +}; > > + > > +static int vpci_bus(void *sysdata) > > The function name should say something about what it does. Maybe > vpci_scan_bus() or something? > > > +{ > > + struct pci_bus *vpci_bus; > > + > > + vpci_bus = pci_scan_bus(VPCI_BUS_NUM, &vpci_ops, sysdata); > > + if (vpci_bus) > > + pr_err("create pci bus\n"); > > + > > + pci_bus_add_devices(vpci_bus); > > + > > + return 0; > > +} > > + > > +/*==================== Virtual PCIe NTB driver ==========================*/ > > + > > +static int vntb_epf_mw_count(struct ntb_dev *ntb, int pidx) > > +{ > > + struct epf_ntb *ndev = ntb_ndev(ntb); > > + > > + return ndev->num_mws; > > +} > > + > > +static int vntb_epf_spad_count(struct ntb_dev *ntb) > > +{ > > + return ntb_ndev(ntb)->spad_count; > > +} > > + > > +static int vntb_epf_peer_mw_count(struct ntb_dev *ntb) > > +{ > > + return ntb_ndev(ntb)->num_mws; > > +} > > + > > +static u64 vntb_epf_db_valid_mask(struct ntb_dev *ntb) > > +{ > > + return BIT_ULL(ntb_ndev(ntb)->db_count) - 1; > > +} > > + > > +static int vntb_epf_db_set_mask(struct ntb_dev *ntb, u64 db_bits) > > +{ > > + return 0; > > +} > > + > > +static int vntb_epf_mw_set_trans(struct ntb_dev *ndev, int pidx, int idx, > > + dma_addr_t addr, resource_size_t size) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + struct pci_epf_bar *epf_bar; > > + enum pci_barno barno; > > + int ret; > > + struct device *dev; > > + > > + dev = &ntb->ntb.dev; > > + barno = ntb->epf_ntb_bar[BAR_MW0 + idx]; > > + epf_bar = &ntb->epf->bar[barno]; > > + epf_bar->phys_addr = addr; > > + epf_bar->barno = barno; > > + epf_bar->size = size; > > + > > + ret = pci_epc_set_bar(ntb->epf->epc, 0, 0, epf_bar); > > + if (ret) { > > + dev_err(dev, "failure set mw trans\n"); > > + return ret; > > + } > > + return 0; > > +} > > + > > +static int vntb_epf_mw_clear_trans(struct ntb_dev *ntb, int pidx, int idx) > > +{ > > + return 0; > > +} > > + > > +static int vntb_epf_peer_mw_get_addr(struct ntb_dev *ndev, int idx, > > + phys_addr_t *base, resource_size_t *size) > > +{ > > + > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + > > + if (base) > > + *base = ntb->vpci_mw_phy[idx]; > > + > > + if (size) > > + *size = ntb->mws_size[idx]; > > + > > + return 0; > > +} > > + > > +static int vntb_epf_link_enable(struct ntb_dev *ntb, > > + enum ntb_speed max_speed, > > + enum ntb_width max_width) > > +{ > > + return 0; > > +} > > + > > +static u32 vntb_epf_spad_read(struct ntb_dev *ndev, int idx) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + int off = ntb->reg->spad_offset, ct = ntb->reg->spad_count * 4; > > + u32 val; > > + void __iomem *base = ntb->reg; > > + > > + val = readl(base + off + ct + idx * 4); > > + return val; > > +} > > + > > +static int vntb_epf_spad_write(struct ntb_dev *ndev, int idx, u32 val) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + struct epf_ntb_ctrl *ctrl = ntb->reg; > > + int off = ctrl->spad_offset, ct = ctrl->spad_count * 4; > > + void __iomem *base = ntb->reg; > > + > > + writel(val, base + off + ct + idx * 4); > > + return 0; > > +} > > + > > +static u32 vntb_epf_peer_spad_read(struct ntb_dev *ndev, int pidx, int idx) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + struct epf_ntb_ctrl *ctrl = ntb->reg; > > + int off = ctrl->spad_offset; > > + void __iomem *base = ntb->reg; > > + u32 val; > > + > > + val = readl(base + off + idx * 4); > > + return val; > > +} > > + > > +static int vntb_epf_peer_spad_write(struct ntb_dev *ndev, int pidx, int idx, u32 val) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + struct epf_ntb_ctrl *ctrl = ntb->reg; > > + int off = ctrl->spad_offset; > > + void __iomem *base = ntb->reg; > > + > > + writel(val, base + off + idx * 4); > > + return 0; > > +} > > + > > +static int vntb_epf_peer_db_set(struct ntb_dev *ndev, u64 db_bits) > > +{ > > + u32 interrupt_num = ffs(db_bits) + 1; > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + u8 func_no, vfunc_no; > > + int ret; > > + > > + func_no = ntb->epf->func_no; > > + vfunc_no = ntb->epf->vfunc_no; > > + > > + ret = pci_epc_raise_irq(ntb->epf->epc, > > + func_no, > > + vfunc_no, > > + PCI_EPC_IRQ_MSI, > > + interrupt_num + 1); > > + if (ret) { > > + dev_err(&ntb->ntb.dev, "intf: Failed to raise IRQ\n"); > > + return ret; > > + } > > + > > + return 0; > > Equivalent to: > > if (ret) > dev_err(...); > > return ret; > > > +} > > + > > +static u64 vntb_epf_db_read(struct ntb_dev *ndev) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + > > + return ntb->db; > > +} > > + > > +static int vntb_epf_mw_get_align(struct ntb_dev *ndev, int pidx, int idx, > > + resource_size_t *addr_align, > > + resource_size_t *size_align, > > + resource_size_t *size_max) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + > > + if (addr_align) > > + *addr_align = SZ_4K; > > + > > + if (size_align) > > + *size_align = 1; > > + > > + if (size_max) > > + *size_max = ntb->mws_size[idx]; > > + > > + return 0; > > +} > > + > > +static u64 vntb_epf_link_is_up(struct ntb_dev *ndev, > > + enum ntb_speed *speed, > > + enum ntb_width *width) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + > > + return ntb->reg->link_status; > > +} > > + > > +static int vntb_epf_db_clear_mask(struct ntb_dev *ndev, u64 db_bits) > > +{ > > + return 0; > > +} > > + > > +static int vntb_epf_db_clear(struct ntb_dev *ndev, u64 db_bits) > > +{ > > + struct epf_ntb *ntb = ntb_ndev(ndev); > > + > > + ntb->db &= ~db_bits; > > + return 0; > > +} > > + > > +static int vntb_epf_link_disable(struct ntb_dev *ntb) > > +{ > > + return 0; > > +} > > + > > +static const struct ntb_dev_ops vntb_epf_ops = { > > + .mw_count = vntb_epf_mw_count, > > + .spad_count = vntb_epf_spad_count, > > + .peer_mw_count = vntb_epf_peer_mw_count, > > + .db_valid_mask = vntb_epf_db_valid_mask, > > + .db_set_mask = vntb_epf_db_set_mask, > > + .mw_set_trans = vntb_epf_mw_set_trans, > > + .mw_clear_trans = vntb_epf_mw_clear_trans, > > + .peer_mw_get_addr = vntb_epf_peer_mw_get_addr, > > + .link_enable = vntb_epf_link_enable, > > + .spad_read = vntb_epf_spad_read, > > + .spad_write = vntb_epf_spad_write, > > + .peer_spad_read = vntb_epf_peer_spad_read, > > + .peer_spad_write = vntb_epf_peer_spad_write, > > + .peer_db_set = vntb_epf_peer_db_set, > > + .db_read = vntb_epf_db_read, > > + .mw_get_align = vntb_epf_mw_get_align, > > + .link_is_up = vntb_epf_link_is_up, > > + .db_clear_mask = vntb_epf_db_clear_mask, > > + .db_clear = vntb_epf_db_clear, > > + .link_disable = vntb_epf_link_disable, > > +}; > > + > > +static int pci_vntb_probe(struct pci_dev *pdev, const struct pci_device_id *id) > > +{ > > + int ret; > > + struct epf_ntb *ndev = (struct epf_ntb *)pdev->sysdata; > > + struct device *dev = &pdev->dev; > > + > > + ndev->ntb.pdev = pdev; > > + ndev->ntb.topo = NTB_TOPO_NONE; > > + ndev->ntb.ops = &vntb_epf_ops; > > + > > + ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); > > + if (ret) { > > + dev_err(dev, "Cannot set DMA mask\n"); > > + return -1; > > return -EINVAL; > > > + } > > + > > + ret = ntb_register_device(&ndev->ntb); > > + if (ret) { > > + dev_err(dev, "Failed to register NTB device\n"); > > + goto err_register_dev; > > + } > > + > > + dev_info(dev, "PCI Virtual NTB driver loaded\n"); > > + return 0; > > + > > +err_register_dev: > > + return -1; > > return -EINVAL; > > > +} > > + > > +static const struct pci_device_id pci_vntb_table[] = { > > + { > > + PCI_DEVICE(VNTB_VID, VNTB_PID), > > + }, > > + {}, > > +}; > > + > > +static struct pci_driver vntb_pci_driver = { > > + .name = "pci-vntb", > > + .id_table = pci_vntb_table, > > + .probe = pci_vntb_probe, > > +}; > > + > > +/* ============ PCIe EPF Driver Bind ====================*/ > > + > > +/** > > + * epf_ntb_bind() - Initialize endpoint controller to provide NTB functionality > > + * @epf: NTB endpoint function device > > + * > > + * Initialize both the endpoint controllers associated with NTB function device. > > + * Invoked when a primary interface or secondary interface is bound to EPC > > + * device. This function will succeed only when EPC is bound to both the > > + * interfaces. > > + */ > > +static int epf_ntb_bind(struct pci_epf *epf) > > +{ > > + struct epf_ntb *ntb = epf_get_drvdata(epf); > > + struct device *dev = &epf->dev; > > + int ret; > > + > > + if (!epf->epc) { > > + dev_dbg(dev, "PRIMARY EPC interface not yet bound\n"); > > + return 0; > > + } > > + > > + ret = epf_ntb_init_epc_bar(ntb); > > + if (ret) { > > + dev_err(dev, "Failed to create NTB EPC\n"); > > + goto err_bar_init; > > + } > > + > > + ret = epf_ntb_config_spad_bar_alloc(ntb); > > + if (ret) { > > + dev_err(dev, "Failed to allocate BAR memory\n"); > > + goto err_bar_alloc; > > + } > > + > > + ret = epf_ntb_epc_init(ntb); > > + if (ret) { > > + dev_err(dev, "Failed to initialize EPC\n"); > > + goto err_bar_alloc; > > + } > > + > > + epf_set_drvdata(epf, ntb); > > + > > + if (pci_register_driver(&vntb_pci_driver)) { > > + dev_err(dev, "failure register vntb pci driver\n"); > > + goto err_bar_alloc; > > + } > > + > > + vpci_bus(ntb); > > + > > + return 0; > > + > > +err_bar_alloc: > > + epf_ntb_config_spad_bar_free(ntb); > > + > > +err_bar_init: > > + epf_ntb_epc_destroy(ntb); > > + > > + return ret; > > +} > > + > > +/** > > + * epf_ntb_unbind() - Cleanup the initialization from epf_ntb_bind() > > + * @epf: NTB endpoint function device > > + * > > + * Cleanup the initialization from epf_ntb_bind() > > + */ > > +static void epf_ntb_unbind(struct pci_epf *epf) > > +{ > > + struct epf_ntb *ntb = epf_get_drvdata(epf); > > + > > + epf_ntb_epc_cleanup(ntb); > > + epf_ntb_config_spad_bar_free(ntb); > > + epf_ntb_epc_destroy(ntb); > > + > > + pci_unregister_driver(&vntb_pci_driver); > > +} > > + > > +// EPF driver probe > > +static struct pci_epf_ops epf_ntb_ops = { > > + .bind = epf_ntb_bind, > > + .unbind = epf_ntb_unbind, > > + .add_cfs = epf_ntb_add_cfs, > > +}; > > + > > +/** > > + * epf_ntb_probe() - Probe NTB function driver > > + * @epf: NTB endpoint function device > > + * > > + * Probe NTB function driver when endpoint function bus detects a NTB > > + * endpoint function. > > + */ > > +static int epf_ntb_probe(struct pci_epf *epf) > > +{ > > + struct epf_ntb *ntb; > > + struct device *dev; > > + > > + dev = &epf->dev; > > + > > + ntb = devm_kzalloc(dev, sizeof(*ntb), GFP_KERNEL); > > + if (!ntb) > > + return -ENOMEM; > > + > > + epf->header = &epf_ntb_header; > > + ntb->epf = epf; > > + epf_set_drvdata(epf, ntb); > > + > > + dev_info(dev, "pci-ep epf driver loaded\n"); > > I think most drivers don't announce when they're loaded because it > really doesn't help the user. > > > + return 0; > > +} > > + > > +static const struct pci_epf_device_id epf_ntb_ids[] = { > > + { > > + .name = "pci_epf_vntb", > > + }, > > + {}, > > +}; > > + > > +static struct pci_epf_driver epf_ntb_driver = { > > + .driver.name = "pci_epf_vntb", > > + .probe = epf_ntb_probe, > > + .id_table = epf_ntb_ids, > > + .ops = &epf_ntb_ops, > > + .owner = THIS_MODULE, > > +}; > > + > > + > > Spurious blank line. > > > +static int __init epf_ntb_init(void) > > +{ > > + int ret; > > + > > + kpcintb_workqueue = alloc_workqueue("kpcintb", WQ_MEM_RECLAIM | > > + WQ_HIGHPRI, 0); > > + ret = pci_epf_register_driver(&epf_ntb_driver); > > + if (ret) { > > + destroy_workqueue(kpcintb_workqueue); > > + pr_err("Failed to register pci epf ntb driver --> %d\n", ret); > > + return ret; > > + } > > + > > + return 0; > > +} > > +module_init(epf_ntb_init); > > + > > +static void __exit epf_ntb_exit(void) > > +{ > > + pci_epf_unregister_driver(&epf_ntb_driver); > > + destroy_workqueue(kpcintb_workqueue); > > +} > > +module_exit(epf_ntb_exit); > > + > > +MODULE_DESCRIPTION("PCI EPF NTB DRIVER"); > > +MODULE_AUTHOR("Frank Li <Frank.li@nxp.com>"); > > +MODULE_LICENSE("GPL v2"); > > -- > > 2.24.0.rc1 > >
diff --git a/drivers/pci/endpoint/functions/Kconfig b/drivers/pci/endpoint/functions/Kconfig index 5f1242ca2f4e4..362555b024e8f 100644 --- a/drivers/pci/endpoint/functions/Kconfig +++ b/drivers/pci/endpoint/functions/Kconfig @@ -25,3 +25,14 @@ config PCI_EPF_NTB device tree. If in doubt, say "N" to disable Endpoint NTB driver. + +config PCI_EPF_VNTB + tristate "PCI Endpoint NTB driver" + depends on PCI_ENDPOINT + select CONFIGFS_FS + help + Select this configuration option to enable the Non-Transparent + Bridge (NTB) driver for PCI Endpoint. NTB driver implements NTB + between PCI host and PCIe Endpoint. + + If in doubt, say "N" to disable Endpoint NTB driver. diff --git a/drivers/pci/endpoint/functions/Makefile b/drivers/pci/endpoint/functions/Makefile index 96ab932a537a2..5c13001deaba1 100644 --- a/drivers/pci/endpoint/functions/Makefile +++ b/drivers/pci/endpoint/functions/Makefile @@ -5,3 +5,4 @@ obj-$(CONFIG_PCI_EPF_TEST) += pci-epf-test.o obj-$(CONFIG_PCI_EPF_NTB) += pci-epf-ntb.o +obj-$(CONFIG_PCI_EPF_VNTB) += pci-epf-vntb.o diff --git a/drivers/pci/endpoint/functions/pci-epf-vntb.c b/drivers/pci/endpoint/functions/pci-epf-vntb.c new file mode 100644 index 0000000000000..ebf7e243eefa4 --- /dev/null +++ b/drivers/pci/endpoint/functions/pci-epf-vntb.c @@ -0,0 +1,1425 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * Endpoint Function Driver to implement Non-Transparent Bridge functionality + * Between PCI RC and EP + * + * Copyright (C) 2020 Texas Instruments + * Copyright (C) 2022 NXP + * + * Based on pci-epf-ntb.c + * Author: Frank Li <Frank.Li@nxp.com> + * Author: Kishon Vijay Abraham I <kishon@ti.com> + */ + +/** + * +------------+ +---------------------------------------+ + * | | | | + * +------------+ | +--------------+ + * | NTB | | | NTB | + * | NetDev | | | NetDev | + * +------------+ | +--------------+ + * | NTB | | | NTB | + * | Transfer | | | Transfer | + * +------------+ | +--------------+ + * | | | | | + * | PCI NTB | | | | + * | EPF | | | | + * | Driver | | | PCI Virtual | + * | | +---------------+ | NTB Driver | + * | | | PCI EP NTB |<------>| | + * | | | FN Driver | | | + * +------------+ +---------------+ +--------------+ + * | | | | | | + * | PCI BUS | <-----> | PCI EP BUS | | Virtual PCI | + * | | PCI | | | BUS | + * +------------+ +---------------+--------+--------------+ + * PCI RC PCI EP + */ + +#include <linux/delay.h> +#include <linux/io.h> +#include <linux/module.h> +#include <linux/slab.h> + +#include <linux/pci-epc.h> +#include <linux/pci-epf.h> +#include <linux/ntb.h> + +static struct workqueue_struct *kpcintb_workqueue; + +#define COMMAND_CONFIGURE_DOORBELL 1 +#define COMMAND_TEARDOWN_DOORBELL 2 +#define COMMAND_CONFIGURE_MW 3 +#define COMMAND_TEARDOWN_MW 4 +#define COMMAND_LINK_UP 5 +#define COMMAND_LINK_DOWN 6 + +#define COMMAND_STATUS_OK 1 +#define COMMAND_STATUS_ERROR 2 + +#define LINK_STATUS_UP BIT(0) + +#define SPAD_COUNT 64 +#define DB_COUNT 4 +#define NTB_MW_OFFSET 2 +#define DB_COUNT_MASK GENMASK(15, 0) +#define MSIX_ENABLE BIT(16) +#define MAX_DB_COUNT 32 +#define MAX_MW 4 + +#define VNTB_VID 0x1957 +#define VNTB_PID 0x080A + +enum epf_ntb_bar { + BAR_CONFIG, + BAR_DB, + BAR_MW0, + BAR_MW1, + BAR_MW2, +}; + +/* + * +--------------------------------------------------+ Base + * | | + * | | + * | | + * | Common Control Register | + * | | + * | | + * | | + * +-----------------------+--------------------------+ Base+span_offset + * | | | + * | Peer Span Space | Span Space | + * | | | + * | | | + * +-----------------------+--------------------------+ Base+span_offset + * | | | +span_count * 4 + * | | | + * | Span Space | Peer Span Space | + * | | | + * +-----------------------+--------------------------+ + * Virtual PCI Pcie Endpoint + * NTB Driver NTB Driver + */ +struct epf_ntb_ctrl { + u32 command; + u32 argument; + u16 command_status; + u16 link_status; + u32 topology; + u64 addr; + u64 size; + u32 num_mws; + u32 reserved; + u32 spad_offset; + u32 spad_count; + u32 db_entry_size; + u32 db_data[MAX_DB_COUNT]; + u32 db_offset[MAX_DB_COUNT]; +} __packed; + +struct epf_ntb { + struct ntb_dev ntb; + struct pci_epf *epf; + struct config_group group; + + u32 num_mws; + u32 db_count; + u32 spad_count; + u64 mws_size[MAX_MW]; + u64 db; + u32 vbus_number; + + bool linkup; + u32 spad_size; + + enum pci_barno epf_ntb_bar[6]; + + struct epf_ntb_ctrl *reg; + + phys_addr_t epf_db_phy; + void __iomem *epf_db; + + phys_addr_t vpci_mw_phy[MAX_MW]; + void __iomem *vpci_mw_addr[MAX_MW]; + + struct delayed_work cmd_handler; +}; + +#define to_epf_ntb(epf_group) container_of((epf_group), struct epf_ntb, group) +#define ntb_ndev(__ntb) container_of(__ntb, struct epf_ntb, ntb) + +static struct pci_epf_header epf_ntb_header = { + .vendorid = PCI_ANY_ID, + .deviceid = PCI_ANY_ID, + .baseclass_code = PCI_BASE_CLASS_MEMORY, + .interrupt_pin = PCI_INTERRUPT_INTA, +}; + +/** + * epf_ntb_link_up() - Raise link_up interrupt to Virtual Host + * @ntb: NTB device that facilitates communication between HOST and VHOST + * @link_up: true or false indicating Link is UP or Down + * + * Once NTB function in HOST invoke ntb_link_enable(), + * this NTB function driver will trigger a link event to vhost. + */ +static int epf_ntb_link_up(struct epf_ntb *ntb, bool link_up) +{ + if (link_up) + ntb->reg->link_status |= LINK_STATUS_UP; + else + ntb->reg->link_status &= ~LINK_STATUS_UP; + + ntb_link_event(&ntb->ntb); + return 0; +} + +/** + * epf_ntb_configure_mw() - Configure the Outbound Address Space for vhost + * to access the memory window of host + * @ntb: NTB device that facilitates communication between host and vhost + * @mw: Index of the memory window (either 0, 1, 2 or 3) + * + * EP Outbound Window + * +--------+ +-----------+ + * | | | | + * | | | | + * | | | | + * | | | | + * | | +-----------+ + * | Virtual| | Memory Win| + * | NTB | -----------> | | + * | Driver | | | + * | | +-----------+ + * | | | | + * | | | | + * +--------+ +-----------+ + * VHost PCI EP + */ +static int epf_ntb_configure_mw(struct epf_ntb *ntb, u32 mw) +{ + phys_addr_t phys_addr; + u8 func_no, vfunc_no; + u64 addr, size; + int ret = 0; + + phys_addr = ntb->vpci_mw_phy[mw]; + addr = ntb->reg->addr; + size = ntb->reg->size; + + func_no = ntb->epf->func_no; + vfunc_no = ntb->epf->vfunc_no; + + ret = pci_epc_map_addr(ntb->epf->epc, func_no, vfunc_no, phys_addr, addr, size); + if (ret) + dev_err(&ntb->epf->epc->dev, + "intf: Failed to map memory window %d address\n", mw); + return ret; +} + +/** + * epf_ntb_teardown_mw() - Teardown the configured OB ATU + * @ntb: NTB device that facilitates communication between HOST and vHOST + * @mw: Index of the memory window (either 0, 1, 2 or 3) + * + * Teardown the configured OB ATU configured in epf_ntb_configure_mw() using + * pci_epc_unmap_addr() + */ +static void epf_ntb_teardown_mw(struct epf_ntb *ntb, u32 mw) +{ + pci_epc_unmap_addr(ntb->epf->epc, + ntb->epf->func_no, + ntb->epf->vfunc_no, + ntb->vpci_mw_phy[mw]); +} + +/** + * epf_ntb_cmd_handler() - Handle commands provided by the NTB Host + * @work: work_struct for the epf_ntb_epc + * + * Workqueue function that gets invoked for the two epf_ntb_epc + * periodically (once every 5ms) to see if it has received any commands + * from NTB host. The host can send commands to configure doorbell or + * configure memory window or to update link status. + */ +static void epf_ntb_cmd_handler(struct work_struct *work) +{ + struct epf_ntb_ctrl *ctrl; + u32 command, argument; + struct epf_ntb *ntb; + struct device *dev; + int ret; + int i; + + ntb = container_of(work, struct epf_ntb, cmd_handler.work); + + for (i = 1; i < ntb->db_count; i++) { + if (readl(ntb->epf_db + i * 4)) { + if (readl(ntb->epf_db + i * 4)) + ntb->db |= 1 << (i - 1); + + ntb_db_event(&ntb->ntb, i); + writel(0, ntb->epf_db + i * 4); + } + } + + ctrl = ntb->reg; + command = ctrl->command; + if (!command) + goto reset_handler; + argument = ctrl->argument; + + ctrl->command = 0; + ctrl->argument = 0; + + ctrl = ntb->reg; + dev = &ntb->epf->dev; + + switch (command) { + case COMMAND_CONFIGURE_DOORBELL: + ctrl->command_status = COMMAND_STATUS_OK; + break; + case COMMAND_TEARDOWN_DOORBELL: + ctrl->command_status = COMMAND_STATUS_OK; + break; + case COMMAND_CONFIGURE_MW: + ret = epf_ntb_configure_mw(ntb, argument); + if (ret < 0) + ctrl->command_status = COMMAND_STATUS_ERROR; + else + ctrl->command_status = COMMAND_STATUS_OK; + break; + case COMMAND_TEARDOWN_MW: + epf_ntb_teardown_mw(ntb, argument); + ctrl->command_status = COMMAND_STATUS_OK; + break; + case COMMAND_LINK_UP: + ntb->linkup = true; + ret = epf_ntb_link_up(ntb, true); + if (ret < 0) + ctrl->command_status = COMMAND_STATUS_ERROR; + else + ctrl->command_status = COMMAND_STATUS_OK; + goto reset_handler; + case COMMAND_LINK_DOWN: + ntb->linkup = false; + ret = epf_ntb_link_up(ntb, false); + if (ret < 0) + ctrl->command_status = COMMAND_STATUS_ERROR; + else + ctrl->command_status = COMMAND_STATUS_OK; + break; + default: + dev_err(dev, "intf UNKNOWN command: %d\n", command); + break; + } + +reset_handler: + queue_delayed_work(kpcintb_workqueue, &ntb->cmd_handler, + msecs_to_jiffies(5)); +} + +/** + * epf_ntb_config_sspad_bar_clear() - Clear Config + Self scratchpad BAR + * @ntb_epc: EPC associated with one of the HOST which holds peer's outbound + * address. + * + * Clear BAR0 of EP CONTROLLER 1 which contains the HOST1's config and + * self scratchpad region (removes inbound ATU configuration). While BAR0 is + * the default self scratchpad BAR, an NTB could have other BARs for self + * scratchpad (because of reserved BARs). This function can get the exact BAR + * used for self scratchpad from epf_ntb_bar[BAR_CONFIG]. + * + * Please note the self scratchpad region and config region is combined to + * a single region and mapped using the same BAR. Also note HOST2's peer + * scratchpad is HOST1's self scratchpad. + */ +static void epf_ntb_config_sspad_bar_clear(struct epf_ntb *ntb) +{ + struct pci_epf_bar *epf_bar; + enum pci_barno barno; + + barno = ntb->epf_ntb_bar[BAR_CONFIG]; + epf_bar = &ntb->epf->bar[barno]; + + pci_epc_clear_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); +} + +/** + * epf_ntb_config_sspad_bar_set() - Set Config + Self scratchpad BAR + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + * Map BAR0 of EP CONTROLLER 1 which contains the HOST1's config and + * self scratchpad region. + * + * Please note the self scratchpad region and config region is combined to + * a single region and mapped using the same BAR. + */ +static int epf_ntb_config_sspad_bar_set(struct epf_ntb *ntb) +{ + struct pci_epf_bar *epf_bar; + enum pci_barno barno; + u8 func_no, vfunc_no; + struct device *dev; + int ret; + + dev = &ntb->epf->dev; + func_no = ntb->epf->func_no; + vfunc_no = ntb->epf->vfunc_no; + barno = ntb->epf_ntb_bar[BAR_CONFIG]; + epf_bar = &ntb->epf->bar[barno]; + + ret = pci_epc_set_bar(ntb->epf->epc, func_no, vfunc_no, epf_bar); + if (ret) { + dev_err(dev, "inft: Config/Status/SPAD BAR set failed\n"); + return ret; + } + return 0; +} + +/** + * epf_ntb_config_spad_bar_free() - Free the physical memory associated with + * config + scratchpad region + * @ntb: NTB device that facilitates communication between HOST and vHOST + */ +static void epf_ntb_config_spad_bar_free(struct epf_ntb *ntb) +{ + enum pci_barno barno; + + barno = ntb->epf_ntb_bar[BAR_CONFIG]; + pci_epf_free_space(ntb->epf, ntb->reg, barno, 0); +} + +/** + * epf_ntb_config_spad_bar_alloc() - Allocate memory for config + scratchpad + * region + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 + * + * Allocate the Local Memory mentioned in the above diagram. The size of + * CONFIG REGION is sizeof(struct epf_ntb_ctrl) and size of SCRATCHPAD REGION + * is obtained from "spad-count" configfs entry. + */ +static int epf_ntb_config_spad_bar_alloc(struct epf_ntb *ntb) +{ + size_t align; + enum pci_barno barno; + struct epf_ntb_ctrl *ctrl; + u32 spad_size, ctrl_size; + u64 size; + struct pci_epf *epf = ntb->epf; + struct device *dev = &epf->dev; + u32 spad_count; + void *base; + int i; + const struct pci_epc_features *epc_features = pci_epc_get_features(epf->epc, + epf->func_no, + epf->vfunc_no); + barno = ntb->epf_ntb_bar[BAR_CONFIG]; + size = epc_features->bar_fixed_size[barno]; + align = epc_features->align; + + if ((!IS_ALIGNED(size, align))) + return -EINVAL; + + spad_count = ntb->spad_count; + + ctrl_size = sizeof(struct epf_ntb_ctrl); + spad_size = 2 * spad_count * 4; + + if (!align) { + ctrl_size = roundup_pow_of_two(ctrl_size); + spad_size = roundup_pow_of_two(spad_size); + } else { + ctrl_size = ALIGN(ctrl_size, align); + spad_size = ALIGN(spad_size, align); + } + + if (!size) + size = ctrl_size + spad_size; + else if (size < ctrl_size + spad_size) + return -EINVAL; + + base = pci_epf_alloc_space(epf, size, barno, align, 0); + if (!base) { + dev_err(dev, "intf: Config/Status/SPAD alloc region fail\n"); + return -ENOMEM; + } + + ntb->reg = base; + + ctrl = ntb->reg; + ctrl->spad_offset = ctrl_size; + + ctrl->spad_count = spad_count; + ctrl->num_mws = ntb->num_mws; + ntb->spad_size = spad_size; + + ctrl->db_entry_size = 4; + + for (i = 0; i < ntb->db_count; i++) { + ntb->reg->db_data[i] = 1 + i; + ntb->reg->db_offset[i] = 0; + } + + return 0; +} + +/** + * epf_ntb_configure_interrupt() - Configure MSI/MSI-X capaiblity + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + * Configure MSI/MSI-X capability for each interface with number of + * interrupts equal to "db_count" configfs entry. + */ +static int epf_ntb_configure_interrupt(struct epf_ntb *ntb) +{ + const struct pci_epc_features *epc_features; + bool msix_capable, msi_capable; + u8 func_no, vfunc_no; + struct device *dev; + u32 db_count; + int ret; + + dev = &ntb->epf->dev; + + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); + msix_capable = epc_features->msix_capable; + msi_capable = epc_features->msi_capable; + + if (!(msix_capable || msi_capable)) { + dev_err(dev, "MSI or MSI-X is required for doorbell\n"); + return -EINVAL; + } + + func_no = ntb->epf->func_no; + vfunc_no = ntb->epf->vfunc_no; + + db_count = ntb->db_count; + if (db_count > MAX_DB_COUNT) { + dev_err(dev, "DB count cannot be more than %d\n", MAX_DB_COUNT); + return -EINVAL; + } + + ntb->db_count = db_count; + + if (msi_capable) { + ret = pci_epc_set_msi(ntb->epf->epc, func_no, vfunc_no, 16); + if (ret) { + dev_err(dev, "intf: MSI configuration failed\n"); + return ret; + } + } + + return 0; +} + +/** + * epf_ntb_db_bar_init() - Configure Doorbell window BARs + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + */ +static int epf_ntb_db_bar_init(struct epf_ntb *ntb) +{ + const struct pci_epc_features *epc_features; + u32 align; + struct device *dev = &ntb->epf->dev; + int ret; + struct pci_epf_bar *epf_bar; + void __iomem *mw_addr; + enum pci_barno barno; + size_t size = 4 * ntb->db_count; + + epc_features = pci_epc_get_features(ntb->epf->epc, + ntb->epf->func_no, + ntb->epf->vfunc_no); + align = epc_features->align; + + if (size < 128) + size = 128; + + if (align) + size = ALIGN(size, align); + else + size = roundup_pow_of_two(size); + + barno = ntb->epf_ntb_bar[BAR_DB]; + + mw_addr = pci_epf_alloc_space(ntb->epf, size, barno, align, 0); + if (!mw_addr) { + dev_err(dev, "intf: Failed to allocate OB address\n"); + return -ENOMEM; + } + + ntb->epf_db = mw_addr; + + epf_bar = &ntb->epf->bar[barno]; + + ret = pci_epc_set_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); + if (ret) { + dev_err(dev, "intf: DoorBell BAR set failed\n"); + goto err_alloc_peer_mem; + } + return ret; + +err_alloc_peer_mem: + pci_epc_mem_free_addr(ntb->epf->epc, epf_bar->phys_addr, mw_addr, epf_bar->size); + return -1; +} + +/** + * epf_ntb_db_bar_clear() - Clear doorbell BAR and free memory + * allocated in peers outbound address space + * @ntb: NTB device that facilitates communication between HOST and vHOST + */ +static void epf_ntb_db_bar_clear(struct epf_ntb *ntb) +{ + enum pci_barno barno; + + barno = ntb->epf_ntb_bar[BAR_DB]; + pci_epf_free_space(ntb->epf, ntb->epf_db, barno, 0); + pci_epc_clear_bar(ntb->epf->epc, + ntb->epf->func_no, + ntb->epf->vfunc_no, + &ntb->epf->bar[barno]); +} + +/** + * epf_ntb_mw_bar_init() - Configure Memory window BARs + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + */ +static int epf_ntb_mw_bar_init(struct epf_ntb *ntb) +{ + int ret = 0; + int i; + u64 size; + enum pci_barno barno; + struct device *dev = &ntb->epf->dev; + + for (i = 0; i < ntb->num_mws; i++) { + + size = ntb->mws_size[i]; + + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; + + ntb->epf->bar[barno].barno = barno; + ntb->epf->bar[barno].size = size; + ntb->epf->bar[barno].addr = 0; + ntb->epf->bar[barno].phys_addr = 0; + ntb->epf->bar[barno].flags |= upper_32_bits(size) ? + PCI_BASE_ADDRESS_MEM_TYPE_64 : + PCI_BASE_ADDRESS_MEM_TYPE_32; + + ret = pci_epc_set_bar(ntb->epf->epc, + ntb->epf->func_no, + ntb->epf->vfunc_no, + &ntb->epf->bar[barno]); + if (ret) { + dev_err(dev, "intf: MW set failed\n"); + goto err_alloc_mem; + } + + /* allocate epc outbound memory windows to vpci vntb device */ + ntb->vpci_mw_addr[i] = pci_epc_mem_alloc_addr(ntb->epf->epc, + &ntb->vpci_mw_phy[i], + size); + if (!ntb->vpci_mw_addr[i]) { + dev_err(dev, "Failed to allocate source address\n"); + goto err_alloc_mem; + } + } + + return ret; +err_alloc_mem: + return ret; +} + +/** + * epf_ntb_mw_bar_clear() - Clear Memory window BARs + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + */ +static void epf_ntb_mw_bar_clear(struct epf_ntb *ntb) +{ + enum pci_barno barno; + int i; + + for (i = 0; i < ntb->num_mws; i++) { + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; + pci_epc_clear_bar(ntb->epf->epc, + ntb->epf->func_no, + ntb->epf->vfunc_no, + &ntb->epf->bar[barno]); + + pci_epc_mem_free_addr(ntb->epf->epc, + ntb->vpci_mw_phy[i], + ntb->vpci_mw_addr[i], + ntb->mws_size[i]); + } +} + +/** + * epf_ntb_epc_destroy() - Cleanup NTB EPC interface + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + * Wrapper for epf_ntb_epc_destroy_interface() to cleanup all the NTB interfaces + */ +static void epf_ntb_epc_destroy(struct epf_ntb *ntb) +{ + pci_epc_remove_epf(ntb->epf->epc, ntb->epf, 0); + pci_epc_put(ntb->epf->epc); +} + +/** + * epf_ntb_init_epc_bar() - Identify BARs to be used for each of the NTB + * constructs (scratchpad region, doorbell, memorywindow) + * @ntb: NTB device that facilitates communication between HOST and vHOST + * + */ +static int epf_ntb_init_epc_bar(struct epf_ntb *ntb) +{ + const struct pci_epc_features *epc_features; + enum pci_barno barno; + enum epf_ntb_bar bar; + struct device *dev; + u32 num_mws; + int i; + + barno = BAR_0; + num_mws = ntb->num_mws; + dev = &ntb->epf->dev; + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); + + /* These are required BARs which are mandatory for NTB functionality */ + for (bar = BAR_CONFIG; bar <= BAR_MW0; bar++, barno++) { + barno = pci_epc_get_next_free_bar(epc_features, barno); + if (barno < 0) { + dev_err(dev, "intf: Fail to get NTB function BAR\n"); + return barno; + } + ntb->epf_ntb_bar[bar] = barno; + } + + /* These are optional BARs which don't impact NTB functionality */ + for (bar = BAR_MW1, i = 1; i < num_mws; bar++, barno++, i++) { + barno = pci_epc_get_next_free_bar(epc_features, barno); + if (barno < 0) { + ntb->num_mws = i; + dev_dbg(dev, "BAR not available for > MW%d\n", i + 1); + } + ntb->epf_ntb_bar[bar] = barno; + } + + return 0; +} + +/** + * epf_ntb_epc_init() - Initialize NTB interface + * @ntb: NTB device that facilitates communication between HOST and vHOST2 + * + * Wrapper to initialize a particular EPC interface and start the workqueue + * to check for commands from host. This function will write to the + * EP controller HW for configuring it. + */ +static int epf_ntb_epc_init(struct epf_ntb *ntb) +{ + u8 func_no, vfunc_no; + struct pci_epc *epc; + struct pci_epf *epf; + struct device *dev; + int ret; + + epf = ntb->epf; + dev = &epf->dev; + epc = epf->epc; + func_no = ntb->epf->func_no; + vfunc_no = ntb->epf->vfunc_no; + + ret = epf_ntb_config_sspad_bar_set(ntb); + if (ret) { + dev_err(dev, "intf: Config/self SPAD BAR init failed"); + return ret; + } + + ret = epf_ntb_configure_interrupt(ntb); + if (ret) { + dev_err(dev, "intf: Interrupt configuration failed\n"); + goto err_config_interrupt; + } + + ret = epf_ntb_db_bar_init(ntb); + if (ret) { + dev_err(dev, "intf: DB BAR init failed\n"); + goto err_db_bar_init; + } + + ret = epf_ntb_mw_bar_init(ntb); + if (ret) { + dev_err(dev, "intf: MW BAR init failed\n"); + goto err_mw_bar_init; + } + + if (vfunc_no <= 1) { + ret = pci_epc_write_header(epc, func_no, vfunc_no, epf->header); + if (ret) { + dev_err(dev, "intf: Configuration header write failed\n"); + goto err_write_header; + } + } + + INIT_DELAYED_WORK(&ntb->cmd_handler, epf_ntb_cmd_handler); + queue_work(kpcintb_workqueue, &ntb->cmd_handler.work); + + return 0; + +err_write_header: + epf_ntb_mw_bar_clear(ntb); +err_mw_bar_init: + epf_ntb_db_bar_clear(ntb); +err_db_bar_init: +err_config_interrupt: + epf_ntb_config_sspad_bar_clear(ntb); + + return ret; +} + + +/** + * epf_ntb_epc_cleanup() - Cleanup all NTB interfaces + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 + * + * Wrapper to cleanup all NTB interfaces. + */ +static void epf_ntb_epc_cleanup(struct epf_ntb *ntb) +{ + epf_ntb_db_bar_clear(ntb); + epf_ntb_mw_bar_clear(ntb); +} + +#define EPF_NTB_R(_name) \ +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ + char *page) \ +{ \ + struct config_group *group = to_config_group(item); \ + struct epf_ntb *ntb = to_epf_ntb(group); \ + \ + return sprintf(page, "%d\n", ntb->_name); \ +} + +#define EPF_NTB_W(_name) \ +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ + const char *page, size_t len) \ +{ \ + struct config_group *group = to_config_group(item); \ + struct epf_ntb *ntb = to_epf_ntb(group); \ + u32 val; \ + int ret; \ + \ + ret = kstrtou32(page, 0, &val); \ + if (ret) \ + return ret; \ + \ + ntb->_name = val; \ + \ + return len; \ +} + +#define EPF_NTB_MW_R(_name) \ +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ + char *page) \ +{ \ + struct config_group *group = to_config_group(item); \ + struct epf_ntb *ntb = to_epf_ntb(group); \ + int win_no; \ + \ + sscanf(#_name, "mw%d", &win_no); \ + \ + return sprintf(page, "%lld\n", ntb->mws_size[win_no - 1]); \ +} + +#define EPF_NTB_MW_W(_name) \ +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ + const char *page, size_t len) \ +{ \ + struct config_group *group = to_config_group(item); \ + struct epf_ntb *ntb = to_epf_ntb(group); \ + struct device *dev = &ntb->epf->dev; \ + int win_no; \ + u64 val; \ + int ret; \ + \ + ret = kstrtou64(page, 0, &val); \ + if (ret) \ + return ret; \ + \ + if (sscanf(#_name, "mw%d", &win_no) != 1) \ + return -EINVAL; \ + \ + if (ntb->num_mws < win_no) { \ + dev_err(dev, "Invalid num_nws: %d value\n", ntb->num_mws); \ + return -EINVAL; \ + } \ + \ + ntb->mws_size[win_no - 1] = val; \ + \ + return len; \ +} + +static ssize_t epf_ntb_num_mws_store(struct config_item *item, + const char *page, size_t len) +{ + struct config_group *group = to_config_group(item); + struct epf_ntb *ntb = to_epf_ntb(group); + u32 val; + int ret; + + ret = kstrtou32(page, 0, &val); + if (ret) + return ret; + + if (val > MAX_MW) + return -EINVAL; + + ntb->num_mws = val; + + return len; +} + +EPF_NTB_R(spad_count) +EPF_NTB_W(spad_count) +EPF_NTB_R(db_count) +EPF_NTB_W(db_count) +EPF_NTB_R(num_mws) +EPF_NTB_R(vbus_number) +EPF_NTB_W(vbus_number) +EPF_NTB_MW_R(mw1) +EPF_NTB_MW_W(mw1) +EPF_NTB_MW_R(mw2) +EPF_NTB_MW_W(mw2) +EPF_NTB_MW_R(mw3) +EPF_NTB_MW_W(mw3) +EPF_NTB_MW_R(mw4) +EPF_NTB_MW_W(mw4) + +CONFIGFS_ATTR(epf_ntb_, spad_count); +CONFIGFS_ATTR(epf_ntb_, db_count); +CONFIGFS_ATTR(epf_ntb_, num_mws); +CONFIGFS_ATTR(epf_ntb_, mw1); +CONFIGFS_ATTR(epf_ntb_, mw2); +CONFIGFS_ATTR(epf_ntb_, mw3); +CONFIGFS_ATTR(epf_ntb_, mw4); +CONFIGFS_ATTR(epf_ntb_, vbus_number); + +static struct configfs_attribute *epf_ntb_attrs[] = { + &epf_ntb_attr_spad_count, + &epf_ntb_attr_db_count, + &epf_ntb_attr_num_mws, + &epf_ntb_attr_mw1, + &epf_ntb_attr_mw2, + &epf_ntb_attr_mw3, + &epf_ntb_attr_mw4, + &epf_ntb_attr_vbus_number, + NULL, +}; + +static const struct config_item_type ntb_group_type = { + .ct_attrs = epf_ntb_attrs, + .ct_owner = THIS_MODULE, +}; + +/** + * epf_ntb_add_cfs() - Add configfs directory specific to NTB + * @epf: NTB endpoint function device + * @group: A pointer to the config_group structure referencing a group of + * config_items of a specific type that belong to a specific sub-system. + * + * Add configfs directory specific to NTB. This directory will hold + * NTB specific properties like db_count, spad_count, num_mws etc., + */ +static struct config_group *epf_ntb_add_cfs(struct pci_epf *epf, + struct config_group *group) +{ + struct epf_ntb *ntb = epf_get_drvdata(epf); + struct config_group *ntb_group = &ntb->group; + struct device *dev = &epf->dev; + + config_group_init_type_name(ntb_group, dev_name(dev), &ntb_group_type); + + return ntb_group; +} + +/*==== virtual PCI bus driver, which only load virutal ntb pci driver ====*/ + +#define VPCI_BUS_NUM 0x10 + +uint32_t pci_space[] = { + (VNTB_VID | (VNTB_PID << 16)), //DeviceID, Vendor ID + 0, // status, Command + 0xffffffff, // Class code, subclass, prog if, revision id + 0x40, //bist, header type, latency Timer, cache line size + 0, //bar 0 + 0, //bar 1 + 0, //bar 2 + 0, //bar 3 + 0, //bar 4 + 0, //bar 5 + 0, //cardbus cis point + 0, //Subsystem ID Subystem vendor id + 0, //ROM Base Address + 0, //Reserved, Cap. Point + 0, //Reserved, + 0, //Max Lat, Min Gnt, interrupt pin, interrupt line +}; + +int pci_read(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 *val) +{ + if (devfn == 0) { + memcpy(val, ((uint8_t *)pci_space) + where, size); + return 0; + } + return -1; +} + +int pci_write(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 val) +{ + return 0; +} + +struct pci_ops vpci_ops = { + .read = pci_read, + .write = pci_write, +}; + +static int vpci_bus(void *sysdata) +{ + struct pci_bus *vpci_bus; + + vpci_bus = pci_scan_bus(VPCI_BUS_NUM, &vpci_ops, sysdata); + if (vpci_bus) + pr_err("create pci bus\n"); + + pci_bus_add_devices(vpci_bus); + + return 0; +} + +/*==================== Virtual PCIe NTB driver ==========================*/ + +static int vntb_epf_mw_count(struct ntb_dev *ntb, int pidx) +{ + struct epf_ntb *ndev = ntb_ndev(ntb); + + return ndev->num_mws; +} + +static int vntb_epf_spad_count(struct ntb_dev *ntb) +{ + return ntb_ndev(ntb)->spad_count; +} + +static int vntb_epf_peer_mw_count(struct ntb_dev *ntb) +{ + return ntb_ndev(ntb)->num_mws; +} + +static u64 vntb_epf_db_valid_mask(struct ntb_dev *ntb) +{ + return BIT_ULL(ntb_ndev(ntb)->db_count) - 1; +} + +static int vntb_epf_db_set_mask(struct ntb_dev *ntb, u64 db_bits) +{ + return 0; +} + +static int vntb_epf_mw_set_trans(struct ntb_dev *ndev, int pidx, int idx, + dma_addr_t addr, resource_size_t size) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + struct pci_epf_bar *epf_bar; + enum pci_barno barno; + int ret; + struct device *dev; + + dev = &ntb->ntb.dev; + barno = ntb->epf_ntb_bar[BAR_MW0 + idx]; + epf_bar = &ntb->epf->bar[barno]; + epf_bar->phys_addr = addr; + epf_bar->barno = barno; + epf_bar->size = size; + + ret = pci_epc_set_bar(ntb->epf->epc, 0, 0, epf_bar); + if (ret) { + dev_err(dev, "failure set mw trans\n"); + return ret; + } + return 0; +} + +static int vntb_epf_mw_clear_trans(struct ntb_dev *ntb, int pidx, int idx) +{ + return 0; +} + +static int vntb_epf_peer_mw_get_addr(struct ntb_dev *ndev, int idx, + phys_addr_t *base, resource_size_t *size) +{ + + struct epf_ntb *ntb = ntb_ndev(ndev); + + if (base) + *base = ntb->vpci_mw_phy[idx]; + + if (size) + *size = ntb->mws_size[idx]; + + return 0; +} + +static int vntb_epf_link_enable(struct ntb_dev *ntb, + enum ntb_speed max_speed, + enum ntb_width max_width) +{ + return 0; +} + +static u32 vntb_epf_spad_read(struct ntb_dev *ndev, int idx) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + int off = ntb->reg->spad_offset, ct = ntb->reg->spad_count * 4; + u32 val; + void __iomem *base = ntb->reg; + + val = readl(base + off + ct + idx * 4); + return val; +} + +static int vntb_epf_spad_write(struct ntb_dev *ndev, int idx, u32 val) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + struct epf_ntb_ctrl *ctrl = ntb->reg; + int off = ctrl->spad_offset, ct = ctrl->spad_count * 4; + void __iomem *base = ntb->reg; + + writel(val, base + off + ct + idx * 4); + return 0; +} + +static u32 vntb_epf_peer_spad_read(struct ntb_dev *ndev, int pidx, int idx) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + struct epf_ntb_ctrl *ctrl = ntb->reg; + int off = ctrl->spad_offset; + void __iomem *base = ntb->reg; + u32 val; + + val = readl(base + off + idx * 4); + return val; +} + +static int vntb_epf_peer_spad_write(struct ntb_dev *ndev, int pidx, int idx, u32 val) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + struct epf_ntb_ctrl *ctrl = ntb->reg; + int off = ctrl->spad_offset; + void __iomem *base = ntb->reg; + + writel(val, base + off + idx * 4); + return 0; +} + +static int vntb_epf_peer_db_set(struct ntb_dev *ndev, u64 db_bits) +{ + u32 interrupt_num = ffs(db_bits) + 1; + struct epf_ntb *ntb = ntb_ndev(ndev); + u8 func_no, vfunc_no; + int ret; + + func_no = ntb->epf->func_no; + vfunc_no = ntb->epf->vfunc_no; + + ret = pci_epc_raise_irq(ntb->epf->epc, + func_no, + vfunc_no, + PCI_EPC_IRQ_MSI, + interrupt_num + 1); + if (ret) { + dev_err(&ntb->ntb.dev, "intf: Failed to raise IRQ\n"); + return ret; + } + + return 0; +} + +static u64 vntb_epf_db_read(struct ntb_dev *ndev) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + + return ntb->db; +} + +static int vntb_epf_mw_get_align(struct ntb_dev *ndev, int pidx, int idx, + resource_size_t *addr_align, + resource_size_t *size_align, + resource_size_t *size_max) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + + if (addr_align) + *addr_align = SZ_4K; + + if (size_align) + *size_align = 1; + + if (size_max) + *size_max = ntb->mws_size[idx]; + + return 0; +} + +static u64 vntb_epf_link_is_up(struct ntb_dev *ndev, + enum ntb_speed *speed, + enum ntb_width *width) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + + return ntb->reg->link_status; +} + +static int vntb_epf_db_clear_mask(struct ntb_dev *ndev, u64 db_bits) +{ + return 0; +} + +static int vntb_epf_db_clear(struct ntb_dev *ndev, u64 db_bits) +{ + struct epf_ntb *ntb = ntb_ndev(ndev); + + ntb->db &= ~db_bits; + return 0; +} + +static int vntb_epf_link_disable(struct ntb_dev *ntb) +{ + return 0; +} + +static const struct ntb_dev_ops vntb_epf_ops = { + .mw_count = vntb_epf_mw_count, + .spad_count = vntb_epf_spad_count, + .peer_mw_count = vntb_epf_peer_mw_count, + .db_valid_mask = vntb_epf_db_valid_mask, + .db_set_mask = vntb_epf_db_set_mask, + .mw_set_trans = vntb_epf_mw_set_trans, + .mw_clear_trans = vntb_epf_mw_clear_trans, + .peer_mw_get_addr = vntb_epf_peer_mw_get_addr, + .link_enable = vntb_epf_link_enable, + .spad_read = vntb_epf_spad_read, + .spad_write = vntb_epf_spad_write, + .peer_spad_read = vntb_epf_peer_spad_read, + .peer_spad_write = vntb_epf_peer_spad_write, + .peer_db_set = vntb_epf_peer_db_set, + .db_read = vntb_epf_db_read, + .mw_get_align = vntb_epf_mw_get_align, + .link_is_up = vntb_epf_link_is_up, + .db_clear_mask = vntb_epf_db_clear_mask, + .db_clear = vntb_epf_db_clear, + .link_disable = vntb_epf_link_disable, +}; + +static int pci_vntb_probe(struct pci_dev *pdev, const struct pci_device_id *id) +{ + int ret; + struct epf_ntb *ndev = (struct epf_ntb *)pdev->sysdata; + struct device *dev = &pdev->dev; + + ndev->ntb.pdev = pdev; + ndev->ntb.topo = NTB_TOPO_NONE; + ndev->ntb.ops = &vntb_epf_ops; + + ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); + if (ret) { + dev_err(dev, "Cannot set DMA mask\n"); + return -1; + } + + ret = ntb_register_device(&ndev->ntb); + if (ret) { + dev_err(dev, "Failed to register NTB device\n"); + goto err_register_dev; + } + + dev_info(dev, "PCI Virtual NTB driver loaded\n"); + return 0; + +err_register_dev: + return -1; +} + +static const struct pci_device_id pci_vntb_table[] = { + { + PCI_DEVICE(VNTB_VID, VNTB_PID), + }, + {}, +}; + +static struct pci_driver vntb_pci_driver = { + .name = "pci-vntb", + .id_table = pci_vntb_table, + .probe = pci_vntb_probe, +}; + +/* ============ PCIe EPF Driver Bind ====================*/ + +/** + * epf_ntb_bind() - Initialize endpoint controller to provide NTB functionality + * @epf: NTB endpoint function device + * + * Initialize both the endpoint controllers associated with NTB function device. + * Invoked when a primary interface or secondary interface is bound to EPC + * device. This function will succeed only when EPC is bound to both the + * interfaces. + */ +static int epf_ntb_bind(struct pci_epf *epf) +{ + struct epf_ntb *ntb = epf_get_drvdata(epf); + struct device *dev = &epf->dev; + int ret; + + if (!epf->epc) { + dev_dbg(dev, "PRIMARY EPC interface not yet bound\n"); + return 0; + } + + ret = epf_ntb_init_epc_bar(ntb); + if (ret) { + dev_err(dev, "Failed to create NTB EPC\n"); + goto err_bar_init; + } + + ret = epf_ntb_config_spad_bar_alloc(ntb); + if (ret) { + dev_err(dev, "Failed to allocate BAR memory\n"); + goto err_bar_alloc; + } + + ret = epf_ntb_epc_init(ntb); + if (ret) { + dev_err(dev, "Failed to initialize EPC\n"); + goto err_bar_alloc; + } + + epf_set_drvdata(epf, ntb); + + if (pci_register_driver(&vntb_pci_driver)) { + dev_err(dev, "failure register vntb pci driver\n"); + goto err_bar_alloc; + } + + vpci_bus(ntb); + + return 0; + +err_bar_alloc: + epf_ntb_config_spad_bar_free(ntb); + +err_bar_init: + epf_ntb_epc_destroy(ntb); + + return ret; +} + +/** + * epf_ntb_unbind() - Cleanup the initialization from epf_ntb_bind() + * @epf: NTB endpoint function device + * + * Cleanup the initialization from epf_ntb_bind() + */ +static void epf_ntb_unbind(struct pci_epf *epf) +{ + struct epf_ntb *ntb = epf_get_drvdata(epf); + + epf_ntb_epc_cleanup(ntb); + epf_ntb_config_spad_bar_free(ntb); + epf_ntb_epc_destroy(ntb); + + pci_unregister_driver(&vntb_pci_driver); +} + +// EPF driver probe +static struct pci_epf_ops epf_ntb_ops = { + .bind = epf_ntb_bind, + .unbind = epf_ntb_unbind, + .add_cfs = epf_ntb_add_cfs, +}; + +/** + * epf_ntb_probe() - Probe NTB function driver + * @epf: NTB endpoint function device + * + * Probe NTB function driver when endpoint function bus detects a NTB + * endpoint function. + */ +static int epf_ntb_probe(struct pci_epf *epf) +{ + struct epf_ntb *ntb; + struct device *dev; + + dev = &epf->dev; + + ntb = devm_kzalloc(dev, sizeof(*ntb), GFP_KERNEL); + if (!ntb) + return -ENOMEM; + + epf->header = &epf_ntb_header; + ntb->epf = epf; + epf_set_drvdata(epf, ntb); + + dev_info(dev, "pci-ep epf driver loaded\n"); + return 0; +} + +static const struct pci_epf_device_id epf_ntb_ids[] = { + { + .name = "pci_epf_vntb", + }, + {}, +}; + +static struct pci_epf_driver epf_ntb_driver = { + .driver.name = "pci_epf_vntb", + .probe = epf_ntb_probe, + .id_table = epf_ntb_ids, + .ops = &epf_ntb_ops, + .owner = THIS_MODULE, +}; + + +static int __init epf_ntb_init(void) +{ + int ret; + + kpcintb_workqueue = alloc_workqueue("kpcintb", WQ_MEM_RECLAIM | + WQ_HIGHPRI, 0); + ret = pci_epf_register_driver(&epf_ntb_driver); + if (ret) { + destroy_workqueue(kpcintb_workqueue); + pr_err("Failed to register pci epf ntb driver --> %d\n", ret); + return ret; + } + + return 0; +} +module_init(epf_ntb_init); + +static void __exit epf_ntb_exit(void) +{ + pci_epf_unregister_driver(&epf_ntb_driver); + destroy_workqueue(kpcintb_workqueue); +} +module_exit(epf_ntb_exit); + +MODULE_DESCRIPTION("PCI EPF NTB DRIVER"); +MODULE_AUTHOR("Frank Li <Frank.li@nxp.com>"); +MODULE_LICENSE("GPL v2");
Add NTB function driver and virtual PCI BUS and Virtual NTB driver to implement communication between PCIe RC and PCIe EP devices ┌────────────┐ ┌─────────────────────────────────────┐ │ │ │ │ ├────────────┤ │ ┌──────────────┤ │ NTB │ │ │ NTB │ │ NetDev │ │ │ NetDev │ ├────────────┤ │ ├──────────────┤ │ NTB │ │ │ NTB │ │ Transfer │ │ │ Transfer │ ├────────────┤ │ ├──────────────┤ │ │ │ │ │ │ PCI NTB │ │ │ │ │ EPF │ │ │ │ │ Driver │ │ │ PCI Virtual │ │ │ ├───────────────┐ │ NTB Driver │ │ │ │ PCI EP NTB │◄────►│ │ │ │ │ FN Driver │ │ │ ├────────────┤ ├───────────────┤ ├──────────────┤ │ │ │ │ │ │ │ PCI BUS │ ◄─────► │ PCI EP BUS │ │ Virtual PCI │ │ │ PCI │ │ │ BUS │ └────────────┘ └───────────────┴──────┴──────────────┘ PCI RC PCI EP This driver include 3 part: 1 PCI EP NTB function driver 2 Virtual PCI bus 3 PCI virutal NTB driver, which is loaded only by above virtual pci bus Signed-off-by: Frank Li <Frank.Li@nxp.com> --- drivers/pci/endpoint/functions/Kconfig | 11 + drivers/pci/endpoint/functions/Makefile | 1 + drivers/pci/endpoint/functions/pci-epf-vntb.c | 1425 +++++++++++++++++ 3 files changed, 1437 insertions(+) create mode 100644 drivers/pci/endpoint/functions/pci-epf-vntb.c