Message ID | 20230602220318.15323-5-brett.creeley@amd.com (mailing list archive) |
---|---|
State | Not Applicable |
Delegated to: | Netdev Maintainers |
Headers | show |
Series | pds_vfio driver | expand |
Context | Check | Description |
---|---|---|
netdev/tree_selection | success | Not a local patch |
> -----Original Message----- > From: Brett Creeley [mailto:brett.creeley@amd.com] > Sent: 02 June 2023 23:03 > To: kvm@vger.kernel.org; netdev@vger.kernel.org; > alex.williamson@redhat.com; jgg@nvidia.com; yishaih@nvidia.com; > Shameerali Kolothum Thodi <shameerali.kolothum.thodi@huawei.com>; > kevin.tian@intel.com > Cc: brett.creeley@amd.com; shannon.nelson@amd.com > Subject: [PATCH v10 vfio 4/7] vfio/pds: Add VFIO live migration support > > Add live migration support via the VFIO subsystem. The migration > implementation aligns with the definition from uapi/vfio.h and uses > the pds_core PF's adminq for device configuration. > > The ability to suspend, resume, and transfer VF device state data is > included along with the required admin queue command structures and > implementations. > > PDS_LM_CMD_SUSPEND and PDS_LM_CMD_SUSPEND_STATUS are added to > support > the VF device suspend operation. > > PDS_LM_CMD_RESUME is added to support the VF device resume operation. > > PDS_LM_CMD_STATUS is added to determine the exact size of the VF > device state data. > > PDS_LM_CMD_SAVE is added to get the VF device state data. > > PDS_LM_CMD_RESTORE is added to restore the VF device with the > previously saved data from PDS_LM_CMD_SAVE. > > PDS_LM_CMD_HOST_VF_STATUS is added to notify the device when > a migration is in/not-in progress from the host's perspective. > > Signed-off-by: Brett Creeley <brett.creeley@amd.com> > Signed-off-by: Shannon Nelson <shannon.nelson@amd.com> > --- > drivers/vfio/pci/pds/Makefile | 1 + > drivers/vfio/pci/pds/cmds.c | 319 ++++++++++++++++++++++++ > drivers/vfio/pci/pds/cmds.h | 8 +- > drivers/vfio/pci/pds/lm.c | 421 > ++++++++++++++++++++++++++++++++ > drivers/vfio/pci/pds/lm.h | 41 ++++ > drivers/vfio/pci/pds/pci_drv.c | 13 + > drivers/vfio/pci/pds/vfio_dev.c | 120 ++++++++- > drivers/vfio/pci/pds/vfio_dev.h | 11 + > include/linux/pds/pds_adminq.h | 217 ++++++++++++++++ > 9 files changed, 1149 insertions(+), 2 deletions(-) > create mode 100644 drivers/vfio/pci/pds/lm.c > create mode 100644 drivers/vfio/pci/pds/lm.h > > diff --git a/drivers/vfio/pci/pds/Makefile b/drivers/vfio/pci/pds/Makefile > index 87581111fa17..dbaf613d3794 100644 > --- a/drivers/vfio/pci/pds/Makefile > +++ b/drivers/vfio/pci/pds/Makefile > @@ -5,5 +5,6 @@ obj-$(CONFIG_PDS_VFIO_PCI) += pds_vfio.o > > pds_vfio-y := \ > cmds.o \ > + lm.o \ > pci_drv.o \ > vfio_dev.o > diff --git a/drivers/vfio/pci/pds/cmds.c b/drivers/vfio/pci/pds/cmds.c > index ae01f5df2f5c..256f458feb58 100644 > --- a/drivers/vfio/pci/pds/cmds.c > +++ b/drivers/vfio/pci/pds/cmds.c > @@ -3,6 +3,7 @@ > > #include <linux/io.h> > #include <linux/types.h> > +#include <linux/delay.h> > > #include <linux/pds/pds_common.h> > #include <linux/pds/pds_core_if.h> > @@ -11,6 +12,34 @@ > #include "vfio_dev.h" > #include "cmds.h" > > +#define SUSPEND_TIMEOUT_S 5 > +#define SUSPEND_CHECK_INTERVAL_MS 1 > + > +static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device > *pds_vfio, > + union pds_core_adminq_cmd *req, > + size_t req_len, > + union pds_core_adminq_comp *resp, > + u64 flags) Why u64? Do we expect more flags to follow? The core interface below only takes a bool(fast_poll) though. Thanks, Shameer > +{ > + union pds_core_adminq_cmd cmd = {}; > + size_t cp_len; > + int err; > + > + /* Wrap the client request */ > + cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD; > + cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id); > + cp_len = min_t(size_t, req_len, sizeof(cmd.client_request.client_cmd)); > + memcpy(cmd.client_request.client_cmd, req, cp_len); > + > + err = pdsc_adminq_post(pds_vfio->pdsc, &cmd, resp, > + !!(flags & PDS_AQ_FLAG_FASTPOLL)); > + if (err && err != -EAGAIN) > + dev_info(pds_vfio_to_dev(pds_vfio), > + "client admin cmd failed: %pe\n", ERR_PTR(err)); > + > + return err; > +} > + > int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio) > { > struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); > @@ -41,3 +70,293 @@ void pds_vfio_unregister_client_cmd(struct > pds_vfio_pci_device *pds_vfio) > > pds_vfio->client_id = 0; > } > + > +static int > +pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_suspend_status = { > + .opcode = PDS_LM_CMD_SUSPEND_STATUS, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + unsigned long time_limit; > + unsigned long time_start; > + unsigned long time_done; > + int err; > + > + time_start = jiffies; > + time_limit = time_start + HZ * SUSPEND_TIMEOUT_S; > + do { > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), > + &comp, PDS_AQ_FLAG_FASTPOLL); > + if (err != -EAGAIN) > + break; > + > + msleep(SUSPEND_CHECK_INTERVAL_MS); > + } while (time_before(jiffies, time_limit)); > + > + time_done = jiffies; > + dev_dbg(dev, "%s: vf%u: Suspend comp received in %d msecs\n", > __func__, > + pds_vfio->vf_id, jiffies_to_msecs(time_done - time_start)); > + > + /* Check the results */ > + if (time_after_eq(time_done, time_limit)) { > + dev_err(dev, "%s: vf%u: Suspend comp timeout\n", __func__, > + pds_vfio->vf_id); > + err = -ETIMEDOUT; > + } > + > + return err; > +} > + > +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_suspend = { > + .opcode = PDS_LM_CMD_SUSPEND, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + int err; > + > + dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id); > + > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), > &comp, > + PDS_AQ_FLAG_FASTPOLL); > + if (err) { > + dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id, > + ERR_PTR(err)); > + return err; > + } > + > + return pds_vfio_suspend_wait_device_cmd(pds_vfio); > +} > + > +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_resume = { > + .opcode = PDS_LM_CMD_RESUME, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + > + dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id); > + > + return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), > &comp, > + 0); > +} > + > +int pds_vfio_get_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, u64 > *size) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_status = { > + .opcode = PDS_LM_CMD_STATUS, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + int err; > + > + dev_dbg(dev, "vf%u: Get migration status\n", pds_vfio->vf_id); > + > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, > 0); > + if (err) > + return err; > + > + *size = le64_to_cpu(comp.lm_status.size); > + return 0; > +} > + > +static int pds_vfio_dma_map_lm_file(struct device *dev, > + enum dma_data_direction dir, > + struct pds_vfio_lm_file *lm_file) > +{ > + struct pds_lm_sg_elem *sgl, *sge; > + struct scatterlist *sg; > + dma_addr_t sgl_addr; > + size_t sgl_size; > + int err; > + int i; > + > + if (!lm_file) > + return -EINVAL; > + > + /* dma map file pages */ > + err = dma_map_sgtable(dev, &lm_file->sg_table, dir, 0); > + if (err) > + return err; > + > + lm_file->num_sge = lm_file->sg_table.nents; > + > + /* alloc sgl */ > + sgl_size = lm_file->num_sge * sizeof(struct pds_lm_sg_elem); > + sgl = kzalloc(sgl_size, GFP_KERNEL); > + if (!sgl) { > + err = -ENOMEM; > + goto out_unmap_sgtable; > + } > + > + /* fill sgl */ > + sge = sgl; > + for_each_sgtable_dma_sg(&lm_file->sg_table, sg, i) { > + sge->addr = cpu_to_le64(sg_dma_address(sg)); > + sge->len = cpu_to_le32(sg_dma_len(sg)); > + dev_dbg(dev, "addr = %llx, len = %u\n", sge->addr, sge->len); > + sge++; > + } > + > + sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE); > + if (dma_mapping_error(dev, sgl_addr)) { > + err = -EIO; > + goto out_free_sgl; > + } > + > + lm_file->sgl = sgl; > + lm_file->sgl_addr = sgl_addr; > + > + return 0; > + > +out_free_sgl: > + kfree(sgl); > +out_unmap_sgtable: > + lm_file->num_sge = 0; > + dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0); > + return err; > +} > + > +static void pds_vfio_dma_unmap_lm_file(struct device *dev, > + enum dma_data_direction dir, > + struct pds_vfio_lm_file *lm_file) > +{ > + if (!lm_file) > + return; > + > + /* free sgl */ > + if (lm_file->sgl) { > + dma_unmap_single(dev, lm_file->sgl_addr, > + lm_file->num_sge * sizeof(*lm_file->sgl), > + DMA_TO_DEVICE); > + kfree(lm_file->sgl); > + lm_file->sgl = NULL; > + lm_file->sgl_addr = DMA_MAPPING_ERROR; > + lm_file->num_sge = 0; > + } > + > + /* dma unmap file pages */ > + dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0); > +} > + > +int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_save = { > + .opcode = PDS_LM_CMD_SAVE, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); > + struct device *pdsc_dev = &pci_physfn(pdev)->dev; > + union pds_core_adminq_comp comp = {}; > + struct pds_vfio_lm_file *lm_file; > + int err; > + > + dev_dbg(&pdev->dev, "vf%u: Get migration state\n", pds_vfio->vf_id); > + > + lm_file = pds_vfio->save_file; > + > + err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_FROM_DEVICE, > lm_file); > + if (err) { > + dev_err(&pdev->dev, "failed to map save migration file: %pe\n", > + ERR_PTR(err)); > + return err; > + } > + > + cmd.lm_save.sgl_addr = cpu_to_le64(lm_file->sgl_addr); > + cmd.lm_save.num_sge = cpu_to_le32(lm_file->num_sge); > + > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, > 0); > + if (err) > + dev_err(&pdev->dev, "failed to get migration state: %pe\n", > + ERR_PTR(err)); > + > + pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file); > + > + return err; > +} > + > +int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_restore = { > + .opcode = PDS_LM_CMD_RESTORE, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); > + struct device *pdsc_dev = &pci_physfn(pdev)->dev; > + union pds_core_adminq_comp comp = {}; > + struct pds_vfio_lm_file *lm_file; > + int err; > + > + dev_dbg(&pdev->dev, "vf%u: Set migration state\n", pds_vfio->vf_id); > + > + lm_file = pds_vfio->restore_file; > + > + err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file); > + if (err) { > + dev_err(&pdev->dev, > + "failed to map restore migration file: %pe\n", > + ERR_PTR(err)); > + return err; > + } > + > + cmd.lm_restore.sgl_addr = cpu_to_le64(lm_file->sgl_addr); > + cmd.lm_restore.num_sge = cpu_to_le32(lm_file->num_sge); > + > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, > 0); > + if (err) > + dev_err(&pdev->dev, "failed to set migration state: %pe\n", > + ERR_PTR(err)); > + > + pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file); > + > + return err; > +} > + > +void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device > *pds_vfio, > + enum pds_lm_host_vf_status vf_status) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_host_vf_status = { > + .opcode = PDS_LM_CMD_HOST_VF_STATUS, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + .status = vf_status, > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + int err; > + > + dev_dbg(dev, "vf%u: Set host VF LM status: %u", pds_vfio->vf_id, > + vf_status); > + if (vf_status != PDS_LM_STA_IN_PROGRESS && > + vf_status != PDS_LM_STA_NONE) { > + dev_warn(dev, "Invalid host VF migration status, %d\n", > + vf_status); > + return; > + } > + > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, > 0); > + if (err) > + dev_warn(dev, "failed to send host VF migration status: %pe\n", > + ERR_PTR(err)); > +} > diff --git a/drivers/vfio/pci/pds/cmds.h b/drivers/vfio/pci/pds/cmds.h > index 4c592afccf89..3d8a5508c733 100644 > --- a/drivers/vfio/pci/pds/cmds.h > +++ b/drivers/vfio/pci/pds/cmds.h > @@ -6,5 +6,11 @@ > > int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio); > void pds_vfio_unregister_client_cmd(struct pds_vfio_pci_device *pds_vfio); > - > +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio); > +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio); > +int pds_vfio_get_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, u64 > *size); > +int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio); > +int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio); > +void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device > *pds_vfio, > + enum pds_lm_host_vf_status vf_status); > #endif /* _CMDS_H_ */ > diff --git a/drivers/vfio/pci/pds/lm.c b/drivers/vfio/pci/pds/lm.c > new file mode 100644 > index 000000000000..c507f39a2339 > --- /dev/null > +++ b/drivers/vfio/pci/pds/lm.c > @@ -0,0 +1,421 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* Copyright(c) 2023 Advanced Micro Devices, Inc. */ > + > +#include <linux/anon_inodes.h> > +#include <linux/file.h> > +#include <linux/fs.h> > +#include <linux/highmem.h> > +#include <linux/vfio.h> > +#include <linux/vfio_pci_core.h> > + > +#include "vfio_dev.h" > +#include "cmds.h" > + > +static struct pds_vfio_lm_file * > +pds_vfio_get_lm_file(const struct file_operations *fops, int flags, u64 size) > +{ > + struct pds_vfio_lm_file *lm_file = NULL; > + unsigned long long npages; > + struct page **pages; > + void *page_mem; > + const void *p; > + > + if (!size) > + return NULL; > + > + /* Alloc file structure */ > + lm_file = kzalloc(sizeof(*lm_file), GFP_KERNEL); > + if (!lm_file) > + return NULL; > + > + /* Create file */ > + lm_file->filep = > + anon_inode_getfile("pds_vfio_lm", fops, lm_file, flags); > + if (!lm_file->filep) > + goto out_free_file; > + > + stream_open(lm_file->filep->f_inode, lm_file->filep); > + mutex_init(&lm_file->lock); > + > + /* prevent file from being released before we are done with it */ > + get_file(lm_file->filep); > + > + /* Allocate memory for file pages */ > + npages = DIV_ROUND_UP_ULL(size, PAGE_SIZE); > + pages = kmalloc_array(npages, sizeof(*pages), GFP_KERNEL); > + if (!pages) > + goto out_put_file; > + > + page_mem = kvzalloc(ALIGN(size, PAGE_SIZE), GFP_KERNEL); > + if (!page_mem) > + goto out_free_pages_array; > + > + p = page_mem - offset_in_page(page_mem); > + for (unsigned long long i = 0; i < npages; i++) { > + if (is_vmalloc_addr(p)) > + pages[i] = vmalloc_to_page(p); > + else > + pages[i] = kmap_to_page((void *)p); > + if (!pages[i]) > + goto out_free_page_mem; > + > + p += PAGE_SIZE; > + } > + > + /* Create scatterlist of file pages to use for DMA mapping later */ > + if (sg_alloc_table_from_pages(&lm_file->sg_table, pages, npages, 0, > + size, GFP_KERNEL)) > + goto out_free_page_mem; > + > + lm_file->size = size; > + lm_file->pages = pages; > + lm_file->npages = npages; > + lm_file->page_mem = page_mem; > + lm_file->alloc_size = npages * PAGE_SIZE; > + > + return lm_file; > + > +out_free_page_mem: > + kvfree(page_mem); > +out_free_pages_array: > + kfree(pages); > +out_put_file: > + fput(lm_file->filep); > + mutex_destroy(&lm_file->lock); > +out_free_file: > + kfree(lm_file); > + > + return NULL; > +} > + > +static void pds_vfio_put_lm_file(struct pds_vfio_lm_file *lm_file) > +{ > + mutex_lock(&lm_file->lock); > + > + lm_file->size = 0; > + lm_file->alloc_size = 0; > + > + /* Free scatter list of file pages */ > + sg_free_table(&lm_file->sg_table); > + > + kvfree(lm_file->page_mem); > + lm_file->page_mem = NULL; > + kfree(lm_file->pages); > + lm_file->pages = NULL; > + > + mutex_unlock(&lm_file->lock); > + > + /* allow file to be released since we are done with it */ > + fput(lm_file->filep); > +} > + > +void pds_vfio_put_save_file(struct pds_vfio_pci_device *pds_vfio) > +{ > + if (!pds_vfio->save_file) > + return; > + > + pds_vfio_put_lm_file(pds_vfio->save_file); > + pds_vfio->save_file = NULL; > +} > + > +void pds_vfio_put_restore_file(struct pds_vfio_pci_device *pds_vfio) > +{ > + if (!pds_vfio->restore_file) > + return; > + > + pds_vfio_put_lm_file(pds_vfio->restore_file); > + pds_vfio->restore_file = NULL; > +} > + > +static struct page *pds_vfio_get_file_page(struct pds_vfio_lm_file *lm_file, > + unsigned long offset) > +{ > + unsigned long cur_offset = 0; > + struct scatterlist *sg; > + unsigned int i; > + > + /* All accesses are sequential */ > + if (offset < lm_file->last_offset || !lm_file->last_offset_sg) { > + lm_file->last_offset = 0; > + lm_file->last_offset_sg = lm_file->sg_table.sgl; > + lm_file->sg_last_entry = 0; > + } > + > + cur_offset = lm_file->last_offset; > + > + for_each_sg(lm_file->last_offset_sg, sg, > + lm_file->sg_table.orig_nents - lm_file->sg_last_entry, i) { > + if (offset < sg->length + cur_offset) { > + lm_file->last_offset_sg = sg; > + lm_file->sg_last_entry += i; > + lm_file->last_offset = cur_offset; > + return nth_page(sg_page(sg), > + (offset - cur_offset) / PAGE_SIZE); > + } > + cur_offset += sg->length; > + } > + > + return NULL; > +} > + > +static int pds_vfio_release_file(struct inode *inode, struct file *filp) > +{ > + struct pds_vfio_lm_file *lm_file = filp->private_data; > + > + mutex_lock(&lm_file->lock); > + lm_file->filep->f_pos = 0; > + lm_file->size = 0; > + mutex_unlock(&lm_file->lock); > + mutex_destroy(&lm_file->lock); > + kfree(lm_file); > + > + return 0; > +} > + > +static ssize_t pds_vfio_save_read(struct file *filp, char __user *buf, > + size_t len, loff_t *pos) > +{ > + struct pds_vfio_lm_file *lm_file = filp->private_data; > + ssize_t done = 0; > + > + if (pos) > + return -ESPIPE; > + pos = &filp->f_pos; > + > + mutex_lock(&lm_file->lock); > + if (*pos > lm_file->size) { > + done = -EINVAL; > + goto out_unlock; > + } > + > + len = min_t(size_t, lm_file->size - *pos, len); > + while (len) { > + size_t page_offset; > + struct page *page; > + size_t page_len; > + u8 *from_buff; > + int err; > + > + page_offset = (*pos) % PAGE_SIZE; > + page = pds_vfio_get_file_page(lm_file, *pos - page_offset); > + if (!page) { > + if (done == 0) > + done = -EINVAL; > + goto out_unlock; > + } > + > + page_len = min_t(size_t, len, PAGE_SIZE - page_offset); > + from_buff = kmap_local_page(page); > + err = copy_to_user(buf, from_buff + page_offset, page_len); > + kunmap_local(from_buff); > + if (err) { > + done = -EFAULT; > + goto out_unlock; > + } > + *pos += page_len; > + len -= page_len; > + done += page_len; > + buf += page_len; > + } > + > +out_unlock: > + mutex_unlock(&lm_file->lock); > + return done; > +} > + > +static const struct file_operations pds_vfio_save_fops = { > + .owner = THIS_MODULE, > + .read = pds_vfio_save_read, > + .release = pds_vfio_release_file, > + .llseek = no_llseek, > +}; > + > +static int pds_vfio_get_save_file(struct pds_vfio_pci_device *pds_vfio) > +{ > + struct device *dev = &pds_vfio->vfio_coredev.pdev->dev; > + struct pds_vfio_lm_file *lm_file; > + int err; > + u64 size; > + > + /* Get live migration state size in this state */ > + err = pds_vfio_get_lm_status_cmd(pds_vfio, &size); > + if (err) { > + dev_err(dev, "failed to get save status: %pe\n", ERR_PTR(err)); > + return err; > + } > + > + dev_dbg(dev, "save status, size = %lld\n", size); > + > + if (!size) { > + dev_err(dev, "invalid state size\n"); > + return -EIO; > + } > + > + lm_file = pds_vfio_get_lm_file(&pds_vfio_save_fops, O_RDONLY, size); > + if (!lm_file) { > + dev_err(dev, "failed to create save file\n"); > + return -ENOENT; > + } > + > + dev_dbg(dev, "size = %lld, alloc_size = %lld, npages = %lld\n", > + lm_file->size, lm_file->alloc_size, lm_file->npages); > + > + pds_vfio->save_file = lm_file; > + > + return 0; > +} > + > +static ssize_t pds_vfio_restore_write(struct file *filp, const char __user > *buf, > + size_t len, loff_t *pos) > +{ > + struct pds_vfio_lm_file *lm_file = filp->private_data; > + loff_t requested_length; > + ssize_t done = 0; > + > + if (pos) > + return -ESPIPE; > + > + pos = &filp->f_pos; > + > + if (*pos < 0 || > + check_add_overflow((loff_t)len, *pos, &requested_length)) > + return -EINVAL; > + > + mutex_lock(&lm_file->lock); > + > + while (len) { > + size_t page_offset; > + struct page *page; > + size_t page_len; > + u8 *to_buff; > + int err; > + > + page_offset = (*pos) % PAGE_SIZE; > + page = pds_vfio_get_file_page(lm_file, *pos - page_offset); > + if (!page) { > + if (done == 0) > + done = -EINVAL; > + goto out_unlock; > + } > + > + page_len = min_t(size_t, len, PAGE_SIZE - page_offset); > + to_buff = kmap_local_page(page); > + err = copy_from_user(to_buff + page_offset, buf, page_len); > + kunmap_local(to_buff); > + if (err) { > + done = -EFAULT; > + goto out_unlock; > + } > + *pos += page_len; > + len -= page_len; > + done += page_len; > + buf += page_len; > + lm_file->size += page_len; > + } > +out_unlock: > + mutex_unlock(&lm_file->lock); > + return done; > +} > + > +static const struct file_operations pds_vfio_restore_fops = { > + .owner = THIS_MODULE, > + .write = pds_vfio_restore_write, > + .release = pds_vfio_release_file, > + .llseek = no_llseek, > +}; > + > +static int pds_vfio_get_restore_file(struct pds_vfio_pci_device *pds_vfio) > +{ > + struct device *dev = &pds_vfio->vfio_coredev.pdev->dev; > + struct pds_vfio_lm_file *lm_file; > + u64 size; > + > + size = sizeof(union pds_lm_dev_state); > + dev_dbg(dev, "restore status, size = %lld\n", size); > + > + if (!size) { > + dev_err(dev, "invalid state size"); > + return -EIO; > + } > + > + lm_file = pds_vfio_get_lm_file(&pds_vfio_restore_fops, O_WRONLY, > size); > + if (!lm_file) { > + dev_err(dev, "failed to create restore file"); > + return -ENOENT; > + } > + pds_vfio->restore_file = lm_file; > + > + return 0; > +} > + > +struct file * > +pds_vfio_step_device_state_locked(struct pds_vfio_pci_device *pds_vfio, > + enum vfio_device_mig_state next) > +{ > + enum vfio_device_mig_state cur = pds_vfio->state; > + int err; > + > + if (cur == VFIO_DEVICE_STATE_STOP && next == > VFIO_DEVICE_STATE_STOP_COPY) { > + err = pds_vfio_get_save_file(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + err = pds_vfio_get_lm_state_cmd(pds_vfio); > + if (err) { > + pds_vfio_put_save_file(pds_vfio); > + return ERR_PTR(err); > + } > + > + return pds_vfio->save_file->filep; > + } > + > + if (cur == VFIO_DEVICE_STATE_STOP_COPY && next == > VFIO_DEVICE_STATE_STOP) { > + pds_vfio_put_save_file(pds_vfio); > + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, > PDS_LM_STA_NONE); > + return NULL; > + } > + > + if (cur == VFIO_DEVICE_STATE_STOP && next == > VFIO_DEVICE_STATE_RESUMING) { > + err = pds_vfio_get_restore_file(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + return pds_vfio->restore_file->filep; > + } > + > + if (cur == VFIO_DEVICE_STATE_RESUMING && next == > VFIO_DEVICE_STATE_STOP) { > + err = pds_vfio_set_lm_state_cmd(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + pds_vfio_put_restore_file(pds_vfio); > + return NULL; > + } > + > + if (cur == VFIO_DEVICE_STATE_RUNNING && next == > VFIO_DEVICE_STATE_RUNNING_P2P) { > + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, > + PDS_LM_STA_IN_PROGRESS); > + err = pds_vfio_suspend_device_cmd(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + return NULL; > + } > + > + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == > VFIO_DEVICE_STATE_RUNNING) { > + err = pds_vfio_resume_device_cmd(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, > PDS_LM_STA_NONE); > + return NULL; > + } > + > + if (cur == VFIO_DEVICE_STATE_STOP && next == > VFIO_DEVICE_STATE_RUNNING_P2P) > + return NULL; > + > + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == > VFIO_DEVICE_STATE_STOP) > + return NULL; > + > + return ERR_PTR(-EINVAL); > +} > diff --git a/drivers/vfio/pci/pds/lm.h b/drivers/vfio/pci/pds/lm.h > new file mode 100644 > index 000000000000..13be893198b7 > --- /dev/null > +++ b/drivers/vfio/pci/pds/lm.h > @@ -0,0 +1,41 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* Copyright(c) 2023 Advanced Micro Devices, Inc. */ > + > +#ifndef _LM_H_ > +#define _LM_H_ > + > +#include <linux/fs.h> > +#include <linux/mutex.h> > +#include <linux/scatterlist.h> > +#include <linux/types.h> > + > +#include <linux/pds/pds_common.h> > +#include <linux/pds/pds_adminq.h> > + > +struct pds_vfio_lm_file { > + struct file *filep; > + struct mutex lock; /* protect live migration data file */ > + u64 size; /* Size with valid data */ > + u64 alloc_size; /* Total allocated size. Always >= len */ > + void *page_mem; /* memory allocated for pages */ > + struct page **pages; /* Backing pages for file */ > + unsigned long long npages; > + struct sg_table sg_table; /* SG table for backing pages */ > + struct pds_lm_sg_elem *sgl; /* DMA mapping */ > + dma_addr_t sgl_addr; > + u16 num_sge; > + struct scatterlist *last_offset_sg; /* Iterator */ > + unsigned int sg_last_entry; > + unsigned long last_offset; > +}; > + > +struct pds_vfio_pci_device; > + > +struct file * > +pds_vfio_step_device_state_locked(struct pds_vfio_pci_device *pds_vfio, > + enum vfio_device_mig_state next); > + > +void pds_vfio_put_save_file(struct pds_vfio_pci_device *pds_vfio); > +void pds_vfio_put_restore_file(struct pds_vfio_pci_device *pds_vfio); > + > +#endif /* _LM_H_ */ > diff --git a/drivers/vfio/pci/pds/pci_drv.c b/drivers/vfio/pci/pds/pci_drv.c > index a49420aa9736..ffd47fa8ede3 100644 > --- a/drivers/vfio/pci/pds/pci_drv.c > +++ b/drivers/vfio/pci/pds/pci_drv.c > @@ -73,11 +73,24 @@ pds_vfio_pci_table[] = { > }; > MODULE_DEVICE_TABLE(pci, pds_vfio_pci_table); > > +static void pds_vfio_pci_aer_reset_done(struct pci_dev *pdev) > +{ > + struct pds_vfio_pci_device *pds_vfio = pds_vfio_pci_drvdata(pdev); > + > + pds_vfio_reset(pds_vfio); > +} > + > +static const struct pci_error_handlers pds_vfio_pci_err_handlers = { > + .reset_done = pds_vfio_pci_aer_reset_done, > + .error_detected = vfio_pci_core_aer_err_detected, > +}; > + > static struct pci_driver pds_vfio_pci_driver = { > .name = KBUILD_MODNAME, > .id_table = pds_vfio_pci_table, > .probe = pds_vfio_pci_probe, > .remove = pds_vfio_pci_remove, > + .err_handler = &pds_vfio_pci_err_handlers, > .driver_managed_dma = true, > }; > > diff --git a/drivers/vfio/pci/pds/vfio_dev.c b/drivers/vfio/pci/pds/vfio_dev.c > index 39771265b78f..2435d8255366 100644 > --- a/drivers/vfio/pci/pds/vfio_dev.c > +++ b/drivers/vfio/pci/pds/vfio_dev.c > @@ -4,6 +4,7 @@ > #include <linux/vfio.h> > #include <linux/vfio_pci_core.h> > > +#include "lm.h" > #include "vfio_dev.h" > > struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio) > @@ -11,6 +12,11 @@ struct pci_dev *pds_vfio_to_pci_dev(struct > pds_vfio_pci_device *pds_vfio) > return pds_vfio->vfio_coredev.pdev; > } > > +struct device *pds_vfio_to_dev(struct pds_vfio_pci_device *pds_vfio) > +{ > + return &pds_vfio_to_pci_dev(pds_vfio)->dev; > +} > + > struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev) > { > struct vfio_pci_core_device *core_device = > dev_get_drvdata(&pdev->dev); > @@ -19,6 +25,98 @@ struct pds_vfio_pci_device > *pds_vfio_pci_drvdata(struct pci_dev *pdev) > vfio_coredev); > } > > +static void pds_vfio_state_mutex_unlock(struct pds_vfio_pci_device > *pds_vfio) > +{ > +again: > + spin_lock(&pds_vfio->reset_lock); > + if (pds_vfio->deferred_reset) { > + pds_vfio->deferred_reset = false; > + if (pds_vfio->state == VFIO_DEVICE_STATE_ERROR) { > + pds_vfio->state = VFIO_DEVICE_STATE_RUNNING; > + pds_vfio_put_restore_file(pds_vfio); > + pds_vfio_put_save_file(pds_vfio); > + } > + spin_unlock(&pds_vfio->reset_lock); > + goto again; > + } > + mutex_unlock(&pds_vfio->state_mutex); > + spin_unlock(&pds_vfio->reset_lock); > +} > + > +void pds_vfio_reset(struct pds_vfio_pci_device *pds_vfio) > +{ > + spin_lock(&pds_vfio->reset_lock); > + pds_vfio->deferred_reset = true; > + if (!mutex_trylock(&pds_vfio->state_mutex)) { > + spin_unlock(&pds_vfio->reset_lock); > + return; > + } > + spin_unlock(&pds_vfio->reset_lock); > + pds_vfio_state_mutex_unlock(pds_vfio); > +} > + > +static struct file * > +pds_vfio_set_device_state(struct vfio_device *vdev, > + enum vfio_device_mig_state new_state) > +{ > + struct pds_vfio_pci_device *pds_vfio = > + container_of(vdev, struct pds_vfio_pci_device, > + vfio_coredev.vdev); > + struct file *res = NULL; > + > + mutex_lock(&pds_vfio->state_mutex); > + while (new_state != pds_vfio->state) { > + enum vfio_device_mig_state next_state; > + > + int err = vfio_mig_get_next_state(vdev, pds_vfio->state, > + new_state, &next_state); > + if (err) { > + res = ERR_PTR(err); > + break; > + } > + > + res = pds_vfio_step_device_state_locked(pds_vfio, next_state); > + if (IS_ERR(res)) > + break; > + > + pds_vfio->state = next_state; > + > + if (WARN_ON(res && new_state != pds_vfio->state)) { > + res = ERR_PTR(-EINVAL); > + break; > + } > + } > + pds_vfio_state_mutex_unlock(pds_vfio); > + > + return res; > +} > + > +static int pds_vfio_get_device_state(struct vfio_device *vdev, > + enum vfio_device_mig_state *current_state) > +{ > + struct pds_vfio_pci_device *pds_vfio = > + container_of(vdev, struct pds_vfio_pci_device, > + vfio_coredev.vdev); > + > + mutex_lock(&pds_vfio->state_mutex); > + *current_state = pds_vfio->state; > + pds_vfio_state_mutex_unlock(pds_vfio); > + return 0; > +} > + > +static int pds_vfio_get_device_state_size(struct vfio_device *vdev, > + unsigned long *stop_copy_length) > +{ > + *stop_copy_length = PDS_LM_DEVICE_STATE_LENGTH; > + return 0; > +} > + > +static const struct vfio_migration_ops pds_vfio_lm_ops = { > + .migration_set_state = pds_vfio_set_device_state, > + .migration_get_state = pds_vfio_get_device_state, > + .migration_get_data_size = pds_vfio_get_device_state_size > +}; > + > static int pds_vfio_init_device(struct vfio_device *vdev) > { > struct pds_vfio_pci_device *pds_vfio = > @@ -34,6 +132,9 @@ static int pds_vfio_init_device(struct vfio_device > *vdev) > pds_vfio->vf_id = pci_iov_vf_id(pdev); > pds_vfio->pci_id = PCI_DEVID(pdev->bus->number, pdev->devfn); > > + vdev->migration_flags = VFIO_MIGRATION_STOP_COPY | > VFIO_MIGRATION_P2P; > + vdev->mig_ops = &pds_vfio_lm_ops; > + > dev_dbg(&pdev->dev, > "%s: PF %#04x VF %#04x (%d) vf_id %d domain %d > pds_vfio %p\n", > __func__, pci_dev_id(pdev->physfn), pds_vfio->pci_id, > @@ -54,17 +155,34 @@ static int pds_vfio_open_device(struct vfio_device > *vdev) > if (err) > return err; > > + mutex_init(&pds_vfio->state_mutex); > + pds_vfio->state = VFIO_DEVICE_STATE_RUNNING; > + > vfio_pci_core_finish_enable(&pds_vfio->vfio_coredev); > > return 0; > } > > +static void pds_vfio_close_device(struct vfio_device *vdev) > +{ > + struct pds_vfio_pci_device *pds_vfio = > + container_of(vdev, struct pds_vfio_pci_device, > + vfio_coredev.vdev); > + > + mutex_lock(&pds_vfio->state_mutex); > + pds_vfio_put_restore_file(pds_vfio); > + pds_vfio_put_save_file(pds_vfio); > + mutex_unlock(&pds_vfio->state_mutex); > + mutex_destroy(&pds_vfio->state_mutex); > + vfio_pci_core_close_device(vdev); > +} > + > static const struct vfio_device_ops pds_vfio_ops = { > .name = "pds-vfio", > .init = pds_vfio_init_device, > .release = vfio_pci_core_release_dev, > .open_device = pds_vfio_open_device, > - .close_device = vfio_pci_core_close_device, > + .close_device = pds_vfio_close_device, > .ioctl = vfio_pci_core_ioctl, > .device_feature = vfio_pci_core_ioctl_feature, > .read = vfio_pci_core_read, > diff --git a/drivers/vfio/pci/pds/vfio_dev.h b/drivers/vfio/pci/pds/vfio_dev.h > index 92e8ff241ca8..df6208a7140b 100644 > --- a/drivers/vfio/pci/pds/vfio_dev.h > +++ b/drivers/vfio/pci/pds/vfio_dev.h > @@ -7,12 +7,21 @@ > #include <linux/pci.h> > #include <linux/vfio_pci_core.h> > > +#include "lm.h" > + > struct pdsc; > > struct pds_vfio_pci_device { > struct vfio_pci_core_device vfio_coredev; > struct pdsc *pdsc; > > + struct pds_vfio_lm_file *save_file; > + struct pds_vfio_lm_file *restore_file; > + struct mutex state_mutex; /* protect migration state */ > + enum vfio_device_mig_state state; > + spinlock_t reset_lock; /* protect reset_done flow */ > + u8 deferred_reset; > + > int vf_id; > int pci_id; > u16 client_id; > @@ -20,7 +29,9 @@ struct pds_vfio_pci_device { > > const struct vfio_device_ops *pds_vfio_ops_info(void); > struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev); > +void pds_vfio_reset(struct pds_vfio_pci_device *pds_vfio); > > struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio); > +struct device *pds_vfio_to_dev(struct pds_vfio_pci_device *pds_vfio); > > #endif /* _VFIO_DEV_H_ */ > diff --git a/include/linux/pds/pds_adminq.h > b/include/linux/pds/pds_adminq.h > index 98a60ce87b92..db6de081f15f 100644 > --- a/include/linux/pds/pds_adminq.h > +++ b/include/linux/pds/pds_adminq.h > @@ -584,6 +584,213 @@ struct pds_core_q_init_comp { > u8 color; > }; > > +#define PDS_LM_DEVICE_STATE_LENGTH 65536 > +#define PDS_LM_CHECK_DEVICE_STATE_LENGTH(X) \ > + PDS_CORE_SIZE_CHECK(union, > PDS_LM_DEVICE_STATE_LENGTH, X) > + > +/* > + * enum pds_lm_cmd_opcode - Live Migration Device commands > + */ > +enum pds_lm_cmd_opcode { > + PDS_LM_CMD_HOST_VF_STATUS = 1, > + > + /* Device state commands */ > + PDS_LM_CMD_STATUS = 16, > + PDS_LM_CMD_SUSPEND = 18, > + PDS_LM_CMD_SUSPEND_STATUS = 19, > + PDS_LM_CMD_RESUME = 20, > + PDS_LM_CMD_SAVE = 21, > + PDS_LM_CMD_RESTORE = 22, > +}; > + > +/** > + * struct pds_lm_cmd - generic command > + * @opcode: Opcode > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + * @rsvd2: Structure padding to 60 Bytes > + */ > +struct pds_lm_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > + u8 rsvd2[56]; > +}; > + > +/** > + * struct pds_lm_comp - generic command completion > + * @status: Status of the command (enum pds_core_status_code) > + * @rsvd: Structure padding to 16 Bytes > + */ > +struct pds_lm_comp { > + u8 status; > + u8 rsvd[15]; > +}; > + > +/** > + * struct pds_lm_status_cmd - STATUS command > + * @opcode: Opcode > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + */ > +struct pds_lm_status_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > +}; > + > +/** > + * struct pds_lm_status_comp - STATUS command completion > + * @status: Status of the command (enum pds_core_status_code) > + * @rsvd: Word boundary padding > + * @comp_index: Index in the desc ring for which this is the > completion > + * @size: Size of the device state > + * @rsvd2: Word boundary padding > + * @color: Color bit > + */ > +struct pds_lm_status_comp { > + u8 status; > + u8 rsvd; > + __le16 comp_index; > + union { > + __le64 size; > + u8 rsvd2[11]; > + } __packed; > + u8 color; > +}; > + > +/** > + * struct pds_lm_suspend_cmd - SUSPEND command > + * @opcode: Opcode PDS_LM_CMD_SUSPEND > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + */ > +struct pds_lm_suspend_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > +}; > + > +/** > + * struct pds_lm_suspend_comp - SUSPEND command completion > + * @status: Status of the command (enum pds_core_status_code) > + * @rsvd: Word boundary padding > + * @comp_index: Index in the desc ring for which this is the > completion > + * @state_size: Size of the device state computed post suspend > + * @rsvd2: Word boundary padding > + * @color: Color bit > + */ > +struct pds_lm_suspend_comp { > + u8 status; > + u8 rsvd; > + __le16 comp_index; > + union { > + __le64 state_size; > + u8 rsvd2[11]; > + } __packed; > + u8 color; > +}; > + > +/** > + * struct pds_lm_suspend_status_cmd - SUSPEND status command > + * @opcode: Opcode PDS_AQ_CMD_LM_SUSPEND_STATUS > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + */ > +struct pds_lm_suspend_status_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > +}; > + > +/** > + * struct pds_lm_resume_cmd - RESUME command > + * @opcode: Opcode PDS_LM_CMD_RESUME > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + */ > +struct pds_lm_resume_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > +}; > + > +/** > + * struct pds_lm_sg_elem - Transmit scatter-gather (SG) descriptor element > + * @addr: DMA address of SG element data buffer > + * @len: Length of SG element data buffer, in bytes > + * @rsvd: Word boundary padding > + */ > +struct pds_lm_sg_elem { > + __le64 addr; > + __le32 len; > + __le16 rsvd[2]; > +}; > + > +/** > + * struct pds_lm_save_cmd - SAVE command > + * @opcode: Opcode PDS_LM_CMD_SAVE > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + * @rsvd2: Word boundary padding > + * @sgl_addr: IOVA address of the SGL to dma the device state > + * @num_sge: Total number of SG elements > + */ > +struct pds_lm_save_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > + u8 rsvd2[4]; > + __le64 sgl_addr; > + __le32 num_sge; > +} __packed; > + > +/** > + * struct pds_lm_restore_cmd - RESTORE command > + * @opcode: Opcode PDS_LM_CMD_RESTORE > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + * @rsvd2: Word boundary padding > + * @sgl_addr: IOVA address of the SGL to dma the device state > + * @num_sge: Total number of SG elements > + */ > +struct pds_lm_restore_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > + u8 rsvd2[4]; > + __le64 sgl_addr; > + __le32 num_sge; > +} __packed; > + > +/** > + * union pds_lm_dev_state - device state information > + * @words: Device state words > + */ > +union pds_lm_dev_state { > + __le32 words[PDS_LM_DEVICE_STATE_LENGTH / sizeof(__le32)]; > +}; > + > +enum pds_lm_host_vf_status { > + PDS_LM_STA_NONE = 0, > + PDS_LM_STA_IN_PROGRESS, > + PDS_LM_STA_MAX, > +}; > + > +/** > + * struct pds_lm_host_vf_status_cmd - HOST_VF_STATUS command > + * @opcode: Opcode PDS_LM_CMD_HOST_VF_STATUS > + * @rsvd: Word boundary padding > + * @vf_id: VF id > + * @status: Current LM status of host VF driver (enum > pds_lm_host_status) > + */ > +struct pds_lm_host_vf_status_cmd { > + u8 opcode; > + u8 rsvd; > + __le16 vf_id; > + u8 status; > +}; > + > union pds_core_adminq_cmd { > u8 opcode; > u8 bytes[64]; > @@ -600,6 +807,14 @@ union pds_core_adminq_cmd { > > struct pds_core_q_identify_cmd q_ident; > struct pds_core_q_init_cmd q_init; > + > + struct pds_lm_suspend_cmd lm_suspend; > + struct pds_lm_suspend_status_cmd lm_suspend_status; > + struct pds_lm_resume_cmd lm_resume; > + struct pds_lm_status_cmd lm_status; > + struct pds_lm_save_cmd lm_save; > + struct pds_lm_restore_cmd lm_restore; > + struct pds_lm_host_vf_status_cmd lm_host_vf_status; > }; > > union pds_core_adminq_comp { > @@ -621,6 +836,8 @@ union pds_core_adminq_comp { > > struct pds_core_q_identify_comp q_ident; > struct pds_core_q_init_comp q_init; > + > + struct pds_lm_status_comp lm_status; > }; > > #ifndef __CHECKER__ > -- > 2.17.1
On 6/15/2023 2:07 PM, Shameerali Kolothum Thodi wrote: > Caution: This message originated from an External Source. Use proper caution when opening attachments, clicking links, or responding. > > >> -----Original Message----- >> From: Brett Creeley [mailto:brett.creeley@amd.com] >> Sent: 02 June 2023 23:03 >> To: kvm@vger.kernel.org; netdev@vger.kernel.org; >> alex.williamson@redhat.com; jgg@nvidia.com; yishaih@nvidia.com; >> Shameerali Kolothum Thodi <shameerali.kolothum.thodi@huawei.com>; >> kevin.tian@intel.com >> Cc: brett.creeley@amd.com; shannon.nelson@amd.com >> Subject: [PATCH v10 vfio 4/7] vfio/pds: Add VFIO live migration support >> >> Add live migration support via the VFIO subsystem. The migration >> implementation aligns with the definition from uapi/vfio.h and uses >> the pds_core PF's adminq for device configuration. >> >> The ability to suspend, resume, and transfer VF device state data is >> included along with the required admin queue command structures and >> implementations. >> >> PDS_LM_CMD_SUSPEND and PDS_LM_CMD_SUSPEND_STATUS are added to >> support >> the VF device suspend operation. >> >> PDS_LM_CMD_RESUME is added to support the VF device resume operation. >> >> PDS_LM_CMD_STATUS is added to determine the exact size of the VF >> device state data. >> >> PDS_LM_CMD_SAVE is added to get the VF device state data. >> >> PDS_LM_CMD_RESTORE is added to restore the VF device with the >> previously saved data from PDS_LM_CMD_SAVE. >> >> PDS_LM_CMD_HOST_VF_STATUS is added to notify the device when >> a migration is in/not-in progress from the host's perspective. >> >> Signed-off-by: Brett Creeley <brett.creeley@amd.com> >> Signed-off-by: Shannon Nelson <shannon.nelson@amd.com> >> --- >> drivers/vfio/pci/pds/Makefile | 1 + >> drivers/vfio/pci/pds/cmds.c | 319 ++++++++++++++++++++++++ >> drivers/vfio/pci/pds/cmds.h | 8 +- >> drivers/vfio/pci/pds/lm.c | 421 >> ++++++++++++++++++++++++++++++++ >> drivers/vfio/pci/pds/lm.h | 41 ++++ >> drivers/vfio/pci/pds/pci_drv.c | 13 + >> drivers/vfio/pci/pds/vfio_dev.c | 120 ++++++++- >> drivers/vfio/pci/pds/vfio_dev.h | 11 + >> include/linux/pds/pds_adminq.h | 217 ++++++++++++++++ >> 9 files changed, 1149 insertions(+), 2 deletions(-) >> create mode 100644 drivers/vfio/pci/pds/lm.c >> create mode 100644 drivers/vfio/pci/pds/lm.h >> >> diff --git a/drivers/vfio/pci/pds/Makefile b/drivers/vfio/pci/pds/Makefile >> index 87581111fa17..dbaf613d3794 100644 >> --- a/drivers/vfio/pci/pds/Makefile >> +++ b/drivers/vfio/pci/pds/Makefile >> @@ -5,5 +5,6 @@ obj-$(CONFIG_PDS_VFIO_PCI) += pds_vfio.o >> >> pds_vfio-y := \ >> cmds.o \ >> + lm.o \ >> pci_drv.o \ >> vfio_dev.o >> diff --git a/drivers/vfio/pci/pds/cmds.c b/drivers/vfio/pci/pds/cmds.c >> index ae01f5df2f5c..256f458feb58 100644 >> --- a/drivers/vfio/pci/pds/cmds.c >> +++ b/drivers/vfio/pci/pds/cmds.c >> @@ -3,6 +3,7 @@ >> >> #include <linux/io.h> >> #include <linux/types.h> >> +#include <linux/delay.h> >> >> #include <linux/pds/pds_common.h> >> #include <linux/pds/pds_core_if.h> >> @@ -11,6 +12,34 @@ >> #include "vfio_dev.h" >> #include "cmds.h" >> >> +#define SUSPEND_TIMEOUT_S 5 >> +#define SUSPEND_CHECK_INTERVAL_MS 1 >> + >> +static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device >> *pds_vfio, >> + union pds_core_adminq_cmd *req, >> + size_t req_len, >> + union pds_core_adminq_comp *resp, >> + u64 flags) > > Why u64? Do we expect more flags to follow? The core interface below > only takes a bool(fast_poll) though. > > Thanks, > Shameer > Shameer, Another good catch. This was leftover from the original set of patches, but using flags is definitely unnecessary in pds_vfio_client_adminq_cmd(). If we ever need more flags I can update then. I will change this to a bool in the next revision. Thanks for the review, Brett >> +{ >> + union pds_core_adminq_cmd cmd = {}; >> + size_t cp_len; >> + int err; >> + >> + /* Wrap the client request */ >> + cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD; >> + cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id); >> + cp_len = min_t(size_t, req_len, sizeof(cmd.client_request.client_cmd)); >> + memcpy(cmd.client_request.client_cmd, req, cp_len); >> + >> + err = pdsc_adminq_post(pds_vfio->pdsc, &cmd, resp, >> + !!(flags & PDS_AQ_FLAG_FASTPOLL)); >> + if (err && err != -EAGAIN) >> + dev_info(pds_vfio_to_dev(pds_vfio), >> + "client admin cmd failed: %pe\n", ERR_PTR(err)); >> + >> + return err; >> +} >> + >> int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio) >> { >> struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); >> @@ -41,3 +70,293 @@ void pds_vfio_unregister_client_cmd(struct >> pds_vfio_pci_device *pds_vfio) >> >> pds_vfio->client_id = 0; >> } >> + >> +static int >> +pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_suspend_status = { >> + .opcode = PDS_LM_CMD_SUSPEND_STATUS, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + unsigned long time_limit; >> + unsigned long time_start; >> + unsigned long time_done; >> + int err; >> + >> + time_start = jiffies; >> + time_limit = time_start + HZ * SUSPEND_TIMEOUT_S; >> + do { >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), >> + &comp, PDS_AQ_FLAG_FASTPOLL); >> + if (err != -EAGAIN) >> + break; >> + >> + msleep(SUSPEND_CHECK_INTERVAL_MS); >> + } while (time_before(jiffies, time_limit)); >> + >> + time_done = jiffies; >> + dev_dbg(dev, "%s: vf%u: Suspend comp received in %d msecs\n", >> __func__, >> + pds_vfio->vf_id, jiffies_to_msecs(time_done - time_start)); >> + >> + /* Check the results */ >> + if (time_after_eq(time_done, time_limit)) { >> + dev_err(dev, "%s: vf%u: Suspend comp timeout\n", __func__, >> + pds_vfio->vf_id); >> + err = -ETIMEDOUT; >> + } >> + >> + return err; >> +} >> + >> +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_suspend = { >> + .opcode = PDS_LM_CMD_SUSPEND, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + int err; >> + >> + dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id); >> + >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), >> &comp, >> + PDS_AQ_FLAG_FASTPOLL); >> + if (err) { >> + dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id, >> + ERR_PTR(err)); >> + return err; >> + } >> + >> + return pds_vfio_suspend_wait_device_cmd(pds_vfio); >> +} >> + >> +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_resume = { >> + .opcode = PDS_LM_CMD_RESUME, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + >> + dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id); >> + >> + return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), >> &comp, >> + 0); >> +} >> + >> +int pds_vfio_get_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, u64 >> *size) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_status = { >> + .opcode = PDS_LM_CMD_STATUS, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + int err; >> + >> + dev_dbg(dev, "vf%u: Get migration status\n", pds_vfio->vf_id); >> + >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, >> 0); >> + if (err) >> + return err; >> + >> + *size = le64_to_cpu(comp.lm_status.size); >> + return 0; >> +} >> + >> +static int pds_vfio_dma_map_lm_file(struct device *dev, >> + enum dma_data_direction dir, >> + struct pds_vfio_lm_file *lm_file) >> +{ >> + struct pds_lm_sg_elem *sgl, *sge; >> + struct scatterlist *sg; >> + dma_addr_t sgl_addr; >> + size_t sgl_size; >> + int err; >> + int i; >> + >> + if (!lm_file) >> + return -EINVAL; >> + >> + /* dma map file pages */ >> + err = dma_map_sgtable(dev, &lm_file->sg_table, dir, 0); >> + if (err) >> + return err; >> + >> + lm_file->num_sge = lm_file->sg_table.nents; >> + >> + /* alloc sgl */ >> + sgl_size = lm_file->num_sge * sizeof(struct pds_lm_sg_elem); >> + sgl = kzalloc(sgl_size, GFP_KERNEL); >> + if (!sgl) { >> + err = -ENOMEM; >> + goto out_unmap_sgtable; >> + } >> + >> + /* fill sgl */ >> + sge = sgl; >> + for_each_sgtable_dma_sg(&lm_file->sg_table, sg, i) { >> + sge->addr = cpu_to_le64(sg_dma_address(sg)); >> + sge->len = cpu_to_le32(sg_dma_len(sg)); >> + dev_dbg(dev, "addr = %llx, len = %u\n", sge->addr, sge->len); >> + sge++; >> + } >> + >> + sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE); >> + if (dma_mapping_error(dev, sgl_addr)) { >> + err = -EIO; >> + goto out_free_sgl; >> + } >> + >> + lm_file->sgl = sgl; >> + lm_file->sgl_addr = sgl_addr; >> + >> + return 0; >> + >> +out_free_sgl: >> + kfree(sgl); >> +out_unmap_sgtable: >> + lm_file->num_sge = 0; >> + dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0); >> + return err; >> +} >> + >> +static void pds_vfio_dma_unmap_lm_file(struct device *dev, >> + enum dma_data_direction dir, >> + struct pds_vfio_lm_file *lm_file) >> +{ >> + if (!lm_file) >> + return; >> + >> + /* free sgl */ >> + if (lm_file->sgl) { >> + dma_unmap_single(dev, lm_file->sgl_addr, >> + lm_file->num_sge * sizeof(*lm_file->sgl), >> + DMA_TO_DEVICE); >> + kfree(lm_file->sgl); >> + lm_file->sgl = NULL; >> + lm_file->sgl_addr = DMA_MAPPING_ERROR; >> + lm_file->num_sge = 0; >> + } >> + >> + /* dma unmap file pages */ >> + dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0); >> +} >> + >> +int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_save = { >> + .opcode = PDS_LM_CMD_SAVE, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); >> + struct device *pdsc_dev = &pci_physfn(pdev)->dev; >> + union pds_core_adminq_comp comp = {}; >> + struct pds_vfio_lm_file *lm_file; >> + int err; >> + >> + dev_dbg(&pdev->dev, "vf%u: Get migration state\n", pds_vfio->vf_id); >> + >> + lm_file = pds_vfio->save_file; >> + >> + err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_FROM_DEVICE, >> lm_file); >> + if (err) { >> + dev_err(&pdev->dev, "failed to map save migration file: %pe\n", >> + ERR_PTR(err)); >> + return err; >> + } >> + >> + cmd.lm_save.sgl_addr = cpu_to_le64(lm_file->sgl_addr); >> + cmd.lm_save.num_sge = cpu_to_le32(lm_file->num_sge); >> + >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, >> 0); >> + if (err) >> + dev_err(&pdev->dev, "failed to get migration state: %pe\n", >> + ERR_PTR(err)); >> + >> + pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file); >> + >> + return err; >> +} >> + >> +int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_restore = { >> + .opcode = PDS_LM_CMD_RESTORE, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); >> + struct device *pdsc_dev = &pci_physfn(pdev)->dev; >> + union pds_core_adminq_comp comp = {}; >> + struct pds_vfio_lm_file *lm_file; >> + int err; >> + >> + dev_dbg(&pdev->dev, "vf%u: Set migration state\n", pds_vfio->vf_id); >> + >> + lm_file = pds_vfio->restore_file; >> + >> + err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file); >> + if (err) { >> + dev_err(&pdev->dev, >> + "failed to map restore migration file: %pe\n", >> + ERR_PTR(err)); >> + return err; >> + } >> + >> + cmd.lm_restore.sgl_addr = cpu_to_le64(lm_file->sgl_addr); >> + cmd.lm_restore.num_sge = cpu_to_le32(lm_file->num_sge); >> + >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, >> 0); >> + if (err) >> + dev_err(&pdev->dev, "failed to set migration state: %pe\n", >> + ERR_PTR(err)); >> + >> + pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file); >> + >> + return err; >> +} >> + >> +void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device >> *pds_vfio, >> + enum pds_lm_host_vf_status vf_status) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_host_vf_status = { >> + .opcode = PDS_LM_CMD_HOST_VF_STATUS, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + .status = vf_status, >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + int err; >> + >> + dev_dbg(dev, "vf%u: Set host VF LM status: %u", pds_vfio->vf_id, >> + vf_status); >> + if (vf_status != PDS_LM_STA_IN_PROGRESS && >> + vf_status != PDS_LM_STA_NONE) { >> + dev_warn(dev, "Invalid host VF migration status, %d\n", >> + vf_status); >> + return; >> + } >> + >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, >> 0); >> + if (err) >> + dev_warn(dev, "failed to send host VF migration status: %pe\n", >> + ERR_PTR(err)); >> +} >> diff --git a/drivers/vfio/pci/pds/cmds.h b/drivers/vfio/pci/pds/cmds.h >> index 4c592afccf89..3d8a5508c733 100644 >> --- a/drivers/vfio/pci/pds/cmds.h >> +++ b/drivers/vfio/pci/pds/cmds.h >> @@ -6,5 +6,11 @@ >> >> int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio); >> void pds_vfio_unregister_client_cmd(struct pds_vfio_pci_device *pds_vfio); >> - >> +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio); >> +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio); >> +int pds_vfio_get_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, u64 >> *size); >> +int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio); >> +int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio); >> +void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device >> *pds_vfio, >> + enum pds_lm_host_vf_status vf_status); >> #endif /* _CMDS_H_ */ >> diff --git a/drivers/vfio/pci/pds/lm.c b/drivers/vfio/pci/pds/lm.c >> new file mode 100644 >> index 000000000000..c507f39a2339 >> --- /dev/null >> +++ b/drivers/vfio/pci/pds/lm.c >> @@ -0,0 +1,421 @@ >> +// SPDX-License-Identifier: GPL-2.0 >> +/* Copyright(c) 2023 Advanced Micro Devices, Inc. */ >> + >> +#include <linux/anon_inodes.h> >> +#include <linux/file.h> >> +#include <linux/fs.h> >> +#include <linux/highmem.h> >> +#include <linux/vfio.h> >> +#include <linux/vfio_pci_core.h> >> + >> +#include "vfio_dev.h" >> +#include "cmds.h" >> + >> +static struct pds_vfio_lm_file * >> +pds_vfio_get_lm_file(const struct file_operations *fops, int flags, u64 size) >> +{ >> + struct pds_vfio_lm_file *lm_file = NULL; >> + unsigned long long npages; >> + struct page **pages; >> + void *page_mem; >> + const void *p; >> + >> + if (!size) >> + return NULL; >> + >> + /* Alloc file structure */ >> + lm_file = kzalloc(sizeof(*lm_file), GFP_KERNEL); >> + if (!lm_file) >> + return NULL; >> + >> + /* Create file */ >> + lm_file->filep = >> + anon_inode_getfile("pds_vfio_lm", fops, lm_file, flags); >> + if (!lm_file->filep) >> + goto out_free_file; >> + >> + stream_open(lm_file->filep->f_inode, lm_file->filep); >> + mutex_init(&lm_file->lock); >> + >> + /* prevent file from being released before we are done with it */ >> + get_file(lm_file->filep); >> + >> + /* Allocate memory for file pages */ >> + npages = DIV_ROUND_UP_ULL(size, PAGE_SIZE); >> + pages = kmalloc_array(npages, sizeof(*pages), GFP_KERNEL); >> + if (!pages) >> + goto out_put_file; >> + >> + page_mem = kvzalloc(ALIGN(size, PAGE_SIZE), GFP_KERNEL); >> + if (!page_mem) >> + goto out_free_pages_array; >> + >> + p = page_mem - offset_in_page(page_mem); >> + for (unsigned long long i = 0; i < npages; i++) { >> + if (is_vmalloc_addr(p)) >> + pages[i] = vmalloc_to_page(p); >> + else >> + pages[i] = kmap_to_page((void *)p); >> + if (!pages[i]) >> + goto out_free_page_mem; >> + >> + p += PAGE_SIZE; >> + } >> + >> + /* Create scatterlist of file pages to use for DMA mapping later */ >> + if (sg_alloc_table_from_pages(&lm_file->sg_table, pages, npages, 0, >> + size, GFP_KERNEL)) >> + goto out_free_page_mem; >> + >> + lm_file->size = size; >> + lm_file->pages = pages; >> + lm_file->npages = npages; >> + lm_file->page_mem = page_mem; >> + lm_file->alloc_size = npages * PAGE_SIZE; >> + >> + return lm_file; >> + >> +out_free_page_mem: >> + kvfree(page_mem); >> +out_free_pages_array: >> + kfree(pages); >> +out_put_file: >> + fput(lm_file->filep); >> + mutex_destroy(&lm_file->lock); >> +out_free_file: >> + kfree(lm_file); >> + >> + return NULL; >> +} >> + >> +static void pds_vfio_put_lm_file(struct pds_vfio_lm_file *lm_file) >> +{ >> + mutex_lock(&lm_file->lock); >> + >> + lm_file->size = 0; >> + lm_file->alloc_size = 0; >> + >> + /* Free scatter list of file pages */ >> + sg_free_table(&lm_file->sg_table); >> + >> + kvfree(lm_file->page_mem); >> + lm_file->page_mem = NULL; >> + kfree(lm_file->pages); >> + lm_file->pages = NULL; >> + >> + mutex_unlock(&lm_file->lock); >> + >> + /* allow file to be released since we are done with it */ >> + fput(lm_file->filep); >> +} >> + >> +void pds_vfio_put_save_file(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + if (!pds_vfio->save_file) >> + return; >> + >> + pds_vfio_put_lm_file(pds_vfio->save_file); >> + pds_vfio->save_file = NULL; >> +} >> + >> +void pds_vfio_put_restore_file(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + if (!pds_vfio->restore_file) >> + return; >> + >> + pds_vfio_put_lm_file(pds_vfio->restore_file); >> + pds_vfio->restore_file = NULL; >> +} >> + >> +static struct page *pds_vfio_get_file_page(struct pds_vfio_lm_file *lm_file, >> + unsigned long offset) >> +{ >> + unsigned long cur_offset = 0; >> + struct scatterlist *sg; >> + unsigned int i; >> + >> + /* All accesses are sequential */ >> + if (offset < lm_file->last_offset || !lm_file->last_offset_sg) { >> + lm_file->last_offset = 0; >> + lm_file->last_offset_sg = lm_file->sg_table.sgl; >> + lm_file->sg_last_entry = 0; >> + } >> + >> + cur_offset = lm_file->last_offset; >> + >> + for_each_sg(lm_file->last_offset_sg, sg, >> + lm_file->sg_table.orig_nents - lm_file->sg_last_entry, i) { >> + if (offset < sg->length + cur_offset) { >> + lm_file->last_offset_sg = sg; >> + lm_file->sg_last_entry += i; >> + lm_file->last_offset = cur_offset; >> + return nth_page(sg_page(sg), >> + (offset - cur_offset) / PAGE_SIZE); >> + } >> + cur_offset += sg->length; >> + } >> + >> + return NULL; >> +} >> + >> +static int pds_vfio_release_file(struct inode *inode, struct file *filp) >> +{ >> + struct pds_vfio_lm_file *lm_file = filp->private_data; >> + >> + mutex_lock(&lm_file->lock); >> + lm_file->filep->f_pos = 0; >> + lm_file->size = 0; >> + mutex_unlock(&lm_file->lock); >> + mutex_destroy(&lm_file->lock); >> + kfree(lm_file); >> + >> + return 0; >> +} >> + >> +static ssize_t pds_vfio_save_read(struct file *filp, char __user *buf, >> + size_t len, loff_t *pos) >> +{ >> + struct pds_vfio_lm_file *lm_file = filp->private_data; >> + ssize_t done = 0; >> + >> + if (pos) >> + return -ESPIPE; >> + pos = &filp->f_pos; >> + >> + mutex_lock(&lm_file->lock); >> + if (*pos > lm_file->size) { >> + done = -EINVAL; >> + goto out_unlock; >> + } >> + >> + len = min_t(size_t, lm_file->size - *pos, len); >> + while (len) { >> + size_t page_offset; >> + struct page *page; >> + size_t page_len; >> + u8 *from_buff; >> + int err; >> + >> + page_offset = (*pos) % PAGE_SIZE; >> + page = pds_vfio_get_file_page(lm_file, *pos - page_offset); >> + if (!page) { >> + if (done == 0) >> + done = -EINVAL; >> + goto out_unlock; >> + } >> + >> + page_len = min_t(size_t, len, PAGE_SIZE - page_offset); >> + from_buff = kmap_local_page(page); >> + err = copy_to_user(buf, from_buff + page_offset, page_len); >> + kunmap_local(from_buff); >> + if (err) { >> + done = -EFAULT; >> + goto out_unlock; >> + } >> + *pos += page_len; >> + len -= page_len; >> + done += page_len; >> + buf += page_len; >> + } >> + >> +out_unlock: >> + mutex_unlock(&lm_file->lock); >> + return done; >> +} >> + >> +static const struct file_operations pds_vfio_save_fops = { >> + .owner = THIS_MODULE, >> + .read = pds_vfio_save_read, >> + .release = pds_vfio_release_file, >> + .llseek = no_llseek, >> +}; >> + >> +static int pds_vfio_get_save_file(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + struct device *dev = &pds_vfio->vfio_coredev.pdev->dev; >> + struct pds_vfio_lm_file *lm_file; >> + int err; >> + u64 size; >> + >> + /* Get live migration state size in this state */ >> + err = pds_vfio_get_lm_status_cmd(pds_vfio, &size); >> + if (err) { >> + dev_err(dev, "failed to get save status: %pe\n", ERR_PTR(err)); >> + return err; >> + } >> + >> + dev_dbg(dev, "save status, size = %lld\n", size); >> + >> + if (!size) { >> + dev_err(dev, "invalid state size\n"); >> + return -EIO; >> + } >> + >> + lm_file = pds_vfio_get_lm_file(&pds_vfio_save_fops, O_RDONLY, size); >> + if (!lm_file) { >> + dev_err(dev, "failed to create save file\n"); >> + return -ENOENT; >> + } >> + >> + dev_dbg(dev, "size = %lld, alloc_size = %lld, npages = %lld\n", >> + lm_file->size, lm_file->alloc_size, lm_file->npages); >> + >> + pds_vfio->save_file = lm_file; >> + >> + return 0; >> +} >> + >> +static ssize_t pds_vfio_restore_write(struct file *filp, const char __user >> *buf, >> + size_t len, loff_t *pos) >> +{ >> + struct pds_vfio_lm_file *lm_file = filp->private_data; >> + loff_t requested_length; >> + ssize_t done = 0; >> + >> + if (pos) >> + return -ESPIPE; >> + >> + pos = &filp->f_pos; >> + >> + if (*pos < 0 || >> + check_add_overflow((loff_t)len, *pos, &requested_length)) >> + return -EINVAL; >> + >> + mutex_lock(&lm_file->lock); >> + >> + while (len) { >> + size_t page_offset; >> + struct page *page; >> + size_t page_len; >> + u8 *to_buff; >> + int err; >> + >> + page_offset = (*pos) % PAGE_SIZE; >> + page = pds_vfio_get_file_page(lm_file, *pos - page_offset); >> + if (!page) { >> + if (done == 0) >> + done = -EINVAL; >> + goto out_unlock; >> + } >> + >> + page_len = min_t(size_t, len, PAGE_SIZE - page_offset); >> + to_buff = kmap_local_page(page); >> + err = copy_from_user(to_buff + page_offset, buf, page_len); >> + kunmap_local(to_buff); >> + if (err) { >> + done = -EFAULT; >> + goto out_unlock; >> + } >> + *pos += page_len; >> + len -= page_len; >> + done += page_len; >> + buf += page_len; >> + lm_file->size += page_len; >> + } >> +out_unlock: >> + mutex_unlock(&lm_file->lock); >> + return done; >> +} >> + >> +static const struct file_operations pds_vfio_restore_fops = { >> + .owner = THIS_MODULE, >> + .write = pds_vfio_restore_write, >> + .release = pds_vfio_release_file, >> + .llseek = no_llseek, >> +}; >> + >> +static int pds_vfio_get_restore_file(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + struct device *dev = &pds_vfio->vfio_coredev.pdev->dev; >> + struct pds_vfio_lm_file *lm_file; >> + u64 size; >> + >> + size = sizeof(union pds_lm_dev_state); >> + dev_dbg(dev, "restore status, size = %lld\n", size); >> + >> + if (!size) { >> + dev_err(dev, "invalid state size"); >> + return -EIO; >> + } >> + >> + lm_file = pds_vfio_get_lm_file(&pds_vfio_restore_fops, O_WRONLY, >> size); >> + if (!lm_file) { >> + dev_err(dev, "failed to create restore file"); >> + return -ENOENT; >> + } >> + pds_vfio->restore_file = lm_file; >> + >> + return 0; >> +} >> + >> +struct file * >> +pds_vfio_step_device_state_locked(struct pds_vfio_pci_device *pds_vfio, >> + enum vfio_device_mig_state next) >> +{ >> + enum vfio_device_mig_state cur = pds_vfio->state; >> + int err; >> + >> + if (cur == VFIO_DEVICE_STATE_STOP && next == >> VFIO_DEVICE_STATE_STOP_COPY) { >> + err = pds_vfio_get_save_file(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + err = pds_vfio_get_lm_state_cmd(pds_vfio); >> + if (err) { >> + pds_vfio_put_save_file(pds_vfio); >> + return ERR_PTR(err); >> + } >> + >> + return pds_vfio->save_file->filep; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_STOP_COPY && next == >> VFIO_DEVICE_STATE_STOP) { >> + pds_vfio_put_save_file(pds_vfio); >> + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, >> PDS_LM_STA_NONE); >> + return NULL; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_STOP && next == >> VFIO_DEVICE_STATE_RESUMING) { >> + err = pds_vfio_get_restore_file(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + return pds_vfio->restore_file->filep; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_RESUMING && next == >> VFIO_DEVICE_STATE_STOP) { >> + err = pds_vfio_set_lm_state_cmd(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + pds_vfio_put_restore_file(pds_vfio); >> + return NULL; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_RUNNING && next == >> VFIO_DEVICE_STATE_RUNNING_P2P) { >> + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, >> + PDS_LM_STA_IN_PROGRESS); >> + err = pds_vfio_suspend_device_cmd(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + return NULL; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == >> VFIO_DEVICE_STATE_RUNNING) { >> + err = pds_vfio_resume_device_cmd(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, >> PDS_LM_STA_NONE); >> + return NULL; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_STOP && next == >> VFIO_DEVICE_STATE_RUNNING_P2P) >> + return NULL; >> + >> + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == >> VFIO_DEVICE_STATE_STOP) >> + return NULL; >> + >> + return ERR_PTR(-EINVAL); >> +} >> diff --git a/drivers/vfio/pci/pds/lm.h b/drivers/vfio/pci/pds/lm.h >> new file mode 100644 >> index 000000000000..13be893198b7 >> --- /dev/null >> +++ b/drivers/vfio/pci/pds/lm.h >> @@ -0,0 +1,41 @@ >> +/* SPDX-License-Identifier: GPL-2.0 */ >> +/* Copyright(c) 2023 Advanced Micro Devices, Inc. */ >> + >> +#ifndef _LM_H_ >> +#define _LM_H_ >> + >> +#include <linux/fs.h> >> +#include <linux/mutex.h> >> +#include <linux/scatterlist.h> >> +#include <linux/types.h> >> + >> +#include <linux/pds/pds_common.h> >> +#include <linux/pds/pds_adminq.h> >> + >> +struct pds_vfio_lm_file { >> + struct file *filep; >> + struct mutex lock; /* protect live migration data file */ >> + u64 size; /* Size with valid data */ >> + u64 alloc_size; /* Total allocated size. Always >= len */ >> + void *page_mem; /* memory allocated for pages */ >> + struct page **pages; /* Backing pages for file */ >> + unsigned long long npages; >> + struct sg_table sg_table; /* SG table for backing pages */ >> + struct pds_lm_sg_elem *sgl; /* DMA mapping */ >> + dma_addr_t sgl_addr; >> + u16 num_sge; >> + struct scatterlist *last_offset_sg; /* Iterator */ >> + unsigned int sg_last_entry; >> + unsigned long last_offset; >> +}; >> + >> +struct pds_vfio_pci_device; >> + >> +struct file * >> +pds_vfio_step_device_state_locked(struct pds_vfio_pci_device *pds_vfio, >> + enum vfio_device_mig_state next); >> + >> +void pds_vfio_put_save_file(struct pds_vfio_pci_device *pds_vfio); >> +void pds_vfio_put_restore_file(struct pds_vfio_pci_device *pds_vfio); >> + >> +#endif /* _LM_H_ */ >> diff --git a/drivers/vfio/pci/pds/pci_drv.c b/drivers/vfio/pci/pds/pci_drv.c >> index a49420aa9736..ffd47fa8ede3 100644 >> --- a/drivers/vfio/pci/pds/pci_drv.c >> +++ b/drivers/vfio/pci/pds/pci_drv.c >> @@ -73,11 +73,24 @@ pds_vfio_pci_table[] = { >> }; >> MODULE_DEVICE_TABLE(pci, pds_vfio_pci_table); >> >> +static void pds_vfio_pci_aer_reset_done(struct pci_dev *pdev) >> +{ >> + struct pds_vfio_pci_device *pds_vfio = pds_vfio_pci_drvdata(pdev); >> + >> + pds_vfio_reset(pds_vfio); >> +} >> + >> +static const struct pci_error_handlers pds_vfio_pci_err_handlers = { >> + .reset_done = pds_vfio_pci_aer_reset_done, >> + .error_detected = vfio_pci_core_aer_err_detected, >> +}; >> + >> static struct pci_driver pds_vfio_pci_driver = { >> .name = KBUILD_MODNAME, >> .id_table = pds_vfio_pci_table, >> .probe = pds_vfio_pci_probe, >> .remove = pds_vfio_pci_remove, >> + .err_handler = &pds_vfio_pci_err_handlers, >> .driver_managed_dma = true, >> }; >> >> diff --git a/drivers/vfio/pci/pds/vfio_dev.c b/drivers/vfio/pci/pds/vfio_dev.c >> index 39771265b78f..2435d8255366 100644 >> --- a/drivers/vfio/pci/pds/vfio_dev.c >> +++ b/drivers/vfio/pci/pds/vfio_dev.c >> @@ -4,6 +4,7 @@ >> #include <linux/vfio.h> >> #include <linux/vfio_pci_core.h> >> >> +#include "lm.h" >> #include "vfio_dev.h" >> >> struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio) >> @@ -11,6 +12,11 @@ struct pci_dev *pds_vfio_to_pci_dev(struct >> pds_vfio_pci_device *pds_vfio) >> return pds_vfio->vfio_coredev.pdev; >> } >> >> +struct device *pds_vfio_to_dev(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + return &pds_vfio_to_pci_dev(pds_vfio)->dev; >> +} >> + >> struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev) >> { >> struct vfio_pci_core_device *core_device = >> dev_get_drvdata(&pdev->dev); >> @@ -19,6 +25,98 @@ struct pds_vfio_pci_device >> *pds_vfio_pci_drvdata(struct pci_dev *pdev) >> vfio_coredev); >> } >> >> +static void pds_vfio_state_mutex_unlock(struct pds_vfio_pci_device >> *pds_vfio) >> +{ >> +again: >> + spin_lock(&pds_vfio->reset_lock); >> + if (pds_vfio->deferred_reset) { >> + pds_vfio->deferred_reset = false; >> + if (pds_vfio->state == VFIO_DEVICE_STATE_ERROR) { >> + pds_vfio->state = VFIO_DEVICE_STATE_RUNNING; >> + pds_vfio_put_restore_file(pds_vfio); >> + pds_vfio_put_save_file(pds_vfio); >> + } >> + spin_unlock(&pds_vfio->reset_lock); >> + goto again; >> + } >> + mutex_unlock(&pds_vfio->state_mutex); >> + spin_unlock(&pds_vfio->reset_lock); >> +} >> + >> +void pds_vfio_reset(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + spin_lock(&pds_vfio->reset_lock); >> + pds_vfio->deferred_reset = true; >> + if (!mutex_trylock(&pds_vfio->state_mutex)) { >> + spin_unlock(&pds_vfio->reset_lock); >> + return; >> + } >> + spin_unlock(&pds_vfio->reset_lock); >> + pds_vfio_state_mutex_unlock(pds_vfio); >> +} >> + >> +static struct file * >> +pds_vfio_set_device_state(struct vfio_device *vdev, >> + enum vfio_device_mig_state new_state) >> +{ >> + struct pds_vfio_pci_device *pds_vfio = >> + container_of(vdev, struct pds_vfio_pci_device, >> + vfio_coredev.vdev); >> + struct file *res = NULL; >> + >> + mutex_lock(&pds_vfio->state_mutex); >> + while (new_state != pds_vfio->state) { >> + enum vfio_device_mig_state next_state; >> + >> + int err = vfio_mig_get_next_state(vdev, pds_vfio->state, >> + new_state, &next_state); >> + if (err) { >> + res = ERR_PTR(err); >> + break; >> + } >> + >> + res = pds_vfio_step_device_state_locked(pds_vfio, next_state); >> + if (IS_ERR(res)) >> + break; >> + >> + pds_vfio->state = next_state; >> + >> + if (WARN_ON(res && new_state != pds_vfio->state)) { >> + res = ERR_PTR(-EINVAL); >> + break; >> + } >> + } >> + pds_vfio_state_mutex_unlock(pds_vfio); >> + >> + return res; >> +} >> + >> +static int pds_vfio_get_device_state(struct vfio_device *vdev, >> + enum vfio_device_mig_state *current_state) >> +{ >> + struct pds_vfio_pci_device *pds_vfio = >> + container_of(vdev, struct pds_vfio_pci_device, >> + vfio_coredev.vdev); >> + >> + mutex_lock(&pds_vfio->state_mutex); >> + *current_state = pds_vfio->state; >> + pds_vfio_state_mutex_unlock(pds_vfio); >> + return 0; >> +} >> + >> +static int pds_vfio_get_device_state_size(struct vfio_device *vdev, >> + unsigned long *stop_copy_length) >> +{ >> + *stop_copy_length = PDS_LM_DEVICE_STATE_LENGTH; >> + return 0; >> +} >> + >> +static const struct vfio_migration_ops pds_vfio_lm_ops = { >> + .migration_set_state = pds_vfio_set_device_state, >> + .migration_get_state = pds_vfio_get_device_state, >> + .migration_get_data_size = pds_vfio_get_device_state_size >> +}; >> + >> static int pds_vfio_init_device(struct vfio_device *vdev) >> { >> struct pds_vfio_pci_device *pds_vfio = >> @@ -34,6 +132,9 @@ static int pds_vfio_init_device(struct vfio_device >> *vdev) >> pds_vfio->vf_id = pci_iov_vf_id(pdev); >> pds_vfio->pci_id = PCI_DEVID(pdev->bus->number, pdev->devfn); >> >> + vdev->migration_flags = VFIO_MIGRATION_STOP_COPY | >> VFIO_MIGRATION_P2P; >> + vdev->mig_ops = &pds_vfio_lm_ops; >> + >> dev_dbg(&pdev->dev, >> "%s: PF %#04x VF %#04x (%d) vf_id %d domain %d >> pds_vfio %p\n", >> __func__, pci_dev_id(pdev->physfn), pds_vfio->pci_id, >> @@ -54,17 +155,34 @@ static int pds_vfio_open_device(struct vfio_device >> *vdev) >> if (err) >> return err; >> >> + mutex_init(&pds_vfio->state_mutex); >> + pds_vfio->state = VFIO_DEVICE_STATE_RUNNING; >> + >> vfio_pci_core_finish_enable(&pds_vfio->vfio_coredev); >> >> return 0; >> } >> >> +static void pds_vfio_close_device(struct vfio_device *vdev) >> +{ >> + struct pds_vfio_pci_device *pds_vfio = >> + container_of(vdev, struct pds_vfio_pci_device, >> + vfio_coredev.vdev); >> + >> + mutex_lock(&pds_vfio->state_mutex); >> + pds_vfio_put_restore_file(pds_vfio); >> + pds_vfio_put_save_file(pds_vfio); >> + mutex_unlock(&pds_vfio->state_mutex); >> + mutex_destroy(&pds_vfio->state_mutex); >> + vfio_pci_core_close_device(vdev); >> +} >> + >> static const struct vfio_device_ops pds_vfio_ops = { >> .name = "pds-vfio", >> .init = pds_vfio_init_device, >> .release = vfio_pci_core_release_dev, >> .open_device = pds_vfio_open_device, >> - .close_device = vfio_pci_core_close_device, >> + .close_device = pds_vfio_close_device, >> .ioctl = vfio_pci_core_ioctl, >> .device_feature = vfio_pci_core_ioctl_feature, >> .read = vfio_pci_core_read, >> diff --git a/drivers/vfio/pci/pds/vfio_dev.h b/drivers/vfio/pci/pds/vfio_dev.h >> index 92e8ff241ca8..df6208a7140b 100644 >> --- a/drivers/vfio/pci/pds/vfio_dev.h >> +++ b/drivers/vfio/pci/pds/vfio_dev.h >> @@ -7,12 +7,21 @@ >> #include <linux/pci.h> >> #include <linux/vfio_pci_core.h> >> >> +#include "lm.h" >> + >> struct pdsc; >> >> struct pds_vfio_pci_device { >> struct vfio_pci_core_device vfio_coredev; >> struct pdsc *pdsc; >> >> + struct pds_vfio_lm_file *save_file; >> + struct pds_vfio_lm_file *restore_file; >> + struct mutex state_mutex; /* protect migration state */ >> + enum vfio_device_mig_state state; >> + spinlock_t reset_lock; /* protect reset_done flow */ >> + u8 deferred_reset; >> + >> int vf_id; >> int pci_id; >> u16 client_id; >> @@ -20,7 +29,9 @@ struct pds_vfio_pci_device { >> >> const struct vfio_device_ops *pds_vfio_ops_info(void); >> struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev); >> +void pds_vfio_reset(struct pds_vfio_pci_device *pds_vfio); >> >> struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio); >> +struct device *pds_vfio_to_dev(struct pds_vfio_pci_device *pds_vfio); >> >> #endif /* _VFIO_DEV_H_ */ >> diff --git a/include/linux/pds/pds_adminq.h >> b/include/linux/pds/pds_adminq.h >> index 98a60ce87b92..db6de081f15f 100644 >> --- a/include/linux/pds/pds_adminq.h >> +++ b/include/linux/pds/pds_adminq.h >> @@ -584,6 +584,213 @@ struct pds_core_q_init_comp { >> u8 color; >> }; >> >> +#define PDS_LM_DEVICE_STATE_LENGTH 65536 >> +#define PDS_LM_CHECK_DEVICE_STATE_LENGTH(X) \ >> + PDS_CORE_SIZE_CHECK(union, >> PDS_LM_DEVICE_STATE_LENGTH, X) >> + >> +/* >> + * enum pds_lm_cmd_opcode - Live Migration Device commands >> + */ >> +enum pds_lm_cmd_opcode { >> + PDS_LM_CMD_HOST_VF_STATUS = 1, >> + >> + /* Device state commands */ >> + PDS_LM_CMD_STATUS = 16, >> + PDS_LM_CMD_SUSPEND = 18, >> + PDS_LM_CMD_SUSPEND_STATUS = 19, >> + PDS_LM_CMD_RESUME = 20, >> + PDS_LM_CMD_SAVE = 21, >> + PDS_LM_CMD_RESTORE = 22, >> +}; >> + >> +/** >> + * struct pds_lm_cmd - generic command >> + * @opcode: Opcode >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + * @rsvd2: Structure padding to 60 Bytes >> + */ >> +struct pds_lm_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> + u8 rsvd2[56]; >> +}; >> + >> +/** >> + * struct pds_lm_comp - generic command completion >> + * @status: Status of the command (enum pds_core_status_code) >> + * @rsvd: Structure padding to 16 Bytes >> + */ >> +struct pds_lm_comp { >> + u8 status; >> + u8 rsvd[15]; >> +}; >> + >> +/** >> + * struct pds_lm_status_cmd - STATUS command >> + * @opcode: Opcode >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + */ >> +struct pds_lm_status_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> +}; >> + >> +/** >> + * struct pds_lm_status_comp - STATUS command completion >> + * @status: Status of the command (enum pds_core_status_code) >> + * @rsvd: Word boundary padding >> + * @comp_index: Index in the desc ring for which this is the >> completion >> + * @size: Size of the device state >> + * @rsvd2: Word boundary padding >> + * @color: Color bit >> + */ >> +struct pds_lm_status_comp { >> + u8 status; >> + u8 rsvd; >> + __le16 comp_index; >> + union { >> + __le64 size; >> + u8 rsvd2[11]; >> + } __packed; >> + u8 color; >> +}; >> + >> +/** >> + * struct pds_lm_suspend_cmd - SUSPEND command >> + * @opcode: Opcode PDS_LM_CMD_SUSPEND >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + */ >> +struct pds_lm_suspend_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> +}; >> + >> +/** >> + * struct pds_lm_suspend_comp - SUSPEND command completion >> + * @status: Status of the command (enum pds_core_status_code) >> + * @rsvd: Word boundary padding >> + * @comp_index: Index in the desc ring for which this is the >> completion >> + * @state_size: Size of the device state computed post suspend >> + * @rsvd2: Word boundary padding >> + * @color: Color bit >> + */ >> +struct pds_lm_suspend_comp { >> + u8 status; >> + u8 rsvd; >> + __le16 comp_index; >> + union { >> + __le64 state_size; >> + u8 rsvd2[11]; >> + } __packed; >> + u8 color; >> +}; >> + >> +/** >> + * struct pds_lm_suspend_status_cmd - SUSPEND status command >> + * @opcode: Opcode PDS_AQ_CMD_LM_SUSPEND_STATUS >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + */ >> +struct pds_lm_suspend_status_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> +}; >> + >> +/** >> + * struct pds_lm_resume_cmd - RESUME command >> + * @opcode: Opcode PDS_LM_CMD_RESUME >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + */ >> +struct pds_lm_resume_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> +}; >> + >> +/** >> + * struct pds_lm_sg_elem - Transmit scatter-gather (SG) descriptor element >> + * @addr: DMA address of SG element data buffer >> + * @len: Length of SG element data buffer, in bytes >> + * @rsvd: Word boundary padding >> + */ >> +struct pds_lm_sg_elem { >> + __le64 addr; >> + __le32 len; >> + __le16 rsvd[2]; >> +}; >> + >> +/** >> + * struct pds_lm_save_cmd - SAVE command >> + * @opcode: Opcode PDS_LM_CMD_SAVE >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + * @rsvd2: Word boundary padding >> + * @sgl_addr: IOVA address of the SGL to dma the device state >> + * @num_sge: Total number of SG elements >> + */ >> +struct pds_lm_save_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> + u8 rsvd2[4]; >> + __le64 sgl_addr; >> + __le32 num_sge; >> +} __packed; >> + >> +/** >> + * struct pds_lm_restore_cmd - RESTORE command >> + * @opcode: Opcode PDS_LM_CMD_RESTORE >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + * @rsvd2: Word boundary padding >> + * @sgl_addr: IOVA address of the SGL to dma the device state >> + * @num_sge: Total number of SG elements >> + */ >> +struct pds_lm_restore_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> + u8 rsvd2[4]; >> + __le64 sgl_addr; >> + __le32 num_sge; >> +} __packed; >> + >> +/** >> + * union pds_lm_dev_state - device state information >> + * @words: Device state words >> + */ >> +union pds_lm_dev_state { >> + __le32 words[PDS_LM_DEVICE_STATE_LENGTH / sizeof(__le32)]; >> +}; >> + >> +enum pds_lm_host_vf_status { >> + PDS_LM_STA_NONE = 0, >> + PDS_LM_STA_IN_PROGRESS, >> + PDS_LM_STA_MAX, >> +}; >> + >> +/** >> + * struct pds_lm_host_vf_status_cmd - HOST_VF_STATUS command >> + * @opcode: Opcode PDS_LM_CMD_HOST_VF_STATUS >> + * @rsvd: Word boundary padding >> + * @vf_id: VF id >> + * @status: Current LM status of host VF driver (enum >> pds_lm_host_status) >> + */ >> +struct pds_lm_host_vf_status_cmd { >> + u8 opcode; >> + u8 rsvd; >> + __le16 vf_id; >> + u8 status; >> +}; >> + >> union pds_core_adminq_cmd { >> u8 opcode; >> u8 bytes[64]; >> @@ -600,6 +807,14 @@ union pds_core_adminq_cmd { >> >> struct pds_core_q_identify_cmd q_ident; >> struct pds_core_q_init_cmd q_init; >> + >> + struct pds_lm_suspend_cmd lm_suspend; >> + struct pds_lm_suspend_status_cmd lm_suspend_status; >> + struct pds_lm_resume_cmd lm_resume; >> + struct pds_lm_status_cmd lm_status; >> + struct pds_lm_save_cmd lm_save; >> + struct pds_lm_restore_cmd lm_restore; >> + struct pds_lm_host_vf_status_cmd lm_host_vf_status; >> }; >> >> union pds_core_adminq_comp { >> @@ -621,6 +836,8 @@ union pds_core_adminq_comp { >> >> struct pds_core_q_identify_comp q_ident; >> struct pds_core_q_init_comp q_init; >> + >> + struct pds_lm_status_comp lm_status; >> }; >> >> #ifndef __CHECKER__ >> -- >> 2.17.1 >
> From: Brett Creeley <brett.creeley@amd.com> > Sent: Saturday, June 3, 2023 6:03 AM > > Add live migration support via the VFIO subsystem. The migration > implementation aligns with the definition from uapi/vfio.h and uses > the pds_core PF's adminq for device configuration. > > The ability to suspend, resume, and transfer VF device state data is > included along with the required admin queue command structures and > implementations. > > PDS_LM_CMD_SUSPEND and PDS_LM_CMD_SUSPEND_STATUS are added to > support > the VF device suspend operation. > > PDS_LM_CMD_RESUME is added to support the VF device resume operation. > > PDS_LM_CMD_STATUS is added to determine the exact size of the VF > device state data. > > PDS_LM_CMD_SAVE is added to get the VF device state data. > > PDS_LM_CMD_RESTORE is added to restore the VF device with the > previously saved data from PDS_LM_CMD_SAVE. > > PDS_LM_CMD_HOST_VF_STATUS is added to notify the device when > a migration is in/not-in progress from the host's perspective. Here is 'the device' referring to the PF or VF? and how would the device use this information? > + > +static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device *pds_vfio, > + union pds_core_adminq_cmd *req, > + size_t req_len, > + union pds_core_adminq_comp *resp, > + u64 flags) > +{ > + union pds_core_adminq_cmd cmd = {}; > + size_t cp_len; > + int err; > + > + /* Wrap the client request */ > + cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD; > + cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id); > + cp_len = min_t(size_t, req_len, > sizeof(cmd.client_request.client_cmd)); 'req_len' is kind of redundant. Looks all the callers use sizeof(req). > +static int > +pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_suspend_status = { > + .opcode = PDS_LM_CMD_SUSPEND_STATUS, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + unsigned long time_limit; > + unsigned long time_start; > + unsigned long time_done; > + int err; > + > + time_start = jiffies; > + time_limit = time_start + HZ * SUSPEND_TIMEOUT_S; > + do { > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, > sizeof(cmd), > + &comp, > PDS_AQ_FLAG_FASTPOLL); > + if (err != -EAGAIN) > + break; > + > + msleep(SUSPEND_CHECK_INTERVAL_MS); > + } while (time_before(jiffies, time_limit)); pds_vfio_client_adminq_cmd() has the exactly same mechanism with 5s timeout and 1ms poll interval when FASTPOLL is set. probably you can introduce another flag to indicate retry on -EAGAIN and then handle it fully in pds_vfio_client_adminq_cmd()? > +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_suspend = { > + .opcode = PDS_LM_CMD_SUSPEND, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + int err; > + > + dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id); > + > + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), > &comp, > + PDS_AQ_FLAG_FASTPOLL); > + if (err) { > + dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id, > + ERR_PTR(err)); > + return err; > + } > + > + return pds_vfio_suspend_wait_device_cmd(pds_vfio); > +} The logic in this function is very confusing. PDS_LM_CMD_SUSPEND has a completion record: +struct pds_lm_suspend_comp { + u8 status; + u8 rsvd; + __le16 comp_index; + union { + __le64 state_size; + u8 rsvd2[11]; + } __packed; + u8 color; Presumably this function can look at the completion record to know whether the suspend request succeeds. Why do you require another wait_device step to query the suspend status? and I have another question. Is it correct to hard-code the 5s timeout in the kernel w/o any input from the VMM? Note the guest has been stopped at this point then very likely the 5s timeout will kill any reasonable SLA which CSPs try to reach hard. Ideally the VMM has an estimation how long a VM can be paused based on SLA, to-be-migrated state size, available network bandwidth, etc. and that hint should be passed to the kernel so any state transition which may violate that expectation can fail quickly to break the migration process and put the VM back to the running state. Jason/Shameer, is there similar concern in mlx/hisilicon drivers? > + > +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio) > +{ > + union pds_core_adminq_cmd cmd = { > + .lm_resume = { > + .opcode = PDS_LM_CMD_RESUME, > + .vf_id = cpu_to_le16(pds_vfio->vf_id), > + }, > + }; > + struct device *dev = pds_vfio_to_dev(pds_vfio); > + union pds_core_adminq_comp comp = {}; > + > + dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id); > + > + return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), > &comp, > + 0); 'resume' is also in the blackout phase when the guest is not running. So presumably FAST_POLL should be set otherwise the max 256ms poll interval (PDSC_ADMINQ_MAX_POLL_INTERVAL) is really inefficient. > + > + if (cur == VFIO_DEVICE_STATE_RUNNING && next == > VFIO_DEVICE_STATE_RUNNING_P2P) { > + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, > + > PDS_LM_STA_IN_PROGRESS); > + err = pds_vfio_suspend_device_cmd(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + return NULL; > + } > + > + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == > VFIO_DEVICE_STATE_RUNNING) { > + err = pds_vfio_resume_device_cmd(pds_vfio); > + if (err) > + return ERR_PTR(err); > + > + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, > PDS_LM_STA_NONE); > + return NULL; > + } > + > + if (cur == VFIO_DEVICE_STATE_STOP && next == > VFIO_DEVICE_STATE_RUNNING_P2P) > + return NULL; > + > + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == > VFIO_DEVICE_STATE_STOP) > + return NULL; I'm not sure whether P2P is actually supported here. By definition P2P means the device is stopped but still responds to p2p request from other devices. If you look at mlx example it uses different cmds between RUNNING->RUNNING_P2P and RUNNING_P2P->STOP. But in your case seems you simply move what is required in STOP into P2P. Probably you can just remove the support of P2P like hisilicon does. > + > +/** > + * struct pds_lm_comp - generic command completion > + * @status: Status of the command (enum pds_core_status_code) > + * @rsvd: Structure padding to 16 Bytes > + */ > +struct pds_lm_comp { > + u8 status; > + u8 rsvd[15]; > +}; not used. Looks most comp structures are defined w/o an user except struct pds_lm_status_comp.
On 6/16/2023 1:06 AM, Tian, Kevin wrote: > Caution: This message originated from an External Source. Use proper caution when opening attachments, clicking links, or responding. > > >> From: Brett Creeley <brett.creeley@amd.com> >> Sent: Saturday, June 3, 2023 6:03 AM >> >> Add live migration support via the VFIO subsystem. The migration >> implementation aligns with the definition from uapi/vfio.h and uses >> the pds_core PF's adminq for device configuration. >> >> The ability to suspend, resume, and transfer VF device state data is >> included along with the required admin queue command structures and >> implementations. >> >> PDS_LM_CMD_SUSPEND and PDS_LM_CMD_SUSPEND_STATUS are added to >> support >> the VF device suspend operation. >> >> PDS_LM_CMD_RESUME is added to support the VF device resume operation. >> >> PDS_LM_CMD_STATUS is added to determine the exact size of the VF >> device state data. >> >> PDS_LM_CMD_SAVE is added to get the VF device state data. >> >> PDS_LM_CMD_RESTORE is added to restore the VF device with the >> previously saved data from PDS_LM_CMD_SAVE. >> >> PDS_LM_CMD_HOST_VF_STATUS is added to notify the device when >> a migration is in/not-in progress from the host's perspective. > > Here is 'the device' referring to the PF or VF? Device is referring to the DSC/firmware not the function. I will clarify the wording here. Thanks. > > and how would the device use this information? > >> + >> +static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device *pds_vfio, >> + union pds_core_adminq_cmd *req, >> + size_t req_len, >> + union pds_core_adminq_comp *resp, >> + u64 flags) >> +{ >> + union pds_core_adminq_cmd cmd = {}; >> + size_t cp_len; >> + int err; >> + >> + /* Wrap the client request */ >> + cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD; >> + cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id); >> + cp_len = min_t(size_t, req_len, >> sizeof(cmd.client_request.client_cmd)); > > 'req_len' is kind of redundant. Looks all the callers use sizeof(req). It does a memcpy based on the min size between req_len and the size of the request. > >> +static int >> +pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_suspend_status = { >> + .opcode = PDS_LM_CMD_SUSPEND_STATUS, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + unsigned long time_limit; >> + unsigned long time_start; >> + unsigned long time_done; >> + int err; >> + >> + time_start = jiffies; >> + time_limit = time_start + HZ * SUSPEND_TIMEOUT_S; >> + do { >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, >> sizeof(cmd), >> + &comp, >> PDS_AQ_FLAG_FASTPOLL); >> + if (err != -EAGAIN) >> + break; >> + >> + msleep(SUSPEND_CHECK_INTERVAL_MS); >> + } while (time_before(jiffies, time_limit)); > > pds_vfio_client_adminq_cmd() has the exactly same mechanism > with 5s timeout and 1ms poll interval when FASTPOLL is set. > > probably you can introduce another flag to indicate retry on > -EAGAIN and then handle it fully in pds_vfio_client_adminq_cmd()? That's the entire purpose of this command. It uses pds_vfio_client_adminq_cmd() to poll the SUSPEND_STATUS. IMHO adding the polling mechanism in pds_vfio_client_adminq_cmd() and using it depending on a flag is just adding to the complexity and not offering any benefit. I plan to keep this function as is to separate the functionality. Thanks. > >> +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_suspend = { >> + .opcode = PDS_LM_CMD_SUSPEND, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + int err; >> + >> + dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id); >> + >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), >> &comp, >> + PDS_AQ_FLAG_FASTPOLL); >> + if (err) { >> + dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id, >> + ERR_PTR(err)); >> + return err; >> + } >> + >> + return pds_vfio_suspend_wait_device_cmd(pds_vfio); >> +} > > The logic in this function is very confusing. > > PDS_LM_CMD_SUSPEND has a completion record: > > +struct pds_lm_suspend_comp { > + u8 status; > + u8 rsvd; > + __le16 comp_index; > + union { > + __le64 state_size; > + u8 rsvd2[11]; > + } __packed; > + u8 color; > > Presumably this function can look at the completion record to know whether > the suspend request succeeds. > > Why do you require another wait_device step to query the suspend status? The driver sends the initial suspend request to tell the DSC/firmware to suspend the VF's data/control path. The DSC/firmware will ack/nack the suspend request in the completion. Then the driver polls the DSC/firmware to find when the VF's data/control path has been fully suspended. When the DSC/firmware isn't done suspending yet it will return -EAGAIN. Otherwise it will return success/failure. I will add some comments clarifying these details. > > and I have another question. Is it correct to hard-code the 5s timeout in > the kernel w/o any input from the VMM? Note the guest has been stopped > at this point then very likely the 5s timeout will kill any reasonable SLA which > CSPs try to reach hard. This gives the device a max of 5 seconds to suspend the VF's data/control path. > > Ideally the VMM has an estimation how long a VM can be paused based on > SLA, to-be-migrated state size, available network bandwidth, etc. and that > hint should be passed to the kernel so any state transition which may violate > that expectation can fail quickly to break the migration process and put the > VM back to the running state. For QEMU there is a parameter that can specify the downtime-limit that's used as you mentioned, but this does not include how long it takes the device to STOP (i.e. suspend the data/control path). > > Jason/Shameer, is there similar concern in mlx/hisilicon drivers? > >> + >> +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio) >> +{ >> + union pds_core_adminq_cmd cmd = { >> + .lm_resume = { >> + .opcode = PDS_LM_CMD_RESUME, >> + .vf_id = cpu_to_le16(pds_vfio->vf_id), >> + }, >> + }; >> + struct device *dev = pds_vfio_to_dev(pds_vfio); >> + union pds_core_adminq_comp comp = {}; >> + >> + dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id); >> + >> + return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), >> &comp, >> + 0); > > 'resume' is also in the blackout phase when the guest is not running. > > So presumably FAST_POLL should be set otherwise the max 256ms > poll interval (PDSC_ADMINQ_MAX_POLL_INTERVAL) is really inefficient. Yeah this is a good catch. I think setting fast_poll = true would be a good idea here. Thanks. > >> + >> + if (cur == VFIO_DEVICE_STATE_RUNNING && next == >> VFIO_DEVICE_STATE_RUNNING_P2P) { >> + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, >> + >> PDS_LM_STA_IN_PROGRESS); >> + err = pds_vfio_suspend_device_cmd(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + return NULL; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == >> VFIO_DEVICE_STATE_RUNNING) { >> + err = pds_vfio_resume_device_cmd(pds_vfio); >> + if (err) >> + return ERR_PTR(err); >> + >> + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, >> PDS_LM_STA_NONE); >> + return NULL; >> + } >> + >> + if (cur == VFIO_DEVICE_STATE_STOP && next == >> VFIO_DEVICE_STATE_RUNNING_P2P) >> + return NULL; >> + >> + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == >> VFIO_DEVICE_STATE_STOP) >> + return NULL; > > I'm not sure whether P2P is actually supported here. By definition > P2P means the device is stopped but still responds to p2p request > from other devices. If you look at mlx example it uses different > cmds between RUNNING->RUNNING_P2P and RUNNING_P2P->STOP. > > But in your case seems you simply move what is required in STOP > into P2P. Probably you can just remove the support of P2P like > hisilicon does. In a previous review it was mentioned that P2P is more or less supported and this is how we are able to support it. Ideally we would not set the P2P feature and just implement the standard STOP/RUNNING states. > >> + >> +/** >> + * struct pds_lm_comp - generic command completion >> + * @status: Status of the command (enum pds_core_status_code) >> + * @rsvd: Structure padding to 16 Bytes >> + */ >> +struct pds_lm_comp { >> + u8 status; >> + u8 rsvd[15]; >> +}; > > not used. Looks most comp structures are defined w/o an user > except struct pds_lm_status_comp. I will look into this. Thanks.
On Fri, Jun 16, 2023 at 08:06:21AM +0000, Tian, Kevin wrote: > Ideally the VMM has an estimation how long a VM can be paused based on > SLA, to-be-migrated state size, available network bandwidth, etc. and that > hint should be passed to the kernel so any state transition which may violate > that expectation can fail quickly to break the migration process and put the > VM back to the running state. > > Jason/Shameer, is there similar concern in mlx/hisilicon drivers? It is handled through the vfio_device_feature_mig_data_size mechanism.. > > + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == > > VFIO_DEVICE_STATE_STOP) > > + return NULL; > > I'm not sure whether P2P is actually supported here. By definition > P2P means the device is stopped but still responds to p2p request > from other devices. If you look at mlx example it uses different > cmds between RUNNING->RUNNING_P2P and RUNNING_P2P->STOP. > > But in your case seems you simply move what is required in STOP > into P2P. Probably you can just remove the support of P2P like > hisilicon does. We want new devices to get their architecture right, they need to support P2P. Didn't we talk about this already and Brett was going to fix it? Jason
> From: Jason Gunthorpe <jgg@nvidia.com> > Sent: Monday, June 19, 2023 8:47 PM > > On Fri, Jun 16, 2023 at 08:06:21AM +0000, Tian, Kevin wrote: > > > Ideally the VMM has an estimation how long a VM can be paused based on > > SLA, to-be-migrated state size, available network bandwidth, etc. and that > > hint should be passed to the kernel so any state transition which may > violate > > that expectation can fail quickly to break the migration process and put the > > VM back to the running state. > > > > Jason/Shameer, is there similar concern in mlx/hisilicon drivers? > > It is handled through the vfio_device_feature_mig_data_size mechanism.. that is only for estimation of copied data. IMHO the stop time when the VM is paused includes both the time of stopping the device and the time of migrating the VM state. For a software-emulated device the time of stopping the device is negligible. But certainly for assigned device the worst-case hard-coded 5s timeout as done in this patch will kill whatever reasonable 'VM dead time' SLA (usually in milliseconds) which CSPs try to meet purely based on the size of copied data. Wouldn't a user-specified stop-device timeout be required to at least allow breaking migration early according to the desired SLA? > > > > + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == > > > VFIO_DEVICE_STATE_STOP) > > > + return NULL; > > > > I'm not sure whether P2P is actually supported here. By definition > > P2P means the device is stopped but still responds to p2p request > > from other devices. If you look at mlx example it uses different > > cmds between RUNNING->RUNNING_P2P and RUNNING_P2P->STOP. > > > > But in your case seems you simply move what is required in STOP > > into P2P. Probably you can just remove the support of P2P like > > hisilicon does. > > We want new devices to get their architecture right, they need to > support P2P. Didn't we talk about this already and Brett was going to > fix it? > Looks it's not fixed since RUNNING_P2P->STOP is a nop in this patch.
> From: Brett Creeley <bcreeley@amd.com> > Sent: Saturday, June 17, 2023 12:45 PM > > On 6/16/2023 1:06 AM, Tian, Kevin wrote: > > Caution: This message originated from an External Source. Use proper > caution when opening attachments, clicking links, or responding. > > > > > >> From: Brett Creeley <brett.creeley@amd.com> > >> Sent: Saturday, June 3, 2023 6:03 AM > >> > >> + > >> +static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device > *pds_vfio, > >> + union pds_core_adminq_cmd *req, > >> + size_t req_len, > >> + union pds_core_adminq_comp *resp, > >> + u64 flags) > >> +{ > >> + union pds_core_adminq_cmd cmd = {}; > >> + size_t cp_len; > >> + int err; > >> + > >> + /* Wrap the client request */ > >> + cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD; > >> + cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id); > >> + cp_len = min_t(size_t, req_len, > >> sizeof(cmd.client_request.client_cmd)); > > > > 'req_len' is kind of redundant. Looks all the callers use sizeof(req). > > It does a memcpy based on the min size between req_len and the size of > the request. If all the callers just pass in sizeof(union) as 'req_len', then it's pointless to do min_t and you can just use sizeof(cmd.client_request.client_cmd) here which is always smaller than or equal to the sizeof(union). > >> + > >> + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), > >> &comp, > >> + PDS_AQ_FLAG_FASTPOLL); > >> + if (err) { > >> + dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id, > >> + ERR_PTR(err)); > >> + return err; > >> + } > >> + > >> + return pds_vfio_suspend_wait_device_cmd(pds_vfio); > >> +} > > > > The logic in this function is very confusing. > > > > PDS_LM_CMD_SUSPEND has a completion record: > > > > +struct pds_lm_suspend_comp { > > + u8 status; > > + u8 rsvd; > > + __le16 comp_index; > > + union { > > + __le64 state_size; > > + u8 rsvd2[11]; > > + } __packed; > > + u8 color; > > > > Presumably this function can look at the completion record to know > whether > > the suspend request succeeds. > > > > Why do you require another wait_device step to query the suspend status? > > The driver sends the initial suspend request to tell the DSC/firmware to > suspend the VF's data/control path. The DSC/firmware will ack/nack the > suspend request in the completion. > > Then the driver polls the DSC/firmware to find when the VF's > data/control path has been fully suspended. When the DSC/firmware isn't > done suspending yet it will return -EAGAIN. Otherwise it will return > success/failure. > > I will add some comments clarifying these details. Yes more comment is welcomed. It's also misleading to have a ' state_size ' field in suspend_comp. In concept the firmware cannot calculate it accurately before the VF is fully suspended.
On Tue, Jun 20, 2023 at 02:02:44AM +0000, Tian, Kevin wrote: > > From: Jason Gunthorpe <jgg@nvidia.com> > > Sent: Monday, June 19, 2023 8:47 PM > > > > On Fri, Jun 16, 2023 at 08:06:21AM +0000, Tian, Kevin wrote: > > > > > Ideally the VMM has an estimation how long a VM can be paused based on > > > SLA, to-be-migrated state size, available network bandwidth, etc. and that > > > hint should be passed to the kernel so any state transition which may > > violate > > > that expectation can fail quickly to break the migration process and put the > > > VM back to the running state. > > > > > > Jason/Shameer, is there similar concern in mlx/hisilicon drivers? > > > > It is handled through the vfio_device_feature_mig_data_size mechanism.. > > that is only for estimation of copied data. > > IMHO the stop time when the VM is paused includes both the time of > stopping the device and the time of migrating the VM state. > > For a software-emulated device the time of stopping the device is negligible. > > But certainly for assigned device the worst-case hard-coded 5s timeout as > done in this patch will kill whatever reasonable 'VM dead time' SLA (usually > in milliseconds) which CSPs try to meet purely based on the size of copied > data. There is not alot that can be done here, the stop time cannot be predicted in advance on these devices - the system relies on the device having a reasonable time window. > Wouldn't a user-specified stop-device timeout be required to at least allow > breaking migration early according to the desired SLA? Not really, the device is going to still execute the stop regardless of the timeout, and when it does the VM will be broken. With a FW approach like this it is pretty stuck, we need the FW to remain in sync as the highest priority. > > We want new devices to get their architecture right, they need to > > support P2P. Didn't we talk about this already and Brett was going to > > fix it? > > Looks it's not fixed since RUNNING_P2P->STOP is a nop in this patch. That could be OK, it needs a comment explaining why it is OK Jason
> From: Jason Gunthorpe <jgg@nvidia.com> > Sent: Tuesday, June 20, 2023 8:31 PM > > On Tue, Jun 20, 2023 at 02:02:44AM +0000, Tian, Kevin wrote: > > > From: Jason Gunthorpe <jgg@nvidia.com> > > > Sent: Monday, June 19, 2023 8:47 PM > > > > > > On Fri, Jun 16, 2023 at 08:06:21AM +0000, Tian, Kevin wrote: > > > > > > > Ideally the VMM has an estimation how long a VM can be paused based > on > > > > SLA, to-be-migrated state size, available network bandwidth, etc. and > that > > > > hint should be passed to the kernel so any state transition which may > > > violate > > > > that expectation can fail quickly to break the migration process and put > the > > > > VM back to the running state. > > > > > > > > Jason/Shameer, is there similar concern in mlx/hisilicon drivers? > > > > > > It is handled through the vfio_device_feature_mig_data_size mechanism.. > > > > that is only for estimation of copied data. > > > > IMHO the stop time when the VM is paused includes both the time of > > stopping the device and the time of migrating the VM state. > > > > For a software-emulated device the time of stopping the device is negligible. > > > > But certainly for assigned device the worst-case hard-coded 5s timeout as > > done in this patch will kill whatever reasonable 'VM dead time' SLA (usually > > in milliseconds) which CSPs try to meet purely based on the size of copied > > data. > > There is not alot that can be done here, the stop time cannot be > predicted in advance on these devices - the system relies on the > device having a reasonable time window. What is the criteria for 'reasonable'? How does CSPs judge that such device can guarantee a *reliable* reasonable window so live migration can be enabled in the production environment? I'm afraid that we are hiding a non-deterministic factor in current protocol. Looking at mlx5 case which has a even larger timeout: [MLX5_TO_CMD_MS] = 60000, > > > Wouldn't a user-specified stop-device timeout be required to at least allow > > breaking migration early according to the desired SLA? > > Not really, the device is going to still execute the stop regardless > of the timeout, and when it does the VM will be broken. > > With a FW approach like this it is pretty stuck, we need the FW to > remain in sync as the highest priority. This makes some sense. But still I don't think it's a good situation where the user has ZERO knowledge about the non-negligible time in the stopping path... > > > > We want new devices to get their architecture right, they need to > > > support P2P. Didn't we talk about this already and Brett was going to > > > fix it? > > > > Looks it's not fixed since RUNNING_P2P->STOP is a nop in this patch. > > That could be OK, it needs a comment explaining why it is OK > Yes, a comment is welcomed. having RUNNING_P2P->STOP as nop kind of suggest that the device has been fully stopped in RUNNING_P2P to meet the definition of the STOP state. But then it violates the definition of RUNNING_P2P.
On Wed, Jun 21, 2023 at 06:49:12AM +0000, Tian, Kevin wrote: > What is the criteria for 'reasonable'? How does CSPs judge that such > device can guarantee a *reliable* reasonable window so live migration > can be enabled in the production environment? The CSP needs to work with the device vendor to understand how it fits into their system, I don't see how we can externalize this kind of detail in a general way. > I'm afraid that we are hiding a non-deterministic factor in current protocol. Yes > But still I don't think it's a good situation where the user has ZERO > knowledge about the non-negligible time in the stopping path... In any sane device design this will be a small period of time. These timeouts should be to protect against a device that has gone wild. Jason
> From: Jason Gunthorpe <jgg@nvidia.com> > Sent: Wednesday, June 21, 2023 9:27 PM > > On Wed, Jun 21, 2023 at 06:49:12AM +0000, Tian, Kevin wrote: > > > What is the criteria for 'reasonable'? How does CSPs judge that such > > device can guarantee a *reliable* reasonable window so live migration > > can be enabled in the production environment? > > The CSP needs to work with the device vendor to understand how it fits > into their system, I don't see how we can externalize this kind of > detail in a general way. > > > I'm afraid that we are hiding a non-deterministic factor in current protocol. > > Yes > > > But still I don't think it's a good situation where the user has ZERO > > knowledge about the non-negligible time in the stopping path... > > In any sane device design this will be a small period of time. These > timeouts should be to protect against a device that has gone wild. > Any example how 'small' it will be (e.g. <1ms)? Should we define a *reasonable* threshold in VFIO community which any new variant driver should provide information to judge against? If the worst-case stop time (assuming the device doesn't go wild) may exceed the threshold then it's time to consider whether a new interface is required to communicate such constraint to userspace. The reason why I keep discussing it is that IMHO achieving negligible stop time is a very challenging task for many accelerators. e.g. IDXD can be stopped only after completing all the pending requests. While it allows software to configure the max pending work size (and a reasonable setting could meet both migration SLA and performance SLA) the worst-case draining latency could be in 10's milliseconds which cannot be ignored by the VMM. Or do you think it's still better left to CSP working with the device vendor even in this case, given the worst-case latency could be affected by many factors hence not something which a kernel driver can accurately estimate? Thanks Kevin
On Mon, Jun 26, 2023 at 07:31:31AM +0000, Tian, Kevin wrote: > > From: Jason Gunthorpe <jgg@nvidia.com> > > Sent: Wednesday, June 21, 2023 9:27 PM > > > > On Wed, Jun 21, 2023 at 06:49:12AM +0000, Tian, Kevin wrote: > > > > > What is the criteria for 'reasonable'? How does CSPs judge that such > > > device can guarantee a *reliable* reasonable window so live migration > > > can be enabled in the production environment? > > > > The CSP needs to work with the device vendor to understand how it fits > > into their system, I don't see how we can externalize this kind of > > detail in a general way. > > > > > I'm afraid that we are hiding a non-deterministic factor in current protocol. > > > > Yes > > > > > But still I don't think it's a good situation where the user has ZERO > > > knowledge about the non-negligible time in the stopping path... > > > > In any sane device design this will be a small period of time. These > > timeouts should be to protect against a device that has gone wild. > > > > Any example how 'small' it will be (e.g. <1ms)? Not personally.. > Should we define a *reasonable* threshold in VFIO community which > any new variant driver should provide information to judge against? Ah, I think we are just too new to get into such details. I think we need some real world experience to see if this is really an issue. > The reason why I keep discussing it is that IMHO achieving negligible > stop time is a very challenging task for many accelerators. e.g. IDXD > can be stopped only after completing all the pending requests. While > it allows software to configure the max pending work size (and a > reasonable setting could meet both migration SLA and performance > SLA) the worst-case draining latency could be in 10's milliseconds which > cannot be ignored by the VMM. Well, what would you report here if you had the opportunity to report something? Some big number? Then what? > Or do you think it's still better left to CSP working with the device vendor > even in this case, given the worst-case latency could be affected by > many factors hence not something which a kernel driver can accurately > estimate? This is my fear, that it is so complicated that reducing it to any sort of cross-vendor data is not feasible. At least I'd like to see someone experiment with what information would be useful to qemu before we add kernel ABI.. Jason
> From: Jason Gunthorpe <jgg@nvidia.com> > Sent: Tuesday, June 27, 2023 2:14 AM > > On Mon, Jun 26, 2023 at 07:31:31AM +0000, Tian, Kevin wrote: > > > From: Jason Gunthorpe <jgg@nvidia.com> > > > Sent: Wednesday, June 21, 2023 9:27 PM > > > > > > On Wed, Jun 21, 2023 at 06:49:12AM +0000, Tian, Kevin wrote: > > > > > > > What is the criteria for 'reasonable'? How does CSPs judge that such > > > > device can guarantee a *reliable* reasonable window so live migration > > > > can be enabled in the production environment? > > > > > > The CSP needs to work with the device vendor to understand how it fits > > > into their system, I don't see how we can externalize this kind of > > > detail in a general way. > > > > > > > I'm afraid that we are hiding a non-deterministic factor in current > protocol. > > > > > > Yes > > > > > > > But still I don't think it's a good situation where the user has ZERO > > > > knowledge about the non-negligible time in the stopping path... > > > > > > In any sane device design this will be a small period of time. These > > > timeouts should be to protect against a device that has gone wild. > > > > > > > Any example how 'small' it will be (e.g. <1ms)? > > Not personally.. > > > Should we define a *reasonable* threshold in VFIO community which > > any new variant driver should provide information to judge against? > > Ah, I think we are just too new to get into such details. I think we > need some real world experience to see if this is really an issue. > > > The reason why I keep discussing it is that IMHO achieving negligible > > stop time is a very challenging task for many accelerators. e.g. IDXD > > can be stopped only after completing all the pending requests. While > > it allows software to configure the max pending work size (and a > > reasonable setting could meet both migration SLA and performance > > SLA) the worst-case draining latency could be in 10's milliseconds which > > cannot be ignored by the VMM. > > Well, what would you report here if you had the opportunity to report > something? Some big number? Then what? > > > Or do you think it's still better left to CSP working with the device vendor > > even in this case, given the worst-case latency could be affected by > > many factors hence not something which a kernel driver can accurately > > estimate? > > This is my fear, that it is so complicated that reducing it to any > sort of cross-vendor data is not feasible. At least I'd like to see > someone experiment with what information would be useful to qemu > before we add kernel ABI.. > OK. make sense.
diff --git a/drivers/vfio/pci/pds/Makefile b/drivers/vfio/pci/pds/Makefile index 87581111fa17..dbaf613d3794 100644 --- a/drivers/vfio/pci/pds/Makefile +++ b/drivers/vfio/pci/pds/Makefile @@ -5,5 +5,6 @@ obj-$(CONFIG_PDS_VFIO_PCI) += pds_vfio.o pds_vfio-y := \ cmds.o \ + lm.o \ pci_drv.o \ vfio_dev.o diff --git a/drivers/vfio/pci/pds/cmds.c b/drivers/vfio/pci/pds/cmds.c index ae01f5df2f5c..256f458feb58 100644 --- a/drivers/vfio/pci/pds/cmds.c +++ b/drivers/vfio/pci/pds/cmds.c @@ -3,6 +3,7 @@ #include <linux/io.h> #include <linux/types.h> +#include <linux/delay.h> #include <linux/pds/pds_common.h> #include <linux/pds/pds_core_if.h> @@ -11,6 +12,34 @@ #include "vfio_dev.h" #include "cmds.h" +#define SUSPEND_TIMEOUT_S 5 +#define SUSPEND_CHECK_INTERVAL_MS 1 + +static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device *pds_vfio, + union pds_core_adminq_cmd *req, + size_t req_len, + union pds_core_adminq_comp *resp, + u64 flags) +{ + union pds_core_adminq_cmd cmd = {}; + size_t cp_len; + int err; + + /* Wrap the client request */ + cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD; + cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id); + cp_len = min_t(size_t, req_len, sizeof(cmd.client_request.client_cmd)); + memcpy(cmd.client_request.client_cmd, req, cp_len); + + err = pdsc_adminq_post(pds_vfio->pdsc, &cmd, resp, + !!(flags & PDS_AQ_FLAG_FASTPOLL)); + if (err && err != -EAGAIN) + dev_info(pds_vfio_to_dev(pds_vfio), + "client admin cmd failed: %pe\n", ERR_PTR(err)); + + return err; +} + int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio) { struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); @@ -41,3 +70,293 @@ void pds_vfio_unregister_client_cmd(struct pds_vfio_pci_device *pds_vfio) pds_vfio->client_id = 0; } + +static int +pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio) +{ + union pds_core_adminq_cmd cmd = { + .lm_suspend_status = { + .opcode = PDS_LM_CMD_SUSPEND_STATUS, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + }, + }; + struct device *dev = pds_vfio_to_dev(pds_vfio); + union pds_core_adminq_comp comp = {}; + unsigned long time_limit; + unsigned long time_start; + unsigned long time_done; + int err; + + time_start = jiffies; + time_limit = time_start + HZ * SUSPEND_TIMEOUT_S; + do { + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), + &comp, PDS_AQ_FLAG_FASTPOLL); + if (err != -EAGAIN) + break; + + msleep(SUSPEND_CHECK_INTERVAL_MS); + } while (time_before(jiffies, time_limit)); + + time_done = jiffies; + dev_dbg(dev, "%s: vf%u: Suspend comp received in %d msecs\n", __func__, + pds_vfio->vf_id, jiffies_to_msecs(time_done - time_start)); + + /* Check the results */ + if (time_after_eq(time_done, time_limit)) { + dev_err(dev, "%s: vf%u: Suspend comp timeout\n", __func__, + pds_vfio->vf_id); + err = -ETIMEDOUT; + } + + return err; +} + +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio) +{ + union pds_core_adminq_cmd cmd = { + .lm_suspend = { + .opcode = PDS_LM_CMD_SUSPEND, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + }, + }; + struct device *dev = pds_vfio_to_dev(pds_vfio); + union pds_core_adminq_comp comp = {}; + int err; + + dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id); + + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, + PDS_AQ_FLAG_FASTPOLL); + if (err) { + dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id, + ERR_PTR(err)); + return err; + } + + return pds_vfio_suspend_wait_device_cmd(pds_vfio); +} + +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio) +{ + union pds_core_adminq_cmd cmd = { + .lm_resume = { + .opcode = PDS_LM_CMD_RESUME, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + }, + }; + struct device *dev = pds_vfio_to_dev(pds_vfio); + union pds_core_adminq_comp comp = {}; + + dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id); + + return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, + 0); +} + +int pds_vfio_get_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, u64 *size) +{ + union pds_core_adminq_cmd cmd = { + .lm_status = { + .opcode = PDS_LM_CMD_STATUS, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + }, + }; + struct device *dev = pds_vfio_to_dev(pds_vfio); + union pds_core_adminq_comp comp = {}; + int err; + + dev_dbg(dev, "vf%u: Get migration status\n", pds_vfio->vf_id); + + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, 0); + if (err) + return err; + + *size = le64_to_cpu(comp.lm_status.size); + return 0; +} + +static int pds_vfio_dma_map_lm_file(struct device *dev, + enum dma_data_direction dir, + struct pds_vfio_lm_file *lm_file) +{ + struct pds_lm_sg_elem *sgl, *sge; + struct scatterlist *sg; + dma_addr_t sgl_addr; + size_t sgl_size; + int err; + int i; + + if (!lm_file) + return -EINVAL; + + /* dma map file pages */ + err = dma_map_sgtable(dev, &lm_file->sg_table, dir, 0); + if (err) + return err; + + lm_file->num_sge = lm_file->sg_table.nents; + + /* alloc sgl */ + sgl_size = lm_file->num_sge * sizeof(struct pds_lm_sg_elem); + sgl = kzalloc(sgl_size, GFP_KERNEL); + if (!sgl) { + err = -ENOMEM; + goto out_unmap_sgtable; + } + + /* fill sgl */ + sge = sgl; + for_each_sgtable_dma_sg(&lm_file->sg_table, sg, i) { + sge->addr = cpu_to_le64(sg_dma_address(sg)); + sge->len = cpu_to_le32(sg_dma_len(sg)); + dev_dbg(dev, "addr = %llx, len = %u\n", sge->addr, sge->len); + sge++; + } + + sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE); + if (dma_mapping_error(dev, sgl_addr)) { + err = -EIO; + goto out_free_sgl; + } + + lm_file->sgl = sgl; + lm_file->sgl_addr = sgl_addr; + + return 0; + +out_free_sgl: + kfree(sgl); +out_unmap_sgtable: + lm_file->num_sge = 0; + dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0); + return err; +} + +static void pds_vfio_dma_unmap_lm_file(struct device *dev, + enum dma_data_direction dir, + struct pds_vfio_lm_file *lm_file) +{ + if (!lm_file) + return; + + /* free sgl */ + if (lm_file->sgl) { + dma_unmap_single(dev, lm_file->sgl_addr, + lm_file->num_sge * sizeof(*lm_file->sgl), + DMA_TO_DEVICE); + kfree(lm_file->sgl); + lm_file->sgl = NULL; + lm_file->sgl_addr = DMA_MAPPING_ERROR; + lm_file->num_sge = 0; + } + + /* dma unmap file pages */ + dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0); +} + +int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio) +{ + union pds_core_adminq_cmd cmd = { + .lm_save = { + .opcode = PDS_LM_CMD_SAVE, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + }, + }; + struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); + struct device *pdsc_dev = &pci_physfn(pdev)->dev; + union pds_core_adminq_comp comp = {}; + struct pds_vfio_lm_file *lm_file; + int err; + + dev_dbg(&pdev->dev, "vf%u: Get migration state\n", pds_vfio->vf_id); + + lm_file = pds_vfio->save_file; + + err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file); + if (err) { + dev_err(&pdev->dev, "failed to map save migration file: %pe\n", + ERR_PTR(err)); + return err; + } + + cmd.lm_save.sgl_addr = cpu_to_le64(lm_file->sgl_addr); + cmd.lm_save.num_sge = cpu_to_le32(lm_file->num_sge); + + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, 0); + if (err) + dev_err(&pdev->dev, "failed to get migration state: %pe\n", + ERR_PTR(err)); + + pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file); + + return err; +} + +int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio) +{ + union pds_core_adminq_cmd cmd = { + .lm_restore = { + .opcode = PDS_LM_CMD_RESTORE, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + }, + }; + struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio); + struct device *pdsc_dev = &pci_physfn(pdev)->dev; + union pds_core_adminq_comp comp = {}; + struct pds_vfio_lm_file *lm_file; + int err; + + dev_dbg(&pdev->dev, "vf%u: Set migration state\n", pds_vfio->vf_id); + + lm_file = pds_vfio->restore_file; + + err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file); + if (err) { + dev_err(&pdev->dev, + "failed to map restore migration file: %pe\n", + ERR_PTR(err)); + return err; + } + + cmd.lm_restore.sgl_addr = cpu_to_le64(lm_file->sgl_addr); + cmd.lm_restore.num_sge = cpu_to_le32(lm_file->num_sge); + + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, 0); + if (err) + dev_err(&pdev->dev, "failed to set migration state: %pe\n", + ERR_PTR(err)); + + pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file); + + return err; +} + +void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, + enum pds_lm_host_vf_status vf_status) +{ + union pds_core_adminq_cmd cmd = { + .lm_host_vf_status = { + .opcode = PDS_LM_CMD_HOST_VF_STATUS, + .vf_id = cpu_to_le16(pds_vfio->vf_id), + .status = vf_status, + }, + }; + struct device *dev = pds_vfio_to_dev(pds_vfio); + union pds_core_adminq_comp comp = {}; + int err; + + dev_dbg(dev, "vf%u: Set host VF LM status: %u", pds_vfio->vf_id, + vf_status); + if (vf_status != PDS_LM_STA_IN_PROGRESS && + vf_status != PDS_LM_STA_NONE) { + dev_warn(dev, "Invalid host VF migration status, %d\n", + vf_status); + return; + } + + err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, sizeof(cmd), &comp, 0); + if (err) + dev_warn(dev, "failed to send host VF migration status: %pe\n", + ERR_PTR(err)); +} diff --git a/drivers/vfio/pci/pds/cmds.h b/drivers/vfio/pci/pds/cmds.h index 4c592afccf89..3d8a5508c733 100644 --- a/drivers/vfio/pci/pds/cmds.h +++ b/drivers/vfio/pci/pds/cmds.h @@ -6,5 +6,11 @@ int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio); void pds_vfio_unregister_client_cmd(struct pds_vfio_pci_device *pds_vfio); - +int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio); +int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio); +int pds_vfio_get_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, u64 *size); +int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio); +int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio); +void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio, + enum pds_lm_host_vf_status vf_status); #endif /* _CMDS_H_ */ diff --git a/drivers/vfio/pci/pds/lm.c b/drivers/vfio/pci/pds/lm.c new file mode 100644 index 000000000000..c507f39a2339 --- /dev/null +++ b/drivers/vfio/pci/pds/lm.c @@ -0,0 +1,421 @@ +// SPDX-License-Identifier: GPL-2.0 +/* Copyright(c) 2023 Advanced Micro Devices, Inc. */ + +#include <linux/anon_inodes.h> +#include <linux/file.h> +#include <linux/fs.h> +#include <linux/highmem.h> +#include <linux/vfio.h> +#include <linux/vfio_pci_core.h> + +#include "vfio_dev.h" +#include "cmds.h" + +static struct pds_vfio_lm_file * +pds_vfio_get_lm_file(const struct file_operations *fops, int flags, u64 size) +{ + struct pds_vfio_lm_file *lm_file = NULL; + unsigned long long npages; + struct page **pages; + void *page_mem; + const void *p; + + if (!size) + return NULL; + + /* Alloc file structure */ + lm_file = kzalloc(sizeof(*lm_file), GFP_KERNEL); + if (!lm_file) + return NULL; + + /* Create file */ + lm_file->filep = + anon_inode_getfile("pds_vfio_lm", fops, lm_file, flags); + if (!lm_file->filep) + goto out_free_file; + + stream_open(lm_file->filep->f_inode, lm_file->filep); + mutex_init(&lm_file->lock); + + /* prevent file from being released before we are done with it */ + get_file(lm_file->filep); + + /* Allocate memory for file pages */ + npages = DIV_ROUND_UP_ULL(size, PAGE_SIZE); + pages = kmalloc_array(npages, sizeof(*pages), GFP_KERNEL); + if (!pages) + goto out_put_file; + + page_mem = kvzalloc(ALIGN(size, PAGE_SIZE), GFP_KERNEL); + if (!page_mem) + goto out_free_pages_array; + + p = page_mem - offset_in_page(page_mem); + for (unsigned long long i = 0; i < npages; i++) { + if (is_vmalloc_addr(p)) + pages[i] = vmalloc_to_page(p); + else + pages[i] = kmap_to_page((void *)p); + if (!pages[i]) + goto out_free_page_mem; + + p += PAGE_SIZE; + } + + /* Create scatterlist of file pages to use for DMA mapping later */ + if (sg_alloc_table_from_pages(&lm_file->sg_table, pages, npages, 0, + size, GFP_KERNEL)) + goto out_free_page_mem; + + lm_file->size = size; + lm_file->pages = pages; + lm_file->npages = npages; + lm_file->page_mem = page_mem; + lm_file->alloc_size = npages * PAGE_SIZE; + + return lm_file; + +out_free_page_mem: + kvfree(page_mem); +out_free_pages_array: + kfree(pages); +out_put_file: + fput(lm_file->filep); + mutex_destroy(&lm_file->lock); +out_free_file: + kfree(lm_file); + + return NULL; +} + +static void pds_vfio_put_lm_file(struct pds_vfio_lm_file *lm_file) +{ + mutex_lock(&lm_file->lock); + + lm_file->size = 0; + lm_file->alloc_size = 0; + + /* Free scatter list of file pages */ + sg_free_table(&lm_file->sg_table); + + kvfree(lm_file->page_mem); + lm_file->page_mem = NULL; + kfree(lm_file->pages); + lm_file->pages = NULL; + + mutex_unlock(&lm_file->lock); + + /* allow file to be released since we are done with it */ + fput(lm_file->filep); +} + +void pds_vfio_put_save_file(struct pds_vfio_pci_device *pds_vfio) +{ + if (!pds_vfio->save_file) + return; + + pds_vfio_put_lm_file(pds_vfio->save_file); + pds_vfio->save_file = NULL; +} + +void pds_vfio_put_restore_file(struct pds_vfio_pci_device *pds_vfio) +{ + if (!pds_vfio->restore_file) + return; + + pds_vfio_put_lm_file(pds_vfio->restore_file); + pds_vfio->restore_file = NULL; +} + +static struct page *pds_vfio_get_file_page(struct pds_vfio_lm_file *lm_file, + unsigned long offset) +{ + unsigned long cur_offset = 0; + struct scatterlist *sg; + unsigned int i; + + /* All accesses are sequential */ + if (offset < lm_file->last_offset || !lm_file->last_offset_sg) { + lm_file->last_offset = 0; + lm_file->last_offset_sg = lm_file->sg_table.sgl; + lm_file->sg_last_entry = 0; + } + + cur_offset = lm_file->last_offset; + + for_each_sg(lm_file->last_offset_sg, sg, + lm_file->sg_table.orig_nents - lm_file->sg_last_entry, i) { + if (offset < sg->length + cur_offset) { + lm_file->last_offset_sg = sg; + lm_file->sg_last_entry += i; + lm_file->last_offset = cur_offset; + return nth_page(sg_page(sg), + (offset - cur_offset) / PAGE_SIZE); + } + cur_offset += sg->length; + } + + return NULL; +} + +static int pds_vfio_release_file(struct inode *inode, struct file *filp) +{ + struct pds_vfio_lm_file *lm_file = filp->private_data; + + mutex_lock(&lm_file->lock); + lm_file->filep->f_pos = 0; + lm_file->size = 0; + mutex_unlock(&lm_file->lock); + mutex_destroy(&lm_file->lock); + kfree(lm_file); + + return 0; +} + +static ssize_t pds_vfio_save_read(struct file *filp, char __user *buf, + size_t len, loff_t *pos) +{ + struct pds_vfio_lm_file *lm_file = filp->private_data; + ssize_t done = 0; + + if (pos) + return -ESPIPE; + pos = &filp->f_pos; + + mutex_lock(&lm_file->lock); + if (*pos > lm_file->size) { + done = -EINVAL; + goto out_unlock; + } + + len = min_t(size_t, lm_file->size - *pos, len); + while (len) { + size_t page_offset; + struct page *page; + size_t page_len; + u8 *from_buff; + int err; + + page_offset = (*pos) % PAGE_SIZE; + page = pds_vfio_get_file_page(lm_file, *pos - page_offset); + if (!page) { + if (done == 0) + done = -EINVAL; + goto out_unlock; + } + + page_len = min_t(size_t, len, PAGE_SIZE - page_offset); + from_buff = kmap_local_page(page); + err = copy_to_user(buf, from_buff + page_offset, page_len); + kunmap_local(from_buff); + if (err) { + done = -EFAULT; + goto out_unlock; + } + *pos += page_len; + len -= page_len; + done += page_len; + buf += page_len; + } + +out_unlock: + mutex_unlock(&lm_file->lock); + return done; +} + +static const struct file_operations pds_vfio_save_fops = { + .owner = THIS_MODULE, + .read = pds_vfio_save_read, + .release = pds_vfio_release_file, + .llseek = no_llseek, +}; + +static int pds_vfio_get_save_file(struct pds_vfio_pci_device *pds_vfio) +{ + struct device *dev = &pds_vfio->vfio_coredev.pdev->dev; + struct pds_vfio_lm_file *lm_file; + int err; + u64 size; + + /* Get live migration state size in this state */ + err = pds_vfio_get_lm_status_cmd(pds_vfio, &size); + if (err) { + dev_err(dev, "failed to get save status: %pe\n", ERR_PTR(err)); + return err; + } + + dev_dbg(dev, "save status, size = %lld\n", size); + + if (!size) { + dev_err(dev, "invalid state size\n"); + return -EIO; + } + + lm_file = pds_vfio_get_lm_file(&pds_vfio_save_fops, O_RDONLY, size); + if (!lm_file) { + dev_err(dev, "failed to create save file\n"); + return -ENOENT; + } + + dev_dbg(dev, "size = %lld, alloc_size = %lld, npages = %lld\n", + lm_file->size, lm_file->alloc_size, lm_file->npages); + + pds_vfio->save_file = lm_file; + + return 0; +} + +static ssize_t pds_vfio_restore_write(struct file *filp, const char __user *buf, + size_t len, loff_t *pos) +{ + struct pds_vfio_lm_file *lm_file = filp->private_data; + loff_t requested_length; + ssize_t done = 0; + + if (pos) + return -ESPIPE; + + pos = &filp->f_pos; + + if (*pos < 0 || + check_add_overflow((loff_t)len, *pos, &requested_length)) + return -EINVAL; + + mutex_lock(&lm_file->lock); + + while (len) { + size_t page_offset; + struct page *page; + size_t page_len; + u8 *to_buff; + int err; + + page_offset = (*pos) % PAGE_SIZE; + page = pds_vfio_get_file_page(lm_file, *pos - page_offset); + if (!page) { + if (done == 0) + done = -EINVAL; + goto out_unlock; + } + + page_len = min_t(size_t, len, PAGE_SIZE - page_offset); + to_buff = kmap_local_page(page); + err = copy_from_user(to_buff + page_offset, buf, page_len); + kunmap_local(to_buff); + if (err) { + done = -EFAULT; + goto out_unlock; + } + *pos += page_len; + len -= page_len; + done += page_len; + buf += page_len; + lm_file->size += page_len; + } +out_unlock: + mutex_unlock(&lm_file->lock); + return done; +} + +static const struct file_operations pds_vfio_restore_fops = { + .owner = THIS_MODULE, + .write = pds_vfio_restore_write, + .release = pds_vfio_release_file, + .llseek = no_llseek, +}; + +static int pds_vfio_get_restore_file(struct pds_vfio_pci_device *pds_vfio) +{ + struct device *dev = &pds_vfio->vfio_coredev.pdev->dev; + struct pds_vfio_lm_file *lm_file; + u64 size; + + size = sizeof(union pds_lm_dev_state); + dev_dbg(dev, "restore status, size = %lld\n", size); + + if (!size) { + dev_err(dev, "invalid state size"); + return -EIO; + } + + lm_file = pds_vfio_get_lm_file(&pds_vfio_restore_fops, O_WRONLY, size); + if (!lm_file) { + dev_err(dev, "failed to create restore file"); + return -ENOENT; + } + pds_vfio->restore_file = lm_file; + + return 0; +} + +struct file * +pds_vfio_step_device_state_locked(struct pds_vfio_pci_device *pds_vfio, + enum vfio_device_mig_state next) +{ + enum vfio_device_mig_state cur = pds_vfio->state; + int err; + + if (cur == VFIO_DEVICE_STATE_STOP && next == VFIO_DEVICE_STATE_STOP_COPY) { + err = pds_vfio_get_save_file(pds_vfio); + if (err) + return ERR_PTR(err); + + err = pds_vfio_get_lm_state_cmd(pds_vfio); + if (err) { + pds_vfio_put_save_file(pds_vfio); + return ERR_PTR(err); + } + + return pds_vfio->save_file->filep; + } + + if (cur == VFIO_DEVICE_STATE_STOP_COPY && next == VFIO_DEVICE_STATE_STOP) { + pds_vfio_put_save_file(pds_vfio); + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, PDS_LM_STA_NONE); + return NULL; + } + + if (cur == VFIO_DEVICE_STATE_STOP && next == VFIO_DEVICE_STATE_RESUMING) { + err = pds_vfio_get_restore_file(pds_vfio); + if (err) + return ERR_PTR(err); + + return pds_vfio->restore_file->filep; + } + + if (cur == VFIO_DEVICE_STATE_RESUMING && next == VFIO_DEVICE_STATE_STOP) { + err = pds_vfio_set_lm_state_cmd(pds_vfio); + if (err) + return ERR_PTR(err); + + pds_vfio_put_restore_file(pds_vfio); + return NULL; + } + + if (cur == VFIO_DEVICE_STATE_RUNNING && next == VFIO_DEVICE_STATE_RUNNING_P2P) { + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, + PDS_LM_STA_IN_PROGRESS); + err = pds_vfio_suspend_device_cmd(pds_vfio); + if (err) + return ERR_PTR(err); + + return NULL; + } + + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == VFIO_DEVICE_STATE_RUNNING) { + err = pds_vfio_resume_device_cmd(pds_vfio); + if (err) + return ERR_PTR(err); + + pds_vfio_send_host_vf_lm_status_cmd(pds_vfio, PDS_LM_STA_NONE); + return NULL; + } + + if (cur == VFIO_DEVICE_STATE_STOP && next == VFIO_DEVICE_STATE_RUNNING_P2P) + return NULL; + + if (cur == VFIO_DEVICE_STATE_RUNNING_P2P && next == VFIO_DEVICE_STATE_STOP) + return NULL; + + return ERR_PTR(-EINVAL); +} diff --git a/drivers/vfio/pci/pds/lm.h b/drivers/vfio/pci/pds/lm.h new file mode 100644 index 000000000000..13be893198b7 --- /dev/null +++ b/drivers/vfio/pci/pds/lm.h @@ -0,0 +1,41 @@ +/* SPDX-License-Identifier: GPL-2.0 */ +/* Copyright(c) 2023 Advanced Micro Devices, Inc. */ + +#ifndef _LM_H_ +#define _LM_H_ + +#include <linux/fs.h> +#include <linux/mutex.h> +#include <linux/scatterlist.h> +#include <linux/types.h> + +#include <linux/pds/pds_common.h> +#include <linux/pds/pds_adminq.h> + +struct pds_vfio_lm_file { + struct file *filep; + struct mutex lock; /* protect live migration data file */ + u64 size; /* Size with valid data */ + u64 alloc_size; /* Total allocated size. Always >= len */ + void *page_mem; /* memory allocated for pages */ + struct page **pages; /* Backing pages for file */ + unsigned long long npages; + struct sg_table sg_table; /* SG table for backing pages */ + struct pds_lm_sg_elem *sgl; /* DMA mapping */ + dma_addr_t sgl_addr; + u16 num_sge; + struct scatterlist *last_offset_sg; /* Iterator */ + unsigned int sg_last_entry; + unsigned long last_offset; +}; + +struct pds_vfio_pci_device; + +struct file * +pds_vfio_step_device_state_locked(struct pds_vfio_pci_device *pds_vfio, + enum vfio_device_mig_state next); + +void pds_vfio_put_save_file(struct pds_vfio_pci_device *pds_vfio); +void pds_vfio_put_restore_file(struct pds_vfio_pci_device *pds_vfio); + +#endif /* _LM_H_ */ diff --git a/drivers/vfio/pci/pds/pci_drv.c b/drivers/vfio/pci/pds/pci_drv.c index a49420aa9736..ffd47fa8ede3 100644 --- a/drivers/vfio/pci/pds/pci_drv.c +++ b/drivers/vfio/pci/pds/pci_drv.c @@ -73,11 +73,24 @@ pds_vfio_pci_table[] = { }; MODULE_DEVICE_TABLE(pci, pds_vfio_pci_table); +static void pds_vfio_pci_aer_reset_done(struct pci_dev *pdev) +{ + struct pds_vfio_pci_device *pds_vfio = pds_vfio_pci_drvdata(pdev); + + pds_vfio_reset(pds_vfio); +} + +static const struct pci_error_handlers pds_vfio_pci_err_handlers = { + .reset_done = pds_vfio_pci_aer_reset_done, + .error_detected = vfio_pci_core_aer_err_detected, +}; + static struct pci_driver pds_vfio_pci_driver = { .name = KBUILD_MODNAME, .id_table = pds_vfio_pci_table, .probe = pds_vfio_pci_probe, .remove = pds_vfio_pci_remove, + .err_handler = &pds_vfio_pci_err_handlers, .driver_managed_dma = true, }; diff --git a/drivers/vfio/pci/pds/vfio_dev.c b/drivers/vfio/pci/pds/vfio_dev.c index 39771265b78f..2435d8255366 100644 --- a/drivers/vfio/pci/pds/vfio_dev.c +++ b/drivers/vfio/pci/pds/vfio_dev.c @@ -4,6 +4,7 @@ #include <linux/vfio.h> #include <linux/vfio_pci_core.h> +#include "lm.h" #include "vfio_dev.h" struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio) @@ -11,6 +12,11 @@ struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio) return pds_vfio->vfio_coredev.pdev; } +struct device *pds_vfio_to_dev(struct pds_vfio_pci_device *pds_vfio) +{ + return &pds_vfio_to_pci_dev(pds_vfio)->dev; +} + struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev) { struct vfio_pci_core_device *core_device = dev_get_drvdata(&pdev->dev); @@ -19,6 +25,98 @@ struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev) vfio_coredev); } +static void pds_vfio_state_mutex_unlock(struct pds_vfio_pci_device *pds_vfio) +{ +again: + spin_lock(&pds_vfio->reset_lock); + if (pds_vfio->deferred_reset) { + pds_vfio->deferred_reset = false; + if (pds_vfio->state == VFIO_DEVICE_STATE_ERROR) { + pds_vfio->state = VFIO_DEVICE_STATE_RUNNING; + pds_vfio_put_restore_file(pds_vfio); + pds_vfio_put_save_file(pds_vfio); + } + spin_unlock(&pds_vfio->reset_lock); + goto again; + } + mutex_unlock(&pds_vfio->state_mutex); + spin_unlock(&pds_vfio->reset_lock); +} + +void pds_vfio_reset(struct pds_vfio_pci_device *pds_vfio) +{ + spin_lock(&pds_vfio->reset_lock); + pds_vfio->deferred_reset = true; + if (!mutex_trylock(&pds_vfio->state_mutex)) { + spin_unlock(&pds_vfio->reset_lock); + return; + } + spin_unlock(&pds_vfio->reset_lock); + pds_vfio_state_mutex_unlock(pds_vfio); +} + +static struct file * +pds_vfio_set_device_state(struct vfio_device *vdev, + enum vfio_device_mig_state new_state) +{ + struct pds_vfio_pci_device *pds_vfio = + container_of(vdev, struct pds_vfio_pci_device, + vfio_coredev.vdev); + struct file *res = NULL; + + mutex_lock(&pds_vfio->state_mutex); + while (new_state != pds_vfio->state) { + enum vfio_device_mig_state next_state; + + int err = vfio_mig_get_next_state(vdev, pds_vfio->state, + new_state, &next_state); + if (err) { + res = ERR_PTR(err); + break; + } + + res = pds_vfio_step_device_state_locked(pds_vfio, next_state); + if (IS_ERR(res)) + break; + + pds_vfio->state = next_state; + + if (WARN_ON(res && new_state != pds_vfio->state)) { + res = ERR_PTR(-EINVAL); + break; + } + } + pds_vfio_state_mutex_unlock(pds_vfio); + + return res; +} + +static int pds_vfio_get_device_state(struct vfio_device *vdev, + enum vfio_device_mig_state *current_state) +{ + struct pds_vfio_pci_device *pds_vfio = + container_of(vdev, struct pds_vfio_pci_device, + vfio_coredev.vdev); + + mutex_lock(&pds_vfio->state_mutex); + *current_state = pds_vfio->state; + pds_vfio_state_mutex_unlock(pds_vfio); + return 0; +} + +static int pds_vfio_get_device_state_size(struct vfio_device *vdev, + unsigned long *stop_copy_length) +{ + *stop_copy_length = PDS_LM_DEVICE_STATE_LENGTH; + return 0; +} + +static const struct vfio_migration_ops pds_vfio_lm_ops = { + .migration_set_state = pds_vfio_set_device_state, + .migration_get_state = pds_vfio_get_device_state, + .migration_get_data_size = pds_vfio_get_device_state_size +}; + static int pds_vfio_init_device(struct vfio_device *vdev) { struct pds_vfio_pci_device *pds_vfio = @@ -34,6 +132,9 @@ static int pds_vfio_init_device(struct vfio_device *vdev) pds_vfio->vf_id = pci_iov_vf_id(pdev); pds_vfio->pci_id = PCI_DEVID(pdev->bus->number, pdev->devfn); + vdev->migration_flags = VFIO_MIGRATION_STOP_COPY | VFIO_MIGRATION_P2P; + vdev->mig_ops = &pds_vfio_lm_ops; + dev_dbg(&pdev->dev, "%s: PF %#04x VF %#04x (%d) vf_id %d domain %d pds_vfio %p\n", __func__, pci_dev_id(pdev->physfn), pds_vfio->pci_id, @@ -54,17 +155,34 @@ static int pds_vfio_open_device(struct vfio_device *vdev) if (err) return err; + mutex_init(&pds_vfio->state_mutex); + pds_vfio->state = VFIO_DEVICE_STATE_RUNNING; + vfio_pci_core_finish_enable(&pds_vfio->vfio_coredev); return 0; } +static void pds_vfio_close_device(struct vfio_device *vdev) +{ + struct pds_vfio_pci_device *pds_vfio = + container_of(vdev, struct pds_vfio_pci_device, + vfio_coredev.vdev); + + mutex_lock(&pds_vfio->state_mutex); + pds_vfio_put_restore_file(pds_vfio); + pds_vfio_put_save_file(pds_vfio); + mutex_unlock(&pds_vfio->state_mutex); + mutex_destroy(&pds_vfio->state_mutex); + vfio_pci_core_close_device(vdev); +} + static const struct vfio_device_ops pds_vfio_ops = { .name = "pds-vfio", .init = pds_vfio_init_device, .release = vfio_pci_core_release_dev, .open_device = pds_vfio_open_device, - .close_device = vfio_pci_core_close_device, + .close_device = pds_vfio_close_device, .ioctl = vfio_pci_core_ioctl, .device_feature = vfio_pci_core_ioctl_feature, .read = vfio_pci_core_read, diff --git a/drivers/vfio/pci/pds/vfio_dev.h b/drivers/vfio/pci/pds/vfio_dev.h index 92e8ff241ca8..df6208a7140b 100644 --- a/drivers/vfio/pci/pds/vfio_dev.h +++ b/drivers/vfio/pci/pds/vfio_dev.h @@ -7,12 +7,21 @@ #include <linux/pci.h> #include <linux/vfio_pci_core.h> +#include "lm.h" + struct pdsc; struct pds_vfio_pci_device { struct vfio_pci_core_device vfio_coredev; struct pdsc *pdsc; + struct pds_vfio_lm_file *save_file; + struct pds_vfio_lm_file *restore_file; + struct mutex state_mutex; /* protect migration state */ + enum vfio_device_mig_state state; + spinlock_t reset_lock; /* protect reset_done flow */ + u8 deferred_reset; + int vf_id; int pci_id; u16 client_id; @@ -20,7 +29,9 @@ struct pds_vfio_pci_device { const struct vfio_device_ops *pds_vfio_ops_info(void); struct pds_vfio_pci_device *pds_vfio_pci_drvdata(struct pci_dev *pdev); +void pds_vfio_reset(struct pds_vfio_pci_device *pds_vfio); struct pci_dev *pds_vfio_to_pci_dev(struct pds_vfio_pci_device *pds_vfio); +struct device *pds_vfio_to_dev(struct pds_vfio_pci_device *pds_vfio); #endif /* _VFIO_DEV_H_ */ diff --git a/include/linux/pds/pds_adminq.h b/include/linux/pds/pds_adminq.h index 98a60ce87b92..db6de081f15f 100644 --- a/include/linux/pds/pds_adminq.h +++ b/include/linux/pds/pds_adminq.h @@ -584,6 +584,213 @@ struct pds_core_q_init_comp { u8 color; }; +#define PDS_LM_DEVICE_STATE_LENGTH 65536 +#define PDS_LM_CHECK_DEVICE_STATE_LENGTH(X) \ + PDS_CORE_SIZE_CHECK(union, PDS_LM_DEVICE_STATE_LENGTH, X) + +/* + * enum pds_lm_cmd_opcode - Live Migration Device commands + */ +enum pds_lm_cmd_opcode { + PDS_LM_CMD_HOST_VF_STATUS = 1, + + /* Device state commands */ + PDS_LM_CMD_STATUS = 16, + PDS_LM_CMD_SUSPEND = 18, + PDS_LM_CMD_SUSPEND_STATUS = 19, + PDS_LM_CMD_RESUME = 20, + PDS_LM_CMD_SAVE = 21, + PDS_LM_CMD_RESTORE = 22, +}; + +/** + * struct pds_lm_cmd - generic command + * @opcode: Opcode + * @rsvd: Word boundary padding + * @vf_id: VF id + * @rsvd2: Structure padding to 60 Bytes + */ +struct pds_lm_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; + u8 rsvd2[56]; +}; + +/** + * struct pds_lm_comp - generic command completion + * @status: Status of the command (enum pds_core_status_code) + * @rsvd: Structure padding to 16 Bytes + */ +struct pds_lm_comp { + u8 status; + u8 rsvd[15]; +}; + +/** + * struct pds_lm_status_cmd - STATUS command + * @opcode: Opcode + * @rsvd: Word boundary padding + * @vf_id: VF id + */ +struct pds_lm_status_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; +}; + +/** + * struct pds_lm_status_comp - STATUS command completion + * @status: Status of the command (enum pds_core_status_code) + * @rsvd: Word boundary padding + * @comp_index: Index in the desc ring for which this is the completion + * @size: Size of the device state + * @rsvd2: Word boundary padding + * @color: Color bit + */ +struct pds_lm_status_comp { + u8 status; + u8 rsvd; + __le16 comp_index; + union { + __le64 size; + u8 rsvd2[11]; + } __packed; + u8 color; +}; + +/** + * struct pds_lm_suspend_cmd - SUSPEND command + * @opcode: Opcode PDS_LM_CMD_SUSPEND + * @rsvd: Word boundary padding + * @vf_id: VF id + */ +struct pds_lm_suspend_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; +}; + +/** + * struct pds_lm_suspend_comp - SUSPEND command completion + * @status: Status of the command (enum pds_core_status_code) + * @rsvd: Word boundary padding + * @comp_index: Index in the desc ring for which this is the completion + * @state_size: Size of the device state computed post suspend + * @rsvd2: Word boundary padding + * @color: Color bit + */ +struct pds_lm_suspend_comp { + u8 status; + u8 rsvd; + __le16 comp_index; + union { + __le64 state_size; + u8 rsvd2[11]; + } __packed; + u8 color; +}; + +/** + * struct pds_lm_suspend_status_cmd - SUSPEND status command + * @opcode: Opcode PDS_AQ_CMD_LM_SUSPEND_STATUS + * @rsvd: Word boundary padding + * @vf_id: VF id + */ +struct pds_lm_suspend_status_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; +}; + +/** + * struct pds_lm_resume_cmd - RESUME command + * @opcode: Opcode PDS_LM_CMD_RESUME + * @rsvd: Word boundary padding + * @vf_id: VF id + */ +struct pds_lm_resume_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; +}; + +/** + * struct pds_lm_sg_elem - Transmit scatter-gather (SG) descriptor element + * @addr: DMA address of SG element data buffer + * @len: Length of SG element data buffer, in bytes + * @rsvd: Word boundary padding + */ +struct pds_lm_sg_elem { + __le64 addr; + __le32 len; + __le16 rsvd[2]; +}; + +/** + * struct pds_lm_save_cmd - SAVE command + * @opcode: Opcode PDS_LM_CMD_SAVE + * @rsvd: Word boundary padding + * @vf_id: VF id + * @rsvd2: Word boundary padding + * @sgl_addr: IOVA address of the SGL to dma the device state + * @num_sge: Total number of SG elements + */ +struct pds_lm_save_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; + u8 rsvd2[4]; + __le64 sgl_addr; + __le32 num_sge; +} __packed; + +/** + * struct pds_lm_restore_cmd - RESTORE command + * @opcode: Opcode PDS_LM_CMD_RESTORE + * @rsvd: Word boundary padding + * @vf_id: VF id + * @rsvd2: Word boundary padding + * @sgl_addr: IOVA address of the SGL to dma the device state + * @num_sge: Total number of SG elements + */ +struct pds_lm_restore_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; + u8 rsvd2[4]; + __le64 sgl_addr; + __le32 num_sge; +} __packed; + +/** + * union pds_lm_dev_state - device state information + * @words: Device state words + */ +union pds_lm_dev_state { + __le32 words[PDS_LM_DEVICE_STATE_LENGTH / sizeof(__le32)]; +}; + +enum pds_lm_host_vf_status { + PDS_LM_STA_NONE = 0, + PDS_LM_STA_IN_PROGRESS, + PDS_LM_STA_MAX, +}; + +/** + * struct pds_lm_host_vf_status_cmd - HOST_VF_STATUS command + * @opcode: Opcode PDS_LM_CMD_HOST_VF_STATUS + * @rsvd: Word boundary padding + * @vf_id: VF id + * @status: Current LM status of host VF driver (enum pds_lm_host_status) + */ +struct pds_lm_host_vf_status_cmd { + u8 opcode; + u8 rsvd; + __le16 vf_id; + u8 status; +}; + union pds_core_adminq_cmd { u8 opcode; u8 bytes[64]; @@ -600,6 +807,14 @@ union pds_core_adminq_cmd { struct pds_core_q_identify_cmd q_ident; struct pds_core_q_init_cmd q_init; + + struct pds_lm_suspend_cmd lm_suspend; + struct pds_lm_suspend_status_cmd lm_suspend_status; + struct pds_lm_resume_cmd lm_resume; + struct pds_lm_status_cmd lm_status; + struct pds_lm_save_cmd lm_save; + struct pds_lm_restore_cmd lm_restore; + struct pds_lm_host_vf_status_cmd lm_host_vf_status; }; union pds_core_adminq_comp { @@ -621,6 +836,8 @@ union pds_core_adminq_comp { struct pds_core_q_identify_comp q_ident; struct pds_core_q_init_comp q_init; + + struct pds_lm_status_comp lm_status; }; #ifndef __CHECKER__