@@ -942,6 +942,7 @@ static struct virtqueue *vu_setup_vq(struct virtio_device *vdev,
{
struct virtio_uml_device *vu_dev = to_virtio_uml_device(vdev);
struct platform_device *pdev = vu_dev->pdev;
+ struct vq_transport_config tp_cfg = {};
struct virtio_uml_vq_info *info;
struct virtqueue *vq;
int num = MAX_SUPPORTED_QUEUE_SIZE;
@@ -955,10 +956,15 @@ static struct virtqueue *vu_setup_vq(struct virtio_device *vdev,
snprintf(info->name, sizeof(info->name), "%s.%d-%s", pdev->name,
pdev->id, cfg->names[cfg->cfg_idx]);
- vq = vring_create_virtqueue(index, num, PAGE_SIZE, vdev, true, true,
- cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false,
- vu_notify,
- cfg->callbacks[cfg->cfg_idx], info->name);
+ tp_cfg.num = num;
+ tp_cfg.vring_align = PAGE_SIZE;
+ tp_cfg.weak_barriers = true;
+ tp_cfg.may_reduce_num = true;
+ tp_cfg.notify = vu_notify;
+
+ cfg->names[cfg->cfg_idx] = info->name;
+
+ vq = vring_create_virtqueue(vdev, index, &tp_cfg, cfg);
if (!vq) {
rc = -ENOMEM;
goto error_create;
@@ -503,6 +503,7 @@ static struct virtqueue *virtio_ccw_setup_vq(struct virtio_device *vdev,
struct virtio_vq_config *cfg)
{
struct virtio_ccw_device *vcdev = to_vc_device(vdev);
+ struct vq_transport_config tp_cfg = {};
bool (*notify)(struct virtqueue *vq);
int err;
struct virtqueue *vq = NULL;
@@ -536,13 +537,14 @@ static struct virtqueue *virtio_ccw_setup_vq(struct virtio_device *vdev,
goto out_err;
}
may_reduce = vcdev->revision > 0;
- vq = vring_create_virtqueue(i, info->num, KVM_VIRTIO_CCW_RING_ALIGN,
- vdev, true, may_reduce,
- cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false,
- notify,
- cfg->callbacks[cfg->cfg_idx],
- cfg->names[cfg->cfg_idx]);
+ tp_cfg.num = info->num;
+ tp_cfg.vring_align = KVM_VIRTIO_CCW_RING_ALIGN;
+ tp_cfg.weak_barriers = true;
+ tp_cfg.may_reduce_num = may_reduce;
+ tp_cfg.notify = notify;
+
+ vq = vring_create_virtqueue(vdev, i, &tp_cfg, cfg);
if (!vq) {
/* For now, we fail if we can't get the requested size. */
dev_warn(&vcdev->cdev->dev, "no vq\n");
@@ -373,6 +373,7 @@ static struct virtqueue *vm_setup_vq(struct virtio_device *vdev, unsigned int in
struct virtio_vq_config *cfg)
{
struct virtio_mmio_device *vm_dev = to_virtio_mmio_device(vdev);
+ struct vq_transport_config tp_cfg = {};
bool (*notify)(struct virtqueue *vq);
struct virtio_mmio_vq_info *info;
struct virtqueue *vq;
@@ -411,13 +412,14 @@ static struct virtqueue *vm_setup_vq(struct virtio_device *vdev, unsigned int in
goto error_new_virtqueue;
}
+ tp_cfg.num = num;
+ tp_cfg.vring_align = VIRTIO_MMIO_VRING_ALIGN;
+ tp_cfg.weak_barriers = true;
+ tp_cfg.may_reduce_num = true;
+ tp_cfg.notify = notify;
+
/* Create the vring */
- vq = vring_create_virtqueue(index, num, VIRTIO_MMIO_VRING_ALIGN, vdev,
- true, true,
- cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false,
- notify,
- cfg->callbacks[cfg->cfg_idx],
- cfg->names[cfg->cfg_idx]);
+ vq = vring_create_virtqueue(vdev, index, &tp_cfg, cfg);
if (!vq) {
err = -ENOMEM;
goto error_new_virtqueue;
@@ -113,6 +113,7 @@ static struct virtqueue *setup_vq(struct virtio_pci_device *vp_dev,
struct virtio_vq_config *cfg,
u16 msix_vec)
{
+ struct vq_transport_config tp_cfg = {};
struct virtqueue *vq;
u16 num;
int err;
@@ -125,14 +126,14 @@ static struct virtqueue *setup_vq(struct virtio_pci_device *vp_dev,
info->msix_vector = msix_vec;
+ tp_cfg.num = num;
+ tp_cfg.vring_align = VIRTIO_PCI_VRING_ALIGN;
+ tp_cfg.weak_barriers = true;
+ tp_cfg.may_reduce_num = false;
+ tp_cfg.notify = vp_notify;
+
/* create the vring */
- vq = vring_create_virtqueue(index, num,
- VIRTIO_PCI_VRING_ALIGN, &vp_dev->vdev,
- true, false,
- cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false,
- vp_notify,
- cfg->callbacks[cfg->cfg_idx],
- cfg->names[cfg->cfg_idx]);
+ vq = vring_create_virtqueue(&vp_dev->vdev, index, &tp_cfg, cfg);
if (!vq)
return ERR_PTR(-ENOMEM);
@@ -535,6 +535,7 @@ static struct virtqueue *setup_vq(struct virtio_pci_device *vp_dev,
{
struct virtio_pci_modern_device *mdev = &vp_dev->mdev;
+ struct vq_transport_config tp_cfg = {};
bool (*notify)(struct virtqueue *vq);
struct virtqueue *vq;
bool is_avq;
@@ -558,14 +559,14 @@ static struct virtqueue *setup_vq(struct virtio_pci_device *vp_dev,
info->msix_vector = msix_vec;
+ tp_cfg.num = num;
+ tp_cfg.vring_align = SMP_CACHE_BYTES;
+ tp_cfg.weak_barriers = true;
+ tp_cfg.may_reduce_num = true;
+ tp_cfg.notify = notify;
+
/* create the vring */
- vq = vring_create_virtqueue(index, num,
- SMP_CACHE_BYTES, &vp_dev->vdev,
- true, true,
- cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false,
- notify,
- cfg->callbacks[cfg->cfg_idx],
- cfg->names[cfg->cfg_idx]);
+ vq = vring_create_virtqueue(&vp_dev->vdev, index, &tp_cfg, cfg);
if (!vq)
return ERR_PTR(-ENOMEM);
@@ -2662,43 +2662,32 @@ static struct virtqueue *__vring_new_virtqueue(unsigned int index,
return &vq->vq;
}
-struct virtqueue *vring_create_virtqueue(
- unsigned int index,
- unsigned int num,
- unsigned int vring_align,
- struct virtio_device *vdev,
- bool weak_barriers,
- bool may_reduce_num,
- bool context,
- bool (*notify)(struct virtqueue *),
- void (*callback)(struct virtqueue *),
- const char *name)
+struct virtqueue *vring_create_virtqueue(struct virtio_device *vdev,
+ unsigned int index,
+ struct vq_transport_config *tp_cfg,
+ struct virtio_vq_config *cfg)
{
+ struct device *dma_dev;
+ unsigned int num;
+ unsigned int vring_align;
+ bool weak_barriers;
+ bool may_reduce_num;
+ bool context;
+ bool (*notify)(struct virtqueue *_);
+ void (*callback)(struct virtqueue *_);
+ const char *name;
- if (virtio_has_feature(vdev, VIRTIO_F_RING_PACKED))
- return vring_create_virtqueue_packed(index, num, vring_align,
- vdev, weak_barriers, may_reduce_num,
- context, notify, callback, name, vdev->dev.parent);
+ dma_dev = tp_cfg->dma_dev ? : vdev->dev.parent;
- return vring_create_virtqueue_split(index, num, vring_align,
- vdev, weak_barriers, may_reduce_num,
- context, notify, callback, name, vdev->dev.parent);
-}
-EXPORT_SYMBOL_GPL(vring_create_virtqueue);
+ num = tp_cfg->num;
+ vring_align = tp_cfg->vring_align;
+ weak_barriers = tp_cfg->weak_barriers;
+ may_reduce_num = tp_cfg->may_reduce_num;
+ notify = tp_cfg->notify;
-struct virtqueue *vring_create_virtqueue_dma(
- unsigned int index,
- unsigned int num,
- unsigned int vring_align,
- struct virtio_device *vdev,
- bool weak_barriers,
- bool may_reduce_num,
- bool context,
- bool (*notify)(struct virtqueue *),
- void (*callback)(struct virtqueue *),
- const char *name,
- struct device *dma_dev)
-{
+ name = cfg->names[cfg->cfg_idx];
+ callback = cfg->callbacks[cfg->cfg_idx];
+ context = cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false;
if (virtio_has_feature(vdev, VIRTIO_F_RING_PACKED))
return vring_create_virtqueue_packed(index, num, vring_align,
@@ -2709,7 +2698,7 @@ struct virtqueue *vring_create_virtqueue_dma(
vdev, weak_barriers, may_reduce_num,
context, notify, callback, name, dma_dev);
}
-EXPORT_SYMBOL_GPL(vring_create_virtqueue_dma);
+EXPORT_SYMBOL_GPL(vring_create_virtqueue);
/**
* virtqueue_resize - resize the vring of vq
@@ -146,8 +146,8 @@ virtio_vdpa_setup_vq(struct virtio_device *vdev, unsigned int index,
{
struct virtio_vdpa_device *vd_dev = to_virtio_vdpa_device(vdev);
struct vdpa_device *vdpa = vd_get_vdpa(vdev);
- struct device *dma_dev;
const struct vdpa_config_ops *ops = vdpa->config;
+ struct vq_transport_config tp_cfg = {};
struct virtio_vdpa_vq_info *info;
bool (*notify)(struct virtqueue *vq) = virtio_vdpa_notify;
struct vdpa_callback cb;
@@ -201,16 +201,17 @@ virtio_vdpa_setup_vq(struct virtio_device *vdev, unsigned int index,
align = ops->get_vq_align(vdpa);
if (ops->get_vq_dma_dev)
- dma_dev = ops->get_vq_dma_dev(vdpa, index);
+ tp_cfg.dma_dev = ops->get_vq_dma_dev(vdpa, index);
else
- dma_dev = vdpa_get_dma_dev(vdpa);
- vq = vring_create_virtqueue_dma(index, max_num, align, vdev,
- true, may_reduce_num,
- cfg->ctx ? cfg->ctx[cfg->cfg_idx] : false,
- notify,
- cfg->callbacks[cfg->cfg_idx],
- cfg->names[cfg->cfg_idx],
- dma_dev);
+ tp_cfg.dma_dev = vdpa_get_dma_dev(vdpa);
+
+ tp_cfg.num = max_num;
+ tp_cfg.vring_align = align;
+ tp_cfg.weak_barriers = true;
+ tp_cfg.may_reduce_num = may_reduce_num;
+ tp_cfg.notify = notify;
+
+ vq = vring_create_virtqueue(vdev, index, &tp_cfg, cfg);
if (!vq) {
err = -ENOMEM;
goto error_new_virtqueue;
@@ -105,7 +105,7 @@ struct virtio_vq_config {
struct virtqueue **vqs;
vq_callback_t **callbacks;
- const char *const *names;
+ const char **names;
const bool *ctx;
struct irq_affinity *desc;
};
@@ -252,7 +252,7 @@ int virtio_find_vqs(struct virtio_device *vdev, unsigned nvqs,
cfg.nvqs = nvqs;
cfg.vqs = vqs;
cfg.callbacks = callbacks;
- cfg.names = names;
+ cfg.names = (const char **)names;
cfg.desc = desc;
return vdev->config->find_vqs(vdev, &cfg);
@@ -269,7 +269,7 @@ int virtio_find_vqs_ctx(struct virtio_device *vdev, unsigned nvqs,
cfg.nvqs = nvqs;
cfg.vqs = vqs;
cfg.callbacks = callbacks;
- cfg.names = names;
+ cfg.names = (const char **)names;
cfg.ctx = ctx;
cfg.desc = desc;
@@ -5,6 +5,7 @@
#include <asm/barrier.h>
#include <linux/irqreturn.h>
#include <uapi/linux/virtio_ring.h>
+#include <linux/virtio_config.h>
/*
* Barriers in virtio are tricky. Non-SMP virtio guests can't assume
@@ -60,38 +61,25 @@ struct virtio_device;
struct virtqueue;
struct device;
+struct vq_transport_config {
+ unsigned int num;
+ unsigned int vring_align;
+ bool weak_barriers;
+ bool may_reduce_num;
+ bool (*notify)(struct virtqueue *vq);
+ struct device *dma_dev;
+};
+
/*
* Creates a virtqueue and allocates the descriptor ring. If
* may_reduce_num is set, then this may allocate a smaller ring than
* expected. The caller should query virtqueue_get_vring_size to learn
* the actual size of the ring.
*/
-struct virtqueue *vring_create_virtqueue(unsigned int index,
- unsigned int num,
- unsigned int vring_align,
- struct virtio_device *vdev,
- bool weak_barriers,
- bool may_reduce_num,
- bool ctx,
- bool (*notify)(struct virtqueue *vq),
- void (*callback)(struct virtqueue *vq),
- const char *name);
-
-/*
- * Creates a virtqueue and allocates the descriptor ring with per
- * virtqueue DMA device.
- */
-struct virtqueue *vring_create_virtqueue_dma(unsigned int index,
- unsigned int num,
- unsigned int vring_align,
- struct virtio_device *vdev,
- bool weak_barriers,
- bool may_reduce_num,
- bool ctx,
- bool (*notify)(struct virtqueue *vq),
- void (*callback)(struct virtqueue *vq),
- const char *name,
- struct device *dma_dev);
+struct virtqueue *vring_create_virtqueue(struct virtio_device *vdev,
+ unsigned int index,
+ struct vq_transport_config *tp_cfg,
+ struct virtio_vq_config *cfg);
/*
* Creates a virtqueue with a standard layout but a caller-allocated
Now, we pass multi parameters to vring_create_virtqueue. These parameters may from transport or from driver. vring_create_virtqueue is called by many places. Every time, we try to add a new parameter, that is difficult. If parameters from the driver, that should directly be passed to vring. Then the vring can access the config from driver directly. If parameters from the transport, we squish the parameters to a structure. That will be helpful to add new parameter. Because the virtio_uml.c changes the name, so change the "names" inside the virtio_vq_config from "const char *const *names" to "const char **names". Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com> --- arch/um/drivers/virtio_uml.c | 14 +++++--- drivers/s390/virtio/virtio_ccw.c | 14 ++++---- drivers/virtio/virtio_mmio.c | 14 ++++---- drivers/virtio/virtio_pci_legacy.c | 15 ++++---- drivers/virtio/virtio_pci_modern.c | 15 ++++---- drivers/virtio/virtio_ring.c | 57 ++++++++++++------------------ drivers/virtio/virtio_vdpa.c | 21 +++++------ include/linux/virtio_config.h | 6 ++-- include/linux/virtio_ring.h | 40 ++++++++------------- 9 files changed, 93 insertions(+), 103 deletions(-)