Message ID | 20230725130709.58207-3-gavinl@nvidia.com (mailing list archive) |
---|---|
State | Superseded |
Delegated to: | Netdev Maintainers |
Headers | show |
Series | virtio_net: add per queue interrupt coalescing support | expand |
On Tue, 2023-07-25 at 16:07 +0300, Gavin Li wrote: > Add interrupt_coalesce config in send_queue and receive_queue to cache user > config. > > Send per virtqueue interrupt moderation config to underlying device in > order to have more efficient interrupt moderation and cpu utilization of > guest VM. > > Additionally, address all the VQs when updating the global configuration, > as now the individual VQs configuration can diverge from the global > configuration. > > Signed-off-by: Gavin Li <gavinl@nvidia.com> > Reviewed-by: Dragos Tatulea <dtatulea@nvidia.com> > Reviewed-by: Jiri Pirko <jiri@nvidia.com> > Acked-by: Michael S. Tsirkin <mst@redhat.com> FTR, this patch is significantly different from the version previously acked/reviewed, I'm unsure if all the reviewers are ok with the new one. [...] > static int virtnet_set_coalesce(struct net_device *dev, > struct ethtool_coalesce *ec, > struct kernel_ethtool_coalesce *kernel_coal, > struct netlink_ext_ack *extack) > { > struct virtnet_info *vi = netdev_priv(dev); > - int ret, i, napi_weight; > + int ret, queue_number, napi_weight; > bool update_napi = false; > > /* Can't change NAPI weight if the link is up */ > napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0; > - if (napi_weight ^ vi->sq[0].napi.weight) { > - if (dev->flags & IFF_UP) > - return -EBUSY; > - else > - update_napi = true; > + for (queue_number = 0; queue_number < vi->max_queue_pairs; queue_number++) { > + ret = virtnet_should_update_vq_weight(dev->flags, napi_weight, > + vi->sq[queue_number].napi.weight, > + &update_napi); > + if (ret) > + return ret; > + > + if (update_napi) { > + /* All queues that belong to [queue_number, queue_count] will be > + * updated for the sake of simplicity, which might not be necessary It looks like the comment above still refers to the old code. Should be: [queue_number, vi->max_queue_pairs] Otherwise LGTM, thanks! Paolo
On Thu, Jul 27, 2023 at 9:28 PM Paolo Abeni <pabeni@redhat.com> wrote: > > On Tue, 2023-07-25 at 16:07 +0300, Gavin Li wrote: > > Add interrupt_coalesce config in send_queue and receive_queue to cache user > > config. > > > > Send per virtqueue interrupt moderation config to underlying device in > > order to have more efficient interrupt moderation and cpu utilization of > > guest VM. > > > > Additionally, address all the VQs when updating the global configuration, > > as now the individual VQs configuration can diverge from the global > > configuration. > > > > Signed-off-by: Gavin Li <gavinl@nvidia.com> > > Reviewed-by: Dragos Tatulea <dtatulea@nvidia.com> > > Reviewed-by: Jiri Pirko <jiri@nvidia.com> > > Acked-by: Michael S. Tsirkin <mst@redhat.com> > > FTR, this patch is significantly different from the version previously > acked/reviewed, I'm unsure if all the reviewers are ok with the new > one. Good point, and I plan to review this no later than next Monday and offer my ack if necessary. Please hold this series now. Thanks > > [...] > > > static int virtnet_set_coalesce(struct net_device *dev, > > struct ethtool_coalesce *ec, > > struct kernel_ethtool_coalesce *kernel_coal, > > struct netlink_ext_ack *extack) > > { > > struct virtnet_info *vi = netdev_priv(dev); > > - int ret, i, napi_weight; > > + int ret, queue_number, napi_weight; > > bool update_napi = false; > > > > /* Can't change NAPI weight if the link is up */ > > napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0; > > - if (napi_weight ^ vi->sq[0].napi.weight) { > > - if (dev->flags & IFF_UP) > > - return -EBUSY; > > - else > > - update_napi = true; > > + for (queue_number = 0; queue_number < vi->max_queue_pairs; queue_number++) { > > + ret = virtnet_should_update_vq_weight(dev->flags, napi_weight, > > + vi->sq[queue_number].napi.weight, > > + &update_napi); > > + if (ret) > > + return ret; > > + > > + if (update_napi) { > > + /* All queues that belong to [queue_number, queue_count] will be > > + * updated for the sake of simplicity, which might not be necessary > > It looks like the comment above still refers to the old code. Should > be: > [queue_number, vi->max_queue_pairs] > > Otherwise LGTM, thanks! > > Paolo >
On Thu, Jul 27, 2023 at 03:28:32PM +0200, Paolo Abeni wrote: > On Tue, 2023-07-25 at 16:07 +0300, Gavin Li wrote: > > Add interrupt_coalesce config in send_queue and receive_queue to cache user > > config. > > > > Send per virtqueue interrupt moderation config to underlying device in > > order to have more efficient interrupt moderation and cpu utilization of > > guest VM. > > > > Additionally, address all the VQs when updating the global configuration, > > as now the individual VQs configuration can diverge from the global > > configuration. > > > > Signed-off-by: Gavin Li <gavinl@nvidia.com> > > Reviewed-by: Dragos Tatulea <dtatulea@nvidia.com> > > Reviewed-by: Jiri Pirko <jiri@nvidia.com> > > Acked-by: Michael S. Tsirkin <mst@redhat.com> > > FTR, this patch is significantly different from the version previously > acked/reviewed, I'm unsure if all the reviewers are ok with the new > one. > > [...] still ok by me Acked-by: Michael S. Tsirkin <mst@redhat.com> let's wait for Jason too. > > static int virtnet_set_coalesce(struct net_device *dev, > > struct ethtool_coalesce *ec, > > struct kernel_ethtool_coalesce *kernel_coal, > > struct netlink_ext_ack *extack) > > { > > struct virtnet_info *vi = netdev_priv(dev); > > - int ret, i, napi_weight; > > + int ret, queue_number, napi_weight; > > bool update_napi = false; > > > > /* Can't change NAPI weight if the link is up */ > > napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0; > > - if (napi_weight ^ vi->sq[0].napi.weight) { > > - if (dev->flags & IFF_UP) > > - return -EBUSY; > > - else > > - update_napi = true; > > + for (queue_number = 0; queue_number < vi->max_queue_pairs; queue_number++) { > > + ret = virtnet_should_update_vq_weight(dev->flags, napi_weight, > > + vi->sq[queue_number].napi.weight, > > + &update_napi); > > + if (ret) > > + return ret; > > + > > + if (update_napi) { > > + /* All queues that belong to [queue_number, queue_count] will be > > + * updated for the sake of simplicity, which might not be necessary > > It looks like the comment above still refers to the old code. Should > be: > [queue_number, vi->max_queue_pairs] > > Otherwise LGTM, thanks! > > Paolo
在 2023/7/25 21:07, Gavin Li 写道: > Add interrupt_coalesce config in send_queue and receive_queue to cache user > config. > > Send per virtqueue interrupt moderation config to underlying device in > order to have more efficient interrupt moderation and cpu utilization of > guest VM. > > Additionally, address all the VQs when updating the global configuration, > as now the individual VQs configuration can diverge from the global > configuration. > > Signed-off-by: Gavin Li <gavinl@nvidia.com> > Reviewed-by: Dragos Tatulea <dtatulea@nvidia.com> > Reviewed-by: Jiri Pirko <jiri@nvidia.com> > Acked-by: Michael S. Tsirkin <mst@redhat.com> > Reviewed-by: Heng Qi <hengqi@linux.alibaba.com> Acked-by: Jason Wang <jasowang@redhat.com> Thanks
在 2023/7/28 13:46, Michael S. Tsirkin 写道: > On Thu, Jul 27, 2023 at 03:28:32PM +0200, Paolo Abeni wrote: >> On Tue, 2023-07-25 at 16:07 +0300, Gavin Li wrote: >>> Add interrupt_coalesce config in send_queue and receive_queue to cache user >>> config. >>> >>> Send per virtqueue interrupt moderation config to underlying device in >>> order to have more efficient interrupt moderation and cpu utilization of >>> guest VM. >>> >>> Additionally, address all the VQs when updating the global configuration, >>> as now the individual VQs configuration can diverge from the global >>> configuration. >>> >>> Signed-off-by: Gavin Li <gavinl@nvidia.com> >>> Reviewed-by: Dragos Tatulea <dtatulea@nvidia.com> >>> Reviewed-by: Jiri Pirko <jiri@nvidia.com> >>> Acked-by: Michael S. Tsirkin <mst@redhat.com> >> FTR, this patch is significantly different from the version previously >> acked/reviewed, I'm unsure if all the reviewers are ok with the new >> one. >> >> [...] > still ok by me > > Acked-by: Michael S. Tsirkin <mst@redhat.com> > > let's wait for Jason too. I'm fine with this series (I've acked each patch). Thanks > >>> static int virtnet_set_coalesce(struct net_device *dev, >>> struct ethtool_coalesce *ec, >>> struct kernel_ethtool_coalesce *kernel_coal, >>> struct netlink_ext_ack *extack) >>> { >>> struct virtnet_info *vi = netdev_priv(dev); >>> - int ret, i, napi_weight; >>> + int ret, queue_number, napi_weight; >>> bool update_napi = false; >>> >>> /* Can't change NAPI weight if the link is up */ >>> napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0; >>> - if (napi_weight ^ vi->sq[0].napi.weight) { >>> - if (dev->flags & IFF_UP) >>> - return -EBUSY; >>> - else >>> - update_napi = true; >>> + for (queue_number = 0; queue_number < vi->max_queue_pairs; queue_number++) { >>> + ret = virtnet_should_update_vq_weight(dev->flags, napi_weight, >>> + vi->sq[queue_number].napi.weight, >>> + &update_napi); >>> + if (ret) >>> + return ret; >>> + >>> + if (update_napi) { >>> + /* All queues that belong to [queue_number, queue_count] will be >>> + * updated for the sake of simplicity, which might not be necessary >> It looks like the comment above still refers to the old code. Should >> be: >> [queue_number, vi->max_queue_pairs] >> >> Otherwise LGTM, thanks! >> >> Paolo
diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c index dd5fec073a27..c185930d7c9d 100644 --- a/drivers/net/virtio_net.c +++ b/drivers/net/virtio_net.c @@ -144,6 +144,8 @@ struct send_queue { struct virtnet_sq_stats stats; + struct virtnet_interrupt_coalesce intr_coal; + struct napi_struct napi; /* Record whether sq is in reset state. */ @@ -161,6 +163,8 @@ struct receive_queue { struct virtnet_rq_stats stats; + struct virtnet_interrupt_coalesce intr_coal; + /* Chain pages by the private ptr. */ struct page *pages; @@ -212,6 +216,7 @@ struct control_buf { struct virtio_net_ctrl_rss rss; struct virtio_net_ctrl_coal_tx coal_tx; struct virtio_net_ctrl_coal_rx coal_rx; + struct virtio_net_ctrl_coal_vq coal_vq; }; struct virtnet_info { @@ -3078,6 +3083,55 @@ static int virtnet_send_notf_coal_cmds(struct virtnet_info *vi, return 0; } +static int virtnet_send_ctrl_coal_vq_cmd(struct virtnet_info *vi, + u16 vqn, u32 max_usecs, u32 max_packets) +{ + struct scatterlist sgs; + + vi->ctrl->coal_vq.vqn = cpu_to_le16(vqn); + vi->ctrl->coal_vq.coal.max_usecs = cpu_to_le32(max_usecs); + vi->ctrl->coal_vq.coal.max_packets = cpu_to_le32(max_packets); + sg_init_one(&sgs, &vi->ctrl->coal_vq, sizeof(vi->ctrl->coal_vq)); + + if (!virtnet_send_command(vi, VIRTIO_NET_CTRL_NOTF_COAL, + VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET, + &sgs)) + return -EINVAL; + + return 0; +} + +static int virtnet_send_notf_coal_vq_cmds(struct virtnet_info *vi, + struct ethtool_coalesce *ec, + u16 queue) +{ + int err; + + if (ec->rx_coalesce_usecs || ec->rx_max_coalesced_frames) { + err = virtnet_send_ctrl_coal_vq_cmd(vi, rxq2vq(queue), + ec->rx_coalesce_usecs, + ec->rx_max_coalesced_frames); + if (err) + return err; + /* Save parameters */ + vi->rq[queue].intr_coal.max_usecs = ec->rx_coalesce_usecs; + vi->rq[queue].intr_coal.max_packets = ec->rx_max_coalesced_frames; + } + + if (ec->tx_coalesce_usecs || ec->tx_max_coalesced_frames) { + err = virtnet_send_ctrl_coal_vq_cmd(vi, txq2vq(queue), + ec->tx_coalesce_usecs, + ec->tx_max_coalesced_frames); + if (err) + return err; + /* Save parameters */ + vi->sq[queue].intr_coal.max_usecs = ec->tx_coalesce_usecs; + vi->sq[queue].intr_coal.max_packets = ec->tx_max_coalesced_frames; + } + + return 0; +} + static int virtnet_coal_params_supported(struct ethtool_coalesce *ec) { /* usecs coalescing is supported only if VIRTIO_NET_F_NOTF_COAL @@ -3093,22 +3147,42 @@ static int virtnet_coal_params_supported(struct ethtool_coalesce *ec) return 0; } +static int virtnet_should_update_vq_weight(int dev_flags, int weight, + int vq_weight, bool *should_update) +{ + if (weight ^ vq_weight) { + if (dev_flags & IFF_UP) + return -EBUSY; + *should_update = true; + } + + return 0; +} + static int virtnet_set_coalesce(struct net_device *dev, struct ethtool_coalesce *ec, struct kernel_ethtool_coalesce *kernel_coal, struct netlink_ext_ack *extack) { struct virtnet_info *vi = netdev_priv(dev); - int ret, i, napi_weight; + int ret, queue_number, napi_weight; bool update_napi = false; /* Can't change NAPI weight if the link is up */ napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0; - if (napi_weight ^ vi->sq[0].napi.weight) { - if (dev->flags & IFF_UP) - return -EBUSY; - else - update_napi = true; + for (queue_number = 0; queue_number < vi->max_queue_pairs; queue_number++) { + ret = virtnet_should_update_vq_weight(dev->flags, napi_weight, + vi->sq[queue_number].napi.weight, + &update_napi); + if (ret) + return ret; + + if (update_napi) { + /* All queues that belong to [queue_number, queue_count] will be + * updated for the sake of simplicity, which might not be necessary + */ + break; + } } if (virtio_has_feature(vi->vdev, VIRTIO_NET_F_NOTF_COAL)) @@ -3120,8 +3194,8 @@ static int virtnet_set_coalesce(struct net_device *dev, return ret; if (update_napi) { - for (i = 0; i < vi->max_queue_pairs; i++) - vi->sq[i].napi.weight = napi_weight; + for (; queue_number < vi->max_queue_pairs; queue_number++) + vi->sq[queue_number].napi.weight = napi_weight; } return ret; @@ -3149,6 +3223,63 @@ static int virtnet_get_coalesce(struct net_device *dev, return 0; } +static int virtnet_set_per_queue_coalesce(struct net_device *dev, + u32 queue, + struct ethtool_coalesce *ec) +{ + struct virtnet_info *vi = netdev_priv(dev); + int ret, napi_weight; + bool update_napi = false; + + if (queue >= vi->max_queue_pairs) + return -EINVAL; + + /* Can't change NAPI weight if the link is up */ + napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0; + ret = virtnet_should_update_vq_weight(dev->flags, napi_weight, + vi->sq[queue].napi.weight, + &update_napi); + if (ret) + return ret; + + if (virtio_has_feature(vi->vdev, VIRTIO_NET_F_VQ_NOTF_COAL)) + ret = virtnet_send_notf_coal_vq_cmds(vi, ec, queue); + else + ret = virtnet_coal_params_supported(ec); + + if (ret) + return ret; + + if (update_napi) + vi->sq[queue].napi.weight = napi_weight; + + return 0; +} + +static int virtnet_get_per_queue_coalesce(struct net_device *dev, + u32 queue, + struct ethtool_coalesce *ec) +{ + struct virtnet_info *vi = netdev_priv(dev); + + if (queue >= vi->max_queue_pairs) + return -EINVAL; + + if (virtio_has_feature(vi->vdev, VIRTIO_NET_F_VQ_NOTF_COAL)) { + ec->rx_coalesce_usecs = vi->rq[queue].intr_coal.max_usecs; + ec->tx_coalesce_usecs = vi->sq[queue].intr_coal.max_usecs; + ec->tx_max_coalesced_frames = vi->sq[queue].intr_coal.max_packets; + ec->rx_max_coalesced_frames = vi->rq[queue].intr_coal.max_packets; + } else { + ec->rx_max_coalesced_frames = 1; + + if (vi->sq[0].napi.weight) + ec->tx_max_coalesced_frames = 1; + } + + return 0; +} + static void virtnet_init_settings(struct net_device *dev) { struct virtnet_info *vi = netdev_priv(dev); @@ -3279,6 +3410,8 @@ static const struct ethtool_ops virtnet_ethtool_ops = { .set_link_ksettings = virtnet_set_link_ksettings, .set_coalesce = virtnet_set_coalesce, .get_coalesce = virtnet_get_coalesce, + .set_per_queue_coalesce = virtnet_set_per_queue_coalesce, + .get_per_queue_coalesce = virtnet_get_per_queue_coalesce, .get_rxfh_key_size = virtnet_get_rxfh_key_size, .get_rxfh_indir_size = virtnet_get_rxfh_indir_size, .get_rxfh = virtnet_get_rxfh, diff --git a/include/uapi/linux/virtio_net.h b/include/uapi/linux/virtio_net.h index 12c1c9699935..cc65ef0f3c3e 100644 --- a/include/uapi/linux/virtio_net.h +++ b/include/uapi/linux/virtio_net.h @@ -56,6 +56,7 @@ #define VIRTIO_NET_F_MQ 22 /* Device supports Receive Flow * Steering */ #define VIRTIO_NET_F_CTRL_MAC_ADDR 23 /* Set MAC address */ +#define VIRTIO_NET_F_VQ_NOTF_COAL 52 /* Device supports virtqueue notification coalescing */ #define VIRTIO_NET_F_NOTF_COAL 53 /* Device supports notifications coalescing */ #define VIRTIO_NET_F_GUEST_USO4 54 /* Guest can handle USOv4 in. */ #define VIRTIO_NET_F_GUEST_USO6 55 /* Guest can handle USOv6 in. */ @@ -391,5 +392,18 @@ struct virtio_net_ctrl_coal_rx { }; #define VIRTIO_NET_CTRL_NOTF_COAL_RX_SET 1 +#define VIRTIO_NET_CTRL_NOTF_COAL_VQ_SET 2 +#define VIRTIO_NET_CTRL_NOTF_COAL_VQ_GET 3 + +struct virtio_net_ctrl_coal { + __le32 max_packets; + __le32 max_usecs; +}; + +struct virtio_net_ctrl_coal_vq { + __le16 vqn; + __le16 reserved; + struct virtio_net_ctrl_coal coal; +}; #endif /* _UAPI_LINUX_VIRTIO_NET_H */