Message ID | 20210902130625.25277-1-weijunji@bytedance.com (mailing list archive) |
---|---|
Headers | show |
Series | VirtIO RDMA | expand |
On Thu, Sep 2, 2021 at 9:07 PM Junji Wei <weijunji@bytedance.com> wrote: > > Hi all, > > This RFC aims to reopen the discussion of Virtio RDMA. > Now this is based on Yuval Shaia's RFC "VirtIO RDMA" > which implemented a frame for Virtio RDMA and a simple > control path (Not sure if Yuval Shaia has any further > plan for it). > > We try to extend this work and implement a simple > data-path and a completed control path. Now this can > work with SEND, RECV and REG_MR in kernel. There is a > simple test module in this patch that can communicate > with ibv_rc_pingpong in rdma-core. > > During doing this work, we have found some problems and > would like to ask for some suggestions from community: I think it would be beneficial if you can post a spec patch. Thanks
> On Sep 3, 2021, at 8:57 AM, Jason Wang <jasowang@redhat.com> wrote: > > On Thu, Sep 2, 2021 at 9:07 PM Junji Wei <weijunji@bytedance.com> wrote: >> >> Hi all, >> >> This RFC aims to reopen the discussion of Virtio RDMA. >> Now this is based on Yuval Shaia's RFC "VirtIO RDMA" >> which implemented a frame for Virtio RDMA and a simple >> control path (Not sure if Yuval Shaia has any further >> plan for it). >> >> We try to extend this work and implement a simple >> data-path and a completed control path. Now this can >> work with SEND, RECV and REG_MR in kernel. There is a >> simple test module in this patch that can communicate >> with ibv_rc_pingpong in rdma-core. >> >> During doing this work, we have found some problems and >> would like to ask for some suggestions from community: > > I think it would be beneficial if you can post a spec patch. Ok, I will do it. Thanks
On Thu, Sep 02, 2021 at 09:06:20PM +0800, Junji Wei wrote: > Hi all, > > This RFC aims to reopen the discussion of Virtio RDMA. > Now this is based on Yuval Shaia's RFC "VirtIO RDMA" > which implemented a frame for Virtio RDMA and a simple > control path (Not sure if Yuval Shaia has any further > plan for it). > > We try to extend this work and implement a simple > data-path and a completed control path. Now this can > work with SEND, RECV and REG_MR in kernel. There is a > simple test module in this patch that can communicate > with ibv_rc_pingpong in rdma-core. > > During doing this work, we have found some problems and > would like to ask for some suggestions from community: These seem like serious problems! Shouldn't these be solved before sending patches? > 1. Each qp need two VQ, but qemu default only support 1024 VQ. > I think it is possible to multiplex the VQ, since the > cmd_post_send carry the qpn in request. QPs and CQs need to have predictable fixed WQE sizes, I don't know how you can reasonably expect to map them to a shared queue. > 2. The virtio-rdma device's gid should equal to host rdma > device's gid. This means that we cannot use gid cache in > rdma subsystem. And theoretically the gid should also equal > to the device's netdev's ip address, how can we deal with > this conflict. You have to follow the correct semantics, the GID flows from the guest into the host and updates the hosts GID table, not the other way around. > 3. How to support DMA mr? The verbs in host cannot support it. > And it seems hard to ping whole guest physical memory in qemu. Either you have to trap the FRWR in the hypervisor and pin the memory, remap the MR, etc or you have to pin the entire guest and rely on something like memory windows to emulate FRWR. > 4. The FRMR api need to set key of MR through IB_WR_REG_MR. > But it is impossible to change a key of mr using uverbs. FRMR is more like memory windows in user space, you can't support it using just regular MRs. > In our implementation, we change the key of WR while post_send, > but this means the MR can only work with SEND and RECV since we > cannot change the key in the remote. Yes, this is not a realistic solution > 5. The GSI is not supported now. And we think it's a problem that > when the host receive a GSI package, it doesn't know which > device it belongs to. Of course, GSI packets are not virtualized. You need to somehow capture GSI messages for the entire GID that the guest is using. We don't have any API to do this in userspace. Jason
> On Sep 15, 2021, at 9:43 PM, Jason Gunthorpe <jgg@nvidia.com> wrote: > > On Thu, Sep 02, 2021 at 09:06:20PM +0800, Junji Wei wrote: >> Hi all, >> >> This RFC aims to reopen the discussion of Virtio RDMA. >> Now this is based on Yuval Shaia's RFC "VirtIO RDMA" >> which implemented a frame for Virtio RDMA and a simple >> control path (Not sure if Yuval Shaia has any further >> plan for it). >> >> We try to extend this work and implement a simple >> data-path and a completed control path. Now this can >> work with SEND, RECV and REG_MR in kernel. There is a >> simple test module in this patch that can communicate >> with ibv_rc_pingpong in rdma-core. >> >> During doing this work, we have found some problems and >> would like to ask for some suggestions from community: > > These seem like serious problems! Shouldn't these be solved before > sending patches? > >> 1. Each qp need two VQ, but qemu default only support 1024 VQ. >> I think it is possible to multiplex the VQ, since the >> cmd_post_send carry the qpn in request. > > QPs and CQs need to have predictable fixed WQE sizes, I don't know how > you can reasonably expect to map them to a shared queue. Yes, it is a bad idea to multiplex the VQ. If we need more VQ, we can extend QEMU and virtio spec. >> 2. The virtio-rdma device's gid should equal to host rdma >> device's gid. This means that we cannot use gid cache in >> rdma subsystem. And theoretically the gid should also equal >> to the device's netdev's ip address, how can we deal with >> this conflict. > > You have to follow the correct semantics, the GID flows from the guest > into the host and updates the hosts GID table, not the other way > around. Sure, this is my misunderstanding. >> 3. How to support DMA mr? The verbs in host cannot support it. >> And it seems hard to ping whole guest physical memory in qemu. > > Either you have to trap the FRWR in the hypervisor and pin the memory, > remap the MR, etc or you have to pin the entire guest and rely on > something like memory windows to emulate FRWR. We want to implement an emulated RDMA device in userspace. Since we can directly access guest's physical memory in QEMU, it will be easy to support DMA mr. >> 4. The FRMR api need to set key of MR through IB_WR_REG_MR. >> But it is impossible to change a key of mr using uverbs. > > FRMR is more like memory windows in user space, you can't support it > using just regular MRs. It is hard to support this using uverbs, but it is easy to support with uRDMA that we can get full control of mrs. >> 5. The GSI is not supported now. And we think it's a problem that >> when the host receive a GSI package, it doesn't know which >> device it belongs to. > > Of course, GSI packets are not virtualized. You need to somehow > capture GSI messages for the entire GID that the guest is using. We > don't have any API to do this in userspace. If we implement uRDMA device in QEMU, there is no need to distinguish which device it belongs to, because there is only one device. Thanks. Junji
On Wed, Sep 22, 2021 at 08:08:44PM +0800, Junji Wei wrote: > > On Sep 15, 2021, at 9:43 PM, Jason Gunthorpe <jgg@nvidia.com> wrote: <...> > >> 4. The FRMR api need to set key of MR through IB_WR_REG_MR. > >> But it is impossible to change a key of mr using uverbs. > > > > FRMR is more like memory windows in user space, you can't support it > > using just regular MRs. > > It is hard to support this using uverbs, but it is easy to support > with uRDMA that we can get full control of mrs. What is uRDMA? Thanks
On Wed, Sep 22, 2021 at 9:06 PM Leon Romanovsky <leon@kernel.org> wrote: > > On Wed, Sep 22, 2021 at 08:08:44PM +0800, Junji Wei wrote: > > > On Sep 15, 2021, at 9:43 PM, Jason Gunthorpe <jgg@nvidia.com> wrote: > > <...> > > > >> 4. The FRMR api need to set key of MR through IB_WR_REG_MR. > > >> But it is impossible to change a key of mr using uverbs. > > > > > > FRMR is more like memory windows in user space, you can't support it > > > using just regular MRs. > > > > It is hard to support this using uverbs, but it is easy to support > > with uRDMA that we can get full control of mrs. > > What is uRDMA? uRDMA is a software implementation of the RoCEv2 protocol like rxe. We will implement it in QEMU with VFIO or DPDK. Thanks. Junji
On Wed, Sep 22, 2021 at 09:37:37PM +0800, 魏俊吉 wrote: > On Wed, Sep 22, 2021 at 9:06 PM Leon Romanovsky <leon@kernel.org> wrote: > > > > On Wed, Sep 22, 2021 at 08:08:44PM +0800, Junji Wei wrote: > > > > On Sep 15, 2021, at 9:43 PM, Jason Gunthorpe <jgg@nvidia.com> wrote: > > > > <...> > > > > > >> 4. The FRMR api need to set key of MR through IB_WR_REG_MR. > > > >> But it is impossible to change a key of mr using uverbs. > > > > > > > > FRMR is more like memory windows in user space, you can't support it > > > > using just regular MRs. > > > > > > It is hard to support this using uverbs, but it is easy to support > > > with uRDMA that we can get full control of mrs. > > > > What is uRDMA? > > uRDMA is a software implementation of the RoCEv2 protocol like rxe. > We will implement it in QEMU with VFIO or DPDK. ok, thanks > > Thanks. > Junji