Message ID | 20230102160317.89851-1-ynachum@amazon.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | [for-rc] RDMA: Fix ib block iterator counter overflow | expand |
On 02/01/2023 18:03, Yonatan Nachum wrote: > When registering a new DMA MR after selecting the best aligned page size > for it, we iterate over the given sglist to split each entry to smaller, > aligned to the selected page size, DMA blocks. > > In given circumstances where the sg entry and page size fit certain sizes > and the sg entry is not aligned to the selected page size, the total size > of the aligned pages we need to cover the sg entry is >= 4GB. Under this > circumstances, while iterating page aligned blocks, the counter responsible > for counting how much we advanced from the start of the sg entry is > overflowed because its type is u32 and we pass 4GB in size. This can > lead to an infinite loop inside the iterator function because in some > cases the overflow prevents the counter to be larger than the size of > the sg entry. > > Fix the presented problem with changing the counter type to u64. > > Backtrace: > [ 192.374329] efa_reg_user_mr_dmabuf > [ 192.376783] efa_register_mr > [ 192.382579] pgsz_bitmap 0xfffff000 rounddown 0x80000000 > [ 192.386423] pg_sz [0x80000000] umem_length[0xc0000000] > [ 192.392657] start 0x0 length 0xc0000000 params.page_shift 31 params.page_num 3 > [ 192.399559] hp_cnt[3], pages_in_hp[524288] > [ 192.403690] umem->sgt_append.sgt.nents[1] > [ 192.407905] number entries: [1], pg_bit: [31] > [ 192.411397] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] > [ 192.415601] biter->__sg_advance [665837568] sg_dma_len[3221225472] > [ 192.419823] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] > [ 192.423976] biter->__sg_advance [2813321216] sg_dma_len[3221225472] > [ 192.428243] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] > [ 192.432397] biter->__sg_advance [665837568] sg_dma_len[3221225472] > > Fixes: a808273a495c Missing the patch subject line, please see: https://www.kernel.org/doc/html/latest/process/submitting-patches.html#describe-your-changes Also, there shouldn't be a blank line here. > > Signed-off-by: Yonatan Nachum <ynachum@amazon.com>
On Mon, Jan 02, 2023 at 06:03:17PM +0200, Yonatan Nachum wrote: > When registering a new DMA MR after selecting the best aligned page size > for it, we iterate over the given sglist to split each entry to smaller, > aligned to the selected page size, DMA blocks. > > In given circumstances where the sg entry and page size fit certain sizes > and the sg entry is not aligned to the selected page size, the total size > of the aligned pages we need to cover the sg entry is >= 4GB. Under this > circumstances, while iterating page aligned blocks, the counter responsible > for counting how much we advanced from the start of the sg entry is > overflowed because its type is u32 and we pass 4GB in size. This can > lead to an infinite loop inside the iterator function because in some > cases the overflow prevents the counter to be larger than the size of > the sg entry. > > Fix the presented problem with changing the counter type to u64. > > Backtrace: > [ 192.374329] efa_reg_user_mr_dmabuf > [ 192.376783] efa_register_mr > [ 192.382579] pgsz_bitmap 0xfffff000 rounddown 0x80000000 > [ 192.386423] pg_sz [0x80000000] umem_length[0xc0000000] > [ 192.392657] start 0x0 length 0xc0000000 params.page_shift 31 params.page_num 3 > [ 192.399559] hp_cnt[3], pages_in_hp[524288] > [ 192.403690] umem->sgt_append.sgt.nents[1] > [ 192.407905] number entries: [1], pg_bit: [31] > [ 192.411397] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] > [ 192.415601] biter->__sg_advance [665837568] sg_dma_len[3221225472] > [ 192.419823] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] > [ 192.423976] biter->__sg_advance [2813321216] sg_dma_len[3221225472] > [ 192.428243] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] > [ 192.432397] biter->__sg_advance [665837568] sg_dma_len[3221225472] > > Fixes: a808273a495c The fixes line is truncated. Please fix and resend. Thanks > > Signed-off-by: Yonatan Nachum <ynachum@amazon.com> > --- > include/rdma/ib_verbs.h | 2 +- > 1 file changed, 1 insertion(+), 1 deletion(-) > > diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h > index 975d6e9efbcb..6821c7951363 100644 > --- a/include/rdma/ib_verbs.h > +++ b/include/rdma/ib_verbs.h > @@ -2836,7 +2836,7 @@ struct ib_block_iter { > struct scatterlist *__sg; /* sg holding the current aligned block */ > dma_addr_t __dma_addr; /* unaligned DMA address of this block */ > unsigned int __sg_nents; /* number of SG entries */ > - unsigned int __sg_advance; /* number of bytes to advance in sg in next step */ > + u64 __sg_advance; /* number of bytes to advance in sg in next step */ > unsigned int __pg_bit; /* alignment of current block */ > }; > > -- > 2.38.1 >
On Mon, Jan 02, 2023 at 06:03:17PM +0200, Yonatan Nachum wrote: > When registering a new DMA MR after selecting the best aligned page size > for it, we iterate over the given sglist to split each entry to smaller, > aligned to the selected page size, DMA blocks. > > In given circumstances where the sg entry and page size fit certain sizes > and the sg entry is not aligned to the selected page size, the total size > of the aligned pages we need to cover the sg entry is >= 4GB. Huh? The max length of a scatterlist element is unsigned int: struct scatterlist { unsigned long page_link; unsigned int offset; unsigned int length; The problem is the construction of the end of iteration test: block_offset = biter->__dma_addr & (BIT_ULL(biter->__pg_bit) - 1); biter->__sg_advance += BIT_ULL(biter->__pg_bit) - block_offset; if (biter->__sg_advance >= sg_dma_len(biter->__sg)) { This should be reworked so it doesn't increment until it knows it is OK and leave the types alone. Jason
diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h index 975d6e9efbcb..6821c7951363 100644 --- a/include/rdma/ib_verbs.h +++ b/include/rdma/ib_verbs.h @@ -2836,7 +2836,7 @@ struct ib_block_iter { struct scatterlist *__sg; /* sg holding the current aligned block */ dma_addr_t __dma_addr; /* unaligned DMA address of this block */ unsigned int __sg_nents; /* number of SG entries */ - unsigned int __sg_advance; /* number of bytes to advance in sg in next step */ + u64 __sg_advance; /* number of bytes to advance in sg in next step */ unsigned int __pg_bit; /* alignment of current block */ };
When registering a new DMA MR after selecting the best aligned page size for it, we iterate over the given sglist to split each entry to smaller, aligned to the selected page size, DMA blocks. In given circumstances where the sg entry and page size fit certain sizes and the sg entry is not aligned to the selected page size, the total size of the aligned pages we need to cover the sg entry is >= 4GB. Under this circumstances, while iterating page aligned blocks, the counter responsible for counting how much we advanced from the start of the sg entry is overflowed because its type is u32 and we pass 4GB in size. This can lead to an infinite loop inside the iterator function because in some cases the overflow prevents the counter to be larger than the size of the sg entry. Fix the presented problem with changing the counter type to u64. Backtrace: [ 192.374329] efa_reg_user_mr_dmabuf [ 192.376783] efa_register_mr [ 192.382579] pgsz_bitmap 0xfffff000 rounddown 0x80000000 [ 192.386423] pg_sz [0x80000000] umem_length[0xc0000000] [ 192.392657] start 0x0 length 0xc0000000 params.page_shift 31 params.page_num 3 [ 192.399559] hp_cnt[3], pages_in_hp[524288] [ 192.403690] umem->sgt_append.sgt.nents[1] [ 192.407905] number entries: [1], pg_bit: [31] [ 192.411397] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] [ 192.415601] biter->__sg_advance [665837568] sg_dma_len[3221225472] [ 192.419823] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] [ 192.423976] biter->__sg_advance [2813321216] sg_dma_len[3221225472] [ 192.428243] biter->__sg_nents [1] biter->__sg [0000000008b0c5d8] [ 192.432397] biter->__sg_advance [665837568] sg_dma_len[3221225472] Fixes: a808273a495c Signed-off-by: Yonatan Nachum <ynachum@amazon.com> --- include/rdma/ib_verbs.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-)