diff mbox series

io_uring: ensure REQ_F_ISREG is set async offload

Message ID a4c89aea-bc3c-b2f6-a1ae-2121e3353d79@kernel.dk (mailing list archive)
State New
Headers show
Series io_uring: ensure REQ_F_ISREG is set async offload | expand

Commit Message

Jens Axboe July 21, 2022, 3:11 p.m. UTC
If we're offloading requests directly to io-wq because IOSQE_ASYNC was
set in the sqe, we can miss hashing writes appropriately because we
haven't set REQ_F_ISREG yet. This can cause a performance regression
with buffered writes, as io-wq then no longer correctly serializes writes
to that file.

Ensure that we set the flags in io_prep_async_work(), which will cause
the io-wq work item to be hashed appropriately.

Fixes: 584b0180f0f4 ("io_uring: move read/write file prep state into actual opcode handler")
Link: https://lore.kernel.org/io-uring/20220608080054.GB22428@xsang-OptiPlex-9020/
Reported-and-tested-by: Yin Fengwei <fengwei.yin@intel.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>

---

Yin, this is the 5.20 version. Once this lands upstream, I'll get the
5.19/18 variant sent to stable as well. Thanks!

Comments

Yin, Fengwei July 21, 2022, 11:43 p.m. UTC | #1
On 7/21/2022 11:11 PM, Jens Axboe wrote:
> If we're offloading requests directly to io-wq because IOSQE_ASYNC was
> set in the sqe, we can miss hashing writes appropriately because we
> haven't set REQ_F_ISREG yet. This can cause a performance regression
> with buffered writes, as io-wq then no longer correctly serializes writes
> to that file.
> 
> Ensure that we set the flags in io_prep_async_work(), which will cause
> the io-wq work item to be hashed appropriately.
> 
> Fixes: 584b0180f0f4 ("io_uring: move read/write file prep state into actual opcode handler")
> Link: https://lore.kernel.org/io-uring/20220608080054.GB22428@xsang-OptiPlex-9020/
> Reported-and-tested-by: Yin Fengwei <fengwei.yin@intel.com>
This issue is reported by (from the original report):

If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang@intel.com>



Regards
Yin, Fengwei

> Signed-off-by: Jens Axboe <axboe@kernel.dk>
> 
> ---
> 
> Yin, this is the 5.20 version. Once this lands upstream, I'll get the
> 5.19/18 variant sent to stable as well. Thanks!
> 
> diff --git a/io_uring/io_uring.c b/io_uring/io_uring.c
> index 624535c62565..4b3fd645d023 100644
> --- a/io_uring/io_uring.c
> +++ b/io_uring/io_uring.c
> @@ -406,6 +406,9 @@ static void io_prep_async_work(struct io_kiocb *req)
>  	if (req->flags & REQ_F_FORCE_ASYNC)
>  		req->work.flags |= IO_WQ_WORK_CONCURRENT;
>  
> +	if (req->file && !io_req_ffs_set(req))
> +		req->flags |= io_file_get_flags(req->file) << REQ_F_SUPPORT_NOWAIT_BIT;
> +
>  	if (req->flags & REQ_F_ISREG) {
>  		if (def->hash_reg_file || (ctx->flags & IORING_SETUP_IOPOLL))
>  			io_wq_hash_work(&req->work, file_inode(req->file));
> diff --git a/io_uring/io_uring.h b/io_uring/io_uring.h
> index 868f45d55543..5db0a60dc04e 100644
> --- a/io_uring/io_uring.h
> +++ b/io_uring/io_uring.h
> @@ -41,6 +41,11 @@ struct file *io_file_get_normal(struct io_kiocb *req, int fd);
>  struct file *io_file_get_fixed(struct io_kiocb *req, int fd,
>  			       unsigned issue_flags);
>  
> +static inline bool io_req_ffs_set(struct io_kiocb *req)
> +{
> +	return req->flags & REQ_F_FIXED_FILE;
> +}
> +
>  bool io_is_uring_fops(struct file *file);
>  bool io_alloc_async_data(struct io_kiocb *req);
>  void io_req_task_work_add(struct io_kiocb *req);
> diff --git a/io_uring/rw.c b/io_uring/rw.c
> index ade3e235f277..c50a0d66d67a 100644
> --- a/io_uring/rw.c
> +++ b/io_uring/rw.c
> @@ -647,11 +647,6 @@ static bool need_read_all(struct io_kiocb *req)
>  		S_ISBLK(file_inode(req->file)->i_mode);
>  }
>  
> -static inline bool io_req_ffs_set(struct io_kiocb *req)
> -{
> -	return req->flags & REQ_F_FIXED_FILE;
> -}
> -
>  static int io_rw_init_file(struct io_kiocb *req, fmode_t mode)
>  {
>  	struct io_rw *rw = io_kiocb_to_cmd(req);
>
Jens Axboe July 22, 2022, 1:36 a.m. UTC | #2
On 7/21/22 5:43 PM, Yin Fengwei wrote:
> 
> On 7/21/2022 11:11 PM, Jens Axboe wrote:
>> If we're offloading requests directly to io-wq because IOSQE_ASYNC was
>> set in the sqe, we can miss hashing writes appropriately because we
>> haven't set REQ_F_ISREG yet. This can cause a performance regression
>> with buffered writes, as io-wq then no longer correctly serializes writes
>> to that file.
>>
>> Ensure that we set the flags in io_prep_async_work(), which will cause
>> the io-wq work item to be hashed appropriately.
>>
>> Fixes: 584b0180f0f4 ("io_uring: move read/write file prep state into actual opcode handler")
>> Link: https://lore.kernel.org/io-uring/20220608080054.GB22428@xsang-OptiPlex-9020/
>> Reported-and-tested-by: Yin Fengwei <fengwei.yin@intel.com>
> This issue is reported by (from the original report):
> 
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <oliver.sang@intel.com>

Sorry, I missed that in the original. I'll be rebasing this branch this
weekend anyway, I'll try and remember to make the edit.
diff mbox series

Patch

diff --git a/io_uring/io_uring.c b/io_uring/io_uring.c
index 624535c62565..4b3fd645d023 100644
--- a/io_uring/io_uring.c
+++ b/io_uring/io_uring.c
@@ -406,6 +406,9 @@  static void io_prep_async_work(struct io_kiocb *req)
 	if (req->flags & REQ_F_FORCE_ASYNC)
 		req->work.flags |= IO_WQ_WORK_CONCURRENT;
 
+	if (req->file && !io_req_ffs_set(req))
+		req->flags |= io_file_get_flags(req->file) << REQ_F_SUPPORT_NOWAIT_BIT;
+
 	if (req->flags & REQ_F_ISREG) {
 		if (def->hash_reg_file || (ctx->flags & IORING_SETUP_IOPOLL))
 			io_wq_hash_work(&req->work, file_inode(req->file));
diff --git a/io_uring/io_uring.h b/io_uring/io_uring.h
index 868f45d55543..5db0a60dc04e 100644
--- a/io_uring/io_uring.h
+++ b/io_uring/io_uring.h
@@ -41,6 +41,11 @@  struct file *io_file_get_normal(struct io_kiocb *req, int fd);
 struct file *io_file_get_fixed(struct io_kiocb *req, int fd,
 			       unsigned issue_flags);
 
+static inline bool io_req_ffs_set(struct io_kiocb *req)
+{
+	return req->flags & REQ_F_FIXED_FILE;
+}
+
 bool io_is_uring_fops(struct file *file);
 bool io_alloc_async_data(struct io_kiocb *req);
 void io_req_task_work_add(struct io_kiocb *req);
diff --git a/io_uring/rw.c b/io_uring/rw.c
index ade3e235f277..c50a0d66d67a 100644
--- a/io_uring/rw.c
+++ b/io_uring/rw.c
@@ -647,11 +647,6 @@  static bool need_read_all(struct io_kiocb *req)
 		S_ISBLK(file_inode(req->file)->i_mode);
 }
 
-static inline bool io_req_ffs_set(struct io_kiocb *req)
-{
-	return req->flags & REQ_F_FIXED_FILE;
-}
-
 static int io_rw_init_file(struct io_kiocb *req, fmode_t mode)
 {
 	struct io_rw *rw = io_kiocb_to_cmd(req);