* [PATCH v2] io_uring: only wake up sq thread while current task is in io worker context
@ 2020-11-12 16:44 Xiaoguang Wang
2020-11-12 16:48 ` Jens Axboe
0 siblings, 1 reply; 2+ messages in thread
From: Xiaoguang Wang @ 2020-11-12 16:44 UTC (permalink / raw)
To: io-uring; +Cc: axboe, joseph.qi
If IORING_SETUP_SQPOLL is enabled, sqes are either handled in sq thread
task context or in io worker task context. If current task context is sq
thread, we don't need to check whether should wake up sq thread.
io_iopoll_req_issued() calls wq_has_sleeper(), which has smp_mb() memory
barrier, before this patch, perf shows obvious overhead:
Samples: 481K of event 'cycles', Event count (approx.): 299807382878
Overhead Comma Shared Object Symbol
3.69% :9630 [kernel.vmlinux] [k] io_issue_sqe
With this patch, perf shows:
Samples: 482K of event 'cycles', Event count (approx.): 299929547283
Overhead Comma Shared Object Symbol
0.70% :4015 [kernel.vmlinux] [k] io_issue_sqe
It shows some obvious improvements.
Signed-off-by: Xiaoguang Wang <[email protected]>
---
fs/io_uring.c | 11 ++++++++---
1 file changed, 8 insertions(+), 3 deletions(-)
diff --git a/fs/io_uring.c b/fs/io_uring.c
index f594c72de777..287efc53f23a 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -2732,7 +2732,7 @@ static void io_complete_rw_iopoll(struct kiocb *kiocb, long res, long res2)
* find it from a io_iopoll_getevents() thread before the issuer is done
* accessing the kiocb cookie.
*/
-static void io_iopoll_req_issued(struct io_kiocb *req)
+static void io_iopoll_req_issued(struct io_kiocb *req, bool in_async)
{
struct io_ring_ctx *ctx = req->ctx;
@@ -2761,7 +2761,12 @@ static void io_iopoll_req_issued(struct io_kiocb *req)
else
list_add_tail(&req->inflight_entry, &ctx->iopoll_list);
- if ((ctx->flags & IORING_SETUP_SQPOLL) &&
+ /*
+ * If IORING_SETUP_SQPOLL is enabled, sqes are either handled in sq thread
+ * task context or in io worker task context. If current task context is
+ * sq thread, we don't need to check whether should wake up sq thread.
+ */
+ if (in_async && (ctx->flags & IORING_SETUP_SQPOLL) &&
wq_has_sleeper(&ctx->sq_data->wait))
wake_up(&ctx->sq_data->wait);
}
@@ -6245,7 +6250,7 @@ static int io_issue_sqe(struct io_kiocb *req, bool force_nonblock,
if (in_async)
mutex_lock(&ctx->uring_lock);
- io_iopoll_req_issued(req);
+ io_iopoll_req_issued(req, in_async);
if (in_async)
mutex_unlock(&ctx->uring_lock);
--
2.17.2
^ permalink raw reply related [flat|nested] 2+ messages in thread
* Re: [PATCH v2] io_uring: only wake up sq thread while current task is in io worker context
2020-11-12 16:44 [PATCH v2] io_uring: only wake up sq thread while current task is in io worker context Xiaoguang Wang
@ 2020-11-12 16:48 ` Jens Axboe
0 siblings, 0 replies; 2+ messages in thread
From: Jens Axboe @ 2020-11-12 16:48 UTC (permalink / raw)
To: Xiaoguang Wang, io-uring; +Cc: joseph.qi
On 11/12/20 9:44 AM, Xiaoguang Wang wrote:
> If IORING_SETUP_SQPOLL is enabled, sqes are either handled in sq thread
> task context or in io worker task context. If current task context is sq
> thread, we don't need to check whether should wake up sq thread.
>
> io_iopoll_req_issued() calls wq_has_sleeper(), which has smp_mb() memory
> barrier, before this patch, perf shows obvious overhead:
> Samples: 481K of event 'cycles', Event count (approx.): 299807382878
> Overhead Comma Shared Object Symbol
> 3.69% :9630 [kernel.vmlinux] [k] io_issue_sqe
>
> With this patch, perf shows:
> Samples: 482K of event 'cycles', Event count (approx.): 299929547283
> Overhead Comma Shared Object Symbol
> 0.70% :4015 [kernel.vmlinux] [k] io_issue_sqe
>
> It shows some obvious improvements.
Applied, thanks.
--
Jens Axboe
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2020-11-12 16:48 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2020-11-12 16:44 [PATCH v2] io_uring: only wake up sq thread while current task is in io worker context Xiaoguang Wang
2020-11-12 16:48 ` Jens Axboe
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox