From: Hao Xu <[email protected]>
To: Pavel Begunkov <[email protected]>,
Jens Axboe <[email protected]>,
[email protected]
Cc: Joseph Qi <[email protected]>
Subject: Re: [PATCH 1/3] io_uring: clean cqe filling functions
Date: Mon, 13 Sep 2021 02:24:17 +0800 [thread overview]
Message-ID: <[email protected]> (raw)
In-Reply-To: <c1c50ac6b6badf319006f580715b8da6438e8e23.1631367587.git.asml.silence@gmail.com>
在 2021/9/11 下午9:52, Pavel Begunkov 写道:
> Split io_cqring_fill_event() into a couple of more targeted functions.
> The first on is io_fill_cqe_aux() for completions that are not
> associated with request completions and doing the ->cq_extra accounting.
> Examples are additional CQEs from multishot poll and rsrc notifications.
>
> The second is io_fill_cqe_req(), should be called when it's a normal
> request completion. Nothing more to it at the moment, will be used in
> later patches.
>
> The last one is inlined __io_fill_cqe() for a finer grained control,
> should be used with caution and in hottest places.
>
> Signed-off-by: Pavel Begunkov <[email protected]>
> ---
> fs/io_uring.c | 58 ++++++++++++++++++++++++++-------------------------
> 1 file changed, 30 insertions(+), 28 deletions(-)
>
> diff --git a/fs/io_uring.c b/fs/io_uring.c
> index e6ccdae189b0..1703130ae8df 100644
> --- a/fs/io_uring.c
> +++ b/fs/io_uring.c
> @@ -1078,8 +1078,8 @@ static void io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
> bool cancel_all);
> static void io_uring_cancel_generic(bool cancel_all, struct io_sq_data *sqd);
>
> -static bool io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data,
> - long res, unsigned int cflags);
> +static void io_fill_cqe_req(struct io_kiocb *req, long res, unsigned int cflags);
> +
> static void io_put_req(struct io_kiocb *req);
> static void io_put_req_deferred(struct io_kiocb *req);
> static void io_dismantle_req(struct io_kiocb *req);
> @@ -1491,7 +1491,7 @@ static void io_kill_timeout(struct io_kiocb *req, int status)
> atomic_set(&req->ctx->cq_timeouts,
> atomic_read(&req->ctx->cq_timeouts) + 1);
> list_del_init(&req->timeout.list);
> - io_cqring_fill_event(req->ctx, req->user_data, status, 0);
> + io_fill_cqe_req(req, status, 0);
> io_put_req_deferred(req);
> }
> }
> @@ -1760,8 +1760,8 @@ static bool io_cqring_event_overflow(struct io_ring_ctx *ctx, u64 user_data,
> return true;
> }
>
> -static inline bool __io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data,
> - long res, unsigned int cflags)
> +static inline bool __io_fill_cqe(struct io_ring_ctx *ctx, u64 user_data,
> + long res, unsigned int cflags)
> {
> struct io_uring_cqe *cqe;
>
> @@ -1782,11 +1782,17 @@ static inline bool __io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data
> return io_cqring_event_overflow(ctx, user_data, res, cflags);
> }
>
> -/* not as hot to bloat with inlining */
> -static noinline bool io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data,
> - long res, unsigned int cflags)
> +static noinline void io_fill_cqe_req(struct io_kiocb *req, long res,
> + unsigned int cflags)
> +{
> + __io_fill_cqe(req->ctx, req->user_data, res, cflags);
> +}
> +
> +static noinline bool io_fill_cqe_aux(struct io_ring_ctx *ctx, u64 user_data,
> + long res, unsigned int cflags)
> {
> - return __io_cqring_fill_event(ctx, user_data, res, cflags);
> + ctx->cq_extra++;
> + return __io_fill_cqe(ctx, user_data, res, cflags);
> }
>
> static void io_req_complete_post(struct io_kiocb *req, long res,
> @@ -1795,7 +1801,7 @@ static void io_req_complete_post(struct io_kiocb *req, long res,
> struct io_ring_ctx *ctx = req->ctx;
>
> spin_lock(&ctx->completion_lock);
> - __io_cqring_fill_event(ctx, req->user_data, res, cflags);
> + __io_fill_cqe(ctx, req->user_data, res, cflags);
> /*
> * If we're the last reference to this request, add to our locked
> * free_list cache.
> @@ -2021,8 +2027,7 @@ static bool io_kill_linked_timeout(struct io_kiocb *req)
> link->timeout.head = NULL;
> if (hrtimer_try_to_cancel(&io->timer) != -1) {
> list_del(&link->timeout.list);
> - io_cqring_fill_event(link->ctx, link->user_data,
> - -ECANCELED, 0);
> + io_fill_cqe_req(link, -ECANCELED, 0);
> io_put_req_deferred(link);
> return true;
> }
> @@ -2046,7 +2051,7 @@ static void io_fail_links(struct io_kiocb *req)
> link->link = NULL;
>
> trace_io_uring_fail_link(req, link);
> - io_cqring_fill_event(link->ctx, link->user_data, res, 0);
> + io_fill_cqe_req(link, res, 0);
> io_put_req_deferred(link);
> link = nxt;
> }
> @@ -2063,8 +2068,7 @@ static bool io_disarm_next(struct io_kiocb *req)
> req->flags &= ~REQ_F_ARM_LTIMEOUT;
> if (link && link->opcode == IORING_OP_LINK_TIMEOUT) {
> io_remove_next_linked(req);
> - io_cqring_fill_event(link->ctx, link->user_data,
> - -ECANCELED, 0);
> + io_fill_cqe_req(link, -ECANCELED, 0);
> io_put_req_deferred(link);
> posted = true;
> }
> @@ -2335,8 +2339,8 @@ static void __io_submit_flush_completions(struct io_ring_ctx *ctx)
> for (i = 0; i < nr; i++) {
> struct io_kiocb *req = state->compl_reqs[i];
>
> - __io_cqring_fill_event(ctx, req->user_data, req->result,
> - req->compl.cflags);
> + __io_fill_cqe(ctx, req->user_data, req->result,
> + req->compl.cflags);
> }
> io_commit_cqring(ctx);
> spin_unlock(&ctx->completion_lock);
> @@ -2454,8 +2458,8 @@ static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events,
> continue;
> }
>
> - __io_cqring_fill_event(ctx, req->user_data, req->result,
> - io_put_rw_kbuf(req));
> + __io_fill_cqe(ctx, req->user_data, req->result,
> + io_put_rw_kbuf(req));
> (*nr_events)++;
>
> if (req_ref_put_and_test(req))
> @@ -5293,13 +5297,12 @@ static bool __io_poll_complete(struct io_kiocb *req, __poll_t mask)
> }
> if (req->poll.events & EPOLLONESHOT)
> flags = 0;
> - if (!io_cqring_fill_event(ctx, req->user_data, error, flags)) {
> + if (!(flags & IORING_CQE_F_MORE)) {
> + io_fill_cqe_req(req, error, flags);
We should check the return value of io_fill_cqe_req() and do
req->poll.done = true if the return value is false, which means ocqe
allocation failed. Though I think the current poll.done logic itself
is not right.(I've changed it in another patch)
> + } else if (!io_fill_cqe_aux(ctx, req->user_data, error, flags)) {
> req->poll.done = true;
> flags = 0;
> }
> - if (flags & IORING_CQE_F_MORE)
> - ctx->cq_extra++;
> -
> return !(flags & IORING_CQE_F_MORE);
> }
>
> @@ -5627,9 +5630,9 @@ static bool io_poll_remove_one(struct io_kiocb *req)
> do_complete = __io_poll_remove_one(req, io_poll_get_single(req), true);
>
> if (do_complete) {
> - io_cqring_fill_event(req->ctx, req->user_data, -ECANCELED, 0);
> - io_commit_cqring(req->ctx);
> req_set_fail(req);
> + io_fill_cqe_req(req, -ECANCELED, 0);
> + io_commit_cqring(req->ctx);
> io_put_req_deferred(req);
> }
> return do_complete;
> @@ -5924,7 +5927,7 @@ static int io_timeout_cancel(struct io_ring_ctx *ctx, __u64 user_data)
> return PTR_ERR(req);
>
> req_set_fail(req);
> - io_cqring_fill_event(ctx, req->user_data, -ECANCELED, 0);
> + io_fill_cqe_req(req, -ECANCELED, 0);
> io_put_req_deferred(req);
> return 0;
> }
> @@ -8122,8 +8125,7 @@ static void __io_rsrc_put_work(struct io_rsrc_node *ref_node)
>
> io_ring_submit_lock(ctx, lock_ring);
> spin_lock(&ctx->completion_lock);
> - io_cqring_fill_event(ctx, prsrc->tag, 0, 0);
> - ctx->cq_extra++;
> + io_fill_cqe_aux(ctx, prsrc->tag, 0, 0);
> io_commit_cqring(ctx);
> spin_unlock(&ctx->completion_lock);
> io_cqring_ev_posted(ctx);
>
next prev parent reply other threads:[~2021-09-12 18:24 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-09-11 13:51 [RFC][PATCH 0/3] allow to skip CQE posting Pavel Begunkov
2021-09-11 13:52 ` [PATCH 1/3] io_uring: clean cqe filling functions Pavel Begunkov
2021-09-12 18:24 ` Hao Xu [this message]
2021-09-12 23:20 ` Pavel Begunkov
2021-09-11 13:52 ` [PATCH 2/3] io_uring: add option to skip CQE posting Pavel Begunkov
2021-09-11 13:52 ` [PATCH 3/3] io_uring: don't spinlock when not posting CQEs Pavel Begunkov
2021-09-11 20:12 ` Hao Xu
2021-09-11 21:10 ` Pavel Begunkov
2021-09-12 18:14 ` Hao Xu
2021-09-12 22:16 ` Pavel Begunkov
2021-09-12 23:49 ` [RFC][PATCH 0/3] allow to skip CQE posting Pavel Begunkov
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=3099ae18-5e15-eb7d-b9b8-ddd1217f1a04@linux.alibaba.com \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox