From: Pavel Begunkov <[email protected]>
To: Ming Lei <[email protected]>
Cc: [email protected], Jens Axboe <[email protected]>,
[email protected]
Subject: Re: [RFC 0/2] optimise local-tw task resheduling
Date: Wed, 15 Mar 2023 16:53:09 +0000 [thread overview]
Message-ID: <[email protected]> (raw)
In-Reply-To: <ZBEvD04sH/[email protected]>
On 3/15/23 02:35, Ming Lei wrote:
> Hi Pavel
>
> On Fri, Mar 10, 2023 at 07:04:14PM +0000, Pavel Begunkov wrote:
>> io_uring extensively uses task_work, but when a task is waiting
>> for multiple CQEs it causes lots of rescheduling. This series
>> is an attempt to optimise it and be a base for future improvements.
>>
>> For some zc network tests eventually waiting for a portion of
>> buffers I've got 10x descrease in the number of context switches,
>> which reduced the CPU consumption more than twice (17% -> 8%).
>> It also helps storage cases, while running fio/t/io_uring against
>> a low performant drive it got 2x descrease of the number of context
>> switches for QD8 and ~4 times for QD32.
>
> ublk uses io_uring_cmd_complete_in_task()(io_req_task_work_add())
> heavily. So I tried this patchset, looks not see obvious change
> on both IOPS and context switches when running 't/io_uring /dev/ublkb0',
> and it is one null ublk target(ublk add -t null -z -u 1 -q 2), IOPS
> is ~2.8M.
Hi Ming,
It's enabled for rw requests and send-zc notifications, but
io_uring_cmd_complete_in_task() is not covered. I'll be enabling
it for more cases, including pass through.
> But ublk applies batch schedule similar with io_uring before calling
> io_uring_cmd_complete_in_task().
The feature doesn't tolerate tw that produce multiple CQEs, so
it can't be applied to this batching and the task would stuck
waiting.
btw, from a quick look it appeared that ublk batching is there
to keep requests together but not to improve batching. And if so,
I think we can get rid of it, rely on io_uring batching and
let ublk to gather its requests from tw list, which sounds
cleaner. I'll elaborate on that later
--
Pavel Begunkov
next prev parent reply other threads:[~2023-03-15 16:56 UTC|newest]
Thread overview: 17+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-03-10 19:04 [RFC 0/2] optimise local-tw task resheduling Pavel Begunkov
2023-03-10 19:04 ` [RFC 1/2] io_uring: add tw add flags Pavel Begunkov
2023-03-10 19:04 ` [RFC 2/2] io_uring: reduce sheduling due to tw Pavel Begunkov
2023-03-11 17:24 ` [RFC 0/2] optimise local-tw task resheduling Jens Axboe
2023-03-11 20:45 ` Pavel Begunkov
2023-03-11 20:53 ` Pavel Begunkov
2023-03-12 15:31 ` Jens Axboe
2023-03-13 3:52 ` Pavel Begunkov
2023-03-12 15:30 ` Jens Axboe
2023-03-13 3:45 ` Pavel Begunkov
2023-03-13 14:16 ` Jens Axboe
2023-03-13 17:50 ` Pavel Begunkov
2023-03-13 22:01 ` Jens Axboe
2023-03-16 12:25 ` Pavel Begunkov
2023-03-15 2:35 ` Ming Lei
2023-03-15 16:53 ` Pavel Begunkov [this message]
2023-03-16 1:25 ` Ming Lei
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox