From: Bernd Schubert <[email protected]>
To: Ming Lei <[email protected]>, Jens Axboe <[email protected]>,
"[email protected]" <[email protected]>,
"[email protected]" <[email protected]>
Cc: "[email protected]" <[email protected]>,
Miklos Szeredi <[email protected]>,
ZiyangZhang <[email protected]>,
Xiaoguang Wang <[email protected]>,
Pavel Begunkov <[email protected]>,
Stefan Hajnoczi <[email protected]>,
Dan Williams <[email protected]>
Subject: Re: [PATCH V6 12/17] block: ublk_drv: cleanup ublk_copy_user_pages
Date: Fri, 31 Mar 2023 16:22:26 +0000 [thread overview]
Message-ID: <[email protected]> (raw)
In-Reply-To: <[email protected]>
On 3/30/23 13:36, Ming Lei wrote:
> Clean up ublk_copy_user_pages() by using iov iter, and code
> gets simplified a lot and becomes much more readable than before.
>
> Signed-off-by: Ming Lei <[email protected]>
> ---
> drivers/block/ublk_drv.c | 112 +++++++++++++++++----------------------
> 1 file changed, 49 insertions(+), 63 deletions(-)
>
> diff --git a/drivers/block/ublk_drv.c b/drivers/block/ublk_drv.c
> index fdccbf5fdaa1..cca0e95a89d8 100644
> --- a/drivers/block/ublk_drv.c
> +++ b/drivers/block/ublk_drv.c
> @@ -419,49 +419,39 @@ static const struct block_device_operations ub_fops = {
>
> #define UBLK_MAX_PIN_PAGES 32
>
> -struct ublk_map_data {
> - const struct request *rq;
> - unsigned long ubuf;
> - unsigned int len;
> -};
> -
> struct ublk_io_iter {
> struct page *pages[UBLK_MAX_PIN_PAGES];
> - unsigned pg_off; /* offset in the 1st page in pages */
> - int nr_pages; /* how many page pointers in pages */
> struct bio *bio;
> struct bvec_iter iter;
> };
>
> -static inline unsigned ublk_copy_io_pages(struct ublk_io_iter *data,
> - unsigned max_bytes, bool to_vm)
> +/* return how many pages are copied */
> +static void ublk_copy_io_pages(struct ublk_io_iter *data,
> + size_t total, size_t pg_off, int dir)
> {
> - const unsigned total = min_t(unsigned, max_bytes,
> - PAGE_SIZE - data->pg_off +
> - ((data->nr_pages - 1) << PAGE_SHIFT));
> unsigned done = 0;
> unsigned pg_idx = 0;
>
> while (done < total) {
> struct bio_vec bv = bio_iter_iovec(data->bio, data->iter);
> - const unsigned int bytes = min3(bv.bv_len, total - done,
> - (unsigned)(PAGE_SIZE - data->pg_off));
> + unsigned int bytes = min3(bv.bv_len, (unsigned)total - done,
> + (unsigned)(PAGE_SIZE - pg_off));
> void *bv_buf = bvec_kmap_local(&bv);
> void *pg_buf = kmap_local_page(data->pages[pg_idx]);
>
> - if (to_vm)
> - memcpy(pg_buf + data->pg_off, bv_buf, bytes);
> + if (dir == ITER_DEST)
> + memcpy(pg_buf + pg_off, bv_buf, bytes);
> else
> - memcpy(bv_buf, pg_buf + data->pg_off, bytes);
> + memcpy(bv_buf, pg_buf + pg_off, bytes);
>
> kunmap_local(pg_buf);
> kunmap_local(bv_buf);
>
> /* advance page array */
> - data->pg_off += bytes;
> - if (data->pg_off == PAGE_SIZE) {
> + pg_off += bytes;
> + if (pg_off == PAGE_SIZE) {
> pg_idx += 1;
> - data->pg_off = 0;
> + pg_off = 0;
> }
>
> done += bytes;
> @@ -475,41 +465,40 @@ static inline unsigned ublk_copy_io_pages(struct ublk_io_iter *data,
> data->iter = data->bio->bi_iter;
> }
> }
> -
> - return done;
> }
>
> -static int ublk_copy_user_pages(struct ublk_map_data *data, bool to_vm)
> +/*
> + * Copy data between request pages and io_iter, and 'offset'
> + * is the start point of linear offset of request.
> + */
> +static size_t ublk_copy_user_pages(const struct request *req,
> + struct iov_iter *uiter, int dir)
> {
> - const unsigned int gup_flags = to_vm ? FOLL_WRITE : 0;
> - const unsigned long start_vm = data->ubuf;
> - unsigned int done = 0;
> struct ublk_io_iter iter = {
> - .pg_off = start_vm & (PAGE_SIZE - 1),
> - .bio = data->rq->bio,
> - .iter = data->rq->bio->bi_iter,
> + .bio = req->bio,
> + .iter = req->bio->bi_iter,
> };
> - const unsigned int nr_pages = round_up(data->len +
> - (start_vm & (PAGE_SIZE - 1)), PAGE_SIZE) >> PAGE_SHIFT;
> -
> - while (done < nr_pages) {
> - const unsigned to_pin = min_t(unsigned, UBLK_MAX_PIN_PAGES,
> - nr_pages - done);
> - unsigned i, len;
> -
> - iter.nr_pages = get_user_pages_fast(start_vm +
> - (done << PAGE_SHIFT), to_pin, gup_flags,
> - iter.pages);
> - if (iter.nr_pages <= 0)
> - return done == 0 ? iter.nr_pages : done;
> - len = ublk_copy_io_pages(&iter, data->len, to_vm);
> - for (i = 0; i < iter.nr_pages; i++) {
> - if (to_vm)
> + size_t done = 0;
> +
> + while (iov_iter_count(uiter) && iter.bio) {
> + unsigned nr_pages;
> + size_t len, off;
> + int i;
> +
> + len = iov_iter_get_pages2(uiter, iter.pages,
> + iov_iter_count(uiter),
> + UBLK_MAX_PIN_PAGES, &off);
> + if (len <= 0)
> + return done;
> +
> + ublk_copy_io_pages(&iter, len, off, dir);
> + nr_pages = DIV_ROUND_UP(len + off, PAGE_SIZE);
> + for (i = 0; i < nr_pages; i++) {
> + if (dir == ITER_DEST)
> set_page_dirty(iter.pages[i]);
> put_page(iter.pages[i]);
> }
> - data->len -= len;
> - done += iter.nr_pages;
> + done += len;
> }
>
> return done;
> @@ -536,15 +525,14 @@ static int ublk_map_io(const struct ublk_queue *ubq, const struct request *req,
> * context is pretty fast, see ublk_pin_user_pages
> */
> if (ublk_need_map_req(req)) {
> - struct ublk_map_data data = {
> - .rq = req,
> - .ubuf = io->addr,
> - .len = rq_bytes,
> - };
> + struct iov_iter iter;
> + struct iovec iov;
> + const int dir = ITER_DEST;
Maybe a comment here that this means "copy to daemon"?
>
> - ublk_copy_user_pages(&data, true);
> + import_single_range(dir, u64_to_user_ptr(io->addr), rq_bytes,
> + &iov, &iter);
>
> - return rq_bytes - data.len;
> + return ublk_copy_user_pages(req, &iter, dir);
> }
> return rq_bytes;
> }
> @@ -556,17 +544,15 @@ static int ublk_unmap_io(const struct ublk_queue *ubq,
> const unsigned int rq_bytes = blk_rq_bytes(req);
>
> if (ublk_need_unmap_req(req)) {
> - struct ublk_map_data data = {
> - .rq = req,
> - .ubuf = io->addr,
> - .len = io->res,
> - };
> + struct iov_iter iter;
> + struct iovec iov;
> + const int dir = ITER_SOURCE;
And here "from daemon"?
>
> WARN_ON_ONCE(io->res > rq_bytes);
>
> - ublk_copy_user_pages(&data, false);
> -
> - return io->res - data.len;
> + import_single_range(dir, u64_to_user_ptr(io->addr), io->res,
> + &iov, &iter);
> + return ublk_copy_user_pages(req, &iter, dir);
> }
> return rq_bytes;
> }
next prev parent reply other threads:[~2023-03-31 16:28 UTC|newest]
Thread overview: 37+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-03-30 11:36 [PATCH V6 00/17] io_uring/ublk: add generic IORING_OP_FUSED_CMD Ming Lei
2023-03-30 11:36 ` [PATCH V6 01/17] io_uring: increase io_kiocb->flags into 64bit Ming Lei
2023-03-30 11:36 ` [PATCH V6 02/17] io_uring: use ctx->cached_sq_head to calculate left sqes Ming Lei
2023-03-30 11:36 ` [PATCH V6 03/17] io_uring: add generic IORING_OP_FUSED_CMD Ming Lei
2023-04-01 14:35 ` Ming Lei
2023-03-30 11:36 ` [PATCH V6 04/17] io_uring: support providing buffer by IORING_OP_FUSED_CMD Ming Lei
2023-03-30 11:36 ` [PATCH V6 05/17] io_uring: support OP_READ/OP_WRITE for fused secondary request Ming Lei
2023-03-30 11:36 ` [PATCH V6 06/17] io_uring: support OP_SEND_ZC/OP_RECV " Ming Lei
2023-03-30 11:36 ` [PATCH V6 07/17] block: ublk_drv: add common exit handling Ming Lei
2023-03-30 11:36 ` [PATCH V6 08/17] block: ublk_drv: don't consider flush request in map/unmap io Ming Lei
2023-03-30 11:36 ` [PATCH V6 09/17] block: ublk_drv: add two helpers to clean up map/unmap request Ming Lei
2023-03-30 11:36 ` [PATCH V6 10/17] block: ublk_drv: clean up several helpers Ming Lei
2023-03-30 11:36 ` [PATCH V6 11/17] block: ublk_drv: cleanup 'struct ublk_map_data' Ming Lei
2023-03-30 11:36 ` [PATCH V6 12/17] block: ublk_drv: cleanup ublk_copy_user_pages Ming Lei
2023-03-31 16:22 ` Bernd Schubert [this message]
2023-03-30 11:36 ` [PATCH V6 13/17] block: ublk_drv: grab request reference when the request is handled by userspace Ming Lei
2023-03-30 11:36 ` [PATCH V6 14/17] block: ublk_drv: support to copy any part of request pages Ming Lei
2023-03-30 11:36 ` [PATCH V6 15/17] block: ublk_drv: add read()/write() support for ublk char device Ming Lei
2023-03-30 11:36 ` [PATCH V6 16/17] block: ublk_drv: don't check buffer in case of zero copy Ming Lei
2023-03-30 11:36 ` [PATCH V6 17/17] block: ublk_drv: apply io_uring FUSED_CMD for supporting " Ming Lei
2023-03-31 19:13 ` Bernd Schubert
2023-04-01 13:19 ` Ming Lei
2023-03-31 19:55 ` Bernd Schubert
2023-04-01 13:22 ` Ming Lei
2023-04-03 9:25 ` Bernd Schubert
2023-04-03 1:11 ` [PATCH V6 00/17] io_uring/ublk: add generic IORING_OP_FUSED_CMD Ming Lei
2023-04-03 1:24 ` Jens Axboe
2023-04-04 7:48 ` Ming Lei
2023-04-03 1:23 ` (subset) " Jens Axboe
2023-04-18 19:38 ` Bernd Schubert
2023-04-19 1:51 ` Ming Lei
2023-04-19 9:56 ` Bernd Schubert
2023-04-19 11:19 ` Ming Lei
2023-04-19 15:42 ` Bernd Schubert
2023-04-20 1:18 ` Pavel Begunkov
2023-04-20 1:38 ` Ming Lei
2023-04-21 22:38 ` Bernd Schubert
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox