public inbox for [email protected]
 help / color / mirror / Atom feed
From: Kanchan Joshi <[email protected]>
To: Ferry Meng <[email protected]>
Cc: [email protected], [email protected],
	[email protected], [email protected]
Subject: Re: [bug report] nvme passthrough: request failed when blocksize exceeding max_hw_sectors
Date: Mon, 26 Jun 2023 19:13:29 +0530	[thread overview]
Message-ID: <20230626134250.GA30651@green245> (raw)
In-Reply-To: <[email protected]>

[-- Attachment #1: Type: text/plain, Size: 3610 bytes --]

On Mon, Jun 26, 2023 at 05:09:59PM +0800, Ferry Meng wrote:
>Hello:
>
>I'm testing the io-uring nvme passthrough via fio. But I have 
>encountered the following issue:
>When I specify 'blocksize' exceeding 128KB (actually the maximum size 
>per request can send 'max_sectors_kb'), the creation of request failed 
>and directly returned -22 (-EINVAL).
>
>For example:
>
># cat fio.job
>
>    [global]
>    ioengine=io_uring_cmd
>    thread=1
>    time_based
>    numjobs=1
>    iodepth=1
>    runtime=120
>    rw=randwrite
>    cmd_type=nvme
>    hipri=1
>
>    [randwrite]
>    bs=132k
>    filename=/dev/ng1n1
>
># fio fio.job
>randwrite: (g=0): rw=randwrite, bs=(R) 132KiB-132KiB, (W) 
>132KiB-132KiB, (T) 132KiB-132KiB, ioengine=io_uring_cmd, iodepth=1
>fio-3.34-10-g2fa0-dirty
>Starting 1 thread
>fio: io_u error on file /dev/ng1n1: Invalid argument: write 
>offset=231584956416, buflen=135168
>fio: pid=148989, err=22/file:io_u.c:1889, func=io_u error, 
>error=Invalid argument
>
>I tracked the position that returns the error val in kernel and dumped 
>calltrace.
>
>[   83.352715] nvme nvme1: 15/0/1 default/read/poll queues
>[   83.363273] nvme nvme1: Ignoring bogus Namespace Identifiers
>[   91.578457] CPU: 14 PID: 3993 Comm: fio Not tainted 
>6.4.0-rc7-00014-g692b7dc87ca6-dirty #2
>[   91.578462] Hardware name: Alibaba Cloud Alibaba Cloud ECS, BIOS 
>2221b89 04/01/2014
>[   91.578463] Call Trace:
>[   91.578476]  <TASK>
>[   91.578478]  dump_stack_lvl+0x36/0x50
>[   91.578484]  ll_back_merge_fn+0x20d/0x320
>[   91.578490]  blk_rq_append_bio+0x6d/0xc0
>[   91.578492]  bio_map_user_iov+0x24a/0x3d0
>[   91.578494]  blk_rq_map_user_iov+0x292/0x680
>[   91.578496]  ? blk_mq_get_tag+0x249/0x280
>[   91.578500]  blk_rq_map_user+0x56/0x80
>[   91.578503]  nvme_map_user_request.isra.15+0x90/0x1e0 [nvme_core]
>[   91.578515]  nvme_uring_cmd_io+0x29d/0x2f0 [nvme_core]
>[   91.578522]  io_uring_cmd+0x89/0x110
>[   91.578526]  ? __pfx_io_uring_cmd+0x10/0x10
>[   91.578528]  io_issue_sqe+0x1e0/0x2d0
>[   91.578530]  io_submit_sqes+0x1e3/0x650
>[   91.578532]  __x64_sys_io_uring_enter+0x2da/0x450
>[   91.578534]  do_syscall_64+0x3b/0x90
>[   91.578537]  entry_SYSCALL_64_after_hwframe+0x72/0xdc
>
>Here in bio_map_user_iov()->blk_rq_append_bio(), I found the error val 
>-EINVAL:
>
>blk_rq_append_bio:
>    ...
>    if (!ll_back_merge_fn(rq, bio, nr_segs))
>        return -EINVAL;
>    rq->biotail->bi_next = bio;
>    ...
>
>And in ll_back_merge_fn(), returns 0 if merge can't happen. It checks 
>the request size:
>ll_back_merge_fn:
>    if (blk_rq_sectors(req) + bio_sectors(bio) >
>        blk_rq_get_max_sectors(req, blk_rq_pos(req))) {
>            req_set_nomerge(req->q, req);
>            return 0;
>    }
>
>The ROOT cause is: In blk_rq_get_max_sectors, it returns 
>'max_hw_sectors' directly(in my device ,it's 256 sector, which means 
>128KB), causing the above inequality to hold true.
>blk_rq_get_max_sectors:
>    ...
>    if (blk_rq_is_passthrough(rq)){
>        return q->limits.max_hw_sectors;
>    }
>    ...
>
>I checked my disk's specs(cat 
>/sys/block/<mydisk>/queue/max_hw_sectors_kb 
>/sys/block/<mydisk>/queue/max_sectors_kb), both are 128KB.So I think 
>this arg causing the issue.
>
>I'm not sure if this is a designed restriction. Or should I have to 
>take care of it in application?

Right, passthrough interface does not abstract the device limits.
This needs to be handled in application.

[-- Attachment #2: Type: text/plain, Size: 0 bytes --]



      reply	other threads:[~2023-06-26 13:46 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <CGME20230626091231epcas5p48cbbb13c9579da9b11d3409c66f8ba71@epcas5p4.samsung.com>
2023-06-26  9:09 ` [bug report] nvme passthrough: request failed when blocksize exceeding max_hw_sectors Ferry Meng
2023-06-26 13:43   ` Kanchan Joshi [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20230626134250.GA30651@green245 \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox