public inbox for [email protected]
 help / color / mirror / Atom feed
From: Ferry Meng <[email protected]>
To: [email protected], [email protected]
Cc: [email protected]
Subject: Problem about sq->khead update and ring full judgement
Date: Fri, 3 Nov 2023 02:43:49 +0800	[thread overview]
Message-ID: <[email protected]> (raw)

Hi all:
     I'm using io_uring in a program, with SQPOLL feature enabled. The 
userspace program will actively count the queue status of urings, the 
programming model is similar to:
     {
         sqe = io_uring_get_sqe();
         if(sqe){
             /* prepare next request */
             queue_count++;
     }


     {
         cqe = io_uring_peek_cqe();
         if(cqe){
             queue_count--;
         }
     }

     In this way, maybe we can assume that " sq_ring_size - queue_count 
= sqe_left_we_can_alloc "?

     Userspace program will compare queue_count with io_uring's sq_size 
: if queue is not full ( queue_count < sq_size), it will try getting new 
sqe(means Initiating a new IO request).

     Now I'm currently coming into a situation where  I/O is very high 
—— Userspace program submit lots of sqes (around 2000) at the same time, 
meanwhile  sq_ring_size is 4096. In kernel, 
__io_sq_thread->io_submit_sqes(), I see that nr(to_submit) is also over 
2000.   At a point, a strange point comes out: Userspace program find 
sq_ring is not full, but Kernel(in fact liburing::io_uring_get_sqe) 
think sq_ring is full.

     After analyzing, I find the reason is: kernel update "sq->khead" 
after submitting "all" sqes. The running of my program is : Before 
kernel update khead, userspace program has received many cqes, causing 
queue_count-- . After decreasing queue_count, user-program thinks 
sq_ring is not full, and try to start new IO requeust. As sq->head is 
not updated, io_uring_get_sqe() returns NULL.

     My questions are:

     1. Is userspace 'queue_count' judgement reasonable? From 
'traditional' point of view, if we want to find sq_ring full or not, we 
can just use io_uring_get_sqe() to check. Maybe this discussion is 
similar to this issue in a way: https://github.com/axboe/liburing/issues/88

     2. I must confess that it's very strange that cqe's average 
receiving latency is shorter than the consumption time of sqe(Now it 
really happens and I'm trying to find why this happened, or it's a just 
some bug). Assuming this scenario is reasonable, it seems that we can 
update sq->khead more often to get higher throughput. When userspace 
gets more 'sensetive', it can send out more IO requests. And it won't 
cause much more overhead if kernel atomically update 'khead' 
a-little-bit more.





             reply	other threads:[~2023-11-02 18:43 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-11-02 18:43 Ferry Meng [this message]
2023-11-02 19:13 ` Problem about sq->khead update and ring full judgement Gabriel Krisman Bertazi

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=7feaba7c-b93c-a136-6438-1de365b5a02a@linux.alibaba.com \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox