From: Bernd Schubert <bernd@bsbernd.com>
To: Joanne Koong <joannelkoong@gmail.com>,
axboe@kernel.dk, miklos@szeredi.hu
Cc: csander@purestorage.com, krisman@suse.de,
io-uring@vger.kernel.org, asml.silence@gmail.com,
xiaobing.li@samsung.com, safinaskar@gmail.com,
linux-fsdevel@vger.kernel.org
Subject: Re: [PATCH v4 19/25] fuse: add io-uring kernel-managed buffer ring
Date: Wed, 28 Jan 2026 22:44:01 +0100 [thread overview]
Message-ID: <2f14fb1a-0ee2-4d86-98be-ed6112ed706d@bsbernd.com> (raw)
In-Reply-To: <20260116233044.1532965-20-joannelkoong@gmail.com>
On 1/17/26 00:30, Joanne Koong wrote:
> Add io-uring kernel-managed buffer ring capability for fuse daemons
> communicating through the io-uring interface.
>
> This has two benefits:
> a) eliminates the overhead of pinning/unpinning user pages and
> translating virtual addresses for every server-kernel interaction
>
> b) reduces the amount of memory needed for the buffers per queue and
> allows buffers to be reused across entries. Incremental buffer
> consumption, when added, will allow a buffer to be used across multiple
> requests.
>
> Buffer ring usage is set on a per-queue basis. In order to use this, the
> daemon needs to have preregistered a kernel-managed buffer ring and a
> fixed buffer at index 0 that will hold all the headers, and set the
> "use_bufring" field during registration. The kernel-managed buffer ring
> will be pinned for the lifetime of the connection.
>
> Signed-off-by: Joanne Koong <joannelkoong@gmail.com>
> ---
> fs/fuse/dev_uring.c | 412 ++++++++++++++++++++++++++++++++------
> fs/fuse/dev_uring_i.h | 31 ++-
> include/uapi/linux/fuse.h | 15 +-
> 3 files changed, 389 insertions(+), 69 deletions(-)
>
> diff --git a/fs/fuse/dev_uring.c b/fs/fuse/dev_uring.c
> index b57871f92d08..40e8c2e6b77c 100644
> --- a/fs/fuse/dev_uring.c
> +++ b/fs/fuse/dev_uring.c
> @@ -10,6 +10,8 @@
> #include "fuse_trace.h"
>
> #include <linux/fs.h>
> +#include <linux/io_uring.h>
> +#include <linux/io_uring/buf.h>
> #include <linux/io_uring/cmd.h>
>
> static bool __read_mostly enable_uring;
> @@ -19,6 +21,8 @@ MODULE_PARM_DESC(enable_uring,
>
> #define FUSE_URING_IOV_SEGS 2 /* header and payload */
>
> +#define FUSE_URING_RINGBUF_GROUP 0
> +#define FUSE_URING_FIXED_HEADERS_OFFSET 0
>
> bool fuse_uring_enabled(void)
> {
> @@ -276,20 +280,45 @@ static struct fuse_ring *fuse_uring_create(struct fuse_conn *fc)
> return res;
> }
>
> -static struct fuse_ring_queue *fuse_uring_create_queue(struct fuse_ring *ring,
> - int qid)
> +static int fuse_uring_buf_ring_setup(struct io_uring_cmd *cmd,
> + struct fuse_ring_queue *queue,
> + unsigned int issue_flags)
> +{
> + int err;
> +
> + err = io_uring_buf_ring_pin(cmd, FUSE_URING_RINGBUF_GROUP, issue_flags,
> + &queue->bufring);
> + if (err)
> + return err;
> +
> + if (!io_uring_is_kmbuf_ring(cmd, FUSE_URING_RINGBUF_GROUP,
> + issue_flags)) {
> + io_uring_buf_ring_unpin(cmd, FUSE_URING_RINGBUF_GROUP,
> + issue_flags);
> + return -EINVAL;
> + }
> +
> + queue->use_bufring = true;
> +
> + return 0;
> +}
> +
> +static struct fuse_ring_queue *
> +fuse_uring_create_queue(struct io_uring_cmd *cmd, struct fuse_ring *ring,
> + int qid, bool use_bufring, unsigned int issue_flags)
> {
> struct fuse_conn *fc = ring->fc;
> struct fuse_ring_queue *queue;
> struct list_head *pq;
> + int err;
>
> queue = kzalloc(sizeof(*queue), GFP_KERNEL_ACCOUNT);
> if (!queue)
> - return NULL;
> + return ERR_PTR(-ENOMEM);
> pq = kcalloc(FUSE_PQ_HASH_SIZE, sizeof(struct list_head), GFP_KERNEL);
> if (!pq) {
> kfree(queue);
> - return NULL;
> + return ERR_PTR(-ENOMEM);
> }
>
> queue->qid = qid;
> @@ -307,6 +336,15 @@ static struct fuse_ring_queue *fuse_uring_create_queue(struct fuse_ring *ring,
> queue->fpq.processing = pq;
> fuse_pqueue_init(&queue->fpq);
>
> + if (use_bufring) {
> + err = fuse_uring_buf_ring_setup(cmd, queue, issue_flags);
> + if (err) {
> + kfree(pq);
> + kfree(queue);
> + return ERR_PTR(err);
> + }
> + }
> +
> spin_lock(&fc->lock);
> if (ring->queues[qid]) {
> spin_unlock(&fc->lock);
> @@ -584,6 +622,35 @@ static int fuse_uring_out_header_has_err(struct fuse_out_header *oh,
> return err;
> }
>
> +static int get_kernel_ring_header(struct fuse_ring_ent *ent,
> + enum fuse_uring_header_type type,
> + struct iov_iter *headers_iter)
> +{
> + size_t offset;
> +
> + switch (type) {
> + case FUSE_URING_HEADER_IN_OUT:
> + /* No offset - start of header */
> + offset = 0;
> + break;
> + case FUSE_URING_HEADER_OP:
> + offset = offsetof(struct fuse_uring_req_header, op_in);
> + break;
> + case FUSE_URING_HEADER_RING_ENT:
> + offset = offsetof(struct fuse_uring_req_header, ring_ent_in_out);
> + break;
> + default:
> + WARN_ONCE(1, "Invalid header type: %d\n", type);
> + return -EINVAL;
> + }
> +
> + *headers_iter = ent->headers_iter;
> + if (offset)
> + iov_iter_advance(headers_iter, offset);
> +
> + return 0;
> +}
> +
> static void __user *get_user_ring_header(struct fuse_ring_ent *ent,
> enum fuse_uring_header_type type)
> {
> @@ -605,17 +672,38 @@ static __always_inline int copy_header_to_ring(struct fuse_ring_ent *ent,
> const void *header,
> size_t header_size)
> {
> - void __user *ring = get_user_ring_header(ent, type);
> + bool use_bufring = ent->queue->use_bufring;
> + int err = 0;
>
> - if (!ring)
> - return -EINVAL;
> + if (use_bufring) {
> + struct iov_iter iter;
> +
> + err = get_kernel_ring_header(ent, type, &iter);
> + if (err)
> + goto done;
> +
> + if (copy_to_iter(header, header_size, &iter) != header_size)
> + err = -EFAULT;
> + } else {
> + void __user *ring = get_user_ring_header(ent, type);
> +
> + if (!ring) {
> + err = -EINVAL;
> + goto done;
> + }
>
> - if (copy_to_user(ring, header, header_size)) {
> - pr_info_ratelimited("Copying header to ring failed.\n");
> - return -EFAULT;
> + if (copy_to_user(ring, header, header_size))
> + err = -EFAULT;
> }
>
> - return 0;
> +done:
> + if (err)
> + pr_info_ratelimited("Copying header to ring failed: "
> + "header_type=%u, header_size=%zu, "
> + "use_bufring=%d\n", type, header_size,
> + use_bufring);
> +
> + return err;
> }
>
> static __always_inline int copy_header_from_ring(struct fuse_ring_ent *ent,
> @@ -623,17 +711,38 @@ static __always_inline int copy_header_from_ring(struct fuse_ring_ent *ent,
> void *header,
> size_t header_size)
> {
> - const void __user *ring = get_user_ring_header(ent, type);
> + bool use_bufring = ent->queue->use_bufring;
> + int err = 0;
>
> - if (!ring)
> - return -EINVAL;
> + if (use_bufring) {
> + struct iov_iter iter;
> +
> + err = get_kernel_ring_header(ent, type, &iter);
> + if (err)
> + goto done;
> +
> + if (copy_from_iter(header, header_size, &iter) != header_size)
> + err = -EFAULT;
> + } else {
> + const void __user *ring = get_user_ring_header(ent, type);
>
> - if (copy_from_user(header, ring, header_size)) {
> - pr_info_ratelimited("Copying header from ring failed.\n");
> - return -EFAULT;
> + if (!ring) {
> + err = -EINVAL;
> + goto done;
> + }
> +
> + if (copy_from_user(header, ring, header_size))
> + err = -EFAULT;
> }
>
> - return 0;
> +done:
> + if (err)
> + pr_info_ratelimited("Copying header from ring failed: "
> + "header_type=%u, header_size=%zu, "
> + "use_bufring=%d\n", type, header_size,
> + use_bufring);
> +
> + return err;
> }
>
> static int setup_fuse_copy_state(struct fuse_copy_state *cs,
> @@ -643,14 +752,23 @@ static int setup_fuse_copy_state(struct fuse_copy_state *cs,
> {
> int err;
>
> - err = import_ubuf(dir, ent->payload, ring->max_payload_sz, iter);
> - if (err) {
> - pr_info_ratelimited("fuse: Import of user buffer failed\n");
> - return err;
> + if (!ent->queue->use_bufring) {
> + err = import_ubuf(dir, ent->payload, ring->max_payload_sz, iter);
> + if (err) {
> + pr_info_ratelimited("fuse: Import of user buffer "
> + "failed\n");
> + return err;
> + }
> }
>
> fuse_copy_init(cs, dir == ITER_DEST, iter);
>
> + if (ent->queue->use_bufring) {
> + cs->is_kaddr = true;
> + cs->len = ent->payload_kvec.iov_len;
> + cs->kaddr = ent->payload_kvec.iov_base;
> + }
> +
> cs->is_uring = true;
> cs->req = req;
>
> @@ -762,6 +880,94 @@ static int fuse_uring_copy_to_ring(struct fuse_ring_ent *ent,
> sizeof(req->in.h));
> }
>
> +static bool fuse_uring_req_has_payload(struct fuse_req *req)
> +{
> + struct fuse_args *args = req->args;
> +
> + return args->in_numargs > 1 || args->out_numargs;
> +}
> +
> +static int fuse_uring_select_buffer(struct fuse_ring_ent *ent,
> + unsigned int issue_flags)
> + __must_hold(&queue->lock)
> +{
> + struct io_br_sel sel;
> + size_t len = 0;
> +
> + lockdep_assert_held(&ent->queue->lock);
> +
> + /* Get a buffer to use for the payload */
> + sel = io_ring_buffer_select(cmd_to_io_kiocb(ent->cmd), &len,
> + ent->queue->bufring, issue_flags);
> + if (sel.val)
> + return sel.val;
> + if (!sel.kaddr)
> + return -ENOENT;
> +
> + ent->payload_kvec.iov_base = sel.kaddr;
> + ent->payload_kvec.iov_len = len;
> + ent->ringbuf_buf_id = sel.buf_id;
> +
> + return 0;
> +}
> +
> +static void fuse_uring_clean_up_buffer(struct fuse_ring_ent *ent,
> + unsigned int issue_flags)
> + __must_hold(&queue->lock)
> +{
> + struct kvec *kvec = &ent->payload_kvec;
> +
> + lockdep_assert_held(&ent->queue->lock);
> +
> + if (!ent->queue->use_bufring || !kvec->iov_base)
> + return;
> +
> + WARN_ON_ONCE(io_uring_kmbuf_recycle(ent->cmd, FUSE_URING_RINGBUF_GROUP,
> + (u64)kvec->iov_base, kvec->iov_len,
> + ent->ringbuf_buf_id, issue_flags));
> +
> + memset(kvec, 0, sizeof(*kvec));
> +}
> +
> +static int fuse_uring_next_req_update_buffer(struct fuse_ring_ent *ent,
> + struct fuse_req *req,
> + unsigned int issue_flags)
> +{
> + bool buffer_selected;
> + bool has_payload;
> +
> + if (!ent->queue->use_bufring)
> + return 0;
> +
> + ent->headers_iter.data_source = false;
> +
> + buffer_selected = ent->payload_kvec.iov_base != NULL;
> + has_payload = fuse_uring_req_has_payload(req);
> +
> + if (has_payload && !buffer_selected)
> + return fuse_uring_select_buffer(ent, issue_flags);
> +
> + if (!has_payload && buffer_selected)
> + fuse_uring_clean_up_buffer(ent, issue_flags);
> +
> + return 0;
> +}
> +
> +static int fuse_uring_prep_buffer(struct fuse_ring_ent *ent,
> + struct fuse_req *req, unsigned issue_flags)
> +{
> + if (!ent->queue->use_bufring)
> + return 0;
> +
> + ent->headers_iter.data_source = false;
> +
> + /* no payload to copy, can skip selecting a buffer */
> + if (!fuse_uring_req_has_payload(req))
> + return 0;
> +
> + return fuse_uring_select_buffer(ent, issue_flags);
> +}
> +
> static int fuse_uring_prepare_send(struct fuse_ring_ent *ent,
> struct fuse_req *req)
> {
> @@ -824,21 +1030,29 @@ static void fuse_uring_add_req_to_ring_ent(struct fuse_ring_ent *ent,
> }
>
> /* Fetch the next fuse request if available */
> -static struct fuse_req *fuse_uring_ent_assign_req(struct fuse_ring_ent *ent)
> +static struct fuse_req *fuse_uring_ent_assign_req(struct fuse_ring_ent *ent,
> + unsigned int issue_flags)
> __must_hold(&queue->lock)
> {
> struct fuse_req *req;
> struct fuse_ring_queue *queue = ent->queue;
> struct list_head *req_queue = &queue->fuse_req_queue;
> + int err;
>
> lockdep_assert_held(&queue->lock);
>
> /* get and assign the next entry while it is still holding the lock */
> req = list_first_entry_or_null(req_queue, struct fuse_req, list);
> - if (req)
> - fuse_uring_add_req_to_ring_ent(ent, req);
> + if (req) {
> + err = fuse_uring_next_req_update_buffer(ent, req, issue_flags);
> + if (!err) {
> + fuse_uring_add_req_to_ring_ent(ent, req);
> + return req;
> + }
> + }
>
> - return req;
> + fuse_uring_clean_up_buffer(ent, issue_flags);
> + return NULL;
> }
>
> /*
> @@ -878,7 +1092,8 @@ static void fuse_uring_commit(struct fuse_ring_ent *ent, struct fuse_req *req,
> * Else, there is no next fuse request and this returns false.
> */
> static bool fuse_uring_get_next_fuse_req(struct fuse_ring_ent *ent,
> - struct fuse_ring_queue *queue)
> + struct fuse_ring_queue *queue,
> + unsigned int issue_flags)
> {
> int err;
> struct fuse_req *req;
> @@ -886,7 +1101,7 @@ static bool fuse_uring_get_next_fuse_req(struct fuse_ring_ent *ent,
> retry:
> spin_lock(&queue->lock);
> fuse_uring_ent_avail(ent, queue);
> - req = fuse_uring_ent_assign_req(ent);
> + req = fuse_uring_ent_assign_req(ent, issue_flags);
> spin_unlock(&queue->lock);
>
> if (req) {
> @@ -927,6 +1142,39 @@ static void fuse_uring_send(struct fuse_ring_ent *ent, struct io_uring_cmd *cmd,
> io_uring_cmd_done(cmd, ret, issue_flags);
> }
>
> +static void fuse_uring_headers_cleanup(struct fuse_ring_ent *ent,
> + unsigned int issue_flags)
> +{
> + if (!ent->queue->use_bufring || !ent->headers_node)
> + return;
> +
> + io_uring_fixed_index_put(ent->cmd, ent->headers_node, issue_flags);
> + ent->headers_node = NULL;
> +}
> +
> +static int fuse_uring_headers_prep(struct fuse_ring_ent *ent, unsigned int dir,
> + unsigned int issue_flags)
> +{
> + size_t header_size = sizeof(struct fuse_uring_req_header);
> + struct io_uring_cmd *cmd = ent->cmd;
> + struct io_rsrc_node *node;
> + unsigned int offset;
> +
> + if (!ent->queue->use_bufring)
> + return 0;
> +
> + offset = ent->fixed_buf_id * header_size;
> +
> + node = io_uring_fixed_index_get(cmd, FUSE_URING_FIXED_HEADERS_OFFSET,
> + offset, header_size, dir,
> + &ent->headers_iter, issue_flags);
> + if (IS_ERR(node))
> + return PTR_ERR(node);
> +
> + ent->headers_node = node;
> + return 0;
> +}
> +
> /* FUSE_URING_CMD_COMMIT_AND_FETCH handler */
> static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags,
> struct fuse_conn *fc)
> @@ -940,6 +1188,7 @@ static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags,
> unsigned int qid = READ_ONCE(cmd_req->qid);
> struct fuse_pqueue *fpq;
> struct fuse_req *req;
> + bool send;
>
> err = -ENOTCONN;
> if (!ring)
> @@ -990,7 +1239,12 @@ static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags,
>
> /* without the queue lock, as other locks are taken */
> fuse_uring_prepare_cancel(cmd, issue_flags, ent);
> - fuse_uring_commit(ent, req, issue_flags);
> +
> + err = fuse_uring_headers_prep(ent, ITER_SOURCE, issue_flags);
> + if (err)
> + fuse_uring_req_end(ent, req, err);
> + else
> + fuse_uring_commit(ent, req, issue_flags);
>
> /*
> * Fetching the next request is absolutely required as queued
> @@ -998,7 +1252,9 @@ static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags,
> * and fetching is done in one step vs legacy fuse, which has separated
> * read (fetch request) and write (commit result).
> */
> - if (fuse_uring_get_next_fuse_req(ent, queue))
> + send = fuse_uring_get_next_fuse_req(ent, queue, issue_flags);
> + fuse_uring_headers_cleanup(ent, issue_flags);
> + if (send)
> fuse_uring_send(ent, cmd, 0, issue_flags);
> return 0;
> }
> @@ -1094,39 +1350,48 @@ fuse_uring_create_ring_ent(struct io_uring_cmd *cmd,
> struct iovec iov[FUSE_URING_IOV_SEGS];
> int err;
>
> + ent = kzalloc(sizeof(*ent), GFP_KERNEL_ACCOUNT);
> + if (!ent)
> + return ERR_PTR(-ENOMEM);
> +
> + INIT_LIST_HEAD(&ent->list);
> +
> + ent->queue = queue;
> +
> + if (queue->use_bufring) {
> + ent->fixed_buf_id = READ_ONCE(cmd->sqe->buf_index);
> + atomic_inc(&ring->queue_refs);
> + return ent;
> + }
> +
> err = fuse_uring_get_iovec_from_sqe(cmd->sqe, iov);
> if (err) {
> pr_info_ratelimited("Failed to get iovec from sqe, err=%d\n",
> err);
> - return ERR_PTR(err);
> + goto error;
> }
>
> err = -EINVAL;
> if (iov[0].iov_len < sizeof(struct fuse_uring_req_header)) {
> pr_info_ratelimited("Invalid header len %zu\n", iov[0].iov_len);
> - return ERR_PTR(err);
> + goto error;
> }
>
> payload_size = iov[1].iov_len;
> if (payload_size < ring->max_payload_sz) {
> pr_info_ratelimited("Invalid req payload len %zu\n",
> payload_size);
> - return ERR_PTR(err);
> + goto error;
> }
> -
> - err = -ENOMEM;
> - ent = kzalloc(sizeof(*ent), GFP_KERNEL_ACCOUNT);
> - if (!ent)
> - return ERR_PTR(err);
> -
> - INIT_LIST_HEAD(&ent->list);
> -
> - ent->queue = queue;
> ent->headers = iov[0].iov_base;
> ent->payload = iov[1].iov_base;
>
> atomic_inc(&ring->queue_refs);
> return ent;
> +
> +error:
> + kfree(ent);
> + return ERR_PTR(err);
> }
>
> /*
> @@ -1137,6 +1402,8 @@ static int fuse_uring_register(struct io_uring_cmd *cmd,
> unsigned int issue_flags, struct fuse_conn *fc)
> {
> const struct fuse_uring_cmd_req *cmd_req = io_uring_sqe_cmd(cmd->sqe);
> + unsigned int init_flags = READ_ONCE(cmd_req->init.flags);
> + bool use_bufring = init_flags & FUSE_URING_BUF_RING;
> struct fuse_ring *ring = smp_load_acquire(&fc->ring);
> struct fuse_ring_queue *queue;
> struct fuse_ring_ent *ent;
> @@ -1157,9 +1424,13 @@ static int fuse_uring_register(struct io_uring_cmd *cmd,
>
> queue = ring->queues[qid];
> if (!queue) {
> - queue = fuse_uring_create_queue(ring, qid);
> - if (!queue)
> - return err;
> + queue = fuse_uring_create_queue(cmd, ring, qid, use_bufring,
> + issue_flags);
> + if (IS_ERR(queue))
> + return PTR_ERR(queue);
> + } else {
> + if (queue->use_bufring != use_bufring)
> + return -EINVAL;
> }
>
> /*
> @@ -1258,15 +1529,19 @@ static void fuse_uring_send_in_task(struct io_tw_req tw_req, io_tw_token_t tw)
> struct io_uring_cmd *cmd = io_uring_cmd_from_tw(tw_req);
> struct fuse_ring_ent *ent = uring_cmd_to_ring_ent(cmd);
> struct fuse_ring_queue *queue = ent->queue;
> + bool send = true;
> int err;
>
> if (!tw.cancel) {
> - err = fuse_uring_prepare_send(ent, ent->fuse_req);
> - if (err) {
> - if (!fuse_uring_get_next_fuse_req(ent, queue))
> - return;
> - err = 0;
> - }
> + if (fuse_uring_headers_prep(ent, ITER_DEST, issue_flags))
> + return;
> +
> + if (fuse_uring_prepare_send(ent, ent->fuse_req))
> + send = fuse_uring_get_next_fuse_req(ent, queue, issue_flags);
> + fuse_uring_headers_cleanup(ent, issue_flags);
> + if (!send)
> + return;
> + err = 0;
> } else {
> err = -ECANCELED;
> }
> @@ -1325,14 +1600,19 @@ void fuse_uring_queue_fuse_req(struct fuse_iqueue *fiq, struct fuse_req *req)
> req->ring_queue = queue;
> ent = list_first_entry_or_null(&queue->ent_avail_queue,
> struct fuse_ring_ent, list);
> - if (ent)
> - fuse_uring_add_req_to_ring_ent(ent, req);
> - else
> - list_add_tail(&req->list, &queue->fuse_req_queue);
> - spin_unlock(&queue->lock);
> + if (ent) {
> + err = fuse_uring_prep_buffer(ent, req, IO_URING_F_UNLOCKED);
> + if (!err) {
> + fuse_uring_add_req_to_ring_ent(ent, req);
> + spin_unlock(&queue->lock);
> + fuse_uring_dispatch_ent(ent);
> + return;
> + }
> + WARN_ON_ONCE(err != -ENOENT);
> + }
>
> - if (ent)
> - fuse_uring_dispatch_ent(ent);
> + list_add_tail(&req->list, &queue->fuse_req_queue);
> + spin_unlock(&queue->lock);
>
> return;
>
> @@ -1350,6 +1630,7 @@ bool fuse_uring_queue_bq_req(struct fuse_req *req)
> struct fuse_ring *ring = fc->ring;
> struct fuse_ring_queue *queue;
> struct fuse_ring_ent *ent = NULL;
> + int err;
>
> queue = fuse_uring_task_to_queue(ring);
> if (!queue)
> @@ -1382,14 +1663,15 @@ bool fuse_uring_queue_bq_req(struct fuse_req *req)
> req = list_first_entry_or_null(&queue->fuse_req_queue, struct fuse_req,
> list);
> if (ent && req) {
> - fuse_uring_add_req_to_ring_ent(ent, req);
> - spin_unlock(&queue->lock);
> -
> - fuse_uring_dispatch_ent(ent);
> - } else {
> - spin_unlock(&queue->lock);
> + err = fuse_uring_prep_buffer(ent, req, IO_URING_F_UNLOCKED);
> + if (!err) {
> + fuse_uring_add_req_to_ring_ent(ent, req);
> + spin_unlock(&queue->lock);
> + fuse_uring_dispatch_ent(ent);
> + return true;
> + }
> }
> -
> + spin_unlock(&queue->lock);
> return true;
> }
>
> diff --git a/fs/fuse/dev_uring_i.h b/fs/fuse/dev_uring_i.h
> index 51a563922ce1..ac6da80c3d70 100644
> --- a/fs/fuse/dev_uring_i.h
> +++ b/fs/fuse/dev_uring_i.h
> @@ -7,6 +7,8 @@
> #ifndef _FS_FUSE_DEV_URING_I_H
> #define _FS_FUSE_DEV_URING_I_H
>
> +#include <linux/uio.h>
> +
> #include "fuse_i.h"
>
> #ifdef CONFIG_FUSE_IO_URING
> @@ -38,9 +40,26 @@ enum fuse_ring_req_state {
>
> /** A fuse ring entry, part of the ring queue */
> struct fuse_ring_ent {
> - /* userspace buffer */
> - struct fuse_uring_req_header __user *headers;
> - void __user *payload;
> + union {
> + /* queue->use_bufring == false */
> + struct {
> + /* userspace buffers */
> + struct fuse_uring_req_header __user *headers;
> + void __user *payload;
> + };
> + /* queue->use_bufring == true */
> + struct {
> + struct iov_iter headers_iter;
> + struct io_rsrc_node *headers_node;
> + struct kvec payload_kvec;
> + /*
> + * This needs to be tracked in order to properly recycle
> + * the buffer when done with it
> + */
> + unsigned int ringbuf_buf_id;
> + unsigned int fixed_buf_id;
> + };
> + };
>
> /* the ring queue that owns the request */
> struct fuse_ring_queue *queue;
> @@ -99,6 +118,12 @@ struct fuse_ring_queue {
> unsigned int active_background;
>
> bool stopped;
> +
> + /* true if kernel-managed buffer ring is used */
> + bool use_bufring: 1;
> +
> + /* synchronized by the queue lock */
> + struct io_buffer_list *bufring;
> };
>
> /**
> diff --git a/include/uapi/linux/fuse.h b/include/uapi/linux/fuse.h
> index c13e1f9a2f12..b49c8d3b9ab6 100644
> --- a/include/uapi/linux/fuse.h
> +++ b/include/uapi/linux/fuse.h
> @@ -240,6 +240,9 @@
> * - add FUSE_COPY_FILE_RANGE_64
> * - add struct fuse_copy_file_range_out
> * - add FUSE_NOTIFY_PRUNE
> + *
> + * 7.46
> + * - add fuse_uring_cmd_req init flags
> */
>
> #ifndef _LINUX_FUSE_H
> @@ -1294,6 +1297,9 @@ enum fuse_uring_cmd {
> FUSE_IO_URING_CMD_COMMIT_AND_FETCH = 2,
> };
>
> +/* fuse_uring_cmd_req init flags */
> +#define FUSE_URING_BUF_RING (1 << 0)
> +
> /**
> * In the 80B command area of the SQE.
> */
> @@ -1305,7 +1311,14 @@ struct fuse_uring_cmd_req {
>
> /* queue the command is for (queue index) */
> uint16_t qid;
> - uint8_t padding[6];
> +
> + union {
> + struct {
> + uint16_t flags;
> + } init;
> + };
> +
I won't manage to review everything of this patch today, but just
noticed this. There is already an unused flags, why don't you use that?
I had edded it exactly for such things.
> + uint8_t padding[4];
> };
>
> #endif /* _LINUX_FUSE_H */
Thanks,
Bernd
next prev parent reply other threads:[~2026-01-28 21:44 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-01-16 23:30 [PATCH v4 00/25] fuse/io-uring: add kernel-managed buffer rings and zero-copy Joanne Koong
2026-01-16 23:30 ` [PATCH v4 01/25] io_uring/kbuf: refactor io_buf_pbuf_register() logic into generic helpers Joanne Koong
2026-01-16 23:30 ` [PATCH v4 02/25] io_uring/kbuf: rename io_unregister_pbuf_ring() to io_unregister_buf_ring() Joanne Koong
2026-01-16 23:30 ` [PATCH v4 03/25] io_uring/kbuf: add support for kernel-managed buffer rings Joanne Koong
2026-01-16 23:30 ` [PATCH v4 04/25] io_uring/kbuf: add mmap " Joanne Koong
2026-01-16 23:30 ` [PATCH v4 05/25] io_uring/kbuf: support kernel-managed buffer rings in buffer selection Joanne Koong
2026-01-16 23:30 ` [PATCH v4 06/25] io_uring/kbuf: add buffer ring pinning/unpinning Joanne Koong
2026-01-16 23:30 ` [PATCH v4 07/25] io_uring/kbuf: add recycling for kernel managed buffer rings Joanne Koong
2026-01-16 23:30 ` [PATCH v4 08/25] io_uring: add io_uring_fixed_index_get() and io_uring_fixed_index_put() Joanne Koong
2026-01-22 21:02 ` Caleb Sander Mateos
2026-01-27 20:05 ` Joanne Koong
2026-01-16 23:30 ` [PATCH v4 09/25] io_uring/kbuf: add io_uring_is_kmbuf_ring() Joanne Koong
2026-01-16 23:30 ` [PATCH v4 10/25] io_uring/kbuf: export io_ring_buffer_select() Joanne Koong
2026-01-16 23:30 ` [PATCH v4 11/25] io_uring/kbuf: return buffer id in buffer selection Joanne Koong
2026-01-16 23:30 ` [PATCH v4 12/25] io_uring/cmd: set selected buffer index in __io_uring_cmd_done() Joanne Koong
2026-01-16 23:30 ` [PATCH v4 13/25] fuse: refactor io-uring logic for getting next fuse request Joanne Koong
2026-01-16 23:30 ` [PATCH v4 14/25] fuse: refactor io-uring header copying to ring Joanne Koong
2026-01-16 23:30 ` [PATCH v4 15/25] fuse: refactor io-uring header copying from ring Joanne Koong
2026-01-27 23:11 ` Bernd Schubert
2026-01-16 23:30 ` [PATCH v4 16/25] fuse: use enum types for header copying Joanne Koong
2026-01-16 23:30 ` [PATCH v4 17/25] fuse: refactor setting up copy state for payload copying Joanne Koong
2026-01-16 23:30 ` [PATCH v4 18/25] fuse: support buffer copying for kernel addresses Joanne Koong
2026-01-27 23:39 ` Bernd Schubert
2026-01-28 0:23 ` Joanne Koong
2026-01-28 21:14 ` Bernd Schubert
2026-01-29 1:16 ` Joanne Koong
2026-01-16 23:30 ` [PATCH v4 19/25] fuse: add io-uring kernel-managed buffer ring Joanne Koong
2026-01-17 5:28 ` kernel test robot
2026-01-28 21:44 ` Bernd Schubert [this message]
2026-01-29 1:15 ` Joanne Koong
2026-01-16 23:30 ` [PATCH v4 20/25] io_uring/rsrc: rename io_buffer_register_bvec()/io_buffer_unregister_bvec() Joanne Koong
2026-01-16 23:30 ` [PATCH v4 21/25] io_uring/rsrc: split io_buffer_register_request() logic Joanne Koong
2026-01-16 23:30 ` [PATCH v4 22/25] io_uring/rsrc: Allow buffer release callback to be optional Joanne Koong
2026-01-16 23:30 ` [PATCH v4 23/25] io_uring/rsrc: add io_buffer_register_bvec() Joanne Koong
2026-01-16 23:30 ` [PATCH v4 24/25] fuse: add zero-copy over io-uring Joanne Koong
2026-01-16 23:30 ` [PATCH v4 25/25] docs: fuse: add io-uring bufring and zero-copy documentation Joanne Koong
2026-01-27 20:12 ` [PATCH v4 00/25] fuse/io-uring: add kernel-managed buffer rings and zero-copy Joanne Koong
2026-01-27 22:44 ` Bernd Schubert
2026-01-27 23:27 ` Joanne Koong
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=2f14fb1a-0ee2-4d86-98be-ed6112ed706d@bsbernd.com \
--to=bernd@bsbernd.com \
--cc=asml.silence@gmail.com \
--cc=axboe@kernel.dk \
--cc=csander@purestorage.com \
--cc=io-uring@vger.kernel.org \
--cc=joannelkoong@gmail.com \
--cc=krisman@suse.de \
--cc=linux-fsdevel@vger.kernel.org \
--cc=miklos@szeredi.hu \
--cc=safinaskar@gmail.com \
--cc=xiaobing.li@samsung.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox