From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-9.8 required=3.0 tests=DKIM_SIGNED,DKIM_VALID, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E1177C35E0B for ; Tue, 25 Feb 2020 16:04:58 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id B9E872072D for ; Tue, 25 Feb 2020 16:04:58 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=kernel-dk.20150623.gappssmtp.com header.i=@kernel-dk.20150623.gappssmtp.com header.b="Vgeww3p/" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730616AbgBYQE6 (ORCPT ); Tue, 25 Feb 2020 11:04:58 -0500 Received: from mail-io1-f68.google.com ([209.85.166.68]:41351 "EHLO mail-io1-f68.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728051AbgBYQE6 (ORCPT ); Tue, 25 Feb 2020 11:04:58 -0500 Received: by mail-io1-f68.google.com with SMTP id m25so2376604ioo.8 for ; Tue, 25 Feb 2020 08:04:56 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=kernel-dk.20150623.gappssmtp.com; s=20150623; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=LCjWZY2DA9SGQMYB8874unldp4qBpu8sVqc5dXs8Kts=; b=Vgeww3p/Z4AA8YB6rkhWq1fHaBx3KGEwB3mUig9HeIxsdA0eTq98Mvk/MH5faWylOB xhE4ousxZxr/nRzuEk5Pvl6q+1RlzlLgeIBSmKVZ7RXj3abvk33JnXc/yMqeB/KoJbu8 oV50cGp95Et2JoKhAJoA2rLDy9E2e6Px+OJkf7xn/HSFP6cqxsNrA8lJQyiRkrnGW9/X 5PivAfYKuQ/OMz67dHc51KjklOlKs+Man2hRSCHYHqx+CPyvaDW0mU5mfD5/iyhbgYiv daBvMg8x3au0UkAwQ98l7GjgDEeMNC2NtMT/H/gDBDE74+/64DbVGJodoT21eWbPVLxe fZcg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=LCjWZY2DA9SGQMYB8874unldp4qBpu8sVqc5dXs8Kts=; b=tCejnk/RJTtQigRBniMZCTuaiUYsT/WDCC+hFWRbi7hD/95Ba1un2L6p7DPGHJXbNx akeTSX1Kw1fmHgQgsx7HnoqRS+t2euLkPk5gJCdU2BPGc2rfWnPpauJ0RaA5OvYRgVxG HrBEBnMJYxZlMr4Q5vSGv2YclEaLTDZsCqNHCdkNqkXxm9K9xL8nvUId1vWzkIO0cK8X eesgl6AY1Wg2jy9VRCff4hUbVF7Fm+2D+Agx6gYVl2vwG1vaJgYabVr+5v5Zvx7NKQ7B 26+ZjAGJlxpIp9Yd7UT5x1NkOQYi/8FeEZD+cpYXZCVY8crktnAiu7phxKByIYIQAxSu DYzQ== X-Gm-Message-State: APjAAAWeqLh4ArEcC+oPM7G6PuqdvBcrp4kQRr+7bm4wlgJ+IFY9ysE4 Xd5qV5yC4W5DjZkRYbmZQybUz6iKmV8= X-Google-Smtp-Source: APXvYqzUsMIyp0Dnk9wFU4H/PKSeCkXBAR1S9gzU6TR8JHPNQgwhAr1OvXKItArJRt5E0xXl+d/NwQ== X-Received: by 2002:a02:a716:: with SMTP id k22mr999681jam.108.1582646695574; Tue, 25 Feb 2020 08:04:55 -0800 (PST) Received: from x1.localdomain ([65.144.74.34]) by smtp.gmail.com with ESMTPSA id k23sm5628100ilg.83.2020.02.25.08.04.54 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 25 Feb 2020 08:04:55 -0800 (PST) From: Jens Axboe To: io-uring@vger.kernel.org Cc: andres@anarazel.de, Jens Axboe Subject: [PATCH 2/3] io_uring: add IORING_OP_PROVIDE_BUFFER Date: Tue, 25 Feb 2020 09:04:50 -0700 Message-Id: <20200225160451.7198-3-axboe@kernel.dk> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20200225160451.7198-1-axboe@kernel.dk> References: <20200225160451.7198-1-axboe@kernel.dk> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Sender: io-uring-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: io-uring@vger.kernel.org IORING_OP_PROVIDE_BUFFER uses the buffer registration infrastructure to support passing in an addr/len that is associated with a buffer ID and buffer group ID. The group ID is used to index and lookup the buffers, while the buffer ID can be used to notify the application which buffer in the group was used. At least for now, no validation is done of the buffer ID. If the application provides buffers within the same group with identical buffer IDs, then it'll have a hard time telling which buffer ID was used. The only restriction is that the buffer ID can be a max of 16-bits in size, so USHRT_MAX is the maximum ID that can be used. Signed-off-by: Jens Axboe --- fs/io_uring.c | 86 ++++++++++++++++++++++++++++++++++- include/uapi/linux/io_uring.h | 1 + 2 files changed, 86 insertions(+), 1 deletion(-) diff --git a/fs/io_uring.c b/fs/io_uring.c index d985da9252a2..d28602f32936 100644 --- a/fs/io_uring.c +++ b/fs/io_uring.c @@ -452,6 +452,14 @@ struct io_splice { unsigned int flags; }; +struct io_provide_buf { + struct file *file; + __u64 addr; + __s32 len; + __u32 gid; + __u16 bid; +}; + struct io_async_connect { struct sockaddr_storage address; }; @@ -577,6 +585,7 @@ struct io_kiocb { struct io_madvise madvise; struct io_epoll epoll; struct io_splice splice; + struct io_provide_buf pbuf; }; struct io_async_ctx *io; @@ -804,7 +813,8 @@ static const struct io_op_def io_op_defs[] = { .needs_file = 1, .hash_reg_file = 1, .unbound_nonreg_file = 1, - } + }, + [IORING_OP_PROVIDE_BUFFER] = {}, }; static void io_wq_submit_work(struct io_wq_work **workptr); @@ -2807,6 +2817,69 @@ static int io_openat(struct io_kiocb *req, struct io_kiocb **nxt, return io_openat2(req, nxt, force_nonblock); } +static int io_provide_buffer_prep(struct io_kiocb *req, + const struct io_uring_sqe *sqe) +{ + struct io_provide_buf *p = &req->pbuf; + u64 off; + + p->addr = READ_ONCE(sqe->addr); + p->len = READ_ONCE(sqe->len); + p->gid = READ_ONCE(sqe->fd); + off = READ_ONCE(sqe->off); + if (off > USHRT_MAX) + return -EINVAL; + p->bid = off; + return 0; +} + +static int io_provide_buffer(struct io_kiocb *req, struct io_kiocb **nxt) +{ + struct io_provide_buf *p = &req->pbuf; + struct io_ring_ctx *ctx = req->ctx; + struct list_head *list; + struct io_buffer *buf; + int ret = 0; + + list = idr_find(&ctx->io_buffer_idr, p->gid); + if (!list) { + list = kmalloc(sizeof(*list), GFP_KERNEL); + if (!list) { + ret = -ENOMEM; + goto out; + } + INIT_LIST_HEAD(list); + ret = idr_alloc(&ctx->io_buffer_idr, list, p->gid, p->gid + 1, + GFP_KERNEL); + if (ret < 0) { + kfree(list); + goto out; + } + } + + buf = kmalloc(sizeof(*buf), GFP_KERNEL); + if (!buf) { + if (list_empty(list)) { + idr_remove(&ctx->io_buffer_idr, p->gid); + kfree(list); + } + ret = -ENOMEM; + goto out; + } + + buf->addr = p->addr; + buf->len = p->len; + buf->bid = p->bid; + list_add(&buf->list, list); + ret = buf->bid; +out: + if (ret < 0) + req_set_fail_links(req); + io_cqring_add_event(req, ret); + io_put_req_find_next(req, nxt); + return 0; +} + static int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) { @@ -4422,6 +4495,9 @@ static int io_req_defer_prep(struct io_kiocb *req, case IORING_OP_SPLICE: ret = io_splice_prep(req, sqe); break; + case IORING_OP_PROVIDE_BUFFER: + ret = io_provide_buffer_prep(req, sqe); + break; default: printk_once(KERN_WARNING "io_uring: unhandled opcode %d\n", req->opcode); @@ -4699,6 +4775,14 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe, } ret = io_splice(req, nxt, force_nonblock); break; + case IORING_OP_PROVIDE_BUFFER: + if (sqe) { + ret = io_provide_buffer_prep(req, sqe); + if (ret) + break; + } + ret = io_provide_buffer(req, nxt); + break; default: ret = -EINVAL; break; diff --git a/include/uapi/linux/io_uring.h b/include/uapi/linux/io_uring.h index 53b36311cdac..7e2af5753bcc 100644 --- a/include/uapi/linux/io_uring.h +++ b/include/uapi/linux/io_uring.h @@ -119,6 +119,7 @@ enum { IORING_OP_OPENAT2, IORING_OP_EPOLL_CTL, IORING_OP_SPLICE, + IORING_OP_PROVIDE_BUFFER, /* this goes last, obviously */ IORING_OP_LAST, -- 2.25.1