From: Pavel Begunkov <[email protected]>
To: [email protected], [email protected],
[email protected], [email protected]
Cc: Jens Axboe <[email protected]>, Alexei Starovoitov <[email protected]>,
Daniel Borkmann <[email protected]>,
Andrii Nakryiko <[email protected]>,
Martin KaFai Lau <[email protected]>, Song Liu <[email protected]>,
Yonghong Song <[email protected]>,
John Fastabend <[email protected]>,
KP Singh <[email protected]>,
Horst Schirmeier <[email protected]>,
"Franz-B . Tuneke" <[email protected]>,
Christian Dietrich <[email protected]>
Subject: [PATCH 10/23] io_uring: add support for multiple CQs
Date: Wed, 19 May 2021 15:13:21 +0100 [thread overview]
Message-ID: <6f8364dc9df59b27b09472f262a545b42f8639a5.1621424513.git.asml.silence@gmail.com> (raw)
In-Reply-To: <[email protected]>
TODO: don't rob all bits from params, use pointer to a struct
Signed-off-by: Pavel Begunkov <[email protected]>
---
fs/io_uring.c | 89 +++++++++++++++++++++++++++--------
include/uapi/linux/io_uring.h | 3 +-
2 files changed, 71 insertions(+), 21 deletions(-)
diff --git a/fs/io_uring.c b/fs/io_uring.c
index f05592ae5f41..067cfb3a6e4a 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -91,6 +91,7 @@
#define IORING_MAX_CQ_ENTRIES (2 * IORING_MAX_ENTRIES)
#define IO_DEFAULT_CQ 0
+#define IO_MAX_CQRINGS 1024
/*
* Shift of 9 is 512 entries, or exactly one page on 64-bit archs
@@ -417,7 +418,7 @@ struct io_ring_ctx {
unsigned long cq_check_overflow;
unsigned cq_extra;
struct wait_queue_head cq_wait;
- struct io_cqring cqs[1];
+ struct io_cqring *cqs;
unsigned int cq_nr;
struct {
@@ -1166,6 +1167,9 @@ static struct io_ring_ctx *io_ring_ctx_alloc(struct io_uring_params *p)
if (!ctx)
return NULL;
+ ctx->cqs = kmalloc_array(p->nr_cq + 1, sizeof(ctx->cqs[0]), GFP_KERNEL);
+ if (!ctx->cqs)
+ goto err;
/*
* Use 5 bits less than the max cq entries, that should give us around
* 32 entries per hash list if totally full and uniformly spread.
@@ -8634,6 +8638,8 @@ static bool io_wait_rsrc_data(struct io_rsrc_data *data)
static void io_ring_ctx_free(struct io_ring_ctx *ctx)
{
+ unsigned int i;
+
io_sq_thread_finish(ctx);
if (ctx->mm_account) {
@@ -8673,6 +8679,9 @@ static void io_ring_ctx_free(struct io_ring_ctx *ctx)
io_mem_free(ctx->rings);
io_mem_free(ctx->sq_sqes);
+ for (i = 1; i < ctx->cq_nr; i++)
+ io_mem_free(ctx->cqs[i].rings);
+ kfree(ctx->cqs);
percpu_ref_exit(&ctx->refs);
free_uid(ctx->user);
@@ -9524,11 +9533,39 @@ static const struct file_operations io_uring_fops = {
#endif
};
+static void __io_init_cqring(struct io_cqring *cq, struct io_rings *rings,
+ unsigned int entries)
+{
+ WRITE_ONCE(rings->cq_ring_entries, entries);
+ WRITE_ONCE(rings->cq_ring_mask, entries - 1);
+
+ cq->cached_tail = 0;
+ cq->rings = rings;
+ cq->entries = entries;
+}
+
+static int io_init_cqring(struct io_cqring *cq, unsigned int entries)
+{
+ struct io_rings *rings;
+ size_t size;
+
+ size = rings_size(0, entries, NULL);
+ if (size == SIZE_MAX)
+ return -EOVERFLOW;
+ rings = io_mem_alloc(size);
+ if (!rings)
+ return -ENOMEM;
+ __io_init_cqring(cq, rings, entries);
+ return 0;
+}
+
static int io_allocate_scq_urings(struct io_ring_ctx *ctx,
struct io_uring_params *p)
{
+ u32 __user *cq_sizes = u64_to_user_ptr(p->cq_sizes);
struct io_rings *rings;
size_t size, sq_array_offset;
+ int i, ret;
/* make sure these are sane, as we already accounted them */
ctx->sq_entries = p->sq_entries;
@@ -9544,30 +9581,43 @@ static int io_allocate_scq_urings(struct io_ring_ctx *ctx,
ctx->rings = rings;
ctx->sq_array = (u32 *)((char *)rings + sq_array_offset);
rings->sq_ring_mask = p->sq_entries - 1;
- rings->cq_ring_mask = p->cq_entries - 1;
rings->sq_ring_entries = p->sq_entries;
- rings->cq_ring_entries = p->cq_entries;
- ctx->cqs[0].cached_tail = 0;
- ctx->cqs[0].rings = rings;
- ctx->cqs[0].entries = p->cq_entries;
+ __io_init_cqring(&ctx->cqs[0], rings, p->cq_entries);
ctx->cq_nr = 1;
size = array_size(sizeof(struct io_uring_sqe), p->sq_entries);
- if (size == SIZE_MAX) {
- io_mem_free(ctx->rings);
- ctx->rings = NULL;
- return -EOVERFLOW;
- }
+ ret = -EOVERFLOW;
+ if (unlikely(size == SIZE_MAX))
+ goto err;
ctx->sq_sqes = io_mem_alloc(size);
- if (!ctx->sq_sqes) {
- io_mem_free(ctx->rings);
- ctx->rings = NULL;
- return -ENOMEM;
+ ret = -ENOMEM;
+ if (unlikely(!ctx->sq_sqes))
+ goto err;
+
+ for (i = 0; i < p->nr_cq; i++, ctx->cq_nr++) {
+ u32 sz;
+ long entries;
+
+ ret = -EFAULT;
+ if (copy_from_user(&sz, &cq_sizes[i], sizeof(sz)))
+ goto err;
+ entries = io_get_cqring_size(p, sz);
+ if (entries < 0) {
+ ret = entries;
+ goto err;
+ }
+ ret = io_init_cqring(&ctx->cqs[i + 1], entries);
+ if (ret)
+ goto err;
}
return 0;
+err:
+ io_mem_free(ctx->rings);
+ ctx->rings = NULL;
+ return ret;
}
static int io_uring_install_fd(struct io_ring_ctx *ctx, struct file *file)
@@ -9653,6 +9703,10 @@ static int io_uring_create(unsigned entries, struct io_uring_params *p,
} else {
p->cq_entries = 2 * p->sq_entries;
}
+ if (p->nr_cq > IO_MAX_CQRINGS)
+ return -EINVAL;
+ if (!p->nr_cq != !p->cq_sizes)
+ return -EINVAL;
ctx = io_ring_ctx_alloc(p);
if (!ctx)
@@ -9744,14 +9798,9 @@ static int io_uring_create(unsigned entries, struct io_uring_params *p,
static long io_uring_setup(u32 entries, struct io_uring_params __user *params)
{
struct io_uring_params p;
- int i;
if (copy_from_user(&p, params, sizeof(p)))
return -EFAULT;
- for (i = 0; i < ARRAY_SIZE(p.resv); i++) {
- if (p.resv[i])
- return -EINVAL;
- }
if (p.flags & ~(IORING_SETUP_IOPOLL | IORING_SETUP_SQPOLL |
IORING_SETUP_SQ_AFF | IORING_SETUP_CQSIZE |
diff --git a/include/uapi/linux/io_uring.h b/include/uapi/linux/io_uring.h
index c2dfb179360a..92b61ca09ea5 100644
--- a/include/uapi/linux/io_uring.h
+++ b/include/uapi/linux/io_uring.h
@@ -263,7 +263,8 @@ struct io_uring_params {
__u32 sq_thread_idle;
__u32 features;
__u32 wq_fd;
- __u32 resv[3];
+ __u32 nr_cq;
+ __u64 cq_sizes;
struct io_sqring_offsets sq_off;
struct io_cqring_offsets cq_off;
};
--
2.31.1
next prev parent reply other threads:[~2021-05-19 14:14 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-05-19 14:13 [RFC v2 00/23] io_uring BPF requests Pavel Begunkov
2021-05-19 14:13 ` [PATCH 01/23] io_uring: shuffle rarely used ctx fields Pavel Begunkov
2021-05-20 21:46 ` Song Liu
2021-05-20 22:46 ` Pavel Begunkov
2021-05-19 14:13 ` [PATCH 02/23] io_uring: localise fixed resources fields Pavel Begunkov
2021-05-19 14:13 ` [PATCH 03/23] io_uring: remove dependency on ring->sq/cq_entries Pavel Begunkov
2021-05-19 14:13 ` [PATCH 04/23] io_uring: deduce cq_mask from cq_entries Pavel Begunkov
2021-05-19 14:13 ` [PATCH 05/23] io_uring: kill cached_cq_overflow Pavel Begunkov
2021-05-19 14:13 ` [PATCH 06/23] io_uring: rename io_get_cqring Pavel Begunkov
2021-05-19 14:13 ` [PATCH 07/23] io_uring: extract struct for CQ Pavel Begunkov
2021-05-19 14:13 ` [PATCH 08/23] io_uring: internally pass CQ indexes Pavel Begunkov
2021-05-19 14:13 ` [PATCH 09/23] io_uring: extract cq size helper Pavel Begunkov
2021-05-19 14:13 ` Pavel Begunkov [this message]
2021-05-19 14:13 ` [PATCH 11/23] io_uring: enable mmap'ing additional CQs Pavel Begunkov
2021-05-19 14:13 ` [PATCH 12/23] bpf: add IOURING program type Pavel Begunkov
2021-05-20 23:34 ` Song Liu
2021-05-21 0:56 ` Pavel Begunkov
2021-05-19 14:13 ` [PATCH 13/23] io_uring: implement bpf prog registration Pavel Begunkov
2021-05-20 23:45 ` Song Liu
2021-05-21 0:43 ` Pavel Begunkov
2021-05-19 14:13 ` [PATCH 14/23] io_uring: add support for bpf requests Pavel Begunkov
2021-05-21 0:42 ` Pavel Begunkov
2021-05-19 14:13 ` [PATCH 15/23] io_uring: enable BPF to submit SQEs Pavel Begunkov
2021-05-21 0:06 ` Song Liu
2021-05-21 1:07 ` Alexei Starovoitov
2021-05-21 9:33 ` Pavel Begunkov
2021-05-19 14:13 ` [PATCH 16/23] io_uring: enable bpf to submit CQEs Pavel Begunkov
2021-05-19 14:13 ` [PATCH 17/23] io_uring: enable bpf to reap CQEs Pavel Begunkov
2021-05-19 14:13 ` [PATCH 18/23] libbpf: support io_uring Pavel Begunkov
2021-05-19 17:38 ` Andrii Nakryiko
2021-05-20 9:58 ` Pavel Begunkov
2021-05-20 17:23 ` Andrii Nakryiko
2021-05-19 14:13 ` [PATCH 19/23] io_uring: pass user_data to bpf executor Pavel Begunkov
2021-05-19 14:13 ` [PATCH 20/23] bpf: Add bpf_copy_to_user() helper Pavel Begunkov
2021-05-19 14:13 ` [PATCH 21/23] io_uring: wire bpf copy to user Pavel Begunkov
2021-05-19 14:13 ` [PATCH 22/23] io_uring: don't wait on CQ exclusively Pavel Begunkov
2021-05-19 14:13 ` [PATCH 23/23] io_uring: enable bpf reqs to wait for CQs Pavel Begunkov
2021-05-21 0:35 ` [RFC v2 00/23] io_uring BPF requests Song Liu
2021-05-21 0:58 ` Pavel Begunkov
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=6f8364dc9df59b27b09472f262a545b42f8639a5.1621424513.git.asml.silence@gmail.com \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
[email protected] \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox