* [PATCH v2 1/2] barriers: add load relaxed
2020-06-21 16:14 [PATCH v2 0/2] Fix hang in io_uring_get_cqe() with iopoll Pavel Begunkov
@ 2020-06-21 16:14 ` Pavel Begunkov
2020-06-21 16:14 ` [PATCH v2 2/2] Fix hang in io_uring_get_cqe() with iopoll Pavel Begunkov
2020-06-21 18:48 ` [PATCH v2 0/2] " Jens Axboe
2 siblings, 0 replies; 5+ messages in thread
From: Pavel Begunkov @ 2020-06-21 16:14 UTC (permalink / raw)
To: Jens Axboe, io-uring
Add io_uring_smp_load_relaxed() for internal use.
Signed-off-by: Pavel Begunkov <[email protected]>
---
src/include/liburing/barrier.h | 4 ++++
1 file changed, 4 insertions(+)
diff --git a/src/include/liburing/barrier.h b/src/include/liburing/barrier.h
index ad69506..6a1aa52 100644
--- a/src/include/liburing/barrier.h
+++ b/src/include/liburing/barrier.h
@@ -47,6 +47,8 @@ do { \
___p1; \
})
+#define io_uring_smp_load_relaxed(p) IO_URING_READ_ONCE(*(p))
+
#else /* defined(__x86_64__) || defined(__i386__) */
/*
* Add arch appropriate definitions. Use built-in atomic operations for
@@ -55,6 +57,8 @@ do { \
#define io_uring_smp_store_release(p, v) \
__atomic_store_n(p, v, __ATOMIC_RELEASE)
#define io_uring_smp_load_acquire(p) __atomic_load_n(p, __ATOMIC_ACQUIRE)
+#define io_uring_smp_load_relaxed(p) __atomic_load_n(p, __ATOMIC_RELAXED)
+
#endif /* defined(__x86_64__) || defined(__i386__) */
#endif /* defined(LIBURING_BARRIER_H) */
--
2.24.0
^ permalink raw reply related [flat|nested] 5+ messages in thread
* [PATCH v2 2/2] Fix hang in io_uring_get_cqe() with iopoll
2020-06-21 16:14 [PATCH v2 0/2] Fix hang in io_uring_get_cqe() with iopoll Pavel Begunkov
2020-06-21 16:14 ` [PATCH v2 1/2] barriers: add load relaxed Pavel Begunkov
@ 2020-06-21 16:14 ` Pavel Begunkov
2020-06-21 18:48 ` [PATCH v2 0/2] " Jens Axboe
2 siblings, 0 replies; 5+ messages in thread
From: Pavel Begunkov @ 2020-06-21 16:14 UTC (permalink / raw)
To: Jens Axboe, io-uring
Because of need_resched() check, io_uring_enter() -> io_iopoll_check()
can return 0 even if @min_complete wasn't satisfied. If that's the
case, __io_uring_get_cqe() sets submit=0 and wait_nr=0, disabling
setting IORING_ENTER_GETEVENTS as well. So, it goes crazy calling
io_uring_enter() in a loop, not actually submitting nor polling.
Set @wait_nr based on actual number of CQEs ready. It doesn't manifest
extra CQEs if any, thus implements __io_uring_cq_ready() with relaxed
semantics.
Signed-off-by: Pavel Begunkov <[email protected]>
---
src/queue.c | 16 +++++++++++++++-
1 file changed, 15 insertions(+), 1 deletion(-)
diff --git a/src/queue.c b/src/queue.c
index 14a0777..d824cfd 100644
--- a/src/queue.c
+++ b/src/queue.c
@@ -32,6 +32,19 @@ static inline bool sq_ring_needs_enter(struct io_uring *ring,
return false;
}
+static inline unsigned int __io_uring_cq_ready(struct io_uring *ring)
+{
+ return io_uring_smp_load_relaxed(ring->cq.ktail) - *ring->cq.khead;
+}
+
+static inline unsigned int io_adjust_wait_nr(struct io_uring *ring,
+ unsigned int to_wait)
+{
+ unsigned int ready = __io_uring_cq_ready(ring);
+
+ return (to_wait <= ready) ? 0 : (to_wait - ready);
+}
+
int __io_uring_get_cqe(struct io_uring *ring, struct io_uring_cqe **cqe_ptr,
unsigned submit, unsigned wait_nr, sigset_t *sigmask)
{
@@ -60,7 +73,8 @@ int __io_uring_get_cqe(struct io_uring *ring, struct io_uring_cqe **cqe_ptr,
err = -errno;
} else if (ret == (int)submit) {
submit = 0;
- wait_nr = 0;
+ if (to_wait)
+ wait_nr = io_adjust_wait_nr(ring, to_wait);
} else {
submit -= ret;
}
--
2.24.0
^ permalink raw reply related [flat|nested] 5+ messages in thread