From: Dmitry Antipov <dmantipov@yandex.ru>
To: Jens Axboe <axboe@kernel.dk>, Pavel Begunkov <asml.silence@gmail.com>
Cc: io-uring@vger.kernel.org, Dmitry Antipov <dmantipov@yandex.ru>
Subject: [PATCH] io_uring: miscellaneous spelling fixes
Date: Tue, 3 Jun 2025 12:20:45 +0300 [thread overview]
Message-ID: <20250603092045.384197-1-dmantipov@yandex.ru> (raw)
Correct spelling here and there as suggested by codespell.
Signed-off-by: Dmitry Antipov <dmantipov@yandex.ru>
---
io_uring/cancel.c | 4 ++--
io_uring/io-wq.c | 2 +-
io_uring/io_uring.c | 8 ++++----
io_uring/notif.c | 2 +-
io_uring/poll.c | 2 +-
io_uring/rw.c | 4 ++--
io_uring/tctx.h | 2 +-
io_uring/uring_cmd.c | 2 +-
io_uring/waitid.c | 2 +-
9 files changed, 14 insertions(+), 14 deletions(-)
diff --git a/io_uring/cancel.c b/io_uring/cancel.c
index 6d57602304df..6d46a0ac278a 100644
--- a/io_uring/cancel.c
+++ b/io_uring/cancel.c
@@ -306,8 +306,8 @@ int io_sync_cancel(struct io_ring_ctx *ctx, void __user *arg)
}
/*
- * Keep looking until we get -ENOENT. we'll get woken everytime
- * every time a request completes and will retry the cancelation.
+ * Keep looking until we get -ENOENT. We'll get woken every
+ * time a request completes and will retry the cancellation.
*/
do {
cd.seq = atomic_inc_return(&ctx->cancel_seq);
diff --git a/io_uring/io-wq.c b/io_uring/io-wq.c
index cd1fcb115739..70fdf174e4a1 100644
--- a/io_uring/io-wq.c
+++ b/io_uring/io-wq.c
@@ -605,7 +605,7 @@ static void io_worker_handle_work(struct io_wq_acct *acct,
work = io_get_next_work(acct, wq);
if (work) {
/*
- * Make sure cancelation can find this, even before
+ * Make sure cancellation can find this, even before
* it becomes the active work. That avoids a window
* where the work has been removed from our general
* work list, but isn't yet discoverable as the
diff --git a/io_uring/io_uring.c b/io_uring/io_uring.c
index c7a9cecf528e..abbb4f3dad88 100644
--- a/io_uring/io_uring.c
+++ b/io_uring/io_uring.c
@@ -1157,7 +1157,7 @@ static void io_req_local_work_add(struct io_kiocb *req, unsigned flags)
BUILD_BUG_ON(IO_CQ_WAKE_FORCE <= IORING_MAX_CQ_ENTRIES);
/*
- * We don't know how many reuqests is there in the link and whether
+ * We don't know how many requests is there in the link and whether
* they can even be queued lazily, fall back to non-lazy.
*/
if (req->flags & IO_REQ_LINK_FLAGS)
@@ -2848,7 +2848,7 @@ static __cold void io_tctx_exit_cb(struct callback_head *cb)
* When @in_cancel, we're in cancellation and it's racy to remove the
* node. It'll be removed by the end of cancellation, just ignore it.
* tctx can be NULL if the queueing of this task_work raced with
- * work cancelation off the exec path.
+ * work cancellation off the exec path.
*/
if (tctx && !atomic_read(&tctx->in_cancel))
io_uring_del_tctx_node((unsigned long)work->ctx);
@@ -2980,7 +2980,7 @@ static __cold void io_ring_ctx_wait_and_kill(struct io_ring_ctx *ctx)
/*
* Use system_unbound_wq to avoid spawning tons of event kworkers
* if we're exiting a ton of rings at the same time. It just adds
- * noise and overhead, there's no discernable change in runtime
+ * noise and overhead, there's no discernible change in runtime
* over using system_wq.
*/
queue_work(iou_wq, &ctx->exit_work);
@@ -3152,7 +3152,7 @@ __cold void io_uring_cancel_generic(bool cancel_all, struct io_sq_data *sqd)
if (!tctx_inflight(tctx, !cancel_all))
break;
- /* read completions before cancelations */
+ /* read completions before cancellations */
inflight = tctx_inflight(tctx, false);
if (!inflight)
break;
diff --git a/io_uring/notif.c b/io_uring/notif.c
index 9a6f6e92d742..93140abebd10 100644
--- a/io_uring/notif.c
+++ b/io_uring/notif.c
@@ -87,7 +87,7 @@ static int io_link_skb(struct sk_buff *skb, struct ubuf_info *uarg)
prev_nd = container_of(prev_uarg, struct io_notif_data, uarg);
prev_notif = cmd_to_io_kiocb(nd);
- /* make sure all noifications can be finished in the same task_work */
+ /* make sure all notifications can be finished in the same task_work */
if (unlikely(notif->ctx != prev_notif->ctx ||
notif->tctx != prev_notif->tctx))
return -EEXIST;
diff --git a/io_uring/poll.c b/io_uring/poll.c
index 0526062e2f81..dafe04dd6915 100644
--- a/io_uring/poll.c
+++ b/io_uring/poll.c
@@ -440,7 +440,7 @@ static bool io_poll_double_prepare(struct io_kiocb *req)
/*
* poll arm might not hold ownership and so race for req->flags with
* io_poll_wake(). There is only one poll entry queued, serialise with
- * it by taking its head lock. As we're still arming the tw hanlder
+ * it by taking its head lock. As we're still arming the tw handler
* is not going to be run, so there are no races with it.
*/
if (head) {
diff --git a/io_uring/rw.c b/io_uring/rw.c
index 710d8cd53ebb..e7e30af269a9 100644
--- a/io_uring/rw.c
+++ b/io_uring/rw.c
@@ -186,7 +186,7 @@ static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags)
* This is really a bug in the core code that does this, any issue
* path should assume that a successful (or -EIOCBQUEUED) return can
* mean that the underlying data can be gone at any time. But that
- * should be fixed seperately, and then this check could be killed.
+ * should be fixed separately, and then this check could be killed.
*/
if (!(req->flags & (REQ_F_REISSUE | REQ_F_REFCOUNT))) {
req->flags &= ~REQ_F_NEED_CLEANUP;
@@ -347,7 +347,7 @@ static int io_prep_rwv(struct io_kiocb *req, const struct io_uring_sqe *sqe,
/*
* Have to do this validation here, as this is in io_read() rw->len
- * might have chanaged due to buffer selection
+ * might have changed due to buffer selection
*/
return io_iov_buffer_select_prep(req);
}
diff --git a/io_uring/tctx.h b/io_uring/tctx.h
index 608e96de70a2..1c10a3a1a00e 100644
--- a/io_uring/tctx.h
+++ b/io_uring/tctx.h
@@ -20,7 +20,7 @@ int io_ringfd_unregister(struct io_ring_ctx *ctx, void __user *__arg,
unsigned nr_args);
/*
- * Note that this task has used io_uring. We use it for cancelation purposes.
+ * Note that this task has used io_uring. We use it for cancellation purposes.
*/
static inline int io_uring_add_tctx_node(struct io_ring_ctx *ctx)
{
diff --git a/io_uring/uring_cmd.c b/io_uring/uring_cmd.c
index 929cad6ee326..40e35e8f8821 100644
--- a/io_uring/uring_cmd.c
+++ b/io_uring/uring_cmd.c
@@ -126,7 +126,7 @@ static void io_uring_cmd_work(struct io_kiocb *req, io_tw_token_t tw)
if (io_should_terminate_tw())
flags |= IO_URING_F_TASK_DEAD;
- /* task_work executor checks the deffered list completion */
+ /* task_work executor checks the deferred list completion */
ioucmd->task_work_cb(ioucmd, flags);
}
diff --git a/io_uring/waitid.c b/io_uring/waitid.c
index e07a94694397..149439fdfcac 100644
--- a/io_uring/waitid.c
+++ b/io_uring/waitid.c
@@ -282,7 +282,7 @@ int io_waitid(struct io_kiocb *req, unsigned int issue_flags)
atomic_set(&iw->refs, 1);
/*
- * Cancel must hold the ctx lock, so there's no risk of cancelation
+ * Cancel must hold the ctx lock, so there's no risk of cancellation
* finding us until a) we remain on the list, and b) the lock is
* dropped. We only need to worry about racing with the wakeup
* callback.
--
2.49.0
next reply other threads:[~2025-06-03 9:29 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-06-03 9:20 Dmitry Antipov [this message]
2025-06-03 13:07 ` [PATCH] io_uring: miscellaneous spelling fixes Jens Axboe
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250603092045.384197-1-dmantipov@yandex.ru \
--to=dmantipov@yandex.ru \
--cc=asml.silence@gmail.com \
--cc=axboe@kernel.dk \
--cc=io-uring@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox