public inbox for [email protected]
 help / color / mirror / Atom feed
From: Zorro Lang <[email protected]>
To: Jens Axboe <[email protected]>
Cc: io-uring <[email protected]>,
	LKML <[email protected]>,
	Linus Torvalds <[email protected]>,
	Dave Chinner <[email protected]>
Subject: Re: [PATCH] io_uring/io-wq: don't clear PF_IO_WORKER on exit
Date: Wed, 14 Jun 2023 09:03:48 +0800	[thread overview]
Message-ID: <20230614010348.2o6wjax543teosph@zlang-mailbox> (raw)
In-Reply-To: <20230614005449.awc2ncxl5lb2eg6m@zlang-mailbox>

On Wed, Jun 14, 2023 at 08:54:49AM +0800, Zorro Lang wrote:
> On Mon, Jun 12, 2023 at 12:11:57PM -0600, Jens Axboe wrote:
> > A recent commit gated the core dumping task exit logic on current->flags
> > remaining consistent in terms of PF_{IO,USER}_WORKER at task exit time.
> > This exposed a problem with the io-wq handling of that, which explicitly
> > clears PF_IO_WORKER before calling do_exit().
> > 
> > The reasons for this manual clear of PF_IO_WORKER is historical, where
> > io-wq used to potentially trigger a sleep on exit. As the io-wq thread
> > is exiting, it should not participate any further accounting. But these
> > days we don't need to rely on current->flags anymore, so we can safely
> > remove the PF_IO_WORKER clearing.
> > 
> > Reported-by: Zorro Lang <[email protected]>
> > Reported-by: Dave Chinner <[email protected]>
> > Reported-by: Linus Torvalds <[email protected]>
> > Link: https://lore.kernel.org/all/[email protected]/
> > Fixes: f9010dbdce91 ("fork, vhost: Use CLONE_THREAD to fix freezer/ps regression")
> > Signed-off-by: Jens Axboe <[email protected]>
> > 
> > ---
> 
> Hi,
> 
> This patch fix the issue I reported. The bug can be reproduced on v6.4-rc6,
> then test passed on v6.4-rc6 with this patch.
> 
> But I found another KASAN bug [1] on aarch64 machine, by running generic/388.
> I hit that 3 times. And hit a panic [2] (once after that kasan bug) on a x86_64
> with pmem device (mount with dax=never), by running geneirc/388 too.

Oh, x86_64 with pmem can reproduce this kasan bug nearly 100%, no matter the
mount option. Then aarch64 (4k page size) can reproduce it too. But I didn't
hit it on ppc64le (64k page size). Other testing jobs haven't been done, currently
I only have these results.

Thanks,
Zorro

> 
> Thanks,
> Zorro
> 
> [1]
> [16505.224912] run fstests generic/388 at 2023-06-13 13:01:12 
> [16507.375208] XFS (vda3): Mounting V5 Filesystem c1139709-d136-4d28-9f24-ba77c4da95cd 
> [16507.390655] XFS (vda3): Ending clean mount 
> [16507.417771] XFS (vda3): User initiated shutdown received. 
> [16507.418371] XFS (vda3): Metadata I/O Error (0x4) detected at xfs_fs_goingdown+0x58/0x180 [xfs] (fs/xfs/xfs_fsops.c:483).  Shutting down filesystem. 
> [16507.419419] XFS (vda3): Please unmount the filesystem and rectify the problem(s) 
> [16507.447760] XFS (vda3): Unmounting Filesystem c1139709-d136-4d28-9f24-ba77c4da95cd 
> [16508.015675] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16508.032248] XFS (vda3): Ending clean mount 
> [16509.082072] XFS (vda3): User initiated shutdown received. 
> [16509.082538] XFS (vda3): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe8/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [16509.083606] XFS (vda3): Please unmount the filesystem and rectify the problem(s) 
> [16509.737031] XFS (vda3): Unmounting Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16510.077278] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16510.387017] XFS (vda3): Starting recovery (logdev: internal) 
> [16510.769548] XFS (vda3): Ending recovery (logdev: internal) 
> [16512.836365] XFS (vda3): User initiated shutdown received. 
> [16512.839914] XFS (vda3): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe8/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [16512.840959] XFS (vda3): Please unmount the filesystem and rectify the problem(s) 
> [16512.999722] ================================================================== 
> [16513.000246] BUG: KASAN: slab-use-after-free in io_wq_worker_running+0x108/0x120 
> [16513.000734] Read of size 4 at addr ffff0000da65b804 by task iou-wrk-1240163/1240168 
> [16513.001232]  
> [16513.001344] CPU: 2 PID: 1240168 Comm: iou-wrk-1240163 Kdump: loaded Not tainted 6.4.0-rc6+ #1 
> [16513.001910] Hardware name: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 
> [16513.002396] Call trace: 
> [16513.002571]  dump_backtrace+0x9c/0x120 
> [16513.002844]  show_stack+0x1c/0x30 
> [16513.003064]  dump_stack_lvl+0xe0/0x168 
> [16513.003323]  print_address_description.constprop.0+0x80/0x390 
> [16513.003721]  print_report+0xa4/0x268 
> [16513.003969]  kasan_report+0x80/0xc0 
> [16513.004216]  __asan_report_load4_noabort+0x1c/0x28 
> [16513.004550]  io_wq_worker_running+0x108/0x120 
> [16513.004864]  schedule+0x178/0x208 
> [16513.005124]  coredump_task_exit+0x1e0/0x380 
> [16513.005434]  do_exit+0x16c/0xbc0 
> [16513.005675]  io_wq_worker+0x4dc/0x828 
> [16513.005936]  ret_from_fork+0x10/0x20 
> [16513.006188]  
> [16513.006312] Allocated by task 1240163: 
> [16513.006585]  kasan_save_stack+0x28/0x50 
> [16513.006868]  kasan_set_track+0x28/0x38 
> [16513.007156]  kasan_save_alloc_info+0x20/0x30 
> [16513.007458]  __kasan_kmalloc+0x98/0xb0 
> [16513.007725]  kmalloc_trace+0x78/0x90 
> [16513.007981]  create_io_worker+0xb0/0x478 
> [16513.008262]  io_wq_create_worker+0x128/0x200 
> [16513.008570]  io_wq_enqueue+0x304/0x6d0 
> [16513.008839]  io_queue_iowq+0x210/0x5a8 
> [16513.009116]  io_queue_async+0x120/0x430 
> [16513.009400]  io_submit_sqes+0xa44/0xfd0 
> [16513.009690]  __do_sys_io_uring_enter+0x554/0xec8 
> [16513.010035]  __arm64_sys_io_uring_enter+0xc4/0x138 
> [16513.010376]  invoke_syscall.constprop.0+0xd8/0x1d8 
> [16513.010719]  el0_svc_common.constprop.0+0x234/0x2c0 
> [16513.011065]  do_el0_svc+0x50/0x108 
> [16513.011314]  el0_svc+0x48/0x120 
> [16513.011538]  el0t_64_sync_handler+0xb4/0x130 
> [16513.011840]  el0t_64_sync+0x17c/0x180 
> [16513.012097]  
> [16513.012206] Freed by task 1239074: 
> [16513.012470]  kasan_save_stack+0x28/0x50 
> [16513.012753]  kasan_set_track+0x28/0x38 
> [16513.013032]  kasan_save_free_info+0x34/0x58 
> [16513.013329]  __kasan_slab_free+0xe4/0x158 
> [16513.013613]  slab_free_freelist_hook+0xf0/0x1d0 
> [16513.013939]  kmem_cache_free_bulk.part.0+0x250/0x4d0 
> [16513.014300]  kmem_cache_free_bulk+0x18/0x28 
> [16513.014600]  kvfree_rcu_bulk+0x338/0x420 
> [16513.014884]  kvfree_rcu_drain_ready+0x1fc/0x5c0 
> [16513.015212]  kfree_rcu_monitor+0x4c/0x980 
> [16513.015503]  process_one_work+0x79c/0x1678 
> [16513.015805]  worker_thread+0x3d8/0xc60 
> [16513.016078]  kthread+0x27c/0x300 
> [16513.016310]  ret_from_fork+0x10/0x20 
> [16513.016574]  
> [16513.016699] Last potentially related work creation: 
> [16513.017097]  kasan_save_stack+0x28/0x50 
> [16513.017396]  __kasan_record_aux_stack+0x98/0xc0 
> [16513.017741]  kasan_record_aux_stack_noalloc+0x10/0x18 
> [16513.018109]  kvfree_call_rcu+0x110/0x458 
> [16513.018416]  io_wq_worker+0x4c8/0x828 
> [16513.018696]  ret_from_fork+0x10/0x20 
> [16513.018972]  
> [16513.019098] Second to last potentially related work creation: 
> [16513.019518]  kasan_save_stack+0x28/0x50 
> [16513.019834]  __kasan_record_aux_stack+0x98/0xc0 
> [16513.020178]  kasan_record_aux_stack_noalloc+0x10/0x18 
> [16513.020559]  insert_work+0x54/0x290 
> [16513.020814]  __queue_work+0x4e4/0xf08 
> [16513.021079]  queue_work_on+0x110/0x160 
> [16513.021346]  xlog_cil_push_now.isra.0+0x154/0x1d8 [xfs] 
> [16513.021881]  xlog_cil_force_seq+0x20c/0x938 [xfs] 
> [16513.022359]  xfs_log_force_seq+0x268/0x840 [xfs] 
> [16513.022829]  xfs_log_force_inode+0xe8/0x140 [xfs] 
> [16513.023302]  xfs_dir_fsync+0x104/0x328 [xfs] 
> [16513.023746]  vfs_fsync_range+0xc4/0x1f0 
> [16513.024020]  do_fsync+0x40/0x88 
> [16513.024245]  __arm64_sys_fsync+0x38/0x58 
> [16513.024524]  invoke_syscall.constprop.0+0xd8/0x1d8 
> [16513.024875]  el0_svc_common.constprop.0+0x234/0x2c0 
> [16513.025221]  do_el0_svc+0x50/0x108 
> [16513.025470]  el0_svc+0x48/0x120 
> [16513.025714]  el0t_64_sync_handler+0xb4/0x130 
> [16513.026033]  el0t_64_sync+0x17c/0x180 
> [16513.026311]  
> [16513.026421] The buggy address belongs to the object at ffff0000da65b800 
> [16513.026421]  which belongs to the cache kmalloc-512 of size 512 
> [16513.027308] The buggy address is located 4 bytes inside of 
> [16513.027308]  freed 512-byte region [ffff0000da65b800, ffff0000da65ba00) 
> [16513.028163]  
> [16513.028282] The buggy address belongs to the physical page: 
> [16513.028687] page:0000000099e24e4d refcount:1 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x11a658 
> [16513.029385] head:0000000099e24e4d order:2 entire_mapcount:0 nr_pages_mapped:0 pincount:0 
> [16513.029957] anon flags: 0x17ffff800010200(slab|head|node=0|zone=2|lastcpupid=0xfffff) 
> [16513.030548] page_type: 0xffffffff() 
> [16513.030816] raw: 017ffff800010200 ffff0000c0002600 0000000000000000 dead000000000001 
> [16513.031374] raw: 0000000000000000 0000000000100010 00000001ffffffff 0000000000000000 
> [16513.031929] page dumped because: kasan: bad access detected 
> [16513.032327]  
> [16513.032441] Memory state around the buggy address: 
> [16513.032821]  ffff0000da65b700: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc 
> [16513.033336]  ffff0000da65b780: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc 
> [16513.033865] >ffff0000da65b800: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb 
> [16513.034378]                    ^ 
> [16513.034609]  ffff0000da65b880: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb 
> [16513.035116]  ffff0000da65b900: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb 
> [16513.035626] ================================================================== 
> [16513.036232] Disabling lock debugging due to kernel taint 
> [16513.262535] XFS (vda3): Unmounting Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16513.623346] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16513.735177] XFS (vda3): Starting recovery (logdev: internal) 
> [16514.226365] XFS (vda3): Ending recovery (logdev: internal) 
> [16514.253119] XFS (vda3): User initiated shutdown received. 
> [16514.253573] XFS (vda3): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe8/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [16514.254595] XFS (vda3): Please unmount the filesystem and rectify the problem(s) 
> [16514.432583] XFS (vda3): Unmounting Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16514.776940] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16514.800218] XFS (vda3): Starting recovery (logdev: internal) 
> [16514.807050] XFS (vda3): Ending recovery (logdev: internal) 
> [16514.826857] XFS (vda3): Unmounting Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16515.205970] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16515.220225] XFS (vda3): Ending clean mount 
> [16517.252566] XFS (vda3): User initiated shutdown received. 
> [16517.252990] XFS (vda3): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe8/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [16517.253979] XFS (vda3): Please unmount the filesystem and rectify the problem(s) 
> [16517.690386] XFS (vda3): Unmounting Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16517.977164] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16518.133228] XFS (vda3): Starting recovery (logdev: internal) 
> [16518.939479] XFS (vda3): Ending recovery (logdev: internal) 
> [16518.963049] XFS (vda3): User initiated shutdown received. 
> [16518.963057] XFS (vda3): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [16518.963481] XFS (vda3): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe8/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [16518.965017] XFS (vda3): Please unmount the filesystem and rectify the problem(s) 
> [16519.204814] XFS (vda3): Unmounting Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16519.463997] XFS (vda3): Mounting V5 Filesystem 0e0c8b4e-9c4b-4804-a5f6-a7909cc599aa 
> [16519.490002] XFS (vda3): Starting recovery (logdev: internal) 
> [16519.494259] XFS (vda3): Ending recovery (logdev: internal) 
> ...
> ...
> 
> [2]
> [ 8365.864086] run fstests generic/388 at 2023-06-13 08:49:12 
> [ 8368.274443] XFS (pmem1): Mounting V5 Filesystem 3c93b02c-97e9-4cc2-9173-73cb6d8c1004 
> [ 8368.293316] XFS (pmem1): Ending clean mount 
> [ 8368.318944] XFS (pmem1): User initiated shutdown received. 
> [ 8368.325440] XFS (pmem1): Metadata I/O Error (0x4) detected at xfs_fs_goingdown+0x4b/0x180 [xfs] (fs/xfs/xfs_fsops.c:483).  Shutting down filesystem. 
> [ 8368.338972] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8368.362196] XFS (pmem1): Unmounting Filesystem 3c93b02c-97e9-4cc2-9173-73cb6d8c1004 
> [ 8368.660094] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8368.678482] XFS (pmem1): Ending clean mount 
> [ 8370.719061] XFS (pmem1): User initiated shutdown received. 
> [ 8370.724599] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8370.738586] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8371.425668] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8371.601187] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8371.644024] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8374.884353] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8376.930483] XFS (pmem1): User initiated shutdown received. 
> [ 8376.930569] iomap_finish_ioend: 1 callbacks suppressed 
> [ 8376.930576] pmem1: writeback error on inode 29392430, offset 3960832, sector 18409168 
> [ 8376.936091] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8376.941475] pmem1: writeback error on inode 29392430, offset 7528448, sector 18425224 
> [ 8376.949302] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8377.575276] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8377.753574] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8377.792831] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8380.328214] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8380.373691] XFS (pmem1): User initiated shutdown received. 
> [ 8380.373702] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8380.373947] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8380.376499] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8380.379233] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8380.417452] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8380.825341] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8380.956173] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8380.983834] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8381.001996] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8382.043732] XFS (pmem1): User initiated shutdown received. 
> [ 8382.044041] pmem1: writeback error on inode 6374535, offset 45056, sector 4070976 
> [ 8382.049348] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8382.070775] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8382.624394] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8382.812699] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8382.846934] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8383.719470] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8383.747107] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8383.868827] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8383.895176] XFS (pmem1): Ending clean mount 
> [ 8383.935475] XFS (pmem1): User initiated shutdown received. 
> [ 8383.935563] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8383.936730] pmem1: writeback error on inode 6294790, offset 602112, sector 4021136 
> [ 8383.938410] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8383.941017] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8383.977798] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8384.390615] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8384.524097] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8384.553122] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8384.575037] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8384.599489] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8384.720213] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8384.749210] XFS (pmem1): Ending clean mount 
> [ 8386.789264] XFS (pmem1): User initiated shutdown received. 
> [ 8386.795506] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8386.808598] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8387.449606] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8387.656847] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8387.696287] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8389.867734] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8389.896231] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8390.023586] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8390.051077] XFS (pmem1): Ending clean mount 
> [ 8392.091249] XFS (pmem1): User initiated shutdown received. 
> [ 8392.097193] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8392.110286] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8392.747591] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8392.970949] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8393.010107] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8392.009668] restraintd[2003]: *** Current Time: Tue Jun 13 08:49:40 2023  Localwatchdog at: Thu Jun 15 06:31:39 2023 
> [ 8394.908469] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8396.954732] XFS (pmem1): User initiated shutdown received. 
> [ 8396.960271] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8396.973361] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8397.214607] ================================================================== 
> [ 8397.221835] BUG: KASAN: slab-use-after-free in io_wq_worker_running+0xd9/0xf0 
> [ 8397.228976] Read of size 4 at addr ffff888609f48404 by task iou-wrk-1340274/1340276 
> [ 8397.236629]  
> [ 8397.238128] CPU: 15 PID: 1340276 Comm: iou-wrk-1340274 Kdump: loaded Not tainted 6.4.0-rc6+ #1 
> [ 8397.246733] Hardware name: Dell Inc. PowerEdge R6515/07PXPY, BIOS 2.3.6 07/06/2021 
> [ 8397.254300] Call Trace: 
> [ 8397.256750]  <TASK> 
> [ 8397.258858]  dump_stack_lvl+0x60/0xb0 
> [ 8397.262525]  print_address_description.constprop.0+0x2c/0x3e0 
> [ 8397.268278]  print_report+0xb5/0x270 
> [ 8397.271859]  ? kasan_addr_to_slab+0x9/0xa0 
> [ 8397.275955]  ? io_wq_worker_running+0xd9/0xf0 
> [ 8397.280317]  kasan_report+0x8c/0xc0 
> [ 8397.283808]  ? io_wq_worker_running+0xd9/0xf0 
> [ 8397.288170]  ? do_exit+0x329/0xf70 
> [ 8397.291572]  io_wq_worker_running+0xd9/0xf0 
> [ 8397.295758]  do_exit+0x329/0xf70 
> [ 8397.298993]  ? __pfx_kvfree_call_rcu+0x10/0x10 
> [ 8397.303438]  ? __pfx_do_exit+0x10/0x10 
> [ 8397.307192]  ? io_wq_worker+0x668/0xa70 
> [ 8397.311031]  io_wq_worker+0x6c7/0xa70 
> [ 8397.314700]  ? __pfx_io_wq_worker+0x10/0x10 
> [ 8397.318881]  ? __pfx___lock_release+0x10/0x10 
> [ 8397.323241]  ? __pfx_do_raw_spin_trylock+0x10/0x10 
> [ 8397.328035]  ? __pfx_io_wq_worker+0x10/0x10 
> [ 8397.332221]  ret_from_fork+0x2c/0x50 
> [ 8397.335803]  </TASK> 
> [ 8397.337992]  
> [ 8397.339488] Allocated by task 1340274: 
> [ 8397.343242]  kasan_save_stack+0x1e/0x40 
> [ 8397.347081]  kasan_set_track+0x21/0x30 
> [ 8397.350834]  __kasan_kmalloc+0x7b/0x90 
> [ 8397.354586]  create_io_worker+0xbe/0x520 
> [ 8397.358513]  io_wq_enqueue+0x57e/0xa30 
> [ 8397.362264]  io_queue_iowq+0x226/0x450 
> [ 8397.366019]  io_queue_async+0x123/0x480 
> [ 8397.369857]  io_submit_sqes+0xac8/0xeb0 
> [ 8397.373695]  __do_sys_io_uring_enter+0x4f3/0x840 
> [ 8397.378314]  do_syscall_64+0x5c/0x90 
> [ 8397.381896]  entry_SYSCALL_64_after_hwframe+0x72/0xdc 
> [ 8397.386947]  
> [ 8397.388448] Freed by task 1339903: 
> [ 8397.391852]  kasan_save_stack+0x1e/0x40 
> [ 8397.395692]  kasan_set_track+0x21/0x30 
> [ 8397.399443]  kasan_save_free_info+0x2a/0x50 
> [ 8397.403631]  __kasan_slab_free+0x106/0x190 
> [ 8397.407729]  slab_free_freelist_hook+0x11d/0x1d0 
> [ 8397.412349]  kmem_cache_free_bulk.part.0+0x1f3/0x3c0 
> [ 8397.417313]  kvfree_rcu_bulk+0x3ce/0x4e0 
> [ 8397.421239]  kvfree_rcu_drain_ready+0x24f/0x760 
> [ 8397.425774]  kfree_rcu_monitor+0x44/0xe10 
> [ 8397.429784]  process_one_work+0x8c2/0x1570 
> [ 8397.433884]  worker_thread+0x5be/0xef0 
> [ 8397.437636]  kthread+0x2f4/0x3d0 
> [ 8397.440868]  ret_from_fork+0x2c/0x50 
> [ 8397.444450]  
> [ 8397.445949] Last potentially related work creation: 
> [ 8397.450827]  kasan_save_stack+0x1e/0x40 
> [ 8397.454668]  __kasan_record_aux_stack+0x97/0xb0 
> [ 8397.459198]  kvfree_call_rcu+0x103/0x480 
> [ 8397.463125]  io_wq_worker+0x6b3/0xa70 
> [ 8397.466789]  ret_from_fork+0x2c/0x50 
> [ 8397.470369]  
> [ 8397.471868] The buggy address belongs to the object at ffff888609f48400 
> [ 8397.471868]  which belongs to the cache kmalloc-512 of size 512 
> [ 8397.484375] The buggy address is located 4 bytes inside of 
> [ 8397.484375]  freed 512-byte region [ffff888609f48400, ffff888609f48600) 
> [ 8397.496447]  
> [ 8397.497946] The buggy address belongs to the physical page: 
> [ 8397.503520] page:00000000f81b3055 refcount:1 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x609f48 
> [ 8397.512907] head:00000000f81b3055 order:3 entire_mapcount:0 nr_pages_mapped:0 pincount:0 
> [ 8397.520990] flags: 0x17ffffc0010200(slab|head|node=0|zone=2|lastcpupid=0x1fffff) 
> [ 8397.528384] page_type: 0xffffffff() 
> [ 8397.531877] raw: 0017ffffc0010200 ffff88860004cc80 dead000000000100 dead000000000122 
> [ 8397.539616] raw: 0000000000000000 0000000000200020 00000001ffffffff 0000000000000000 
> [ 8397.547354] page dumped because: kasan: bad access detected 
> [ 8397.552927]  
> [ 8397.554425] Memory state around the buggy address: 
> [ 8397.559218]  ffff888609f48300: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc 
> [ 8397.566437]  ffff888609f48380: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc 
> [ 8397.573657] >ffff888609f48400: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb 
> [ 8397.580877]                    ^ 
> [ 8397.584109]  ffff888609f48480: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb 
> [ 8397.591329]  ffff888609f48500: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb 
> [ 8397.598546] ================================================================== 
> [ 8397.605801] Disabling lock debugging due to kernel taint 
> [ 8397.852031] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8397.997724] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8398.024623] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8399.109910] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8399.127813] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8399.205606] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8399.225964] XFS (pmem1): Ending clean mount 
> [ 8400.251290] XFS (pmem1): User initiated shutdown received. 
> [ 8400.256812] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8400.270262] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8400.604837] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8400.762347] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8400.788685] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8401.455999] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8403.485499] XFS (pmem1): User initiated shutdown received. 
> [ 8403.491496] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8403.504570] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8403.918900] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8404.071291] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8404.102643] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8406.143287] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8406.161472] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8406.242454] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8406.262542] XFS (pmem1): Ending clean mount 
> [ 8406.287327] XFS (pmem1): User initiated shutdown received. 
> [ 8406.287350] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8406.287468] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8406.292852] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8406.293069] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8406.524860] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8406.626770] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8406.647060] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8406.663675] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8406.680155] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8406.756538] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8406.777642] XFS (pmem1): Ending clean mount 
> [ 8406.802489] XFS (pmem1): User initiated shutdown received. 
> [ 8406.802512] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8406.803211] XFS (pmem1): xfs_imap_lookup: xfs_ialloc_read_agi() returned error -5, agno 0 
> [ 8406.808007] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8406.837523] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8407.073851] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8407.182295] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8407.201895] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8407.214975] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8407.231390] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8407.317269] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8407.340883] XFS (pmem1): Ending clean mount 
> [ 8409.367221] XFS (pmem1): User initiated shutdown received. 
> [ 8409.372743] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8409.386050] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8409.769341] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8409.934824] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8409.964592] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8411.426464] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8412.455356] XFS (pmem1): User initiated shutdown received. 
> [ 8412.460875] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8412.473949] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8412.810227] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8412.973145] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8412.998560] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8413.455386] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8414.483856] XFS (pmem1): User initiated shutdown received. 
> [ 8414.489379] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8414.502520] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8414.848931] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8415.007610] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8415.033772] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8415.443346] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8415.461155] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8415.553187] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8415.573764] XFS (pmem1): Ending clean mount 
> [ 8416.598523] XFS (pmem1): User initiated shutdown received. 
> [ 8416.604049] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8416.617127] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8416.934096] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.106652] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.128596] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8417.254168] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8417.271833] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.346746] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.369479] XFS (pmem1): Ending clean mount 
> [ 8417.395549] XFS (pmem1): User initiated shutdown received. 
> [ 8417.395688] pmem1: writeback error on inode 6639562, offset 245760, sector 4396792 
> [ 8417.401089] pmem1: writeback error on inode 6639562, offset 520192, sector 4397328 
> [ 8417.401099] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8417.408700] pmem1: writeback error on inode 379851, offset 696320, sector 1794184 
> [ 8417.416234] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8417.444058] pmem1: writeback error on inode 8519238, offset 1024000, sector 5373600 
> [ 8417.444069] pmem1: writeback error on inode 8519238, offset 1622016, sector 5406856 
> [ 8417.645652] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.765394] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.786756] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8417.807733] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8417.824673] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.912097] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8417.934861] XFS (pmem1): Ending clean mount 
> [ 8418.960516] XFS (pmem1): User initiated shutdown received. 
> [ 8418.966034] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8418.979113] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8419.309546] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8419.470480] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8419.494082] XFS (pmem1): Starting recovery (logdev: internal) 
> [ 8419.654738] XFS (pmem1): Ending recovery (logdev: internal) 
> [ 8419.672303] XFS (pmem1): Unmounting Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8419.758252] XFS (pmem1): Mounting V5 Filesystem ce96d037-325c-4e7d-9515-221f48f2ee33 
> [ 8419.780023] XFS (pmem1): Ending clean mount 
> [ 8421.311883] BUG: kernel NULL pointer dereference, address: 0000000000000008 
> [ 8421.318856] #PF: supervisor read access in kernel mode 
> [ 8421.323992] #PF: error_code(0x0000) - not-present page 
> [ 8421.329132] PGD 9327c1067 P4D 9327c1067 PUD 92593f067 PMD 0  
> [ 8421.334799] Oops: 0000 [#1] PREEMPT SMP KASAN NOPTI 
> [ 8421.339678] CPU: 2 PID: 1340923 Comm: fsstress Kdump: loaded Tainted: G    B              6.4.0-rc6+ #1 
> [ 8421.349063] Hardware name: Dell Inc. PowerEdge R6515/07PXPY, BIOS 2.3.6 07/06/2021 
> [ 8421.356629] RIP: 0010:qlist_free_all+0xd3/0x1a0 
> [ 8421.361162] Code: 03 05 d9 71 ad 02 48 8b 48 08 48 89 c2 f6 c1 01 0f 85 ba 00 00 00 0f 1f 44 00 00 48 8b 02 f6 c4 02 b8 00 00 00 00 48 0f 44 d0 <4c> 8b 72 08 e9 50 ff ff ff 49 83 7e 48 00 0f 85 68 ff ff ff 41 f7 
> [ 8421.379906] RSP: 0018:ffffc9000a867438 EFLAGS: 00010246 
> [ 8421.385133] RAX: 0000000000000000 RBX: ffff888b117c0800 RCX: ffffea002c45f008 
> [ 8421.392267] RDX: 0000000000000000 RSI: ffffea00223c6400 RDI: 0000000040000000 
> [ 8421.399399] RBP: 0000000000000000 R08: ffff888b117c0800 R09: 0000000000200012 
> [ 8421.406530] R10: ffffea00223c6400 R11: 0000000000000000 R12: dffffc0000000000 
> [ 8421.413663] R13: ffffc9000a867478 R14: 0000000000000000 R15: ffff88888f194c00 
> [ 8421.420795] FS:  00007f5b21032740(0000) GS:ffff88902de00000(0000) knlGS:0000000000000000 
> [ 8421.428881] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033 
> [ 8421.434627] CR2: 0000000000000008 CR3: 00000007f8c72000 CR4: 0000000000350ee0 
> [ 8421.441759] Call Trace: 
> [ 8421.444213]  <TASK> 
> [ 8421.446318]  ? __die+0x20/0x70 
> [ 8421.449378]  ? page_fault_oops+0x111/0x220 
> [ 8421.453477]  ? __pfx_page_fault_oops+0x10/0x10 
> [ 8421.457923]  ? do_raw_spin_trylock+0xb5/0x180 
> [ 8421.462285]  ? do_user_addr_fault+0xa5/0xda0 
> [ 8421.466554]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.470483]  ? exc_page_fault+0x5a/0xe0 
> [ 8421.474320]  ? asm_exc_page_fault+0x22/0x30 
> [ 8421.478509]  ? qlist_free_all+0xd3/0x1a0 
> [ 8421.482433]  ? qlist_free_all+0x6d/0x1a0 
> [ 8421.486360]  kasan_quarantine_reduce+0x18b/0x1d0 
> [ 8421.490979]  __kasan_slab_alloc+0x45/0x70 
> [ 8421.494989]  kmem_cache_alloc+0x17c/0x3d0 
> [ 8421.499006]  xfs_inobt_init_common+0x88/0x370 [xfs] 
> [ 8421.504082]  xfs_inobt_cur+0xf6/0x2d0 [xfs] 
> [ 8421.508460]  xfs_iwalk_run_callbacks+0x345/0x540 [xfs] 
> [ 8421.513808]  xfs_iwalk_ag+0x60b/0x890 [xfs] 
> [ 8421.518191]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.522120]  ? __pfx_xfs_iwalk_ag+0x10/0x10 [xfs] 
> [ 8421.527020]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.530947]  ? __kmalloc+0xf9/0x190 
> [ 8421.534441]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.538364]  ? lock_release+0x258/0x300 
> [ 8421.542207]  xfs_iwalk+0x233/0x3e0 [xfs] 
> [ 8421.546324]  ? __pfx_xfs_iwalk+0x10/0x10 [xfs] 
> [ 8421.550967]  ? kmem_cache_alloc+0x38d/0x3d0 
> [ 8421.555155]  ? __pfx_xfs_bulkstat_iwalk+0x10/0x10 [xfs] 
> [ 8421.560570]  ? xfs_trans_alloc_empty+0x7d/0xb0 [xfs] 
> [ 8421.565734]  ? __pfx_xfs_trans_alloc_empty+0x10/0x10 [xfs] 
> [ 8421.571413]  xfs_bulkstat+0x2cd/0x460 [xfs] 
> [ 8421.575797]  ? __pfx_xfs_bulkstat+0x10/0x10 [xfs] 
> [ 8421.580710]  ? __pfx_xfs_fsbulkstat_one_fmt+0x10/0x10 [xfs] 
> [ 8421.586474]  ? __might_fault+0xc5/0x170 
> [ 8421.590312]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.594237]  ? lock_release+0x258/0x300 
> [ 8421.598080]  xfs_ioc_fsbulkstat.isra.0+0x214/0x380 [xfs] 
> [ 8421.603608]  ? __pfx_xfs_ioc_fsbulkstat.isra.0+0x10/0x10 [xfs] 
> [ 8421.609630]  ? xfs_ioc_fsbulkstat.isra.0+0x26d/0x380 [xfs] 
> [ 8421.615308]  ? __pfx_xfs_ioc_fsbulkstat.isra.0+0x10/0x10 [xfs] 
> [ 8421.621332]  ? lock_release+0x258/0x300 
> [ 8421.625174]  xfs_file_ioctl+0xfe0/0x1360 [xfs] 
> [ 8421.629825]  ? __pfx_xfs_file_ioctl+0x10/0x10 [xfs] 
> [ 8421.634895]  ? __pfx_xfs_file_ioctl+0x10/0x10 [xfs] 
> [ 8421.639961]  ? xfs_file_ioctl+0xfe0/0x1360 [xfs] 
> [ 8421.644775]  ? __pfx_xfs_file_ioctl+0x10/0x10 [xfs] 
> [ 8421.649843]  ? __pfx_do_raw_spin_trylock+0x10/0x10 
> [ 8421.654634]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.658559]  ? lock_acquire+0x4d6/0x5e0 
> [ 8421.662400]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.666325]  ? lock_acquire+0x4d6/0x5e0 
> [ 8421.670165]  ? __pfx_lock_acquire+0x10/0x10 
> [ 8421.674352]  ? __pfx_do_vfs_ioctl+0x10/0x10 
> [ 8421.678540]  ? ioctl_has_perm.constprop.0.isra.0+0x26a/0x430 
> [ 8421.684194]  ? __pfx_lock_acquire+0x10/0x10 
> [ 8421.688383]  ? __pfx_ioctl_has_perm.constprop.0.isra.0+0x10/0x10 
> [ 8421.694388]  ? check_prev_add+0x1bf5/0x2110 
> [ 8421.698574]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.702501]  ? rcu_is_watching+0x11/0xb0 
> [ 8421.706424]  ? lock_release+0x258/0x300 
> [ 8421.710266]  ? __fget_files+0x1d0/0x3f0 
> [ 8421.714105]  ? security_file_ioctl+0x50/0x90 
> [ 8421.718377]  __x64_sys_ioctl+0x12b/0x1a0 
> [ 8421.722302]  do_syscall_64+0x5c/0x90 
> [ 8421.725882]  ? trace_hardirqs_on_prepare+0x19/0x30 
> [ 8421.730675]  ? do_syscall_64+0x69/0x90 
> [ 8421.734427]  ? trace_irq_enable.constprop.0+0x13d/0x180 
> [ 8421.739652]  ? trace_hardirqs_on_prepare+0x19/0x30 
> [ 8421.744444]  ? do_syscall_64+0x69/0x90 
> [ 8421.748197]  ? do_syscall_64+0x69/0x90 
> [ 8421.751949]  entry_SYSCALL_64_after_hwframe+0x72/0xdc 
> [ 8421.757002] RIP: 0033:0x7f5b20e3ec6b 
> [ 8421.760582] Code: 73 01 c3 48 8b 0d b5 b1 1b 00 f7 d8 64 89 01 48 83 c8 ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa b8 10 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 85 b1 1b 00 f7 d8 64 89 01 48 
> [ 8421.779326] RSP: 002b:00007ffca12282d8 EFLAGS: 00000246 ORIG_RAX: 0000000000000010 
> [ 8421.786892] RAX: ffffffffffffffda RBX: 0000000000001dd8 RCX: 00007f5b20e3ec6b 
> [ 8421.794025] RDX: 00007ffca12282f0 RSI: ffffffffc0205865 RDI: 0000000000000004 
> [ 8421.801156] RBP: 0000000001186400 R08: 0000000000000002 R09: 0000000000000000 
> [ 8421.805134] XFS (pmem1): User initiated shutdown received. 
> [ 8421.808290] R10: 0000000000000000 R11: 0000000000000246 R12: 00000000000002fc 
> [ 8421.808292] R13: 00000000000004fe R14: 0000000000000004 R15: 00007f5b210326c0 
> [ 8421.808299]  </TASK> 
> [ 8421.808300] Modules linked in: ext4 
> [ 8421.813804] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xe1/0x180 [xfs] (fs/xfs/xfs_fsops.c:486).  Shutting down filesystem. 
> [ 8421.820916]  mbcache jbd2 loop rfkill sunrpc intel_rapl_msr intel_rapl_common amd64_edac edac_mce_amd 
> [ 8421.828065] XFS (pmem1): Please unmount the filesystem and rectify the problem(s) 
> [ 8421.830251]  kvm_amd mlx5_ib ipmi_ssif kvm mgag200 dell_smbios i2c_algo_bit drm_shmem_helper dcdbas ib_uverbs drm_kms_helper irqbypass nd_pmem wmi_bmof dax_pmem dell_wmi_descriptor pcspkr rapl syscopyarea sysfillrect sysimgblt ib_core acpi_ipmi ptdma k10temp i2c_piix4 ipmi_si ipmi_devintf ipmi_msghandler acpi_power_meter drm fuse xfs libcrc32c sd_mod t10_pi sg mlx5_core ahci crct10dif_pclmul libahci mlxfw crc32_pclmul crc32c_intel tls libata ghash_clmulni_intel tg3 psample ccp megaraid_sas pci_hyperv_intf sp5100_tco wmi dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_debug] 
> [ 8421.915124] CR2: 0000000000000008 
> [ 8421.918441] ---[ end trace 0000000000000000 ]--- 
> [ 8421.923060] RIP: 0010:qlist_free_all+0xd3/0x1a0 
> [ 8421.927593] Code: 03 05 d9 71 ad 02 48 8b 48 08 48 89 c2 f6 c1 01 0f 85 ba 00 00 00 0f 1f 44 00 00 48 8b 02 f6 c4 02 b8 00 00 00 00 48 0f 44 d0 <4c> 8b 72 08 e9 50 ff ff ff 49 83 7e 48 00 0f 85 68 ff ff ff 41 f7 
> [ 8421.946339] RSP: 0018:ffffc9000a867438 EFLAGS: 00010246 
> [ 8421.951565] RAX: 0000000000000000 RBX: ffff888b117c0800 RCX: ffffea002c45f008 
> [ 8421.958695] RDX: 0000000000000000 RSI: ffffea00223c6400 RDI: 0000000040000000 
> [ 8421.965829] RBP: 0000000000000000 R08: ffff888b117c0800 R09: 0000000000200012 
> [ 8421.972962] R10: ffffea00223c6400 R11: 0000000000000000 R12: dffffc0000000000 
> [ 8421.980093] R13: ffffc9000a867478 R14: 0000000000000000 R15: ffff88888f194c00 
> [ 8421.987227] FS:  00007f5b21032740(0000) GS:ffff88902de00000(0000) knlGS:0000000000000000 
> [ 8421.995312] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033 
> [ 8422.001057] CR2: 0000000000000008 CR3: 00000007f8c72000 CR4: 0000000000350ee0 
> [ 8422.008192] note: fsstress[1340923] exited with irqs disabled 
> ...
> ...
> 
> > 
> > diff --git a/io_uring/io-wq.c b/io_uring/io-wq.c
> > index b2715988791e..fe38eb0cbc82 100644
> > --- a/io_uring/io-wq.c
> > +++ b/io_uring/io-wq.c
> > @@ -221,9 +221,6 @@ static void io_worker_exit(struct io_worker *worker)
> >  	raw_spin_unlock(&wq->lock);
> >  	io_wq_dec_running(worker);
> >  	worker->flags = 0;
> > -	preempt_disable();
> > -	current->flags &= ~PF_IO_WORKER;
> > -	preempt_enable();
> >  
> >  	kfree_rcu(worker, rcu);
> >  	io_worker_ref_put(wq);
> > 
> > -- 
> > Jens Axboe
> > 


  reply	other threads:[~2023-06-14  1:04 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-06-12 18:11 [PATCH] io_uring/io-wq: don't clear PF_IO_WORKER on exit Jens Axboe
2023-06-14  0:54 ` Zorro Lang
2023-06-14  1:03   ` Zorro Lang [this message]
2023-06-14  1:14   ` Jens Axboe
2023-06-14  4:49     ` Zorro Lang
2023-06-14 17:44     ` Linus Torvalds
2023-06-14 19:25       ` Jens Axboe
2023-06-15  2:22     ` Zorro Lang
2023-06-15  2:23       ` Jens Axboe

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20230614010348.2o6wjax543teosph@zlang-mailbox \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    [email protected] \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox