Closed
Description
Hi Jens,
I got a test failed with a potential deadlock at <rsrc_tags>
.
Kernel: 5.14.0-rc3-bluetea-test-00249-gc7d102232649
[ 9257.317473] Running test 232c93d07b74-test:
[ 9261.828744] Running test 35fa71a030ca-test:
[ 9321.858788] Running test 500f9fbadef8-test:
[ 9323.308544] Running test 7ad0e4b2f83c-test:
[ 9324.419221] Running test 8a9973408177-test:
[ 9325.635902] Running test 917257daa0fe-test:
[ 9325.755814] Running test a0908ae19763-test:
[ 9325.871406] Running test a4c0b3decb33-test:
[ 9340.684624] Running test accept:
[ 9342.070441] Running test accept-link:
[ 9342.389549] Running test accept-reuse:
[ 9342.511045] Running test accept-test:
[ 9342.627728] Running test across-fork:
[ 9342.745884] Running test splice:
[ 9342.872690] Running test b19062a56726-test:
[ 9342.981963] Running test b5837bd5311d-test:
[ 9343.100657] Running test ce593a6c480a-test:
[ 9344.229048] Running test close-opath:
[ 9344.339398] Running test connect:
[ 9344.363080] TCP: request_sock_TCP: Possible SYN flooding on port 29878. Sending cookies. Check SNMP counters.
[ 9344.453743] Running test cq-full:
[ 9344.563967] Running test cq-overflow:
[ 9355.410900] Running test cq-peek-batch:
[ 9355.522211] Running test cq-ready:
[ 9355.635496] Running test cq-size:
[ 9355.747607] Running test d4ae271dfaae-test:
[ 9356.041987] Running test d77a67ed5f27-test:
[ 9356.303955] Running test defer:
[ 9359.413848] Running test double-poll-crash:
[ 9359.528898] Running test eeed8b54e0df-test:
[ 9359.683170] Running test eventfd:
[ 9359.814646] Running test eventfd-disable:
[ 9359.945202] Running test eventfd-ring:
[ 9360.092271] Running test fadvise:
[ 9360.413969] Running test fallocate:
[ 9360.561731] Running test fc2a85cb02ef-test:
[ 9360.677307] Running test file-register:
[ 9378.660480] Running test file-update:
[ 9381.894220] Running test files-exit-hang-poll:
[ 9382.996801] Running test files-exit-hang-timeout:
[ 9384.095303] Running test fixed-link:
[ 9384.194950] Running test fsync:
[ 9384.422513] Running test hardlink:
[ 9384.528068] Running test io-cancel:
[ 9387.144482] Running test io_uring_enter:
[ 9387.506670] Running test io_uring_register:
[ 9389.784601] Running test io_uring_setup:
[ 9389.892289] Running test iopoll:
[ 9402.967112] Running test lfs-openat:
[ 9403.070135] Running test lfs-openat-write:
[ 9403.173498] Running test link:
[ 9403.313011] Running test link-timeout:
[ 9404.916786] Running test link_drain:
[ 9405.841861] Running test madvise:
[ 9406.152754] Running test mkdir:
[ 9406.256519] Running test multicqes_drain:
[ 9451.376548] Running test nop:
[ 9451.477609] Running test nop-all-sizes:
[ 9451.622920] Running test open-close:
[ 9451.868377] Running test openat2:
[ 9452.092030] Running test personality:
[ 9452.197029] Running test pipe-eof:
[ 9452.301000] Running test pipe-reuse:
[ 9452.409544] Running test poll:
[ 9452.512898] Running test poll-cancel:
[ 9452.613946] Running test poll-cancel-ton:
[ 9453.008100] Running test poll-link:
[ 9453.316323] Running test poll-many:
[ 9465.760035] Running test poll-mshot-update:
[ 9474.184772] Running test poll-ring:
[ 9474.294344] Running test poll-v-poll:
[ 9474.815263] Running test probe:
[ 9474.925634] Running test read-write:
[ 9519.635522] Running test register-restrictions:
[ 9519.995146] Running test rename:
[ 9520.108175] Running test ring-leak:
[ 9520.246420] Running test ring-leak2:
[ 9521.360344] Running test rw_merge_test:
[ 9521.470459] Running test self:
[ 9521.582026] Running test send_recv:
[ 9521.702652] Running test send_recvmsg:
[ 9521.820930] Running test shared-wq:
[ 9521.934517] Running test short-read:
[ 9522.136810] Running test shutdown:
[ 9522.253045] Running test sigfd-deadlock:
[ 9522.372489] Running test socket-rw:
[ 9522.485295] Running test socket-rw-eagain:
[ 9522.597577] Running test sq-full:
[ 9522.707849] Running test sq-poll-dup:
[ 9526.798436] Running test sq-poll-kthread:
[ 9529.006681] Running test sq-poll-share:
[ 9536.551806] Running test sqpoll-disable-exit:
[ 9537.131130] Running test sqpoll-exit-hang:
[ 9538.244750] Running test sqpoll-cancel-hang:
[ 9539.352283] Running test sqpoll-sleep:
[ 9539.466045] Running test sq-space_left:
[ 9539.576885] Running test stdout:
[ 9541.760714] Running test submit-reuse:
[ 9545.929462] Running test symlink:
[ 9546.033826] Running test teardowns:
[ 9546.236248] Running test thread-exit:
[ 9546.423347] Running test timeout:
[ 9552.550575] Running test timeout-new:
[ 9555.062722] Running test timeout-overflow:
[ 9555.168925] Running test unlink:
[ 9555.268847] Running test wakeup-hang:
[ 9557.369264] Running test sendmsg_fs_cve:
[ 9557.476769] Running test rsrc_tags:
[ 9568.748103] ======================================================
[ 9568.748111] WARNING: possible circular locking dependency detected
[ 9568.748120] 5.14.0-rc3-bluetea-test-00249-gc7d102232649 #5 Tainted: G OE
[ 9568.748130] ------------------------------------------------------
[ 9568.748136] kworker/2:4/2684 is trying to acquire lock:
[ 9568.748145] ffff88814bb1c0a8 (&ctx->uring_lock){+.+.}-{3:3}, at: io_rsrc_put_work+0x13d/0x1a0
[ 9568.748183]
but task is already holding lock:
[ 9568.748190] ffffc90001c6be70 ((work_completion)(&(&ctx->rsrc_put_work)->work)){+.+.}-{0:0}, at: process_one_work+0x1bc/0x530
[ 9568.748219]
which lock already depends on the new lock.
[ 9568.748226]
the existing dependency chain (in reverse order) is:
[ 9568.748232]
-> #1 ((work_completion)(&(&ctx->rsrc_put_work)->work)){+.+.}-{0:0}:
[ 9568.748249] __flush_work+0x31b/0x490
[ 9568.748263] io_rsrc_ref_quiesce.part.0.constprop.0+0x35/0xb0
[ 9568.748276] __do_sys_io_uring_register+0x45b/0x1060
[ 9568.748289] do_syscall_64+0x35/0xb0
[ 9568.748301] entry_SYSCALL_64_after_hwframe+0x44/0xae
[ 9568.748316]
-> #0 (&ctx->uring_lock){+.+.}-{3:3}:
[ 9568.748332] __lock_acquire+0x119a/0x1e10
[ 9568.748344] lock_acquire+0xc8/0x2f0
[ 9568.748355] __mutex_lock+0x86/0x740
[ 9568.748365] io_rsrc_put_work+0x13d/0x1a0
[ 9568.748376] process_one_work+0x236/0x530
[ 9568.748388] worker_thread+0x52/0x3b0
[ 9568.748400] kthread+0x135/0x160
[ 9568.748411] ret_from_fork+0x1f/0x30
[ 9568.748426]
other info that might help us debug this:
[ 9568.748432] Possible unsafe locking scenario:
[ 9568.748438] CPU0 CPU1
[ 9568.748443] ---- ----
[ 9568.748448] lock((work_completion)(&(&ctx->rsrc_put_work)->work));
[ 9568.748459] lock(&ctx->uring_lock);
[ 9568.748469] lock((work_completion)(&(&ctx->rsrc_put_work)->work));
[ 9568.748479] lock(&ctx->uring_lock);
[ 9568.748488]
*** DEADLOCK ***
[ 9568.748494] 2 locks held by kworker/2:4/2684:
[ 9568.748503] #0: ffff88810004d938 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x1bc/0x530
[ 9568.748531] #1: ffffc90001c6be70 ((work_completion)(&(&ctx->rsrc_put_work)->work)){+.+.}-{0:0}, at: process_one_work+0x1bc/0x530
[ 9568.748558]
stack backtrace:
[ 9568.748566] CPU: 2 PID: 2684 Comm: kworker/2:4 Tainted: G OE 5.14.0-rc3-bluetea-test-00249-gc7d102232649 #5
[ 9568.748580] Hardware name: Acer Aspire ES1-421/OLVIA_BE, BIOS V1.05 07/02/2015
[ 9568.748589] Workqueue: events io_rsrc_put_work
[ 9568.748605] Call Trace:
[ 9568.748620] dump_stack_lvl+0x6a/0x9a
[ 9568.748639] check_noncircular+0xfe/0x110
[ 9568.748660] __lock_acquire+0x119a/0x1e10
[ 9568.748680] lock_acquire+0xc8/0x2f0
[ 9568.748692] ? io_rsrc_put_work+0x13d/0x1a0
[ 9568.748712] __mutex_lock+0x86/0x740
[ 9568.748724] ? io_rsrc_put_work+0x13d/0x1a0
[ 9568.748738] ? io_rsrc_put_work+0x13d/0x1a0
[ 9568.748756] ? io_rsrc_put_work+0x13d/0x1a0
[ 9568.748767] ? process_one_work+0x1ce/0x530
[ 9568.748779] io_rsrc_put_work+0x13d/0x1a0
[ 9568.748796] process_one_work+0x236/0x530
[ 9568.748816] worker_thread+0x52/0x3b0
[ 9568.748830] ? process_one_work+0x530/0x530
[ 9568.748845] kthread+0x135/0x160
[ 9568.748857] ? set_kthread_struct+0x40/0x40
[ 9568.748872] ret_from_fork+0x1f/0x30
[ 9579.199425] Running test statx:
[ 9579.337521] Running test sq-full-cpp:
Cc: @isilence
Metadata
Metadata
Assignees
Labels
No labels