From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-12.3 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI, MENTIONS_GIT_HOSTING,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 9ED58C433ED for ; Thu, 22 Apr 2021 14:27:07 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 6293C6144A for ; Thu, 22 Apr 2021 14:27:07 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236428AbhDVO1k (ORCPT ); Thu, 22 Apr 2021 10:27:40 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:36086 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S236398AbhDVO1i (ORCPT ); Thu, 22 Apr 2021 10:27:38 -0400 Received: from hr2.samba.org (hr2.samba.org [IPv6:2a01:4f8:192:486::2:0]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 745A0C06174A; Thu, 22 Apr 2021 07:27:01 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=samba.org; s=42; h=Date:Message-ID:From:Cc:To; bh=4Kb9zfRmA2ZmFuTqllA0SoHgtgUqa0tGA+vIBLYFziw=; b=sm00KaolnOUY/MttJ+l1D+2AQR xIG86ZNCWPUvBbDFabP5V3SkAUe2jbxaRFOt78RC36JP7g5aO9GJkZAqKu9Ssl5EAlZwhY7U4xNXY 5qiW0E2vGE0RnHvazIBxCWoqTT+1d5FYe6U1cOfF97kDsLhHfIsrCf/moo1ToXXoWoxYAEQiKhGcD 8ML6riR3uyR3DepNGjry2tjHVFh6TWKrGmdpITy27nl4dXtCf55WuXa1OO6rCN49Doi9GnXGOtypB K8XBZAIWcf1vmdSHDHhn/nbyvutiT8VyY5LOzF81RAwSiroZ88rQGdXqaPCR+YKMMWAYTNnTjdAzy AHk77OX/PPoW+PY7cebiU+NpFwV1JpeXL7E9z7OH85qytu/ha78EK1W+8eERrf7solt0Oen3v4LUD 3uRdXyXSyOHa9f/5BLgvAOkHJ4CNVB7oWwFccik6U/fBSp1f0ru/Ilf4q5enR+HWj9uhLk7yD4BgG 6I91CvMLM+813cekxJpBdEgx; Received: from [127.0.0.2] (localhost [127.0.0.1]) by hr2.samba.org with esmtpsa (TLS1.3:ECDHE_RSA_CHACHA20_POLY1305:256) (Exim) id 1lZaID-0000dI-Sn; Thu, 22 Apr 2021 14:26:57 +0000 To: Steven Rostedt , Ingo Molnar Cc: linux-trace-devel@vger.kernel.org, io-uring From: Stefan Metzmacher Subject: Tracing busy processes/threads freezes/stalls the whole machine Message-ID: <293cfb1d-8a53-21e1-83c1-cdb6e2f32c65@samba.org> Date: Thu, 22 Apr 2021 16:26:57 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.7.1 MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: io-uring@vger.kernel.org Hi Steven, hi Ingo, I recently tried to analyze the performance of Samba using io-uring. I was using ubuntu 20.04 with the 5.10.0-1023-oem kernel, which is based on v5.10.25, see: https://kernel.ubuntu.com/git/kernel-ppa/mirror/ubuntu-oem-5.10-focal.git/log/?h=oem-5.10-prep trace-cmd is at version 2.8.3-4build1. In order to find the bottleneck I tried to use (trace-cmd is at version 2.8.3-4build1): trace-cmd -e all -P ${pid_of_io_uring_worker} As a result the server was completely dead immediately. I tried to reproduce this in a virtual machine (inside virtualbox). I used a modified 'io_uring-cp' that loops forever, see: https://github.com/metze-samba/liburing/commit/5e98efed053baf03521692e786c1c55690b04d8e When I run './io_uring-cp-forever link-cp.c file', I see a 'io_wq_manager' and a 'io_wqe_worker-0' kernel thread, while './io_uring-cp-forever link-cp.c file' as well as 'io_wqe_worker-0' consume about 25% cpu each. When I run 'trace-cmd -e all -P $pid' for 'io_wqe_worker-0' or 'io_wq_manager' I can reproduce the problem, then I found that the same seems to happen for also for other kernel threads e.g. '[kworker/1:1-events]', it seems that it happens for all kernel threads, which are not completely idle. Which this: From 'top': 1341 root 20 0 2512 576 508 S 33,4 0,1 0:10.39 io_uring-cp-for 1343 root 20 0 0 0 0 R 29,8 0,0 0:08.43 io_wqe_worker-0 7 root 20 0 0 0 0 I 0,3 0,0 0:00.31 kworker/0:1-events PID 5 is [kworker/0:0-ata_sff] # trace-cmd record -e all -P 5' Hit Ctrl^C to stop recording ^CCPU0 data recorded at offset=0x7b8000 0 bytes in size CPU1 data recorded at offset=0x7b8000 69632 bytes in size # But # trace-cmd record -e all -P 7 => machine unresponsive (no blinking cursor on the console anymore) On the host 'top' shows that the VirtualBoxVM cpu emulator thread 'EMT-1' uses 100% cpu, so I guess the guest kernel is in something like an endless recursion loop. Maybe a trace function recursing to itself? On the same VM I tried a 5.12rc8 kernel and there I can also reproduce the problem. I also managed to reproduce the problem without io-uring, just using: while true; do cat linux-image-5.12.0-rc8-dbg_5.12.0-rc8-22_amd64.deb > /dev/null; done in order to keep some kernel threads moving. This happens with 5.12rc8 and 5.10.0-1023-oem, but I was not able to reproduce any of this using the 5.8.0-50-generic kernel, see https://kernel.ubuntu.com/git/ubuntu/ubuntu-focal.git/log/?h=Ubuntu-hwe-5.8-5.8.0-50.56_20.04.1 I was also able to reproduce this with a ubuntu 21.04 vm using the 5.11.0-14-generic kernel, see: https://kernel.ubuntu.com/git/ubuntu/ubuntu-hirsute.git/log/?h=Ubuntu-5.11.0-14.15 On this one I only managed to reproduce the problem with './io_uring-cp-forever link-cp.c file', but not with 'while true; do cat linux-image-5.12.0-rc8-dbg_5.12.0-rc8-22_amd64.deb > /dev/null; done' So it seems the problem was introduced after 5.8 and is not really related to io-uring. And it may not be purely related to kernel threads. With this on 5.12-rc8 (again): └─tmux: server,903 ├─bash,904 │ └─io_uring-cp-for,925 link-cp.c file │ ├─{iou-mgr-925},926 │ └─{iou-wrk-925},927 └─bash,929 └─pstree,938 -a -t -p I was able to to trace once: root@ub1704-166:~# trace-cmd record -e all -P 925 Hit Ctrl^C to stop recording ^CCPU0 data recorded at offset=0x7b8000 10842112 bytes in size CPU1 data recorded at offset=0x120f000 36450304 bytes in size But the 2nd run reproduced the problem: root@ub1704-166:~# trace-cmd record -e all -P 925 I was also able to reproduce it with: while true; do cat linux-image-5.12.0-rc8-dbg_5.12.0-rc8-22_amd64.deb > /dev/null; done and pidofcat=$(pidof cat); echo $pidofcat; trace-cmd record -e all -P $pidofcat So it seems any busy thread (user or kernel) triggers the problem. Any ideas what has changed after 5.8? Thanks! metze