summaryrefslogtreecommitdiffstats
path: root/tools/arch
diff options
context:
space:
mode:
authorSteven Rostedt (Google) <rostedt@goodmis.org>2023-05-31 15:24:19 +0200
committerBorislav Petkov (AMD) <bp@alien8.de>2023-06-14 18:50:00 +0200
commit9350a629e839ca1c2b529a83a916cf2370bd1c64 (patch)
tree98fec5701d857871dda0824f7c8d394b6a4b56a3 /tools/arch
parentx86/nospec: Shorten RESET_CALL_DEPTH (diff)
downloadlinux-9350a629e839ca1c2b529a83a916cf2370bd1c64.tar.xz
linux-9350a629e839ca1c2b529a83a916cf2370bd1c64.zip
x86/alternatives: Add cond_resched() to text_poke_bp_batch()
Debugging in the kernel has started slowing down the kernel by a noticeable amount. The ftrace start up tests are triggering the softlockup watchdog on some boxes. This is caused by the start up tests that enable function and function graph tracing several times. Sprinkling cond_resched() just in the start up test code was not enough to stop the softlockup from triggering. It would sometimes trigger in the text_poke_bp_batch() code. When function tracing enables all functions, it will call text_poke_queue() to queue the places that need to be patched. Every 256 entries will do a "flush" that calls text_poke_bp_batch() to do the update of the 256 locations. As this is in a scheduleable context, calling cond_resched() at the start of text_poke_bp_batch() will ensure that other tasks could get a chance to run while the patching is happening. This keeps the softlockup from triggering in the start up tests. Signed-off-by: Steven Rostedt (Google) <rostedt@goodmis.org> Signed-off-by: Borislav Petkov (AMD) <bp@alien8.de> Acked-by: Masami Hiramatsu (Google) <mhiramat@kernel.org> Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: https://lore.kernel.org/r/20230531092419.4d051374@rorschach.local.home
Diffstat (limited to 'tools/arch')
0 files changed, 0 insertions, 0 deletions