diff options
author | David Miller <davem@davemloft.net> | 2020-02-24 15:01:43 +0100 |
---|---|---|
committer | Alexei Starovoitov <ast@kernel.org> | 2020-02-25 01:20:09 +0100 |
commit | 3d9f773cf2876c01a505b9fe27270901d464e90a (patch) | |
tree | d5ed0350db81b38cf8c0234d7c066f65595addff /net/kcm | |
parent | bpf: Replace cant_sleep() with cant_migrate() (diff) | |
download | linux-3d9f773cf2876c01a505b9fe27270901d464e90a.tar.xz linux-3d9f773cf2876c01a505b9fe27270901d464e90a.zip |
bpf: Use bpf_prog_run_pin_on_cpu() at simple call sites.
All of these cases are strictly of the form:
preempt_disable();
BPF_PROG_RUN(...);
preempt_enable();
Replace this with bpf_prog_run_pin_on_cpu() which wraps BPF_PROG_RUN()
with:
migrate_disable();
BPF_PROG_RUN(...);
migrate_enable();
On non RT enabled kernels this maps to preempt_disable/enable() and on RT
enabled kernels this solely prevents migration, which is sufficient as
there is no requirement to prevent reentrancy to any BPF program from a
preempting task. The only requirement is that the program stays on the same
CPU.
Therefore, this is a trivially correct transformation.
The seccomp loop does not need protection over the loop. It only needs
protection per BPF filter program
[ tglx: Converted to bpf_prog_run_pin_on_cpu() ]
Signed-off-by: David S. Miller <davem@davemloft.net>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
Link: https://lore.kernel.org/bpf/20200224145643.691493094@linutronix.de
Diffstat (limited to 'net/kcm')
-rw-r--r-- | net/kcm/kcmsock.c | 4 |
1 files changed, 1 insertions, 3 deletions
diff --git a/net/kcm/kcmsock.c b/net/kcm/kcmsock.c index ea9e73428ed9..56fac24a627a 100644 --- a/net/kcm/kcmsock.c +++ b/net/kcm/kcmsock.c @@ -380,9 +380,7 @@ static int kcm_parse_func_strparser(struct strparser *strp, struct sk_buff *skb) struct bpf_prog *prog = psock->bpf_prog; int res; - preempt_disable(); - res = BPF_PROG_RUN(prog, skb); - preempt_enable(); + res = bpf_prog_run_pin_on_cpu(prog, skb); return res; } |