bpf, arm64: remove prefetch insn in xadd mapping
authorDaniel Borkmann <daniel@iogearbox.net>
Fri, 26 Apr 2019 19:48:21 +0000 (21:48 +0200)
committerAlexei Starovoitov <ast@kernel.org>
Sat, 27 Apr 2019 01:53:15 +0000 (18:53 -0700)
Prefetch-with-intent-to-write is currently part of the XADD mapping in
the AArch64 JIT and follows the kernel's implementation of atomic_add.
This may interfere with other threads executing the LDXR/STXR loop,
leading to potential starvation and fairness issues. Drop the optional
prefetch instruction.

Fixes: 85f68fe89832 ("bpf, arm64: implement jiting of BPF_XADD")
Reported-by: Will Deacon <will.deacon@arm.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jean-Philippe Brucker <jean-philippe.brucker@arm.com>
Acked-by: Will Deacon <will.deacon@arm.com>
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
arch/arm64/net/bpf_jit.h
arch/arm64/net/bpf_jit_comp.c

index 783de51a6c4e5b33725d4904be6b09011dd48605..6c881659ee8a1d39db8de6a90b97e994a720d90b 100644 (file)
 #define A64_STXR(sf, Rt, Rn, Rs) \
        A64_LSX(sf, Rt, Rn, Rs, STORE_EX)
 
-/* Prefetch */
-#define A64_PRFM(Rn, type, target, policy) \
-       aarch64_insn_gen_prefetch(Rn, AARCH64_INSN_PRFM_TYPE_##type, \
-                                 AARCH64_INSN_PRFM_TARGET_##target, \
-                                 AARCH64_INSN_PRFM_POLICY_##policy)
-
 /* Add/subtract (immediate) */
 #define A64_ADDSUB_IMM(sf, Rd, Rn, imm12, type) \
        aarch64_insn_gen_add_sub_imm(Rd, Rn, imm12, \
index aaddc0217e73dfce56b44f2e7f19c26a64ac9c2b..a1420626fca272a39a66f975ef5233c398cef617 100644 (file)
@@ -762,7 +762,6 @@ emit_cond_jmp:
        case BPF_STX | BPF_XADD | BPF_DW:
                emit_a64_mov_i(1, tmp, off, ctx);
                emit(A64_ADD(1, tmp, tmp, dst), ctx);
-               emit(A64_PRFM(tmp, PST, L1, STRM), ctx);
                emit(A64_LDXR(isdw, tmp2, tmp), ctx);
                emit(A64_ADD(isdw, tmp2, tmp2, src), ctx);
                emit(A64_STXR(isdw, tmp2, tmp, tmp3), ctx);