1e6e57d9b3
Writing the outer loop of an LL/SC sequence using do {...} while
constructs potentially allows the compiler to hoist memory accesses
between the STXR and the branch back to the LDXR. On CPUs that do not
guarantee forward progress of LL/SC loops when faced with memory
accesses to the same ERG (up to 2k) between the failed STXR and the
branch back, we may end up livelocking.
This patch avoids this issue in our percpu atomics by rewriting the
outer loop as part of the LL/SC inline assembly block.
Cc: <stable@vger.kernel.org>
Fixes: f97fc81079
("arm64: percpu: Implement this_cpu operations")
Reviewed-by: Mark Rutland <mark.rutland@arm.com>
Tested-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Will Deacon <will.deacon@arm.com>
270 lines
7.3 KiB
C
270 lines
7.3 KiB
C
/*
|
|
* Copyright (C) 2013 ARM Ltd.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
#ifndef __ASM_PERCPU_H
|
|
#define __ASM_PERCPU_H
|
|
|
|
static inline void set_my_cpu_offset(unsigned long off)
|
|
{
|
|
asm volatile("msr tpidr_el1, %0" :: "r" (off) : "memory");
|
|
}
|
|
|
|
static inline unsigned long __my_cpu_offset(void)
|
|
{
|
|
unsigned long off;
|
|
|
|
/*
|
|
* We want to allow caching the value, so avoid using volatile and
|
|
* instead use a fake stack read to hazard against barrier().
|
|
*/
|
|
asm("mrs %0, tpidr_el1" : "=r" (off) :
|
|
"Q" (*(const unsigned long *)current_stack_pointer));
|
|
|
|
return off;
|
|
}
|
|
#define __my_cpu_offset __my_cpu_offset()
|
|
|
|
#define PERCPU_OP(op, asm_op) \
|
|
static inline unsigned long __percpu_##op(void *ptr, \
|
|
unsigned long val, int size) \
|
|
{ \
|
|
unsigned long loop, ret; \
|
|
\
|
|
switch (size) { \
|
|
case 1: \
|
|
asm ("//__per_cpu_" #op "_1\n" \
|
|
"1: ldxrb %w[ret], %[ptr]\n" \
|
|
#asm_op " %w[ret], %w[ret], %w[val]\n" \
|
|
" stxrb %w[loop], %w[ret], %[ptr]\n" \
|
|
" cbnz %w[loop], 1b" \
|
|
: [loop] "=&r" (loop), [ret] "=&r" (ret), \
|
|
[ptr] "+Q"(*(u8 *)ptr) \
|
|
: [val] "Ir" (val)); \
|
|
break; \
|
|
case 2: \
|
|
asm ("//__per_cpu_" #op "_2\n" \
|
|
"1: ldxrh %w[ret], %[ptr]\n" \
|
|
#asm_op " %w[ret], %w[ret], %w[val]\n" \
|
|
" stxrh %w[loop], %w[ret], %[ptr]\n" \
|
|
" cbnz %w[loop], 1b" \
|
|
: [loop] "=&r" (loop), [ret] "=&r" (ret), \
|
|
[ptr] "+Q"(*(u16 *)ptr) \
|
|
: [val] "Ir" (val)); \
|
|
break; \
|
|
case 4: \
|
|
asm ("//__per_cpu_" #op "_4\n" \
|
|
"1: ldxr %w[ret], %[ptr]\n" \
|
|
#asm_op " %w[ret], %w[ret], %w[val]\n" \
|
|
" stxr %w[loop], %w[ret], %[ptr]\n" \
|
|
" cbnz %w[loop], 1b" \
|
|
: [loop] "=&r" (loop), [ret] "=&r" (ret), \
|
|
[ptr] "+Q"(*(u32 *)ptr) \
|
|
: [val] "Ir" (val)); \
|
|
break; \
|
|
case 8: \
|
|
asm ("//__per_cpu_" #op "_8\n" \
|
|
"1: ldxr %[ret], %[ptr]\n" \
|
|
#asm_op " %[ret], %[ret], %[val]\n" \
|
|
" stxr %w[loop], %[ret], %[ptr]\n" \
|
|
" cbnz %w[loop], 1b" \
|
|
: [loop] "=&r" (loop), [ret] "=&r" (ret), \
|
|
[ptr] "+Q"(*(u64 *)ptr) \
|
|
: [val] "Ir" (val)); \
|
|
break; \
|
|
default: \
|
|
BUILD_BUG(); \
|
|
} \
|
|
\
|
|
return ret; \
|
|
}
|
|
|
|
PERCPU_OP(add, add)
|
|
PERCPU_OP(and, and)
|
|
PERCPU_OP(or, orr)
|
|
#undef PERCPU_OP
|
|
|
|
static inline unsigned long __percpu_read(void *ptr, int size)
|
|
{
|
|
unsigned long ret;
|
|
|
|
switch (size) {
|
|
case 1:
|
|
ret = ACCESS_ONCE(*(u8 *)ptr);
|
|
break;
|
|
case 2:
|
|
ret = ACCESS_ONCE(*(u16 *)ptr);
|
|
break;
|
|
case 4:
|
|
ret = ACCESS_ONCE(*(u32 *)ptr);
|
|
break;
|
|
case 8:
|
|
ret = ACCESS_ONCE(*(u64 *)ptr);
|
|
break;
|
|
default:
|
|
BUILD_BUG();
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline void __percpu_write(void *ptr, unsigned long val, int size)
|
|
{
|
|
switch (size) {
|
|
case 1:
|
|
ACCESS_ONCE(*(u8 *)ptr) = (u8)val;
|
|
break;
|
|
case 2:
|
|
ACCESS_ONCE(*(u16 *)ptr) = (u16)val;
|
|
break;
|
|
case 4:
|
|
ACCESS_ONCE(*(u32 *)ptr) = (u32)val;
|
|
break;
|
|
case 8:
|
|
ACCESS_ONCE(*(u64 *)ptr) = (u64)val;
|
|
break;
|
|
default:
|
|
BUILD_BUG();
|
|
}
|
|
}
|
|
|
|
static inline unsigned long __percpu_xchg(void *ptr, unsigned long val,
|
|
int size)
|
|
{
|
|
unsigned long ret, loop;
|
|
|
|
switch (size) {
|
|
case 1:
|
|
asm ("//__percpu_xchg_1\n"
|
|
"1: ldxrb %w[ret], %[ptr]\n"
|
|
" stxrb %w[loop], %w[val], %[ptr]\n"
|
|
" cbnz %w[loop], 1b"
|
|
: [loop] "=&r"(loop), [ret] "=&r"(ret),
|
|
[ptr] "+Q"(*(u8 *)ptr)
|
|
: [val] "r" (val));
|
|
break;
|
|
case 2:
|
|
asm ("//__percpu_xchg_2\n"
|
|
"1: ldxrh %w[ret], %[ptr]\n"
|
|
" stxrh %w[loop], %w[val], %[ptr]\n"
|
|
" cbnz %w[loop], 1b"
|
|
: [loop] "=&r"(loop), [ret] "=&r"(ret),
|
|
[ptr] "+Q"(*(u16 *)ptr)
|
|
: [val] "r" (val));
|
|
break;
|
|
case 4:
|
|
asm ("//__percpu_xchg_4\n"
|
|
"1: ldxr %w[ret], %[ptr]\n"
|
|
" stxr %w[loop], %w[val], %[ptr]\n"
|
|
" cbnz %w[loop], 1b"
|
|
: [loop] "=&r"(loop), [ret] "=&r"(ret),
|
|
[ptr] "+Q"(*(u32 *)ptr)
|
|
: [val] "r" (val));
|
|
break;
|
|
case 8:
|
|
asm ("//__percpu_xchg_8\n"
|
|
"1: ldxr %[ret], %[ptr]\n"
|
|
" stxr %w[loop], %[val], %[ptr]\n"
|
|
" cbnz %w[loop], 1b"
|
|
: [loop] "=&r"(loop), [ret] "=&r"(ret),
|
|
[ptr] "+Q"(*(u64 *)ptr)
|
|
: [val] "r" (val));
|
|
break;
|
|
default:
|
|
BUILD_BUG();
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
#define _percpu_read(pcp) \
|
|
({ \
|
|
typeof(pcp) __retval; \
|
|
preempt_disable_notrace(); \
|
|
__retval = (typeof(pcp))__percpu_read(raw_cpu_ptr(&(pcp)), \
|
|
sizeof(pcp)); \
|
|
preempt_enable_notrace(); \
|
|
__retval; \
|
|
})
|
|
|
|
#define _percpu_write(pcp, val) \
|
|
do { \
|
|
preempt_disable_notrace(); \
|
|
__percpu_write(raw_cpu_ptr(&(pcp)), (unsigned long)(val), \
|
|
sizeof(pcp)); \
|
|
preempt_enable_notrace(); \
|
|
} while(0) \
|
|
|
|
#define _pcp_protect(operation, pcp, val) \
|
|
({ \
|
|
typeof(pcp) __retval; \
|
|
preempt_disable(); \
|
|
__retval = (typeof(pcp))operation(raw_cpu_ptr(&(pcp)), \
|
|
(val), sizeof(pcp)); \
|
|
preempt_enable(); \
|
|
__retval; \
|
|
})
|
|
|
|
#define _percpu_add(pcp, val) \
|
|
_pcp_protect(__percpu_add, pcp, val)
|
|
|
|
#define _percpu_add_return(pcp, val) _percpu_add(pcp, val)
|
|
|
|
#define _percpu_and(pcp, val) \
|
|
_pcp_protect(__percpu_and, pcp, val)
|
|
|
|
#define _percpu_or(pcp, val) \
|
|
_pcp_protect(__percpu_or, pcp, val)
|
|
|
|
#define _percpu_xchg(pcp, val) (typeof(pcp)) \
|
|
_pcp_protect(__percpu_xchg, pcp, (unsigned long)(val))
|
|
|
|
#define this_cpu_add_1(pcp, val) _percpu_add(pcp, val)
|
|
#define this_cpu_add_2(pcp, val) _percpu_add(pcp, val)
|
|
#define this_cpu_add_4(pcp, val) _percpu_add(pcp, val)
|
|
#define this_cpu_add_8(pcp, val) _percpu_add(pcp, val)
|
|
|
|
#define this_cpu_add_return_1(pcp, val) _percpu_add_return(pcp, val)
|
|
#define this_cpu_add_return_2(pcp, val) _percpu_add_return(pcp, val)
|
|
#define this_cpu_add_return_4(pcp, val) _percpu_add_return(pcp, val)
|
|
#define this_cpu_add_return_8(pcp, val) _percpu_add_return(pcp, val)
|
|
|
|
#define this_cpu_and_1(pcp, val) _percpu_and(pcp, val)
|
|
#define this_cpu_and_2(pcp, val) _percpu_and(pcp, val)
|
|
#define this_cpu_and_4(pcp, val) _percpu_and(pcp, val)
|
|
#define this_cpu_and_8(pcp, val) _percpu_and(pcp, val)
|
|
|
|
#define this_cpu_or_1(pcp, val) _percpu_or(pcp, val)
|
|
#define this_cpu_or_2(pcp, val) _percpu_or(pcp, val)
|
|
#define this_cpu_or_4(pcp, val) _percpu_or(pcp, val)
|
|
#define this_cpu_or_8(pcp, val) _percpu_or(pcp, val)
|
|
|
|
#define this_cpu_read_1(pcp) _percpu_read(pcp)
|
|
#define this_cpu_read_2(pcp) _percpu_read(pcp)
|
|
#define this_cpu_read_4(pcp) _percpu_read(pcp)
|
|
#define this_cpu_read_8(pcp) _percpu_read(pcp)
|
|
|
|
#define this_cpu_write_1(pcp, val) _percpu_write(pcp, val)
|
|
#define this_cpu_write_2(pcp, val) _percpu_write(pcp, val)
|
|
#define this_cpu_write_4(pcp, val) _percpu_write(pcp, val)
|
|
#define this_cpu_write_8(pcp, val) _percpu_write(pcp, val)
|
|
|
|
#define this_cpu_xchg_1(pcp, val) _percpu_xchg(pcp, val)
|
|
#define this_cpu_xchg_2(pcp, val) _percpu_xchg(pcp, val)
|
|
#define this_cpu_xchg_4(pcp, val) _percpu_xchg(pcp, val)
|
|
#define this_cpu_xchg_8(pcp, val) _percpu_xchg(pcp, val)
|
|
|
|
#include <asm-generic/percpu.h>
|
|
|
|
#endif /* __ASM_PERCPU_H */
|