diff options
author | Will Deacon <will.deacon@arm.com> | 2015-05-29 20:28:44 +0300 |
---|---|---|
committer | Will Deacon <will.deacon@arm.com> | 2015-07-27 13:08:40 +0300 |
commit | 4b3dc9679cf779339d9049800803dfc3c83433d1 (patch) | |
tree | 8c04c26a614240f3142b7afafc071719fc3ea0f5 /arch/arm64/include/asm/percpu.h | |
parent | 52da443ec4d0a807b720527eb474f9c2878cd671 (diff) | |
download | linux-4b3dc9679cf779339d9049800803dfc3c83433d1.tar.xz |
arm64: force CONFIG_SMP=y and remove redundant #ifdefs
Nobody seems to be producing !SMP systems anymore, so this is just
becoming a source of kernel bugs, particularly if people want to use
coherent DMA with non-shared pages.
This patch forces CONFIG_SMP=y for arm64, removing a modest amount of
code in the process.
Signed-off-by: Will Deacon <will.deacon@arm.com>
Diffstat (limited to 'arch/arm64/include/asm/percpu.h')
-rw-r--r-- | arch/arm64/include/asm/percpu.h | 8 |
1 files changed, 0 insertions, 8 deletions
diff --git a/arch/arm64/include/asm/percpu.h b/arch/arm64/include/asm/percpu.h index 4fde8c1df97f..0a456bef8c79 100644 --- a/arch/arm64/include/asm/percpu.h +++ b/arch/arm64/include/asm/percpu.h @@ -16,8 +16,6 @@ #ifndef __ASM_PERCPU_H #define __ASM_PERCPU_H -#ifdef CONFIG_SMP - static inline void set_my_cpu_offset(unsigned long off) { asm volatile("msr tpidr_el1, %0" :: "r" (off) : "memory"); @@ -38,12 +36,6 @@ static inline unsigned long __my_cpu_offset(void) } #define __my_cpu_offset __my_cpu_offset() -#else /* !CONFIG_SMP */ - -#define set_my_cpu_offset(x) do { } while (0) - -#endif /* CONFIG_SMP */ - #define PERCPU_OP(op, asm_op) \ static inline unsigned long __percpu_##op(void *ptr, \ unsigned long val, int size) \ |