abe1ee3a22
This patch changes the remaining direct references to .data.page_aligned in C and assembly code to use the macros in include/linux/linkage.h. Signed-off-by: Tim Abbott <tabbott@ksplice.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Ingo Molnar <mingo@redhat.com> Cc: H. Peter Anvin <hpa@zytor.com> Cc: Haavard Skinnemoen <hskinnemoen@atmel.com> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Paul Mackerras <paulus@samba.org> Cc: Martin Schwidefsky <schwidefsky@de.ibm.com> Signed-off-by: Sam Ravnborg <sam@ravnborg.org>
22 lines
556 B
C
22 lines
556 B
C
#ifndef _ASM_X86_CACHE_H
|
|
#define _ASM_X86_CACHE_H
|
|
|
|
#include <linux/linkage.h>
|
|
|
|
/* L1 cache line size */
|
|
#define L1_CACHE_SHIFT (CONFIG_X86_L1_CACHE_SHIFT)
|
|
#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT)
|
|
|
|
#define __read_mostly __attribute__((__section__(".data.read_mostly")))
|
|
|
|
#ifdef CONFIG_X86_VSMP
|
|
/* vSMP Internode cacheline shift */
|
|
#define INTERNODE_CACHE_SHIFT (12)
|
|
#ifdef CONFIG_SMP
|
|
#define __cacheline_aligned_in_smp \
|
|
__attribute__((__aligned__(1 << (INTERNODE_CACHE_SHIFT)))) \
|
|
__page_aligned_data
|
|
#endif
|
|
#endif
|
|
|
|
#endif /* _ASM_X86_CACHE_H */
|