Loading Documentation/kernel-parameters.txt +6 −5 Original line number Diff line number Diff line Loading @@ -1919,11 +1919,12 @@ and is between 256 and 4096 characters. It is defined in the file Format: { 0 | 1 } See arch/parisc/kernel/pdc_chassis.c percpu_alloc= [X86] Select which percpu first chunk allocator to use. Allowed values are one of "lpage", "embed" and "4k". See comments in arch/x86/kernel/setup_percpu.c for details on each allocator. This parameter is primarily for debugging and performance comparison. percpu_alloc= Select which percpu first chunk allocator to use. Currently supported values are "embed" and "page". Archs may support subset or none of the selections. See comments in mm/percpu.c for details on each allocator. This parameter is primarily for debugging and performance comparison. pf. [PARIDE] See Documentation/blockdev/paride.txt. Loading Makefile +1 −1 Original line number Diff line number Diff line Loading @@ -325,7 +325,7 @@ CHECKFLAGS := -D__linux__ -Dlinux -D__STDC__ -Dunix -D__unix__ \ MODFLAGS = -DMODULE CFLAGS_MODULE = $(MODFLAGS) AFLAGS_MODULE = $(MODFLAGS) LDFLAGS_MODULE = LDFLAGS_MODULE = -T $(srctree)/scripts/module-common.lds CFLAGS_KERNEL = AFLAGS_KERNEL = CFLAGS_GCOV = -fprofile-arcs -ftest-coverage Loading arch/alpha/include/asm/percpu.h +8 −92 Original line number Diff line number Diff line #ifndef __ALPHA_PERCPU_H #define __ALPHA_PERCPU_H #include <linux/compiler.h> #include <linux/threads.h> #include <linux/percpu-defs.h> /* * Determine the real variable name from the name visible in the * kernel sources. */ #define per_cpu_var(var) per_cpu__##var #ifdef CONFIG_SMP /* * per_cpu_offset() is the offset that has to be added to a * percpu variable to get to the instance for a certain processor. */ extern unsigned long __per_cpu_offset[NR_CPUS]; #define per_cpu_offset(x) (__per_cpu_offset[x]) #define __my_cpu_offset per_cpu_offset(raw_smp_processor_id()) #ifdef CONFIG_DEBUG_PREEMPT #define my_cpu_offset per_cpu_offset(smp_processor_id()) #else #define my_cpu_offset __my_cpu_offset #endif #ifndef MODULE #define SHIFT_PERCPU_PTR(var, offset) RELOC_HIDE(&per_cpu_var(var), (offset)) #define PER_CPU_DEF_ATTRIBUTES #else /* * To calculate addresses of locally defined variables, GCC uses 32-bit * displacement from the GP. Which doesn't work for per cpu variables in * modules, as an offset to the kernel per cpu area is way above 4G. * To calculate addresses of locally defined variables, GCC uses * 32-bit displacement from the GP. Which doesn't work for per cpu * variables in modules, as an offset to the kernel per cpu area is * way above 4G. * * This forces allocation of a GOT entry for per cpu variable using * ldq instruction with a 'literal' relocation. */ #define SHIFT_PERCPU_PTR(var, offset) ({ \ extern int simple_identifier_##var(void); \ unsigned long __ptr, tmp_gp; \ asm ( "br %1, 1f \n\ 1: ldgp %1, 0(%1) \n\ ldq %0, per_cpu__" #var"(%1)\t!literal" \ : "=&r"(__ptr), "=&r"(tmp_gp)); \ (typeof(&per_cpu_var(var)))(__ptr + (offset)); }) #define PER_CPU_DEF_ATTRIBUTES __used #endif /* MODULE */ /* * A percpu variable may point to a discarded regions. The following are * established ways to produce a usable pointer from the percpu variable * offset. * Always use weak definitions for percpu variables in modules. */ #define per_cpu(var, cpu) \ (*SHIFT_PERCPU_PTR(var, per_cpu_offset(cpu))) #define __get_cpu_var(var) \ (*SHIFT_PERCPU_PTR(var, my_cpu_offset)) #define __raw_get_cpu_var(var) \ (*SHIFT_PERCPU_PTR(var, __my_cpu_offset)) #else /* ! SMP */ #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu_var(var))) #define __get_cpu_var(var) per_cpu_var(var) #define __raw_get_cpu_var(var) per_cpu_var(var) #define PER_CPU_DEF_ATTRIBUTES #endif /* SMP */ #ifdef CONFIG_SMP #define PER_CPU_BASE_SECTION ".data.percpu" #else #define PER_CPU_BASE_SECTION ".data" #endif #ifdef CONFIG_SMP #ifdef MODULE #define PER_CPU_SHARED_ALIGNED_SECTION "" #else #define PER_CPU_SHARED_ALIGNED_SECTION ".shared_aligned" #endif #define PER_CPU_FIRST_SECTION ".first" #else #define PER_CPU_SHARED_ALIGNED_SECTION "" #define PER_CPU_FIRST_SECTION "" #if defined(MODULE) && defined(CONFIG_SMP) #define ARCH_NEEDS_WEAK_PER_CPU #endif #define PER_CPU_ATTRIBUTES #include <asm-generic/percpu.h> #endif /* __ALPHA_PERCPU_H */ arch/alpha/include/asm/tlbflush.h +1 −0 Original line number Diff line number Diff line Loading @@ -2,6 +2,7 @@ #define _ALPHA_TLBFLUSH_H #include <linux/mm.h> #include <linux/sched.h> #include <asm/compiler.h> #include <asm/pgalloc.h> Loading arch/alpha/kernel/vmlinux.lds.S +2 −7 Original line number Diff line number Diff line Loading @@ -134,13 +134,6 @@ SECTIONS __bss_stop = .; _end = .; /* Sections to be discarded */ /DISCARD/ : { EXIT_TEXT EXIT_DATA *(.exitcall.exit) } .mdebug 0 : { *(.mdebug) } Loading @@ -150,4 +143,6 @@ SECTIONS STABS_DEBUG DWARF_DEBUG DISCARDS } Loading
Documentation/kernel-parameters.txt +6 −5 Original line number Diff line number Diff line Loading @@ -1919,11 +1919,12 @@ and is between 256 and 4096 characters. It is defined in the file Format: { 0 | 1 } See arch/parisc/kernel/pdc_chassis.c percpu_alloc= [X86] Select which percpu first chunk allocator to use. Allowed values are one of "lpage", "embed" and "4k". See comments in arch/x86/kernel/setup_percpu.c for details on each allocator. This parameter is primarily for debugging and performance comparison. percpu_alloc= Select which percpu first chunk allocator to use. Currently supported values are "embed" and "page". Archs may support subset or none of the selections. See comments in mm/percpu.c for details on each allocator. This parameter is primarily for debugging and performance comparison. pf. [PARIDE] See Documentation/blockdev/paride.txt. Loading
Makefile +1 −1 Original line number Diff line number Diff line Loading @@ -325,7 +325,7 @@ CHECKFLAGS := -D__linux__ -Dlinux -D__STDC__ -Dunix -D__unix__ \ MODFLAGS = -DMODULE CFLAGS_MODULE = $(MODFLAGS) AFLAGS_MODULE = $(MODFLAGS) LDFLAGS_MODULE = LDFLAGS_MODULE = -T $(srctree)/scripts/module-common.lds CFLAGS_KERNEL = AFLAGS_KERNEL = CFLAGS_GCOV = -fprofile-arcs -ftest-coverage Loading
arch/alpha/include/asm/percpu.h +8 −92 Original line number Diff line number Diff line #ifndef __ALPHA_PERCPU_H #define __ALPHA_PERCPU_H #include <linux/compiler.h> #include <linux/threads.h> #include <linux/percpu-defs.h> /* * Determine the real variable name from the name visible in the * kernel sources. */ #define per_cpu_var(var) per_cpu__##var #ifdef CONFIG_SMP /* * per_cpu_offset() is the offset that has to be added to a * percpu variable to get to the instance for a certain processor. */ extern unsigned long __per_cpu_offset[NR_CPUS]; #define per_cpu_offset(x) (__per_cpu_offset[x]) #define __my_cpu_offset per_cpu_offset(raw_smp_processor_id()) #ifdef CONFIG_DEBUG_PREEMPT #define my_cpu_offset per_cpu_offset(smp_processor_id()) #else #define my_cpu_offset __my_cpu_offset #endif #ifndef MODULE #define SHIFT_PERCPU_PTR(var, offset) RELOC_HIDE(&per_cpu_var(var), (offset)) #define PER_CPU_DEF_ATTRIBUTES #else /* * To calculate addresses of locally defined variables, GCC uses 32-bit * displacement from the GP. Which doesn't work for per cpu variables in * modules, as an offset to the kernel per cpu area is way above 4G. * To calculate addresses of locally defined variables, GCC uses * 32-bit displacement from the GP. Which doesn't work for per cpu * variables in modules, as an offset to the kernel per cpu area is * way above 4G. * * This forces allocation of a GOT entry for per cpu variable using * ldq instruction with a 'literal' relocation. */ #define SHIFT_PERCPU_PTR(var, offset) ({ \ extern int simple_identifier_##var(void); \ unsigned long __ptr, tmp_gp; \ asm ( "br %1, 1f \n\ 1: ldgp %1, 0(%1) \n\ ldq %0, per_cpu__" #var"(%1)\t!literal" \ : "=&r"(__ptr), "=&r"(tmp_gp)); \ (typeof(&per_cpu_var(var)))(__ptr + (offset)); }) #define PER_CPU_DEF_ATTRIBUTES __used #endif /* MODULE */ /* * A percpu variable may point to a discarded regions. The following are * established ways to produce a usable pointer from the percpu variable * offset. * Always use weak definitions for percpu variables in modules. */ #define per_cpu(var, cpu) \ (*SHIFT_PERCPU_PTR(var, per_cpu_offset(cpu))) #define __get_cpu_var(var) \ (*SHIFT_PERCPU_PTR(var, my_cpu_offset)) #define __raw_get_cpu_var(var) \ (*SHIFT_PERCPU_PTR(var, __my_cpu_offset)) #else /* ! SMP */ #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu_var(var))) #define __get_cpu_var(var) per_cpu_var(var) #define __raw_get_cpu_var(var) per_cpu_var(var) #define PER_CPU_DEF_ATTRIBUTES #endif /* SMP */ #ifdef CONFIG_SMP #define PER_CPU_BASE_SECTION ".data.percpu" #else #define PER_CPU_BASE_SECTION ".data" #endif #ifdef CONFIG_SMP #ifdef MODULE #define PER_CPU_SHARED_ALIGNED_SECTION "" #else #define PER_CPU_SHARED_ALIGNED_SECTION ".shared_aligned" #endif #define PER_CPU_FIRST_SECTION ".first" #else #define PER_CPU_SHARED_ALIGNED_SECTION "" #define PER_CPU_FIRST_SECTION "" #if defined(MODULE) && defined(CONFIG_SMP) #define ARCH_NEEDS_WEAK_PER_CPU #endif #define PER_CPU_ATTRIBUTES #include <asm-generic/percpu.h> #endif /* __ALPHA_PERCPU_H */
arch/alpha/include/asm/tlbflush.h +1 −0 Original line number Diff line number Diff line Loading @@ -2,6 +2,7 @@ #define _ALPHA_TLBFLUSH_H #include <linux/mm.h> #include <linux/sched.h> #include <asm/compiler.h> #include <asm/pgalloc.h> Loading
arch/alpha/kernel/vmlinux.lds.S +2 −7 Original line number Diff line number Diff line Loading @@ -134,13 +134,6 @@ SECTIONS __bss_stop = .; _end = .; /* Sections to be discarded */ /DISCARD/ : { EXIT_TEXT EXIT_DATA *(.exitcall.exit) } .mdebug 0 : { *(.mdebug) } Loading @@ -150,4 +143,6 @@ SECTIONS STABS_DEBUG DWARF_DEBUG DISCARDS }