Loading arch/powerpc/Kconfig +1 −1 Original line number Diff line number Diff line Loading @@ -130,7 +130,7 @@ config PPC select ARCH_HAS_PTE_SPECIAL select ARCH_HAS_MEMBARRIER_CALLBACKS select ARCH_HAS_SCALED_CPUTIME if VIRT_CPU_ACCOUNTING_NATIVE && PPC_BOOK3S_64 select ARCH_HAS_STRICT_KERNEL_RWX if ((PPC_BOOK3S_64 || PPC32) && !HIBERNATION) select ARCH_HAS_STRICT_KERNEL_RWX if (PPC32 && !HIBERNATION) select ARCH_HAS_TICK_BROADCAST if GENERIC_CLOCKEVENTS_BROADCAST select ARCH_HAS_UACCESS_FLUSHCACHE select ARCH_HAS_UACCESS_MCSAFE if PPC64 Loading arch/powerpc/include/asm/book3s/32/hash.h +4 −4 Original line number Diff line number Diff line Loading @@ -17,9 +17,9 @@ * updating the accessed and modified bits in the page table tree. */ #define _PAGE_USER 0x001 /* usermode access allowed */ #define _PAGE_RW 0x002 /* software: user write access allowed */ #define _PAGE_PRESENT 0x004 /* software: pte contains a translation */ #define _PAGE_PRESENT 0x001 /* software: pte contains a translation */ #define _PAGE_HASHPTE 0x002 /* hash_page has made an HPTE for this pte */ #define _PAGE_USER 0x004 /* usermode access allowed */ #define _PAGE_GUARDED 0x008 /* G: prohibit speculative access */ #define _PAGE_COHERENT 0x010 /* M: enforce memory coherence (SMP systems) */ #define _PAGE_NO_CACHE 0x020 /* I: cache inhibit */ Loading @@ -27,7 +27,7 @@ #define _PAGE_DIRTY 0x080 /* C: page changed */ #define _PAGE_ACCESSED 0x100 /* R: page referenced */ #define _PAGE_EXEC 0x200 /* software: exec allowed */ #define _PAGE_HASHPTE 0x400 /* hash_page has made an HPTE for this pte */ #define _PAGE_RW 0x400 /* software: user write access allowed */ #define _PAGE_SPECIAL 0x800 /* software: Special page */ #ifdef CONFIG_PTE_64BIT Loading arch/powerpc/include/asm/book3s/32/kup.h +1 −1 Original line number Diff line number Diff line Loading @@ -75,7 +75,7 @@ .macro kuap_check current, gpr #ifdef CONFIG_PPC_KUAP_DEBUG lwz \gpr2, KUAP(thread) lwz \gpr, KUAP(thread) 999: twnei \gpr, 0 EMIT_BUG_ENTRY 999b, __FILE__, __LINE__, (BUGFLAG_WARNING | BUGFLAG_ONCE) #endif Loading arch/powerpc/include/asm/hw_irq.h +19 −1 Original line number Diff line number Diff line Loading @@ -250,9 +250,27 @@ static inline bool arch_irqs_disabled(void) } \ } while(0) static inline bool __lazy_irq_pending(u8 irq_happened) { return !!(irq_happened & ~PACA_IRQ_HARD_DIS); } /* * Check if a lazy IRQ is pending. Should be called with IRQs hard disabled. */ static inline bool lazy_irq_pending(void) { return !!(get_paca()->irq_happened & ~PACA_IRQ_HARD_DIS); return __lazy_irq_pending(get_paca()->irq_happened); } /* * Check if a lazy IRQ is pending, with no debugging checks. * Should be called with IRQs hard disabled. * For use in RI disabled code or other constrained situations. */ static inline bool lazy_irq_pending_nocheck(void) { return __lazy_irq_pending(local_paca->irq_happened); } /* Loading arch/powerpc/kernel/entry_32.S +1 −1 Original line number Diff line number Diff line Loading @@ -732,7 +732,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_SPE) stw r10,_CCR(r1) stw r1,KSP(r3) /* Set old stack pointer */ kuap_check r2, r4 kuap_check r2, r0 #ifdef CONFIG_SMP /* We need a sync somewhere here to make sure that if the * previous task gets rescheduled on another CPU, it sees all Loading Loading
arch/powerpc/Kconfig +1 −1 Original line number Diff line number Diff line Loading @@ -130,7 +130,7 @@ config PPC select ARCH_HAS_PTE_SPECIAL select ARCH_HAS_MEMBARRIER_CALLBACKS select ARCH_HAS_SCALED_CPUTIME if VIRT_CPU_ACCOUNTING_NATIVE && PPC_BOOK3S_64 select ARCH_HAS_STRICT_KERNEL_RWX if ((PPC_BOOK3S_64 || PPC32) && !HIBERNATION) select ARCH_HAS_STRICT_KERNEL_RWX if (PPC32 && !HIBERNATION) select ARCH_HAS_TICK_BROADCAST if GENERIC_CLOCKEVENTS_BROADCAST select ARCH_HAS_UACCESS_FLUSHCACHE select ARCH_HAS_UACCESS_MCSAFE if PPC64 Loading
arch/powerpc/include/asm/book3s/32/hash.h +4 −4 Original line number Diff line number Diff line Loading @@ -17,9 +17,9 @@ * updating the accessed and modified bits in the page table tree. */ #define _PAGE_USER 0x001 /* usermode access allowed */ #define _PAGE_RW 0x002 /* software: user write access allowed */ #define _PAGE_PRESENT 0x004 /* software: pte contains a translation */ #define _PAGE_PRESENT 0x001 /* software: pte contains a translation */ #define _PAGE_HASHPTE 0x002 /* hash_page has made an HPTE for this pte */ #define _PAGE_USER 0x004 /* usermode access allowed */ #define _PAGE_GUARDED 0x008 /* G: prohibit speculative access */ #define _PAGE_COHERENT 0x010 /* M: enforce memory coherence (SMP systems) */ #define _PAGE_NO_CACHE 0x020 /* I: cache inhibit */ Loading @@ -27,7 +27,7 @@ #define _PAGE_DIRTY 0x080 /* C: page changed */ #define _PAGE_ACCESSED 0x100 /* R: page referenced */ #define _PAGE_EXEC 0x200 /* software: exec allowed */ #define _PAGE_HASHPTE 0x400 /* hash_page has made an HPTE for this pte */ #define _PAGE_RW 0x400 /* software: user write access allowed */ #define _PAGE_SPECIAL 0x800 /* software: Special page */ #ifdef CONFIG_PTE_64BIT Loading
arch/powerpc/include/asm/book3s/32/kup.h +1 −1 Original line number Diff line number Diff line Loading @@ -75,7 +75,7 @@ .macro kuap_check current, gpr #ifdef CONFIG_PPC_KUAP_DEBUG lwz \gpr2, KUAP(thread) lwz \gpr, KUAP(thread) 999: twnei \gpr, 0 EMIT_BUG_ENTRY 999b, __FILE__, __LINE__, (BUGFLAG_WARNING | BUGFLAG_ONCE) #endif Loading
arch/powerpc/include/asm/hw_irq.h +19 −1 Original line number Diff line number Diff line Loading @@ -250,9 +250,27 @@ static inline bool arch_irqs_disabled(void) } \ } while(0) static inline bool __lazy_irq_pending(u8 irq_happened) { return !!(irq_happened & ~PACA_IRQ_HARD_DIS); } /* * Check if a lazy IRQ is pending. Should be called with IRQs hard disabled. */ static inline bool lazy_irq_pending(void) { return !!(get_paca()->irq_happened & ~PACA_IRQ_HARD_DIS); return __lazy_irq_pending(get_paca()->irq_happened); } /* * Check if a lazy IRQ is pending, with no debugging checks. * Should be called with IRQs hard disabled. * For use in RI disabled code or other constrained situations. */ static inline bool lazy_irq_pending_nocheck(void) { return __lazy_irq_pending(local_paca->irq_happened); } /* Loading
arch/powerpc/kernel/entry_32.S +1 −1 Original line number Diff line number Diff line Loading @@ -732,7 +732,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_SPE) stw r10,_CCR(r1) stw r1,KSP(r3) /* Set old stack pointer */ kuap_check r2, r4 kuap_check r2, r0 #ifdef CONFIG_SMP /* We need a sync somewhere here to make sure that if the * previous task gets rescheduled on another CPU, it sees all Loading