Viewing file: kup.h (3.68 KB) -rw-r--r-- Select action/file-type: (+) | (+) | (+) | Code (+) | Session (+) | (+) | SDB (+) | (+) | (+) | (+) | (+) | (+) |
/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _ASM_POWERPC_BOOK3S_32_KUP_H #define _ASM_POWERPC_BOOK3S_32_KUP_H
#include <asm/bug.h> #include <asm/book3s/32/mmu-hash.h> #include <asm/mmu.h> #include <asm/synch.h>
#ifndef __ASSEMBLY__
#ifdef CONFIG_PPC_KUAP
#include <linux/sched.h>
#define KUAP_NONE (~0UL)
static __always_inline void kuap_lock_one(unsigned long addr) { mtsr(mfsr(addr) | SR_KS, addr); isync(); /* Context sync required after mtsr() */ }
static __always_inline void kuap_unlock_one(unsigned long addr) { mtsr(mfsr(addr) & ~SR_KS, addr); isync(); /* Context sync required after mtsr() */ }
static __always_inline void uaccess_begin_32s(unsigned long addr) { unsigned long tmp;
asm volatile(ASM_MMU_FTR_IFSET( "mfsrin %0, %1;" "rlwinm %0, %0, 0, %2;" "mtsrin %0, %1;" "isync", "", %3) : "=&r"(tmp) : "r"(addr), "i"(~SR_KS), "i"(MMU_FTR_KUAP) : "memory"); }
static __always_inline void uaccess_end_32s(unsigned long addr) { unsigned long tmp;
asm volatile(ASM_MMU_FTR_IFSET( "mfsrin %0, %1;" "oris %0, %0, %2;" "mtsrin %0, %1;" "isync", "", %3) : "=&r"(tmp) : "r"(addr), "i"(SR_KS >> 16), "i"(MMU_FTR_KUAP) : "memory"); }
static __always_inline void __kuap_save_and_lock(struct pt_regs *regs) { unsigned long kuap = current->thread.kuap;
regs->kuap = kuap; if (unlikely(kuap == KUAP_NONE)) return;
current->thread.kuap = KUAP_NONE; kuap_lock_one(kuap); } #define __kuap_save_and_lock __kuap_save_and_lock
static __always_inline void kuap_user_restore(struct pt_regs *regs) { }
static __always_inline void __kuap_kernel_restore(struct pt_regs *regs, unsigned long kuap) { if (unlikely(kuap != KUAP_NONE)) { current->thread.kuap = KUAP_NONE; kuap_lock_one(kuap); }
if (likely(regs->kuap == KUAP_NONE)) return;
current->thread.kuap = regs->kuap;
kuap_unlock_one(regs->kuap); }
static __always_inline unsigned long __kuap_get_and_assert_locked(void) { unsigned long kuap = current->thread.kuap;
WARN_ON_ONCE(IS_ENABLED(CONFIG_PPC_KUAP_DEBUG) && kuap != KUAP_NONE);
return kuap; } #define __kuap_get_and_assert_locked __kuap_get_and_assert_locked
static __always_inline void allow_user_access(void __user *to, const void __user *from, u32 size, unsigned long dir) { BUILD_BUG_ON(!__builtin_constant_p(dir));
if (!(dir & KUAP_WRITE)) return;
current->thread.kuap = (__force u32)to; uaccess_begin_32s((__force u32)to); }
static __always_inline void prevent_user_access(unsigned long dir) { u32 kuap = current->thread.kuap;
BUILD_BUG_ON(!__builtin_constant_p(dir));
if (!(dir & KUAP_WRITE)) return;
current->thread.kuap = KUAP_NONE; uaccess_end_32s(kuap); }
static __always_inline unsigned long prevent_user_access_return(void) { unsigned long flags = current->thread.kuap;
if (flags != KUAP_NONE) { current->thread.kuap = KUAP_NONE; uaccess_end_32s(flags); }
return flags; }
static __always_inline void restore_user_access(unsigned long flags) { if (flags != KUAP_NONE) { current->thread.kuap = flags; uaccess_begin_32s(flags); } }
static __always_inline bool __bad_kuap_fault(struct pt_regs *regs, unsigned long address, bool is_write) { unsigned long kuap = regs->kuap;
if (!is_write) return false; if (kuap == KUAP_NONE) return true;
/* * If faulting address doesn't match unlocked segment, change segment. * In case of unaligned store crossing two segments, emulate store. */ if ((kuap ^ address) & 0xf0000000) { if (!(kuap & 0x0fffffff) && address > kuap - 4 && fix_alignment(regs)) { regs_add_return_ip(regs, 4); emulate_single_step(regs); } else { regs->kuap = address; } }
return false; }
#endif /* CONFIG_PPC_KUAP */
#endif /* __ASSEMBLY__ */
#endif /* _ASM_POWERPC_BOOK3S_32_KUP_H */
|