Loading...
/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _ASM_SCS_H #define _ASM_SCS_H #ifdef __ASSEMBLY__ #include <asm/asm-offsets.h> #include <asm/sysreg.h> #ifdef CONFIG_SHADOW_CALL_STACK scs_sp .req x18 .macro scs_load_current get_current_task scs_sp ldr scs_sp, [scs_sp, #TSK_TI_SCS_SP] .endm .macro scs_save tsk str scs_sp, [\tsk, #TSK_TI_SCS_SP] .endm #else .macro scs_load_current .endm .macro scs_save tsk .endm #endif /* CONFIG_SHADOW_CALL_STACK */ #else #include <linux/scs.h> #include <asm/cpufeature.h> #ifdef CONFIG_UNWIND_PATCH_PAC_INTO_SCS static inline bool should_patch_pac_into_scs(void) { u64 reg; /* * We only enable the shadow call stack dynamically if we are running * on a system that does not implement PAC or BTI. PAC and SCS provide * roughly the same level of protection, and BTI relies on the PACIASP * instructions serving as landing pads, preventing us from patching * those instructions into something else. */ reg = read_sysreg_s(SYS_ID_AA64ISAR1_EL1); if (SYS_FIELD_GET(ID_AA64ISAR1_EL1, APA, reg) | SYS_FIELD_GET(ID_AA64ISAR1_EL1, API, reg)) return false; reg = read_sysreg_s(SYS_ID_AA64ISAR2_EL1); if (SYS_FIELD_GET(ID_AA64ISAR2_EL1, APA3, reg)) return false; if (IS_ENABLED(CONFIG_ARM64_BTI_KERNEL)) { reg = read_sysreg_s(SYS_ID_AA64PFR1_EL1); if (reg & (0xf << ID_AA64PFR1_EL1_BT_SHIFT)) return false; } return true; } static inline void dynamic_scs_init(void) { if (should_patch_pac_into_scs()) { pr_info("Enabling dynamic shadow call stack\n"); static_branch_enable(&dynamic_scs_enabled); } } #else static inline void dynamic_scs_init(void) {} #endif int scs_patch(const u8 eh_frame[], int size); asmlinkage void scs_patch_vmlinux(void); #endif /* __ASSEMBLY __ */ #endif /* _ASM_SCS_H */ |