Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 | // SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2015 Linaro Ltd <ard.biesheuvel@linaro.org> */ #include <linux/efi.h> #include <linux/memblock.h> #include <linux/screen_info.h> #include <asm/efi.h> #include <asm/mach/map.h> #include <asm/mmu_context.h> static int __init set_permissions(pte_t *ptep, unsigned long addr, void *data) { efi_memory_desc_t *md = data; pte_t pte = *ptep; if (md->attribute & EFI_MEMORY_RO) pte = set_pte_bit(pte, __pgprot(L_PTE_RDONLY)); if (md->attribute & EFI_MEMORY_XP) pte = set_pte_bit(pte, __pgprot(L_PTE_XN)); set_pte_ext(ptep, pte, PTE_EXT_NG); return 0; } int __init efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool ignored) { unsigned long base, size; base = md->virt_addr; size = md->num_pages << EFI_PAGE_SHIFT; /* * We can only use apply_to_page_range() if we can guarantee that the * entire region was mapped using pages. This should be the case if the * region does not cover any naturally aligned SECTION_SIZE sized * blocks. */ if (round_down(base + size, SECTION_SIZE) < round_up(base, SECTION_SIZE) + SECTION_SIZE) return apply_to_page_range(mm, base, size, set_permissions, md); return 0; } int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) { struct map_desc desc = { .virtual = md->virt_addr, .pfn = __phys_to_pfn(md->phys_addr), .length = md->num_pages * EFI_PAGE_SIZE, }; /* * Order is important here: memory regions may have all of the * bits below set (and usually do), so we check them in order of * preference. */ if (md->attribute & EFI_MEMORY_WB) desc.type = MT_MEMORY_RWX; else if (md->attribute & EFI_MEMORY_WT) desc.type = MT_MEMORY_RWX_NONCACHED; else if (md->attribute & EFI_MEMORY_WC) desc.type = MT_DEVICE_WC; else desc.type = MT_DEVICE; create_mapping_late(mm, &desc, true); /* * If stricter permissions were specified, apply them now. */ if (md->attribute & (EFI_MEMORY_RO | EFI_MEMORY_XP)) return efi_set_mapping_permissions(mm, md, false); return 0; } static unsigned long __initdata cpu_state_table = EFI_INVALID_TABLE_ADDR; const efi_config_table_type_t efi_arch_tables[] __initconst = { {LINUX_EFI_ARM_CPU_STATE_TABLE_GUID, &cpu_state_table}, {} }; static void __init load_cpu_state_table(void) { if (cpu_state_table != EFI_INVALID_TABLE_ADDR) { struct efi_arm_entry_state *state; bool dump_state = true; state = early_memremap_ro(cpu_state_table, sizeof(struct efi_arm_entry_state)); if (state == NULL) { pr_warn("Unable to map CPU entry state table.\n"); return; } if ((state->sctlr_before_ebs & 1) == 0) pr_warn(FW_BUG "EFI stub was entered with MMU and Dcache disabled, please fix your firmware!\n"); else if ((state->sctlr_after_ebs & 1) == 0) pr_warn(FW_BUG "ExitBootServices() returned with MMU and Dcache disabled, please fix your firmware!\n"); else dump_state = false; if (dump_state || efi_enabled(EFI_DBG)) { pr_info("CPSR at EFI stub entry : 0x%08x\n", state->cpsr_before_ebs); pr_info("SCTLR at EFI stub entry : 0x%08x\n", state->sctlr_before_ebs); pr_info("CPSR after ExitBootServices() : 0x%08x\n", state->cpsr_after_ebs); pr_info("SCTLR after ExitBootServices(): 0x%08x\n", state->sctlr_after_ebs); } early_memunmap(state, sizeof(struct efi_arm_entry_state)); } } void __init arm_efi_init(void) { efi_init(); /* ARM does not permit early mappings to persist across paging_init() */ efi_memmap_unmap(); load_cpu_state_table(); } |