123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142 |
- /*
- * Extensible Firmware Interface
- *
- * Based on Extensible Firmware Interface Specification version 2.4
- *
- * Copyright (C) 2013, 2014 Linaro Ltd.
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License version 2 as
- * published by the Free Software Foundation.
- *
- */
- #include <linux/dmi.h>
- #include <linux/efi.h>
- #include <linux/init.h>
- #include <asm/efi.h>
- /*
- * Only regions of type EFI_RUNTIME_SERVICES_CODE need to be
- * executable, everything else can be mapped with the XN bits
- * set. Also take the new (optional) RO/XP bits into account.
- */
- static __init pteval_t create_mapping_protection(efi_memory_desc_t *md)
- {
- u64 attr = md->attribute;
- u32 type = md->type;
- if (type == EFI_MEMORY_MAPPED_IO)
- return PROT_DEVICE_nGnRE;
- if (WARN_ONCE(!PAGE_ALIGNED(md->phys_addr),
- "UEFI Runtime regions are not aligned to 64 KB -- buggy firmware?"))
- /*
- * If the region is not aligned to the page size of the OS, we
- * can not use strict permissions, since that would also affect
- * the mapping attributes of the adjacent regions.
- */
- return pgprot_val(PAGE_KERNEL_EXEC);
- /* R-- */
- if ((attr & (EFI_MEMORY_XP | EFI_MEMORY_RO)) ==
- (EFI_MEMORY_XP | EFI_MEMORY_RO))
- return pgprot_val(PAGE_KERNEL_RO);
- /* R-X */
- if (attr & EFI_MEMORY_RO)
- return pgprot_val(PAGE_KERNEL_ROX);
- /* RW- */
- if (attr & EFI_MEMORY_XP || type != EFI_RUNTIME_SERVICES_CODE)
- return pgprot_val(PAGE_KERNEL);
- /* RWX */
- return pgprot_val(PAGE_KERNEL_EXEC);
- }
- /* we will fill this structure from the stub, so don't put it in .bss */
- struct screen_info screen_info __section(.data);
- int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
- {
- pteval_t prot_val = create_mapping_protection(md);
- bool allow_block_mappings = (md->type != EFI_RUNTIME_SERVICES_CODE &&
- md->type != EFI_RUNTIME_SERVICES_DATA);
- if (!PAGE_ALIGNED(md->phys_addr) ||
- !PAGE_ALIGNED(md->num_pages << EFI_PAGE_SHIFT)) {
- /*
- * If the end address of this region is not aligned to page
- * size, the mapping is rounded up, and may end up sharing a
- * page frame with the next UEFI memory region. If we create
- * a block entry now, we may need to split it again when mapping
- * the next region, and support for that is going to be removed
- * from the MMU routines. So avoid block mappings altogether in
- * that case.
- */
- allow_block_mappings = false;
- }
- create_pgd_mapping(mm, md->phys_addr, md->virt_addr,
- md->num_pages << EFI_PAGE_SHIFT,
- __pgprot(prot_val | PTE_NG), allow_block_mappings);
- return 0;
- }
- static int __init set_permissions(pte_t *ptep, pgtable_t token,
- unsigned long addr, void *data)
- {
- efi_memory_desc_t *md = data;
- pte_t pte = *ptep;
- if (md->attribute & EFI_MEMORY_RO)
- pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
- if (md->attribute & EFI_MEMORY_XP)
- pte = set_pte_bit(pte, __pgprot(PTE_PXN));
- set_pte(ptep, pte);
- return 0;
- }
- int __init efi_set_mapping_permissions(struct mm_struct *mm,
- efi_memory_desc_t *md)
- {
- BUG_ON(md->type != EFI_RUNTIME_SERVICES_CODE &&
- md->type != EFI_RUNTIME_SERVICES_DATA);
- /*
- * Calling apply_to_page_range() is only safe on regions that are
- * guaranteed to be mapped down to pages. Since we are only called
- * for regions that have been mapped using efi_create_mapping() above
- * (and this is checked by the generic Memory Attributes table parsing
- * routines), there is no need to check that again here.
- */
- return apply_to_page_range(mm, md->virt_addr,
- md->num_pages << EFI_PAGE_SHIFT,
- set_permissions, md);
- }
- static int __init arm64_dmi_init(void)
- {
- /*
- * On arm64, DMI depends on UEFI, and dmi_scan_machine() needs to
- * be called early because dmi_id_init(), which is an arch_initcall
- * itself, depends on dmi_scan_machine() having been called already.
- */
- dmi_scan_machine();
- if (dmi_available)
- dmi_set_dump_stack_arch_desc();
- return 0;
- }
- core_initcall(arm64_dmi_init);
- /*
- * UpdateCapsule() depends on the system being shutdown via
- * ResetSystem().
- */
- bool efi_poweroff_required(void)
- {
- return efi_enabled(EFI_RUNTIME_SERVICES);
- }
|