| /* |
| * Copyright (c) 2016-2018, ARM Limited and Contributors. All rights reserved. |
| * |
| * SPDX-License-Identifier: BSD-3-Clause |
| */ |
| |
| #include <assert.h> |
| |
| #include <platform_def.h> |
| |
| #include <arch.h> |
| #include <arch_helpers.h> |
| #include <lib/utils.h> |
| #include <lib/xlat_tables/xlat_tables_arch.h> |
| #include <lib/xlat_tables/xlat_tables.h> |
| |
| #include "../xlat_tables_private.h" |
| |
| #if (ARM_ARCH_MAJOR == 7) && !defined(ARMV7_SUPPORTS_LARGE_PAGE_ADDRESSING) |
| #error ARMv7 target does not support LPAE MMU descriptors |
| #endif |
| |
| #define XLAT_TABLE_LEVEL_BASE \ |
| GET_XLAT_TABLE_LEVEL_BASE(PLAT_VIRT_ADDR_SPACE_SIZE) |
| |
| #define NUM_BASE_LEVEL_ENTRIES \ |
| GET_NUM_BASE_LEVEL_ENTRIES(PLAT_VIRT_ADDR_SPACE_SIZE) |
| |
| static uint64_t base_xlation_table[NUM_BASE_LEVEL_ENTRIES] |
| __aligned(NUM_BASE_LEVEL_ENTRIES * sizeof(uint64_t)); |
| |
| #if ENABLE_ASSERTIONS |
| static unsigned long long get_max_supported_pa(void) |
| { |
| /* Physical address space size for long descriptor format. */ |
| return (1ULL << 40) - 1ULL; |
| } |
| #endif /* ENABLE_ASSERTIONS */ |
| |
| unsigned int xlat_arch_current_el(void) |
| { |
| /* |
| * If EL3 is in AArch32 mode, all secure PL1 modes (Monitor, System, |
| * SVC, Abort, UND, IRQ and FIQ modes) execute at EL3. |
| */ |
| return 3U; |
| } |
| |
| uint64_t xlat_arch_get_xn_desc(unsigned int el __unused) |
| { |
| return UPPER_ATTRS(XN); |
| } |
| |
| void init_xlat_tables(void) |
| { |
| unsigned long long max_pa; |
| uintptr_t max_va; |
| |
| assert(PLAT_VIRT_ADDR_SPACE_SIZE >= MIN_VIRT_ADDR_SPACE_SIZE); |
| assert(PLAT_VIRT_ADDR_SPACE_SIZE <= MAX_VIRT_ADDR_SPACE_SIZE); |
| assert(IS_POWER_OF_TWO(PLAT_VIRT_ADDR_SPACE_SIZE)); |
| |
| print_mmap(); |
| init_xlation_table(0U, base_xlation_table, XLAT_TABLE_LEVEL_BASE, |
| &max_va, &max_pa); |
| |
| assert(max_va <= (PLAT_VIRT_ADDR_SPACE_SIZE - 1U)); |
| assert(max_pa <= (PLAT_PHY_ADDR_SPACE_SIZE - 1U)); |
| assert((PLAT_PHY_ADDR_SPACE_SIZE - 1U) <= get_max_supported_pa()); |
| } |
| |
| void enable_mmu_svc_mon(unsigned int flags) |
| { |
| unsigned int mair0, ttbcr, sctlr; |
| uint64_t ttbr0; |
| |
| assert(IS_IN_SECURE()); |
| assert((read_sctlr() & SCTLR_M_BIT) == 0U); |
| |
| /* Set attributes in the right indices of the MAIR */ |
| mair0 = MAIR0_ATTR_SET(ATTR_DEVICE, ATTR_DEVICE_INDEX); |
| mair0 |= MAIR0_ATTR_SET(ATTR_IWBWA_OWBWA_NTR, |
| ATTR_IWBWA_OWBWA_NTR_INDEX); |
| mair0 |= MAIR0_ATTR_SET(ATTR_NON_CACHEABLE, |
| ATTR_NON_CACHEABLE_INDEX); |
| write_mair0(mair0); |
| |
| /* Invalidate TLBs at the current exception level */ |
| tlbiall(); |
| |
| /* |
| * Set TTBCR bits as well. Set TTBR0 table properties. Disable TTBR1. |
| */ |
| int t0sz = 32 - __builtin_ctzll(PLAT_VIRT_ADDR_SPACE_SIZE); |
| |
| if ((flags & XLAT_TABLE_NC) != 0U) { |
| /* Inner & outer non-cacheable non-shareable. */ |
| ttbcr = TTBCR_EAE_BIT | |
| TTBCR_SH0_NON_SHAREABLE | TTBCR_RGN0_OUTER_NC | |
| TTBCR_RGN0_INNER_NC | (uint32_t) t0sz; |
| } else { |
| /* Inner & outer WBWA & shareable. */ |
| ttbcr = TTBCR_EAE_BIT | |
| TTBCR_SH0_INNER_SHAREABLE | TTBCR_RGN0_OUTER_WBA | |
| TTBCR_RGN0_INNER_WBA | (uint32_t) t0sz; |
| } |
| ttbcr |= TTBCR_EPD1_BIT; |
| write_ttbcr(ttbcr); |
| |
| /* Set TTBR0 bits as well */ |
| ttbr0 = (uintptr_t) base_xlation_table; |
| write64_ttbr0(ttbr0); |
| write64_ttbr1(0U); |
| |
| /* |
| * Ensure all translation table writes have drained |
| * into memory, the TLB invalidation is complete, |
| * and translation register writes are committed |
| * before enabling the MMU |
| */ |
| dsbish(); |
| isb(); |
| |
| sctlr = read_sctlr(); |
| sctlr |= SCTLR_WXN_BIT | SCTLR_M_BIT; |
| |
| if ((flags & DISABLE_DCACHE) != 0U) |
| sctlr &= ~SCTLR_C_BIT; |
| else |
| sctlr |= SCTLR_C_BIT; |
| |
| write_sctlr(sctlr); |
| |
| /* Ensure the MMU enable takes effect immediately */ |
| isb(); |
| } |
| |
| void enable_mmu_direct_svc_mon(unsigned int flags) |
| { |
| enable_mmu_svc_mon(flags); |
| } |