Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 | // SPDX-License-Identifier: GPL-2.0+ /* * (C) Copyright 2002 * Wolfgang Denk, DENX Software Engineering, wd@denx.de. */ /* for now: just dummy functions to satisfy the linker */ #include <common.h> #include <cpu_func.h> #include <log.h> #include <malloc.h> #include <asm/cache.h> DECLARE_GLOBAL_DATA_PTR; /* * Flush range from all levels of d-cache/unified-cache. * Affects the range [start, start + size - 1]. */ __weak void flush_cache(unsigned long start, unsigned long size) { flush_dcache_range(start, start + size); } /* * Default implementation: * do a range flush for the entire range */ __weak void flush_dcache_all(void) { flush_cache(0, ~0); } /* * Default implementation of enable_caches() * Real implementation should be in platform code */ __weak void enable_caches(void) { puts("WARNING: Caches not enabled\n"); } __weak void invalidate_dcache_range(unsigned long start, unsigned long stop) { /* An empty stub, real implementation should be in platform code */ } __weak void flush_dcache_range(unsigned long start, unsigned long stop) { /* An empty stub, real implementation should be in platform code */ } int check_cache_range(unsigned long start, unsigned long stop) { int ok = 1; if (start & (CONFIG_SYS_CACHELINE_SIZE - 1)) ok = 0; if (stop & (CONFIG_SYS_CACHELINE_SIZE - 1)) ok = 0; if (!ok) { warn_non_spl("CACHE: Misaligned operation at range [%08lx, %08lx]\n", start, stop); } return ok; } #ifdef CONFIG_SYS_NONCACHED_MEMORY /* * Reserve one MMU section worth of address space below the malloc() area that * will be mapped uncached. */ static unsigned long noncached_start; static unsigned long noncached_end; static unsigned long noncached_next; void noncached_set_region(void) { #if !CONFIG_IS_ENABLED(SYS_DCACHE_OFF) mmu_set_region_dcache_behaviour(noncached_start, noncached_end - noncached_start, DCACHE_OFF); #endif } void noncached_init(void) { phys_addr_t start, end; size_t size; /* If this calculation changes, update board_f.c:reserve_noncached() */ end = ALIGN(mem_malloc_start, MMU_SECTION_SIZE) - MMU_SECTION_SIZE; size = ALIGN(CONFIG_SYS_NONCACHED_MEMORY, MMU_SECTION_SIZE); start = end - size; debug("mapping memory %pa-%pa non-cached\n", &start, &end); noncached_start = start; noncached_end = end; noncached_next = start; noncached_set_region(); } phys_addr_t noncached_alloc(size_t size, size_t align) { phys_addr_t next = ALIGN(noncached_next, align); if (next >= noncached_end || (noncached_end - next) < size) return 0; debug("allocated %zu bytes of uncached memory @%pa\n", size, &next); noncached_next = next + size; return next; } #endif /* CONFIG_SYS_NONCACHED_MEMORY */ #if CONFIG_IS_ENABLED(SYS_THUMB_BUILD) void invalidate_l2_cache(void) { unsigned int val = 0; asm volatile("mcr p15, 1, %0, c15, c11, 0 @ invl l2 cache" : : "r" (val) : "cc"); isb(); } #endif int arch_reserve_mmu(void) { return arm_reserve_mmu(); } __weak int arm_reserve_mmu(void) { #if !(CONFIG_IS_ENABLED(SYS_ICACHE_OFF) && CONFIG_IS_ENABLED(SYS_DCACHE_OFF)) /* reserve TLB table */ gd->arch.tlb_size = PGTABLE_SIZE; gd->relocaddr -= gd->arch.tlb_size; /* round down to next 64 kB limit */ gd->relocaddr &= ~(0x10000 - 1); gd->arch.tlb_addr = gd->relocaddr; debug("TLB table from %08lx to %08lx\n", gd->arch.tlb_addr, gd->arch.tlb_addr + gd->arch.tlb_size); #ifdef CONFIG_SYS_MEM_RESERVE_SECURE /* * Record allocated tlb_addr in case gd->tlb_addr to be overwritten * with location within secure ram. */ gd->arch.tlb_allocated = gd->arch.tlb_addr; #endif #endif return 0; } |