Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 | /* SPDX-License-Identifier: GPL-2.0+ */ /* * Copyright (C) 2005 - 2013 Tensilica Inc. * Copyright (C) 2014 - 2016 Cadence Design Systems Inc. */ #ifndef _XTENSA_ASMMACRO_H #define _XTENSA_ASMMACRO_H #include <asm/arch/core.h> /* * Function entry and return macros for supported ABIs. */ #if defined(__XTENSA_WINDOWED_ABI__) #define abi_entry entry sp, 16 #define abi_ret retw #elif defined(__XTENSA_CALL0_ABI__) #define abi_entry #define abi_ret ret #else #error Unsupported Xtensa ABI #endif /* * Some little helpers for loops. Use zero-overhead-loops * where applicable and if supported by the processor. * * __loopi ar, at, size, inc * ar register initialized with the start address * at scratch register used by macro * size size immediate value * inc increment * * __loops ar, as, at, inc_log2[, mask_log2][, cond][, ncond] * ar register initialized with the start address * as register initialized with the size * at scratch register use by macro * inc_log2 increment [in log2] * mask_log2 mask [in log2] * cond true condition (used in loop'cond') * ncond false condition (used in b'ncond') * * __loop as * restart loop. 'as' register must not have been modified! * * __endla ar, as, incr * ar start address (modified) * as scratch register used by __loops/__loopi macros or * end address used by __loopt macro * inc increment */ #if XCHAL_HAVE_LOOPS .macro __loopi ar, at, size, incr movi \at, ((\size + \incr - 1) / (\incr)) loop \at, 99f .endm .macro __loops ar, as, at, incr_log2, mask_log2, cond, ncond .ifgt \incr_log2 - 1 addi \at, \as, (1 << \incr_log2) - 1 .ifnc \mask_log2, extui \at, \at, \incr_log2, \mask_log2 .else srli \at, \at, \incr_log2 .endif .endif loop\cond \at, 99f .endm .macro __loopt ar, as, at, incr_log2 sub \at, \as, \ar .ifgt \incr_log2 - 1 addi \at, \at, (1 << \incr_log2) - 1 srli \at, \at, \incr_log2 .endif loop \at, 99f .endm .macro __loop as loop \as, 99f .endm .macro __endl ar, as 99: .endm #else .macro __loopi ar, at, size, incr movi \at, ((\size + \incr - 1) / (\incr)) addi \at, \ar, \size 98: .endm .macro __loops ar, as, at, incr_log2, mask_log2, cond, ncond .ifnc \mask_log2, extui \at, \as, \incr_log2, \mask_log2 .else .ifnc \ncond, srli \at, \as, \incr_log2 .endif .endif .ifnc \ncond, b\ncond \at, 99f .endif .ifnc \mask_log2, slli \at, \at, \incr_log2 add \at, \ar, \at .else add \at, \ar, \as .endif 98: .endm .macro __loopt ar, as, at, incr_log2 98: .endm .macro __loop as 98: .endm .macro __endl ar, as bltu \ar, \as, 98b 99: .endm #endif .macro __endla ar, as, incr addi \ar, \ar, \incr __endl \ar \as .endm #endif /* _XTENSA_ASMMACRO_H */ |