David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame^] | 1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 2 | /* |
| 3 | * Copyright (C) 2013 Imagination Technologies |
| 4 | * Author: Paul Burton <paul.burton@mips.com> |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 5 | */ |
| 6 | #ifndef _ASM_MSA_H |
| 7 | #define _ASM_MSA_H |
| 8 | |
| 9 | #include <asm/mipsregs.h> |
| 10 | |
| 11 | #ifndef __ASSEMBLY__ |
| 12 | |
| 13 | #include <asm/inst.h> |
| 14 | |
| 15 | extern void _save_msa(struct task_struct *); |
| 16 | extern void _restore_msa(struct task_struct *); |
| 17 | extern void _init_msa_upper(void); |
| 18 | |
| 19 | extern void read_msa_wr_b(unsigned idx, union fpureg *to); |
| 20 | extern void read_msa_wr_h(unsigned idx, union fpureg *to); |
| 21 | extern void read_msa_wr_w(unsigned idx, union fpureg *to); |
| 22 | extern void read_msa_wr_d(unsigned idx, union fpureg *to); |
| 23 | |
| 24 | /** |
| 25 | * read_msa_wr() - Read a single MSA vector register |
| 26 | * @idx: The index of the vector register to read |
| 27 | * @to: The FPU register union to store the registers value in |
| 28 | * @fmt: The format of the data in the vector register |
| 29 | * |
| 30 | * Read the value of MSA vector register idx into the FPU register |
| 31 | * union to, using the format fmt. |
| 32 | */ |
| 33 | static inline void read_msa_wr(unsigned idx, union fpureg *to, |
| 34 | enum msa_2b_fmt fmt) |
| 35 | { |
| 36 | switch (fmt) { |
| 37 | case msa_fmt_b: |
| 38 | read_msa_wr_b(idx, to); |
| 39 | break; |
| 40 | |
| 41 | case msa_fmt_h: |
| 42 | read_msa_wr_h(idx, to); |
| 43 | break; |
| 44 | |
| 45 | case msa_fmt_w: |
| 46 | read_msa_wr_w(idx, to); |
| 47 | break; |
| 48 | |
| 49 | case msa_fmt_d: |
| 50 | read_msa_wr_d(idx, to); |
| 51 | break; |
| 52 | |
| 53 | default: |
| 54 | BUG(); |
| 55 | } |
| 56 | } |
| 57 | |
| 58 | extern void write_msa_wr_b(unsigned idx, union fpureg *from); |
| 59 | extern void write_msa_wr_h(unsigned idx, union fpureg *from); |
| 60 | extern void write_msa_wr_w(unsigned idx, union fpureg *from); |
| 61 | extern void write_msa_wr_d(unsigned idx, union fpureg *from); |
| 62 | |
| 63 | /** |
| 64 | * write_msa_wr() - Write a single MSA vector register |
| 65 | * @idx: The index of the vector register to write |
| 66 | * @from: The FPU register union to take the registers value from |
| 67 | * @fmt: The format of the data in the vector register |
| 68 | * |
| 69 | * Write the value from the FPU register union from into MSA vector |
| 70 | * register idx, using the format fmt. |
| 71 | */ |
| 72 | static inline void write_msa_wr(unsigned idx, union fpureg *from, |
| 73 | enum msa_2b_fmt fmt) |
| 74 | { |
| 75 | switch (fmt) { |
| 76 | case msa_fmt_b: |
| 77 | write_msa_wr_b(idx, from); |
| 78 | break; |
| 79 | |
| 80 | case msa_fmt_h: |
| 81 | write_msa_wr_h(idx, from); |
| 82 | break; |
| 83 | |
| 84 | case msa_fmt_w: |
| 85 | write_msa_wr_w(idx, from); |
| 86 | break; |
| 87 | |
| 88 | case msa_fmt_d: |
| 89 | write_msa_wr_d(idx, from); |
| 90 | break; |
| 91 | |
| 92 | default: |
| 93 | BUG(); |
| 94 | } |
| 95 | } |
| 96 | |
| 97 | static inline void enable_msa(void) |
| 98 | { |
| 99 | if (cpu_has_msa) { |
| 100 | set_c0_config5(MIPS_CONF5_MSAEN); |
| 101 | enable_fpu_hazard(); |
| 102 | } |
| 103 | } |
| 104 | |
| 105 | static inline void disable_msa(void) |
| 106 | { |
| 107 | if (cpu_has_msa) { |
| 108 | clear_c0_config5(MIPS_CONF5_MSAEN); |
| 109 | disable_fpu_hazard(); |
| 110 | } |
| 111 | } |
| 112 | |
| 113 | static inline int is_msa_enabled(void) |
| 114 | { |
| 115 | if (!cpu_has_msa) |
| 116 | return 0; |
| 117 | |
| 118 | return read_c0_config5() & MIPS_CONF5_MSAEN; |
| 119 | } |
| 120 | |
| 121 | static inline int thread_msa_context_live(void) |
| 122 | { |
| 123 | /* |
| 124 | * Check cpu_has_msa only if it's a constant. This will allow the |
| 125 | * compiler to optimise out code for CPUs without MSA without adding |
| 126 | * an extra redundant check for CPUs with MSA. |
| 127 | */ |
| 128 | if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa) |
| 129 | return 0; |
| 130 | |
| 131 | return test_thread_flag(TIF_MSA_CTX_LIVE); |
| 132 | } |
| 133 | |
| 134 | static inline void save_msa(struct task_struct *t) |
| 135 | { |
| 136 | if (cpu_has_msa) |
| 137 | _save_msa(t); |
| 138 | } |
| 139 | |
| 140 | static inline void restore_msa(struct task_struct *t) |
| 141 | { |
| 142 | if (cpu_has_msa) |
| 143 | _restore_msa(t); |
| 144 | } |
| 145 | |
| 146 | static inline void init_msa_upper(void) |
| 147 | { |
| 148 | /* |
| 149 | * Check cpu_has_msa only if it's a constant. This will allow the |
| 150 | * compiler to optimise out code for CPUs without MSA without adding |
| 151 | * an extra redundant check for CPUs with MSA. |
| 152 | */ |
| 153 | if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa) |
| 154 | return; |
| 155 | |
| 156 | _init_msa_upper(); |
| 157 | } |
| 158 | |
| 159 | #ifndef TOOLCHAIN_SUPPORTS_MSA |
| 160 | /* |
| 161 | * Define assembler macros using .word for the c[ft]cmsa instructions in order |
| 162 | * to allow compilation with toolchains that do not support MSA. Once all |
| 163 | * toolchains in use support MSA these can be removed. |
| 164 | */ |
| 165 | _ASM_MACRO_2R(cfcmsa, rd, cs, |
| 166 | _ASM_INSN_IF_MIPS(0x787e0019 | __cs << 11 | __rd << 6) |
| 167 | _ASM_INSN32_IF_MM(0x587e0016 | __cs << 11 | __rd << 6)); |
| 168 | _ASM_MACRO_2R(ctcmsa, cd, rs, |
| 169 | _ASM_INSN_IF_MIPS(0x783e0019 | __rs << 11 | __cd << 6) |
| 170 | _ASM_INSN32_IF_MM(0x583e0016 | __rs << 11 | __cd << 6)); |
| 171 | #define _ASM_SET_MSA "" |
| 172 | #else /* TOOLCHAIN_SUPPORTS_MSA */ |
| 173 | #define _ASM_SET_MSA ".set\tfp=64\n\t" \ |
| 174 | ".set\tmsa\n\t" |
| 175 | #endif |
| 176 | |
| 177 | #define __BUILD_MSA_CTL_REG(name, cs) \ |
| 178 | static inline unsigned int read_msa_##name(void) \ |
| 179 | { \ |
| 180 | unsigned int reg; \ |
| 181 | __asm__ __volatile__( \ |
| 182 | " .set push\n" \ |
| 183 | _ASM_SET_MSA \ |
| 184 | " cfcmsa %0, $" #cs "\n" \ |
| 185 | " .set pop\n" \ |
| 186 | : "=r"(reg)); \ |
| 187 | return reg; \ |
| 188 | } \ |
| 189 | \ |
| 190 | static inline void write_msa_##name(unsigned int val) \ |
| 191 | { \ |
| 192 | __asm__ __volatile__( \ |
| 193 | " .set push\n" \ |
| 194 | _ASM_SET_MSA \ |
| 195 | " ctcmsa $" #cs ", %0\n" \ |
| 196 | " .set pop\n" \ |
| 197 | : : "r"(val)); \ |
| 198 | } |
| 199 | |
| 200 | __BUILD_MSA_CTL_REG(ir, 0) |
| 201 | __BUILD_MSA_CTL_REG(csr, 1) |
| 202 | __BUILD_MSA_CTL_REG(access, 2) |
| 203 | __BUILD_MSA_CTL_REG(save, 3) |
| 204 | __BUILD_MSA_CTL_REG(modify, 4) |
| 205 | __BUILD_MSA_CTL_REG(request, 5) |
| 206 | __BUILD_MSA_CTL_REG(map, 6) |
| 207 | __BUILD_MSA_CTL_REG(unmap, 7) |
| 208 | |
| 209 | #endif /* !__ASSEMBLY__ */ |
| 210 | |
| 211 | #define MSA_IR 0 |
| 212 | #define MSA_CSR 1 |
| 213 | #define MSA_ACCESS 2 |
| 214 | #define MSA_SAVE 3 |
| 215 | #define MSA_MODIFY 4 |
| 216 | #define MSA_REQUEST 5 |
| 217 | #define MSA_MAP 6 |
| 218 | #define MSA_UNMAP 7 |
| 219 | |
| 220 | /* MSA Implementation Register (MSAIR) */ |
| 221 | #define MSA_IR_REVB 0 |
| 222 | #define MSA_IR_REVF (_ULCAST_(0xff) << MSA_IR_REVB) |
| 223 | #define MSA_IR_PROCB 8 |
| 224 | #define MSA_IR_PROCF (_ULCAST_(0xff) << MSA_IR_PROCB) |
| 225 | #define MSA_IR_WRPB 16 |
| 226 | #define MSA_IR_WRPF (_ULCAST_(0x1) << MSA_IR_WRPB) |
| 227 | |
| 228 | /* MSA Control & Status Register (MSACSR) */ |
| 229 | #define MSA_CSR_RMB 0 |
| 230 | #define MSA_CSR_RMF (_ULCAST_(0x3) << MSA_CSR_RMB) |
| 231 | #define MSA_CSR_RM_NEAREST 0 |
| 232 | #define MSA_CSR_RM_TO_ZERO 1 |
| 233 | #define MSA_CSR_RM_TO_POS 2 |
| 234 | #define MSA_CSR_RM_TO_NEG 3 |
| 235 | #define MSA_CSR_FLAGSB 2 |
| 236 | #define MSA_CSR_FLAGSF (_ULCAST_(0x1f) << MSA_CSR_FLAGSB) |
| 237 | #define MSA_CSR_FLAGS_IB 2 |
| 238 | #define MSA_CSR_FLAGS_IF (_ULCAST_(0x1) << MSA_CSR_FLAGS_IB) |
| 239 | #define MSA_CSR_FLAGS_UB 3 |
| 240 | #define MSA_CSR_FLAGS_UF (_ULCAST_(0x1) << MSA_CSR_FLAGS_UB) |
| 241 | #define MSA_CSR_FLAGS_OB 4 |
| 242 | #define MSA_CSR_FLAGS_OF (_ULCAST_(0x1) << MSA_CSR_FLAGS_OB) |
| 243 | #define MSA_CSR_FLAGS_ZB 5 |
| 244 | #define MSA_CSR_FLAGS_ZF (_ULCAST_(0x1) << MSA_CSR_FLAGS_ZB) |
| 245 | #define MSA_CSR_FLAGS_VB 6 |
| 246 | #define MSA_CSR_FLAGS_VF (_ULCAST_(0x1) << MSA_CSR_FLAGS_VB) |
| 247 | #define MSA_CSR_ENABLESB 7 |
| 248 | #define MSA_CSR_ENABLESF (_ULCAST_(0x1f) << MSA_CSR_ENABLESB) |
| 249 | #define MSA_CSR_ENABLES_IB 7 |
| 250 | #define MSA_CSR_ENABLES_IF (_ULCAST_(0x1) << MSA_CSR_ENABLES_IB) |
| 251 | #define MSA_CSR_ENABLES_UB 8 |
| 252 | #define MSA_CSR_ENABLES_UF (_ULCAST_(0x1) << MSA_CSR_ENABLES_UB) |
| 253 | #define MSA_CSR_ENABLES_OB 9 |
| 254 | #define MSA_CSR_ENABLES_OF (_ULCAST_(0x1) << MSA_CSR_ENABLES_OB) |
| 255 | #define MSA_CSR_ENABLES_ZB 10 |
| 256 | #define MSA_CSR_ENABLES_ZF (_ULCAST_(0x1) << MSA_CSR_ENABLES_ZB) |
| 257 | #define MSA_CSR_ENABLES_VB 11 |
| 258 | #define MSA_CSR_ENABLES_VF (_ULCAST_(0x1) << MSA_CSR_ENABLES_VB) |
| 259 | #define MSA_CSR_CAUSEB 12 |
| 260 | #define MSA_CSR_CAUSEF (_ULCAST_(0x3f) << MSA_CSR_CAUSEB) |
| 261 | #define MSA_CSR_CAUSE_IB 12 |
| 262 | #define MSA_CSR_CAUSE_IF (_ULCAST_(0x1) << MSA_CSR_CAUSE_IB) |
| 263 | #define MSA_CSR_CAUSE_UB 13 |
| 264 | #define MSA_CSR_CAUSE_UF (_ULCAST_(0x1) << MSA_CSR_CAUSE_UB) |
| 265 | #define MSA_CSR_CAUSE_OB 14 |
| 266 | #define MSA_CSR_CAUSE_OF (_ULCAST_(0x1) << MSA_CSR_CAUSE_OB) |
| 267 | #define MSA_CSR_CAUSE_ZB 15 |
| 268 | #define MSA_CSR_CAUSE_ZF (_ULCAST_(0x1) << MSA_CSR_CAUSE_ZB) |
| 269 | #define MSA_CSR_CAUSE_VB 16 |
| 270 | #define MSA_CSR_CAUSE_VF (_ULCAST_(0x1) << MSA_CSR_CAUSE_VB) |
| 271 | #define MSA_CSR_CAUSE_EB 17 |
| 272 | #define MSA_CSR_CAUSE_EF (_ULCAST_(0x1) << MSA_CSR_CAUSE_EB) |
| 273 | #define MSA_CSR_NXB 18 |
| 274 | #define MSA_CSR_NXF (_ULCAST_(0x1) << MSA_CSR_NXB) |
| 275 | #define MSA_CSR_FSB 24 |
| 276 | #define MSA_CSR_FSF (_ULCAST_(0x1) << MSA_CSR_FSB) |
| 277 | |
| 278 | #endif /* _ASM_MSA_H */ |