Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0 |
| 2 | |
| 3 | #include <linux/compiler.h> |
| 4 | #include <linux/context_tracking.h> |
| 5 | #include <linux/errno.h> |
| 6 | #include <linux/nospec.h> |
| 7 | #include <linux/ptrace.h> |
| 8 | #include <linux/syscalls.h> |
| 9 | |
| 10 | #include <asm/daifflags.h> |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 11 | #include <asm/debug-monitors.h> |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 12 | #include <asm/fpsimd.h> |
| 13 | #include <asm/syscall.h> |
| 14 | #include <asm/thread_info.h> |
| 15 | #include <asm/unistd.h> |
| 16 | |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 17 | long compat_arm_syscall(struct pt_regs *regs, int scno); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 18 | long sys_ni_syscall(void); |
| 19 | |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 20 | static long do_ni_syscall(struct pt_regs *regs, int scno) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 21 | { |
| 22 | #ifdef CONFIG_COMPAT |
| 23 | long ret; |
| 24 | if (is_compat_task()) { |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 25 | ret = compat_arm_syscall(regs, scno); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 26 | if (ret != -ENOSYS) |
| 27 | return ret; |
| 28 | } |
| 29 | #endif |
| 30 | |
| 31 | return sys_ni_syscall(); |
| 32 | } |
| 33 | |
| 34 | static long __invoke_syscall(struct pt_regs *regs, syscall_fn_t syscall_fn) |
| 35 | { |
| 36 | return syscall_fn(regs); |
| 37 | } |
| 38 | |
| 39 | static void invoke_syscall(struct pt_regs *regs, unsigned int scno, |
| 40 | unsigned int sc_nr, |
| 41 | const syscall_fn_t syscall_table[]) |
| 42 | { |
| 43 | long ret; |
| 44 | |
| 45 | if (scno < sc_nr) { |
| 46 | syscall_fn_t syscall_fn; |
| 47 | syscall_fn = syscall_table[array_index_nospec(scno, sc_nr)]; |
| 48 | ret = __invoke_syscall(regs, syscall_fn); |
| 49 | } else { |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 50 | ret = do_ni_syscall(regs, scno); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 51 | } |
| 52 | |
Olivier Deprez | 0e64123 | 2021-09-23 10:07:05 +0200 | [diff] [blame] | 53 | syscall_set_return_value(current, regs, 0, ret); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 54 | } |
| 55 | |
| 56 | static inline bool has_syscall_work(unsigned long flags) |
| 57 | { |
| 58 | return unlikely(flags & _TIF_SYSCALL_WORK); |
| 59 | } |
| 60 | |
| 61 | int syscall_trace_enter(struct pt_regs *regs); |
| 62 | void syscall_trace_exit(struct pt_regs *regs); |
| 63 | |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 64 | #ifdef CONFIG_ARM64_ERRATUM_1463225 |
| 65 | DECLARE_PER_CPU(int, __in_cortex_a76_erratum_1463225_wa); |
| 66 | |
| 67 | static void cortex_a76_erratum_1463225_svc_handler(void) |
| 68 | { |
| 69 | u32 reg, val; |
| 70 | |
| 71 | if (!unlikely(test_thread_flag(TIF_SINGLESTEP))) |
| 72 | return; |
| 73 | |
| 74 | if (!unlikely(this_cpu_has_cap(ARM64_WORKAROUND_1463225))) |
| 75 | return; |
| 76 | |
| 77 | __this_cpu_write(__in_cortex_a76_erratum_1463225_wa, 1); |
| 78 | reg = read_sysreg(mdscr_el1); |
| 79 | val = reg | DBG_MDSCR_SS | DBG_MDSCR_KDE; |
| 80 | write_sysreg(val, mdscr_el1); |
| 81 | asm volatile("msr daifclr, #8"); |
| 82 | isb(); |
| 83 | |
| 84 | /* We will have taken a single-step exception by this point */ |
| 85 | |
| 86 | write_sysreg(reg, mdscr_el1); |
| 87 | __this_cpu_write(__in_cortex_a76_erratum_1463225_wa, 0); |
| 88 | } |
| 89 | #else |
| 90 | static void cortex_a76_erratum_1463225_svc_handler(void) { } |
| 91 | #endif /* CONFIG_ARM64_ERRATUM_1463225 */ |
| 92 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 93 | static void el0_svc_common(struct pt_regs *regs, int scno, int sc_nr, |
| 94 | const syscall_fn_t syscall_table[]) |
| 95 | { |
| 96 | unsigned long flags = current_thread_info()->flags; |
| 97 | |
| 98 | regs->orig_x0 = regs->regs[0]; |
| 99 | regs->syscallno = scno; |
| 100 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 101 | /* |
| 102 | * BTI note: |
| 103 | * The architecture does not guarantee that SPSR.BTYPE is zero |
| 104 | * on taking an SVC, so we could return to userspace with a |
| 105 | * non-zero BTYPE after the syscall. |
| 106 | * |
| 107 | * This shouldn't matter except when userspace is explicitly |
| 108 | * doing something stupid, such as setting PROT_BTI on a page |
| 109 | * that lacks conforming BTI/PACIxSP instructions, falling |
| 110 | * through from one executable page to another with differing |
| 111 | * PROT_BTI, or messing with BTYPE via ptrace: in such cases, |
| 112 | * userspace should not be surprised if a SIGILL occurs on |
| 113 | * syscall return. |
| 114 | * |
| 115 | * So, don't touch regs->pstate & PSR_BTYPE_MASK here. |
| 116 | * (Similarly for HVC and SMC elsewhere.) |
| 117 | */ |
| 118 | |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 119 | cortex_a76_erratum_1463225_svc_handler(); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 120 | local_daif_restore(DAIF_PROCCTX); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 121 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 122 | if (system_supports_mte() && (flags & _TIF_MTE_ASYNC_FAULT)) { |
| 123 | /* |
| 124 | * Process the asynchronous tag check fault before the actual |
| 125 | * syscall. do_notify_resume() will send a signal to userspace |
| 126 | * before the syscall is restarted. |
| 127 | */ |
| 128 | syscall_set_return_value(current, regs, -ERESTARTNOINTR, 0); |
| 129 | return; |
| 130 | } |
| 131 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 132 | if (has_syscall_work(flags)) { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 133 | /* |
| 134 | * The de-facto standard way to skip a system call using ptrace |
| 135 | * is to set the system call to -1 (NO_SYSCALL) and set x0 to a |
| 136 | * suitable error code for consumption by userspace. However, |
| 137 | * this cannot be distinguished from a user-issued syscall(-1) |
| 138 | * and so we must set x0 to -ENOSYS here in case the tracer doesn't |
| 139 | * issue the skip and we fall into trace_exit with x0 preserved. |
| 140 | * |
| 141 | * This is slightly odd because it also means that if a tracer |
| 142 | * sets the system call number to -1 but does not initialise x0, |
| 143 | * then x0 will be preserved for all system calls apart from a |
| 144 | * user-issued syscall(-1). However, requesting a skip and not |
| 145 | * setting the return value is unlikely to do anything sensible |
| 146 | * anyway. |
| 147 | */ |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 148 | if (scno == NO_SYSCALL) |
Olivier Deprez | 0e64123 | 2021-09-23 10:07:05 +0200 | [diff] [blame] | 149 | syscall_set_return_value(current, regs, -ENOSYS, 0); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 150 | scno = syscall_trace_enter(regs); |
| 151 | if (scno == NO_SYSCALL) |
| 152 | goto trace_exit; |
| 153 | } |
| 154 | |
| 155 | invoke_syscall(regs, scno, sc_nr, syscall_table); |
| 156 | |
| 157 | /* |
| 158 | * The tracing status may have changed under our feet, so we have to |
| 159 | * check again. However, if we were tracing entry, then we always trace |
| 160 | * exit regardless, as the old entry assembly did. |
| 161 | */ |
| 162 | if (!has_syscall_work(flags) && !IS_ENABLED(CONFIG_DEBUG_RSEQ)) { |
| 163 | local_daif_mask(); |
| 164 | flags = current_thread_info()->flags; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 165 | if (!has_syscall_work(flags) && !(flags & _TIF_SINGLESTEP)) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 166 | return; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 167 | local_daif_restore(DAIF_PROCCTX); |
| 168 | } |
| 169 | |
| 170 | trace_exit: |
| 171 | syscall_trace_exit(regs); |
| 172 | } |
| 173 | |
| 174 | static inline void sve_user_discard(void) |
| 175 | { |
| 176 | if (!system_supports_sve()) |
| 177 | return; |
| 178 | |
| 179 | clear_thread_flag(TIF_SVE); |
| 180 | |
| 181 | /* |
| 182 | * task_fpsimd_load() won't be called to update CPACR_EL1 in |
| 183 | * ret_to_user unless TIF_FOREIGN_FPSTATE is still set, which only |
| 184 | * happens if a context switch or kernel_neon_begin() or context |
| 185 | * modification (sigreturn, ptrace) intervenes. |
| 186 | * So, ensure that CPACR_EL1 is already correct for the fast-path case. |
| 187 | */ |
| 188 | sve_user_disable(); |
| 189 | } |
| 190 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 191 | void do_el0_svc(struct pt_regs *regs) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 192 | { |
| 193 | sve_user_discard(); |
| 194 | el0_svc_common(regs, regs->regs[8], __NR_syscalls, sys_call_table); |
| 195 | } |
| 196 | |
| 197 | #ifdef CONFIG_COMPAT |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 198 | void do_el0_svc_compat(struct pt_regs *regs) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 199 | { |
| 200 | el0_svc_common(regs, regs->regs[7], __NR_compat_syscalls, |
| 201 | compat_sys_call_table); |
| 202 | } |
| 203 | #endif |