1 | // SPDX-License-Identifier: GPL-2.0 |
2 | |
3 | #include <linux/compiler.h> |
4 | #include <linux/context_tracking.h> |
5 | #include <linux/errno.h> |
6 | #include <linux/nospec.h> |
7 | #include <linux/ptrace.h> |
8 | #include <linux/randomize_kstack.h> |
9 | #include <linux/syscalls.h> |
10 | |
11 | #include <asm/debug-monitors.h> |
12 | #include <asm/exception.h> |
13 | #include <asm/fpsimd.h> |
14 | #include <asm/syscall.h> |
15 | #include <asm/thread_info.h> |
16 | #include <asm/unistd.h> |
17 | #include <asm/unistd_compat_32.h> |
18 | |
19 | long compat_arm_syscall(struct pt_regs *regs, int scno); |
20 | long sys_ni_syscall(void); |
21 | |
22 | static long do_ni_syscall(struct pt_regs *regs, int scno) |
23 | { |
24 | if (is_compat_task()) { |
25 | long ret = compat_arm_syscall(regs, scno); |
26 | if (ret != -ENOSYS) |
27 | return ret; |
28 | } |
29 | |
30 | return sys_ni_syscall(); |
31 | } |
32 | |
33 | static long __invoke_syscall(struct pt_regs *regs, syscall_fn_t syscall_fn) |
34 | { |
35 | return syscall_fn(regs); |
36 | } |
37 | |
38 | static void invoke_syscall(struct pt_regs *regs, unsigned int scno, |
39 | unsigned int sc_nr, |
40 | const syscall_fn_t syscall_table[]) |
41 | { |
42 | long ret; |
43 | |
44 | add_random_kstack_offset(); |
45 | |
46 | if (scno < sc_nr) { |
47 | syscall_fn_t syscall_fn; |
48 | syscall_fn = syscall_table[array_index_nospec(scno, sc_nr)]; |
49 | ret = __invoke_syscall(regs, syscall_fn: syscall_fn); |
50 | } else { |
51 | ret = do_ni_syscall(regs, scno); |
52 | } |
53 | |
54 | syscall_set_return_value(current, regs, error: 0, val: ret); |
55 | |
56 | /* |
57 | * This value will get limited by KSTACK_OFFSET_MAX(), which is 10 |
58 | * bits. The actual entropy will be further reduced by the compiler |
59 | * when applying stack alignment constraints: the AAPCS mandates a |
60 | * 16-byte aligned SP at function boundaries, which will remove the |
61 | * 4 low bits from any entropy chosen here. |
62 | * |
63 | * The resulting 6 bits of entropy is seen in SP[9:4]. |
64 | */ |
65 | choose_random_kstack_offset(get_random_u16()); |
66 | } |
67 | |
68 | static inline bool has_syscall_work(unsigned long flags) |
69 | { |
70 | return unlikely(flags & _TIF_SYSCALL_WORK); |
71 | } |
72 | |
73 | static void el0_svc_common(struct pt_regs *regs, int scno, int sc_nr, |
74 | const syscall_fn_t syscall_table[]) |
75 | { |
76 | unsigned long flags = read_thread_flags(); |
77 | |
78 | regs->orig_x0 = regs->regs[0]; |
79 | regs->syscallno = scno; |
80 | |
81 | /* |
82 | * BTI note: |
83 | * The architecture does not guarantee that SPSR.BTYPE is zero |
84 | * on taking an SVC, so we could return to userspace with a |
85 | * non-zero BTYPE after the syscall. |
86 | * |
87 | * This shouldn't matter except when userspace is explicitly |
88 | * doing something stupid, such as setting PROT_BTI on a page |
89 | * that lacks conforming BTI/PACIxSP instructions, falling |
90 | * through from one executable page to another with differing |
91 | * PROT_BTI, or messing with BTYPE via ptrace: in such cases, |
92 | * userspace should not be surprised if a SIGILL occurs on |
93 | * syscall return. |
94 | * |
95 | * So, don't touch regs->pstate & PSR_BTYPE_MASK here. |
96 | * (Similarly for HVC and SMC elsewhere.) |
97 | */ |
98 | |
99 | if (flags & _TIF_MTE_ASYNC_FAULT) { |
100 | /* |
101 | * Process the asynchronous tag check fault before the actual |
102 | * syscall. do_notify_resume() will send a signal to userspace |
103 | * before the syscall is restarted. |
104 | */ |
105 | syscall_set_return_value(current, regs, error: -ERESTARTNOINTR, val: 0); |
106 | return; |
107 | } |
108 | |
109 | if (has_syscall_work(flags)) { |
110 | /* |
111 | * The de-facto standard way to skip a system call using ptrace |
112 | * is to set the system call to -1 (NO_SYSCALL) and set x0 to a |
113 | * suitable error code for consumption by userspace. However, |
114 | * this cannot be distinguished from a user-issued syscall(-1) |
115 | * and so we must set x0 to -ENOSYS here in case the tracer doesn't |
116 | * issue the skip and we fall into trace_exit with x0 preserved. |
117 | * |
118 | * This is slightly odd because it also means that if a tracer |
119 | * sets the system call number to -1 but does not initialise x0, |
120 | * then x0 will be preserved for all system calls apart from a |
121 | * user-issued syscall(-1). However, requesting a skip and not |
122 | * setting the return value is unlikely to do anything sensible |
123 | * anyway. |
124 | */ |
125 | if (scno == NO_SYSCALL) |
126 | syscall_set_return_value(current, regs, error: -ENOSYS, val: 0); |
127 | scno = syscall_trace_enter(regs); |
128 | if (scno == NO_SYSCALL) |
129 | goto trace_exit; |
130 | } |
131 | |
132 | invoke_syscall(regs, scno, sc_nr, syscall_table); |
133 | |
134 | /* |
135 | * The tracing status may have changed under our feet, so we have to |
136 | * check again. However, if we were tracing entry, then we always trace |
137 | * exit regardless, as the old entry assembly did. |
138 | */ |
139 | if (!has_syscall_work(flags) && !IS_ENABLED(CONFIG_DEBUG_RSEQ)) { |
140 | flags = read_thread_flags(); |
141 | if (!has_syscall_work(flags) && !(flags & _TIF_SINGLESTEP)) |
142 | return; |
143 | } |
144 | |
145 | trace_exit: |
146 | syscall_trace_exit(regs); |
147 | } |
148 | |
149 | void do_el0_svc(struct pt_regs *regs) |
150 | { |
151 | el0_svc_common(regs, scno: regs->regs[8], __NR_syscalls, syscall_table: sys_call_table); |
152 | } |
153 | |
154 | #ifdef CONFIG_COMPAT |
155 | void do_el0_svc_compat(struct pt_regs *regs) |
156 | { |
157 | el0_svc_common(regs, scno: regs->regs[7], sc_nr: __NR_compat32_syscalls, |
158 | syscall_table: compat_sys_call_table); |
159 | } |
160 | #endif |
161 | |