blob: 1dd851397bd909a915447d75975d2ab6db52413b [file] [log] [blame]
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (C) 1991, 1992 Linus Torvalds
4 * Copyright (C) 2000, 2001, 2002 Andi Kleen, SuSE Labs
5 */
6#include <linux/sched/debug.h>
7#include <linux/kallsyms.h>
8#include <linux/kprobes.h>
9#include <linux/uaccess.h>
10#include <linux/hardirq.h>
11#include <linux/kdebug.h>
12#include <linux/export.h>
13#include <linux/ptrace.h>
14#include <linux/kexec.h>
15#include <linux/sysfs.h>
16#include <linux/bug.h>
17#include <linux/nmi.h>
18
David Brazdil0f672f62019-12-10 10:32:29 +000019#include <asm/cpu_entry_area.h>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000020#include <asm/stacktrace.h>
21
David Brazdil0f672f62019-12-10 10:32:29 +000022static const char * const exception_stack_names[] = {
23 [ ESTACK_DF ] = "#DF",
24 [ ESTACK_NMI ] = "NMI",
David Brazdil0f672f62019-12-10 10:32:29 +000025 [ ESTACK_DB ] = "#DB",
26 [ ESTACK_MCE ] = "#MC",
Olivier Deprez157378f2022-04-04 15:47:50 +020027 [ ESTACK_VC ] = "#VC",
28 [ ESTACK_VC2 ] = "#VC2",
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000029};
30
31const char *stack_type_name(enum stack_type type)
32{
David Brazdil0f672f62019-12-10 10:32:29 +000033 BUILD_BUG_ON(N_EXCEPTION_STACKS != 6);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000034
35 if (type == STACK_TYPE_IRQ)
36 return "IRQ";
37
38 if (type == STACK_TYPE_ENTRY) {
39 /*
40 * On 64-bit, we have a generic entry stack that we
41 * use for all the kernel entry points, including
42 * SYSENTER.
43 */
44 return "ENTRY_TRAMPOLINE";
45 }
46
47 if (type >= STACK_TYPE_EXCEPTION && type <= STACK_TYPE_EXCEPTION_LAST)
48 return exception_stack_names[type - STACK_TYPE_EXCEPTION];
49
50 return NULL;
51}
52
David Brazdil0f672f62019-12-10 10:32:29 +000053/**
54 * struct estack_pages - Page descriptor for exception stacks
55 * @offs: Offset from the start of the exception stack area
56 * @size: Size of the exception stack
57 * @type: Type to store in the stack_info struct
58 */
59struct estack_pages {
60 u32 offs;
61 u16 size;
62 u16 type;
63};
64
65#define EPAGERANGE(st) \
66 [PFN_DOWN(CEA_ESTACK_OFFS(st)) ... \
67 PFN_DOWN(CEA_ESTACK_OFFS(st) + CEA_ESTACK_SIZE(st) - 1)] = { \
68 .offs = CEA_ESTACK_OFFS(st), \
69 .size = CEA_ESTACK_SIZE(st), \
70 .type = STACK_TYPE_EXCEPTION + ESTACK_ ##st, }
71
72/*
73 * Array of exception stack page descriptors. If the stack is larger than
74 * PAGE_SIZE, all pages covering a particular stack will have the same
75 * info. The guard pages including the not mapped DB2 stack are zeroed
76 * out.
77 */
78static const
79struct estack_pages estack_pages[CEA_ESTACK_PAGES] ____cacheline_aligned = {
80 EPAGERANGE(DF),
81 EPAGERANGE(NMI),
David Brazdil0f672f62019-12-10 10:32:29 +000082 EPAGERANGE(DB),
83 EPAGERANGE(MCE),
Olivier Deprez157378f2022-04-04 15:47:50 +020084 EPAGERANGE(VC),
85 EPAGERANGE(VC2),
David Brazdil0f672f62019-12-10 10:32:29 +000086};
87
Olivier Deprez157378f2022-04-04 15:47:50 +020088static __always_inline bool in_exception_stack(unsigned long *stack, struct stack_info *info)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000089{
David Brazdil0f672f62019-12-10 10:32:29 +000090 unsigned long begin, end, stk = (unsigned long)stack;
91 const struct estack_pages *ep;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000092 struct pt_regs *regs;
David Brazdil0f672f62019-12-10 10:32:29 +000093 unsigned int k;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000094
David Brazdil0f672f62019-12-10 10:32:29 +000095 BUILD_BUG_ON(N_EXCEPTION_STACKS != 6);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000096
David Brazdil0f672f62019-12-10 10:32:29 +000097 begin = (unsigned long)__this_cpu_read(cea_exception_stacks);
98 /*
99 * Handle the case where stack trace is collected _before_
100 * cea_exception_stacks had been initialized.
101 */
102 if (!begin)
103 return false;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000104
David Brazdil0f672f62019-12-10 10:32:29 +0000105 end = begin + sizeof(struct cea_exception_stacks);
106 /* Bail if @stack is outside the exception stack area. */
107 if (stk < begin || stk >= end)
108 return false;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000109
David Brazdil0f672f62019-12-10 10:32:29 +0000110 /* Calc page offset from start of exception stacks */
111 k = (stk - begin) >> PAGE_SHIFT;
112 /* Lookup the page descriptor */
113 ep = &estack_pages[k];
114 /* Guard page? */
115 if (!ep->size)
116 return false;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000117
David Brazdil0f672f62019-12-10 10:32:29 +0000118 begin += (unsigned long)ep->offs;
119 end = begin + (unsigned long)ep->size;
120 regs = (struct pt_regs *)end - 1;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000121
David Brazdil0f672f62019-12-10 10:32:29 +0000122 info->type = ep->type;
123 info->begin = (unsigned long *)begin;
124 info->end = (unsigned long *)end;
125 info->next_sp = (unsigned long *)regs->sp;
126 return true;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000127}
128
Olivier Deprez157378f2022-04-04 15:47:50 +0200129static __always_inline bool in_irq_stack(unsigned long *stack, struct stack_info *info)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000130{
David Brazdil0f672f62019-12-10 10:32:29 +0000131 unsigned long *end = (unsigned long *)this_cpu_read(hardirq_stack_ptr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000132 unsigned long *begin = end - (IRQ_STACK_SIZE / sizeof(long));
133
134 /*
135 * This is a software stack, so 'end' can be a valid stack pointer.
136 * It just means the stack is empty.
137 */
David Brazdil0f672f62019-12-10 10:32:29 +0000138 if (stack < begin || stack >= end)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000139 return false;
140
141 info->type = STACK_TYPE_IRQ;
142 info->begin = begin;
143 info->end = end;
144
145 /*
146 * The next stack pointer is the first thing pushed by the entry code
147 * after switching to the irq stack.
148 */
149 info->next_sp = (unsigned long *)*(end - 1);
150
151 return true;
152}
153
Olivier Deprez157378f2022-04-04 15:47:50 +0200154bool noinstr get_stack_info_noinstr(unsigned long *stack, struct task_struct *task,
155 struct stack_info *info)
156{
157 if (in_task_stack(stack, task, info))
158 return true;
159
160 if (task != current)
161 return false;
162
163 if (in_exception_stack(stack, info))
164 return true;
165
166 if (in_irq_stack(stack, info))
167 return true;
168
169 if (in_entry_stack(stack, info))
170 return true;
171
172 return false;
173}
174
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000175int get_stack_info(unsigned long *stack, struct task_struct *task,
176 struct stack_info *info, unsigned long *visit_mask)
177{
Olivier Deprez157378f2022-04-04 15:47:50 +0200178 task = task ? : current;
179
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000180 if (!stack)
181 goto unknown;
182
Olivier Deprez157378f2022-04-04 15:47:50 +0200183 if (!get_stack_info_noinstr(stack, task, info))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000184 goto unknown;
185
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000186 /*
187 * Make sure we don't iterate through any given stack more than once.
188 * If it comes up a second time then there's something wrong going on:
189 * just break out and report an unknown stack type.
190 */
191 if (visit_mask) {
192 if (*visit_mask & (1UL << info->type)) {
Olivier Deprez157378f2022-04-04 15:47:50 +0200193 if (task == current)
194 printk_deferred_once(KERN_WARNING "WARNING: stack recursion on stack type %d\n", info->type);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000195 goto unknown;
196 }
197 *visit_mask |= 1UL << info->type;
198 }
199
200 return 0;
201
202unknown:
203 info->type = STACK_TYPE_UNKNOWN;
204 return -EINVAL;
205}