[RFC PATCH v5 1/2] arm64: Introduce stack trace reliability checks in the unwinder

From: madvenka
Date: Wed May 26 2021 - 17:49:31 EST


From: "Madhavan T. Venkataraman" <madvenka@xxxxxxxxxxxxxxxxxxx>

The unwinder should check for the presence of various features and
conditions that can render the stack trace unreliable and mark the
the stack trace as unreliable for the benefit of the caller.

Introduce the first reliability check - If a return PC is not a valid
kernel text address, consider the stack trace unreliable. It could be
some generated code.

Other reliability checks will be added in the future.

Signed-off-by: Madhavan T. Venkataraman <madvenka@xxxxxxxxxxxxxxxxxxx>
---
arch/arm64/include/asm/stacktrace.h | 9 +++++++
arch/arm64/kernel/stacktrace.c | 38 +++++++++++++++++++++++++----
2 files changed, 42 insertions(+), 5 deletions(-)

diff --git a/arch/arm64/include/asm/stacktrace.h b/arch/arm64/include/asm/stacktrace.h
index eb29b1fe8255..4c822ef7f588 100644
--- a/arch/arm64/include/asm/stacktrace.h
+++ b/arch/arm64/include/asm/stacktrace.h
@@ -49,6 +49,13 @@ struct stack_info {
*
* @graph: When FUNCTION_GRAPH_TRACER is selected, holds the index of a
* replacement lr value in the ftrace graph stack.
+ *
+ * @reliable: Is this stack frame reliable? There are several checks that
+ * need to be performed in unwind_frame() before a stack frame
+ * is truly reliable. Until all the checks are present, this flag
+ * is just a place holder. Once all the checks are implemented,
+ * this comment will be updated and the flag can be used by the
+ * caller of unwind_frame().
*/
struct stackframe {
unsigned long fp;
@@ -59,6 +66,7 @@ struct stackframe {
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
int graph;
#endif
+ bool reliable;
};

extern int unwind_frame(struct task_struct *tsk, struct stackframe *frame);
@@ -169,6 +177,7 @@ static inline void start_backtrace(struct stackframe *frame,
bitmap_zero(frame->stacks_done, __NR_STACK_TYPES);
frame->prev_fp = 0;
frame->prev_type = STACK_TYPE_UNKNOWN;
+ frame->reliable = true;
}

#endif /* __ASM_STACKTRACE_H */
diff --git a/arch/arm64/kernel/stacktrace.c b/arch/arm64/kernel/stacktrace.c
index d55bdfb7789c..9061375c8785 100644
--- a/arch/arm64/kernel/stacktrace.c
+++ b/arch/arm64/kernel/stacktrace.c
@@ -44,21 +44,29 @@ int notrace unwind_frame(struct task_struct *tsk, struct stackframe *frame)
unsigned long fp = frame->fp;
struct stack_info info;

+ frame->reliable = true;
+
/* Terminal record; nothing to unwind */
if (!fp)
return -ENOENT;

- if (fp & 0xf)
+ if (fp & 0xf) {
+ frame->reliable = false;
return -EINVAL;
+ }

if (!tsk)
tsk = current;

- if (!on_accessible_stack(tsk, fp, &info))
+ if (!on_accessible_stack(tsk, fp, &info)) {
+ frame->reliable = false;
return -EINVAL;
+ }

- if (test_bit(info.type, frame->stacks_done))
+ if (test_bit(info.type, frame->stacks_done)) {
+ frame->reliable = false;
return -EINVAL;
+ }

/*
* As stacks grow downward, any valid record on the same stack must be
@@ -74,8 +82,10 @@ int notrace unwind_frame(struct task_struct *tsk, struct stackframe *frame)
* stack.
*/
if (info.type == frame->prev_type) {
- if (fp <= frame->prev_fp)
+ if (fp <= frame->prev_fp) {
+ frame->reliable = false;
return -EINVAL;
+ }
} else {
set_bit(frame->prev_type, frame->stacks_done);
}
@@ -100,14 +110,32 @@ int notrace unwind_frame(struct task_struct *tsk, struct stackframe *frame)
* So replace it to an original value.
*/
ret_stack = ftrace_graph_get_ret_stack(tsk, frame->graph++);
- if (WARN_ON_ONCE(!ret_stack))
+ if (WARN_ON_ONCE(!ret_stack)) {
+ frame->reliable = false;
return -EINVAL;
+ }
frame->pc = ret_stack->ret;
}
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */

frame->pc = ptrauth_strip_insn_pac(frame->pc);

+ /*
+ * Check the return PC for conditions that make unwinding unreliable.
+ * In each case, mark the stack trace as such.
+ */
+
+ /*
+ * Make sure that the return address is a proper kernel text address.
+ * A NULL or invalid return address could mean:
+ *
+ * - generated code such as eBPF and optprobe trampolines
+ * - Foreign code (e.g. EFI runtime services)
+ * - Procedure Linkage Table (PLT) entries and veneer functions
+ */
+ if (!__kernel_text_address(frame->pc))
+ frame->reliable = false;
+
return 0;
}
NOKPROBE_SYMBOL(unwind_frame);
--
2.25.1