| /* |
| * Copyright (C) 2009 Matt Fleming |
| * |
| * Based, in part, on kernel/time/clocksource.c. |
| * |
| * This file provides arbitration code for stack unwinders. |
| * |
| * Multiple stack unwinders can be available on a system, usually with |
| * the most accurate unwinder being the currently active one. |
| */ |
| #include <linux/errno.h> |
| #include <linux/list.h> |
| #include <linux/spinlock.h> |
| #include <asm/unwinder.h> |
| #include <asm/atomic.h> |
| |
| /* |
| * This is the most basic stack unwinder an architecture can |
| * provide. For architectures without reliable frame pointers, e.g. |
| * RISC CPUs, it can be implemented by looking through the stack for |
| * addresses that lie within the kernel text section. |
| * |
| * Other CPUs, e.g. x86, can use their frame pointer register to |
| * construct more accurate stack traces. |
| */ |
| static struct list_head unwinder_list; |
| static struct unwinder stack_reader = { |
| .name = "stack-reader", |
| .dump = stack_reader_dump, |
| .rating = 50, |
| .list = { |
| .next = &unwinder_list, |
| .prev = &unwinder_list, |
| }, |
| }; |
| |
| /* |
| * "curr_unwinder" points to the stack unwinder currently in use. This |
| * is the unwinder with the highest rating. |
| * |
| * "unwinder_list" is a linked-list of all available unwinders, sorted |
| * by rating. |
| * |
| * All modifications of "curr_unwinder" and "unwinder_list" must be |
| * performed whilst holding "unwinder_lock". |
| */ |
| static struct unwinder *curr_unwinder = &stack_reader; |
| |
| static struct list_head unwinder_list = { |
| .next = &stack_reader.list, |
| .prev = &stack_reader.list, |
| }; |
| |
| static DEFINE_SPINLOCK(unwinder_lock); |
| |
| /** |
| * select_unwinder - Select the best registered stack unwinder. |
| * |
| * Private function. Must hold unwinder_lock when called. |
| * |
| * Select the stack unwinder with the best rating. This is useful for |
| * setting up curr_unwinder. |
| */ |
| static struct unwinder *select_unwinder(void) |
| { |
| struct unwinder *best; |
| |
| if (list_empty(&unwinder_list)) |
| return NULL; |
| |
| best = list_entry(unwinder_list.next, struct unwinder, list); |
| if (best == curr_unwinder) |
| return NULL; |
| |
| return best; |
| } |
| |
| /* |
| * Enqueue the stack unwinder sorted by rating. |
| */ |
| static int unwinder_enqueue(struct unwinder *ops) |
| { |
| struct list_head *tmp, *entry = &unwinder_list; |
| |
| list_for_each(tmp, &unwinder_list) { |
| struct unwinder *o; |
| |
| o = list_entry(tmp, struct unwinder, list); |
| if (o == ops) |
| return -EBUSY; |
| /* Keep track of the place, where to insert */ |
| if (o->rating >= ops->rating) |
| entry = tmp; |
| } |
| list_add(&ops->list, entry); |
| |
| return 0; |
| } |
| |
| /** |
| * unwinder_register - Used to install new stack unwinder |
| * @u: unwinder to be registered |
| * |
| * Install the new stack unwinder on the unwinder list, which is sorted |
| * by rating. |
| * |
| * Returns -EBUSY if registration fails, zero otherwise. |
| */ |
| int unwinder_register(struct unwinder *u) |
| { |
| unsigned long flags; |
| int ret; |
| |
| spin_lock_irqsave(&unwinder_lock, flags); |
| ret = unwinder_enqueue(u); |
| if (!ret) |
| curr_unwinder = select_unwinder(); |
| spin_unlock_irqrestore(&unwinder_lock, flags); |
| |
| return ret; |
| } |
| |
| int unwinder_faulted = 0; |
| |
| /* |
| * Unwind the call stack and pass information to the stacktrace_ops |
| * functions. Also handle the case where we need to switch to a new |
| * stack dumper because the current one faulted unexpectedly. |
| */ |
| void unwind_stack(struct task_struct *task, struct pt_regs *regs, |
| unsigned long *sp, const struct stacktrace_ops *ops, |
| void *data) |
| { |
| unsigned long flags; |
| |
| /* |
| * The problem with unwinders with high ratings is that they are |
| * inherently more complicated than the simple ones with lower |
| * ratings. We are therefore more likely to fault in the |
| * complicated ones, e.g. hitting BUG()s. If we fault in the |
| * code for the current stack unwinder we try to downgrade to |
| * one with a lower rating. |
| * |
| * Hopefully this will give us a semi-reliable stacktrace so we |
| * can diagnose why curr_unwinder->dump() faulted. |
| */ |
| if (unwinder_faulted) { |
| spin_lock_irqsave(&unwinder_lock, flags); |
| |
| /* Make sure no one beat us to changing the unwinder */ |
| if (unwinder_faulted && !list_is_singular(&unwinder_list)) { |
| list_del(&curr_unwinder->list); |
| curr_unwinder = select_unwinder(); |
| |
| unwinder_faulted = 0; |
| } |
| |
| spin_unlock_irqrestore(&unwinder_lock, flags); |
| } |
| |
| curr_unwinder->dump(task, regs, sp, ops, data); |
| } |
| |
| /* |
| * Trap handler for UWINDER_BUG() statements. We must switch to the |
| * unwinder with the next highest rating. |
| */ |
| BUILD_TRAP_HANDLER(unwinder) |
| { |
| insn_size_t insn; |
| TRAP_HANDLER_DECL; |
| |
| /* Rewind */ |
| regs->pc -= instruction_size(ctrl_inw(regs->pc - 4)); |
| insn = *(insn_size_t *)instruction_pointer(regs); |
| |
| /* Switch unwinders when unwind_stack() is called */ |
| unwinder_faulted = 1; |
| |
| #ifdef CONFIG_BUG |
| handle_BUG(regs); |
| #endif |
| } |