123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165 |
- /**
- * @file backtrace.c
- *
- * @remark Copyright 2002 OProfile authors
- * @remark Read the file COPYING
- *
- * @author John Levon
- * @author David Smith
- */
- #include <linux/oprofile.h>
- #include <linux/sched.h>
- #include <linux/mm.h>
- #include <linux/compat.h>
- #include <linux/highmem.h>
- #include <asm/ptrace.h>
- #include <asm/uaccess.h>
- #include <asm/stacktrace.h>
- static int backtrace_stack(void *data, char *name)
- {
- /* Yes, we want all stacks */
- return 0;
- }
- static void backtrace_address(void *data, unsigned long addr, int reliable)
- {
- unsigned int *depth = data;
- if ((*depth)--)
- oprofile_add_trace(addr);
- }
- static struct stacktrace_ops backtrace_ops = {
- .stack = backtrace_stack,
- .address = backtrace_address,
- .walk_stack = print_context_stack,
- };
- /* from arch/x86/kernel/cpu/perf_event.c: */
- /*
- * best effort, GUP based copy_from_user() that assumes IRQ or NMI context
- */
- static unsigned long
- copy_from_user_nmi(void *to, const void __user *from, unsigned long n)
- {
- unsigned long offset, addr = (unsigned long)from;
- unsigned long size, len = 0;
- struct page *page;
- void *map;
- int ret;
- do {
- ret = __get_user_pages_fast(addr, 1, 0, &page);
- if (!ret)
- break;
- offset = addr & (PAGE_SIZE - 1);
- size = min(PAGE_SIZE - offset, n - len);
- map = kmap_atomic(page);
- memcpy(to, map+offset, size);
- kunmap_atomic(map);
- put_page(page);
- len += size;
- to += size;
- addr += size;
- } while (len < n);
- return len;
- }
- #ifdef CONFIG_COMPAT
- static struct stack_frame_ia32 *
- dump_user_backtrace_32(struct stack_frame_ia32 *head)
- {
- /* Also check accessibility of one struct frame_head beyond: */
- struct stack_frame_ia32 bufhead[2];
- struct stack_frame_ia32 *fp;
- unsigned long bytes;
- bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead));
- if (bytes != sizeof(bufhead))
- return NULL;
- fp = (struct stack_frame_ia32 *) compat_ptr(bufhead[0].next_frame);
- oprofile_add_trace(bufhead[0].return_address);
- /* frame pointers should strictly progress back up the stack
- * (towards higher addresses) */
- if (head >= fp)
- return NULL;
- return fp;
- }
- static inline int
- x86_backtrace_32(struct pt_regs * const regs, unsigned int depth)
- {
- struct stack_frame_ia32 *head;
- /* User process is 32-bit */
- if (!current || !test_thread_flag(TIF_IA32))
- return 0;
- head = (struct stack_frame_ia32 *) regs->bp;
- while (depth-- && head)
- head = dump_user_backtrace_32(head);
- return 1;
- }
- #else
- static inline int
- x86_backtrace_32(struct pt_regs * const regs, unsigned int depth)
- {
- return 0;
- }
- #endif /* CONFIG_COMPAT */
- static struct stack_frame *dump_user_backtrace(struct stack_frame *head)
- {
- /* Also check accessibility of one struct frame_head beyond: */
- struct stack_frame bufhead[2];
- unsigned long bytes;
- bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead));
- if (bytes != sizeof(bufhead))
- return NULL;
- oprofile_add_trace(bufhead[0].return_address);
- /* frame pointers should strictly progress back up the stack
- * (towards higher addresses) */
- if (head >= bufhead[0].next_frame)
- return NULL;
- return bufhead[0].next_frame;
- }
- void
- x86_backtrace(struct pt_regs * const regs, unsigned int depth)
- {
- struct stack_frame *head = (struct stack_frame *)frame_pointer(regs);
- if (!user_mode_vm(regs)) {
- unsigned long stack = kernel_stack_pointer(regs);
- if (depth)
- dump_trace(NULL, regs, (unsigned long *)stack, 0,
- &backtrace_ops, &depth);
- return;
- }
- if (x86_backtrace_32(regs, depth))
- return;
- while (depth-- && head)
- head = dump_user_backtrace(head);
- }
|