sh: Function graph tracer support
Add both dynamic and static function graph tracer support for sh. Signed-off-by: Matt Fleming <matt@console-pimps.org> Signed-off-by: Paul Mundt <lethal@linux-sh.org>
This commit is contained in:
parent
b99610fb9c
commit
327933f5d6
7 changed files with 244 additions and 1 deletions
|
@ -33,6 +33,7 @@ config SUPERH32
|
||||||
select HAVE_DYNAMIC_FTRACE
|
select HAVE_DYNAMIC_FTRACE
|
||||||
select HAVE_FUNCTION_TRACE_MCOUNT_TEST
|
select HAVE_FUNCTION_TRACE_MCOUNT_TEST
|
||||||
select HAVE_FTRACE_SYSCALLS
|
select HAVE_FTRACE_SYSCALLS
|
||||||
|
select HAVE_FUNCTION_GRAPH_TRACER
|
||||||
select HAVE_ARCH_KGDB
|
select HAVE_ARCH_KGDB
|
||||||
select ARCH_HIBERNATION_POSSIBLE if MMU
|
select ARCH_HIBERNATION_POSSIBLE if MMU
|
||||||
|
|
||||||
|
|
|
@ -13,8 +13,11 @@ extern void mcount(void);
|
||||||
#ifdef CONFIG_DYNAMIC_FTRACE
|
#ifdef CONFIG_DYNAMIC_FTRACE
|
||||||
#define CALL_ADDR ((long)(ftrace_call))
|
#define CALL_ADDR ((long)(ftrace_call))
|
||||||
#define STUB_ADDR ((long)(ftrace_stub))
|
#define STUB_ADDR ((long)(ftrace_stub))
|
||||||
|
#define GRAPH_ADDR ((long)(ftrace_graph_call))
|
||||||
|
#define CALLER_ADDR ((long)(ftrace_caller))
|
||||||
|
|
||||||
#define MCOUNT_INSN_OFFSET ((STUB_ADDR - CALL_ADDR) - 4)
|
#define MCOUNT_INSN_OFFSET ((STUB_ADDR - CALL_ADDR) - 4)
|
||||||
|
#define GRAPH_INSN_OFFSET ((CALLER_ADDR - GRAPH_ADDR) - 4)
|
||||||
|
|
||||||
struct dyn_arch_ftrace {
|
struct dyn_arch_ftrace {
|
||||||
/* No extra data needed on sh */
|
/* No extra data needed on sh */
|
||||||
|
|
|
@ -30,6 +30,7 @@ obj-$(CONFIG_KPROBES) += kprobes.o
|
||||||
obj-$(CONFIG_GENERIC_GPIO) += gpio.o
|
obj-$(CONFIG_GENERIC_GPIO) += gpio.o
|
||||||
obj-$(CONFIG_DYNAMIC_FTRACE) += ftrace.o
|
obj-$(CONFIG_DYNAMIC_FTRACE) += ftrace.o
|
||||||
obj-$(CONFIG_FTRACE_SYSCALLS) += ftrace.o
|
obj-$(CONFIG_FTRACE_SYSCALLS) += ftrace.o
|
||||||
|
obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += ftrace.o
|
||||||
obj-$(CONFIG_DUMP_CODE) += disassemble.o
|
obj-$(CONFIG_DUMP_CODE) += disassemble.o
|
||||||
obj-$(CONFIG_HIBERNATION) += swsusp.o
|
obj-$(CONFIG_HIBERNATION) += swsusp.o
|
||||||
|
|
||||||
|
|
|
@ -16,11 +16,13 @@
|
||||||
#include <linux/string.h>
|
#include <linux/string.h>
|
||||||
#include <linux/init.h>
|
#include <linux/init.h>
|
||||||
#include <linux/io.h>
|
#include <linux/io.h>
|
||||||
|
#include <linux/kernel.h>
|
||||||
#include <asm/ftrace.h>
|
#include <asm/ftrace.h>
|
||||||
#include <asm/cacheflush.h>
|
#include <asm/cacheflush.h>
|
||||||
#include <asm/unistd.h>
|
#include <asm/unistd.h>
|
||||||
#include <trace/syscall.h>
|
#include <trace/syscall.h>
|
||||||
|
|
||||||
|
#ifdef CONFIG_DYNAMIC_FTRACE
|
||||||
static unsigned char ftrace_replaced_code[MCOUNT_INSN_SIZE];
|
static unsigned char ftrace_replaced_code[MCOUNT_INSN_SIZE];
|
||||||
|
|
||||||
static unsigned char ftrace_nop[4];
|
static unsigned char ftrace_nop[4];
|
||||||
|
@ -133,6 +135,126 @@ int __init ftrace_dyn_arch_init(void *data)
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
#endif /* CONFIG_DYNAMIC_FTRACE */
|
||||||
|
|
||||||
|
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||||
|
#ifdef CONFIG_DYNAMIC_FTRACE
|
||||||
|
extern void ftrace_graph_call(void);
|
||||||
|
|
||||||
|
static int ftrace_mod(unsigned long ip, unsigned long old_addr,
|
||||||
|
unsigned long new_addr)
|
||||||
|
{
|
||||||
|
unsigned char code[MCOUNT_INSN_SIZE];
|
||||||
|
|
||||||
|
if (probe_kernel_read(code, (void *)ip, MCOUNT_INSN_SIZE))
|
||||||
|
return -EFAULT;
|
||||||
|
|
||||||
|
if (old_addr != __raw_readl((unsigned long *)code))
|
||||||
|
return -EINVAL;
|
||||||
|
|
||||||
|
__raw_writel(new_addr, ip);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
int ftrace_enable_ftrace_graph_caller(void)
|
||||||
|
{
|
||||||
|
unsigned long ip, old_addr, new_addr;
|
||||||
|
|
||||||
|
ip = (unsigned long)(&ftrace_graph_call) + GRAPH_INSN_OFFSET;
|
||||||
|
old_addr = (unsigned long)(&skip_trace);
|
||||||
|
new_addr = (unsigned long)(&ftrace_graph_caller);
|
||||||
|
|
||||||
|
return ftrace_mod(ip, old_addr, new_addr);
|
||||||
|
}
|
||||||
|
|
||||||
|
int ftrace_disable_ftrace_graph_caller(void)
|
||||||
|
{
|
||||||
|
unsigned long ip, old_addr, new_addr;
|
||||||
|
|
||||||
|
ip = (unsigned long)(&ftrace_graph_call) + GRAPH_INSN_OFFSET;
|
||||||
|
old_addr = (unsigned long)(&ftrace_graph_caller);
|
||||||
|
new_addr = (unsigned long)(&skip_trace);
|
||||||
|
|
||||||
|
return ftrace_mod(ip, old_addr, new_addr);
|
||||||
|
}
|
||||||
|
#endif /* CONFIG_DYNAMIC_FTRACE */
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Hook the return address and push it in the stack of return addrs
|
||||||
|
* in the current thread info.
|
||||||
|
*
|
||||||
|
* This is the main routine for the function graph tracer. The function
|
||||||
|
* graph tracer essentially works like this:
|
||||||
|
*
|
||||||
|
* parent is the stack address containing self_addr's return address.
|
||||||
|
* We pull the real return address out of parent and store it in
|
||||||
|
* current's ret_stack. Then, we replace the return address on the stack
|
||||||
|
* with the address of return_to_handler. self_addr is the function that
|
||||||
|
* called mcount.
|
||||||
|
*
|
||||||
|
* When self_addr returns, it will jump to return_to_handler which calls
|
||||||
|
* ftrace_return_to_handler. ftrace_return_to_handler will pull the real
|
||||||
|
* return address off of current's ret_stack and jump to it.
|
||||||
|
*/
|
||||||
|
void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr)
|
||||||
|
{
|
||||||
|
unsigned long old;
|
||||||
|
int faulted, err;
|
||||||
|
struct ftrace_graph_ent trace;
|
||||||
|
unsigned long return_hooker = (unsigned long)&return_to_handler;
|
||||||
|
|
||||||
|
if (unlikely(atomic_read(¤t->tracing_graph_pause)))
|
||||||
|
return;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Protect against fault, even if it shouldn't
|
||||||
|
* happen. This tool is too much intrusive to
|
||||||
|
* ignore such a protection.
|
||||||
|
*/
|
||||||
|
__asm__ __volatile__(
|
||||||
|
"1: \n\t"
|
||||||
|
"mov.l @%2, %0 \n\t"
|
||||||
|
"2: \n\t"
|
||||||
|
"mov.l %3, @%2 \n\t"
|
||||||
|
"mov #0, %1 \n\t"
|
||||||
|
"3: \n\t"
|
||||||
|
".section .fixup, \"ax\" \n\t"
|
||||||
|
"4: \n\t"
|
||||||
|
"mov.l 5f, %0 \n\t"
|
||||||
|
"jmp @%0 \n\t"
|
||||||
|
" mov #1, %1 \n\t"
|
||||||
|
".balign 4 \n\t"
|
||||||
|
"5: .long 3b \n\t"
|
||||||
|
".previous \n\t"
|
||||||
|
".section __ex_table,\"a\" \n\t"
|
||||||
|
".long 1b, 4b \n\t"
|
||||||
|
".long 2b, 4b \n\t"
|
||||||
|
".previous \n\t"
|
||||||
|
: "=&r" (old), "=r" (faulted)
|
||||||
|
: "r" (parent), "r" (return_hooker)
|
||||||
|
);
|
||||||
|
|
||||||
|
if (unlikely(faulted)) {
|
||||||
|
ftrace_graph_stop();
|
||||||
|
WARN_ON(1);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
err = ftrace_push_return_trace(old, self_addr, &trace.depth, 0);
|
||||||
|
if (err == -EBUSY) {
|
||||||
|
__raw_writel(old, parent);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
trace.func = self_addr;
|
||||||
|
|
||||||
|
/* Only trace if the calling function expects to */
|
||||||
|
if (!ftrace_graph_entry(&trace)) {
|
||||||
|
current->curr_ret_stack--;
|
||||||
|
__raw_writel(old, parent);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||||
|
|
||||||
#ifdef CONFIG_FTRACE_SYSCALLS
|
#ifdef CONFIG_FTRACE_SYSCALLS
|
||||||
|
|
||||||
|
|
0
arch/sh/kernel/vmlinux_64.lds.S
Normal file
0
arch/sh/kernel/vmlinux_64.lds.S
Normal file
|
@ -25,6 +25,7 @@ memcpy-$(CONFIG_CPU_SH4) := memcpy-sh4.o
|
||||||
|
|
||||||
lib-$(CONFIG_MMU) += copy_page.o clear_page.o
|
lib-$(CONFIG_MMU) += copy_page.o clear_page.o
|
||||||
lib-$(CONFIG_FUNCTION_TRACER) += mcount.o
|
lib-$(CONFIG_FUNCTION_TRACER) += mcount.o
|
||||||
|
lib-$(CONFIG_FUNCTION_GRAPH_TRACER) += mcount.o
|
||||||
lib-y += $(memcpy-y) $(udivsi3-y)
|
lib-y += $(memcpy-y) $(udivsi3-y)
|
||||||
|
|
||||||
EXTRA_CFLAGS += -Werror
|
EXTRA_CFLAGS += -Werror
|
||||||
|
|
|
@ -111,14 +111,62 @@ mcount_call:
|
||||||
jsr @r6
|
jsr @r6
|
||||||
nop
|
nop
|
||||||
|
|
||||||
|
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||||
|
mov.l .Lftrace_graph_return, r6
|
||||||
|
mov.l .Lftrace_stub, r7
|
||||||
|
cmp/eq r6, r7
|
||||||
|
bt 1f
|
||||||
|
|
||||||
|
mov.l .Lftrace_graph_caller, r0
|
||||||
|
jmp @r0
|
||||||
|
nop
|
||||||
|
|
||||||
|
1:
|
||||||
|
mov.l .Lftrace_graph_entry, r6
|
||||||
|
mov.l .Lftrace_graph_entry_stub, r7
|
||||||
|
cmp/eq r6, r7
|
||||||
|
bt skip_trace
|
||||||
|
|
||||||
|
mov.l .Lftrace_graph_caller, r0
|
||||||
|
jmp @r0
|
||||||
|
nop
|
||||||
|
|
||||||
|
.align 2
|
||||||
|
.Lftrace_graph_return:
|
||||||
|
.long ftrace_graph_return
|
||||||
|
.Lftrace_graph_entry:
|
||||||
|
.long ftrace_graph_entry
|
||||||
|
.Lftrace_graph_entry_stub:
|
||||||
|
.long ftrace_graph_entry_stub
|
||||||
|
.Lftrace_graph_caller:
|
||||||
|
.long ftrace_graph_caller
|
||||||
|
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||||
|
|
||||||
|
.globl skip_trace
|
||||||
skip_trace:
|
skip_trace:
|
||||||
MCOUNT_LEAVE()
|
MCOUNT_LEAVE()
|
||||||
|
|
||||||
.align 2
|
.align 2
|
||||||
.Lftrace_trace_function:
|
.Lftrace_trace_function:
|
||||||
.long ftrace_trace_function
|
.long ftrace_trace_function
|
||||||
|
|
||||||
#ifdef CONFIG_DYNAMIC_FTRACE
|
#ifdef CONFIG_DYNAMIC_FTRACE
|
||||||
|
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||||
|
/*
|
||||||
|
* NOTE: Do not move either ftrace_graph_call or ftrace_caller
|
||||||
|
* as this will affect the calculation of GRAPH_INSN_OFFSET.
|
||||||
|
*/
|
||||||
|
.globl ftrace_graph_call
|
||||||
|
ftrace_graph_call:
|
||||||
|
mov.l .Lskip_trace, r0
|
||||||
|
jmp @r0
|
||||||
|
nop
|
||||||
|
|
||||||
|
.align 2
|
||||||
|
.Lskip_trace:
|
||||||
|
.long skip_trace
|
||||||
|
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||||
|
|
||||||
.globl ftrace_caller
|
.globl ftrace_caller
|
||||||
ftrace_caller:
|
ftrace_caller:
|
||||||
mov.l .Lfunction_trace_stop, r0
|
mov.l .Lfunction_trace_stop, r0
|
||||||
|
@ -136,7 +184,12 @@ ftrace_call:
|
||||||
jsr @r6
|
jsr @r6
|
||||||
nop
|
nop
|
||||||
|
|
||||||
|
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||||
|
bra ftrace_graph_call
|
||||||
|
nop
|
||||||
|
#else
|
||||||
MCOUNT_LEAVE()
|
MCOUNT_LEAVE()
|
||||||
|
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||||
#endif /* CONFIG_DYNAMIC_FTRACE */
|
#endif /* CONFIG_DYNAMIC_FTRACE */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -188,3 +241,65 @@ stack_panic:
|
||||||
.Lpanic_str:
|
.Lpanic_str:
|
||||||
.string "Stack error"
|
.string "Stack error"
|
||||||
#endif /* CONFIG_STACK_DEBUG */
|
#endif /* CONFIG_STACK_DEBUG */
|
||||||
|
|
||||||
|
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||||
|
.globl ftrace_graph_caller
|
||||||
|
ftrace_graph_caller:
|
||||||
|
mov.l 2f, r0
|
||||||
|
mov.l @r0, r0
|
||||||
|
tst r0, r0
|
||||||
|
bt 1f
|
||||||
|
|
||||||
|
mov.l 3f, r1
|
||||||
|
jmp @r1
|
||||||
|
nop
|
||||||
|
1:
|
||||||
|
/*
|
||||||
|
* MCOUNT_ENTER() pushed 5 registers onto the stack, so
|
||||||
|
* the stack address containing our return address is
|
||||||
|
* r15 + 20.
|
||||||
|
*/
|
||||||
|
mov #20, r0
|
||||||
|
add r15, r0
|
||||||
|
mov r0, r4
|
||||||
|
|
||||||
|
mov.l .Lprepare_ftrace_return, r0
|
||||||
|
jsr @r0
|
||||||
|
nop
|
||||||
|
|
||||||
|
MCOUNT_LEAVE()
|
||||||
|
|
||||||
|
.align 2
|
||||||
|
2: .long function_trace_stop
|
||||||
|
3: .long skip_trace
|
||||||
|
.Lprepare_ftrace_return:
|
||||||
|
.long prepare_ftrace_return
|
||||||
|
|
||||||
|
.globl return_to_handler
|
||||||
|
return_to_handler:
|
||||||
|
/*
|
||||||
|
* Save the return values.
|
||||||
|
*/
|
||||||
|
mov.l r0, @-r15
|
||||||
|
mov.l r1, @-r15
|
||||||
|
|
||||||
|
mov #0, r4
|
||||||
|
|
||||||
|
mov.l .Lftrace_return_to_handler, r0
|
||||||
|
jsr @r0
|
||||||
|
nop
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The return value from ftrace_return_handler has the real
|
||||||
|
* address that we should return to.
|
||||||
|
*/
|
||||||
|
lds r0, pr
|
||||||
|
mov.l @r15+, r1
|
||||||
|
rts
|
||||||
|
mov.l @r15+, r0
|
||||||
|
|
||||||
|
|
||||||
|
.align 2
|
||||||
|
.Lftrace_return_to_handler:
|
||||||
|
.long ftrace_return_to_handler
|
||||||
|
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||||
|
|
Loading…
Reference in a new issue