1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889 |
- /*
- * Copyright IBM Corp. 2008,2009
- *
- * Author(s): Heiko Carstens <heiko.carstens@de.ibm.com>,
- *
- */
- #include <asm/asm-offsets.h>
- .globl ftrace_stub
- ftrace_stub:
- br %r14
- .globl _mcount
- _mcount:
- #ifdef CONFIG_DYNAMIC_FTRACE
- br %r14
- .data
- .globl ftrace_dyn_func
- ftrace_dyn_func:
- .long ftrace_stub
- .previous
- .globl ftrace_caller
- ftrace_caller:
- #endif
- stm %r2,%r5,16(%r15)
- bras %r1,2f
- #ifdef CONFIG_DYNAMIC_FTRACE
- 0: .long ftrace_dyn_func
- #else
- 0: .long ftrace_trace_function
- #endif
- 1: .long function_trace_stop
- 2: l %r2,1b-0b(%r1)
- icm %r2,0xf,0(%r2)
- jnz 3f
- st %r14,56(%r15)
- lr %r0,%r15
- ahi %r15,-96
- l %r3,100(%r15)
- la %r2,0(%r14)
- st %r0,__SF_BACKCHAIN(%r15)
- la %r3,0(%r3)
- l %r14,0b-0b(%r1)
- l %r14,0(%r14)
- basr %r14,%r14
- #ifdef CONFIG_FUNCTION_GRAPH_TRACER
- #ifdef CONFIG_DYNAMIC_FTRACE
- .globl ftrace_graph_caller
- ftrace_graph_caller:
- # This unconditional branch gets runtime patched. Change only if
- # you know what you are doing. See ftrace_enable_graph_caller().
- j 1f
- #endif
- bras %r1,0f
- .long prepare_ftrace_return
- 0: l %r2,152(%r15)
- l %r4,0(%r1)
- l %r3,100(%r15)
- basr %r14,%r4
- st %r2,100(%r15)
- 1:
- #endif
- ahi %r15,96
- l %r14,56(%r15)
- 3: lm %r2,%r5,16(%r15)
- br %r14
- #ifdef CONFIG_FUNCTION_GRAPH_TRACER
- .globl return_to_handler
- return_to_handler:
- stm %r2,%r5,16(%r15)
- st %r14,56(%r15)
- lr %r0,%r15
- ahi %r15,-96
- st %r0,__SF_BACKCHAIN(%r15)
- bras %r1,0f
- .long ftrace_return_to_handler
- 0: l %r2,0b-0b(%r1)
- basr %r14,%r2
- lr %r14,%r2
- ahi %r15,96
- lm %r2,%r5,16(%r15)
- br %r14
- #endif
|