On Wed, 2008-05-14 at 23:49 -0400, Steven Rostedt wrote: > plain text document attachment (ftrace-powerpc-port.patch) > This patch adds full support for ftrace for PowerPC (both 64 and 32 bit). > This includes dynamic tracing and function filtering.
Hi Steven, Just a few comments inline .. > Index: linux-sched-devel.git/arch/powerpc/kernel/Makefile > =================================================================== > --- linux-sched-devel.git.orig/arch/powerpc/kernel/Makefile 2008-05-14 > 19:30:53.000000000 -0700 > +++ linux-sched-devel.git/arch/powerpc/kernel/Makefile 2008-05-14 > 19:31:56.000000000 -0700 > @@ -12,6 +12,18 @@ CFLAGS_prom_init.o += -fPIC > CFLAGS_btext.o += -fPIC > endif > > +ifdef CONFIG_FTRACE > +# Do not trace early boot code > +CFLAGS_REMOVE_cputable.o = -pg > +CFLAGS_REMOVE_prom_init.o = -pg Why do we not want to trace early boot? Just because it's not useful? > Index: linux-sched-devel.git/arch/powerpc/kernel/entry_32.S > =================================================================== > --- linux-sched-devel.git.orig/arch/powerpc/kernel/entry_32.S 2008-05-14 > 19:30:50.000000000 -0700 > +++ linux-sched-devel.git/arch/powerpc/kernel/entry_32.S 2008-05-14 > 19:31:56.000000000 -0700 > @@ -1035,3 +1035,133 @@ machine_check_in_rtas: > /* XXX load up BATs and panic */ > .. snip > +_GLOBAL(mcount) > +_GLOBAL(_mcount) > + stwu r1,-48(r1) > + stw r3, 12(r1) > + stw r4, 16(r1) > + stw r5, 20(r1) > + stw r6, 24(r1) > + mflr r3 > + lwz r4, 52(r1) > + mfcr r5 > + stw r7, 28(r1) > + stw r8, 32(r1) > + stw r9, 36(r1) > + stw r10,40(r1) > + stw r3, 44(r1) > + stw r5, 8(r1) > + > + LOAD_REG_ADDR(r5, ftrace_trace_function) > +#if 0 > + mtctr r3 > + mr r1, r5 > + bctrl > +#endif > + lwz r5,0(r5) > +#if 1 > + mtctr r5 > + bctrl > +#else > + bl ftrace_stub > +#endif #if 0, #if 1 ? > Index: linux-sched-devel.git/arch/powerpc/kernel/ftrace.c > =================================================================== > --- /dev/null 1970-01-01 00:00:00.000000000 +0000 > +++ linux-sched-devel.git/arch/powerpc/kernel/ftrace.c 2008-05-14 > 19:31:56.000000000 -0700 > @@ -0,0 +1,165 @@ > +/* > + * Code for replacing ftrace calls with jumps. > + * > + * Copyright (C) 2007-2008 Steven Rostedt <[EMAIL PROTECTED]> > + * > + * Thanks goes out to P.A. Semi, Inc for supplying me with a PPC64 box. > + * > + */ > + > +#include <linux/spinlock.h> > +#include <linux/hardirq.h> > +#include <linux/ftrace.h> > +#include <linux/percpu.h> > +#include <linux/init.h> > +#include <linux/list.h> > + > +#include <asm/cacheflush.h> > + > +#define CALL_BACK 4 I don't grok what you're doing with CALL_BACK, you add it in places and subtract in others - and it looks like you could do neither, but I haven't gone over it in detail. > +static unsigned int ftrace_nop = 0x60000000; I should really add a #define for that. > +#ifdef CONFIG_PPC32 > +# define GET_ADDR(addr) addr > +#else > +/* PowerPC64's functions are data that points to the functions */ > +# define GET_ADDR(addr) *(unsigned long *)addr > +#endif And that. .. snip > +notrace unsigned char *ftrace_call_replace(unsigned long ip, unsigned long > addr) > +{ > + static unsigned int op; > + > + addr = GET_ADDR(addr); > + > + /* Set to "bl addr" */ > + op = 0x48000001 | (ftrace_calc_offset(ip, addr) & 0x03fffffe); 0x03fffffe should be 0x03fffffc, if you set bit 1 you'll end with a "bla" instruction, ie. branch absolute and link. That shouldn't happen as long as ip and addr are properly aligned, but still. In fact I think you should just use create_function_call() or create_branch() from include/asm-powerpc/system.h > +#ifdef CONFIG_PPC64 > +# define _ASM_ALIGN " .align 3 " > +# define _ASM_PTR " .llong " > +#else > +# define _ASM_ALIGN " .align 2 " > +# define _ASM_PTR " .long " > +#endif We already have a #define for .long, it's called PPC_LONG (asm/asm-compat.h) Perhaps we should add one for .align, PPC_LONG_ALIGN or something? > +notrace int > +ftrace_modify_code(unsigned long ip, unsigned char *old_code, > + unsigned char *new_code) > +{ > + unsigned replaced; > + unsigned old = *(unsigned *)old_code; > + unsigned new = *(unsigned *)new_code; > + int faulted = 0; > + > + /* move the IP back to the start of the call */ > + ip -= CALL_BACK; > + > + /* > + * Note: Due to modules and __init, code can > + * disappear and change, we need to protect against faulting > + * as well as code changing. > + * > + * No real locking needed, this code is run through > + * kstop_machine. > + */ > + asm volatile ( > + "1: lwz %1, 0(%2)\n" > + " cmpw %1, %5\n" > + " bne 2f\n" > + " stwu %3, 0(%2)\n" > + "2:\n" > + ".section .fixup, \"ax\"\n" > + "3: li %0, 1\n" > + " b 2b\n" > + ".previous\n" > + ".section __ex_table,\"a\"\n" > + _ASM_ALIGN "\n" > + _ASM_PTR "1b, 3b\n" > + ".previous" Or perhaps we just need a macro for adding exception table entries. > + : "=r"(faulted), "=r"(replaced) > + : "r"(ip), "r"(new), > + "0"(faulted), "r"(old) > + : "memory"); > + > + if (replaced != old && replaced != new) > + faulted = 2; > + > + if (!faulted) > + flush_icache_range(ip, ip + 8); > + > + return faulted; > +} > Index: linux-sched-devel.git/arch/powerpc/kernel/setup_32.c > =================================================================== > --- linux-sched-devel.git.orig/arch/powerpc/kernel/setup_32.c 2008-05-14 > 19:30:50.000000000 -0700 > +++ linux-sched-devel.git/arch/powerpc/kernel/setup_32.c 2008-05-14 > 19:31:56.000000000 -0700 > @@ -47,6 +47,11 @@ > #include <asm/kgdb.h> > #endif > > +#ifdef CONFIG_FTRACE > +extern void _mcount(void); > +EXPORT_SYMBOL(_mcount); > +#endif Can you please put the extern in a header, and the EXPORT_SYMBOL in arch/powerpc/kernel/ftrace.c? > Index: linux-sched-devel.git/arch/powerpc/kernel/setup_64.c > =================================================================== > --- linux-sched-devel.git.orig/arch/powerpc/kernel/setup_64.c 2008-05-14 > 19:30:50.000000000 -0700 > +++ linux-sched-devel.git/arch/powerpc/kernel/setup_64.c 2008-05-14 > 19:31:56.000000000 -0700 > @@ -85,6 +85,11 @@ struct ppc64_caches ppc64_caches = { > }; > EXPORT_SYMBOL_GPL(ppc64_caches); > > +#ifdef CONFIG_FTRACE > +extern void _mcount(void); > +EXPORT_SYMBOL(_mcount); > +#endif Ditto. cheers -- Michael Ellerman OzLabs, IBM Australia Development Lab wwweb: http://michael.ellerman.id.au phone: +61 2 6212 1183 (tie line 70 21183) We do not inherit the earth from our ancestors, we borrow it from our children. - S.M.A.R.T Person
signature.asc
Description: This is a digitally signed message part
_______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@ozlabs.org https://ozlabs.org/mailman/listinfo/linuxppc-dev