diff options
author | Frederic Weisbecker <fweisbec@gmail.com> | 2009-09-28 19:12:49 +0400 |
---|---|---|
committer | Frederic Weisbecker <fweisbec@gmail.com> | 2009-09-28 20:00:48 +0400 |
commit | 925936ebf35a95c290e010b784c962164e6728f3 (patch) | |
tree | e0bae3a2655f84dcf238c9acacd158e5ce422b32 /arch/sh/include/asm/entry-macros.S | |
parent | 3f6fe06dbf67b46d36fedec502300e04dffeb67a (diff) | |
download | linux-925936ebf35a95c290e010b784c962164e6728f3.tar.xz |
tracing: Pushdown the bkl tracepoints calls
Currently we are calling the bkl tracepoint callbacks just before the
bkl lock/unlock operations, ie the tracepoint call is not inside a
lock_kernel() function but inside a lock_kernel() macro. Hence the
bkl trace event header must be included from smp_lock.h. This raises
some nasty circular header dependencies:
linux/smp_lock.h -> trace/events/bkl.h -> trace/define_trace.h
-> trace/ftrace.h -> linux/ftrace_event.h -> linux/hardirq.h
-> linux/smp_lock.h
This results in incomplete event declarations, spurious event
definitions and other kind of funny behaviours.
This is hardly fixable without ugly workarounds. So instead, we push
the file name, line number and function name as lock_kernel()
parameters, so that we only deal with the trace event header from
lib/kernel_lock.c
This adds two parameters to lock_kernel() and unlock_kernel() but
it should be fine wrt to performances because this pair dos not seem
to be called in fast paths.
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Li Zefan <lizf@cn.fujitsu.com>
Diffstat (limited to 'arch/sh/include/asm/entry-macros.S')
0 files changed, 0 insertions, 0 deletions