summaryrefslogtreecommitdiffstats
path: root/lib/kernel_lock.c
diff options
context:
space:
mode:
Diffstat (limited to 'lib/kernel_lock.c')
-rw-r--r--lib/kernel_lock.c15
1 files changed, 11 insertions, 4 deletions
diff --git a/lib/kernel_lock.c b/lib/kernel_lock.c
index 5c10b2e1fd0..4ebfa5a164d 100644
--- a/lib/kernel_lock.c
+++ b/lib/kernel_lock.c
@@ -8,9 +8,11 @@
#include <linux/module.h>
#include <linux/kallsyms.h>
#include <linux/semaphore.h>
-#define CREATE_TRACE_POINTS
#include <linux/smp_lock.h>
+#define CREATE_TRACE_POINTS
+#include <trace/events/bkl.h>
+
/*
* The 'big kernel lock'
*
@@ -114,19 +116,24 @@ static inline void __unlock_kernel(void)
* This cannot happen asynchronously, so we only need to
* worry about other CPU's.
*/
-void __lockfunc _lock_kernel(void)
+void __lockfunc _lock_kernel(const char *func, const char *file, int line)
{
- int depth = current->lock_depth+1;
+ int depth = current->lock_depth + 1;
+
+ trace_lock_kernel(func, file, line);
+
if (likely(!depth))
__lock_kernel();
current->lock_depth = depth;
}
-void __lockfunc _unlock_kernel(void)
+void __lockfunc _unlock_kernel(const char *func, const char *file, int line)
{
BUG_ON(current->lock_depth < 0);
if (likely(--current->lock_depth < 0))
__unlock_kernel();
+
+ trace_unlock_kernel(func, file, line);
}
EXPORT_SYMBOL(_lock_kernel);