/* * kernel/lockdep_proc.c * * Runtime locking correctness validator * * Started by Ingo Molnar: * * Copyright (C) 2006 Red Hat, Inc., Ingo Molnar <mingo@redhat.com> * * Code for /proc/lockdep and /proc/lockdep_stats: * */ #include <linux/module.h> #include <linux/proc_fs.h> #include <linux/seq_file.h> #include <linux/kallsyms.h> #include <linux/debug_locks.h> #include "lockdep_internals.h" static void *l_next(struct seq_file *m, void *v, loff_t *pos) { struct lock_class *class = v; (*pos)++; if (class->lock_entry.next != &all_lock_classes) class = list_entry(class->lock_entry.next, struct lock_class, lock_entry); else class = NULL; m->private = class; return class; } static void *l_start(struct seq_file *m, loff_t *pos) { struct lock_class *class = m->private; if (&class->lock_entry == all_lock_classes.next) seq_printf(m, "all lock classes:\n"); return class; } static void l_stop(struct seq_file *m, void *v) { } static unsigned long count_forward_deps(struct lock_class *class) { struct lock_list *entry; unsigned long ret = 1; /* * Recurse this class's dependency list: */ list_for_each_entry(entry, &class->locks_after, entry) ret += count_forward_deps(entry->class); return ret; } static unsigned long count_backward_deps(struct lock_class *class) { struct lock_list *entry; unsigned long ret = 1; /* * Recurse this class's dependency list: */ list_for_each_entry(entry, &class->locks_before, entry) ret += count_backward_deps(entry->class); return ret; } static void print_name(struct seq_file *m, struct lock_class *class) { char str[128]; const char *name = class->name; if (!name) { name = __get_key_name(class->key, str); seq_printf(m, "%s", name); } else{ seq_printf(m, "%s", name); if (class->name_version > 1) seq_printf(m, "#%d", class->name_version); if (class->subclass) seq_printf(m, "/%d", class->subclass); } } static int l_show(struct seq_file *m, void *v) { unsigned long nr_forward_deps, nr_backward_deps; struct lock_class *class = m->private; struct lock_list *entry; char c1, c2, c3, c4; seq_printf(m, "%p", class->key); #ifdef CONFIG_DEBUG_LOCKDEP seq_printf(m, " OPS:%8ld", class->ops); #endif nr_forward_deps = count_forward_deps(class); seq_printf(m, " FD:%5ld", nr_forward_deps); nr_backward_deps = count_backward_deps(class); seq_printf(m, " BD:%5ld", nr_backward_deps); get_usage_chars(class, &c1, &c2, &c3, &c4); seq_printf(m, " %c%c%c%c", c1, c2, c3, c4); seq_printf(m, ": "); print_name(m, class); seq_puts(m, "\n"); list_for_each_entry(entry, &class->locks_after, entry) { if (entry->distance == 1) { seq_printf(m, " -> [%p] ", entry->class); print_name(m, entry->class); seq_puts(m, "\n"); } } seq_puts(m, "\n"); return 0; } static const struct seq_operations lockdep_ops = { .start = l_start, .next = l_next, .stop = l_stop, .show = l_show, }; static int lockdep_open(struct inode *inode, struct file *file) { int res = seq_open(file, &lockdep_ops); if (!res) { struct seq_file *m = file->private_data; if (!list_empty(&all_lock_classes)) m->private = list_entry(all_lock_classes.next, struct lock_class, lock_entry); else m->private = NULL; } return res; } static const struct file_operations proc_lockdep_operations = { .open = lockdep_open, .read = seq_read, .llseek = seq_lseek, .release = seq_release, }; static void lockdep_stats_debug_show(struct seq_file *m) { #ifdef CONFIG_DEBUG_LOCKDEP unsigned int hi1 = debug_atomic_read(&hardirqs_on_events), hi2 = debug_atomic_read(&hardirqs_off_events), hr1 = debug_atomic_read(&redundant_hardirqs_on), hr2 = debug_atomic_read(&redundant_hardirqs_off), si1 = debug_atomic_read(&softirqs_on_events), si2 = debug_atomic_read(&softirqs_off_events), sr1 = debug_atomic_read(&redundant_softirqs_on), sr2 = debug_atomic_read(&redundant_softirqs_off); seq_printf(m, " chain lookup misses: %11u\n", debug_atomic_read(&chain_lookup_misses)); seq_printf(m, " chain lookup hits: %11u\n", debug_atomic_read(&chain_lookup_hits)); seq_printf(m, " cyclic checks: %11u\n", debug_atomic_read(&nr_cyclic_checks)); seq_printf(m, " cyclic-check recursions: %11u\n", debug_atomic_read(&nr_cyclic_check_recursions)); seq_printf(m, " find-mask forwards checks: %11u\n", debug_atomic_read(&nr_find_usage_forwards_checks)); seq_printf(m, " find-mask forwards recursions: %11u\n", debug_atomic_read(&nr_find_usage_forwards_recursions)); seq_printf(m, " find-mask backwards checks: %11u\n", debug_atomic_read(&nr_find_usage_backwards_checks)); seq_printf(m, " find-mask backwards recursions:%11u\n", debug_atomic_read(&nr_find_usage_backwards_recursions)); seq_printf(m, " hardirq on events: %11u\n", hi1); seq_printf(m, " hardirq off events: %11u\n", hi2); seq_printf(m, " redundant hardirq ons: %11u\n", hr1); seq_printf(m, " redundant hardirq offs: %11u\n", hr2); seq_printf(m, " softirq on events: %11u\n", si1); seq_printf(m, " softirq off events: %11u\n", si2); seq_printf(m, " redundant softirq ons: %11u\n", sr1); seq_printf(m, " redundant softirq offs: %11u\n", sr2); #endif } static int lockdep_stats_show(struct seq_file *m, void *v) { struct lock_class *class; unsigned long nr_unused = 0, nr_uncategorized = 0, nr_irq_safe = 0, nr_irq_unsafe = 0, nr_softirq_safe = 0, nr_softirq_unsafe = 0, nr_hardirq_safe = 0, nr_hardirq_unsafe = 0, nr_irq_read_safe = 0, nr_irq_read_unsafe = 0, nr_softirq_read_safe = 0, nr_softirq_read_unsafe = 0, nr_hardirq_read_safe = 0, nr_hardirq_read_unsafe = 0, sum_forward_deps = 0, factor = 0; list_for_each_entry(class, &all_lock_classes, lock_entry) { if (class->usage_mask == 0) nr_unused++; if (class->usage_mask == LOCKF_USED) nr_uncategorized++; if (class->usage_mask & LOCKF_USED_IN_IRQ) nr_irq_safe++; if (class->usage_mask & LOCKF_ENABLED_IRQS) nr_irq_unsafe++; if (class->usage_mask & LOCKF_USED_IN_SOFTIRQ) nr_softirq_safe++; if (class->usage_mask & LOCKF_ENABLED_SOFTIRQS) nr_softirq_unsafe++; if (class->usage_mask & LOCKF_USED_IN_HARDIRQ) nr_hardirq_safe++; if (class->usage_mask & LOCKF_ENABLED_HARDIRQS) nr_hardirq_unsafe++; if (class->usage_mask & LOCKF_USED_IN_IRQ_READ) nr_irq_read_safe++; if (class->usage_mask & LOCKF_ENABLED_IRQS_READ) nr_irq_read_unsafe++; if (class->usage_mask & LOCKF_USED_IN_SOFTIRQ_READ) nr_softirq_read_safe++; if (class->usage_mask & LOCKF_ENABLED_SOFTIRQS_READ) nr_softirq_read_unsafe++; if (class->usage_mask & LOCKF_USED_IN_HARDIRQ_READ) nr_hardirq_read_safe++; if (class->usage_mask & LOCKF_ENABLED_HARDIRQS_READ) nr_hardirq_read_unsafe++; sum_forward_deps += count_forward_deps(class); } #ifdef CONFIG_DEBUG_LOCKDEP DEBUG_LOCKS_WARN_ON(debug_atomic_read(&nr_unused_locks) != nr_unused); #endif seq_printf(m, " lock-classes: %11lu [max: %lu]\n", nr_lock_classes, MAX_LOCKDEP_KEYS); seq_printf(m, " direct dependencies: %11lu [max: %lu]\n", nr_list_entries, MAX_LOCKDEP_ENTRIES); seq_printf(m, " indirect dependencies: %11lu\n", sum_forward_deps); /* * Total number of dependencies: * * All irq-safe locks may nest inside irq-unsafe locks, * plus all the other known dependencies: */ seq_printf(m, " all direct dependencies: %11lu\n", nr_irq_unsafe * nr_irq_safe + nr_hardirq_unsafe * nr_hardirq_safe + nr_list_entries); /* * Estimated factor between direct and indirect * dependencies: */ if (nr_list_entries) factor = sum_forward_deps / nr_list_entries; seq_printf(m, " dependency chains: %11lu [max: %lu]\n", nr_lock_chains, MAX_LOCKDEP_CHAINS); #ifdef CONFIG_TRACE_IRQFLAGS seq_printf(m, " in-hardirq chains: %11u\n", nr_hardirq_chains); seq_printf(m, " in-softirq chains: %11u\n", nr_softirq_chains); #endif seq_printf(m, " in-process chains: %11u\n", nr_process_chains); seq_printf(m, " stack-trace entries: %11lu [max: %lu]\n", nr_stack_trace_entries, MAX_STACK_TRACE_ENTRIES); seq_printf(m, " combined max dependencies: %11u\n", (nr_hardirq_chains + 1) * (nr_softirq_chains + 1) * (nr_process_chains + 1) ); seq_printf(m, " hardirq-safe locks: %11lu\n", nr_hardirq_safe); seq_printf(m, " hardirq-unsafe locks: %11lu\n", nr_hardirq_unsafe); seq_printf(m, " softirq-safe locks: %11lu\n", nr_softirq_safe); seq_printf(m, " softirq-unsafe locks: %11lu\n", nr_softirq_unsafe); seq_printf(m, " irq-safe locks: %11lu\n", nr_irq_safe); seq_printf(m, " irq-unsafe locks: %11lu\n", nr_irq_unsafe); seq_printf(m, " hardirq-read-safe locks: %11lu\n", nr_hardirq_read_safe); seq_printf(m, " hardirq-read-unsafe locks: %11lu\n", nr_hardirq_read_unsafe); seq_printf(m, " softirq-read-safe locks: %11lu\n", nr_softirq_read_safe); seq_printf(m, " softirq-read-unsafe locks: %11lu\n", nr_softirq_read_unsafe); seq_printf(m, " irq-read-safe locks: %11lu\n", nr_irq_read_safe); seq_printf(m, " irq-read-unsafe locks: %11lu\n", nr_irq_read_unsafe); seq_printf(m, " uncategorized locks: %11lu\n", nr_uncategorized); seq_printf(m, " unused locks: %11lu\n", nr_unused); seq_printf(m, " max locking depth: %11u\n", max_lockdep_depth); seq_printf(m, " max recursion depth: %11u\n", max_recursion_depth); lockdep_stats_debug_show(m); seq_printf(m, " debug_locks: %11u\n", debug_locks); return 0; } static int lockdep_stats_open(struct inode *inode, struct file *file) { return single_open(file, lockdep_stats_show, NULL); } static const struct file_operations proc_lockdep_stats_operations = { .open = lockdep_stats_open, .read = seq_read, .llseek = seq_lseek, .release = seq_release, }; static int __init lockdep_proc_init(void) { struct proc_dir_entry *entry; entry = create_proc_entry("lockdep", S_IRUSR, NULL); if (entry) entry->proc_fops = &proc_lockdep_operations; entry = create_proc_entry("lockdep_stats", S_IRUSR, NULL); if (entry) entry->proc_fops = &proc_lockdep_stats_operations; return 0; } __initcall(lockdep_proc_init);