Message ID | 20180320160258.11381-1-daniel.vetter@ffwll.ch (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Tue, 20 Mar 2018, Daniel Vetter wrote: > Sometimes it's really easy to know which object has gone boom and > where the offending code is, and sometimes it's really hard. One case > we're trying to hunt down is when module unload catches a live debug > object, with a module with lots of them. > > Capture a full stack trace from debug_object_init() and dump it when > there's a problem. > > FIXME: Should we have a separate Kconfig knob for the backtraces, > they're quite expensive? Atm I'm just selecting it for the general > debug object stuff. Just make it boot time enabled. We really want to be able to switch it off. > +#define ODEBUG_STACKDEPTH 32 Do we really need it that deep? I doubt that. > +static noinline depot_stack_handle_t save_stack(struct debug_obj *obj) > +{ > + unsigned long entries[ODEBUG_STACKDEPTH]; > + struct stack_trace trace = { > + .entries = entries, > + .max_entries = ODEBUG_STACKDEPTH, > + .skip = 2 > + }; > + > + save_stack_trace(&trace); > + if (trace.nr_entries != 0 && > + trace.entries[trace.nr_entries-1] == ULONG_MAX) > + trace.nr_entries--; > + > + /* May be called under spinlock, so avoid sleeping */ > + return depot_save_stack(&trace, GFP_NOWAIT); I really don't like the idea of unconditional allocations in that code path. We went great length to make it perform halfways sane with the pool. Though we should actually have per cpu pools to avoid the lock contention, but thats a different problem. I'd rather make the storage a fixed size allocation which is appended to the debug object. Something like this: struct debug_obj_trace { struct stack_trace trace; unsigned long entries[ODEBUG_STACKDEPTH]; }; struct debug_obj { unsigned int astate; void *object; struct debug_obj_descr *descr; struct debug_obj_trace trace[0]; }; void __init debug_objects_mem_init(void) { unsigned long objsize = sizeof (struct debug_obj); if (!debug_objects_enabled) return; if (debug_objects_trace_stack) objsize += sizeof(struct debug_obj_trace); obj_cache = kmem_cache_create("debug_objects_cache", objsize, 0, SLAB_DEBUG_OBJECTS, NULL); __debug_object_init(void *addr, struct debug_obj_descr *descr, int onstack) { .... case ODEBUG_STATE_NONE: case ODEBUG_STATE_INIT: case ODEBUG_STATE_INACTIVE: if (debug_object_trace_stack) { obj->trace[0].trace.entries = obj->trace[0].entries; save_stack_trace(&trace[0].trace); } That means we need to size the static objects which are used during early boot with stack under the assumption that stack tracing is enabled, but that's simple enough. Hmm? Thanks, tglx
On Tue, Mar 20, 2018 at 8:44 PM, Thomas Gleixner <tglx@linutronix.de> wrote: > On Tue, 20 Mar 2018, Daniel Vetter wrote: > >> Sometimes it's really easy to know which object has gone boom and >> where the offending code is, and sometimes it's really hard. One case >> we're trying to hunt down is when module unload catches a live debug >> object, with a module with lots of them. >> >> Capture a full stack trace from debug_object_init() and dump it when >> there's a problem. >> >> FIXME: Should we have a separate Kconfig knob for the backtraces, >> they're quite expensive? Atm I'm just selecting it for the general >> debug object stuff. > > Just make it boot time enabled. We really want to be able to switch it off. > >> +#define ODEBUG_STACKDEPTH 32 > > Do we really need it that deep? I doubt that. Entirely arbitrary, I just needed this to start hunting a rare crash we're seeing in CI and stitched something together. I agree we probably don't need that much. >> +static noinline depot_stack_handle_t save_stack(struct debug_obj *obj) >> +{ >> + unsigned long entries[ODEBUG_STACKDEPTH]; >> + struct stack_trace trace = { >> + .entries = entries, >> + .max_entries = ODEBUG_STACKDEPTH, >> + .skip = 2 >> + }; >> + >> + save_stack_trace(&trace); >> + if (trace.nr_entries != 0 && >> + trace.entries[trace.nr_entries-1] == ULONG_MAX) >> + trace.nr_entries--; >> + >> + /* May be called under spinlock, so avoid sleeping */ >> + return depot_save_stack(&trace, GFP_NOWAIT); > > I really don't like the idea of unconditional allocations in that code > path. We went great length to make it perform halfways sane with the > pool. Though we should actually have per cpu pools to avoid the lock > contention, but thats a different problem. > > I'd rather make the storage a fixed size allocation which is appended > to the debug object. Something like this: > > struct debug_obj_trace { > struct stack_trace trace; > unsigned long entries[ODEBUG_STACKDEPTH]; > }; > > struct debug_obj { > unsigned int astate; > void *object; > struct debug_obj_descr *descr; > struct debug_obj_trace trace[0]; > }; > > void __init debug_objects_mem_init(void) > { > unsigned long objsize = sizeof (struct debug_obj); > > if (!debug_objects_enabled) > return; > > if (debug_objects_trace_stack) > objsize += sizeof(struct debug_obj_trace); > > obj_cache = kmem_cache_create("debug_objects_cache", objsize, 0, > SLAB_DEBUG_OBJECTS, NULL); > > __debug_object_init(void *addr, struct debug_obj_descr *descr, int onstack) > { > .... > case ODEBUG_STATE_NONE: > case ODEBUG_STATE_INIT: > case ODEBUG_STATE_INACTIVE: > if (debug_object_trace_stack) { > obj->trace[0].trace.entries = obj->trace[0].entries; > save_stack_trace(&trace[0].trace); > } > > That means we need to size the static objects which are used during early > boot with stack under the assumption that stack tracing is enabled, but > that's simple enough. > > Hmm? Yeah looks reasonable, and gives us an easy Kconfig/boot-time option to enable/disable it. I'm a bit swamped, but I'll try to respin. Thanks very much for the look and suggesting a cleaner integration approach - the allocations and recursion into the stack depot really did not result in clean code. Thanks, Daniel > Thanks, > > tglx
diff --git a/include/linux/debugobjects.h b/include/linux/debugobjects.h index afc416e5dcab..d3a6ca1a7756 100644 --- a/include/linux/debugobjects.h +++ b/include/linux/debugobjects.h @@ -4,6 +4,7 @@ #include <linux/list.h> #include <linux/spinlock.h> +#include <linux/stackdepot.h> enum debug_obj_state { ODEBUG_STATE_NONE, @@ -31,6 +32,7 @@ struct debug_obj { unsigned int astate; void *object; struct debug_obj_descr *descr; + depot_stack_handle_t stack; }; /** diff --git a/lib/Kconfig.debug b/lib/Kconfig.debug index 64155e310a9f..894dd792e771 100644 --- a/lib/Kconfig.debug +++ b/lib/Kconfig.debug @@ -442,6 +442,7 @@ source mm/Kconfig.debug config DEBUG_OBJECTS bool "Debug object operations" depends on DEBUG_KERNEL + select STACKDEPOT help If you say Y here, additional code will be inserted into the kernel to track the life time of various objects and validate diff --git a/lib/debugobjects.c b/lib/debugobjects.c index 2f5349c6e81a..2acad7150bee 100644 --- a/lib/debugobjects.c +++ b/lib/debugobjects.c @@ -19,6 +19,7 @@ #include <linux/slab.h> #include <linux/hash.h> #include <linux/kmemleak.h> +#include <linux/stacktrace.h> #define ODEBUG_HASH_BITS 14 #define ODEBUG_HASH_SIZE (1 << ODEBUG_HASH_BITS) @@ -30,6 +31,8 @@ #define ODEBUG_CHUNK_SIZE (1 << ODEBUG_CHUNK_SHIFT) #define ODEBUG_CHUNK_MASK (~(ODEBUG_CHUNK_SIZE - 1)) +#define ODEBUG_STACKDEPTH 32 + struct debug_bucket { struct hlist_head list; raw_spinlock_t lock; @@ -280,15 +283,24 @@ static void debug_print_object(struct debug_obj *obj, char *msg) { struct debug_obj_descr *descr = obj->descr; static int limit; + unsigned long entries[ODEBUG_STACKDEPTH]; + struct stack_trace trace = { + .entries = entries, + .max_entries = ODEBUG_STACKDEPTH + }; + if (limit < 5 && descr != descr_test) { void *hint = descr->debug_hint ? descr->debug_hint(obj->object) : NULL; limit++; + depot_fetch_stack(obj->stack, &trace); WARN(1, KERN_ERR "ODEBUG: %s %s (active state %u) " "object type: %s hint: %pS\n", msg, obj_states[obj->state], obj->astate, descr->name, hint); + pr_err("ODEBUG: debug object originally initialized at:\n"); + print_stack_trace(&trace, 2); } debug_objects_warnings++; } @@ -328,6 +340,24 @@ static void debug_object_is_on_stack(void *addr, int onstack) WARN_ON(1); } +static noinline depot_stack_handle_t save_stack(struct debug_obj *obj) +{ + unsigned long entries[ODEBUG_STACKDEPTH]; + struct stack_trace trace = { + .entries = entries, + .max_entries = ODEBUG_STACKDEPTH, + .skip = 2 + }; + + save_stack_trace(&trace); + if (trace.nr_entries != 0 && + trace.entries[trace.nr_entries-1] == ULONG_MAX) + trace.nr_entries--; + + /* May be called under spinlock, so avoid sleeping */ + return depot_save_stack(&trace, GFP_NOWAIT); +} + static void __debug_object_init(void *addr, struct debug_obj_descr *descr, int onstack) { @@ -344,14 +374,29 @@ __debug_object_init(void *addr, struct debug_obj_descr *descr, int onstack) obj = lookup_object(addr, db); if (!obj) { - obj = alloc_object(addr, db, descr); + depot_stack_handle_t stack; + + /* + * must drop locks while storing the stack trace to avoid + * recursive deadlock through depot_save_stack + * allocating/freeing memory. + */ + raw_spin_unlock_irqrestore(&db->lock, flags); + stack = save_stack(obj); + raw_spin_lock_irqsave(&db->lock, flags); + + obj = lookup_object(addr, db); if (!obj) { - debug_objects_enabled = 0; - raw_spin_unlock_irqrestore(&db->lock, flags); - debug_objects_oom(); - return; + obj = alloc_object(addr, db, descr); + if (!obj) { + debug_objects_enabled = 0; + raw_spin_unlock_irqrestore(&db->lock, flags); + debug_objects_oom(); + return; + } + debug_object_is_on_stack(addr, onstack); + obj->stack = stack; } - debug_object_is_on_stack(addr, onstack); } switch (obj->state) {
Sometimes it's really easy to know which object has gone boom and where the offending code is, and sometimes it's really hard. One case we're trying to hunt down is when module unload catches a live debug object, with a module with lots of them. Capture a full stack trace from debug_object_init() and dump it when there's a problem. FIXME: Should we have a separate Kconfig knob for the backtraces, they're quite expensive? Atm I'm just selecting it for the general debug object stuff. v2: Drop the locks for gathering&storing the backtrace. This is required because depot_save_stack can call free_pages (to drop it's preallocation), which can call debug_check_no_obj_freed, which will recurse into the db->lock spinlocks. Cc: Philippe Ombredanne <pombredanne@nexb.com> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Kate Stewart <kstewart@linuxfoundation.org> Cc: Daniel Vetter <daniel.vetter@ffwll.ch> Cc: Waiman Long <longman@redhat.com> Signed-off-by: Daniel Vetter <daniel.vetter@intel.com> --- include/linux/debugobjects.h | 2 ++ lib/Kconfig.debug | 1 + lib/debugobjects.c | 57 +++++++++++++++++++++++++++++++++++++++----- 3 files changed, 54 insertions(+), 6 deletions(-)