Merge branch 'sh/dwarf-unwinder'
Conflicts: arch/sh/kernel/cpu/sh3/entry.S
This commit is contained in:
commit
fa9d3b4da5
10 changed files with 265 additions and 130 deletions
|
@ -2,6 +2,7 @@
|
||||||
#define __ASM_SH_BUG_H
|
#define __ASM_SH_BUG_H
|
||||||
|
|
||||||
#define TRAPA_BUG_OPCODE 0xc33e /* trapa #0x3e */
|
#define TRAPA_BUG_OPCODE 0xc33e /* trapa #0x3e */
|
||||||
|
#define BUGFLAG_UNWINDER (1 << 1)
|
||||||
|
|
||||||
#ifdef CONFIG_GENERIC_BUG
|
#ifdef CONFIG_GENERIC_BUG
|
||||||
#define HAVE_ARCH_BUG
|
#define HAVE_ARCH_BUG
|
||||||
|
@ -72,6 +73,36 @@ do { \
|
||||||
unlikely(__ret_warn_on); \
|
unlikely(__ret_warn_on); \
|
||||||
})
|
})
|
||||||
|
|
||||||
|
#define UNWINDER_BUG() \
|
||||||
|
do { \
|
||||||
|
__asm__ __volatile__ ( \
|
||||||
|
"1:\t.short %O0\n" \
|
||||||
|
_EMIT_BUG_ENTRY \
|
||||||
|
: \
|
||||||
|
: "n" (TRAPA_BUG_OPCODE), \
|
||||||
|
"i" (__FILE__), \
|
||||||
|
"i" (__LINE__), \
|
||||||
|
"i" (BUGFLAG_UNWINDER), \
|
||||||
|
"i" (sizeof(struct bug_entry))); \
|
||||||
|
} while (0)
|
||||||
|
|
||||||
|
#define UNWINDER_BUG_ON(x) ({ \
|
||||||
|
int __ret_unwinder_on = !!(x); \
|
||||||
|
if (__builtin_constant_p(__ret_unwinder_on)) { \
|
||||||
|
if (__ret_unwinder_on) \
|
||||||
|
UNWINDER_BUG(); \
|
||||||
|
} else { \
|
||||||
|
if (unlikely(__ret_unwinder_on)) \
|
||||||
|
UNWINDER_BUG(); \
|
||||||
|
} \
|
||||||
|
unlikely(__ret_unwinder_on); \
|
||||||
|
})
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
#define UNWINDER_BUG BUG
|
||||||
|
#define UNWINDER_BUG_ON BUG_ON
|
||||||
|
|
||||||
#endif /* CONFIG_GENERIC_BUG */
|
#endif /* CONFIG_GENERIC_BUG */
|
||||||
|
|
||||||
#include <asm-generic/bug.h>
|
#include <asm-generic/bug.h>
|
||||||
|
|
|
@ -265,10 +265,7 @@ struct dwarf_frame {
|
||||||
|
|
||||||
unsigned long pc;
|
unsigned long pc;
|
||||||
|
|
||||||
struct dwarf_reg *regs;
|
struct list_head reg_list;
|
||||||
unsigned int num_regs; /* how many regs are allocated? */
|
|
||||||
|
|
||||||
unsigned int depth; /* what level are we in the callstack? */
|
|
||||||
|
|
||||||
unsigned long cfa;
|
unsigned long cfa;
|
||||||
|
|
||||||
|
@ -292,20 +289,15 @@ struct dwarf_frame {
|
||||||
* @flags: Describes how to calculate the value of this register
|
* @flags: Describes how to calculate the value of this register
|
||||||
*/
|
*/
|
||||||
struct dwarf_reg {
|
struct dwarf_reg {
|
||||||
|
struct list_head link;
|
||||||
|
|
||||||
|
unsigned int number;
|
||||||
|
|
||||||
unsigned long addr;
|
unsigned long addr;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
#define DWARF_REG_OFFSET (1 << 0)
|
#define DWARF_REG_OFFSET (1 << 0)
|
||||||
};
|
#define DWARF_VAL_OFFSET (1 << 1)
|
||||||
|
#define DWARF_UNDEFINED (1 << 2)
|
||||||
/**
|
|
||||||
* dwarf_stack - a DWARF stack contains a collection of DWARF frames
|
|
||||||
* @depth: the number of frames in the stack
|
|
||||||
* @level: an array of DWARF frames, indexed by stack level
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
struct dwarf_stack {
|
|
||||||
unsigned int depth;
|
|
||||||
struct dwarf_frame **level;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -370,17 +362,16 @@ static inline unsigned int DW_CFA_operand(unsigned long insn)
|
||||||
#define DW_EXT_HI 0xffffffff
|
#define DW_EXT_HI 0xffffffff
|
||||||
#define DW_EXT_DWARF64 DW_EXT_HI
|
#define DW_EXT_DWARF64 DW_EXT_HI
|
||||||
|
|
||||||
extern void dwarf_unwinder_init(void);
|
|
||||||
|
|
||||||
extern struct dwarf_frame *dwarf_unwind_stack(unsigned long,
|
extern struct dwarf_frame *dwarf_unwind_stack(unsigned long,
|
||||||
struct dwarf_frame *);
|
struct dwarf_frame *);
|
||||||
#endif /* __ASSEMBLY__ */
|
#endif /* !__ASSEMBLY__ */
|
||||||
|
|
||||||
#define CFI_STARTPROC .cfi_startproc
|
#define CFI_STARTPROC .cfi_startproc
|
||||||
#define CFI_ENDPROC .cfi_endproc
|
#define CFI_ENDPROC .cfi_endproc
|
||||||
#define CFI_DEF_CFA .cfi_def_cfa
|
#define CFI_DEF_CFA .cfi_def_cfa
|
||||||
#define CFI_REGISTER .cfi_register
|
#define CFI_REGISTER .cfi_register
|
||||||
#define CFI_REL_OFFSET .cfi_rel_offset
|
#define CFI_REL_OFFSET .cfi_rel_offset
|
||||||
|
#define CFI_UNDEFINED .cfi_undefined
|
||||||
|
|
||||||
#else
|
#else
|
||||||
|
|
||||||
|
@ -394,6 +385,7 @@ extern struct dwarf_frame *dwarf_unwind_stack(unsigned long,
|
||||||
#define CFI_DEF_CFA CFI_IGNORE
|
#define CFI_DEF_CFA CFI_IGNORE
|
||||||
#define CFI_REGISTER CFI_IGNORE
|
#define CFI_REGISTER CFI_IGNORE
|
||||||
#define CFI_REL_OFFSET CFI_IGNORE
|
#define CFI_REL_OFFSET CFI_IGNORE
|
||||||
|
#define CFI_UNDEFINED CFI_IGNORE
|
||||||
|
|
||||||
#ifndef __ASSEMBLY__
|
#ifndef __ASSEMBLY__
|
||||||
static inline void dwarf_unwinder_init(void)
|
static inline void dwarf_unwinder_init(void)
|
||||||
|
|
|
@ -181,6 +181,11 @@ BUILD_TRAP_HANDLER(breakpoint);
|
||||||
BUILD_TRAP_HANDLER(singlestep);
|
BUILD_TRAP_HANDLER(singlestep);
|
||||||
BUILD_TRAP_HANDLER(fpu_error);
|
BUILD_TRAP_HANDLER(fpu_error);
|
||||||
BUILD_TRAP_HANDLER(fpu_state_restore);
|
BUILD_TRAP_HANDLER(fpu_state_restore);
|
||||||
|
BUILD_TRAP_HANDLER(unwinder);
|
||||||
|
|
||||||
|
#ifdef CONFIG_BUG
|
||||||
|
extern void handle_BUG(struct pt_regs *);
|
||||||
|
#endif
|
||||||
|
|
||||||
#define arch_align_stack(x) (x)
|
#define arch_align_stack(x) (x)
|
||||||
|
|
||||||
|
|
|
@ -22,4 +22,10 @@ extern void stack_reader_dump(struct task_struct *, struct pt_regs *,
|
||||||
unsigned long *, const struct stacktrace_ops *,
|
unsigned long *, const struct stacktrace_ops *,
|
||||||
void *);
|
void *);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Used by fault handling code to signal to the unwinder code that it
|
||||||
|
* should switch to a different unwinder.
|
||||||
|
*/
|
||||||
|
extern int unwinder_faulted;
|
||||||
|
|
||||||
#endif /* _LINUX_UNWINDER_H */
|
#endif /* _LINUX_UNWINDER_H */
|
||||||
|
|
|
@ -508,6 +508,8 @@ ENTRY(handle_interrupt)
|
||||||
bsr save_regs ! needs original pr value in k3
|
bsr save_regs ! needs original pr value in k3
|
||||||
mov #-1, k2 ! default vector kept in k2
|
mov #-1, k2 ! default vector kept in k2
|
||||||
|
|
||||||
|
setup_frame_reg
|
||||||
|
|
||||||
stc sr, r0 ! get status register
|
stc sr, r0 ! get status register
|
||||||
shlr2 r0
|
shlr2 r0
|
||||||
and #0x3c, r0
|
and #0x3c, r0
|
||||||
|
|
|
@ -11,12 +11,14 @@
|
||||||
*
|
*
|
||||||
* TODO:
|
* TODO:
|
||||||
* - DWARF64 doesn't work.
|
* - DWARF64 doesn't work.
|
||||||
|
* - Registers with DWARF_VAL_OFFSET rules aren't handled properly.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/* #define DEBUG */
|
/* #define DEBUG */
|
||||||
#include <linux/kernel.h>
|
#include <linux/kernel.h>
|
||||||
#include <linux/io.h>
|
#include <linux/io.h>
|
||||||
#include <linux/list.h>
|
#include <linux/list.h>
|
||||||
|
#include <linux/mempool.h>
|
||||||
#include <linux/mm.h>
|
#include <linux/mm.h>
|
||||||
#include <asm/dwarf.h>
|
#include <asm/dwarf.h>
|
||||||
#include <asm/unwinder.h>
|
#include <asm/unwinder.h>
|
||||||
|
@ -25,55 +27,89 @@
|
||||||
#include <asm/dwarf.h>
|
#include <asm/dwarf.h>
|
||||||
#include <asm/stacktrace.h>
|
#include <asm/stacktrace.h>
|
||||||
|
|
||||||
|
/* Reserve enough memory for two stack frames */
|
||||||
|
#define DWARF_FRAME_MIN_REQ 2
|
||||||
|
/* ... with 4 registers per frame. */
|
||||||
|
#define DWARF_REG_MIN_REQ (DWARF_FRAME_MIN_REQ * 4)
|
||||||
|
|
||||||
|
static struct kmem_cache *dwarf_frame_cachep;
|
||||||
|
static mempool_t *dwarf_frame_pool;
|
||||||
|
|
||||||
|
static struct kmem_cache *dwarf_reg_cachep;
|
||||||
|
static mempool_t *dwarf_reg_pool;
|
||||||
|
|
||||||
static LIST_HEAD(dwarf_cie_list);
|
static LIST_HEAD(dwarf_cie_list);
|
||||||
DEFINE_SPINLOCK(dwarf_cie_lock);
|
static DEFINE_SPINLOCK(dwarf_cie_lock);
|
||||||
|
|
||||||
static LIST_HEAD(dwarf_fde_list);
|
static LIST_HEAD(dwarf_fde_list);
|
||||||
DEFINE_SPINLOCK(dwarf_fde_lock);
|
static DEFINE_SPINLOCK(dwarf_fde_lock);
|
||||||
|
|
||||||
static struct dwarf_cie *cached_cie;
|
static struct dwarf_cie *cached_cie;
|
||||||
|
|
||||||
/*
|
/**
|
||||||
* Figure out whether we need to allocate some dwarf registers. If dwarf
|
* dwarf_frame_alloc_reg - allocate memory for a DWARF register
|
||||||
* registers have already been allocated then we may need to realloc
|
* @frame: the DWARF frame whose list of registers we insert on
|
||||||
* them. "reg" is a register number that we need to be able to access
|
* @reg_num: the register number
|
||||||
* after this call.
|
|
||||||
*
|
*
|
||||||
* Register numbers start at zero, therefore we need to allocate space
|
* Allocate space for, and initialise, a dwarf reg from
|
||||||
* for "reg" + 1 registers.
|
* dwarf_reg_pool and insert it onto the (unsorted) linked-list of
|
||||||
|
* dwarf registers for @frame.
|
||||||
|
*
|
||||||
|
* Return the initialised DWARF reg.
|
||||||
*/
|
*/
|
||||||
static void dwarf_frame_alloc_regs(struct dwarf_frame *frame,
|
static struct dwarf_reg *dwarf_frame_alloc_reg(struct dwarf_frame *frame,
|
||||||
unsigned int reg)
|
unsigned int reg_num)
|
||||||
{
|
{
|
||||||
struct dwarf_reg *regs;
|
struct dwarf_reg *reg;
|
||||||
unsigned int num_regs = reg + 1;
|
|
||||||
size_t new_size;
|
|
||||||
size_t old_size;
|
|
||||||
|
|
||||||
new_size = num_regs * sizeof(*regs);
|
reg = mempool_alloc(dwarf_reg_pool, GFP_ATOMIC);
|
||||||
old_size = frame->num_regs * sizeof(*regs);
|
if (!reg) {
|
||||||
|
printk(KERN_WARNING "Unable to allocate a DWARF register\n");
|
||||||
/* Fast path: don't allocate any regs if we've already got enough. */
|
|
||||||
if (frame->num_regs >= num_regs)
|
|
||||||
return;
|
|
||||||
|
|
||||||
regs = kzalloc(new_size, GFP_ATOMIC);
|
|
||||||
if (!regs) {
|
|
||||||
printk(KERN_WARNING "Unable to allocate DWARF registers\n");
|
|
||||||
/*
|
/*
|
||||||
* Let's just bomb hard here, we have no way to
|
* Let's just bomb hard here, we have no way to
|
||||||
* gracefully recover.
|
* gracefully recover.
|
||||||
*/
|
*/
|
||||||
BUG();
|
UNWINDER_BUG();
|
||||||
}
|
}
|
||||||
|
|
||||||
if (frame->regs) {
|
reg->number = reg_num;
|
||||||
memcpy(regs, frame->regs, old_size);
|
reg->addr = 0;
|
||||||
kfree(frame->regs);
|
reg->flags = 0;
|
||||||
|
|
||||||
|
list_add(®->link, &frame->reg_list);
|
||||||
|
|
||||||
|
return reg;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void dwarf_frame_free_regs(struct dwarf_frame *frame)
|
||||||
|
{
|
||||||
|
struct dwarf_reg *reg, *n;
|
||||||
|
|
||||||
|
list_for_each_entry_safe(reg, n, &frame->reg_list, link) {
|
||||||
|
list_del(®->link);
|
||||||
|
mempool_free(reg, dwarf_reg_pool);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* dwarf_frame_reg - return a DWARF register
|
||||||
|
* @frame: the DWARF frame to search in for @reg_num
|
||||||
|
* @reg_num: the register number to search for
|
||||||
|
*
|
||||||
|
* Lookup and return the dwarf reg @reg_num for this frame. Return
|
||||||
|
* NULL if @reg_num is an register invalid number.
|
||||||
|
*/
|
||||||
|
static struct dwarf_reg *dwarf_frame_reg(struct dwarf_frame *frame,
|
||||||
|
unsigned int reg_num)
|
||||||
|
{
|
||||||
|
struct dwarf_reg *reg;
|
||||||
|
|
||||||
|
list_for_each_entry(reg, &frame->reg_list, link) {
|
||||||
|
if (reg->number == reg_num)
|
||||||
|
return reg;
|
||||||
}
|
}
|
||||||
|
|
||||||
frame->regs = regs;
|
return NULL;
|
||||||
frame->num_regs = num_regs;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -196,7 +232,7 @@ static int dwarf_read_encoded_value(char *addr, unsigned long *val,
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
pr_debug("encoding=0x%x\n", (encoding & 0x70));
|
pr_debug("encoding=0x%x\n", (encoding & 0x70));
|
||||||
BUG();
|
UNWINDER_BUG();
|
||||||
}
|
}
|
||||||
|
|
||||||
if ((encoding & 0x07) == 0x00)
|
if ((encoding & 0x07) == 0x00)
|
||||||
|
@ -211,7 +247,7 @@ static int dwarf_read_encoded_value(char *addr, unsigned long *val,
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
pr_debug("encoding=0x%x\n", encoding);
|
pr_debug("encoding=0x%x\n", encoding);
|
||||||
BUG();
|
UNWINDER_BUG();
|
||||||
}
|
}
|
||||||
|
|
||||||
return count;
|
return count;
|
||||||
|
@ -264,7 +300,7 @@ static inline int dwarf_entry_len(char *addr, unsigned long *len)
|
||||||
*/
|
*/
|
||||||
static struct dwarf_cie *dwarf_lookup_cie(unsigned long cie_ptr)
|
static struct dwarf_cie *dwarf_lookup_cie(unsigned long cie_ptr)
|
||||||
{
|
{
|
||||||
struct dwarf_cie *cie, *n;
|
struct dwarf_cie *cie;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
spin_lock_irqsave(&dwarf_cie_lock, flags);
|
spin_lock_irqsave(&dwarf_cie_lock, flags);
|
||||||
|
@ -278,7 +314,7 @@ static struct dwarf_cie *dwarf_lookup_cie(unsigned long cie_ptr)
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
list_for_each_entry_safe(cie, n, &dwarf_cie_list, link) {
|
list_for_each_entry(cie, &dwarf_cie_list, link) {
|
||||||
if (cie->cie_pointer == cie_ptr) {
|
if (cie->cie_pointer == cie_ptr) {
|
||||||
cached_cie = cie;
|
cached_cie = cie;
|
||||||
break;
|
break;
|
||||||
|
@ -299,11 +335,12 @@ static struct dwarf_cie *dwarf_lookup_cie(unsigned long cie_ptr)
|
||||||
*/
|
*/
|
||||||
struct dwarf_fde *dwarf_lookup_fde(unsigned long pc)
|
struct dwarf_fde *dwarf_lookup_fde(unsigned long pc)
|
||||||
{
|
{
|
||||||
|
struct dwarf_fde *fde;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct dwarf_fde *fde, *n;
|
|
||||||
|
|
||||||
spin_lock_irqsave(&dwarf_fde_lock, flags);
|
spin_lock_irqsave(&dwarf_fde_lock, flags);
|
||||||
list_for_each_entry_safe(fde, n, &dwarf_fde_list, link) {
|
|
||||||
|
list_for_each_entry(fde, &dwarf_fde_list, link) {
|
||||||
unsigned long start, end;
|
unsigned long start, end;
|
||||||
|
|
||||||
start = fde->initial_location;
|
start = fde->initial_location;
|
||||||
|
@ -346,6 +383,7 @@ static int dwarf_cfa_execute_insns(unsigned char *insn_start,
|
||||||
unsigned char insn;
|
unsigned char insn;
|
||||||
unsigned char *current_insn;
|
unsigned char *current_insn;
|
||||||
unsigned int count, delta, reg, expr_len, offset;
|
unsigned int count, delta, reg, expr_len, offset;
|
||||||
|
struct dwarf_reg *regp;
|
||||||
|
|
||||||
current_insn = insn_start;
|
current_insn = insn_start;
|
||||||
|
|
||||||
|
@ -368,9 +406,9 @@ static int dwarf_cfa_execute_insns(unsigned char *insn_start,
|
||||||
count = dwarf_read_uleb128(current_insn, &offset);
|
count = dwarf_read_uleb128(current_insn, &offset);
|
||||||
current_insn += count;
|
current_insn += count;
|
||||||
offset *= cie->data_alignment_factor;
|
offset *= cie->data_alignment_factor;
|
||||||
dwarf_frame_alloc_regs(frame, reg);
|
regp = dwarf_frame_alloc_reg(frame, reg);
|
||||||
frame->regs[reg].addr = offset;
|
regp->addr = offset;
|
||||||
frame->regs[reg].flags |= DWARF_REG_OFFSET;
|
regp->flags |= DWARF_REG_OFFSET;
|
||||||
continue;
|
continue;
|
||||||
/* NOTREACHED */
|
/* NOTREACHED */
|
||||||
case DW_CFA_restore:
|
case DW_CFA_restore:
|
||||||
|
@ -414,6 +452,8 @@ static int dwarf_cfa_execute_insns(unsigned char *insn_start,
|
||||||
case DW_CFA_undefined:
|
case DW_CFA_undefined:
|
||||||
count = dwarf_read_uleb128(current_insn, ®);
|
count = dwarf_read_uleb128(current_insn, ®);
|
||||||
current_insn += count;
|
current_insn += count;
|
||||||
|
regp = dwarf_frame_alloc_reg(frame, reg);
|
||||||
|
regp->flags |= DWARF_UNDEFINED;
|
||||||
break;
|
break;
|
||||||
case DW_CFA_def_cfa:
|
case DW_CFA_def_cfa:
|
||||||
count = dwarf_read_uleb128(current_insn,
|
count = dwarf_read_uleb128(current_insn,
|
||||||
|
@ -452,17 +492,18 @@ static int dwarf_cfa_execute_insns(unsigned char *insn_start,
|
||||||
count = dwarf_read_leb128(current_insn, &offset);
|
count = dwarf_read_leb128(current_insn, &offset);
|
||||||
current_insn += count;
|
current_insn += count;
|
||||||
offset *= cie->data_alignment_factor;
|
offset *= cie->data_alignment_factor;
|
||||||
dwarf_frame_alloc_regs(frame, reg);
|
regp = dwarf_frame_alloc_reg(frame, reg);
|
||||||
frame->regs[reg].flags |= DWARF_REG_OFFSET;
|
regp->flags |= DWARF_REG_OFFSET;
|
||||||
frame->regs[reg].addr = offset;
|
regp->addr = offset;
|
||||||
break;
|
break;
|
||||||
case DW_CFA_val_offset:
|
case DW_CFA_val_offset:
|
||||||
count = dwarf_read_uleb128(current_insn, ®);
|
count = dwarf_read_uleb128(current_insn, ®);
|
||||||
current_insn += count;
|
current_insn += count;
|
||||||
count = dwarf_read_leb128(current_insn, &offset);
|
count = dwarf_read_leb128(current_insn, &offset);
|
||||||
offset *= cie->data_alignment_factor;
|
offset *= cie->data_alignment_factor;
|
||||||
frame->regs[reg].flags |= DWARF_REG_OFFSET;
|
regp = dwarf_frame_alloc_reg(frame, reg);
|
||||||
frame->regs[reg].addr = offset;
|
regp->flags |= DWARF_VAL_OFFSET;
|
||||||
|
regp->addr = offset;
|
||||||
break;
|
break;
|
||||||
case DW_CFA_GNU_args_size:
|
case DW_CFA_GNU_args_size:
|
||||||
count = dwarf_read_uleb128(current_insn, &offset);
|
count = dwarf_read_uleb128(current_insn, &offset);
|
||||||
|
@ -473,12 +514,14 @@ static int dwarf_cfa_execute_insns(unsigned char *insn_start,
|
||||||
current_insn += count;
|
current_insn += count;
|
||||||
count = dwarf_read_uleb128(current_insn, &offset);
|
count = dwarf_read_uleb128(current_insn, &offset);
|
||||||
offset *= cie->data_alignment_factor;
|
offset *= cie->data_alignment_factor;
|
||||||
dwarf_frame_alloc_regs(frame, reg);
|
|
||||||
frame->regs[reg].flags |= DWARF_REG_OFFSET;
|
regp = dwarf_frame_alloc_reg(frame, reg);
|
||||||
frame->regs[reg].addr = -offset;
|
regp->flags |= DWARF_REG_OFFSET;
|
||||||
|
regp->addr = -offset;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
pr_debug("unhandled DWARF instruction 0x%x\n", insn);
|
pr_debug("unhandled DWARF instruction 0x%x\n", insn);
|
||||||
|
UNWINDER_BUG();
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -495,14 +538,14 @@ static int dwarf_cfa_execute_insns(unsigned char *insn_start,
|
||||||
* on the callstack. Each of the lower (older) stack frames are
|
* on the callstack. Each of the lower (older) stack frames are
|
||||||
* linked via the "prev" member.
|
* linked via the "prev" member.
|
||||||
*/
|
*/
|
||||||
struct dwarf_frame *dwarf_unwind_stack(unsigned long pc,
|
struct dwarf_frame * dwarf_unwind_stack(unsigned long pc,
|
||||||
struct dwarf_frame *prev)
|
struct dwarf_frame *prev)
|
||||||
{
|
{
|
||||||
struct dwarf_frame *frame;
|
struct dwarf_frame *frame;
|
||||||
struct dwarf_cie *cie;
|
struct dwarf_cie *cie;
|
||||||
struct dwarf_fde *fde;
|
struct dwarf_fde *fde;
|
||||||
|
struct dwarf_reg *reg;
|
||||||
unsigned long addr;
|
unsigned long addr;
|
||||||
int i, offset;
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If this is the first invocation of this recursive function we
|
* If this is the first invocation of this recursive function we
|
||||||
|
@ -515,11 +558,16 @@ struct dwarf_frame *dwarf_unwind_stack(unsigned long pc,
|
||||||
if (!pc && !prev)
|
if (!pc && !prev)
|
||||||
pc = (unsigned long)current_text_addr();
|
pc = (unsigned long)current_text_addr();
|
||||||
|
|
||||||
frame = kzalloc(sizeof(*frame), GFP_ATOMIC);
|
frame = mempool_alloc(dwarf_frame_pool, GFP_ATOMIC);
|
||||||
if (!frame)
|
if (!frame) {
|
||||||
return NULL;
|
printk(KERN_ERR "Unable to allocate a dwarf frame\n");
|
||||||
|
UNWINDER_BUG();
|
||||||
|
}
|
||||||
|
|
||||||
|
INIT_LIST_HEAD(&frame->reg_list);
|
||||||
|
frame->flags = 0;
|
||||||
frame->prev = prev;
|
frame->prev = prev;
|
||||||
|
frame->return_addr = 0;
|
||||||
|
|
||||||
fde = dwarf_lookup_fde(pc);
|
fde = dwarf_lookup_fde(pc);
|
||||||
if (!fde) {
|
if (!fde) {
|
||||||
|
@ -539,7 +587,7 @@ struct dwarf_frame *dwarf_unwind_stack(unsigned long pc,
|
||||||
* case above, which sucks because we could print a
|
* case above, which sucks because we could print a
|
||||||
* warning here.
|
* warning here.
|
||||||
*/
|
*/
|
||||||
return NULL;
|
goto bail;
|
||||||
}
|
}
|
||||||
|
|
||||||
cie = dwarf_lookup_cie(fde->cie_pointer);
|
cie = dwarf_lookup_cie(fde->cie_pointer);
|
||||||
|
@ -559,10 +607,11 @@ struct dwarf_frame *dwarf_unwind_stack(unsigned long pc,
|
||||||
switch (frame->flags) {
|
switch (frame->flags) {
|
||||||
case DWARF_FRAME_CFA_REG_OFFSET:
|
case DWARF_FRAME_CFA_REG_OFFSET:
|
||||||
if (prev) {
|
if (prev) {
|
||||||
BUG_ON(!prev->regs[frame->cfa_register].flags);
|
reg = dwarf_frame_reg(prev, frame->cfa_register);
|
||||||
|
UNWINDER_BUG_ON(!reg);
|
||||||
|
UNWINDER_BUG_ON(reg->flags != DWARF_REG_OFFSET);
|
||||||
|
|
||||||
addr = prev->cfa;
|
addr = prev->cfa + reg->addr;
|
||||||
addr += prev->regs[frame->cfa_register].addr;
|
|
||||||
frame->cfa = __raw_readl(addr);
|
frame->cfa = __raw_readl(addr);
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
|
@ -579,27 +628,30 @@ struct dwarf_frame *dwarf_unwind_stack(unsigned long pc,
|
||||||
frame->cfa += frame->cfa_offset;
|
frame->cfa += frame->cfa_offset;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
BUG();
|
UNWINDER_BUG();
|
||||||
}
|
}
|
||||||
|
|
||||||
/* If we haven't seen the return address reg, we're screwed. */
|
reg = dwarf_frame_reg(frame, DWARF_ARCH_RA_REG);
|
||||||
BUG_ON(!frame->regs[DWARF_ARCH_RA_REG].flags);
|
|
||||||
|
|
||||||
for (i = 0; i <= frame->num_regs; i++) {
|
/*
|
||||||
struct dwarf_reg *reg = &frame->regs[i];
|
* If we haven't seen the return address register or the return
|
||||||
|
* address column is undefined then we must assume that this is
|
||||||
|
* the end of the callstack.
|
||||||
|
*/
|
||||||
|
if (!reg || reg->flags == DWARF_UNDEFINED)
|
||||||
|
goto bail;
|
||||||
|
|
||||||
if (!reg->flags)
|
UNWINDER_BUG_ON(reg->flags != DWARF_REG_OFFSET);
|
||||||
continue;
|
|
||||||
|
|
||||||
offset = reg->addr;
|
addr = frame->cfa + reg->addr;
|
||||||
offset += frame->cfa;
|
|
||||||
}
|
|
||||||
|
|
||||||
addr = frame->cfa + frame->regs[DWARF_ARCH_RA_REG].addr;
|
|
||||||
frame->return_addr = __raw_readl(addr);
|
frame->return_addr = __raw_readl(addr);
|
||||||
|
|
||||||
frame->next = dwarf_unwind_stack(frame->return_addr, frame);
|
|
||||||
return frame;
|
return frame;
|
||||||
|
|
||||||
|
bail:
|
||||||
|
dwarf_frame_free_regs(frame);
|
||||||
|
mempool_free(frame, dwarf_frame_pool);
|
||||||
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
|
static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
|
||||||
|
@ -624,7 +676,7 @@ static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
|
||||||
cie->cie_pointer = (unsigned long)entry;
|
cie->cie_pointer = (unsigned long)entry;
|
||||||
|
|
||||||
cie->version = *(char *)p++;
|
cie->version = *(char *)p++;
|
||||||
BUG_ON(cie->version != 1);
|
UNWINDER_BUG_ON(cie->version != 1);
|
||||||
|
|
||||||
cie->augmentation = p;
|
cie->augmentation = p;
|
||||||
p += strlen(cie->augmentation) + 1;
|
p += strlen(cie->augmentation) + 1;
|
||||||
|
@ -654,7 +706,7 @@ static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
|
||||||
count = dwarf_read_uleb128(p, &length);
|
count = dwarf_read_uleb128(p, &length);
|
||||||
p += count;
|
p += count;
|
||||||
|
|
||||||
BUG_ON((unsigned char *)p > end);
|
UNWINDER_BUG_ON((unsigned char *)p > end);
|
||||||
|
|
||||||
cie->initial_instructions = p + length;
|
cie->initial_instructions = p + length;
|
||||||
cie->augmentation++;
|
cie->augmentation++;
|
||||||
|
@ -682,16 +734,16 @@ static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
|
||||||
* routine in the CIE
|
* routine in the CIE
|
||||||
* augmentation.
|
* augmentation.
|
||||||
*/
|
*/
|
||||||
BUG();
|
UNWINDER_BUG();
|
||||||
} else if (*cie->augmentation == 'S') {
|
} else if (*cie->augmentation == 'S') {
|
||||||
BUG();
|
UNWINDER_BUG();
|
||||||
} else {
|
} else {
|
||||||
/*
|
/*
|
||||||
* Unknown augmentation. Assume
|
* Unknown augmentation. Assume
|
||||||
* 'z' augmentation.
|
* 'z' augmentation.
|
||||||
*/
|
*/
|
||||||
p = cie->initial_instructions;
|
p = cie->initial_instructions;
|
||||||
BUG_ON(!p);
|
UNWINDER_BUG_ON(!p);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -708,7 +760,8 @@ static int dwarf_parse_cie(void *entry, void *p, unsigned long len,
|
||||||
}
|
}
|
||||||
|
|
||||||
static int dwarf_parse_fde(void *entry, u32 entry_type,
|
static int dwarf_parse_fde(void *entry, u32 entry_type,
|
||||||
void *start, unsigned long len)
|
void *start, unsigned long len,
|
||||||
|
unsigned char *end)
|
||||||
{
|
{
|
||||||
struct dwarf_fde *fde;
|
struct dwarf_fde *fde;
|
||||||
struct dwarf_cie *cie;
|
struct dwarf_cie *cie;
|
||||||
|
@ -755,7 +808,7 @@ static int dwarf_parse_fde(void *entry, u32 entry_type,
|
||||||
|
|
||||||
/* Call frame instructions. */
|
/* Call frame instructions. */
|
||||||
fde->instructions = p;
|
fde->instructions = p;
|
||||||
fde->end = start + len;
|
fde->end = end;
|
||||||
|
|
||||||
/* Add to list. */
|
/* Add to list. */
|
||||||
spin_lock_irqsave(&dwarf_fde_lock, flags);
|
spin_lock_irqsave(&dwarf_fde_lock, flags);
|
||||||
|
@ -765,17 +818,33 @@ static int dwarf_parse_fde(void *entry, u32 entry_type,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dwarf_unwinder_dump(struct task_struct *task, struct pt_regs *regs,
|
static void dwarf_unwinder_dump(struct task_struct *task,
|
||||||
|
struct pt_regs *regs,
|
||||||
unsigned long *sp,
|
unsigned long *sp,
|
||||||
const struct stacktrace_ops *ops, void *data)
|
const struct stacktrace_ops *ops,
|
||||||
|
void *data)
|
||||||
{
|
{
|
||||||
struct dwarf_frame *frame;
|
struct dwarf_frame *frame, *_frame;
|
||||||
|
unsigned long return_addr;
|
||||||
|
|
||||||
frame = dwarf_unwind_stack(0, NULL);
|
_frame = NULL;
|
||||||
|
return_addr = 0;
|
||||||
|
|
||||||
while (frame && frame->return_addr) {
|
while (1) {
|
||||||
ops->address(data, frame->return_addr, 1);
|
frame = dwarf_unwind_stack(return_addr, _frame);
|
||||||
frame = frame->next;
|
|
||||||
|
if (_frame) {
|
||||||
|
dwarf_frame_free_regs(_frame);
|
||||||
|
mempool_free(_frame, dwarf_frame_pool);
|
||||||
|
}
|
||||||
|
|
||||||
|
_frame = frame;
|
||||||
|
|
||||||
|
if (!frame || !frame->return_addr)
|
||||||
|
break;
|
||||||
|
|
||||||
|
return_addr = frame->return_addr;
|
||||||
|
ops->address(data, return_addr, 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -787,24 +856,22 @@ static struct unwinder dwarf_unwinder = {
|
||||||
|
|
||||||
static void dwarf_unwinder_cleanup(void)
|
static void dwarf_unwinder_cleanup(void)
|
||||||
{
|
{
|
||||||
struct dwarf_cie *cie, *m;
|
struct dwarf_cie *cie;
|
||||||
struct dwarf_fde *fde, *n;
|
struct dwarf_fde *fde;
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Deallocate all the memory allocated for the DWARF unwinder.
|
* Deallocate all the memory allocated for the DWARF unwinder.
|
||||||
* Traverse all the FDE/CIE lists and remove and free all the
|
* Traverse all the FDE/CIE lists and remove and free all the
|
||||||
* memory associated with those data structures.
|
* memory associated with those data structures.
|
||||||
*/
|
*/
|
||||||
spin_lock_irqsave(&dwarf_cie_lock, flags);
|
list_for_each_entry(cie, &dwarf_cie_list, link)
|
||||||
list_for_each_entry_safe(cie, m, &dwarf_cie_list, link)
|
|
||||||
kfree(cie);
|
kfree(cie);
|
||||||
spin_unlock_irqrestore(&dwarf_cie_lock, flags);
|
|
||||||
|
|
||||||
spin_lock_irqsave(&dwarf_fde_lock, flags);
|
list_for_each_entry(fde, &dwarf_fde_list, link)
|
||||||
list_for_each_entry_safe(fde, n, &dwarf_fde_list, link)
|
|
||||||
kfree(fde);
|
kfree(fde);
|
||||||
spin_unlock_irqrestore(&dwarf_fde_lock, flags);
|
|
||||||
|
kmem_cache_destroy(dwarf_reg_cachep);
|
||||||
|
kmem_cache_destroy(dwarf_frame_cachep);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -816,7 +883,7 @@ static void dwarf_unwinder_cleanup(void)
|
||||||
* easy to lookup the FDE for a given PC, so we build a list of FDE
|
* easy to lookup the FDE for a given PC, so we build a list of FDE
|
||||||
* and CIE entries that make it easier.
|
* and CIE entries that make it easier.
|
||||||
*/
|
*/
|
||||||
void dwarf_unwinder_init(void)
|
static int __init dwarf_unwinder_init(void)
|
||||||
{
|
{
|
||||||
u32 entry_type;
|
u32 entry_type;
|
||||||
void *p, *entry;
|
void *p, *entry;
|
||||||
|
@ -831,6 +898,21 @@ void dwarf_unwinder_init(void)
|
||||||
f_entries = 0;
|
f_entries = 0;
|
||||||
entry = &__start_eh_frame;
|
entry = &__start_eh_frame;
|
||||||
|
|
||||||
|
dwarf_frame_cachep = kmem_cache_create("dwarf_frames",
|
||||||
|
sizeof(struct dwarf_frame), 0, SLAB_PANIC, NULL);
|
||||||
|
dwarf_reg_cachep = kmem_cache_create("dwarf_regs",
|
||||||
|
sizeof(struct dwarf_reg), 0, SLAB_PANIC, NULL);
|
||||||
|
|
||||||
|
dwarf_frame_pool = mempool_create(DWARF_FRAME_MIN_REQ,
|
||||||
|
mempool_alloc_slab,
|
||||||
|
mempool_free_slab,
|
||||||
|
dwarf_frame_cachep);
|
||||||
|
|
||||||
|
dwarf_reg_pool = mempool_create(DWARF_REG_MIN_REQ,
|
||||||
|
mempool_alloc_slab,
|
||||||
|
mempool_free_slab,
|
||||||
|
dwarf_reg_cachep);
|
||||||
|
|
||||||
while ((char *)entry < __stop_eh_frame) {
|
while ((char *)entry < __stop_eh_frame) {
|
||||||
p = entry;
|
p = entry;
|
||||||
|
|
||||||
|
@ -860,7 +942,7 @@ void dwarf_unwinder_init(void)
|
||||||
else
|
else
|
||||||
c_entries++;
|
c_entries++;
|
||||||
} else {
|
} else {
|
||||||
err = dwarf_parse_fde(entry, entry_type, p, len);
|
err = dwarf_parse_fde(entry, entry_type, p, len, end);
|
||||||
if (err < 0)
|
if (err < 0)
|
||||||
goto out;
|
goto out;
|
||||||
else
|
else
|
||||||
|
@ -877,9 +959,11 @@ void dwarf_unwinder_init(void)
|
||||||
if (err)
|
if (err)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
return;
|
return 0;
|
||||||
|
|
||||||
out:
|
out:
|
||||||
printk(KERN_ERR "Failed to initialise DWARF unwinder: %d\n", err);
|
printk(KERN_ERR "Failed to initialise DWARF unwinder: %d\n", err);
|
||||||
dwarf_unwinder_cleanup();
|
dwarf_unwinder_cleanup();
|
||||||
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
early_initcall(dwarf_unwinder_init);
|
||||||
|
|
|
@ -14,7 +14,6 @@
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
#include <asm/machvec.h>
|
#include <asm/machvec.h>
|
||||||
#include <asm/uaccess.h>
|
#include <asm/uaccess.h>
|
||||||
#include <asm/dwarf.h>
|
|
||||||
#include <asm/thread_info.h>
|
#include <asm/thread_info.h>
|
||||||
#include <cpu/mmu_context.h>
|
#include <cpu/mmu_context.h>
|
||||||
|
|
||||||
|
@ -262,9 +261,6 @@ void __init init_IRQ(void)
|
||||||
sh_mv.mv_init_irq();
|
sh_mv.mv_init_irq();
|
||||||
|
|
||||||
irq_ctx_init(smp_processor_id());
|
irq_ctx_init(smp_processor_id());
|
||||||
|
|
||||||
/* This needs to be early, but not too early.. */
|
|
||||||
dwarf_unwinder_init();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_SPARSE_IRQ
|
#ifdef CONFIG_SPARSE_IRQ
|
||||||
|
|
|
@ -5,18 +5,32 @@
|
||||||
#include <linux/signal.h>
|
#include <linux/signal.h>
|
||||||
#include <linux/sched.h>
|
#include <linux/sched.h>
|
||||||
#include <linux/uaccess.h>
|
#include <linux/uaccess.h>
|
||||||
|
#include <asm/unwinder.h>
|
||||||
#include <asm/system.h>
|
#include <asm/system.h>
|
||||||
|
|
||||||
#ifdef CONFIG_BUG
|
#ifdef CONFIG_BUG
|
||||||
static void handle_BUG(struct pt_regs *regs)
|
void handle_BUG(struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
|
const struct bug_entry *bug;
|
||||||
|
unsigned long bugaddr = regs->pc;
|
||||||
enum bug_trap_type tt;
|
enum bug_trap_type tt;
|
||||||
tt = report_bug(regs->pc, regs);
|
|
||||||
|
if (!is_valid_bugaddr(bugaddr))
|
||||||
|
goto invalid;
|
||||||
|
|
||||||
|
bug = find_bug(bugaddr);
|
||||||
|
|
||||||
|
/* Switch unwinders when unwind_stack() is called */
|
||||||
|
if (bug->flags & BUGFLAG_UNWINDER)
|
||||||
|
unwinder_faulted = 1;
|
||||||
|
|
||||||
|
tt = report_bug(bugaddr, regs);
|
||||||
if (tt == BUG_TRAP_TYPE_WARN) {
|
if (tt == BUG_TRAP_TYPE_WARN) {
|
||||||
regs->pc += instruction_size(regs->pc);
|
regs->pc += instruction_size(bugaddr);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
invalid:
|
||||||
die("Kernel BUG", regs, TRAPA_BUG_OPCODE & 0xff);
|
die("Kernel BUG", regs, TRAPA_BUG_OPCODE & 0xff);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -28,8 +42,10 @@ int is_valid_bugaddr(unsigned long addr)
|
||||||
return 0;
|
return 0;
|
||||||
if (probe_kernel_address((insn_size_t *)addr, opcode))
|
if (probe_kernel_address((insn_size_t *)addr, opcode))
|
||||||
return 0;
|
return 0;
|
||||||
|
if (opcode == TRAPA_BUG_OPCODE)
|
||||||
|
return 1;
|
||||||
|
|
||||||
return opcode == TRAPA_BUG_OPCODE;
|
return 0;
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
@ -136,6 +136,7 @@ static void die_if_no_fixup(const char * str, struct pt_regs * regs, long err)
|
||||||
regs->pc = fixup->fixup;
|
regs->pc = fixup->fixup;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
die(str, regs, err);
|
die(str, regs, err);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -11,6 +11,7 @@
|
||||||
#include <linux/errno.h>
|
#include <linux/errno.h>
|
||||||
#include <linux/list.h>
|
#include <linux/list.h>
|
||||||
#include <linux/spinlock.h>
|
#include <linux/spinlock.h>
|
||||||
|
#include <linux/module.h>
|
||||||
#include <asm/unwinder.h>
|
#include <asm/unwinder.h>
|
||||||
#include <asm/atomic.h>
|
#include <asm/atomic.h>
|
||||||
|
|
||||||
|
@ -53,8 +54,6 @@ static struct list_head unwinder_list = {
|
||||||
|
|
||||||
static DEFINE_SPINLOCK(unwinder_lock);
|
static DEFINE_SPINLOCK(unwinder_lock);
|
||||||
|
|
||||||
static atomic_t unwinder_running = ATOMIC_INIT(0);
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* select_unwinder - Select the best registered stack unwinder.
|
* select_unwinder - Select the best registered stack unwinder.
|
||||||
*
|
*
|
||||||
|
@ -122,6 +121,8 @@ int unwinder_register(struct unwinder *u)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int unwinder_faulted = 0;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Unwind the call stack and pass information to the stacktrace_ops
|
* Unwind the call stack and pass information to the stacktrace_ops
|
||||||
* functions. Also handle the case where we need to switch to a new
|
* functions. Also handle the case where we need to switch to a new
|
||||||
|
@ -144,19 +145,20 @@ void unwind_stack(struct task_struct *task, struct pt_regs *regs,
|
||||||
* Hopefully this will give us a semi-reliable stacktrace so we
|
* Hopefully this will give us a semi-reliable stacktrace so we
|
||||||
* can diagnose why curr_unwinder->dump() faulted.
|
* can diagnose why curr_unwinder->dump() faulted.
|
||||||
*/
|
*/
|
||||||
if (atomic_inc_return(&unwinder_running) != 1) {
|
if (unwinder_faulted) {
|
||||||
spin_lock_irqsave(&unwinder_lock, flags);
|
spin_lock_irqsave(&unwinder_lock, flags);
|
||||||
|
|
||||||
if (!list_is_singular(&unwinder_list)) {
|
/* Make sure no one beat us to changing the unwinder */
|
||||||
|
if (unwinder_faulted && !list_is_singular(&unwinder_list)) {
|
||||||
list_del(&curr_unwinder->list);
|
list_del(&curr_unwinder->list);
|
||||||
curr_unwinder = select_unwinder();
|
curr_unwinder = select_unwinder();
|
||||||
|
|
||||||
|
unwinder_faulted = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_unlock_irqrestore(&unwinder_lock, flags);
|
spin_unlock_irqrestore(&unwinder_lock, flags);
|
||||||
atomic_dec(&unwinder_running);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
curr_unwinder->dump(task, regs, sp, ops, data);
|
curr_unwinder->dump(task, regs, sp, ops, data);
|
||||||
|
|
||||||
atomic_dec(&unwinder_running);
|
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(unwind_stack);
|
||||||
|
|
Loading…
Reference in a new issue