kernel-fxtec-pro1x/include/asm-v850/processor.h
Chase Venters f6dc8c5b8e [PATCH] Make cpu_relax() imply barrier() on all arches
During the recent discussion of taking 'volatile' off of the spinlock, I
noticed that while most arches #define cpu_relax() such that it implies
barrier(), some arches define cpu_relax() to be empty.

This patch changes the definition of cpu_relax() for frv, h8300, m68knommu,
sh, sh64, v850 and xtensa from an empty while(0) to the compiler barrier().

Signed-off-by: Chase Venters <chase.venters@clientec.com>
Acked-by: Arjan van de Ven <arjan@Linux.intel.com>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-07-12 12:56:45 -07:00

120 lines
3.2 KiB
C

/*
* include/asm-v850/processor.h
*
* Copyright (C) 2001,02,03 NEC Electronics Corporation
* Copyright (C) 2001,02,03 Miles Bader <miles@gnu.org>
*
* This file is subject to the terms and conditions of the GNU General
* Public License. See the file COPYING in the main directory of this
* archive for more details.
*
* Written by Miles Bader <miles@gnu.org>
*/
#ifndef __V850_PROCESSOR_H__
#define __V850_PROCESSOR_H__
#ifndef __ASSEMBLY__ /* <linux/thread_info.h> is not asm-safe. */
#include <linux/thread_info.h>
#endif
#include <linux/compiler.h>
#include <asm/ptrace.h>
#include <asm/entry.h>
/* Some code expects `segment' stuff to be defined here. */
#include <asm/segment.h>
/*
* The only places this is used seem to be horrible bletcherous kludges,
* so we just define it to be as large as possible.
*/
#define TASK_SIZE (0xFFFFFFFF)
/*
* This decides where the kernel will search for a free chunk of vm
* space during mmap's. We won't be using it.
*/
#define TASK_UNMAPPED_BASE 0
#ifndef __ASSEMBLY__
/*
* Default implementation of macro that returns current
* instruction pointer ("program counter").
*/
#define current_text_addr() ({ __label__ _l; _l: &&_l;})
/* If you change this, you must change the associated assembly-languages
constants defined below, THREAD_*. */
struct thread_struct {
/* kernel stack pointer (must be first field in structure) */
unsigned long ksp;
};
#define INIT_THREAD { sizeof init_stack + (unsigned long)init_stack }
/* Do necessary setup to start up a newly executed thread. */
static inline void start_thread (struct pt_regs *regs,
unsigned long pc, unsigned long usp)
{
regs->pc = pc;
regs->gpr[GPR_SP] = usp;
regs->kernel_mode = 0;
}
/* Free all resources held by a thread. */
static inline void release_thread (struct task_struct *dead_task)
{
}
/* Prepare to copy thread state - unlazy all lazy status */
#define prepare_to_copy(tsk) do { } while (0)
extern int kernel_thread (int (*fn)(void *), void * arg, unsigned long flags);
/* Free current thread data structures etc. */
static inline void exit_thread (void)
{
}
/* Return the registers saved during context-switch by the currently
not-running thread T. Note that this only includes some registers!
See entry.S for details. */
#define thread_saved_regs(t) \
((struct pt_regs*)((t)->thread.ksp + STATE_SAVE_PT_OFFSET))
/* Return saved (kernel) PC of a blocked thread. Actually, we return the
LP register, because the thread is actually blocked in switch_thread,
and we're interested in the PC it will _return_ to. */
#define thread_saved_pc(t) (thread_saved_regs(t)->gpr[GPR_LP])
unsigned long get_wchan (struct task_struct *p);
/* Return some info about the user process TASK. */
#define task_tos(task) ((unsigned long)task_stack_page(task) + THREAD_SIZE)
#define task_pt_regs(task) ((struct pt_regs *)task_tos (task) - 1)
#define task_sp(task) (task_pt_regs (task)->gpr[GPR_SP])
#define task_pc(task) (task_pt_regs (task)->pc)
/* Grotty old names for some. */
#define KSTK_EIP(task) task_pc (task)
#define KSTK_ESP(task) task_sp (task)
#define cpu_relax() barrier()
#else /* __ASSEMBLY__ */
#define THREAD_KSP 0
#endif /* !__ASSEMBLY__ */
#endif /* __V850_PROCESSOR_H__ */