newlib/libgloss/sparc_leon/asm-leon/leonbare_kernel.h

439 lines
18 KiB
C
Raw Normal View History

2011-12-15 Konrad Eisele <konrad@gaisler.com> * configure.in: Add SPARC LEON support. * configure: Regenerated. * sparc_leon/asm-leon/amba.h, sparc_leon/asm-leon/asmmacro.h, sparc_leon/asm-leon/clock.h, sparc_leon/asm-leon/contextswitch.h, sparc_leon/asm-leon/elfmacro.h, sparc_leon/asm-leon/head.h, sparc_leon/asm-leon/irq.h, sparc_leon/asm-leon/jiffies.h, sparc_leon/asm-leon/lambapp.h, sparc_leon/asm-leon/lambapp_devs.h, sparc_leon/asm-leon/leon.h, sparc_leon/asm-leon/leon3.h, sparc_leon/asm-leon/leonbare_debug.h, sparc_leon/asm-leon/leonbare_kernel.h, sparc_leon/asm-leon/leonbare_kernel_queue.h, sparc_leon/asm-leon/leoncompat.h, sparc_leon/asm-leon/leondbg.h, sparc_leon/asm-leon/leonstack.h, sparc_leon/asm-leon/liblocks.h, sparc_leon/asm-leon/linkage.h, sparc_leon/asm-leon/param.h, sparc_leon/asm-leon/queue.h, sparc_leon/asm-leon/spinlock.h, sparc_leon/asm-leon/stack.h, sparc_leon/asm-leon/time.h, sparc_leon/asm-leon/timer.h, sparc_leon/asm-leon/types.h, sparc_leon/asm-leon/winmacros.h: New file. * sparc_leon/Makefile.in, sparc_leon/_exit.c, sparc_leon/amba.c, sparc_leon/amba_dbg.c, sparc_leon/amba_driver.c, sparc_leon/amba_scan.c, sparc_leon/asm-leon, sparc_leon/bdinit.S, sparc_leon/busscan.S, sparc_leon/cacheA.S, sparc_leon/catch_interrupt.c, sparc_leon/catch_interrupt_mvt.c, sparc_leon/catch_interrupt_pending.c, sparc_leon/catch_interrupt_svt.c, sparc_leon/configure.in, sparc_leon/console.c, sparc_leon/console_dbg.c, sparc_leon/console_init.c, sparc_leon/contextswitch.c, sparc_leon/contextswitch_asm.S, sparc_leon/crt0.S, sparc_leon/crti.S, sparc_leon/crtn.S, sparc_leon/etrap.S, sparc_leon/etrap_fast.S, sparc_leon/fpu.S, sparc_leon/gettimeofday.c, sparc_leon/initcalls.c, sparc_leon/io.c, sparc_leon/irqinstall.S, sparc_leon/irqtrap.S, sparc_leon/irqtrap_fast.S, sparc_leon/jiffies.c, sparc_leon/kernel.c, sparc_leon/kernel_context.S, sparc_leon/kernel_debug.c, sparc_leon/kernel_debug_var.c, sparc_leon/kernel_mm.c, sparc_leon/kernel_mutex.c, sparc_leon/kernel_queue.c, sparc_leon/kernel_sched.c, sparc_leon/kernel_thread.c, sparc_leon/lcpuinit.S, sparc_leon/locore.S, sparc_leon/locore_atexit.c, sparc_leon/locore_clean.S, sparc_leon/locore_mvt.S, sparc_leon/locore_mvt_reset.S, sparc_leon/locore_svt.S, sparc_leon/locore_svt_reset.S, sparc_leon/locore_svtdisp.S, sparc_leon/locore_var.S, sparc_leon/locore_var_svt.S, sparc_leon/mmu_asm.S, sparc_leon/mutex.c, sparc_leon/nocache.S, sparc_leon/pnpinit.c, sparc_leon/pnpinit_malloc.c, sparc_leon/pnpinit_simple.c, sparc_leon/regwin.S, sparc_leon/regwin_patch.c, sparc_leon/regwin_slow.S, sparc_leon/regwinflush.S, sparc_leon/rtc.c, sparc_leon/rtrap.S, sparc_leon/rtrap_fast.S, sparc_leon/stop.S, sparc_leon/timer.c, sparc_leon/times.c: New file * sparc_leon/configure: Regenerate
2011-12-15 23:58:40 +01:00
/*
* Copyright (c) 2011 Aeroflex Gaisler
*
* BSD license:
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the "Software"), to deal
* in the Software without restriction, including without limitation the rights
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
* copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
* THE SOFTWARE.
*/
#ifndef __LEONBARE_KERNEL_H__
#define __LEONBARE_KERNEL_H__
#include <asm-leon/contextswitch.h>
#include <asm-leon/leonbare_debug.h>
#include <asm-leon/leon.h>
#ifndef __ASSEMBLER__
#include <asm-leon/leonbare_kernel_queue.h>
#include <reent.h>
#endif
#include "irq.h"
#define LEONBARE_RUNQ_READY_NR (2) /* queue 0-1 for ready */
#define LEONBARE_RUNQ_SUSPENDED_IDX (2) /* queue 2 for suspended */
#define LEONBARE_RUNQ_PREPARE_IDX (3) /* LEONBARE_RUNQ_READY_NR times queues */
#define LEONBARE_RUNQ_KILLED_IDX (LEONBARE_RUNQ_PREPARE_IDX+LEONBARE_RUNQ_READY_NR) /* queue 2 for killed threads */
#define LEONBARE_RUNQ_NR (LEONBARE_RUNQ_KILLED_IDX+1)
#define LEONBARE_RUNQ_ISREADY(idx) ((idx) >= 0 && (idx) < LEONBARE_RUNQ_READY_NR)
#define LEONBARE_RUNQ_ISPREPARE(idx) ((idx) >= LEONBARE_RUNQ_PREPARE_IDX && (idx) < LEONBARE_RUNQ_PREPARE_IDX+LEONBARE_RUNQ_READY_NR)
#define LEONBARE_RUNQ_ISSUSPEND(idx) ((idx) == LEONBARE_RUNQ_SUSPENDED_IDX)
#define LEONBARE_RUNQ_ISKILLED(idx) ((idx) == LEONBARE_RUNQ_KILLED_IDX)
#ifndef __ASSEMBLER__
#ifndef NULL
#define NULL ((void *)0)
#endif
#define MACRO_BEGIN do {
#define MACRO_END } while (0)
#define optbarrier() __asm__ __volatile__("": : :"memory")
typedef struct leonbare_thread_ctx
{
unsigned long sf_locals[8];
unsigned long sf_ins[8];
unsigned long outs[8];
unsigned long globals[8];
unsigned long psr;
unsigned long wim;
unsigned long magic;
unsigned long fpu;
/* size aligned to 8 */
} leonbare_thread_ctx_t;
#define LEONBARE_THREAD_CTX_SZ sizeof(struct leonbare_thread_ctx)
typedef
LBTAILQ_HEAD (leonbare_mutex_queue, leonbare_mutex) *
leonbare_mutex_queue_t;
#endif
#define LEONBARE_THREAD_OFFSET_CTX 0
#ifndef __ASSEMBLER__
struct leonbare_thread_protect
{
unsigned int runq;
unsigned int krp_runq_depth;
unsigned int krp_k_depth;
struct leonbare_mutex *krp_m;
unsigned int krp_m_depth;
unsigned int krp_flags;;
unsigned int krp_flags_depth;
};
#define LEONBARE_INT_DISABLE_DECL unsigned long _irq_flags = leonbare_disable_traps();
#define LEONBARE_INT_ENABLE_DECL leonbare_enable_traps(_irq_flags);
#define leonbare_setu32p(a,v) leonbare_leon23_storenocache(a,v)
#define leonbare_setu32(a,v) leonbare_leon23_storenocache(a,v)
#define leonbare_getu32(a) leonbare_leon23_loadnocache(a)
#define LEONBARE_KERNEL_UNCACHED
#ifndef LEONBARE_KERNEL_UNCACHED
#define LEONBARE_KERNEL_SETU32P(a,v) (a=v)
#define LEONBARE_KERNEL_SETU32(a,v) (a=v) /* uncached version should return v */
#define LEONBARE_KERNEL_GETU32(a) (a)
#define LEONBARE_KERNEL_GETU32P(a) (a)
#define LEONBARE_KERNEL_GETI32(a) (a)
#define LEONBARE_KERNEL_GETU32P_CAST(a,typ) ((typ)(a))
#define LEONBARE_KERNEL_GETU32P_BARE(a) (*(a)) /* uncached: no & */
#define LEONBARE_KERNEL_SETU32P_BARE(a,v) (*(a) = v) /* uncached: no & */
#else
#define LEONBARE_KERNEL_SETU32P(a,v) (leonbare_setu32p(&a,v))
#define LEONBARE_KERNEL_SETU32(a,v) (leonbare_setu32p(&a,v)) /* uncached version should return v */
#define LEONBARE_KERNEL_GETU32(a) (leonbare_getu32(&a))
#define LEONBARE_KERNEL_GETU32P(a) ((void *)leonbare_getu32(&a))
#define LEONBARE_KERNEL_GETI32(a) (leonbare_getu32(&a))
#define LEONBARE_KERNEL_GETU32P_CAST(a,typ) ((typ)(LEONBARE_KERNEL_GETU32P(a)))
#define LEONBARE_KERNEL_GETU32P_BARE(a) ((void *)leonbare_getu32(a)) /* uncached: no & */
#define LEONBARE_KERNEL_SETU32P_BARE(a,v) (leonbare_setu32p(a,v)) /* uncached: no & */
#endif
#define LEONBARE_SMP_SPINLOCK_AQUIRE(l)
#define LEONBARE_SMP_SPINLOCK_RELEASE(l)
#define LEONBARE_ISQ_ISDISABLED ((leon23_getpsr() & SPARC_PSR_PIL_MASK) == SPARC_PSR_PIL_MASK)
#define _LEONBARE_PROTECT_IRQ_START \
if (LEONBARE_KR_CURRENT->th_prot.krp_flags_depth++) { \
LBPASSERT((LEONBARE_ISQ_ISDISABLED),"Internal error: Recursiv IRQ protection with irq's enabled",0); \
} else { \
LEONBARE_KR_CURRENT->th_prot.krp_flags = leonbare_disable_traps(); \
}
#define _LEONBARE_PROTECT_IRQ_END \
if (--LEONBARE_KR_CURRENT->th_prot.krp_flags_depth) { \
LBPASSERT((LEONBARE_ISQ_ISDISABLED),"Internal error: Recursiv IRQ protection with irq's enabled",0); \
} else { \
leonbare_enable_traps(LEONBARE_KR_CURRENT->th_prot.krp_flags); \
}
#define _LEONBARE_PROTECT_MUTEXSTRUCT_START(m) \
if (LEONBARE_KR_CURRENT->th_prot.krp_m_depth++) { \
LBPASSERT((LEONBARE_KR_CURRENT->th_prot.krp_m == m),"Mutex protection only allowed for one mutex at a time",0); \
} else { \
LEONBARE_SMP_SPINLOCK_AQUIRE(m->smp_lock); \
LEONBARE_KR_CURRENT->th_prot.krp_m = m; \
}
#define _LEONBARE_PROTECT_MUTEXSTRUCT_END(m) \
LBPASSERT((LEONBARE_KR_CURRENT->th_prot.krp_m == m),"Mutex protection only allowed for one mutex at a time",0); \
if ((--LEONBARE_KR_CURRENT->th_prot.krp_m_depth) == 0) { \
LEONBARE_SMP_SPINLOCK_RELEASE(m->smp_lock); \
}
#define _LEONBARE_PROTECT_KERNEL_START \
if (LEONBARE_KR_CURRENT->th_prot.krp_k_depth++ == 0) { \
LEONBARE_SMP_SPINLOCK_AQUIRE(LEONBARE_KR_LOCK); \
}
#define _LEONBARE_PROTECT_KERNEL_END \
if ((--LEONBARE_KR_CURRENT->th_prot.krp_k_depth) == 0) { \
LEONBARE_SMP_SPINLOCK_RELEASE(LEONBARE_KR_LOCK); \
}
#define LEONBARE_PROTECT_MUTEXSTRUCT_START(m) \
_LEONBARE_PROTECT_IRQ_START; \
_LEONBARE_PROTECT_MUTEXSTRUCT_START(m)
#define LEONBARE_PROTECT_MUTEXSTRUCT_END(m) \
_LEONBARE_PROTECT_MUTEXSTRUCT_END(m) \
_LEONBARE_PROTECT_IRQ_END;
#define LEONBARE_PROTECT_KERNEL_START() \
_LEONBARE_PROTECT_IRQ_START; \
_LEONBARE_PROTECT_KERNEL_START;
#define LEONBARE_PROTECT_KERNEL_END() \
_LEONBARE_PROTECT_KERNEL_END; \
_LEONBARE_PROTECT_IRQ_END;
typedef struct leonbare_thread
{
struct leonbare_thread_ctx th_ctx;
unsigned int th_flags;
int th_account; /* how many ticks the thread stays in the readyqueue for one round */
int th_caccount; /* current value of th_account, updated on reinsertion */
unsigned int th_pri_idx; /* ready queue index */
unsigned int th_runq_idx; /* ready queue index index */
unsigned int th_runq_which; /* 0: ready queue, 1: ready prepare queue */
char *th_name;
int th_result;
int (*th_func) (void *);
void *th_arg;
char *th_stack_base;
unsigned int th_stack_size;
struct _reent th_reent; /* reentrant structure for newlib */
struct _reent *th_reentp; /* pointer to eather pt_reent or global reent */
struct leonbare_thread_protect th_prot;
LBTAILQ_ENTRY (leonbare_thread) th_runq;
LBTAILQ_ENTRY (leonbare_thread) th_allq;
LBTAILQ_ENTRY (leonbare_thread) th_mutex;
struct leonbare_mutex_queue th_mutex_locked;
} *leonbare_thread_t __attribute__ ((aligned (8)));
#define LEONBARE_TH_FLAGS_get(c) LEONBARE_KERNEL_GETU32((c)->th_flags)
#define LEONBARE_TH_ACCOUNT_get(c) LEONBARE_KERNEL_GETI32((c)->th_account)
#define LEONBARE_TH_CACCOUNT_get(c) LEONBARE_KERNEL_GETI32((c)->th_caccount)
#define LEONBARE_TH_PRI_IDX_get(c) LEONBARE_KERNEL_GETU32((c)->th_pri_idx)
#define LEONBARE_TH_RUNQ_IDX_get(c) LEONBARE_KERNEL_GETU32((c)->th_runq_idx)
#define LEONBARE_TH_RUNQ_WHICH_get(c) LEONBARE_KERNEL_GETU32((c)->th_runq_which)
#define LEONBARE_TH_NAME_get(c) LEONBARE_KERNEL_GETU32P((c)->th_name)
#define LEONBARE_TH_RESULT_get(c) LEONBARE_KERNEL_GETI32((c)->th_result)
#define LEONBARE_TH_FUNC_get(c) LEONBARE_KERNEL_GETU32((c)->th_func)
#define LEONBARE_TH_ARG_get(c) LEONBARE_KERNEL_GETU32((c)->th_arg)
#define LEONBARE_TH_STACK_BASE_get(c) LEONBARE_KERNEL_GETU32P((c)->th_stack_base)
#define LEONBARE_TH_STACK_SIZE_get(c) LEONBARE_KERNEL_GETU32((c)->th_stack_size)
#define LEONBARE_TH_REENTP_get(c) LEONBARE_KERNEL_GETU32P((c)->th_reentp)
#define LEONBARE_TH_NAME(c) (c->th_name)
#define LEONBARE_TH_NAME_DBG(c) (LEONBARE_TH_NAME(c) ? LEONBARE_TH_NAME(c) : "<unknown>")
#define LEONBARE_REENT_SET(p) ((_impure_ptr=(p)->th_reentp)==_impure_ptr)
#define LEONBARE_TH_READY (1<<0)
#define LEONBARE_TH_SUSPENDED (1<<1)
#define LEONBARE_TH_TERMINATED (1<<2)
#define LEONBARE_TH_FINISHED (1<<3)
#define LEONBARE_TH_SATEMASK (LEONBARE_TH_READY | \
LEONBARE_TH_SUSPENDED | \
LEONBARE_TH_TERMINATED | \
LEONBARE_TH_FINISHED)
#define LEONBARE_TH_SETSTATE(c,f) c->th_flags = ((c->th_flags & ~LEONBARE_TH_SATEMASK) | (f & LEONBARE_TH_SATEMASK))
#define LEONBARE_TH_ORSTATE(c,f) c->th_flags |= (f & LEONBARE_TH_SATEMASK)
typedef LBTAILQ_HEAD (leonbare_thread_queue,
leonbare_thread) * leonbare_thread_queue_t;
extern struct leonbare_kernel leonbare_kernel;
#define KERNEL_GLOBAL leonbare_kernel
typedef struct leonbare_kernel
{
leonbare_thread_t kr_cur, kr_next;
struct leonbare_thread_queue kr_runq[LEONBARE_RUNQ_NR];
struct leonbare_thread_queue kr_allq;
struct leonbare_mutex_queue kr_allm;
int kr_is_inkernel, kr_need_schedule, kr_is_preemption, kr_runq_which;
int kr_protect_flags;
} leonbare_kernel_t __attribute__ ((aligned (8)));
#define LEONBARE_KR_CURRENT (KERNEL_GLOBAL.kr_cur)
#define LEONBARE_KR_NEXT (KERNEL_GLOBAL.kr_next)
#define LEONBARE_KR_RUNQ(i) (&(KERNEL_GLOBAL.kr_runq[i]))
#define LEONBARE_KR_RUNQ_WHICH (KERNEL_GLOBAL.kr_runq_which)
#define LEONBARE_KR_ALLQ (&(KERNEL_GLOBAL.kr_allq))
#define LEONBARE_KR_ALLM (&(KERNEL_GLOBAL.kr_allm))
#define LEONBARE_KR_IS_IN_KERNEL (KERNEL_GLOBAL.kr_is_inkernel)
#define LEONBARE_KR_IS_PREEMPTION (KERNEL_GLOBAL.kr_is_preemption)
#define LEONBARE_KR_NEED_SCHEDULE (LEONBARE_KR_CURRENT != LEONBARE_KR_NEXT)
#define LEONBARE_STACKALIGN(sp) ((((unsigned int)sp) + 7) & ~7)
/* context switching macros, implemented via setjmp/longjmp plus saving errno */
#define SAVE_CONTEXT(t) ( _leonbare_kernel_savecontext((t), 0) )
#define RESTORE_CONTEXT(t) _leonbare_kernel_switchto((t), 1)
#define KERNEL_SCHEDULE(f,retval) \
MACRO_BEGIN \
LEONBARE_KR_IS_IN_KERNEL--; \
if (LEONBARE_KR_IS_IN_KERNEL == 0 && LEONBARE_KR_NEED_SCHEDULE) { \
LEONBARE_KR_IS_IN_KERNEL++; \
if ((f) && (SAVE_CONTEXT(LEONBARE_KR_CURRENT) == 0)) { \
leonbare_sched(); \
} \
optbarrier(); \
LEONBARE_KR_IS_IN_KERNEL--; \
} \
MACRO_END
#define KERNEL_ENTER LEONBARE_KR_IS_IN_KERNEL++;
#define KERNEL_EXIT(f,ret) KERNEL_SCHEDULE(f,ret)
int leonbare_thread_init ();
int leonbare_thread_create (struct leonbare_thread *thread, char *stack,
int stacksize);
int leonbare_sched_update ();
leonbare_thread_t leonbare_sched_paytime ();
void leonbare_sched_insert (struct leonbare_thread *thread, int head,
int prepare);
unsigned int leonbare_sched ();
unsigned int reschedule ();
unsigned int _leonbare_kernel_switchto (struct leonbare_thread *old,
struct leonbare_thread *new);
#define LEONBARE_STACK_DEFINE(n,size) unsigned char n[size] __attribute__((aligned(8)));
#define LEONBARE_STACK_SIZE_DEFAULT 1024*20
typedef struct leonbare_mutex
{
unsigned int mx_owner_cnt;
leonbare_thread_t mx_owner;
struct leonbare_thread_queue mx_threads;
LBTAILQ_ENTRY (leonbare_mutex) mx_allm;
LBTAILQ_ENTRY (leonbare_mutex) mx_locked;
} *leonbare_mutex_t;
#define LEONBARE_MUTEX_OWNER_GET(m) LEONBARE_KERNEL_GETU32(m->mx_owner)
#define LEONBARE_MUTEX_OWNER_SET(m,o) LEONBARE_KERNEL_SETU32(m->mx_owner,o)
#define LEONBARE_MUTEX_OWNER_CNT_GET(m) LEONBARE_KERNEL_GETU32(m->mx_owner_cnt)
#define LEONBARE_MUTEX_OWNER_CNT_SET(m,o) LEONBARE_KERNEL_SETU32(m->mx_owner_cnt,o)
#define LEONBARE_MUTEX_LOCK_TIMEOUT -1
#define LEONBARE_MUTEX_LOCK_OK 0
#define LEONBARE_MUTEX_LOCK_ERROR 1
#define LEONBARE_MUTEX_UNLOCK_OK 0
#define LEONBARE_MUTEX_UNLOCK_ERROR 1
#define LEONBARE_PROTECT_DECL(flags) unsigned long flags;
#define LEONBARE_PROTECT_KERNEL(flags) flags = leonbare_disable_traps();
#define LEONBARE_UNPROTECT_KERNEL(flags) leonbare_enable_traps(flags);
#define LEONBARE_PROTECT_MUTEX(flags,m) flags = leonbare_disable_traps();
#define LEONBARE_UNPROTECT_MUTEX(flags,m) leonbare_enable_traps(flags);
#else
#define LEONBARE_THREAD_CTX_STORE_LOCALS(base_reg) \
std %l0, [%base_reg + LEONBARE_THREAD_CTX_STACK_L0]; \
std %l2, [%base_reg + LEONBARE_THREAD_CTX_STACK_L2]; \
std %l4, [%base_reg + LEONBARE_THREAD_CTX_STACK_L4]; \
std %l6, [%base_reg + LEONBARE_THREAD_CTX_STACK_L6];
#define LEONBARE_THREAD_CTX_STORE_INS(base_reg) \
std %i0, [%base_reg + LEONBARE_THREAD_CTX_STACK_I0]; \
std %i2, [%base_reg + LEONBARE_THREAD_CTX_STACK_I2]; \
std %i4, [%base_reg + LEONBARE_THREAD_CTX_STACK_I4]; \
std %i6, [%base_reg + LEONBARE_THREAD_CTX_STACK_I6];
#define LEONBARE_THREAD_CTX_STORE_OUTS(base_reg) \
std %o0, [%base_reg + LEONBARE_THREAD_CTX_STACK_O0]; \
std %o2, [%base_reg + LEONBARE_THREAD_CTX_STACK_O2]; \
std %o4, [%base_reg + LEONBARE_THREAD_CTX_STACK_O4]; \
std %o6, [%base_reg + LEONBARE_THREAD_CTX_STACK_O6];
#define LEONBARE_THREAD_CTX_STORE_GLOBALS(base_reg) \
st %g1, [%base_reg + LEONBARE_THREAD_CTX_STACK_G1]; \
std %g2, [%base_reg + LEONBARE_THREAD_CTX_STACK_G2]; \
std %g4, [%base_reg + LEONBARE_THREAD_CTX_STACK_G4]; \
std %g6, [%base_reg + LEONBARE_THREAD_CTX_STACK_G6];
#define LEONBARE_THREAD_CTX_LOAD_LOCALS(base_reg) \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_L0], %l0; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_L2], %l2; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_L4], %l4; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_L6], %l6;
#define LEONBARE_THREAD_CTX_LOAD_INS(base_reg) \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_I0], %i0; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_I2], %i2; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_I4], %i4; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_I6], %i6;
#define LEONBARE_THREAD_CTX_LOAD_OUTS(base_reg) \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_O0], %o0; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_O2], %o2; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_O4], %o4; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_O6], %o6;
#define LEONBARE_THREAD_CTX_LOAD_GLOBALS(base_reg) \
ld [%base_reg + LEONBARE_THREAD_CTX_STACK_G1], %g1; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_G2], %g2; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_G4], %g4; \
ldd [%base_reg + LEONBARE_THREAD_CTX_STACK_G6], %g6;
#define LEONBARE_THREAD_CTX_STACK_L0 (0*8*4)
#define LEONBARE_THREAD_CTX_STACK_L2 (LEONBARE_THREAD_CTX_STACK_L0+(2*4))
#define LEONBARE_THREAD_CTX_STACK_L4 (LEONBARE_THREAD_CTX_STACK_L0+(4*4))
#define LEONBARE_THREAD_CTX_STACK_L6 (LEONBARE_THREAD_CTX_STACK_L0+(6*4))
#define LEONBARE_THREAD_CTX_STACK_I0 (1*8*4)
#define LEONBARE_THREAD_CTX_STACK_I2 (LEONBARE_THREAD_CTX_STACK_I0+(2*4))
#define LEONBARE_THREAD_CTX_STACK_I4 (LEONBARE_THREAD_CTX_STACK_I0+(4*4))
#define LEONBARE_THREAD_CTX_STACK_I6 (LEONBARE_THREAD_CTX_STACK_I0+(6*4))
#define LEONBARE_THREAD_CTX_STACK_O0 (2*8*4)
#define LEONBARE_THREAD_CTX_STACK_O2 (LEONBARE_THREAD_CTX_STACK_O0+(2*4))
#define LEONBARE_THREAD_CTX_STACK_O4 (LEONBARE_THREAD_CTX_STACK_O0+(4*4))
#define LEONBARE_THREAD_CTX_STACK_O6 (LEONBARE_THREAD_CTX_STACK_O0+(6*4))
#define LEONBARE_THREAD_CTX_STACK_G0 (3*8*4)
#define LEONBARE_THREAD_CTX_STACK_G1 (LEONBARE_THREAD_CTX_STACK_G0+(1*4))
#define LEONBARE_THREAD_CTX_STACK_G2 (LEONBARE_THREAD_CTX_STACK_G0+(2*4))
#define LEONBARE_THREAD_CTX_STACK_G4 (LEONBARE_THREAD_CTX_STACK_G0+(4*4))
#define LEONBARE_THREAD_CTX_STACK_G6 (LEONBARE_THREAD_CTX_STACK_G0+(6*4))
#define LEONBARE_THREAD_CTX_STACK_PSR (4*8*4)
#define LEONBARE_THREAD_CTX_STACK_WIM (LEONBARE_THREAD_CTX_STACK_PSR+4)
#define LEONBARE_THREAD_CTX_STACK_MAGIC (LEONBARE_THREAD_CTX_STACK_PSR+8)
#define LEONBARE_THREAD_CTX_STACK_FPU (LEONBARE_THREAD_CTX_STACK_PSR+12)
#define LEONBARE_THREAD_CTX_SZ (LEONBARE_THREAD_CTX_STACK_PSR+16)
#endif /* __ASSEMBLER__ */
# define LEONBARE_STOPALL \
LBDEBUG_HEADER_PRINTF(LBDEBUG_ALWAYS_NR,"Stopped at %s(%d), possibly not implemented yet\n",__FUNCTION__,__LINE__); \
_leonbare_Stop();
#define LEONBARE_THREAD_CTX_MAGIC 0x1234
#ifdef LBDEBUG_DO_ASSERT
#define LEONBARE_VERIFYIRQDISABLED() LBPASSERT(((leon23_getpsr() & SPARC_PSR_PIL_MASK) == SPARC_PSR_PIL_MASK),"Irq must be disabled (pil==0xf)\n",0)
#define LEONBARE_VERIFYSCHED() leonbare_sched_verify()
#else
#define LEONBARE_VERIFYIRQDISABLED()
#define LEONBARE_VERIFYSCHED()
#endif
#define LEONBARE_PRINTQUEUES() if (PDEBUG_FLAGS_CHECK(LBDEBUG_QUEUE_NR)) { leonbare_sched_printqueue(); }
#endif