OpenCores
URL https://opencores.org/ocsvn/or1k_soc_on_altera_embedded_dev_kit/or1k_soc_on_altera_embedded_dev_kit/trunk

Subversion Repositories or1k_soc_on_altera_embedded_dev_kit

[/] [or1k_soc_on_altera_embedded_dev_kit/] [trunk/] [linux-2.6/] [linux-2.6.24/] [arch/] [sparc/] [kernel/] [etrap.S] - Rev 3

Compare with Previous | Blame | View Log

/* $Id: etrap.S,v 1.31 2000/01/08 16:38:18 anton Exp $
 * etrap.S: Sparc trap window preparation for entry into the
 *          Linux kernel.
 *
 * Copyright (C) 1995 David S. Miller (davem@caip.rutgers.edu)
 */

#include <asm/head.h>
#include <asm/asi.h>
#include <asm/contregs.h>
#include <asm/page.h>
#include <asm/psr.h>
#include <asm/ptrace.h>
#include <asm/winmacro.h>
#include <asm/asmmacro.h>
#include <asm/thread_info.h>

/* Registers to not touch at all. */
#define t_psr        l0 /* Set by caller */
#define t_pc         l1 /* Set by caller */
#define t_npc        l2 /* Set by caller */
#define t_wim        l3 /* Set by caller */
#define t_twinmask   l4 /* Set at beginning of this entry routine. */
#define t_kstack     l5 /* Set right before pt_regs frame is built */
#define t_retpc      l6 /* If you change this, change winmacro.h header file */
#define t_systable   l7 /* Never touch this, could be the syscall table ptr. */
#define curptr       g6 /* Set after pt_regs frame is built */

        .text
        .align 4

        /* SEVEN WINDOW PATCH INSTRUCTIONS */
        .globl  tsetup_7win_patch1, tsetup_7win_patch2
        .globl  tsetup_7win_patch3, tsetup_7win_patch4
        .globl  tsetup_7win_patch5, tsetup_7win_patch6
tsetup_7win_patch1:     sll     %t_wim, 0x6, %t_wim
tsetup_7win_patch2:     and     %g2, 0x7f, %g2
tsetup_7win_patch3:     and     %g2, 0x7f, %g2
tsetup_7win_patch4:     and     %g1, 0x7f, %g1
tsetup_7win_patch5:     sll     %t_wim, 0x6, %t_wim
tsetup_7win_patch6:     and     %g2, 0x7f, %g2
        /* END OF PATCH INSTRUCTIONS */

        /* At trap time, interrupts and all generic traps do the
         * following:
         *
         * rd   %psr, %l0
         * b    some_handler
         * rd   %wim, %l3
         * nop
         *
         * Then 'some_handler' if it needs a trap frame (ie. it has
         * to call c-code and the trap cannot be handled in-window)
         * then it does the SAVE_ALL macro in entry.S which does
         *
         * sethi        %hi(trap_setup), %l4
         * jmpl         %l4 + %lo(trap_setup), %l6
         * nop
         */

        /* 2 3 4  window number
         * -----
         * O T S  mnemonic
         *
         * O == Current window before trap
         * T == Window entered when trap occurred
         * S == Window we will need to save if (1<<T) == %wim
         *
         * Before execution gets here, it must be guaranteed that
         * %l0 contains trap time %psr, %l1 and %l2 contain the
         * trap pc and npc, and %l3 contains the trap time %wim.
         */

        .globl  trap_setup, tsetup_patch1, tsetup_patch2
        .globl  tsetup_patch3, tsetup_patch4
        .globl  tsetup_patch5, tsetup_patch6
trap_setup:
        /* Calculate mask of trap window.  See if from user
         * or kernel and branch conditionally.
         */
        mov     1, %t_twinmask
        andcc   %t_psr, PSR_PS, %g0              ! fromsupv_p = (psr & PSR_PS)
        be      trap_setup_from_user             ! nope, from user mode
         sll    %t_twinmask, %t_psr, %t_twinmask ! t_twinmask = (1 << psr)

        /* From kernel, allocate more kernel stack and
         * build a pt_regs trap frame.
         */
        sub     %fp, (STACKFRAME_SZ + TRACEREG_SZ), %t_kstack
        STORE_PT_ALL(t_kstack, t_psr, t_pc, t_npc, g2)

        /* See if we are in the trap window. */
        andcc   %t_twinmask, %t_wim, %g0
        bne     trap_setup_kernel_spill         ! in trap window, clean up
         nop

        /* Trap from kernel with a window available.
         * Just do it...
         */
        jmpl    %t_retpc + 0x8, %g0     ! return to caller
         mov    %t_kstack, %sp          ! jump onto new stack

trap_setup_kernel_spill:
        ld      [%curptr + TI_UWINMASK], %g1
        orcc    %g0, %g1, %g0
        bne     trap_setup_user_spill   ! there are some user windows, yuck
        /* Spill from kernel, but only kernel windows, adjust
         * %wim and go.
         */
         srl    %t_wim, 0x1, %g2        ! begin computation of new %wim
tsetup_patch1:
        sll     %t_wim, 0x7, %t_wim     ! patched on 7 window Sparcs
        or      %t_wim, %g2, %g2
tsetup_patch2:
        and     %g2, 0xff, %g2          ! patched on 7 window Sparcs

        save    %g0, %g0, %g0

        /* Set new %wim value */
        wr      %g2, 0x0, %wim

        /* Save the kernel window onto the corresponding stack. */
        STORE_WINDOW(sp)

        restore %g0, %g0, %g0

        jmpl    %t_retpc + 0x8, %g0     ! return to caller
         mov    %t_kstack, %sp          ! and onto new kernel stack

#define STACK_OFFSET (THREAD_SIZE - TRACEREG_SZ - STACKFRAME_SZ)

trap_setup_from_user:
        /* We can't use %curptr yet. */
        LOAD_CURRENT(t_kstack, t_twinmask)

        sethi   %hi(STACK_OFFSET), %t_twinmask
        or      %t_twinmask, %lo(STACK_OFFSET), %t_twinmask
        add     %t_kstack, %t_twinmask, %t_kstack

        mov     1, %t_twinmask
        sll     %t_twinmask, %t_psr, %t_twinmask ! t_twinmask = (1 << psr)

        /* Build pt_regs frame. */
        STORE_PT_ALL(t_kstack, t_psr, t_pc, t_npc, g2)

#if 0
        /* If we're sure every task_struct is THREAD_SIZE aligned,
           we can speed this up. */
        sethi   %hi(STACK_OFFSET), %curptr
        or      %curptr, %lo(STACK_OFFSET), %curptr
        sub     %t_kstack, %curptr, %curptr
#else
        sethi   %hi(~(THREAD_SIZE - 1)), %curptr
        and     %t_kstack, %curptr, %curptr
#endif

        /* Clear current_thread_info->w_saved */
        st      %g0, [%curptr + TI_W_SAVED]

        /* See if we are in the trap window. */
        andcc   %t_twinmask, %t_wim, %g0
        bne     trap_setup_user_spill           ! yep we are
         orn    %g0, %t_twinmask, %g1           ! negate trap win mask into %g1

        /* Trap from user, but not into the invalid window.
         * Calculate new umask.  The way this works is,
         * any window from the %wim at trap time until
         * the window right before the one we are in now,
         * is a user window.  A diagram:
         *
         *      7 6 5 4 3 2 1 0    window number
         *      ---------------
         *        I     L T        mnemonic
         *
         * Window 'I' is the invalid window in our example,
         * window 'L' is the window the user was in when
         * the trap occurred, window T is the trap window
         * we are in now.  So therefore, windows 5, 4 and
         * 3 are user windows.  The following sequence
         * computes the user winmask to represent this.
         */
        subcc   %t_wim, %t_twinmask, %g2
        bneg,a  1f
         sub    %g2, 0x1, %g2
1:
        andn    %g2, %t_twinmask, %g2
tsetup_patch3:
        and     %g2, 0xff, %g2                  ! patched on 7win Sparcs
        st      %g2, [%curptr + TI_UWINMASK]    ! store new umask

        jmpl    %t_retpc + 0x8, %g0             ! return to caller
         mov    %t_kstack, %sp                  ! and onto kernel stack

trap_setup_user_spill:
        /* A spill occurred from either kernel or user mode
         * and there exist some user windows to deal with.
         * A mask of the currently valid user windows
         * is in %g1 upon entry to here.
         */

tsetup_patch4:
        and     %g1, 0xff, %g1          ! patched on 7win Sparcs, mask
        srl     %t_wim, 0x1, %g2        ! compute new %wim
tsetup_patch5:
        sll     %t_wim, 0x7, %t_wim     ! patched on 7win Sparcs
        or      %t_wim, %g2, %g2        ! %g2 is new %wim
tsetup_patch6:
        and     %g2, 0xff, %g2          ! patched on 7win Sparcs
        andn    %g1, %g2, %g1           ! clear this bit in %g1
        st      %g1, [%curptr + TI_UWINMASK]

        save    %g0, %g0, %g0

        wr      %g2, 0x0, %wim

        /* Call MMU-architecture dependent stack checking
         * routine.
         */
        .globl  tsetup_mmu_patchme
tsetup_mmu_patchme:
        b       tsetup_sun4c_stackchk
         andcc  %sp, 0x7, %g0

        /* Architecture specific stack checking routines.  When either
         * of these routines are called, the globals are free to use
         * as they have been safely stashed on the new kernel stack
         * pointer.  Thus the definition below for simplicity.
         */
#define glob_tmp     g1

        .globl  tsetup_sun4c_stackchk
tsetup_sun4c_stackchk:
        /* Done by caller: andcc %sp, 0x7, %g0 */
        bne     trap_setup_user_stack_is_bolixed
         sra    %sp, 29, %glob_tmp

        add     %glob_tmp, 0x1, %glob_tmp
        andncc  %glob_tmp, 0x1, %g0
        bne     trap_setup_user_stack_is_bolixed
         and    %sp, 0xfff, %glob_tmp           ! delay slot

        /* See if our dump area will be on more than one
         * page.
         */
        add     %glob_tmp, 0x38, %glob_tmp
        andncc  %glob_tmp, 0xff8, %g0
        be      tsetup_sun4c_onepage            ! only one page to check
         lda    [%sp] ASI_PTE, %glob_tmp        ! have to check first page anyways

tsetup_sun4c_twopages:
        /* Is first page ok permission wise? */
        srl     %glob_tmp, 29, %glob_tmp
        cmp     %glob_tmp, 0x6
        bne     trap_setup_user_stack_is_bolixed
         add    %sp, 0x38, %glob_tmp            /* Is second page in vma hole? */

        sra     %glob_tmp, 29, %glob_tmp
        add     %glob_tmp, 0x1, %glob_tmp
        andncc  %glob_tmp, 0x1, %g0
        bne     trap_setup_user_stack_is_bolixed
         add    %sp, 0x38, %glob_tmp

        lda     [%glob_tmp] ASI_PTE, %glob_tmp

tsetup_sun4c_onepage:
        srl     %glob_tmp, 29, %glob_tmp
        cmp     %glob_tmp, 0x6                          ! can user write to it?
        bne     trap_setup_user_stack_is_bolixed        ! failure
         nop

        STORE_WINDOW(sp)

        restore %g0, %g0, %g0

        jmpl    %t_retpc + 0x8, %g0
         mov    %t_kstack, %sp

        .globl  tsetup_srmmu_stackchk
tsetup_srmmu_stackchk:
        /* Check results of callers andcc %sp, 0x7, %g0 */
        bne     trap_setup_user_stack_is_bolixed
         sethi   %hi(PAGE_OFFSET), %glob_tmp

        cmp     %glob_tmp, %sp
        bleu,a  1f
         lda    [%g0] ASI_M_MMUREGS, %glob_tmp          ! read MMU control

trap_setup_user_stack_is_bolixed:
        /* From user/kernel into invalid window w/bad user
         * stack. Save bad user stack, and return to caller.
         */
        SAVE_BOLIXED_USER_STACK(curptr, g3)
        restore %g0, %g0, %g0

        jmpl    %t_retpc + 0x8, %g0
         mov    %t_kstack, %sp

1:
        /* Clear the fault status and turn on the no_fault bit. */
        or      %glob_tmp, 0x2, %glob_tmp               ! or in no_fault bit
        sta     %glob_tmp, [%g0] ASI_M_MMUREGS          ! set it

        /* Dump the registers and cross fingers. */
        STORE_WINDOW(sp)

        /* Clear the no_fault bit and check the status. */
        andn    %glob_tmp, 0x2, %glob_tmp
        sta     %glob_tmp, [%g0] ASI_M_MMUREGS
        mov     AC_M_SFAR, %glob_tmp
        lda     [%glob_tmp] ASI_M_MMUREGS, %g0
        mov     AC_M_SFSR, %glob_tmp
        lda     [%glob_tmp] ASI_M_MMUREGS, %glob_tmp    ! save away status of winstore
        andcc   %glob_tmp, 0x2, %g0                     ! did we fault?
        bne     trap_setup_user_stack_is_bolixed        ! failure
         nop

        restore %g0, %g0, %g0

        jmpl    %t_retpc + 0x8, %g0
         mov    %t_kstack, %sp

Compare with Previous | Blame | View Log

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.