blob: 3ec4d40118864d483d74ec0b6f05d57b36deb323 [file] [log] [blame]
/*
* Copyright (c) 2003-2006 The Regents of The University of Michigan
* Copyright (c) 1992-1995 Hewlett-Packard Development Company
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met: redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer;
* redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution;
* neither the name of the copyright holders nor the names of its
* contributors may be used to endorse or promote products derived from
* this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
* Authors: Ali G. Saidi
* Nathan L. Binkert
*/
// modified to use the Hudson style "impure.h" instead of ev5_impure.sdl
// since we don't have a mechanism to expand the data structures.... pb Nov/95
#include "ev5_defs.h"
#include "ev5_impure.h"
#include "ev5_alpha_defs.h"
#include "ev5_paldef.h"
#include "ev5_osfalpha_defs.h"
#include "fromHudsonMacros.h"
#include "fromHudsonOsf.h"
#include "dc21164FromGasSources.h"
#define DEBUGSTORE(c) nop
#define DEBUG_EXC_ADDR()\
bsr r25, put_exc_addr; \
DEBUGSTORE(13) ; \
DEBUGSTORE(10)
// This is the fix for the user-mode super page references causing the
// machine to crash.
#define hw_rei_spe hw_rei
#define vmaj 1
#define vmin 18
#define vms_pal 1
#define osf_pal 2
#define pal_type osf_pal
#define osfpal_version_l ((pal_type<<16) | (vmaj<<8) | (vmin<<0))
///////////////////////////
// PALtemp register usage
///////////////////////////
// The EV5 Ibox holds 24 PALtemp registers. This maps the OSF PAL usage
// for these PALtemps:
//
// pt0 local scratch
// pt1 local scratch
// pt2 entUna pt_entUna
// pt3 CPU specific impure area pointer pt_impure
// pt4 memory management temp
// pt5 memory management temp
// pt6 memory management temp
// pt7 entIF pt_entIF
// pt8 intmask pt_intmask
// pt9 entSys pt_entSys
// pt10
// pt11 entInt pt_entInt
// pt12 entArith pt_entArith
// pt13 reserved for system specific PAL
// pt14 reserved for system specific PAL
// pt15 reserved for system specific PAL
// pt16 MISC: scratch ! WHAMI<7:0> ! 0 0 0 MCES<4:0> pt_misc, pt_whami,
// pt_mces
// pt17 sysval pt_sysval
// pt18 usp pt_usp
// pt19 ksp pt_ksp
// pt20 PTBR pt_ptbr
// pt21 entMM pt_entMM
// pt22 kgp pt_kgp
// pt23 PCBB pt_pcbb
//
//
/////////////////////////////
// PALshadow register usage
/////////////////////////////
//
// EV5 shadows R8-R14 and R25 when in PALmode and ICSR<shadow_enable> = 1.
// This maps the OSF PAL usage of R8 - R14 and R25:
//
// r8 ITBmiss/DTBmiss scratch
// r9 ITBmiss/DTBmiss scratch
// r10 ITBmiss/DTBmiss scratch
// r11 PS
// r12 local scratch
// r13 local scratch
// r14 local scratch
// r25 local scratch
//
// .sbttl "PALcode configuration options"
// There are a number of options that may be assembled into this version of
// PALcode. They should be adjusted in a prefix assembly file (i.e. do not edit
// the following). The options that can be adjusted cause the resultant PALcode
// to reflect the desired target system.
// multiprocessor support can be enabled for a max of n processors by
// setting the following to the number of processors on the system.
// Note that this is really the max cpuid.
#define max_cpuid 1
#ifndef max_cpuid
#define max_cpuid 8
#endif
#define osf_svmin 1
#define osfpal_version_h ((max_cpuid<<16) | (osf_svmin<<0))
//
// RESET - Reset Trap Entry Point
//
// RESET - offset 0000
// Entry:
// Vectored into via hardware trap on reset, or branched to
// on swppal.
//
// r0 = whami
// r1 = pal_base
// r2 = base of scratch area
// r3 = halt code
//
//
// Function:
//
//
.text 0
. = 0x0000
.globl _start
.globl Pal_Base
_start:
Pal_Base:
HDW_VECTOR(PAL_RESET_ENTRY)
Trap_Reset:
nop
/*
* store into r1
*/
br r1,sys_reset
// Specify PAL version info as a constant
// at a known location (reset + 8).
.long osfpal_version_l // <pal_type@16> ! <vmaj@8> ! <vmin@0>
.long osfpal_version_h // <max_cpuid@16> ! <osf_svmin@0>
.long 0
.long 0
pal_impure_start:
.quad 0
pal_debug_ptr:
.quad 0 // reserved for debug pointer ; 20
//
// IACCVIO - Istream Access Violation Trap Entry Point
//
// IACCVIO - offset 0080
// Entry:
// Vectored into via hardware trap on Istream access violation or sign check error on PC.
//
// Function:
// Build stack frame
// a0 <- Faulting VA
// a1 <- MMCSR (1 for ACV)
// a2 <- -1 (for ifetch fault)
// vector via entMM
//
HDW_VECTOR(PAL_IACCVIO_ENTRY)
Trap_Iaccvio:
DEBUGSTORE(0x42)
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
bis r11, r31, r12 // Save PS
bge r25, TRAP_IACCVIO_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r12 // Set new PS
mfpr r30, pt_ksp
TRAP_IACCVIO_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
mfpr r14, exc_addr // get pc
stq r16, osfsf_a0(sp) // save regs
bic r14, 3, r16 // pass pc/va as a0
stq r17, osfsf_a1(sp) // a1
or r31, mmcsr_c_acv, r17 // pass mm_csr as a1
stq r18, osfsf_a2(sp) // a2
mfpr r13, pt_entmm // get entry point
stq r11, osfsf_ps(sp) // save old ps
bis r12, r31, r11 // update ps
stq r16, osfsf_pc(sp) // save pc
stq r29, osfsf_gp(sp) // save gp
mtpr r13, exc_addr // load exc_addr with entMM
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kgp
subq r31, 1, r18 // pass flag of istream, as a2
hw_rei_spe
//
// INTERRUPT - Interrupt Trap Entry Point
//
// INTERRUPT - offset 0100
// Entry:
// Vectored into via trap on hardware interrupt
//
// Function:
// check for halt interrupt
// check for passive release (current ipl geq requestor)
// if necessary, switch to kernel mode push stack frame,
// update ps (including current mode and ipl copies), sp, and gp
// pass the interrupt info to the system module
//
//
HDW_VECTOR(PAL_INTERRUPT_ENTRY)
Trap_Interrupt:
mfpr r13, ev5__intid // Fetch level of interruptor
mfpr r25, ev5__isr // Fetch interrupt summary register
srl r25, isr_v_hlt, r9 // Get HLT bit
mfpr r14, ev5__ipl
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kern
blbs r9, sys_halt_interrupt // halt_interrupt if HLT bit set
cmple r13, r14, r8 // R8 = 1 if intid .less than or eql. ipl
bne r8, sys_passive_release // Passive release is current rupt is lt or eq ipl
and r11, osfps_m_mode, r10 // get mode bit
beq r10, TRAP_INTERRUPT_10_ // Skip stack swap in kernel
mtpr r30, pt_usp // save user stack
mfpr r30, pt_ksp // get kern stack
TRAP_INTERRUPT_10_:
lda sp, (0-osfsf_c_size)(sp)// allocate stack space
mfpr r14, exc_addr // get pc
stq r11, osfsf_ps(sp) // save ps
stq r14, osfsf_pc(sp) // save pc
stq r29, osfsf_gp(sp) // push gp
stq r16, osfsf_a0(sp) // a0
// pvc_violate 354 // ps is cleared anyway, if store to stack faults.
mtpr r31, ev5__ps // Set Ibox current mode to kernel
stq r17, osfsf_a1(sp) // a1
stq r18, osfsf_a2(sp) // a2
subq r13, 0x11, r12 // Start to translate from EV5IPL->OSFIPL
srl r12, 1, r8 // 1d, 1e: ipl 6. 1f: ipl 7.
subq r13, 0x1d, r9 // Check for 1d, 1e, 1f
cmovge r9, r8, r12 // if .ge. 1d, then take shifted value
bis r12, r31, r11 // set new ps
mfpr r12, pt_intmask
and r11, osfps_m_ipl, r14 // Isolate just new ipl (not really needed, since all non-ipl bits zeroed already)
/*
* Lance had space problems. We don't.
*/
extbl r12, r14, r14 // Translate new OSFIPL->EV5IPL
mfpr r29, pt_kgp // update gp
mtpr r14, ev5__ipl // load the new IPL into Ibox
br r31, sys_interrupt // Go handle interrupt
//
// ITBMISS - Istream TBmiss Trap Entry Point
//
// ITBMISS - offset 0180
// Entry:
// Vectored into via hardware trap on Istream translation buffer miss.
//
// Function:
// Do a virtual fetch of the PTE, and fill the ITB if the PTE is valid.
// Can trap into DTBMISS_DOUBLE.
// This routine can use the PALshadow registers r8, r9, and r10
//
//
HDW_VECTOR(PAL_ITB_MISS_ENTRY)
Trap_Itbmiss:
// Real MM mapping
nop
mfpr r8, ev5__ifault_va_form // Get virtual address of PTE.
nop
mfpr r10, exc_addr // Get PC of faulting instruction in case of DTBmiss.
pal_itb_ldq:
ld_vpte r8, 0(r8) // Get PTE, traps to DTBMISS_DOUBLE in case of TBmiss
mtpr r10, exc_addr // Restore exc_address if there was a trap.
mfpr r31, ev5__va // Unlock VA in case there was a double miss
nop
and r8, osfpte_m_foe, r25 // Look for FOE set.
blbc r8, invalid_ipte_handler // PTE not valid.
nop
bne r25, foe_ipte_handler // FOE is set
nop
mtpr r8, ev5__itb_pte // Ibox remembers the VA, load the PTE into the ITB.
hw_rei_stall //
//
// DTBMISS_SINGLE - Dstream Single TBmiss Trap Entry Point
//
// DTBMISS_SINGLE - offset 0200
// Entry:
// Vectored into via hardware trap on Dstream single translation
// buffer miss.
//
// Function:
// Do a virtual fetch of the PTE, and fill the DTB if the PTE is valid.
// Can trap into DTBMISS_DOUBLE.
// This routine can use the PALshadow registers r8, r9, and r10
//
HDW_VECTOR(PAL_DTB_MISS_ENTRY)
Trap_Dtbmiss_Single:
mfpr r8, ev5__va_form // Get virtual address of PTE - 1 cycle delay. E0.
mfpr r10, exc_addr // Get PC of faulting instruction in case of error. E1.
// DEBUGSTORE(0x45)
// DEBUG_EXC_ADDR()
// Real MM mapping
mfpr r9, ev5__mm_stat // Get read/write bit. E0.
mtpr r10, pt6 // Stash exc_addr away
pal_dtb_ldq:
ld_vpte r8, 0(r8) // Get PTE, traps to DTBMISS_DOUBLE in case of TBmiss
nop // Pad MF VA
mfpr r10, ev5__va // Get original faulting VA for TB load. E0.
nop
mtpr r8, ev5__dtb_pte // Write DTB PTE part. E0.
blbc r8, invalid_dpte_handler // Handle invalid PTE
mtpr r10, ev5__dtb_tag // Write DTB TAG part, completes DTB load. No virt ref for 3 cycles.
mfpr r10, pt6
// Following 2 instructions take 2 cycles
mtpr r10, exc_addr // Return linkage in case we trapped. E1.
mfpr r31, pt0 // Pad the write to dtb_tag
hw_rei // Done, return
//
// DTBMISS_DOUBLE - Dstream Double TBmiss Trap Entry Point
//
//
// DTBMISS_DOUBLE - offset 0280
// Entry:
// Vectored into via hardware trap on Double TBmiss from single
// miss flows.
//
// r8 - faulting VA
// r9 - original MMstat
// r10 - original exc_addr (both itb,dtb miss)
// pt6 - original exc_addr (dtb miss flow only)
// VA IPR - locked with original faulting VA
//
// Function:
// Get PTE, if valid load TB and return.
// If not valid then take TNV/ACV exception.
//
// pt4 and pt5 are reserved for this flow.
//
//
//
HDW_VECTOR(PAL_DOUBLE_MISS_ENTRY)
Trap_Dtbmiss_double:
mtpr r8, pt4 // save r8 to do exc_addr check
mfpr r8, exc_addr
blbc r8, Trap_Dtbmiss_Single //if not in palmode, should be in the single routine, dummy!
mfpr r8, pt4 // restore r8
nop
mtpr r22, pt5 // Get some scratch space. E1.
// Due to virtual scheme, we can skip the first lookup and go
// right to fetch of level 2 PTE
sll r8, (64-((2*page_seg_size_bits)+page_offset_size_bits)), r22 // Clean off upper bits of VA
mtpr r21, pt4 // Get some scratch space. E1.
srl r22, 61-page_seg_size_bits, r22 // Get Va<seg1>*8
mfpr r21, pt_ptbr // Get physical address of the page table.
nop
addq r21, r22, r21 // Index into page table for level 2 PTE.
sll r8, (64-((1*page_seg_size_bits)+page_offset_size_bits)), r22 // Clean off upper bits of VA
ldq_p r21, 0(r21) // Get level 2 PTE (addr<2:0> ignored)
srl r22, 61-page_seg_size_bits, r22 // Get Va<seg1>*8
blbc r21, double_pte_inv // Check for Invalid PTE.
srl r21, 32, r21 // extract PFN from PTE
sll r21, page_offset_size_bits, r21 // get PFN * 2^13 for add to <seg3>*8
addq r21, r22, r21 // Index into page table for level 3 PTE.
nop
ldq_p r21, 0(r21) // Get level 3 PTE (addr<2:0> ignored)
blbc r21, double_pte_inv // Check for invalid PTE.
mtpr r21, ev5__dtb_pte // Write the PTE. E0.
mfpr r22, pt5 // Restore scratch register
mtpr r8, ev5__dtb_tag // Write the TAG. E0. No virtual references in subsequent 3 cycles.
mfpr r21, pt4 // Restore scratch register
nop // Pad write to tag.
nop
nop // Pad write to tag.
nop
hw_rei
//
// UNALIGN -- Dstream unalign trap
//
// UNALIGN - offset 0300
// Entry:
// Vectored into via hardware trap on unaligned Dstream reference.
//
// Function:
// Build stack frame
// a0 <- Faulting VA
// a1 <- Opcode
// a2 <- src/dst register number
// vector via entUna
//
HDW_VECTOR(PAL_UNALIGN_ENTRY)
Trap_Unalign:
/* DEBUGSTORE(0x47)*/
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
mfpr r8, ev5__mm_stat // Get mmstat --ok to use r8, no tbmiss
mfpr r14, exc_addr // get pc
srl r8, mm_stat_v_ra, r13 // Shift Ra field to ls bits
blbs r14, pal_pal_bug_check // Bugcheck if unaligned in PAL
blbs r8, UNALIGN_NO_DISMISS // lsb only set on store or fetch_m
// not set, must be a load
and r13, 0x1F, r8 // isolate ra
cmpeq r8, 0x1F, r8 // check for r31/F31
bne r8, dfault_fetch_ldr31_err // if its a load to r31 or f31 -- dismiss the fault
UNALIGN_NO_DISMISS:
bis r11, r31, r12 // Save PS
bge r25, UNALIGN_NO_DISMISS_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r12 // Set new PS
mfpr r30, pt_ksp
UNALIGN_NO_DISMISS_10_:
mfpr r25, ev5__va // Unlock VA
lda sp, 0-osfsf_c_size(sp)// allocate stack space
mtpr r25, pt0 // Stash VA
stq r18, osfsf_a2(sp) // a2
stq r11, osfsf_ps(sp) // save old ps
srl r13, mm_stat_v_opcode-mm_stat_v_ra, r25// Isolate opcode
stq r29, osfsf_gp(sp) // save gp
addq r14, 4, r14 // inc PC past the ld/st
stq r17, osfsf_a1(sp) // a1
and r25, mm_stat_m_opcode, r17// Clean opocde for a1
stq r16, osfsf_a0(sp) // save regs
mfpr r16, pt0 // a0 <- va/unlock
stq r14, osfsf_pc(sp) // save pc
mfpr r25, pt_entuna // get entry point
bis r12, r31, r11 // update ps
br r31, unalign_trap_cont
//
// DFAULT - Dstream Fault Trap Entry Point
//
// DFAULT - offset 0380
// Entry:
// Vectored into via hardware trap on dstream fault or sign check
// error on DVA.
//
// Function:
// Ignore faults on FETCH/FETCH_M
// Check for DFAULT in PAL
// Build stack frame
// a0 <- Faulting VA
// a1 <- MMCSR (1 for ACV, 2 for FOR, 4 for FOW)
// a2 <- R/W
// vector via entMM
//
//
HDW_VECTOR(PAL_D_FAULT_ENTRY)
Trap_Dfault:
// DEBUGSTORE(0x48)
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
mfpr r13, ev5__mm_stat // Get mmstat
mfpr r8, exc_addr // get pc, preserve r14
srl r13, mm_stat_v_opcode, r9 // Shift opcode field to ls bits
blbs r8, dfault_in_pal
bis r8, r31, r14 // move exc_addr to correct place
bis r11, r31, r12 // Save PS
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
and r9, mm_stat_m_opcode, r9 // Clean all but opcode
cmpeq r9, evx_opc_sync, r9 // Is the opcode fetch/fetchm?
bne r9, dfault_fetch_ldr31_err // Yes, dismiss the fault
//dismiss exception if load to r31/f31
blbs r13, dfault_no_dismiss // mm_stat<0> set on store or fetchm
// not a store or fetch, must be a load
srl r13, mm_stat_v_ra, r9 // Shift rnum to low bits
and r9, 0x1F, r9 // isolate rnum
nop
cmpeq r9, 0x1F, r9 // Is the rnum r31 or f31?
bne r9, dfault_fetch_ldr31_err // Yes, dismiss the fault
dfault_no_dismiss:
and r13, 0xf, r13 // Clean extra bits in mm_stat
bge r25, dfault_trap_cont // no stack swap needed if cm=kern
mtpr r30, pt_usp // save user stack
bis r31, r31, r12 // Set new PS
mfpr r30, pt_ksp
br r31, dfault_trap_cont
//
// MCHK - Machine Check Trap Entry Point
//
// MCHK - offset 0400
// Entry:
// Vectored into via hardware trap on machine check.
//
// Function:
//
//
HDW_VECTOR(PAL_MCHK_ENTRY)
Trap_Mchk:
DEBUGSTORE(0x49)
mtpr r31, ic_flush_ctl // Flush the Icache
br r31, sys_machine_check
//
// OPCDEC - Illegal Opcode Trap Entry Point
//
// OPCDEC - offset 0480
// Entry:
// Vectored into via hardware trap on illegal opcode.
//
// Build stack frame
// a0 <- code
// a1 <- unpred
// a2 <- unpred
// vector via entIF
//
//
HDW_VECTOR(PAL_OPCDEC_ENTRY)
Trap_Opcdec:
DEBUGSTORE(0x4a)
//simos DEBUG_EXC_ADDR()
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
mfpr r14, exc_addr // get pc
blbs r14, pal_pal_bug_check // check opcdec in palmode
bis r11, r31, r12 // Save PS
bge r25, TRAP_OPCDEC_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r12 // Set new PS
mfpr r30, pt_ksp
TRAP_OPCDEC_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
addq r14, 4, r14 // inc pc
stq r16, osfsf_a0(sp) // save regs
bis r31, osf_a0_opdec, r16 // set a0
stq r11, osfsf_ps(sp) // save old ps
mfpr r13, pt_entif // get entry point
stq r18, osfsf_a2(sp) // a2
stq r17, osfsf_a1(sp) // a1
stq r29, osfsf_gp(sp) // save gp
stq r14, osfsf_pc(sp) // save pc
bis r12, r31, r11 // update ps
mtpr r13, exc_addr // load exc_addr with entIF
// 1 cycle to hw_rei, E1
mfpr r29, pt_kgp // get the kgp, E1
hw_rei_spe // done, E1
//
// ARITH - Arithmetic Exception Trap Entry Point
//
// ARITH - offset 0500
// Entry:
// Vectored into via hardware trap on arithmetic excpetion.
//
// Function:
// Build stack frame
// a0 <- exc_sum
// a1 <- exc_mask
// a2 <- unpred
// vector via entArith
//
//
HDW_VECTOR(PAL_ARITH_ENTRY)
Trap_Arith:
DEBUGSTORE(0x4b)
and r11, osfps_m_mode, r12 // get mode bit
mfpr r31, ev5__va // unlock mbox
bis r11, r31, r25 // save ps
mfpr r14, exc_addr // get pc
nop
blbs r14, pal_pal_bug_check // arith trap from PAL
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
beq r12, TRAP_ARITH_10_ // if zero we are in kern now
bis r31, r31, r25 // set the new ps
mtpr r30, pt_usp // save user stack
nop
mfpr r30, pt_ksp // get kern stack
TRAP_ARITH_10_: lda sp, 0-osfsf_c_size(sp) // allocate stack space
mtpr r31, ev5__ps // Set Ibox current mode to kernel
nop // Pad current mode write and stq
mfpr r13, ev5__exc_sum // get the exc_sum
mfpr r12, pt_entarith
stq r14, osfsf_pc(sp) // save pc
stq r17, osfsf_a1(sp)
mfpr r17, ev5__exc_mask // Get exception register mask IPR - no mtpr exc_sum in next cycle
stq r11, osfsf_ps(sp) // save ps
bis r25, r31, r11 // set new ps
stq r16, osfsf_a0(sp) // save regs
srl r13, exc_sum_v_swc, r16 // shift data to correct position
stq r18, osfsf_a2(sp)
// pvc_violate 354 // ok, but make sure reads of exc_mask/sum are not in same trap shadow
mtpr r31, ev5__exc_sum // Unlock exc_sum and exc_mask
stq r29, osfsf_gp(sp)
mtpr r12, exc_addr // Set new PC - 1 bubble to hw_rei - E1
mfpr r29, pt_kgp // get the kern gp - E1
hw_rei_spe // done - E1
//
// FEN - Illegal Floating Point Operation Trap Entry Point
//
// FEN - offset 0580
// Entry:
// Vectored into via hardware trap on illegal FP op.
//
// Function:
// Build stack frame
// a0 <- code
// a1 <- unpred
// a2 <- unpred
// vector via entIF
//
//
HDW_VECTOR(PAL_FEN_ENTRY)
Trap_Fen:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
mfpr r14, exc_addr // get pc
blbs r14, pal_pal_bug_check // check opcdec in palmode
mfpr r13, ev5__icsr
nop
bis r11, r31, r12 // Save PS
bge r25, TRAP_FEN_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r12 // Set new PS
mfpr r30, pt_ksp
TRAP_FEN_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
srl r13, icsr_v_fpe, r25 // Shift FP enable to bit 0
stq r16, osfsf_a0(sp) // save regs
mfpr r13, pt_entif // get entry point
stq r18, osfsf_a2(sp) // a2
stq r11, osfsf_ps(sp) // save old ps
stq r29, osfsf_gp(sp) // save gp
bis r12, r31, r11 // set new ps
stq r17, osfsf_a1(sp) // a1
blbs r25,fen_to_opcdec // If FP is enabled, this is really OPCDEC.
bis r31, osf_a0_fen, r16 // set a0
stq r14, osfsf_pc(sp) // save pc
mtpr r13, exc_addr // load exc_addr with entIF
// 1 cycle to hw_rei -E1
mfpr r29, pt_kgp // get the kgp -E1
hw_rei_spe // done -E1
// FEN trap was taken, but the fault is really opcdec.
ALIGN_BRANCH
fen_to_opcdec:
addq r14, 4, r14 // save PC+4
bis r31, osf_a0_opdec, r16 // set a0
stq r14, osfsf_pc(sp) // save pc
mtpr r13, exc_addr // load exc_addr with entIF
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kgp
hw_rei_spe // done
//////////////////////////////////////////////////////////////////////////////
// Misc handlers - Start area for misc code.
//////////////////////////////////////////////////////////////////////////////
//
// dfault_trap_cont
// A dfault trap has been taken. The sp has been updated if necessary.
// Push a stack frame a vector via entMM.
//
// Current state:
// r12 - new PS
// r13 - MMstat
// VA - locked
//
//
ALIGN_BLOCK
dfault_trap_cont:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
mfpr r25, ev5__va // Fetch VA/unlock
stq r18, osfsf_a2(sp) // a2
and r13, 1, r18 // Clean r/w bit for a2
stq r16, osfsf_a0(sp) // save regs
bis r25, r31, r16 // a0 <- va
stq r17, osfsf_a1(sp) // a1
srl r13, 1, r17 // shift fault bits to right position
stq r11, osfsf_ps(sp) // save old ps
bis r12, r31, r11 // update ps
stq r14, osfsf_pc(sp) // save pc
mfpr r25, pt_entmm // get entry point
stq r29, osfsf_gp(sp) // save gp
cmovlbs r17, 1, r17 // a2. acv overrides fox.
mtpr r25, exc_addr // load exc_addr with entMM
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kgp
hw_rei_spe // done
//
//unalign_trap_cont
// An unalign trap has been taken. Just need to finish up a few things.
//
// Current state:
// r25 - entUna
// r13 - shifted MMstat
//
//
ALIGN_BLOCK
unalign_trap_cont:
mtpr r25, exc_addr // load exc_addr with entUna
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kgp
and r13, mm_stat_m_ra, r18 // Clean Ra for a2
hw_rei_spe // done
//
// dfault_in_pal
// Dfault trap was taken, exc_addr points to a PAL PC.
// r9 - mmstat<opcode> right justified
// r8 - exception address
//
// These are the cases:
// opcode was STQ -- from a stack builder, KSP not valid halt
// r14 - original exc_addr
// r11 - original PS
// opcode was STL_C -- rti or retsys clear lock_flag by stack write,
// KSP not valid halt
// r11 - original PS
// r14 - original exc_addr
// opcode was LDQ -- retsys or rti stack read, KSP not valid halt
// r11 - original PS
// r14 - original exc_addr
// opcode was HW_LD -- itbmiss or dtbmiss, bugcheck due to fault on page tables
// r10 - original exc_addr
// r11 - original PS
//
//
//
ALIGN_BLOCK
dfault_in_pal:
DEBUGSTORE(0x50)
bic r8, 3, r8 // Clean PC
mfpr r9, pal_base
mfpr r31, va // unlock VA
// if not real_mm, should never get here from miss flows
subq r9, r8, r8 // pal_base - offset
lda r9, pal_itb_ldq-pal_base(r8)
nop
beq r9, dfault_do_bugcheck
lda r9, pal_dtb_ldq-pal_base(r8)
beq r9, dfault_do_bugcheck
//
// KSP invalid halt case --
ksp_inval_halt:
DEBUGSTORE(76)
bic r11, osfps_m_mode, r11 // set ps to kernel mode
mtpr r0, pt0
mtpr r31, dtb_cm // Make sure that the CM IPRs are all kernel mode
mtpr r31, ips
mtpr r14, exc_addr // Set PC to instruction that caused trouble
bsr r0, pal_update_pcb // update the pcb
lda r0, hlt_c_ksp_inval(r31) // set halt code to hw halt
br r31, sys_enter_console // enter the console
ALIGN_BRANCH
dfault_do_bugcheck:
bis r10, r31, r14 // bugcheck expects exc_addr in r14
br r31, pal_pal_bug_check
//
// dfault_fetch_ldr31_err - ignore faults on fetch(m) and loads to r31/f31
// On entry -
// r14 - exc_addr
// VA is locked
//
//
ALIGN_BLOCK
dfault_fetch_ldr31_err:
mtpr r11, ev5__dtb_cm
mtpr r11, ev5__ps // Make sure ps hasn't changed
mfpr r31, va // unlock the mbox
addq r14, 4, r14 // inc the pc to skip the fetch
mtpr r14, exc_addr // give ibox new PC
mfpr r31, pt0 // pad exc_addr write
hw_rei
ALIGN_BLOCK
//
// sys_from_kern
// callsys from kernel mode - OS bugcheck machine check
//
//
sys_from_kern:
mfpr r14, exc_addr // PC points to call_pal
subq r14, 4, r14
lda r25, mchk_c_os_bugcheck(r31) // fetch mchk code
br r31, pal_pal_mchk
// Continuation of long call_pal flows
//
// wrent_tbl
// Table to write *int in paltemps.
// 4 instructions/entry
// r16 has new value
//
//
ALIGN_BLOCK
wrent_tbl:
//orig pvc_jsr wrent, dest=1
nop
mtpr r16, pt_entint
mfpr r31, pt0 // Pad for mt->mf paltemp rule
hw_rei
//orig pvc_jsr wrent, dest=1
nop
mtpr r16, pt_entarith
mfpr r31, pt0 // Pad for mt->mf paltemp rule
hw_rei
//orig pvc_jsr wrent, dest=1
nop
mtpr r16, pt_entmm
mfpr r31, pt0 // Pad for mt->mf paltemp rule
hw_rei
//orig pvc_jsr wrent, dest=1
nop
mtpr r16, pt_entif
mfpr r31, pt0 // Pad for mt->mf paltemp rule
hw_rei
//orig pvc_jsr wrent, dest=1
nop
mtpr r16, pt_entuna
mfpr r31, pt0 // Pad for mt->mf paltemp rule
hw_rei
//orig pvc_jsr wrent, dest=1
nop
mtpr r16, pt_entsys
mfpr r31, pt0 // Pad for mt->mf paltemp rule
hw_rei
ALIGN_BLOCK
//
// tbi_tbl
// Table to do tbi instructions
// 4 instructions per entry
//
tbi_tbl:
// -2 tbia
//orig pvc_jsr tbi, dest=1
mtpr r31, ev5__dtb_ia // Flush DTB
mtpr r31, ev5__itb_ia // Flush ITB
hw_rei_stall
nop // Pad table
// -1 tbiap
//orig pvc_jsr tbi, dest=1
mtpr r31, ev5__dtb_iap // Flush DTB
mtpr r31, ev5__itb_iap // Flush ITB
hw_rei_stall
nop // Pad table
// 0 unused
//orig pvc_jsr tbi, dest=1
hw_rei // Pad table
nop
nop
nop
// 1 tbisi
//orig pvc_jsr tbi, dest=1
nop
nop
mtpr r17, ev5__itb_is // Flush ITB
hw_rei_stall
// 2 tbisd
//orig pvc_jsr tbi, dest=1
mtpr r17, ev5__dtb_is // Flush DTB.
nop
nop
hw_rei_stall
// 3 tbis
//orig pvc_jsr tbi, dest=1
mtpr r17, ev5__dtb_is // Flush DTB
br r31, tbi_finish
ALIGN_BRANCH
tbi_finish:
mtpr r17, ev5__itb_is // Flush ITB
hw_rei_stall
ALIGN_BLOCK
//
// bpt_bchk_common:
// Finish up the bpt/bchk instructions
//
bpt_bchk_common:
stq r18, osfsf_a2(sp) // a2
mfpr r13, pt_entif // get entry point
stq r12, osfsf_ps(sp) // save old ps
stq r14, osfsf_pc(sp) // save pc
stq r29, osfsf_gp(sp) // save gp
mtpr r13, exc_addr // load exc_addr with entIF
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kgp
hw_rei_spe // done
ALIGN_BLOCK
//
// rti_to_user
// Finish up the rti instruction
//
rti_to_user:
mtpr r11, ev5__dtb_cm // set Mbox current mode - no virt ref for 2 cycles
mtpr r11, ev5__ps // set Ibox current mode - 2 bubble to hw_rei
mtpr r31, ev5__ipl // set the ipl. No hw_rei for 2 cycles
mtpr r25, pt_ksp // save off incase RTI to user
mfpr r30, pt_usp
hw_rei_spe // and back
ALIGN_BLOCK
//
// rti_to_kern
// Finish up the rti instruction
//
rti_to_kern:
and r12, osfps_m_ipl, r11 // clean ps
mfpr r12, pt_intmask // get int mask
extbl r12, r11, r12 // get mask for this ipl
mtpr r25, pt_ksp // save off incase RTI to user
mtpr r12, ev5__ipl // set the new ipl.
or r25, r31, sp // sp
// pvc_violate 217 // possible hidden mt->mf ipl not a problem in callpals
hw_rei
ALIGN_BLOCK
//
// swpctx_cont
// Finish up the swpctx instruction
//
swpctx_cont:
bic r25, r24, r25 // clean icsr<FPE,PMP>
sll r12, icsr_v_fpe, r12 // shift new fen to pos
ldq_p r14, osfpcb_q_mmptr(r16)// get new mmptr
srl r22, osfpcb_v_pme, r22 // get pme down to bit 0
or r25, r12, r25 // icsr with new fen
srl r23, 32, r24 // move asn to low asn pos
and r22, 1, r22
sll r24, itb_asn_v_asn, r12
sll r22, icsr_v_pmp, r22
nop
or r25, r22, r25 // icsr with new pme
sll r24, dtb_asn_v_asn, r24
subl r23, r13, r13 // gen new cc offset
mtpr r12, itb_asn // no hw_rei_stall in 0,1,2,3,4
mtpr r24, dtb_asn // Load up new ASN
mtpr r25, icsr // write the icsr
sll r14, page_offset_size_bits, r14 // Move PTBR into internal position.
ldq_p r25, osfpcb_q_usp(r16) // get new usp
insll r13, 4, r13 // >> 32
// pvc_violate 379 // ldq_p can't trap except replay. only problem if mf same ipr in same shadow
mtpr r14, pt_ptbr // load the new ptbr
mtpr r13, cc // set new offset
ldq_p r30, osfpcb_q_ksp(r16) // get new ksp
// pvc_violate 379 // ldq_p can't trap except replay. only problem if mf same ipr in same shadow
mtpr r25, pt_usp // save usp
no_pm_change_10_: hw_rei_stall // back we go
ALIGN_BLOCK
//
// swppal_cont - finish up the swppal call_pal
//
swppal_cont:
mfpr r2, pt_misc // get misc bits
sll r0, pt_misc_v_switch, r0 // get the "I've switched" bit
or r2, r0, r2 // set the bit
mtpr r31, ev5__alt_mode // ensure alt_mode set to 0 (kernel)
mtpr r2, pt_misc // update the chip
or r3, r31, r4
mfpr r3, pt_impure // pass pointer to the impure area in r3
//orig fix_impure_ipr r3 // adjust impure pointer for ipr read
//orig restore_reg1 bc_ctl, r1, r3, ipr=1 // pass cns_bc_ctl in r1
//orig restore_reg1 bc_config, r2, r3, ipr=1 // pass cns_bc_config in r2
//orig unfix_impure_ipr r3 // restore impure pointer
lda r3, CNS_Q_IPR(r3)
RESTORE_SHADOW(r1,CNS_Q_BC_CTL,r3);
RESTORE_SHADOW(r1,CNS_Q_BC_CFG,r3);
lda r3, -CNS_Q_IPR(r3)
or r31, r31, r0 // set status to success
// pvc_violate 1007
jmp r31, (r4) // and call our friend, it's her problem now
swppal_fail:
addq r0, 1, r0 // set unknown pal or not loaded
hw_rei // and return
// .sbttl "Memory management"
ALIGN_BLOCK
//
//foe_ipte_handler
// IFOE detected on level 3 pte, sort out FOE vs ACV
//
// on entry:
// with
// R8 = pte
// R10 = pc
//
// Function
// Determine TNV vs ACV vs FOE. Build stack and dispatch
// Will not be here if TNV.
//
foe_ipte_handler:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
bis r11, r31, r12 // Save PS for stack write
bge r25, foe_ipte_handler_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
srl r8, osfpte_v_ure-osfpte_v_kre, r8 // move pte user bits to kern
nop
foe_ipte_handler_10_: srl r8, osfpte_v_kre, r25 // get kre to <0>
lda sp, 0-osfsf_c_size(sp)// allocate stack space
or r10, r31, r14 // Save pc/va in case TBmiss or fault on stack
mfpr r13, pt_entmm // get entry point
stq r16, osfsf_a0(sp) // a0
or r14, r31, r16 // pass pc/va as a0
stq r17, osfsf_a1(sp) // a1
nop
stq r18, osfsf_a2(sp) // a2
lda r17, mmcsr_c_acv(r31) // assume ACV
stq r16, osfsf_pc(sp) // save pc
cmovlbs r25, mmcsr_c_foe, r17 // otherwise FOE
stq r12, osfsf_ps(sp) // save ps
subq r31, 1, r18 // pass flag of istream as a2
stq r29, osfsf_gp(sp)
mtpr r13, exc_addr // set vector address
mfpr r29, pt_kgp // load kgp
hw_rei_spe // out to exec
ALIGN_BLOCK
//
//invalid_ipte_handler
// TNV detected on level 3 pte, sort out TNV vs ACV
//
// on entry:
// with
// R8 = pte
// R10 = pc
//
// Function
// Determine TNV vs ACV. Build stack and dispatch.
//
invalid_ipte_handler:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
bis r11, r31, r12 // Save PS for stack write
bge r25, invalid_ipte_handler_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
srl r8, osfpte_v_ure-osfpte_v_kre, r8 // move pte user bits to kern
nop
invalid_ipte_handler_10_: srl r8, osfpte_v_kre, r25 // get kre to <0>
lda sp, 0-osfsf_c_size(sp)// allocate stack space
or r10, r31, r14 // Save pc/va in case TBmiss on stack
mfpr r13, pt_entmm // get entry point
stq r16, osfsf_a0(sp) // a0
or r14, r31, r16 // pass pc/va as a0
stq r17, osfsf_a1(sp) // a1
nop
stq r18, osfsf_a2(sp) // a2
and r25, 1, r17 // Isolate kre
stq r16, osfsf_pc(sp) // save pc
xor r17, 1, r17 // map to acv/tnv as a1
stq r12, osfsf_ps(sp) // save ps
subq r31, 1, r18 // pass flag of istream as a2
stq r29, osfsf_gp(sp)
mtpr r13, exc_addr // set vector address
mfpr r29, pt_kgp // load kgp
hw_rei_spe // out to exec
ALIGN_BLOCK
//
//invalid_dpte_handler
// INVALID detected on level 3 pte, sort out TNV vs ACV
//
// on entry:
// with
// R10 = va
// R8 = pte
// R9 = mm_stat
// PT6 = pc
//
// Function
// Determine TNV vs ACV. Build stack and dispatch
//
invalid_dpte_handler:
mfpr r12, pt6
blbs r12, tnv_in_pal // Special handler if original faulting reference was in PALmode
bis r12, r31, r14 // save PC in case of tbmiss or fault
srl r9, mm_stat_v_opcode, r25 // shift opc to <0>
mtpr r11, pt0 // Save PS for stack write
and r25, mm_stat_m_opcode, r25 // isolate opcode
cmpeq r25, evx_opc_sync, r25 // is it FETCH/FETCH_M?
blbs r25, nmiss_fetch_ldr31_err // yes
//dismiss exception if load to r31/f31
blbs r9, invalid_dpte_no_dismiss // mm_stat<0> set on store or fetchm
// not a store or fetch, must be a load
srl r9, mm_stat_v_ra, r25 // Shift rnum to low bits
and r25, 0x1F, r25 // isolate rnum
nop
cmpeq r25, 0x1F, r25 // Is the rnum r31 or f31?
bne r25, nmiss_fetch_ldr31_err // Yes, dismiss the fault
invalid_dpte_no_dismiss:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
bge r25, invalid_dpte_no_dismiss_10_ // no stack swap needed if cm=kern
srl r8, osfpte_v_ure-osfpte_v_kre, r8 // move pte user bits to kern
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
invalid_dpte_no_dismiss_10_: srl r8, osfpte_v_kre, r12 // get kre to <0>
lda sp, 0-osfsf_c_size(sp)// allocate stack space
or r10, r31, r25 // Save va in case TBmiss on stack
and r9, 1, r13 // save r/w flag
stq r16, osfsf_a0(sp) // a0
or r25, r31, r16 // pass va as a0
stq r17, osfsf_a1(sp) // a1
or r31, mmcsr_c_acv, r17 // assume acv
srl r12, osfpte_v_kwe-osfpte_v_kre, r25 // get write enable to <0>
stq r29, osfsf_gp(sp)
stq r18, osfsf_a2(sp) // a2
cmovlbs r13, r25, r12 // if write access move acv based on write enable
or r13, r31, r18 // pass flag of dstream access and read vs write
mfpr r25, pt0 // get ps
stq r14, osfsf_pc(sp) // save pc
mfpr r13, pt_entmm // get entry point
stq r25, osfsf_ps(sp) // save ps
mtpr r13, exc_addr // set vector address
mfpr r29, pt_kgp // load kgp
cmovlbs r12, mmcsr_c_tnv, r17 // make p2 be tnv if access ok else acv
hw_rei_spe // out to exec
//
//
// We come here if we are erring on a dtb_miss, and the instr is a
// fetch, fetch_m, of load to r31/f31.
// The PC is incremented, and we return to the program.
// essentially ignoring the instruction and error.
//
//
ALIGN_BLOCK
nmiss_fetch_ldr31_err:
mfpr r12, pt6
addq r12, 4, r12 // bump pc to pc+4
mtpr r12, exc_addr // and set entry point
mfpr r31, pt0 // pad exc_addr write
hw_rei //
ALIGN_BLOCK
//
// double_pte_inv
// We had a single tbmiss which turned into a double tbmiss which found
// an invalid PTE. Return to single miss with a fake pte, and the invalid
// single miss flow will report the error.
//
// on entry:
// r21 PTE
// r22 available
// VA IPR locked with original fault VA
// pt4 saved r21
// pt5 saved r22
// pt6 original exc_addr
//
// on return to tbmiss flow:
// r8 fake PTE
//
//
//
double_pte_inv:
srl r21, osfpte_v_kre, r21 // get the kre bit to <0>
mfpr r22, exc_addr // get the pc
lda r22, 4(r22) // inc the pc
lda r8, osfpte_m_prot(r31) // make a fake pte with xre and xwe set
cmovlbc r21, r31, r8 // set to all 0 for acv if pte<kre> is 0
mtpr r22, exc_addr // set for rei
mfpr r21, pt4 // restore regs
mfpr r22, pt5 // restore regs
hw_rei // back to tb miss
ALIGN_BLOCK
//
//tnv_in_pal
// The only places in pal that ld or store are the
// stack builders, rti or retsys. Any of these mean we
// need to take a ksp not valid halt.
//
//
tnv_in_pal:
br r31, ksp_inval_halt
// .sbttl "Icache flush routines"
ALIGN_BLOCK
//
// Common Icache flush routine.
//
//
//
pal_ic_flush:
nop
mtpr r31, ev5__ic_flush_ctl // Icache flush - E1
nop
nop
// Now, do 44 NOPs. 3RFB prefetches (24) + IC buffer,IB,slot,issue (20)
nop
nop
nop
nop
nop
nop
nop
nop
nop
nop // 10
nop
nop
nop
nop
nop
nop
nop
nop
nop
nop // 20
nop
nop
nop
nop
nop
nop
nop
nop
nop
nop // 30
nop
nop
nop
nop
nop
nop
nop
nop
nop
nop // 40
nop
nop
one_cycle_and_hw_rei:
nop
nop
hw_rei_stall
ALIGN_BLOCK
//
//osfpal_calpal_opcdec
// Here for all opcdec CALL_PALs
//
// Build stack frame
// a0 <- code
// a1 <- unpred
// a2 <- unpred
// vector via entIF
//
//
osfpal_calpal_opcdec:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
mfpr r14, exc_addr // get pc
nop
bis r11, r31, r12 // Save PS for stack write
bge r25, osfpal_calpal_opcdec_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
osfpal_calpal_opcdec_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
nop
stq r16, osfsf_a0(sp) // save regs
bis r31, osf_a0_opdec, r16 // set a0
stq r18, osfsf_a2(sp) // a2
mfpr r13, pt_entif // get entry point
stq r12, osfsf_ps(sp) // save old ps
stq r17, osfsf_a1(sp) // a1
stq r14, osfsf_pc(sp) // save pc
nop
stq r29, osfsf_gp(sp) // save gp
mtpr r13, exc_addr // load exc_addr with entIF
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kgp
hw_rei_spe // done
//
//pal_update_pcb
// Update the PCB with the current SP, AST, and CC info
//
// r0 - return linkage
//
ALIGN_BLOCK
pal_update_pcb:
mfpr r12, pt_pcbb // get pcbb
and r11, osfps_m_mode, r25 // get mode
beq r25, pal_update_pcb_10_ // in kern? no need to update user sp
mtpr r30, pt_usp // save user stack
stq_p r30, osfpcb_q_usp(r12) // store usp
br r31, pal_update_pcb_20_ // join common
pal_update_pcb_10_: stq_p r30, osfpcb_q_ksp(r12) // store ksp
pal_update_pcb_20_: rpcc r13 // get cyccounter
srl r13, 32, r14 // move offset
addl r13, r14, r14 // merge for new time
stl_p r14, osfpcb_l_cc(r12) // save time
//orig pvc_jsr updpcb, bsr=1, dest=1
ret r31, (r0)
//
// pal_save_state
//
// Function
// All chip state saved, all PT's, SR's FR's, IPR's
//
//
// Regs' on entry...
//
// R0 = halt code
// pt0 = r0
// R1 = pointer to impure
// pt4 = r1
// R3 = return addr
// pt5 = r3
//
// register usage:
// r0 = halt_code
// r1 = addr of impure area
// r3 = return_address
// r4 = scratch
//
//
ALIGN_BLOCK
.globl pal_save_state
pal_save_state:
//
//
// start of implementation independent save routine
//
// the impure area is larger than the addressibility of hw_ld and hw_st
// therefore, we need to play some games: The impure area
// is informally divided into the "machine independent" part and the
// "machine dependent" part. The state that will be saved in the
// "machine independent" part are gpr's, fpr's, hlt, flag, mchkflag (use (un)fix_impure_gpr macros).
// All others will be in the "machine dependent" part (use (un)fix_impure_ipr macros).
// The impure pointer will need to be adjusted by a different offset for each. The store/restore_reg
// macros will automagically adjust the offset correctly.
//
// The distributed code is commented out and followed by corresponding SRC code.
// Beware: SAVE_IPR and RESTORE_IPR blow away r0(v0)
//orig fix_impure_gpr r1 // adjust impure area pointer for stores to "gpr" part of impure area
lda r1, 0x200(r1) // Point to center of CPU segment
//orig store_reg1 flag, r31, r1, ipr=1 // clear dump area flag
SAVE_GPR(r31,CNS_Q_FLAG,r1) // Clear the valid flag
//orig store_reg1 hlt, r0, r1, ipr=1
SAVE_GPR(r0,CNS_Q_HALT,r1) // Save the halt code
mfpr r0, pt0 // get r0 back //orig
//orig store_reg1 0, r0, r1 // save r0
SAVE_GPR(r0,CNS_Q_GPR+0x00,r1) // Save r0
mfpr r0, pt4 // get r1 back //orig
//orig store_reg1 1, r0, r1 // save r1
SAVE_GPR(r0,CNS_Q_GPR+0x08,r1) // Save r1
//orig store_reg 2 // save r2
SAVE_GPR(r2,CNS_Q_GPR+0x10,r1) // Save r2
mfpr r0, pt5 // get r3 back //orig
//orig store_reg1 3, r0, r1 // save r3
SAVE_GPR(r0,CNS_Q_GPR+0x18,r1) // Save r3
// reason code has been saved
// r0 has been saved
// r1 has been saved
// r2 has been saved
// r3 has been saved
// pt0, pt4, pt5 have been lost
//
// Get out of shadow mode
//
mfpr r2, icsr // Get icsr
ldah r0, (1<<(icsr_v_sde-16))(r31)
bic r2, r0, r0 // ICSR with SDE clear
mtpr r0, icsr // Turn off SDE
mfpr r31, pt0 // SDE bubble cycle 1
mfpr r31, pt0 // SDE bubble cycle 2
mfpr r31, pt0 // SDE bubble cycle 3
nop
// save integer regs R4-r31
SAVE_GPR(r4,CNS_Q_GPR+0x20,r1)
SAVE_GPR(r5,CNS_Q_GPR+0x28,r1)
SAVE_GPR(r6,CNS_Q_GPR+0x30,r1)
SAVE_GPR(r7,CNS_Q_GPR+0x38,r1)
SAVE_GPR(r8,CNS_Q_GPR+0x40,r1)
SAVE_GPR(r9,CNS_Q_GPR+0x48,r1)
SAVE_GPR(r10,CNS_Q_GPR+0x50,r1)
SAVE_GPR(r11,CNS_Q_GPR+0x58,r1)
SAVE_GPR(r12,CNS_Q_GPR+0x60,r1)
SAVE_GPR(r13,CNS_Q_GPR+0x68,r1)
SAVE_GPR(r14,CNS_Q_GPR+0x70,r1)
SAVE_GPR(r15,CNS_Q_GPR+0x78,r1)
SAVE_GPR(r16,CNS_Q_GPR+0x80,r1)
SAVE_GPR(r17,CNS_Q_GPR+0x88,r1)
SAVE_GPR(r18,CNS_Q_GPR+0x90,r1)
SAVE_GPR(r19,CNS_Q_GPR+0x98,r1)
SAVE_GPR(r20,CNS_Q_GPR+0xA0,r1)
SAVE_GPR(r21,CNS_Q_GPR+0xA8,r1)
SAVE_GPR(r22,CNS_Q_GPR+0xB0,r1)
SAVE_GPR(r23,CNS_Q_GPR+0xB8,r1)
SAVE_GPR(r24,CNS_Q_GPR+0xC0,r1)
SAVE_GPR(r25,CNS_Q_GPR+0xC8,r1)
SAVE_GPR(r26,CNS_Q_GPR+0xD0,r1)
SAVE_GPR(r27,CNS_Q_GPR+0xD8,r1)
SAVE_GPR(r28,CNS_Q_GPR+0xE0,r1)
SAVE_GPR(r29,CNS_Q_GPR+0xE8,r1)
SAVE_GPR(r30,CNS_Q_GPR+0xF0,r1)
SAVE_GPR(r31,CNS_Q_GPR+0xF8,r1)
// save all paltemp regs except pt0
//orig unfix_impure_gpr r1 // adjust impure area pointer for gpr stores
//orig fix_impure_ipr r1 // adjust impure area pointer for pt stores
lda r1, -0x200(r1) // Restore the impure base address.
lda r1, CNS_Q_IPR(r1) // Point to the base of IPR area.
SAVE_IPR(pt0,CNS_Q_PT+0x00,r1) // the osf code didn't save/restore palTemp 0 ?? pboyle
SAVE_IPR(pt1,CNS_Q_PT+0x08,r1)
SAVE_IPR(pt2,CNS_Q_PT+0x10,r1)
SAVE_IPR(pt3,CNS_Q_PT+0x18,r1)
SAVE_IPR(pt4,CNS_Q_PT+0x20,r1)
SAVE_IPR(pt5,CNS_Q_PT+0x28,r1)
SAVE_IPR(pt6,CNS_Q_PT+0x30,r1)
SAVE_IPR(pt7,CNS_Q_PT+0x38,r1)
SAVE_IPR(pt8,CNS_Q_PT+0x40,r1)
SAVE_IPR(pt9,CNS_Q_PT+0x48,r1)
SAVE_IPR(pt10,CNS_Q_PT+0x50,r1)
SAVE_IPR(pt11,CNS_Q_PT+0x58,r1)
SAVE_IPR(pt12,CNS_Q_PT+0x60,r1)
SAVE_IPR(pt13,CNS_Q_PT+0x68,r1)
SAVE_IPR(pt14,CNS_Q_PT+0x70,r1)
SAVE_IPR(pt15,CNS_Q_PT+0x78,r1)
SAVE_IPR(pt16,CNS_Q_PT+0x80,r1)
SAVE_IPR(pt17,CNS_Q_PT+0x88,r1)
SAVE_IPR(pt18,CNS_Q_PT+0x90,r1)
SAVE_IPR(pt19,CNS_Q_PT+0x98,r1)
SAVE_IPR(pt20,CNS_Q_PT+0xA0,r1)
SAVE_IPR(pt21,CNS_Q_PT+0xA8,r1)
SAVE_IPR(pt22,CNS_Q_PT+0xB0,r1)
SAVE_IPR(pt23,CNS_Q_PT+0xB8,r1)
// Restore shadow mode
mfpr r31, pt0 // pad write to icsr out of shadow of store (trap does not abort write)
mfpr r31, pt0
mtpr r2, icsr // Restore original ICSR
mfpr r31, pt0 // SDE bubble cycle 1
mfpr r31, pt0 // SDE bubble cycle 2
mfpr r31, pt0 // SDE bubble cycle 3
nop
// save all integer shadow regs
SAVE_SHADOW( r8,CNS_Q_SHADOW+0x00,r1) // also called p0...p7 in the Hudson code
SAVE_SHADOW( r9,CNS_Q_SHADOW+0x08,r1)
SAVE_SHADOW(r10,CNS_Q_SHADOW+0x10,r1)
SAVE_SHADOW(r11,CNS_Q_SHADOW+0x18,r1)
SAVE_SHADOW(r12,CNS_Q_SHADOW+0x20,r1)
SAVE_SHADOW(r13,CNS_Q_SHADOW+0x28,r1)
SAVE_SHADOW(r14,CNS_Q_SHADOW+0x30,r1)
SAVE_SHADOW(r25,CNS_Q_SHADOW+0x38,r1)
SAVE_IPR(excAddr,CNS_Q_EXC_ADDR,r1)
SAVE_IPR(palBase,CNS_Q_PAL_BASE,r1)
SAVE_IPR(mmStat,CNS_Q_MM_STAT,r1)
SAVE_IPR(va,CNS_Q_VA,r1)
SAVE_IPR(icsr,CNS_Q_ICSR,r1)
SAVE_IPR(ipl,CNS_Q_IPL,r1)
SAVE_IPR(ips,CNS_Q_IPS,r1)
SAVE_IPR(itbAsn,CNS_Q_ITB_ASN,r1)
SAVE_IPR(aster,CNS_Q_ASTER,r1)
SAVE_IPR(astrr,CNS_Q_ASTRR,r1)
SAVE_IPR(sirr,CNS_Q_SIRR,r1)
SAVE_IPR(isr,CNS_Q_ISR,r1)
SAVE_IPR(iVptBr,CNS_Q_IVPTBR,r1)
SAVE_IPR(mcsr,CNS_Q_MCSR,r1)
SAVE_IPR(dcMode,CNS_Q_DC_MODE,r1)
//orig pvc_violate 379 // mf maf_mode after a store ok (pvc doesn't distinguish ld from st)
//orig store_reg maf_mode, ipr=1 // save ipr -- no mbox instructions for
//orig // PVC violation applies only to
pvc$osf35$379: // loads. HW_ST ok here, so ignore
SAVE_IPR(mafMode,CNS_Q_MAF_MODE,r1) // MBOX INST->MF MAF_MODE IN 0,1,2
//the following iprs are informational only -- will not be restored
SAVE_IPR(icPerr,CNS_Q_ICPERR_STAT,r1)
SAVE_IPR(PmCtr,CNS_Q_PM_CTR,r1)
SAVE_IPR(intId,CNS_Q_INT_ID,r1)
SAVE_IPR(excSum,CNS_Q_EXC_SUM,r1)
SAVE_IPR(excMask,CNS_Q_EXC_MASK,r1)
ldah r14, 0xFFF0(zero)
zap r14, 0xE0, r14 // Get base address of CBOX IPRs
NOP // Pad mfpr dcPerr out of shadow of
NOP // last store
NOP
SAVE_IPR(dcPerr,CNS_Q_DCPERR_STAT,r1)
// read cbox ipr state
mb
ldq_p r2, scCtl(r14)
ldq_p r13, ldLock(r14)
ldq_p r4, scAddr(r14)
ldq_p r5, eiAddr(r14)
ldq_p r6, bcTagAddr(r14)
ldq_p r7, fillSyn(r14)
bis r5, r4, zero // Make sure all loads complete before
bis r7, r6, zero // reading registers that unlock them.
ldq_p r8, scStat(r14) // Unlocks scAddr.
ldq_p r9, eiStat(r14) // Unlocks eiAddr, bcTagAddr, fillSyn.
ldq_p zero, eiStat(r14) // Make sure it is really unlocked.
mb
// save cbox ipr state
SAVE_SHADOW(r2,CNS_Q_SC_CTL,r1);
SAVE_SHADOW(r13,CNS_Q_LD_LOCK,r1);
SAVE_SHADOW(r4,CNS_Q_SC_ADDR,r1);
SAVE_SHADOW(r5,CNS_Q_EI_ADDR,r1);
SAVE_SHADOW(r6,CNS_Q_BC_TAG_ADDR,r1);
SAVE_SHADOW(r7,CNS_Q_FILL_SYN,r1);
SAVE_SHADOW(r8,CNS_Q_SC_STAT,r1);
SAVE_SHADOW(r9,CNS_Q_EI_STAT,r1);
//bc_config? sl_rcv?
// restore impure base
//orig unfix_impure_ipr r1
lda r1, -CNS_Q_IPR(r1)
// save all floating regs
mfpr r0, icsr // get icsr
or r31, 1, r2 // get a one
sll r2, icsr_v_fpe, r2 // Shift it into ICSR<FPE> position
or r2, r0, r0 // set FEN on
mtpr r0, icsr // write to icsr, enabling FEN
// map the save area virtually
mtpr r31, dtbIa // Clear all DTB entries
srl r1, va_s_off, r0 // Clean off byte-within-page offset
sll r0, pte_v_pfn, r0 // Shift to form PFN
lda r0, pte_m_prot(r0) // Set all read/write enable bits
mtpr r0, dtbPte // Load the PTE and set valid
mtpr r1, dtbTag // Write the PTE and tag into the DTB
// map the next page too - in case the impure area crosses a page boundary
lda r4, (1<<va_s_off)(r1) // Generate address for next page
srl r4, va_s_off, r0 // Clean off byte-within-page offset
sll r0, pte_v_pfn, r0 // Shift to form PFN
lda r0, pte_m_prot(r0) // Set all read/write enable bits
mtpr r0, dtbPte // Load the PTE and set valid
mtpr r4, dtbTag // Write the PTE and tag into the DTB
sll r31, 0, r31 // stall cycle 1
sll r31, 0, r31 // stall cycle 2
sll r31, 0, r31 // stall cycle 3
nop
// add offset for saving fpr regs
//orig fix_impure_gpr r1
lda r1, 0x200(r1) // Point to center of CPU segment
// now save the regs - F0-F31
mf_fpcr f0 // original
SAVE_FPR(f0,CNS_Q_FPR+0x00,r1)
SAVE_FPR(f1,CNS_Q_FPR+0x08,r1)
SAVE_FPR(f2,CNS_Q_FPR+0x10,r1)
SAVE_FPR(f3,CNS_Q_FPR+0x18,r1)
SAVE_FPR(f4,CNS_Q_FPR+0x20,r1)
SAVE_FPR(f5,CNS_Q_FPR+0x28,r1)
SAVE_FPR(f6,CNS_Q_FPR+0x30,r1)
SAVE_FPR(f7,CNS_Q_FPR+0x38,r1)
SAVE_FPR(f8,CNS_Q_FPR+0x40,r1)
SAVE_FPR(f9,CNS_Q_FPR+0x48,r1)
SAVE_FPR(f10,CNS_Q_FPR+0x50,r1)
SAVE_FPR(f11,CNS_Q_FPR+0x58,r1)
SAVE_FPR(f12,CNS_Q_FPR+0x60,r1)
SAVE_FPR(f13,CNS_Q_FPR+0x68,r1)
SAVE_FPR(f14,CNS_Q_FPR+0x70,r1)
SAVE_FPR(f15,CNS_Q_FPR+0x78,r1)
SAVE_FPR(f16,CNS_Q_FPR+0x80,r1)
SAVE_FPR(f17,CNS_Q_FPR+0x88,r1)
SAVE_FPR(f18,CNS_Q_FPR+0x90,r1)
SAVE_FPR(f19,CNS_Q_FPR+0x98,r1)
SAVE_FPR(f20,CNS_Q_FPR+0xA0,r1)
SAVE_FPR(f21,CNS_Q_FPR+0xA8,r1)
SAVE_FPR(f22,CNS_Q_FPR+0xB0,r1)
SAVE_FPR(f23,CNS_Q_FPR+0xB8,r1)
SAVE_FPR(f24,CNS_Q_FPR+0xC0,r1)
SAVE_FPR(f25,CNS_Q_FPR+0xC8,r1)
SAVE_FPR(f26,CNS_Q_FPR+0xD0,r1)
SAVE_FPR(f27,CNS_Q_FPR+0xD8,r1)
SAVE_FPR(f28,CNS_Q_FPR+0xE0,r1)
SAVE_FPR(f29,CNS_Q_FPR+0xE8,r1)
SAVE_FPR(f30,CNS_Q_FPR+0xF0,r1)
SAVE_FPR(f31,CNS_Q_FPR+0xF8,r1)
//switch impure offset from gpr to ipr---
//orig unfix_impure_gpr r1
//orig fix_impure_ipr r1
//orig store_reg1 fpcsr, f0, r1, fpcsr=1
SAVE_FPR(f0,CNS_Q_FPCSR,r1) // fpcsr loaded above into f0 -- can it reach
lda r1, -0x200(r1) // Restore the impure base address
// and back to gpr ---
//orig unfix_impure_ipr r1
//orig fix_impure_gpr r1
//orig lda r0, cns_mchksize(r31) // get size of mchk area
//orig store_reg1 mchkflag, r0, r1, ipr=1
//orig mb
lda r1, CNS_Q_IPR(r1) // Point to base of IPR area again
// save this using the IPR base (it is closer) not the GRP base as they used...pb
lda r0, MACHINE_CHECK_SIZE(r31) // get size of mchk area
SAVE_SHADOW(r0,CNS_Q_MCHK,r1);
mb
//orig or r31, 1, r0 // get a one
//orig store_reg1 flag, r0, r1, ipr=1 // set dump area flag
//orig mb
lda r1, -CNS_Q_IPR(r1) // back to the base
lda r1, 0x200(r1) // Point to center of CPU segment
or r31, 1, r0 // get a one
SAVE_GPR(r0,CNS_Q_FLAG,r1) // // set dump area valid flag
mb
// restore impure area base
//orig unfix_impure_gpr r1
lda r1, -0x200(r1) // Point to center of CPU segment
mtpr r31, dtb_ia // clear the dtb
mtpr r31, itb_ia // clear the itb
//orig pvc_jsr savsta, bsr=1, dest=1
ret r31, (r3) // and back we go
// .sbttl "PAL_RESTORE_STATE"
//
//
// Pal_restore_state
//
//
// register usage:
// r1 = addr of impure area
// r3 = return_address
// all other regs are scratchable, as they are about to
// be reloaded from ram.
//
// Function:
// All chip state restored, all SRs, FRs, PTs, IPRs
// *** except R1, R3, PT0, PT4, PT5 ***
//
//
ALIGN_BLOCK
pal_restore_state:
//need to restore sc_ctl,bc_ctl,bc_config??? if so, need to figure out a safe way to do so.
// map the console io area virtually
mtpr r31, dtbIa // Clear all DTB entries
srl r1, va_s_off, r0 // Clean off byte-within-page offset
sll r0, pte_v_pfn, r0 // Shift to form PFN
lda r0, pte_m_prot(r0) // Set all read/write enable bits
mtpr r0, dtbPte // Load the PTE and set valid
mtpr r1, dtbTag // Write the PTE and tag into the DTB
// map the next page too, in case impure area crosses page boundary
lda r4, (1<<VA_S_OFF)(r1) // Generate address for next page
srl r4, va_s_off, r0 // Clean off byte-within-page offset
sll r0, pte_v_pfn, r0 // Shift to form PFN
lda r0, pte_m_prot(r0) // Set all read/write enable bits
mtpr r0, dtbPte // Load the PTE and set valid
mtpr r4, dtbTag // Write the PTE and tag into the DTB
// save all floating regs
mfpr r0, icsr // Get current ICSR
bis zero, 1, r2 // Get a '1'
or r2, (1<<(icsr_v_sde-icsr_v_fpe)), r2
sll r2, icsr_v_fpe, r2 // Shift bits into position
bis r2, r2, r0 // Set ICSR<SDE> and ICSR<FPE>
mtpr r0, icsr // Update the chip
mfpr r31, pt0 // FPE bubble cycle 1 //orig
mfpr r31, pt0 // FPE bubble cycle 2 //orig
mfpr r31, pt0 // FPE bubble cycle 3 //orig
//orig fix_impure_ipr r1
//orig restore_reg1 fpcsr, f0, r1, fpcsr=1
//orig mt_fpcr f0
//orig
//orig unfix_impure_ipr r1
//orig fix_impure_gpr r1 // adjust impure pointer offset for gpr access
lda r1, 200(r1) // Point to base of IPR area again
RESTORE_FPR(f0,CNS_Q_FPCSR,r1) // can it reach?? pb
mt_fpcr f0 // original
lda r1, 0x200(r1) // point to center of CPU segment
// restore all floating regs
RESTORE_FPR(f0,CNS_Q_FPR+0x00,r1)
RESTORE_FPR(f1,CNS_Q_FPR+0x08,r1)
RESTORE_FPR(f2,CNS_Q_FPR+0x10,r1)
RESTORE_FPR(f3,CNS_Q_FPR+0x18,r1)
RESTORE_FPR(f4,CNS_Q_FPR+0x20,r1)
RESTORE_FPR(f5,CNS_Q_FPR+0x28,r1)
RESTORE_FPR(f6,CNS_Q_FPR+0x30,r1)
RESTORE_FPR(f7,CNS_Q_FPR+0x38,r1)
RESTORE_FPR(f8,CNS_Q_FPR+0x40,r1)
RESTORE_FPR(f9,CNS_Q_FPR+0x48,r1)
RESTORE_FPR(f10,CNS_Q_FPR+0x50,r1)
RESTORE_FPR(f11,CNS_Q_FPR+0x58,r1)
RESTORE_FPR(f12,CNS_Q_FPR+0x60,r1)
RESTORE_FPR(f13,CNS_Q_FPR+0x68,r1)
RESTORE_FPR(f14,CNS_Q_FPR+0x70,r1)
RESTORE_FPR(f15,CNS_Q_FPR+0x78,r1)
RESTORE_FPR(f16,CNS_Q_FPR+0x80,r1)
RESTORE_FPR(f17,CNS_Q_FPR+0x88,r1)
RESTORE_FPR(f18,CNS_Q_FPR+0x90,r1)
RESTORE_FPR(f19,CNS_Q_FPR+0x98,r1)
RESTORE_FPR(f20,CNS_Q_FPR+0xA0,r1)
RESTORE_FPR(f21,CNS_Q_FPR+0xA8,r1)
RESTORE_FPR(f22,CNS_Q_FPR+0xB0,r1)
RESTORE_FPR(f23,CNS_Q_FPR+0xB8,r1)
RESTORE_FPR(f24,CNS_Q_FPR+0xC0,r1)
RESTORE_FPR(f25,CNS_Q_FPR+0xC8,r1)
RESTORE_FPR(f26,CNS_Q_FPR+0xD0,r1)
RESTORE_FPR(f27,CNS_Q_FPR+0xD8,r1)
RESTORE_FPR(f28,CNS_Q_FPR+0xE0,r1)
RESTORE_FPR(f29,CNS_Q_FPR+0xE8,r1)
RESTORE_FPR(f30,CNS_Q_FPR+0xF0,r1)
RESTORE_FPR(f31,CNS_Q_FPR+0xF8,r1)
// switch impure pointer from gpr to ipr area --
//orig unfix_impure_gpr r1
//orig fix_impure_ipr r1
lda r1, -0x200(r1) // Restore base address of impure area.
lda r1, CNS_Q_IPR(r1) // Point to base of IPR area.
// restore all pal regs
RESTORE_IPR(pt0,CNS_Q_PT+0x00,r1) // the osf code didn't save/restore palTemp 0 ?? pboyle
RESTORE_IPR(pt1,CNS_Q_PT+0x08,r1)
RESTORE_IPR(pt2,CNS_Q_PT+0x10,r1)
RESTORE_IPR(pt3,CNS_Q_PT+0x18,r1)
RESTORE_IPR(pt4,CNS_Q_PT+0x20,r1)
RESTORE_IPR(pt5,CNS_Q_PT+0x28,r1)
RESTORE_IPR(pt6,CNS_Q_PT+0x30,r1)
RESTORE_IPR(pt7,CNS_Q_PT+0x38,r1)
RESTORE_IPR(pt8,CNS_Q_PT+0x40,r1)
RESTORE_IPR(pt9,CNS_Q_PT+0x48,r1)
RESTORE_IPR(pt10,CNS_Q_PT+0x50,r1)
RESTORE_IPR(pt11,CNS_Q_PT+0x58,r1)
RESTORE_IPR(pt12,CNS_Q_PT+0x60,r1)
RESTORE_IPR(pt13,CNS_Q_PT+0x68,r1)
RESTORE_IPR(pt14,CNS_Q_PT+0x70,r1)
RESTORE_IPR(pt15,CNS_Q_PT+0x78,r1)
RESTORE_IPR(pt16,CNS_Q_PT+0x80,r1)
RESTORE_IPR(pt17,CNS_Q_PT+0x88,r1)
RESTORE_IPR(pt18,CNS_Q_PT+0x90,r1)
RESTORE_IPR(pt19,CNS_Q_PT+0x98,r1)
RESTORE_IPR(pt20,CNS_Q_PT+0xA0,r1)
RESTORE_IPR(pt21,CNS_Q_PT+0xA8,r1)
RESTORE_IPR(pt22,CNS_Q_PT+0xB0,r1)
RESTORE_IPR(pt23,CNS_Q_PT+0xB8,r1)
//orig restore_reg exc_addr, ipr=1 // restore ipr
//orig restore_reg pal_base, ipr=1 // restore ipr
//orig restore_reg ipl, ipr=1 // restore ipr
//orig restore_reg ps, ipr=1 // restore ipr
//orig mtpr r0, dtb_cm // set current mode in mbox too
//orig restore_reg itb_asn, ipr=1
//orig srl r0, itb_asn_v_asn, r0
//orig sll r0, dtb_asn_v_asn, r0
//orig mtpr r0, dtb_asn // set ASN in Mbox too
//orig restore_reg ivptbr, ipr=1
//orig mtpr r0, mvptbr // use ivptbr value to restore mvptbr
//orig restore_reg mcsr, ipr=1
//orig restore_reg aster, ipr=1
//orig restore_reg astrr, ipr=1
//orig restore_reg sirr, ipr=1
//orig restore_reg maf_mode, ipr=1 // no mbox instruction for 3 cycles
//orig mfpr r31, pt0 // (may issue with mt maf_mode)
//orig mfpr r31, pt0 // bubble cycle 1
//orig mfpr r31, pt0 // bubble cycle 2
//orig mfpr r31, pt0 // bubble cycle 3
//orig mfpr r31, pt0 // (may issue with following ld)
// r0 gets the value of RESTORE_IPR in the macro and this code uses this side effect (gag)
RESTORE_IPR(excAddr,CNS_Q_EXC_ADDR,r1)
RESTORE_IPR(palBase,CNS_Q_PAL_BASE,r1)
RESTORE_IPR(ipl,CNS_Q_IPL,r1)
RESTORE_IPR(ips,CNS_Q_IPS,r1)
mtpr r0, dtbCm // Set Mbox current mode too.
RESTORE_IPR(itbAsn,CNS_Q_ITB_ASN,r1)
srl r0, 4, r0
sll r0, 57, r0
mtpr r0, dtbAsn // Set Mbox ASN too
RESTORE_IPR(iVptBr,CNS_Q_IVPTBR,r1)
mtpr r0, mVptBr // Set Mbox VptBr too
RESTORE_IPR(mcsr,CNS_Q_MCSR,r1)
RESTORE_IPR(aster,CNS_Q_ASTER,r1)
RESTORE_IPR(astrr,CNS_Q_ASTRR,r1)
RESTORE_IPR(sirr,CNS_Q_SIRR,r1)
RESTORE_IPR(mafMode,CNS_Q_MAF_MODE,r1)
STALL
STALL
STALL
STALL
STALL
// restore all integer shadow regs
RESTORE_SHADOW( r8,CNS_Q_SHADOW+0x00,r1) // also called p0...p7 in the Hudson code
RESTORE_SHADOW( r9,CNS_Q_SHADOW+0x08,r1)
RESTORE_SHADOW(r10,CNS_Q_SHADOW+0x10,r1)
RESTORE_SHADOW(r11,CNS_Q_SHADOW+0x18,r1)
RESTORE_SHADOW(r12,CNS_Q_SHADOW+0x20,r1)
RESTORE_SHADOW(r13,CNS_Q_SHADOW+0x28,r1)
RESTORE_SHADOW(r14,CNS_Q_SHADOW+0x30,r1)
RESTORE_SHADOW(r25,CNS_Q_SHADOW+0x38,r1)
RESTORE_IPR(dcMode,CNS_Q_DC_MODE,r1)
//
// Get out of shadow mode
//
mfpr r31, pt0 // pad last load to icsr write (in case of replay, icsr will be written anyway)
mfpr r31, pt0 // ""
mfpr r0, icsr // Get icsr
ldah r2, (1<<(ICSR_V_SDE-16))(r31) // Get a one in SHADOW_ENABLE bit location
bic r0, r2, r2 // ICSR with SDE clear
mtpr r2, icsr // Turn off SDE - no palshadow rd/wr for 3 bubble cycles
mfpr r31, pt0 // SDE bubble cycle 1
mfpr r31, pt0 // SDE bubble cycle 2
mfpr r31, pt0 // SDE bubble cycle 3
nop
// switch impure pointer from ipr to gpr area --
//orig unfix_impure_ipr r1
//orig fix_impure_gpr r1
// Restore GPRs (r0, r2 are restored later, r1 and r3 are trashed) ...
lda r1, -CNS_Q_IPR(r1) // Restore base address of impure area
lda r1, 0x200(r1) // Point to center of CPU segment
// restore all integer regs
RESTORE_GPR(r4,CNS_Q_GPR+0x20,r1)
RESTORE_GPR(r5,CNS_Q_GPR+0x28,r1)
RESTORE_GPR(r6,CNS_Q_GPR+0x30,r1)
RESTORE_GPR(r7,CNS_Q_GPR+0x38,r1)
RESTORE_GPR(r8,CNS_Q_GPR+0x40,r1)
RESTORE_GPR(r9,CNS_Q_GPR+0x48,r1)
RESTORE_GPR(r10,CNS_Q_GPR+0x50,r1)
RESTORE_GPR(r11,CNS_Q_GPR+0x58,r1)
RESTORE_GPR(r12,CNS_Q_GPR+0x60,r1)
RESTORE_GPR(r13,CNS_Q_GPR+0x68,r1)
RESTORE_GPR(r14,CNS_Q_GPR+0x70,r1)
RESTORE_GPR(r15,CNS_Q_GPR+0x78,r1)
RESTORE_GPR(r16,CNS_Q_GPR+0x80,r1)
RESTORE_GPR(r17,CNS_Q_GPR+0x88,r1)
RESTORE_GPR(r18,CNS_Q_GPR+0x90,r1)
RESTORE_GPR(r19,CNS_Q_GPR+0x98,r1)
RESTORE_GPR(r20,CNS_Q_GPR+0xA0,r1)
RESTORE_GPR(r21,CNS_Q_GPR+0xA8,r1)
RESTORE_GPR(r22,CNS_Q_GPR+0xB0,r1)
RESTORE_GPR(r23,CNS_Q_GPR+0xB8,r1)
RESTORE_GPR(r24,CNS_Q_GPR+0xC0,r1)
RESTORE_GPR(r25,CNS_Q_GPR+0xC8,r1)
RESTORE_GPR(r26,CNS_Q_GPR+0xD0,r1)
RESTORE_GPR(r27,CNS_Q_GPR+0xD8,r1)
RESTORE_GPR(r28,CNS_Q_GPR+0xE0,r1)
RESTORE_GPR(r29,CNS_Q_GPR+0xE8,r1)
RESTORE_GPR(r30,CNS_Q_GPR+0xF0,r1)
RESTORE_GPR(r31,CNS_Q_GPR+0xF8,r1)
//orig // switch impure pointer from gpr to ipr area --
//orig unfix_impure_gpr r1
//orig fix_impure_ipr r1
//orig restore_reg icsr, ipr=1 // restore original icsr- 4 bubbles to hw_rei
lda t0, -0x200(t0) // Restore base address of impure area.
lda t0, CNS_Q_IPR(t0) // Point to base of IPR area again.
RESTORE_IPR(icsr,CNS_Q_ICSR,r1)
//orig // and back again --
//orig unfix_impure_ipr r1
//orig fix_impure_gpr r1
//orig store_reg1 flag, r31, r1, ipr=1 // clear dump area valid flag
//orig mb
lda t0, -CNS_Q_IPR(t0) // Back to base of impure area again,
lda t0, 0x200(t0) // and back to center of CPU segment
SAVE_GPR(r31,CNS_Q_FLAG,r1) // Clear the dump area valid flag
mb
//orig // and back we go
//orig// restore_reg 3
//orig restore_reg 2
//orig// restore_reg 1
//orig restore_reg 0
//orig // restore impure area base
//orig unfix_impure_gpr r1
RESTORE_GPR(r2,CNS_Q_GPR+0x10,r1)
RESTORE_GPR(r0,CNS_Q_GPR+0x00,r1)
lda r1, -0x200(r1) // Restore impure base address
mfpr r31, pt0 // stall for ldq_p above //orig
mtpr r31, dtb_ia // clear the tb //orig
mtpr r31, itb_ia // clear the itb //orig
//orig pvc_jsr rststa, bsr=1, dest=1
ret r31, (r3) // back we go //orig
//
// pal_pal_bug_check -- code has found a bugcheck situation.
// Set things up and join common machine check flow.
//
// Input:
// r14 - exc_addr
//
// On exit:
// pt0 - saved r0
// pt1 - saved r1
// pt4 - saved r4
// pt5 - saved r5
// pt6 - saved r6
// pt10 - saved exc_addr
// pt_misc<47:32> - mchk code
// pt_misc<31:16> - scb vector
// r14 - base of Cbox IPRs in IO space
// MCES<mchk> is set
//
ALIGN_BLOCK
.globl pal_pal_bug_check_from_int
pal_pal_bug_check_from_int:
DEBUGSTORE(0x79)
//simos DEBUG_EXC_ADDR()
DEBUGSTORE(0x20)
//simos bsr r25, put_hex
lda r25, mchk_c_bugcheck(r31)
addq r25, 1, r25 // set flag indicating we came from interrupt and stack is already pushed
br r31, pal_pal_mchk
nop
pal_pal_bug_check:
lda r25, mchk_c_bugcheck(r31)
pal_pal_mchk:
sll r25, 32, r25 // Move mchk code to position
mtpr r14, pt10 // Stash exc_addr
mtpr r14, exc_addr
mfpr r12, pt_misc // Get MCES and scratch
zap r12, 0x3c, r12
or r12, r25, r12 // Combine mchk code
lda r25, scb_v_procmchk(r31) // Get SCB vector
sll r25, 16, r25 // Move SCBv to position
or r12, r25, r25 // Combine SCBv
mtpr r0, pt0 // Stash for scratch
bis r25, mces_m_mchk, r25 // Set MCES<MCHK> bit
mtpr r25, pt_misc // Save mchk code!scbv!whami!mces
ldah r14, 0xfff0(r31)
mtpr r1, pt1 // Stash for scratch
zap r14, 0xE0, r14 // Get Cbox IPR base
mtpr r4, pt4
mtpr r5, pt5
mtpr r6, pt6
blbs r12, sys_double_machine_check // MCHK halt if double machine check
br r31, sys_mchk_collect_iprs // Join common machine check flow
// align_to_call_pal_section
// Align to address of first call_pal entry point - 2000
//
// HALT - PALcode for HALT instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// GO to console code
//
//
.text 1
// . = 0x2000
CALL_PAL_PRIV(PAL_HALT_ENTRY)
call_pal_halt:
mfpr r31, pt0 // Pad exc_addr read
mfpr r31, pt0
mfpr r12, exc_addr // get PC
subq r12, 4, r12 // Point to the HALT
mtpr r12, exc_addr
mtpr r0, pt0
//orig pvc_jsr updpcb, bsr=1
bsr r0, pal_update_pcb // update the pcb
lda r0, hlt_c_sw_halt(r31) // set halt code to sw halt
br r31, sys_enter_console // enter the console
//
// CFLUSH - PALcode for CFLUSH instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// R16 - contains the PFN of the page to be flushed
//
// Function:
// Flush all Dstream caches of 1 entire page
// The CFLUSH routine is in the system specific module.
//
//
CALL_PAL_PRIV(PAL_CFLUSH_ENTRY)
Call_Pal_Cflush:
br r31, sys_cflush
//
// DRAINA - PALcode for DRAINA instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
// Implicit TRAPB performed by hardware.
//
// Function:
// Stall instruction issue until all prior instructions are guaranteed to
// complete without incurring aborts. For the EV5 implementation, this
// means waiting until all pending DREADS are returned.
//
//
CALL_PAL_PRIV(PAL_DRAINA_ENTRY)
Call_Pal_Draina:
ldah r14, 0x100(r31) // Init counter. Value?
nop
DRAINA_LOOP:
subq r14, 1, r14 // Decrement counter
mfpr r13, ev5__maf_mode // Fetch status bit
srl r13, maf_mode_v_dread_pending, r13
ble r14, DRAINA_LOOP_TOO_LONG
nop
blbs r13, DRAINA_LOOP // Wait until all DREADS clear
hw_rei
DRAINA_LOOP_TOO_LONG:
br r31, call_pal_halt
// CALL_PAL OPCDECs
CALL_PAL_PRIV(0x0003)
CallPal_OpcDec03:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0004)
CallPal_OpcDec04:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0005)
CallPal_OpcDec05:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0006)
CallPal_OpcDec06:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0007)
CallPal_OpcDec07:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0008)
CallPal_OpcDec08:
br r31, osfpal_calpal_opcdec
//
// CSERVE - PALcode for CSERVE instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Various functions for private use of console software
//
// option selector in r0
// arguments in r16....
// The CSERVE routine is in the system specific module.
//
//
CALL_PAL_PRIV(PAL_CSERVE_ENTRY)
Call_Pal_Cserve:
br r31, sys_cserve
//
// swppal - PALcode for swppal instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
// Vectored into via hardware PALcode instruction dispatch.
// R16 contains the new PAL identifier
// R17:R21 contain implementation-specific entry parameters
//
// R0 receives status:
// 0 success (PAL was switched)
// 1 unknown PAL variant
// 2 known PAL variant, but PAL not loaded
//
//
// Function:
// Swap control to another PAL.
//
CALL_PAL_PRIV(PAL_SWPPAL_ENTRY)
Call_Pal_Swppal:
cmpule r16, 255, r0 // see if a kibble was passed
cmoveq r16, r16, r0 // if r16=0 then a valid address (ECO 59)
or r16, r31, r3 // set r3 incase this is a address
blbc r0, swppal_cont // nope, try it as an address
cmpeq r16, 2, r0 // is it our friend OSF?
blbc r0, swppal_fail // nope, don't know this fellow
br r2, CALL_PAL_SWPPAL_10_ // tis our buddy OSF
// .global osfpal_hw_entry_reset
// .weak osfpal_hw_entry_reset
// .long <osfpal_hw_entry_reset-pal_start>
//orig halt // don't know how to get the address here - kludge ok, load pal at 0
.long 0 // ?? hack upon hack...pb
CALL_PAL_SWPPAL_10_: ldl_p r3, 0(r2) // fetch target addr
// ble r3, swppal_fail ; if OSF not linked in say not loaded.
mfpr r2, pal_base // fetch pal base
addq r2, r3, r3 // add pal base
lda r2, 0x3FFF(r31) // get pal base checker mask
and r3, r2, r2 // any funky bits set?
cmpeq r2, 0, r0 //
blbc r0, swppal_fail // return unknown if bad bit set.
br r31, swppal_cont
// .sbttl "CALL_PAL OPCDECs"
CALL_PAL_PRIV(0x000B)
CallPal_OpcDec0B:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x000C)
CallPal_OpcDec0C:
br r31, osfpal_calpal_opcdec
//
// wripir - PALcode for wripir instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
// r16 = processor number to interrupt
//
// Function:
// IPIR <- R16
// Handled in system-specific code
//
// Exit:
// interprocessor interrupt is recorded on the target processor
// and is initiated when the proper enabling conditions are present.
//
CALL_PAL_PRIV(PAL_WRIPIR_ENTRY)
Call_Pal_Wrpir:
br r31, sys_wripir
// .sbttl "CALL_PAL OPCDECs"
CALL_PAL_PRIV(0x000E)
CallPal_OpcDec0E:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x000F)
CallPal_OpcDec0F:
br r31, osfpal_calpal_opcdec
//
// rdmces - PALcode for rdmces instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// R0 <- ZEXT(MCES)
//
CALL_PAL_PRIV(PAL_RDMCES_ENTRY)
Call_Pal_Rdmces:
mfpr r0, pt_mces // Read from PALtemp
and r0, mces_m_all, r0 // Clear other bits
hw_rei
//
// wrmces - PALcode for wrmces instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// If {R16<0> EQ 1} then MCES<0> <- 0 (MCHK)
// If {R16<1> EQ 1} then MCES<1> <- 0 (SCE)
// If {R16<2> EQ 1} then MCES<2> <- 0 (PCE)
// MCES<3> <- R16<3> (DPC)
// MCES<4> <- R16<4> (DSC)
//
//
CALL_PAL_PRIV(PAL_WRMCES_ENTRY)
Call_Pal_Wrmces:
and r16, ((1<<mces_v_mchk) | (1<<mces_v_sce) | (1<<mces_v_pce)), r13 // Isolate MCHK, SCE, PCE
mfpr r14, pt_mces // Get current value
ornot r31, r13, r13 // Flip all the bits
and r16, ((1<<mces_v_dpc) | (1<<mces_v_dsc)), r17
and r14, r13, r1 // Update MCHK, SCE, PCE
bic r1, ((1<<mces_v_dpc) | (1<<mces_v_dsc)), r1 // Clear old DPC, DSC
or r1, r17, r1 // Update DPC and DSC
mtpr r1, pt_mces // Write MCES back
nop // Pad to fix PT write->read restriction
nop
hw_rei
// CALL_PAL OPCDECs
CALL_PAL_PRIV(0x0012)
CallPal_OpcDec12:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0013)
CallPal_OpcDec13:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0014)
CallPal_OpcDec14:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0015)
CallPal_OpcDec15:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0016)
CallPal_OpcDec16:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0017)
CallPal_OpcDec17:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0018)
CallPal_OpcDec18:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0019)
CallPal_OpcDec19:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x001A)
CallPal_OpcDec1A:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x001B)
CallPal_OpcDec1B:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x001C)
CallPal_OpcDec1C:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x001D)
CallPal_OpcDec1D:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x001E)
CallPal_OpcDec1E:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x001F)
CallPal_OpcDec1F:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0020)
CallPal_OpcDec20:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0021)
CallPal_OpcDec21:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0022)
CallPal_OpcDec22:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0023)
CallPal_OpcDec23:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0024)
CallPal_OpcDec24:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0025)
CallPal_OpcDec25:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0026)
CallPal_OpcDec26:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0027)
CallPal_OpcDec27:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0028)
CallPal_OpcDec28:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x0029)
CallPal_OpcDec29:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x002A)
CallPal_OpcDec2A:
br r31, osfpal_calpal_opcdec
//
// wrfen - PALcode for wrfen instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// a0<0> -> ICSR<FPE>
// Store new FEN in PCB
// Final value of t0 (r1), t8..t10 (r22..r24) and a0 (r16)
// are UNPREDICTABLE
//
// Issue: What about pending FP loads when FEN goes from on->off????
//
CALL_PAL_PRIV(PAL_WRFEN_ENTRY)
Call_Pal_Wrfen:
or r31, 1, r13 // Get a one
mfpr r1, ev5__icsr // Get current FPE
sll r13, icsr_v_fpe, r13 // shift 1 to icsr<fpe> spot, e0
and r16, 1, r16 // clean new fen
sll r16, icsr_v_fpe, r12 // shift new fen to correct bit position
bic r1, r13, r1 // zero icsr<fpe>
or r1, r12, r1 // Or new FEN into ICSR
mfpr r12, pt_pcbb // Get PCBB - E1
mtpr r1, ev5__icsr // write new ICSR. 3 Bubble cycles to HW_REI
stl_p r16, osfpcb_q_fen(r12) // Store FEN in PCB.
mfpr r31, pt0 // Pad ICSR<FPE> write.
mfpr r31, pt0
mfpr r31, pt0
// pvc_violate 225 // cuz PVC can't distinguish which bits changed
hw_rei
CALL_PAL_PRIV(0x002C)
CallPal_OpcDec2C:
br r31, osfpal_calpal_opcdec
//
// wrvptpr - PALcode for wrvptpr instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// vptptr <- a0 (r16)
//
CALL_PAL_PRIV(PAL_WRVPTPTR_ENTRY)
Call_Pal_Wrvptptr:
mtpr r16, ev5__mvptbr // Load Mbox copy
mtpr r16, ev5__ivptbr // Load Ibox copy
nop // Pad IPR write
nop
hw_rei
CALL_PAL_PRIV(0x002E)
CallPal_OpcDec2E:
br r31, osfpal_calpal_opcdec
CALL_PAL_PRIV(0x002F)
CallPal_OpcDec2F:
br r31, osfpal_calpal_opcdec
//
// swpctx - PALcode for swpctx instruction
//
// Entry:
// hardware dispatch via callPal instruction
// R16 -> new pcb
//
// Function:
// dynamic state moved to old pcb
// new state loaded from new pcb
// pcbb pointer set
// old pcbb returned in R0
//
// Note: need to add perf monitor stuff
//
CALL_PAL_PRIV(PAL_SWPCTX_ENTRY)
Call_Pal_Swpctx:
rpcc r13 // get cyccounter
mfpr r0, pt_pcbb // get pcbb
ldq_p r22, osfpcb_q_fen(r16) // get new fen/pme
ldq_p r23, osfpcb_l_cc(r16) // get new asn
srl r13, 32, r25 // move offset
mfpr r24, pt_usp // get usp
stq_p r30, osfpcb_q_ksp(r0) // store old ksp
// pvc_violate 379 // stq_p can't trap except replay. only problem if mf same ipr in same shadow.
mtpr r16, pt_pcbb // set new pcbb
stq_p r24, osfpcb_q_usp(r0) // store usp
addl r13, r25, r25 // merge for new time
stl_p r25, osfpcb_l_cc(r0) // save time
ldah r24, (1<<(icsr_v_fpe-16))(r31)
and r22, 1, r12 // isolate fen
mfpr r25, icsr // get current icsr
lda r24, (1<<icsr_v_pmp)(r24)
br r31, swpctx_cont
//
// wrval - PALcode for wrval instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// sysvalue <- a0 (r16)
//
CALL_PAL_PRIV(PAL_WRVAL_ENTRY)
Call_Pal_Wrval:
nop
mtpr r16, pt_sysval // Pad paltemp write
nop
nop
hw_rei
//
// rdval - PALcode for rdval instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// v0 (r0) <- sysvalue
//
CALL_PAL_PRIV(PAL_RDVAL_ENTRY)
Call_Pal_Rdval:
nop
mfpr r0, pt_sysval
nop
hw_rei
//
// tbi - PALcode for tbi instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// TB invalidate
// r16/a0 = TBI type
// r17/a1 = Va for TBISx instructions
//
CALL_PAL_PRIV(PAL_TBI_ENTRY)
Call_Pal_Tbi:
addq r16, 2, r16 // change range to 0-2
br r23, CALL_PAL_tbi_10_ // get our address
CALL_PAL_tbi_10_: cmpult r16, 6, r22 // see if in range
lda r23, tbi_tbl-CALL_PAL_tbi_10_(r23) // set base to start of table
sll r16, 4, r16 // * 16
blbc r22, CALL_PAL_tbi_30_ // go rei, if not
addq r23, r16, r23 // addr of our code
//orig pvc_jsr tbi
jmp r31, (r23) // and go do it
CALL_PAL_tbi_30_:
hw_rei
nop
//
// wrent - PALcode for wrent instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Update ent* in paltemps
// r16/a0 = Address of entry routine
// r17/a1 = Entry Number 0..5
//
// r22, r23 trashed
//
CALL_PAL_PRIV(PAL_WRENT_ENTRY)
Call_Pal_Wrent:
cmpult r17, 6, r22 // see if in range
br r23, CALL_PAL_wrent_10_ // get our address
CALL_PAL_wrent_10_: bic r16, 3, r16 // clean pc
blbc r22, CALL_PAL_wrent_30_ // go rei, if not in range
lda r23, wrent_tbl-CALL_PAL_wrent_10_(r23) // set base to start of table
sll r17, 4, r17 // *16
addq r17, r23, r23 // Get address in table
//orig pvc_jsr wrent
jmp r31, (r23) // and go do it
CALL_PAL_wrent_30_:
hw_rei // out of range, just return
//
// swpipl - PALcode for swpipl instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// v0 (r0) <- PS<IPL>
// PS<IPL> <- a0<2:0> (r16)
//
// t8 (r22) is scratch
//
CALL_PAL_PRIV(PAL_SWPIPL_ENTRY)
Call_Pal_Swpipl:
and r16, osfps_m_ipl, r16 // clean New ipl
mfpr r22, pt_intmask // get int mask
extbl r22, r16, r22 // get mask for this ipl
bis r11, r31, r0 // return old ipl
bis r16, r31, r11 // set new ps
mtpr r22, ev5__ipl // set new mask
mfpr r31, pt0 // pad ipl write
mfpr r31, pt0 // pad ipl write
hw_rei // back
//
// rdps - PALcode for rdps instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// v0 (r0) <- ps
//
CALL_PAL_PRIV(PAL_RDPS_ENTRY)
Call_Pal_Rdps:
bis r11, r31, r0 // Fetch PALshadow PS
nop // Must be 2 cycles long
hw_rei
//
// wrkgp - PALcode for wrkgp instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// kgp <- a0 (r16)
//
CALL_PAL_PRIV(PAL_WRKGP_ENTRY)
Call_Pal_Wrkgp:
nop
mtpr r16, pt_kgp
nop // Pad for pt write->read restriction
nop
hw_rei
//
// wrusp - PALcode for wrusp instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// usp <- a0 (r16)
//
CALL_PAL_PRIV(PAL_WRUSP_ENTRY)
Call_Pal_Wrusp:
nop
mtpr r16, pt_usp
nop // Pad possible pt write->read restriction
nop
hw_rei
//
// wrperfmon - PALcode for wrperfmon instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
//
// Function:
// Various control functions for the onchip performance counters
//
// option selector in r16
// option argument in r17
// returned status in r0
//
//
// r16 = 0 Disable performance monitoring for one or more cpu's
// r17 = 0 disable no counters
// r17 = bitmask disable counters specified in bit mask (1=disable)
//
// r16 = 1 Enable performance monitoring for one or more cpu's
// r17 = 0 enable no counters
// r17 = bitmask enable counters specified in bit mask (1=enable)
//
// r16 = 2 Mux select for one or more cpu's
// r17 = Mux selection (cpu specific)
// <24:19> bc_ctl<pm_mux_sel> field (see spec)
// <31>,<7:4>,<3:0> pmctr <sel0>,<sel1>,<sel2> fields (see spec)
//
// r16 = 3 Options
// r17 = (cpu specific)
// <0> = 0 log all processes
// <0> = 1 log only selected processes
// <30,9,8> mode select - ku,kp,kk
//
// r16 = 4 Interrupt frequency select
// r17 = (cpu specific) indicates interrupt frequencies desired for each
// counter, with "zero interrupts" being an option
// frequency info in r17 bits as defined by PMCTR_CTL<FRQx> below
//
// r16 = 5 Read Counters
// r17 = na
// r0 = value (same format as ev5 pmctr)
// <0> = 0 Read failed
// <0> = 1 Read succeeded
//
// r16 = 6 Write Counters
// r17 = value (same format as ev5 pmctr; all counters written simultaneously)
//
// r16 = 7 Enable performance monitoring for one or more cpu's and reset counter to 0
// r17 = 0 enable no counters
// r17 = bitmask enable & clear counters specified in bit mask (1=enable & clear)
//
//=============================================================================
//Assumptions:
//PMCTR_CTL:
//
// <15:14> CTL0 -- encoded frequency select and enable - CTR0
// <13:12> CTL1 -- " - CTR1
// <11:10> CTL2 -- " - CTR2
//
// <9:8> FRQ0 -- frequency select for CTR0 (no enable info)
// <7:6> FRQ1 -- frequency select for CTR1
// <5:4> FRQ2 -- frequency select for CTR2
//
// <0> all vs. select processes (0=all,1=select)
//
// where
// FRQx<1:0>
// 0 1 disable interrupt
// 1 0 frequency = 65536 (16384 for ctr2)
// 1 1 frequency = 256
// note: FRQx<1:0> = 00 will keep counters from ever being enabled.
//
//=============================================================================
//
CALL_PAL_PRIV(0x0039)
// unsupported in Hudson code .. pboyle Nov/95
CALL_PAL_Wrperfmon:
// "real" performance monitoring code
cmpeq r16, 1, r0 // check for enable
bne r0, perfmon_en // br if requested to enable
cmpeq r16, 2, r0 // check for mux ctl
bne r0, perfmon_muxctl // br if request to set mux controls
cmpeq r16, 3, r0 // check for options
bne r0, perfmon_ctl // br if request to set options
cmpeq r16, 4, r0 // check for interrupt frequency select
bne r0, perfmon_freq // br if request to change frequency select
cmpeq r16, 5, r0 // check for counter read request
bne r0, perfmon_rd // br if request to read counters
cmpeq r16, 6, r0 // check for counter write request
bne r0, perfmon_wr // br if request to write counters
cmpeq r16, 7, r0 // check for counter clear/enable request
bne r0, perfmon_enclr // br if request to clear/enable counters
beq r16, perfmon_dis // br if requested to disable (r16=0)
br r31, perfmon_unknown // br if unknown request
//
// rdusp - PALcode for rdusp instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// v0 (r0) <- usp
//
CALL_PAL_PRIV(PAL_RDUSP_ENTRY)
Call_Pal_Rdusp:
nop
mfpr r0, pt_usp
hw_rei
CALL_PAL_PRIV(0x003B)
CallPal_OpcDec3B:
br r31, osfpal_calpal_opcdec
//
// whami - PALcode for whami instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// v0 (r0) <- whami
//
CALL_PAL_PRIV(PAL_WHAMI_ENTRY)
Call_Pal_Whami:
nop
mfpr r0, pt_whami // Get Whami
extbl r0, 1, r0 // Isolate just whami bits
hw_rei
//
// retsys - PALcode for retsys instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
// 00(sp) contains return pc
// 08(sp) contains r29
//
// Function:
// Return from system call.
// mode switched from kern to user.
// stacks swapped, ugp, upc restored.
// r23, r25 junked
//
CALL_PAL_PRIV(PAL_RETSYS_ENTRY)
Call_Pal_Retsys:
lda r25, osfsf_c_size(sp) // pop stack
bis r25, r31, r14 // touch r25 & r14 to stall mf exc_addr
mfpr r14, exc_addr // save exc_addr in case of fault
ldq r23, osfsf_pc(sp) // get pc
ldq r29, osfsf_gp(sp) // get gp
stl_c r31, -4(sp) // clear lock_flag
lda r11, 1<<osfps_v_mode(r31)// new PS:mode=user
mfpr r30, pt_usp // get users stack
bic r23, 3, r23 // clean return pc
mtpr r31, ev5__ipl // zero ibox IPL - 2 bubbles to hw_rei
mtpr r11, ev5__dtb_cm // set Mbox current mode - no virt ref for 2 cycles
mtpr r11, ev5__ps // set Ibox current mode - 2 bubble to hw_rei
mtpr r23, exc_addr // set return address - 1 bubble to hw_rei
mtpr r25, pt_ksp // save kern stack
rc r31 // clear inter_flag
// pvc_violate 248 // possible hidden mt->mf pt violation ok in callpal
hw_rei_spe // and back
CALL_PAL_PRIV(0x003E)
CallPal_OpcDec3E:
br r31, osfpal_calpal_opcdec
//
// rti - PALcode for rti instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// 00(sp) -> ps
// 08(sp) -> pc
// 16(sp) -> r29 (gp)
// 24(sp) -> r16 (a0)
// 32(sp) -> r17 (a1)
// 40(sp) -> r18 (a3)
//
CALL_PAL_PRIV(PAL_RTI_ENTRY)
/* called once by platform_tlaser */
.globl Call_Pal_Rti
Call_Pal_Rti:
lda r25, osfsf_c_size(sp) // get updated sp
bis r25, r31, r14 // touch r14,r25 to stall mf exc_addr
mfpr r14, exc_addr // save PC in case of fault
rc r31 // clear intr_flag
ldq r12, -6*8(r25) // get ps
ldq r13, -5*8(r25) // pc
ldq r18, -1*8(r25) // a2
ldq r17, -2*8(r25) // a1
ldq r16, -3*8(r25) // a0
ldq r29, -4*8(r25) // gp
bic r13, 3, r13 // clean return pc
stl_c r31, -4(r25) // clear lock_flag
and r12, osfps_m_mode, r11 // get mode
mtpr r13, exc_addr // set return address
beq r11, rti_to_kern // br if rti to Kern
br r31, rti_to_user // out of call_pal space
///////////////////////////////////////////////////
// Start the Unprivileged CALL_PAL Entry Points
///////////////////////////////////////////////////
//
// bpt - PALcode for bpt instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Build stack frame
// a0 <- code
// a1 <- unpred
// a2 <- unpred
// vector via entIF
//
//
//
.text 1
// . = 0x3000
CALL_PAL_UNPRIV(PAL_BPT_ENTRY)
Call_Pal_Bpt:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
bis r11, r31, r12 // Save PS for stack write
bge r25, CALL_PAL_bpt_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
CALL_PAL_bpt_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
mfpr r14, exc_addr // get pc
stq r16, osfsf_a0(sp) // save regs
bis r31, osf_a0_bpt, r16 // set a0
stq r17, osfsf_a1(sp) // a1
br r31, bpt_bchk_common // out of call_pal space
//
// bugchk - PALcode for bugchk instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Build stack frame
// a0 <- code
// a1 <- unpred
// a2 <- unpred
// vector via entIF
//
//
//
CALL_PAL_UNPRIV(PAL_BUGCHK_ENTRY)
Call_Pal_Bugchk:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
bis r11, r31, r12 // Save PS for stack write
bge r25, CALL_PAL_bugchk_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
CALL_PAL_bugchk_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
mfpr r14, exc_addr // get pc
stq r16, osfsf_a0(sp) // save regs
bis r31, osf_a0_bugchk, r16 // set a0
stq r17, osfsf_a1(sp) // a1
br r31, bpt_bchk_common // out of call_pal space
CALL_PAL_UNPRIV(0x0082)
CallPal_OpcDec82:
br r31, osfpal_calpal_opcdec
//
// callsys - PALcode for callsys instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Switch mode to kernel and build a callsys stack frame.
// sp = ksp
// gp = kgp
// t8 - t10 (r22-r24) trashed
//
//
//
CALL_PAL_UNPRIV(PAL_CALLSYS_ENTRY)
Call_Pal_Callsys:
and r11, osfps_m_mode, r24 // get mode
mfpr r22, pt_ksp // get ksp
beq r24, sys_from_kern // sysCall from kern is not allowed
mfpr r12, pt_entsys // get address of callSys routine
//
// from here on we know we are in user going to Kern
//
mtpr r31, ev5__dtb_cm // set Mbox current mode - no virt ref for 2 cycles
mtpr r31, ev5__ps // set Ibox current mode - 2 bubble to hw_rei
bis r31, r31, r11 // PS=0 (mode=kern)
mfpr r23, exc_addr // get pc
mtpr r30, pt_usp // save usp
lda sp, 0-osfsf_c_size(r22)// set new sp
stq r29, osfsf_gp(sp) // save user gp/r29
stq r24, osfsf_ps(sp) // save ps
stq r23, osfsf_pc(sp) // save pc
mtpr r12, exc_addr // set address
// 1 cycle to hw_rei
mfpr r29, pt_kgp // get the kern gp/r29
hw_rei_spe // and off we go!
CALL_PAL_UNPRIV(0x0084)
CallPal_OpcDec84:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0085)
CallPal_OpcDec85:
br r31, osfpal_calpal_opcdec
//
// imb - PALcode for imb instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Flush the writebuffer and flush the Icache
//
//
//
CALL_PAL_UNPRIV(PAL_IMB_ENTRY)
Call_Pal_Imb:
mb // Clear the writebuffer
mfpr r31, ev5__mcsr // Sync with clear
nop
nop
br r31, pal_ic_flush // Flush Icache
// CALL_PAL OPCDECs
CALL_PAL_UNPRIV(0x0087)
CallPal_OpcDec87:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0088)
CallPal_OpcDec88:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0089)
CallPal_OpcDec89:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x008A)
CallPal_OpcDec8A:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x008B)
CallPal_OpcDec8B:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x008C)
CallPal_OpcDec8C:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x008D)
CallPal_OpcDec8D:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x008E)
CallPal_OpcDec8E:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x008F)
CallPal_OpcDec8F:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0090)
CallPal_OpcDec90:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0091)
CallPal_OpcDec91:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0092)
CallPal_OpcDec92:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0093)
CallPal_OpcDec93:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0094)
CallPal_OpcDec94:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0095)
CallPal_OpcDec95:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0096)
CallPal_OpcDec96:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0097)
CallPal_OpcDec97:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0098)
CallPal_OpcDec98:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x0099)
CallPal_OpcDec99:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x009A)
CallPal_OpcDec9A:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x009B)
CallPal_OpcDec9B:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x009C)
CallPal_OpcDec9C:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x009D)
CallPal_OpcDec9D:
br r31, osfpal_calpal_opcdec
//
// rdunique - PALcode for rdunique instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// v0 (r0) <- unique
//
//
//
CALL_PAL_UNPRIV(PAL_RDUNIQUE_ENTRY)
CALL_PALrdunique_:
mfpr r0, pt_pcbb // get pcb pointer
ldq_p r0, osfpcb_q_unique(r0) // get new value
hw_rei
//
// wrunique - PALcode for wrunique instruction
//
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// unique <- a0 (r16)
//
//
//
CALL_PAL_UNPRIV(PAL_WRUNIQUE_ENTRY)
CALL_PAL_Wrunique:
nop
mfpr r12, pt_pcbb // get pcb pointer
stq_p r16, osfpcb_q_unique(r12)// get new value
nop // Pad palshadow write
hw_rei // back
// CALL_PAL OPCDECs
CALL_PAL_UNPRIV(0x00A0)
CallPal_OpcDecA0:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A1)
CallPal_OpcDecA1:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A2)
CallPal_OpcDecA2:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A3)
CallPal_OpcDecA3:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A4)
CallPal_OpcDecA4:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A5)
CallPal_OpcDecA5:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A6)
CallPal_OpcDecA6:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A7)
CallPal_OpcDecA7:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A8)
CallPal_OpcDecA8:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00A9)
CallPal_OpcDecA9:
br r31, osfpal_calpal_opcdec
//
// gentrap - PALcode for gentrap instruction
//
// CALL_PAL_gentrap:
// Entry:
// Vectored into via hardware PALcode instruction dispatch.
//
// Function:
// Build stack frame
// a0 <- code
// a1 <- unpred
// a2 <- unpred
// vector via entIF
//
//
CALL_PAL_UNPRIV(0x00AA)
// unsupported in Hudson code .. pboyle Nov/95
CALL_PAL_gentrap:
sll r11, 63-osfps_v_mode, r25 // Shift mode up to MS bit
mtpr r31, ev5__ps // Set Ibox current mode to kernel
bis r11, r31, r12 // Save PS for stack write
bge r25, CALL_PAL_gentrap_10_ // no stack swap needed if cm=kern
mtpr r31, ev5__dtb_cm // Set Mbox current mode to kernel -
// no virt ref for next 2 cycles
mtpr r30, pt_usp // save user stack
bis r31, r31, r11 // Set new PS
mfpr r30, pt_ksp
CALL_PAL_gentrap_10_:
lda sp, 0-osfsf_c_size(sp)// allocate stack space
mfpr r14, exc_addr // get pc
stq r16, osfsf_a0(sp) // save regs
bis r31, osf_a0_gentrap, r16// set a0
stq r17, osfsf_a1(sp) // a1
br r31, bpt_bchk_common // out of call_pal space
// CALL_PAL OPCDECs
CALL_PAL_UNPRIV(0x00AB)
CallPal_OpcDecAB:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00AC)
CallPal_OpcDecAC:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00AD)
CallPal_OpcDecAD:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00AE)
CallPal_OpcDecAE:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00AF)
CallPal_OpcDecAF:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B0)
CallPal_OpcDecB0:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B1)
CallPal_OpcDecB1:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B2)
CallPal_OpcDecB2:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B3)
CallPal_OpcDecB3:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B4)
CallPal_OpcDecB4:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B5)
CallPal_OpcDecB5:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B6)
CallPal_OpcDecB6:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B7)
CallPal_OpcDecB7:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B8)
CallPal_OpcDecB8:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00B9)
CallPal_OpcDecB9:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00BA)
CallPal_OpcDecBA:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00BB)
CallPal_OpcDecBB:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00BC)
CallPal_OpcDecBC:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00BD)
CallPal_OpcDecBD:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00BE)
CallPal_OpcDecBE:
br r31, osfpal_calpal_opcdec
CALL_PAL_UNPRIV(0x00BF)
CallPal_OpcDecBF:
// MODIFIED BY EGH 2/25/04
br r31, copypal_impl
/*======================================================================*/
/* OSF/1 CALL_PAL CONTINUATION AREA */
/*======================================================================*/
.text 2
. = 0x4000
// Continuation of MTPR_PERFMON
ALIGN_BLOCK
// "real" performance monitoring code
// mux ctl
perfmon_muxctl:
lda r8, 1(r31) // get a 1
sll r8, pmctr_v_sel0, r8 // move to sel0 position
or r8, ((0xf<<pmctr_v_sel1) | (0xf<<pmctr_v_sel2)), r8 // build mux select mask
and r17, r8, r25 // isolate pmctr mux select bits
mfpr r0, ev5__pmctr
bic r0, r8, r0 // clear old mux select bits
or r0,r25, r25 // or in new mux select bits
mtpr r25, ev5__pmctr
// ok, now tackle cbox mux selects
ldah r14, 0xfff0(r31)
zap r14, 0xE0, r14 // Get Cbox IPR base
//orig get_bc_ctl_shadow r16 // bc_ctl returned in lower longword
// adapted from ev5_pal_macros.mar
mfpr r16, pt_impure
lda r16, CNS_Q_IPR(r16)
RESTORE_SHADOW(r16,CNS_Q_BC_CTL,r16);
lda r8, 0x3F(r31) // build mux select mask
sll r8, bc_ctl_v_pm_mux_sel, r8
and r17, r8, r25 // isolate bc_ctl mux select bits
bic r16, r8, r16 // isolate old mux select bits
or r16, r25, r25 // create new bc_ctl
mb // clear out cbox for future ipr write
stq_p r25, ev5__bc_ctl(r14) // store to cbox ipr
mb // clear out cbox for future ipr write
//orig update_bc_ctl_shadow r25, r16 // r25=value, r16-overwritten with adjusted impure ptr
// adapted from ev5_pal_macros.mar