us3_cheetahplus_asm.s revision 7c478bd95313f5f23a4c958a745db2134aa03244
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License, Version 1.0 only
* (the "License"). You may not use this file except in compliance
* with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright 2005 Sun Microsystems, Inc. All rights reserved.
* Use is subject to license terms.
*
* Assembly code support for the Cheetah+ module
*/
#pragma ident "%Z%%M% %I% %E% SMI"
#if !defined(lint)
#include "assym.h"
#endif /* lint */
#include <sys/asm_linkage.h>
#include <sys/mmu.h>
#include <vm/hat_sfmmu.h>
#include <sys/machparam.h>
#include <sys/machcpuvar.h>
#include <sys/machthread.h>
#include <sys/machtrap.h>
#include <sys/privregs.h>
#include <sys/asm_linkage.h>
#include <sys/trap.h>
#include <sys/cheetahregs.h>
#include <sys/us3_module.h>
#include <sys/xc_impl.h>
#include <sys/intreg.h>
#include <sys/async.h>
#include <sys/clock.h>
#include <sys/cheetahasm.h>
#ifdef TRAPTRACE
#include <sys/traptrace.h>
#endif /* TRAPTRACE */
#if !defined(lint)
/* BEGIN CSTYLED */
/*
* Cheetah+ version to reflush an Ecache line by index.
*
* By default we assume the Ecache is 2-way so we flush both
* ways. Even if the cache is direct-mapped no harm will come
* from performing the flush twice, apart from perhaps a performance
* penalty.
*
* XXX - scr2 not used.
*/
#define ECACHE_REFLUSH_LINE(ec_set_size, index, scr2) \
ldxa [index]ASI_EC_DIAG, %g0; \
ldxa [index + ec_set_size]ASI_EC_DIAG, %g0;
/*
* Cheetah+ version of ecache_flush_line. Uses Cheetah+ Ecache Displacement
* Flush feature.
*/
#define ECACHE_FLUSH_LINE(physaddr, ec_set_size, scr1, scr2) \
sub ec_set_size, 1, scr1; \
and physaddr, scr1, scr1; \
set CHP_ECACHE_IDX_DISP_FLUSH, scr2; \
or scr2, scr1, scr1; \
ECACHE_REFLUSH_LINE(ec_set_size, scr1, scr2)
/* END CSTYLED */
/*
* Panther version to reflush a line from both the L2 cache and L3
* cache by the respective indexes. Flushes all ways of the line from
* each cache.
*
* l2_index Index into the L2$ of the line to be flushed. This
* register will not be modified by this routine.
* l3_index Index into the L3$ of the line to be flushed. This
* register will not be modified by this routine.
* scr2 scratch register.
* scr3 scratch register.
*
*/
#define PN_ECACHE_REFLUSH_LINE(l2_index, l3_index, scr2, scr3) \
set PN_L2_MAX_SET, scr2; \
set PN_L2_SET_SIZE, scr3; \
1: \
ldxa [l2_index + scr2]ASI_L2_TAG, %g0; \
cmp scr2, %g0; \
bg,a 1b; \
sub scr2, scr3, scr2; \
set PN_L3_MAX_SET, scr2; \
set PN_L3_SET_SIZE, scr3; \
2: \
ldxa [l3_index + scr2]ASI_EC_DIAG, %g0; \
cmp scr2, %g0; \
bg,a 2b; \
sub scr2, scr3, scr2;
/*
* Panther version of ecache_flush_line. Flushes the line corresponding
* to physaddr from both the L2 cache and the L3 cache.
*
* physaddr Input: Physical address to flush.
* Output: Physical address to flush (preserved).
* l2_idx_out Input: scratch register.
* Output: Index into the L2$ of the line to be flushed.
* l3_idx_out Input: scratch register.
* Output: Index into the L3$ of the line to be flushed.
* scr3 scratch register.
* scr4 scratch register.
*
*/
#define PN_ECACHE_FLUSH_LINE(physaddr, l2_idx_out, l3_idx_out, scr3, scr4) \
set PN_L3_SET_SIZE, l2_idx_out; \
sub l2_idx_out, 1, l2_idx_out; \
and physaddr, l2_idx_out, l3_idx_out; \
set PN_L3_IDX_DISP_FLUSH, l2_idx_out; \
or l2_idx_out, l3_idx_out, l3_idx_out; \
set PN_L2_SET_SIZE, l2_idx_out; \
sub l2_idx_out, 1, l2_idx_out; \
and physaddr, l2_idx_out, l2_idx_out; \
set PN_L2_IDX_DISP_FLUSH, scr3; \
or l2_idx_out, scr3, l2_idx_out; \
PN_ECACHE_REFLUSH_LINE(l2_idx_out, l3_idx_out, scr3, scr4)
#endif /* !lint */
/*
* Fast ECC error at TL>0 handler
* We get here via trap 70 at TL>0->Software trap 0 at TL>0. We enter
* this routine with %g1 and %g2 already saved in %tpc, %tnpc and %tstate.
* For a complete description of the Fast ECC at TL>0 handling see the
* comment block "Cheetah/Cheetah+ Fast ECC at TL>0 trap strategy" in
* us3_common_asm.s
*/
#if defined(lint)
void
fast_ecc_tl1_err(void)
{}
#else /* lint */
.section ".text"
.align 64
ENTRY_NP(fast_ecc_tl1_err)
/*
* This macro turns off the D$/I$ if they are on and saves their
* original state in ch_err_tl1_tmp, saves all the %g registers in the
* ch_err_tl1_data structure, updates the ch_err_tl1_flags and saves
* the %tpc in ch_err_tl1_tpc. At the end of this macro, %g1 will
* point to the ch_err_tl1_data structure and the original D$/I$ state
* will be saved in ch_err_tl1_tmp. All %g registers except for %g1
* will be available.
*/
CH_ERR_TL1_FECC_ENTER;
/*
* Get the diagnostic logout data. %g4 must be initialized to
* current CEEN state, %g5 must point to logout structure in
* ch_err_tl1_data_t. %g3 will contain the nesting count upon
* return.
*/
ldxa [%g0]ASI_ESTATE_ERR, %g4
and %g4, EN_REG_CEEN, %g4
add %g1, CH_ERR_TL1_LOGOUT, %g5
DO_TL1_CPU_LOGOUT(%g3, %g2, %g4, %g5, %g6, %g3, %g4)
/*
* If the logout nesting count is exceeded, we're probably
* not making any progress, try to panic instead.
*/
cmp %g3, CLO_NESTING_MAX
bge fecc_tl1_err
nop
/*
* Save the current CEEN and NCEEN state in %g7 and turn them off
* before flushing the Ecache.
*/
ldxa [%g0]ASI_ESTATE_ERR, %g7
andn %g7, EN_REG_CEEN | EN_REG_NCEEN, %g5
stxa %g5, [%g0]ASI_ESTATE_ERR
membar #Sync
/*
* Flush the Ecache, using the largest possible cache size with the
* smallest possible line size since we can't get the actual sizes
* from the cpu_node due to DTLB misses.
*/
PN_L2_FLUSHALL(%g3, %g4, %g5)
set CH_ECACHE_MAX_SIZE, %g4
set CH_ECACHE_MIN_LSIZE, %g5
GET_CPU_IMPL(%g6)
cmp %g6, PANTHER_IMPL
bne %xcc, 2f
nop
set PN_L3_SIZE, %g4
2:
mov %g6, %g3
CHP_ECACHE_FLUSHALL(%g4, %g5, %g3)
/*
* Restore CEEN and NCEEN to the previous state.
*/
stxa %g7, [%g0]ASI_ESTATE_ERR
membar #Sync
/*
* If we turned off the D$, then flush it and turn it back on.
*/
ldxa [%g1 + CH_ERR_TL1_TMP]%asi, %g3
andcc %g3, CH_ERR_TSTATE_DC_ON, %g0
bz %xcc, 3f
nop
/*
* Flush the D$.
*/
ASM_LD(%g4, dcache_size)
ASM_LD(%g5, dcache_linesize)
CH_DCACHE_FLUSHALL(%g4, %g5, %g6)
/*
* Turn the D$ back on.
*/
ldxa [%g0]ASI_DCU, %g3
or %g3, DCU_DC, %g3
stxa %g3, [%g0]ASI_DCU
membar #Sync
3:
/*
* If we turned off the I$, then flush it and turn it back on.
*/
ldxa [%g1 + CH_ERR_TL1_TMP]%asi, %g3
andcc %g3, CH_ERR_TSTATE_IC_ON, %g0
bz %xcc, 4f
nop
/*
* Flush the I$. Panther has different I$ parameters, and we
* can't access the logout I$ params without possibly generating
* a MMU miss.
*/
GET_CPU_IMPL(%g6)
set PN_ICACHE_SIZE, %g3
set CH_ICACHE_SIZE, %g4
mov CH_ICACHE_LSIZE, %g5
cmp %g6, PANTHER_IMPL
movz %xcc, %g3, %g4
movz %xcc, PN_ICACHE_LSIZE, %g5
CH_ICACHE_FLUSHALL(%g4, %g5, %g6, %g3)
/*
* Turn the I$ back on. Changing DCU_IC requires flush.
*/
ldxa [%g0]ASI_DCU, %g3
or %g3, DCU_IC, %g3
stxa %g3, [%g0]ASI_DCU
flush %g0
4:
#ifdef TRAPTRACE
/*
* Get current trap trace entry physical pointer.
*/
CPU_INDEX(%g6, %g5)
sll %g6, TRAPTR_SIZE_SHIFT, %g6
set trap_trace_ctl, %g5
add %g6, %g5, %g6
ld [%g6 + TRAPTR_LIMIT], %g5
tst %g5
be %icc, skip_traptrace
nop
ldx [%g6 + TRAPTR_PBASE], %g5
ld [%g6 + TRAPTR_OFFSET], %g4
add %g5, %g4, %g5
/*
* Create trap trace entry.
*/
rd %asi, %g7
wr %g0, TRAPTR_ASI, %asi
rd STICK, %g4
stxa %g4, [%g5 + TRAP_ENT_TICK]%asi
rdpr %tl, %g4
stha %g4, [%g5 + TRAP_ENT_TL]%asi
rdpr %tt, %g4
stha %g4, [%g5 + TRAP_ENT_TT]%asi
rdpr %tpc, %g4
stna %g4, [%g5 + TRAP_ENT_TPC]%asi
rdpr %tstate, %g4
stxa %g4, [%g5 + TRAP_ENT_TSTATE]%asi
stna %sp, [%g5 + TRAP_ENT_SP]%asi
stna %g0, [%g5 + TRAP_ENT_TR]%asi
wr %g0, %g7, %asi
ldxa [%g1 + CH_ERR_TL1_SDW_AFAR]%asi, %g3
ldxa [%g1 + CH_ERR_TL1_SDW_AFSR]%asi, %g4
wr %g0, TRAPTR_ASI, %asi
stna %g3, [%g5 + TRAP_ENT_F1]%asi
stna %g4, [%g5 + TRAP_ENT_F2]%asi
wr %g0, %g7, %asi
ldxa [%g1 + CH_ERR_TL1_AFAR]%asi, %g3
ldxa [%g1 + CH_ERR_TL1_AFSR]%asi, %g4
wr %g0, TRAPTR_ASI, %asi
stna %g3, [%g5 + TRAP_ENT_F3]%asi
stna %g4, [%g5 + TRAP_ENT_F4]%asi
wr %g0, %g7, %asi
/*
* Advance trap trace pointer.
*/
ld [%g6 + TRAPTR_OFFSET], %g5
ld [%g6 + TRAPTR_LIMIT], %g4
st %g5, [%g6 + TRAPTR_LAST_OFFSET]
add %g5, TRAP_ENT_SIZE, %g5
sub %g4, TRAP_ENT_SIZE, %g4
cmp %g5, %g4
movge %icc, 0, %g5
st %g5, [%g6 + TRAPTR_OFFSET]
skip_traptrace:
#endif /* TRAPTRACE */
/*
* If nesting count is not zero, skip all the AFSR/AFAR
* handling and just do the necessary cache-flushing.
*/
ldxa [%g1 + CH_ERR_TL1_NEST_CNT]%asi, %g2
brnz %g2, 6f
nop
/*
* If a UCU or L3_UCU followed by a WDU has occurred go ahead
* and panic since a UE will occur (on the retry) before the
* UCU and WDU messages are enqueued.
*/
ldxa [%g1 + CH_ERR_TL1_SDW_AFSR]%asi, %g3
set 1, %g4
sllx %g4, C_AFSR_UCU_SHIFT, %g4
btst %g4, %g3 ! UCU in original shadow AFSR?
bnz %xcc, 5f
mov 1, %g4
ldxa [%g1 + CH_ERR_TL1_SDW_AFSR_EXT]%asi, %g3
sllx %g4, C_AFSR_L3_UCU_SHIFT, %g4
btst %g4, %g3 ! L3_UCU in original shadow AFSR_EXT?
bz %xcc, 6f
nop
5:
ldxa [%g1 + CH_ERR_TL1_AFSR]%asi, %g4 ! original AFSR
ldxa [%g0]ASI_AFSR, %g3 ! current AFSR
or %g3, %g4, %g3 ! %g3 = original + current AFSR
set 1, %g4
sllx %g4, C_AFSR_WDU_SHIFT, %g4
btst %g4, %g3 ! WDU in original or current AFSR?
bnz %xcc, fecc_tl1_err
nop
6:
/*
* We fall into this macro if we've successfully logged the error in
* the ch_err_tl1_data structure and want the PIL15 softint to pick
* it up and log it. %g1 must point to the ch_err_tl1_data structure.
* Restores the %g registers and issues retry.
*/
CH_ERR_TL1_EXIT;
/*
* Establish panic exit label.
*/
CH_ERR_TL1_PANIC_EXIT(fecc_tl1_err);
SET_SIZE(fast_ecc_tl1_err)
#endif /* lint */
#if defined(lint)
/*
* scrubphys - Pass in the aligned physical memory address
* that you want to scrub, along with the ecache set size.
*
* 1) Displacement flush the E$ line corresponding to %addr.
* The first ldxa guarantees that the %addr is no longer in
* M, O, or E (goes to I or S (if instruction fetch also happens).
* 2) "Write" the data using a CAS %addr,%g0,%g0.
* The casxa guarantees a transition from I to M or S to M.
* 3) Displacement flush the E$ line corresponding to %addr.
* The second ldxa pushes the M line out of the ecache, into the
* writeback buffers, on the way to memory.
* 4) The "membar #Sync" pushes the cache line out of the writeback
* buffers onto the bus, on the way to dram finally.
*
* This is a modified version of the algorithm suggested by Gary Lauterbach.
* In theory the CAS %addr,%g0,%g0 is supposed to mark the addr's cache line
* as modified, but then we found out that for spitfire, if it misses in the
* E$ it will probably install as an M, but if it hits in the E$, then it
* will stay E, if the store doesn't happen. So the first displacement flush
* should ensure that the CAS will miss in the E$. Arrgh.
*/
/* ARGSUSED */
void
scrubphys(uint64_t paddr, int ecache_set_size)
{}
#else /* lint */
ENTRY(scrubphys)
rdpr %pstate, %o4
andn %o4, PSTATE_IE | PSTATE_AM, %o5
wrpr %o5, %g0, %pstate ! clear IE, AM bits
GET_CPU_IMPL(%o5) ! Panther Ecache is flushed differently
cmp %o5, PANTHER_IMPL
bne scrubphys_1
nop
PN_ECACHE_FLUSH_LINE(%o0, %o1, %o2, %o3, %o5)
casxa [%o0]ASI_MEM, %g0, %g0
PN_ECACHE_REFLUSH_LINE(%o1, %o2, %o3, %o0)
b scrubphys_2
nop
scrubphys_1:
ECACHE_FLUSH_LINE(%o0, %o1, %o2, %o3)
casxa [%o0]ASI_MEM, %g0, %g0
ECACHE_REFLUSH_LINE(%o1, %o2, %o3)
scrubphys_2:
wrpr %g0, %o4, %pstate ! restore earlier pstate register value
retl
membar #Sync ! move the data out of the load buffer
SET_SIZE(scrubphys)
#endif /* lint */
#if defined(lint)
/*
* clearphys - Pass in the aligned physical memory address
* that you want to push out, as a ecache_linesize byte block of zeros,
* from the ecache zero-filled.
*/
/* ARGSUSED */
void
clearphys(uint64_t paddr, int ecache_set_size, int ecache_linesize)
{
}
#else /* lint */
ENTRY(clearphys)
/* turn off IE, AM bits */
rdpr %pstate, %o4
andn %o4, PSTATE_IE | PSTATE_AM, %o5
wrpr %o5, %g0, %pstate
/* turn off NCEEN */
ldxa [%g0]ASI_ESTATE_ERR, %o5
andn %o5, EN_REG_NCEEN, %o3
stxa %o3, [%g0]ASI_ESTATE_ERR
membar #Sync
/* zero the E$ line */
clearphys_1:
subcc %o2, 8, %o2
bge clearphys_1
stxa %g0, [%o0 + %o2]ASI_MEM
GET_CPU_IMPL(%o3) ! Panther Ecache is flushed differently
cmp %o3, PANTHER_IMPL
bne clearphys_2
nop
PN_ECACHE_FLUSH_LINE(%o0, %o1, %o2, %o3, %g1)
casxa [%o0]ASI_MEM, %g0, %g0
PN_ECACHE_REFLUSH_LINE(%o1, %o2, %o3, %o0)
b clearphys_3
nop
clearphys_2:
ECACHE_FLUSH_LINE(%o0, %o1, %o2, %o3)
casxa [%o0]ASI_MEM, %g0, %g0
ECACHE_REFLUSH_LINE(%o1, %o2, %o3)
clearphys_3:
/* clear the AFSR */
ldxa [%g0]ASI_AFSR, %o1
stxa %o1, [%g0]ASI_AFSR
membar #Sync
/* turn NCEEN back on */
stxa %o5, [%g0]ASI_ESTATE_ERR
membar #Sync
/* return and re-enable IE and AM */
retl
wrpr %g0, %o4, %pstate
SET_SIZE(clearphys)
#endif /* lint */
#if defined(lint)
/*
* Cheetah+ Ecache displacement flush the specified line from the E$
*
* For Panther, this means flushing the specified line from both the
* L2 cache and L3 cache.
*
* Register usage:
* %o0 - 64 bit physical address for flushing
* %o1 - Ecache set size
*/
/*ARGSUSED*/
void
ecache_flush_line(uint64_t flushaddr, int ec_set_size)
{
}
#else /* lint */
ENTRY(ecache_flush_line)
GET_CPU_IMPL(%o3) ! Panther Ecache is flushed differently
cmp %o3, PANTHER_IMPL
bne ecache_flush_line_1
nop
PN_ECACHE_FLUSH_LINE(%o0, %o1, %o2, %o3, %o4)
b ecache_flush_line_2
nop
ecache_flush_line_1:
ECACHE_FLUSH_LINE(%o0, %o1, %o2, %o3)
ecache_flush_line_2:
retl
nop
SET_SIZE(ecache_flush_line)
#endif /* lint */
#if defined(lint)
void
set_afsr_ext(uint64_t afsr_ext)
{
afsr_ext = afsr_ext;
}
#else /* lint */
ENTRY(set_afsr_ext)
set ASI_AFSR_EXT_VA, %o1
stxa %o0, [%o1]ASI_AFSR ! afsr_ext reg
membar #Sync
retl
nop
SET_SIZE(set_afsr_ext)
#endif /* lint */
#if defined(lint)
/*
* The CPU jumps here from the MMU exception handler if an ITLB parity
* error is detected and we are running on Panther.
*
* In this routine we collect diagnostic information and write it to our
* logout structure (if possible) and clear all ITLB entries that may have
* caused our parity trap.
* Then we call cpu_tlb_parity_error via systrap in order to drop down to TL0
* and log any error messages. As for parameters to cpu_tlb_parity_error, we
* send two:
*
* %g2 - Contains the VA whose lookup in the ITLB caused the parity error
* %g3 - Contains the tlo_info field of the pn_tlb_logout logout struct,
* regardless of whether or not we actually used the logout struct.
*
* In the TL0 handler (cpu_tlb_parity_error) we will compare those two
* parameters to the data contained in the logout structure in order to
* determine whether the logout information is valid for this particular
* error or not.
*/
void
itlb_parity_trap(void)
{}
#else /* lint */
ENTRY_NP(itlb_parity_trap)
/*
* Collect important information about the trap which will be
* used as a parameter to the TL0 handler.
*/
wr %g0, ASI_IMMU, %asi
rdpr %tpc, %g2 ! VA that caused the IMMU trap
ldxa [MMU_TAG_ACCESS_EXT]%asi, %g3 ! read the trap VA page size
set PN_ITLB_PGSZ_MASK, %g4
and %g3, %g4, %g3
ldxa [MMU_TAG_ACCESS]%asi, %g4
set TAGREAD_CTX_MASK, %g5
and %g4, %g5, %g4
or %g4, %g3, %g3 ! 'or' in the trap context and
mov 1, %g4 ! add the IMMU flag to complete
sllx %g4, PN_TLO_INFO_IMMU_SHIFT, %g4
or %g4, %g3, %g3 ! the tlo_info field for logout
stxa %g0,[MMU_SFSR]%asi ! clear the SFSR
membar #Sync
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
*
* Next, we calculate the TLB index value for the failing VA.
*/
mov %g2, %g4 ! We need the ITLB index
set PN_ITLB_PGSZ_MASK, %g5
and %g3, %g5, %g5
srlx %g5, PN_ITLB_PGSZ_SHIFT, %g5
PN_GET_TLB_INDEX(%g4, %g5) ! %g4 has the index
sllx %g4, PN_TLB_ACC_IDX_SHIFT, %g4 ! shift the index into place
set PN_ITLB_T512, %g5
or %g4, %g5, %g4 ! and add in the TLB ID
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
* %g4 - contains the TLB access index value for the
* VA/PgSz in question
*
* Check to see if the logout structure is available.
*/
set CHPR_TLB_LOGOUT, %g6
GET_CPU_PRIVATE_PTR(%g6, %g1, %g5, itlb_parity_trap_1)
set LOGOUT_INVALID_U32, %g6
sllx %g6, 32, %g6 ! if our logout structure is
set LOGOUT_INVALID_L32, %g5 ! unavailable or if it is
or %g5, %g6, %g5 ! already being used, then we
ldx [%g1 + PN_TLO_ADDR], %g6 ! don't collect any diagnostic
cmp %g6, %g5 ! information before clearing
bne itlb_parity_trap_1 ! and logging the error.
nop
/*
* Record the logout information. %g4 contains our index + TLB ID
* for use in ASI_ITLB_ACCESS and ASI_ITLB_TAGREAD. %g1 contains
* the pointer to our logout struct.
*/
stx %g3, [%g1 + PN_TLO_INFO]
stx %g2, [%g1 + PN_TLO_ADDR]
stx %g2, [%g1 + PN_TLO_PC] ! %tpc == fault addr for IMMU
add %g1, PN_TLO_ITLB_TTE, %g1 ! move up the pointer
ldxa [%g4]ASI_ITLB_ACCESS, %g5 ! read the data
stx %g5, [%g1 + CH_TLO_TTE_DATA] ! store it away
ldxa [%g4]ASI_ITLB_TAGREAD, %g5 ! read the tag
stx %g5, [%g1 + CH_TLO_TTE_TAG] ! store it away
set PN_TLB_ACC_WAY_BIT, %g6 ! same thing again for way 1
or %g4, %g6, %g4
add %g1, CH_TLO_TTE_SIZE, %g1 ! move up the pointer
ldxa [%g4]ASI_ITLB_ACCESS, %g5 ! read the data
stx %g5, [%g1 + CH_TLO_TTE_DATA] ! store it away
ldxa [%g4]ASI_ITLB_TAGREAD, %g5 ! read the tag
stx %g5, [%g1 + CH_TLO_TTE_TAG] ! store it away
andn %g4, %g6, %g4 ! back to way 0
itlb_parity_trap_1:
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
* %g4 - contains the TLB access index value for the
* VA/PgSz in question
*
* Here we will clear the errors from the TLB.
*/
set MMU_TAG_ACCESS, %g5 ! We write a TTE tag value of
stxa %g0, [%g5]ASI_IMMU ! 0 as it will be invalid.
stxa %g0, [%g4]ASI_ITLB_ACCESS ! Write the data and tag
membar #Sync
set PN_TLB_ACC_WAY_BIT, %g6 ! same thing again for way 1
or %g4, %g6, %g4
stxa %g0, [%g4]ASI_ITLB_ACCESS ! Write same data and tag
membar #Sync
sethi %hi(FLUSH_ADDR), %g6 ! PRM says we need to issue a
flush %g6 ! flush after writing MMU regs
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
*
* Call cpu_tlb_parity_error via systrap at PIL 14 unless we're
* already at PIL 15. */
set cpu_tlb_parity_error, %g1
rdpr %pil, %g4
cmp %g4, PIL_14
movl %icc, PIL_14, %g4
ba sys_trap
nop
SET_SIZE(itlb_parity_trap)
#endif /* lint */
#if defined(lint)
/*
* The CPU jumps here from the MMU exception handler if a DTLB parity
* error is detected and we are running on Panther.
*
* In this routine we collect diagnostic information and write it to our
* logout structure (if possible) and clear all DTLB entries that may have
* caused our parity trap.
* Then we call cpu_tlb_parity_error via systrap in order to drop down to TL0
* and log any error messages. As for parameters to cpu_tlb_parity_error, we
* send two:
*
* %g2 - Contains the VA whose lookup in the DTLB caused the parity error
* %g3 - Contains the tlo_info field of the pn_tlb_logout logout struct,
* regardless of whether or not we actually used the logout struct.
*
* In the TL0 handler (cpu_tlb_parity_error) we will compare those two
* parameters to the data contained in the logout structure in order to
* determine whether the logout information is valid for this particular
* error or not.
*/
void
dtlb_parity_trap(void)
{}
#else /* lint */
ENTRY_NP(dtlb_parity_trap)
/*
* Collect important information about the trap which will be
* used as a parameter to the TL0 handler.
*/
wr %g0, ASI_DMMU, %asi
ldxa [MMU_SFAR]%asi, %g2 ! VA that caused the IMMU trap
ldxa [MMU_TAG_ACCESS_EXT]%asi, %g3 ! read the trap VA page sizes
set PN_DTLB_PGSZ_MASK, %g4
and %g3, %g4, %g3
ldxa [MMU_TAG_ACCESS]%asi, %g4
set TAGREAD_CTX_MASK, %g5 ! 'or' in the trap context
and %g4, %g5, %g4 ! to complete the tlo_info
or %g4, %g3, %g3 ! field for logout
stxa %g0,[MMU_SFSR]%asi ! clear the SFSR
membar #Sync
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
*
* Calculate the TLB index values for the failing VA. Since the T512
* TLBs can be configured for different page sizes, we need to find
* the index into each one separately.
*/
mov %g2, %g4 ! First we get the DTLB_0 index
set PN_DTLB_PGSZ0_MASK, %g5
and %g3, %g5, %g5
srlx %g5, PN_DTLB_PGSZ0_SHIFT, %g5
PN_GET_TLB_INDEX(%g4, %g5) ! %g4 has the DTLB_0 index
sllx %g4, PN_TLB_ACC_IDX_SHIFT, %g4 ! shift the index into place
set PN_DTLB_T512_0, %g5
or %g4, %g5, %g4 ! and add in the TLB ID
mov %g2, %g7 ! Next we get the DTLB_1 index
set PN_DTLB_PGSZ1_MASK, %g5
and %g3, %g5, %g5
srlx %g5, PN_DTLB_PGSZ1_SHIFT, %g5
PN_GET_TLB_INDEX(%g7, %g5) ! %g7 has the DTLB_1 index
sllx %g7, PN_TLB_ACC_IDX_SHIFT, %g7 ! shift the index into place
set PN_DTLB_T512_1, %g5
or %g7, %g5, %g7 ! and add in the TLB ID
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
* %g4 - contains the T512_0 access index value for the
* VA/PgSz in question
* %g7 - contains the T512_1 access index value for the
* VA/PgSz in question
*
* If this trap happened at TL>0, then we don't want to mess
* with the normal logout struct since that could caused a TLB
* miss.
*/
rdpr %tl, %g6 ! read current trap level
cmp %g6, 1 ! skip over the tl>1 code
ble dtlb_parity_trap_1 ! if TL <= 1.
nop
/*
* If we are here, then the trap happened at TL>1. Simply
* update our tlo_info field and then skip to the TLB flush
* code.
*/
mov 1, %g6
sllx %g6, PN_TLO_INFO_TL1_SHIFT, %g6
or %g6, %g3, %g3
ba dtlb_parity_trap_2
nop
dtlb_parity_trap_1:
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
* %g4 - contains the T512_0 access index value for the
* VA/PgSz in question
* %g7 - contains the T512_1 access index value for the
* VA/PgSz in question
*
* Check to see if the logout structure is available.
*/
set CHPR_TLB_LOGOUT, %g6
GET_CPU_PRIVATE_PTR(%g6, %g1, %g5, dtlb_parity_trap_2)
set LOGOUT_INVALID_U32, %g6
sllx %g6, 32, %g6 ! if our logout structure is
set LOGOUT_INVALID_L32, %g5 ! unavailable or if it is
or %g5, %g6, %g5 ! already being used, then we
ldx [%g1 + PN_TLO_ADDR], %g6 ! don't collect any diagnostic
cmp %g6, %g5 ! information before clearing
bne dtlb_parity_trap_2 ! and logging the error.
nop
/*
* Record the logout information. %g4 contains our DTLB_0
* index + TLB ID and %g7 contains our DTLB_1 index + TLB ID
* both of which will be used for ASI_DTLB_ACCESS and
* ASI_DTLB_TAGREAD. %g1 contains the pointer to our logout
* struct.
*/
stx %g3, [%g1 + PN_TLO_INFO]
stx %g2, [%g1 + PN_TLO_ADDR]
rdpr %tpc, %g5
stx %g5, [%g1 + PN_TLO_PC]
add %g1, PN_TLO_DTLB_TTE, %g1 ! move up the pointer
ldxa [%g4]ASI_DTLB_ACCESS, %g5 ! read the data from DTLB_0
stx %g5, [%g1 + CH_TLO_TTE_DATA] ! way 0 and store it away
ldxa [%g4]ASI_DTLB_TAGREAD, %g5 ! read the tag from DTLB_0
stx %g5, [%g1 + CH_TLO_TTE_TAG] ! way 0 and store it away
ldxa [%g7]ASI_DTLB_ACCESS, %g5 ! now repeat for DTLB_1 way 0
stx %g5, [%g1 + (CH_TLO_TTE_DATA + (CH_TLO_TTE_SIZE * 2))]
ldxa [%g7]ASI_DTLB_TAGREAD, %g5
stx %g5, [%g1 + (CH_TLO_TTE_TAG + (CH_TLO_TTE_SIZE * 2))]
set PN_TLB_ACC_WAY_BIT, %g6 ! same thing again for way 1
or %g4, %g6, %g4 ! of each TLB.
or %g7, %g6, %g7
add %g1, CH_TLO_TTE_SIZE, %g1 ! move up the pointer
ldxa [%g4]ASI_DTLB_ACCESS, %g5 ! read the data from DTLB_0
stx %g5, [%g1 + CH_TLO_TTE_DATA] ! way 1 and store it away
ldxa [%g4]ASI_DTLB_TAGREAD, %g5 ! read the tag from DTLB_0
stx %g5, [%g1 + CH_TLO_TTE_TAG] ! way 1 and store it away
ldxa [%g7]ASI_DTLB_ACCESS, %g5 ! now repeat for DTLB_1 way 1
stx %g5, [%g1 + (CH_TLO_TTE_DATA + (CH_TLO_TTE_SIZE * 2))]
ldxa [%g7]ASI_DTLB_TAGREAD, %g5
stx %g5, [%g1 + (CH_TLO_TTE_TAG + (CH_TLO_TTE_SIZE * 2))]
andn %g4, %g6, %g4 ! back to way 0
andn %g7, %g6, %g7 ! back to way 0
dtlb_parity_trap_2:
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
* %g4 - contains the T512_0 access index value for the
* VA/PgSz in question
* %g7 - contains the T512_1 access index value for the
* VA/PgSz in question
*
* Here we will clear the errors from the DTLB.
*/
set MMU_TAG_ACCESS, %g5 ! We write a TTE tag value of
stxa %g0, [%g5]ASI_DMMU ! 0 as it will be invalid.
stxa %g0, [%g4]ASI_DTLB_ACCESS ! Write the data and tag.
stxa %g0, [%g7]ASI_DTLB_ACCESS ! Now repeat for DTLB_1 way 0
membar #Sync
set PN_TLB_ACC_WAY_BIT, %g6 ! same thing again for way 1
or %g4, %g6, %g4
or %g7, %g6, %g7
stxa %g0, [%g4]ASI_DTLB_ACCESS ! Write same data and tag.
stxa %g0, [%g7]ASI_DTLB_ACCESS ! Now repeat for DTLB_1 way 0
membar #Sync
sethi %hi(FLUSH_ADDR), %g6 ! PRM says we need to issue a
flush %g6 ! flush after writing MMU regs
/*
* at this point:
* %g2 - contains the VA whose lookup caused the trap
* %g3 - contains the tlo_info field
*
* Call cpu_tlb_parity_error via systrap at PIL 14 unless we're
* already at PIL 15. We do this even for TL>1 traps since
* those will lead to a system panic.
*/
set cpu_tlb_parity_error, %g1
rdpr %pil, %g4
cmp %g4, PIL_14
movl %icc, PIL_14, %g4
ba sys_trap
nop
SET_SIZE(dtlb_parity_trap)
#endif /* lint */
#if defined(lint)
/*
* Calculates the Panther TLB index based on a virtual address and page size
*
* Register usage:
* %o0 - virtual address whose index we want
* %o1 - Page Size of the TLB in question as encoded in the
* ASI_[D|I]MMU_TAG_ACCESS_EXT register.
*/
uint64_t
pn_get_tlb_index(uint64_t va, uint64_t pg_sz)
{
return ((va + pg_sz)-(va + pg_sz));
}
#else /* lint */
ENTRY(pn_get_tlb_index)
PN_GET_TLB_INDEX(%o0, %o1)
retl
nop
SET_SIZE(pn_get_tlb_index)
#endif /* lint */
#if defined(lint)
/*
* For Panther CPUs we need to flush the IPB after any I$ or D$
* parity errors are detected.
*/
void
flush_ipb(void)
{ return; }
#else /* lint */
ENTRY(flush_ipb)
clr %o0
flush_ipb_1:
stxa %g0, [%o0]ASI_IPB_TAG
membar #Sync
cmp %o0, PN_IPB_TAG_ADDR_MAX
blt flush_ipb_1
add %o0, PN_IPB_TAG_ADDR_LINESIZE, %o0
sethi %hi(FLUSH_ADDR), %o0
flush %o0
retl
nop
SET_SIZE(flush_ipb)
#endif /* lint */