/*
- * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2005 Apple Computer, Inc. All rights reserved.
*
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_LICENSE_OSREFERENCE_HEADER_START@
*
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License"). You may not use this file except in compliance with the
- * License. Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
- *
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
- * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
- * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
- * License for the specific language governing rights and limitations
- * under the License.
- *
- * @APPLE_LICENSE_HEADER_END@
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the
+ * License may not be used to create, or enable the creation or
+ * redistribution of, unlawful or unlicensed copies of an Apple operating
+ * system, or to circumvent, violate, or enable the circumvention or
+ * violation of, any terms of an Apple operating system software license
+ * agreement.
+ *
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this
+ * file.
+ *
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
+ * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
+ *
+ * @APPLE_LICENSE_OSREFERENCE_HEADER_END@
*/
/*
* @OSF_COPYRIGHT@
#include <debug.h>
#include <mach_assert.h>
#include <mach/exception_types.h>
+#include <mach/kern_return.h>
#include <mach/ppc/vm_param.h>
#include <assym.s>
#include <ppc/proc_reg.h>
#include <ppc/trap.h>
#include <ppc/exception.h>
-#include <ppc/spl.h>
+#include <ppc/savearea.h>
#define VERIFYSAVE 0
#define FPVECDBG 0
-
+#define FPFLOOD 0
+#define INSTRUMENT 0
+
/*
* thandler(type)
*
/*
* If pcb.ksp == 0 then the kernel stack is already busy,
- * we save ppc_saved state below the current stack pointer,
+ * we make a stack frame
* leaving enough space for the 'red zone' in case the
* trapped thread was in the middle of saving state below
* its stack pointer.
*
- * otherwise we save a ppc_saved_state in the pcb, and switch to
+ * otherwise we make a stack frame and
* the kernel stack (setting pcb.ksp to 0)
*
* on return, we do the reverse, the last state is popped from the pcb
.align 5
.globl EXT(thandler)
-LEXT(thandler) /* Trap handler */
-
-#if 0
-;
-; NOTE: This trap will hang VPC running Windows98 (and probably others)...
-;
- lwz r25,savedar(r4) ; (TEST/DEBUG)
- cmplwi r25,0x298 ; (TEST/DEBUG)
-
-deadloop: addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- addi r25,r25,1 ; (TEST/DEBUG)
- beq- deadloop ; (TEST/DEBUG)
-#endif
+LEXT(thandler) ; Trap handler
- mfsprg r25,0 /* Get the per_proc */
+ mfsprg r13,1 ; Get the current activation
+ lwz r25,ACT_PER_PROC(r13) ; Get the per_proc block
- lwz r1,PP_ISTACKPTR(r25) ; Get interrupt stack pointer
+ lwz r1,PP_ISTACKPTR(r25) ; Get interrupt stack pointer
- lwz r6,PP_CPU_DATA(r25) /* Get point to cpu specific data */
- cmpwi cr0,r1,0 ; Are we on interrupt stack?
- lwz r6,CPU_ACTIVE_THREAD(r6) /* Get the pointer to the currently active thread */
- beq- cr0,EXT(ihandler) ; If on interrupt stack, treat this as interrupt...
- lwz r13,THREAD_TOP_ACT(r6) /* Point to the active activation */
- lwz r26,ACT_MACT_SPF(r13) ; Get special flags
- lwz r8,ACT_MACT_PCB(r13) /* Get the last savearea used */
+ cmpwi cr0,r1,0 ; Are we on interrupt stack?
+ mr r6,r13
+ beq- cr0,EXT(ihandler) ; If on interrupt stack, treat this as interrupt...
+ lwz r26,ACT_MACT_SPF(r13) ; Get special flags
+ lwz r8,ACT_MACT_PCB(r13) ; Get the last savearea used
rlwinm. r26,r26,0,bbThreadbit,bbThreadbit ; Do we have Blue Box Assist active?
- lwz r1,ACT_MACT_KSP(r13) ; Get the top of kernel stack
- bnel- checkassist /* See if we should assist this */
- stw r4,ACT_MACT_PCB(r13) /* Point to our savearea */
- stw r8,SAVprev(r4) /* Queue the new save area in the front */
+ lwz r1,ACT_MACT_KSP(r13) ; Get the top of kernel stack
+ bnel- checkassist ; See if we should assist this
+ stw r4,ACT_MACT_PCB(r13) ; Point to our savearea
+ stw r8,SAVprev+4(r4) ; Queue the new save area in the front
#if VERIFYSAVE
- bl versave ; (TEST/DEBUG)
+ bl versave ; (TEST/DEBUG)
#endif
- lwz r9,THREAD_KERNEL_STACK(r6) ; Get our kernel stack start
- cmpwi cr1,r1,0 ; Are we already on kernel stack?
- stw r13,SAVact(r4) ; Mark the savearea as belonging to this activation
- lwz r26,saver1(r4) ; Get the stack at interrupt time
+ lwz r9,THREAD_KERNEL_STACK(r6) ; Get our kernel stack start
+ cmpwi cr1,r1,0 ; Are we already on kernel stack?
+ stw r13,SAVact(r4) ; Mark the savearea as belonging to this activation
+ lwz r26,saver1+4(r4) ; Get the stack at interrupt time
- bne+ cr1,.L_kstackfree ; We are not on kernel stack yet...
+ bne+ cr1,.L_kstackfree ; We are not on kernel stack yet...
- subi r1,r26,FM_REDZONE ; Make a red zone on interrupt time kernel stack
+ subi r1,r26,FM_REDZONE ; Make a red zone on interrupt time kernel stack
.L_kstackfree:
- lwz r7,savesrr1(r4) /* Pick up the entry MSR */
- sub r9,r1,r9 ; Get displacment into the kernel stack
- li r0,0 /* Make this 0 */
- cmplwi cr2,r9,KERNEL_STACK_SIZE ; Do we still have room on the stack?
- beq cr1,.L_state_on_kstack /* using above test for pcb/stack */
+ lwz r31,savesrr1+4(r4) ; Pick up the entry MSR
+ sub r9,r1,r9 ; Get displacment into the kernel stack
+ li r0,0 ; Make this 0
+ rlwinm. r0,r9,0,28,31 ; Verify that we have a 16-byte aligned stack (and get a 0)
+ cmplwi cr2,r9,KERNEL_STACK_SIZE ; Do we still have room on the stack?
+ beq cr1,.L_state_on_kstack ; using above test for pcb/stack
- stw r0,ACT_MACT_KSP(r13) /* Show that we have taken the stack */
+ stw r0,ACT_MACT_KSP(r13) ; Show that we have taken the stack
.L_state_on_kstack:
- rlwinm. r6,r7,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on?
- bgt- cr2,kernelStackBad ; Kernel stack is bogus...
-kernelStackNotBad: ; Not really
- beq+ tvecoff ; Vector off, do not save vrsave...
- lwz r3,savevrsave(r4) ; Get the VRSAVE register
- stw r3,liveVRS(r25) ; Set the live value
-
-tvecoff: rlwinm. r3,r7,0,MSR_FP_BIT,MSR_FP_BIT ; Was floating point on?
- subi r1,r1,FM_SIZE /* Push a header onto the current stack */
- beq+ tfpoff /* Floating point was off... */
- lwz r3,savexfpscr(r4) ; Grab the just saved FPSCR
- stw r3,liveFPSCR(r25) ; Make it the live copy
-
-tfpoff: stw r26,FM_BACKPTR(r1) ; Link back to the previous frame
+ lwz r9,savevrsave(r4) ; Get the VRSAVE register
+ bne-- kernelStackUnaligned ; Stack is unaligned...
+ rlwinm. r6,r31,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on?
+ subi r1,r1,FM_SIZE ; Push a header onto the current stack
+ bgt-- cr2,kernelStackBad ; Kernel stack is bogus...
+
+kernelStackNotBad: ; Vector was off
+ beq++ tvecoff ; Vector off, do not save vrsave...
+ stw r9,liveVRS(r25) ; Set the live value
+
+tvecoff: stw r26,FM_BACKPTR(r1) ; Link back to the previous frame
#if DEBUG
/* If debugging, we need two frames, the first being a dummy
* which links back to the trapped routine. The second is
* that which the C routine below will need
*/
- lwz r3,savesrr0(r4) /* Get the point of interruption */
- stw r3,FM_LR_SAVE(r1) /* save old instr ptr as LR value */
- stwu r1, -FM_SIZE(r1) /* and make new frame */
+ lwz r3,savesrr0+4(r4) ; Get the point of interruption
+ stw r3,FM_LR_SAVE(r1) ; save old instr ptr as LR value
+ stwu r1, -FM_SIZE(r1) ; and make new frame
#endif /* DEBUG */
+ mr r30,r4
+ lwz r3,SAVtime+4(r4)
+ addi r4,r13,SYSTEM_TIMER
+ bl EXT(timer_event)
/* call trap handler proper, with
- * ARG0 = type (not yet, holds pcb ptr)
- * ARG1 = saved_state ptr (already there)
- * ARG2 = dsisr (already there)
- * ARG3 = dar (already there)
+ * ARG0 = type
+ * ARG1 = saved_state ptr
+ * ARG2 = dsisr
+ * ARG3 = dar
*/
-
- lwz r3,saveexception(r4) /* Get the exception code */
- lwz r0,ACT_MACT_SPF(r13) ; Get the special flags
+ mr r4,r30
+ lwz r3,saveexception(r30) ; Get the exception code
+ lwz r0,ACT_MACT_SPF(r13) ; Get the special flags
- addi r5,r3,-T_DATA_ACCESS ; Adjust to start of range
+ addi r5,r3,-T_DATA_ACCESS ; Adjust to start of range
rlwinm. r0,r0,0,runningVMbit,runningVMbit ; Are we in VM state? (cr0_eq == 0 if yes)
- cmplwi cr2,r5,T_RUNMODE_TRACE-T_DATA_ACCESS ; Are we still in range? (cr_gt if not)
+ cmplwi cr2,r5,T_TRACE-T_DATA_ACCESS ; Are we still in range? (cr_gt if not)
- lwz r5,savedsisr(r4) /* Get the saved DSISR */
+ lwz r5,savedsisr(r4) ; Get the saved DSISR
- crnor cr7_eq,cr0_eq,cr2_gt ; We should intercept if in VM and is a true trap (cr7_eq == 1 if yes)
- rlwinm. r0,r7,0,MSR_PR_BIT,MSR_PR_BIT ; Are we trapping from supervisor state? (cr0_eq == 1 if yes)
+ crnor cr7_eq,cr0_eq,cr2_gt ; We should intercept if in VM and is a true trap (cr7_eq == 1 if yes)
+ rlwinm. r0,r31,0,MSR_PR_BIT,MSR_PR_BIT ; Are we trapping from supervisor state? (cr0_eq == 1 if yes)
- cmpi cr2,r3,T_PREEMPT ; Is this a preemption?
+ cmpi cr2,r3,T_PREEMPT ; Is this a preemption?
+
+ beq-- .L_check_VM
+ stw r4,ACT_MACT_UPCB(r13) ; Store user savearea
+.L_check_VM:
- crandc cr0_eq,cr7_eq,cr0_eq ; Do not intercept if we are in the kernel (cr0_eq == 1 if yes)
+ crandc cr0_eq,cr7_eq,cr0_eq ; Do not intercept if we are in the kernel (cr0_eq == 1 if yes)
- lwz r6,savedar(r4) /* Get the DAR */
+ lwz r6,savedar(r4) ; Get the DAR (top)
+ lwz r7,savedar+4(r4) ; Get the DAR (bottom)
- beq- cr2, .L_call_trap /* Don't turn on interrupts for T_PREEMPT */
- beq- exitFromVM ; Any true trap but T_MACHINE_CHECK exits us from the VM...
+ beq- cr2,.L_call_trap ; Do not turn on interrupts for T_PREEMPT
+ beq- exitFromVM ; Any true trap but T_MACHINE_CHECK exits us from the VM...
/* syscall exception might warp here if there's nothing left
* to do except generate a trap
*/
.L_call_trap:
-#if 0
- lis r0,HIGH_ADDR(CutTrace) /* (TEST/DEBUG) */
- oris r0,r0,LOW_ADDR(CutTrace) /* (TEST/DEBUG) */
- sc /* (TEST/DEBUG) */
-#endif
+#if FPFLOOD
+ stfd f31,emfp31(r25) ; (TEST/DEBUG)
+#endif
+
bl EXT(trap)
+ lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable
+ mfmsr r7 ; Get the MSR
+ ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE
+ andc r7,r7,r10 ; Turn off VEC, FP, and EE
+ mtmsr r7 ; Disable for interrupts
+ mfsprg r8,1 ; Get the current activation
+ lwz r10,ACT_PER_PROC(r8) ; Get the per_proc block
/*
- * Ok, return from C function
- *
* This is also the point where new threads come when they are created.
* The new thread is setup to look like a thread that took an
* interrupt and went immediatly into trap.
- *
*/
thread_return:
+ lwz r11,SAVflags(r3) ; Get the flags of the current savearea
+ lwz r0,savesrr1+4(r3) ; Get the MSR we are going to
+ lwz r4,SAVprev+4(r3) ; Pick up the previous savearea
+ mfsprg r8,1 ; Get the current thread
+ rlwinm r11,r11,0,15,13 ; Clear the syscall flag
+ rlwinm. r0,r0,0,MSR_PR_BIT,MSR_PR_BIT ; Are we going to the user?
+ mr r1,r8
+ stw r11,SAVflags(r3) ; Save back the flags (with reset stack cleared)
+
+ lwz r5,THREAD_KERNEL_STACK(r1) ; Get the base pointer to the stack
+ stw r4,ACT_MACT_PCB(r8) ; Point to the previous savearea (or 0 if none)
+ addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE ; Reset to empty
- mfmsr r7 /* Get the MSR */
- lwz r4,SAVprev(r3) /* Pick up the previous savearea */
- rlwinm r7,r7,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Clear the interrupt enable mask */
- lwz r11,SAVflags(r3) /* Get the flags of the current savearea */
- mtmsr r7 /* Disable for interrupts */
-
- mfsprg r10,0 /* Restore the per_proc info */
-
- lwz r8,savesrr1(r3) ; Get the MSR we are going to
- lwz r1,PP_CPU_DATA(r10) /* Get the CPU data area */
- rlwinm r11,r11,0,15,13 /* Clear the syscall flag */
- lwz r1,CPU_ACTIVE_THREAD(r1) /* and the active thread */
- rlwinm. r8,r8,0,MSR_PR_BIT,MSR_PR_BIT ; Are we going to the user?
- lwz r8,THREAD_TOP_ACT(r1) /* Now find the current activation */
- stw r11,SAVflags(r3) /* Save back the flags (with reset stack cleared) */
-
-#if 0
- lis r0,HIGH_ADDR(CutTrace) /* (TEST/DEBUG) */
- oris r0,r0,LOW_ADDR(CutTrace) /* (TEST/DEBUG) */
- sc /* (TEST/DEBUG) */
-#endif
- stw r4,ACT_MACT_PCB(r8) /* Point to the previous savearea (or 0 if none) */
-
- beq- chkfac ; We are not leaving the kernel yet...
+ beq-- chkfac ; We are not leaving the kernel yet...
- lwz r5,THREAD_KERNEL_STACK(r1) /* Get the base pointer to the stack */
- addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE /* Reset to empty */
- stw r5,ACT_MACT_KSP(r8) /* Save the empty stack pointer */
- b chkfac /* Go end it all... */
+ stw r5,ACT_MACT_KSP(r8) ; Save the empty stack pointer
+ b chkfac ; Go end it all...
;
kernelStackBad:
- lwz r3,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top
+ lwz r3,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top
subi r3,r3,KERNEL_STACK_SIZE-FM_SIZE ; Adjust to start of stack
- sub r3,r1,r3 ; Get displacement into debug stack
+ sub r3,r1,r3 ; Get displacement into debug stack
cmplwi cr2,r3,KERNEL_STACK_SIZE-FM_SIZE ; Check if we are on debug stack
- blt+ cr2,kernelStackNotBad ; Yeah, that is ok too...
+ blt+ cr2,kernelStackNotBad ; Yeah, that is ok too...
+
+ lis r0,hi16(Choke) ; Choke code
+ ori r0,r0,lo16(Choke) ; and the rest
+ li r3,failStack ; Bad stack code
+ sc ; System ABEND
- lis r0,hi16(Choke) ; Choke code
- ori r0,r0,lo16(Choke) ; and the rest
- li r3,failStack ; Bad stack code
- sc ; System ABEND
+kernelStackUnaligned:
+ lis r0,hi16(Choke) ; Choke code
+ ori r0,r0,lo16(Choke) ; and the rest
+ li r3,failUnalignedStk ; Unaligned stack code
+ sc ; System ABEND
/*
.align 5
.globl EXT(shandler)
-LEXT(shandler) /* System call handler */
-
- mfsprg r25,0 /* Get the per proc area */
- lwz r0,saver0(r4) /* Get the original syscall number */
- lwz r17,PP_ISTACKPTR(r25) ; Get interrupt stack pointer
- rlwinm r15,r0,0,0,19 ; Clear the bottom of call number for fast check
- lwz r16,PP_CPU_DATA(r25) /* Assume we need this */
- mr. r17,r17 ; Are we on interrupt stack?
- lwz r7,savesrr1(r4) ; Get the SRR1 value
- beq- EXT(ihandler) ; On interrupt stack, not allowed...
+LEXT(shandler) ; System call handler
+
+ lwz r7,savesrr1+4(r4) ; Get the SRR1 value
+ mfsprg r13,1 ; Get the current activation
+ lwz r25,ACT_PER_PROC(r13) ; Get the per_proc block
+ lwz r0,saver0+4(r4) ; Get the original syscall number
+ lwz r17,PP_ISTACKPTR(r25) ; Get interrupt stack pointer
+ rlwinm r15,r0,0,0,19 ; Clear the bottom of call number for fast check
+ mr. r17,r17 ; Are we on interrupt stack?
+ lwz r9,savevrsave(r4) ; Get the VRsave register
+ beq-- EXT(ihandler) ; On interrupt stack, not allowed...
rlwinm. r6,r7,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on?
- lwz r16,CPU_ACTIVE_THREAD(r16) /* Get the thread pointer */
-
- beq+ svecoff ; Vector off, do not save vrsave...
- lwz r6,savevrsave(r4) ; Get the VRSAVE register
- stw r6,liveVRS(r25) ; Set the live value
-
-svecoff: rlwinm. r6,r7,0,MSR_FP_BIT,MSR_FP_BIT ; Was floating point on?
- lwz r13,THREAD_TOP_ACT(r16) /* Pick up the active thread */
- beq+ sfpoff ; Skip if floating point is off...
- lwz r9,savexfpscr(r4) ; Grab the just saved FPSCR
- stw r9,liveFPSCR(r25) ; Make it the live copy
+ mr r16,r13
+ beq++ svecoff ; Vector off, do not save vrsave...
+ stw r9,liveVRS(r25) ; Set the live value
+;
; Check if SCs are being redirected for the BlueBox or to VMM
+;
-sfpoff: lwz r6,ACT_MACT_SPF(r13) ; Pick up activation special flags
- mtcrf 0x41,r6 ; Check special flags
- crmove cr6_eq,runningVMbit ; Remember if we are in VMM
- bf+ bbNoMachSCbit,noassist ; Take branch if SCs are not redirected
- lwz r26,ACT_MACT_BEDA(r13) ; Pick up the pointer to the blue box exception area
- b EXT(atomic_switch_syscall) ; Go to the assist...
-
-noassist: cmplwi r15,0x7000 /* Do we have a fast path trap? */
- lwz r14,ACT_MACT_PCB(r13) /* Now point to the PCB */
- beql+ fastpath /* We think it's a fastpath... */
-
- lwz r1,ACT_MACT_KSP(r13) /* Get the kernel stack pointer */
+svecoff: lwz r6,ACT_MACT_SPF(r13) ; Pick up activation special flags
+ mtcrf 0x40,r6 ; Check special flags
+ mtcrf 0x01,r6 ; Check special flags
+ crmove cr6_eq,runningVMbit ; Remember if we are in VMM
+ bne++ cr6,sVMchecked ; Not running VM
+ lwz r18,spcFlags(r25) ; Load per_proc special flags
+ rlwinm. r18,r18,0,FamVMmodebit,FamVMmodebit ; Is FamVMmodebit set?
+ beq sVMchecked ; Not in FAM
+ cmpwi r0,0x6004 ; Is it vmm_dispatch syscall:
+ bne sVMchecked
+ lwz r26,saver3+4(r4) ; Get the original syscall number
+ cmpwi cr6,r26,kvmmExitToHost ; vmm_exit_to_host request
+sVMchecked:
+ bf++ bbNoMachSCbit,noassist ; Take branch if SCs are not redirected
+ lwz r26,ACT_MACT_BEDA(r13) ; Pick up the pointer to the blue box exception area
+ b EXT(atomic_switch_syscall) ; Go to the assist...
+
+noassist: cmplwi r15,0x7000 ; Do we have a fast path trap?
+ lwz r14,ACT_MACT_PCB(r13) ; Now point to the PCB
+ beql fastpath ; We think it is a fastpath...
+
+ lwz r1,ACT_MACT_KSP(r13) ; Get the kernel stack pointer
#if DEBUG
- mr. r1,r1 /* Are we already on the kernel stack? */
- li r3,T_SYSTEM_CALL /* Yup, pretend we had an interrupt... */
- beq- EXT(ihandler) /* Bad boy, bad boy... What'cha gonna do when they come for you? */
+ mr. r1,r1 ; Are we already on the kernel stack?
+ li r3,T_SYSTEM_CALL ; Yup, pretend we had an interrupt...
+ beq- EXT(ihandler) ; Bad boy, bad boy... What cha gonna do when they come for you?
#endif /* DEBUG */
- stw r4,ACT_MACT_PCB(r13) /* Point to our savearea */
- li r0,0 /* Clear this out */
- stw r14,SAVprev(r4) /* Queue the new save area in the front */
- stw r13,SAVact(r4) /* Point the savearea at its activation */
+ stw r4,ACT_MACT_PCB(r13) ; Point to our savearea
+ stw r4,ACT_MACT_UPCB(r13) ; Store user savearea
+ li r0,0 ; Clear this out
+ stw r14,SAVprev+4(r4) ; Queue the new save area in the front
+ stw r13,SAVact(r4) ; Point the savearea at its activation
#if VERIFYSAVE
- bl versave ; (TEST/DEBUG)
+ bl versave ; (TEST/DEBUG)
#endif
- mr r30,r4 /* Save pointer to the new context savearea */
- lwz r15,saver1(r4) /* Grab interrupt time stack */
- stw r0,ACT_MACT_KSP(r13) /* Mark stack as busy with 0 val */
- stw r15,FM_BACKPTR(r1) /* Link backwards */
+ lwz r15,saver1+4(r4) ; Grab interrupt time stack
+ mr r30,r4 ; Save pointer to the new context savearea
+ stw r0,ACT_MACT_KSP(r13) ; Mark stack as busy with 0 val
+ stw r15,FM_BACKPTR(r1) ; Link stack frame backwards
+
+ lwz r3,SAVtime+4(r30)
+ addi r4,r13,SYSTEM_TIMER
+ bl EXT(timer_event)
#if DEBUG
- /* If debugging, we need two frames, the first being a dummy
- * which links back to the trapped routine. The second is
- * that which the C routine below will need
- */
- lwz r8,savesrr0(r30) /* Get the point of interruption */
- stw r8,FM_LR_SAVE(r1) /* save old instr ptr as LR value */
- stwu r1, -FM_SIZE(r1) /* and make new frame */
+/* If debugging, we need two frames, the first being a dummy
+ * which links back to the trapped routine. The second is
+ * that which the C routine below will need
+ */
+ lwz r8,savesrr0+4(r30) ; Get the point of interruption
+ stw r8,FM_LR_SAVE(r1) ; Save old instr ptr as LR value
+ stwu r1, -FM_SIZE(r1) ; and make new frame
#endif /* DEBUG */
- mfmsr r11 /* Get the MSR */
- lwz r15,SAVflags(r4) /* Get the savearea flags */
- ori r11,r11,lo16(MASK(MSR_EE)) /* Turn on interruption enabled bit */
- lwz r0,saver0(r30) ; Get R0 back
- oris r15,r15,SAVsyscall >> 16 /* Mark that it this is a syscall */
- rlwinm r10,r0,0,0,19 ; Keep only the top part
- stwu r1,-(FM_SIZE+ARG_SIZE)(r1) /* Make a stack frame */
- cmplwi r10,0x6000 ; Is it the special ppc-only guy?
- stw r15,SAVflags(r30) /* Save syscall marker */
- beq- cr6,exitFromVM ; It is time to exit from alternate context...
-
- beq- ppcscall ; Call the ppc-only system call handler...
-
- mtmsr r11 /* Enable interruptions */
-
- /* Call a function that can print out our syscall info */
- /* Note that we don't care about any volatiles yet */
mr r4,r30
- bl EXT(syscall_trace)
-
- lwz r0,saver0(r30) /* Get the system call selector */
- mr. r0,r0 /* What kind is it? */
- blt- .L_kernel_syscall /* -ve syscall - go to kernel */
- /* +ve syscall - go to server */
- cmpwi cr0,r0,0x7FFA
- beq- .L_notify_interrupt_syscall
-
-#ifdef MACH_BSD
- mr r3,r30 /* Get PCB/savearea */
- lwz r4,saver4(r30) /* Restore r4 */
- lwz r5,saver5(r30) /* Restore r5 */
- lwz r6,saver6(r30) /* Restore r6 */
- lwz r7,saver7(r30) /* Restore r7 */
- lwz r8,saver8(r30) /* Restore r8 */
- lwz r9,saver9(r30) /* Restore r9 */
- lwz r10,saver10(r30) /* Restore r10 */
- bl EXT(unix_syscall) /* Check out unix... */
+
+ lwz r15,SAVflags(r30) ; Get the savearea flags
+ lwz r0,saver0+4(r30) ; Get R0 back
+ mfmsr r11 ; Get the MSR
+ stwu r1,-(FM_SIZE+ARG_SIZE+MUNGE_ARGS_SIZE)(r1) ; Make a stack frame
+ ori r11,r11,lo16(MASK(MSR_EE)) ; Turn on interruption enabled bit
+ rlwinm r10,r0,0,0,19 ; Keep only the top part
+ oris r15,r15,SAVsyscall >> 16 ; Mark that it this is a syscall
+ cmplwi r10,0x6000 ; Is it the special ppc-only guy?
+ stw r15,SAVflags(r30) ; Save syscall marker
+ beq-- cr6,exitFromVM ; It is time to exit from alternate context...
+
+ beq-- ppcscall ; Call the ppc-only system call handler...
+
+ mr. r0,r0 ; What kind is it?
+ mtmsr r11 ; Enable interruptions
+
+ blt-- .L_kernel_syscall ; System call number if negative, this is a mach call...
+
+ lwz r8,ACT_TASK(r13) ; Get our task
+ cmpwi cr0,r0,0x7FFA ; Special blue box call?
+ beq-- .L_notify_interrupt_syscall ; Yeah, call it...
+
+ lwz r7,TASK_SYSCALLS_UNIX(r8) ; Get the current count
+ mr r3,r30 ; Get PCB/savearea
+ mr r4,r13 ; current activation
+ addi r7,r7,1 ; Bump it
+ stw r7,TASK_SYSCALLS_UNIX(r8) ; Save it
+
+#if FPFLOOD
+ stfd f31,emfp31(r25) ; (TEST/DEBUG)
#endif
+ bl EXT(unix_syscall) ; Check out unix...
+
.L_call_server_syscall_exception:
- li r3,EXC_SYSCALL /* doexception(EXC_SYSCALL, num, 1) */
+ li r3,EXC_SYSCALL ; doexception(EXC_SYSCALL, num, 1)
.L_call_server_exception:
- mr r4,r0 /* Set syscall selector */
+ mr r4,r0 ; Set syscall selector
li r5,1
- b EXT(doexception) /* Go away, never to return... */
-
-/* The above, but with EXC_MACH_SYSCALL */
-.L_call_server_mach_syscall:
- li r3,EXC_MACH_SYSCALL
- b .L_call_server_exception /* Join the common above... */
+ b EXT(doexception) ; Go away, never to return...
.L_notify_interrupt_syscall:
- lwz r3,saver3(r30) ; Get the new PC address to pass in
+ lwz r3,saver3+4(r30) ; Get the new PC address to pass in
bl EXT(syscall_notify_interrupt)
- b .L_syscall_return
+/*
+ * Ok, return from C function, R3 = return value
+ *
+ * saved state is still in R30 and the active thread is in R16 .
+ */
+ mr r31,r16 ; Move the current thread pointer
+ stw r3,saver3+4(r30) ; Stash the return code
+ b .L_thread_syscall_ret_check_ast
;
; Handle PPC-only system call interface
; and the savearea/pcb as the first parameter.
; It is up to the callee to enable interruptions if
; they should be. We are in a state here where
-; both interrupts and preemption is ok, but because we could
+; both interrupts and preemption are ok, but because we could
; be calling diagnostic code we will not enable.
;
; Also, the callee is responsible for finding any parameters
; Note: the last option is intended for special diagnostics calls that
; want the thread to return and execute before checking for preemption.
;
+; NOTE: Both R16 (thread) and R30 (savearea) need to be preserved over this call!!!!
+;
+
+ .align 5
-ppcscall: rlwinm r11,r0,2,18,29 ; Make an index into the table
- lis r10,hi16(EXT(PPCcalls)) ; Get PPC-only system call table
- cmplwi r11,PPCcallmax ; See if we are too big
- ori r10,r10,lo16(EXT(PPCcalls)) ; Merge in low half
+ppcscall: rlwinm r11,r0,2,18,29 ; Make an index into the table
+ lis r10,hi16(EXT(PPCcalls)) ; Get PPC-only system call table
+ cmplwi r11,PPCcallmax ; See if we are too big
+ ori r10,r10,lo16(EXT(PPCcalls)) ; Merge in low half
bgt- .L_call_server_syscall_exception ; Bogus call...
- lwzx r11,r10,r11 ; Get function address
+ lwzx r11,r10,r11 ; Get function address
;
; Note: make sure we do not change the savearea in R30 to
; depend upon it being there.
;
- mr r3,r30 ; Pass the savearea
- mr r4,r13 ; Pass the activation
- mr. r11,r11 ; See if there is a function here
- mtlr r11 ; Set the function address
+ mr r3,r30 ; Pass the savearea
+ mr r4,r13 ; Pass the activation
+ mr. r11,r11 ; See if there is a function here
+ mtctr r11 ; Set the function address
beq- .L_call_server_syscall_exception ; Disabled call...
- blrl ; Call it
-
-
+#if INSTRUMENT
+ mfspr r4,pmc1 ; Get stamp
+ stw r4,0x6100+(9*16)+0x0(0) ; Save it
+ mfspr r4,pmc2 ; Get stamp
+ stw r4,0x6100+(9*16)+0x4(0) ; Save it
+ mfspr r4,pmc3 ; Get stamp
+ stw r4,0x6100+(9*16)+0x8(0) ; Save it
+ mfspr r4,pmc4 ; Get stamp
+ stw r4,0x6100+(9*16)+0xC(0) ; Save it
+#endif
+ bctrl ; Call it
+
.globl EXT(ppcscret)
+
LEXT(ppcscret)
- mr. r3,r3 ; See what we should do
- mr r31,r16 ; Restore the current thread pointer
+ mr. r3,r3 ; See what we should do
+ mr r31,r16 ; Restore the current thread pointer
bgt+ .L_thread_syscall_ret_check_ast ; Take normal AST checking return....
- mfsprg r10,0 ; Get the per_proc
- blt+ .L_thread_syscall_return ; Return, but no ASTs....
- lwz r0,saver0(r30) ; Restore the system call number
+ mfsprg r10,1 ; Get the current activation
+ lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block
+ blt+ .L_thread_syscall_return ; Return, but no ASTs....
+ lwz r0,saver0+4(r30) ; Restore the system call number
b .L_call_server_syscall_exception ; Go to common exit...
+
+/*
+ * we get here for mach system calls
+ * when kdebug tracing is enabled
+ */
+
+ksystrace:
+ mr r4,r30 ; Pass in saved state
+ bl EXT(syscall_trace)
+
+ cmplw r31,r29 ; Is this syscall in the table?
+ add r31,r27,r28 ; Point right to the syscall table entry
+
+ bge- .L_call_server_syscall_exception ; The syscall number is invalid
+
+ lwz r0,savesrr1(r30) ; Get the saved srr1
+ rlwinm. r0,r0,0,MSR_SF_BIT,MSR_SF_BIT ; Test for 64 bit caller
+ lwz r0,MACH_TRAP_ARG_MUNGE32(r31) ; Pick up the 32 bit munge function address
+ beq-- .L_ksystrace_munge
+ lwz r0,MACH_TRAP_ARG_MUNGE64(r31) ; Pick up the 64 bit munge function address
+
+.L_ksystrace_munge:
+ cmplwi r0,0 ; do we have a munger to call?
+ mtctr r0 ; Set the function call address
+ addi r3,r30,saver3 ; Pointer to args from save area
+ addi r4,r1,FM_ARG0+ARG_SIZE ; Pointer for munged args
+ beq-- .L_ksystrace_trapcall ; just make the trap call
+ bctrl ; Call the munge function
+
+.L_ksystrace_trapcall:
+ lwz r0,MACH_TRAP_FUNCTION(r31) ; Pick up the function address
+ mtctr r0 ; Set the function call address
+ addi r3,r1,FM_ARG0+ARG_SIZE ; Pointer to munged args
+ bctrl
+
+ mr r4,r30 ; Pass in the savearea
+ bl EXT(syscall_trace_end) ; Trace the exit of the system call
+ b .L_mach_return
+
+
+
/* Once here, we know that the syscall was -ve
* we should still have r1=ksp,
* r16 = pointer to current thread,
* r0 = syscall number
* r30 = pointer to saved state (in pcb)
*/
-.L_kernel_syscall:
- neg r31, r0 /* Make number +ve and put in r31*/
-
- /* If out of range, call server with syscall exception */
- addis r29, 0, HIGH_CADDR(EXT(mach_trap_count))
- addi r29, r29, LOW_ADDR(EXT(mach_trap_count))
- lwz r29, 0(r29)
- cmp cr0, r31, r29
- bge- cr0, .L_call_server_syscall_exception
-
- addis r29, 0, HIGH_CADDR(EXT(mach_trap_table))
- addi r29, r29, LOW_ADDR(EXT(mach_trap_table))
-
- /* multiply the trap number to get offset into table */
- slwi r31, r31, MACH_TRAP_OFFSET_POW2
-
- /* r31 now holds offset into table of our trap entry,
- * add on the table base, and it then holds pointer to entry
- */
- add r31, r31, r29
-
- /* If the function is kern_invalid, prepare to send an exception.
- This is messy, but parallels the x86. We need it for task_by_pid,
- at least. */
- lis r29, HIGH_CADDR(EXT(kern_invalid))
- addi r29, r29, LOW_ADDR(EXT(kern_invalid))
- lwz r0, MACH_TRAP_FUNCTION(r31)
- cmp cr0, r0, r29
- beq- .L_call_server_syscall_exception
-
- /* get arg count. If argc > 8 then not all args were in regs,
- * so we must perform copyin.
- */
- lwz r29, MACH_TRAP_ARGC(r31)
- cmpwi cr0, r29, 8
- ble+ .L_syscall_got_args
-
-/* argc > 8 - perform a copyin */
-/* if the syscall came from kernel space, we can just copy */
-
- lwz r0,savesrr1(r30) /* Pick up exception time MSR */
- andi. r0,r0,MASK(MSR_PR) /* Check the priv bit */
- bne+ .L_syscall_arg_copyin /* We're not priviliged... */
-
-/* we came from a privilaged task, just do a copy */
-/* get user's stack pointer */
-
- lwz r28,saver1(r30) /* Get the stack pointer */
-
- subi r29,r29,8 /* Get the number of arguments to copy */
-
- addi r28,r28,COPYIN_ARG0_OFFSET-4 /* Point to source - 4 */
- addi r27,r1,FM_ARG0-4 /* Point to sink - 4 */
-
-.L_syscall_copy_word_loop:
- addic. r29,r29,-1 /* Count down the number of arguments left */
- lwz r0,4(r28) /* Pick up the argument from the stack */
- addi r28,r28,4 /* Point to the next source */
- stw r0,4(r27) /* Store the argument */
- addi r27,r27,4 /* Point to the next sink */
- bne+ .L_syscall_copy_word_loop /* Move all arguments... */
- b .L_syscall_got_args /* Go call it now... */
-
-
-/* we came from a user task, pay the price of a real copyin */
-/* set recovery point */
-
-.L_syscall_arg_copyin:
- lwz r8,ACT_VMMAP(r13) ; Get the vm_map for this activation
- lis r28,hi16(.L_syscall_copyin_recover)
- lwz r8,VMMAP_PMAP(r8) ; Get the pmap
- ori r28,r28,lo16(.L_syscall_copyin_recover)
- addi r8,r8,PMAP_SEGS ; Point to the pmap SR slots
- stw r28,THREAD_RECOVER(r16) /* R16 still holds thread ptr */
-
-/* We can manipulate the COPYIN segment register quite easily
- * here, but we've also got to make sure we don't go over a
- * segment boundary - hence some mess.
- * Registers from 12-29 are free for our use.
- */
-
+ .align 5
- lwz r28,saver1(r30) /* Get the stack pointer */
- subi r29,r29,8 /* Get the number of arguments to copy */
- addi r28,r28,COPYIN_ARG0_OFFSET /* Set source in user land */
-
-/* set up SR_COPYIN to allow us to copy, we may need to loop
- * around if we change segments. We know that this previously
- * pointed to user space, so the sid doesn't need setting.
- */
-
- rlwinm r7,r28,6,26,29 ; Get index to the segment slot
-
-.L_syscall_copyin_seg_loop:
-
-
- lwzx r10,r8,r7 ; Get the source SR value
- rlwinm r26,r28,0,4,31 ; Clear the segment number from source address
- mtsr SR_COPYIN,r10 ; Set the copyin SR
- isync
-
- oris r26,r26,(SR_COPYIN_NUM << (28-16)) ; Insert the copyin segment number into source address
-
-/* Make r27 point to address-4 of where we will store copied args */
- addi r27,r1,FM_ARG0-4
-
-.L_syscall_copyin_word_loop:
-
- lwz r0,0(r26) /* MAY CAUSE PAGE FAULT! */
- subi r29,r29,1 ; Decrement count
- addi r26,r26,4 ; Bump input
- stw r0,4(r27) ; Save the copied in word
- mr. r29,r29 ; Are they all moved?
- addi r27,r27,4 ; Bump output
- beq+ .L_syscall_copyin_done ; Escape if we are done...
-
- rlwinm. r0,r26,0,4,29 ; Did we just step into a new segment?
- addi r28,r28,4 ; Bump up user state address also
- bne+ .L_syscall_copyin_word_loop ; We are still on the same segment...
-
- addi r7,r7,4 ; Bump to next slot
- b .L_syscall_copyin_seg_loop /* On new segment! remap */
-
-/* Don't bother restoring SR_COPYIN, we can leave it trashed */
-/* clear thread recovery as we're done touching user data */
-
-.L_syscall_copyin_done:
- li r0,0
- stw r0,THREAD_RECOVER(r16) /* R16 still holds thread ptr */
-
-.L_syscall_got_args:
- lwz r8,ACT_TASK(r13) /* Get our task */
- lis r10,hi16(EXT(c_syscalls_mach)) /* Get top half of counter address */
- lwz r7,TASK_SYSCALLS_MACH(r8) ; Get the current count
- lwz r3,saver3(r30) /* Restore r3 */
- addi r7,r7,1 ; Bump it
- ori r10,r10,lo16(EXT(c_syscalls_mach)) /* Get low half of counter address */
- stw r7,TASK_SYSCALLS_MACH(r8) ; Save it
- lwz r4,saver4(r30) /* Restore r4 */
- lwz r9,0(r10) /* Get counter */
- lwz r5,saver5(r30) /* Restore r5 */
- lwz r6,saver6(r30) /* Restore r6 */
- addi r9,r9,1 /* Add 1 */
- lwz r7,saver7(r30) /* Restore r7 */
- lwz r8,saver8(r30) /* Restore r8 */
- stw r9,0(r10) /* Save it back */
- lwz r9,saver9(r30) /* Restore r9 */
- lwz r10,saver10(r30) /* Restore r10 */
-
- lwz r0,MACH_TRAP_FUNCTION(r31)
-
-/* calling this function, all the callee-saved registers are
- * still valid except for r30 and r31 which are in the PCB
- * r30 holds pointer to saved state (ie. pcb)
- * r31 is scrap
- */
- mtctr r0
- bctrl /* perform the actual syscall */
+.L_kernel_syscall:
+;
+; Call a function that can print out our syscall info
+; Note that we don t care about any volatiles yet
+;
+ lwz r10,ACT_TASK(r13) ; Get our task
+ lwz r0,saver0+4(r30)
+ lis r8,hi16(EXT(kdebug_enable)) ; Get top of kdebug_enable
+ lis r28,hi16(EXT(mach_trap_table)) ; Get address of table
+ ori r8,r8,lo16(EXT(kdebug_enable)) ; Get bottom of kdebug_enable
+ lwz r8,0(r8) ; Get kdebug_enable
+
+ lwz r7,TASK_SYSCALLS_MACH(r10) ; Get the current count
+ neg r31,r0 ; Make this positive
+ mr r3,r31 ; save it
+ slwi r27,r3,4 ; multiply by 16
+ slwi r3,r3,2 ; and the original by 4
+ ori r28,r28,lo16(EXT(mach_trap_table)) ; Get address of table
+ add r27,r27,r3 ; for a total of 20x (5 words/entry)
+ addi r7,r7,1 ; Bump TASK_SYSCALLS_MACH count
+ cmplwi r8,0 ; Is kdebug_enable non-zero
+ stw r7,TASK_SYSCALLS_MACH(r10) ; Save count
+ bne-- ksystrace ; yes, tracing enabled
+
+ cmplwi r31,MACH_TRAP_TABLE_COUNT ; Is this syscall in the table?
+ add r31,r27,r28 ; Point right to the syscall table entry
+
+ bge-- .L_call_server_syscall_exception ; The syscall number is invalid
+
+ lwz r0,savesrr1(r30) ; Get the saved srr1
+ rlwinm. r0,r0,0,MSR_SF_BIT,MSR_SF_BIT ; Test for 64 bit caller
+ lwz r0,MACH_TRAP_ARG_MUNGE32(r31) ; Pick up the 32 bit munge function address
+ beq-- .L_kernel_syscall_munge
+ lwz r0,MACH_TRAP_ARG_MUNGE64(r31) ; Pick up the 64 bit munge function address
+
+.L_kernel_syscall_munge:
+ cmplwi r0,0 ; test for null munger
+ mtctr r0 ; Set the function call address
+ addi r3,r30,saver3 ; Pointer to args from save area
+ addi r4,r1,FM_ARG0+ARG_SIZE ; Pointer for munged args
+ beq-- .L_kernel_syscall_trapcall ; null munger - skip to trap call
+ bctrl ; Call the munge function
+
+.L_kernel_syscall_trapcall:
+ lwz r0,MACH_TRAP_FUNCTION(r31) ; Pick up the function address
+ mtctr r0 ; Set the function call address
+ addi r3,r1,FM_ARG0+ARG_SIZE ; Pointer to munged args
+
+#if FPFLOOD
+ stfd f31,emfp31(r25) ; (TEST/DEBUG)
+#endif
-/* 'standard' syscall returns here - INTERRUPTS ARE STILL ON */
+ bctrl
-/* r3 contains value that we're going to return to the user
- */
/*
- * Ok, return from C function, ARG0 = return value
+ * Ok, return from C function, R3 = return value
*
* get the active thread's PCB pointer and thus pointer to user state
- * saved state is still in R30 and the active thread is in R16 .
+ * saved state is still in R30 and the active thread is in R16
*/
-/* Store return value into saved state structure, since
- * we need to pick up the value from here later - the
- * syscall may perform a thread_set_syscall_return
+.L_mach_return:
+ srawi r0,r3,31 ; properly extend the return code
+ cmpi cr0,r3,KERN_INVALID_ARGUMENT ; deal with invalid system calls
+ mr r31,r16 ; Move the current thread pointer
+ stw r0, saver3(r30) ; stash the high part of the return code
+ stw r3,saver3+4(r30) ; Stash the low part of the return code
+ beq-- cr0,.L_mach_invalid_ret ; otherwise fall through into the normal return path
+.L_mach_invalid_arg:
+
+
+/* 'standard' syscall returns here - INTERRUPTS ARE STILL ON
+ * the syscall may perform a thread_set_syscall_return
* followed by a thread_exception_return, ending up
* at thread_syscall_return below, with SS_R3 having
* been set up already
- */
-
-/* When we are here, r16 should point to the current thread,
+ *
+ * When we are here, r31 should point to the current thread,
* r30 should point to the current pcb
+ * r3 contains value that we're going to return to the user
+ * which has already been stored back into the save area
*/
-
-/* save off return value, we must load it
- * back anyway for thread_exception_return
- * TODO NMGS put in register?
- */
-.L_syscall_return:
- mr r31,r16 /* Move the current thread pointer */
- stw r3,saver3(r30) /* Stash the return code */
-
- /* Call a function that records the end of */
- /* the mach system call */
- mr r4,r30
- bl EXT(syscall_trace_end)
-
-#if 0
- lis r0,HIGH_ADDR(CutTrace) /* (TEST/DEBUG) */
- mr r4,r31 /* (TEST/DEBUG) */
- oris r0,r0,LOW_ADDR(CutTrace) /* (TEST/DEBUG) */
- mr r5,r30 /* (TEST/DEBUG) */
- sc /* (TEST/DEBUG) */
-#endif
-
+
.L_thread_syscall_ret_check_ast:
- mfmsr r12 /* Get the current MSR */
- rlwinm r12,r12,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Turn off interruptions enable bit */
- mtmsr r12 /* Turn interruptions off */
+ lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable
+ mfmsr r12 ; Get the current MSR
+ ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE
+ andc r12,r12,r10 ; Turn off VEC, FP, and EE
+ mtmsr r12 ; Turn interruptions off
- mfsprg r10,0 /* Get the per_processor block */
+ mfsprg r10,1 ; Get the current activation
+ lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block
/* Check to see if there's an outstanding AST */
- lwz r4,PP_NEED_AST(r10)
- lwz r4,0(r4)
- cmpi cr0,r4, 0
- beq cr0,.L_syscall_no_ast
+ lwz r4,PP_PENDING_AST(r10)
+ cmpi cr0,r4, 0 ; Any pending asts?
+ beq++ cr0,.L_syscall_no_ast ; Nope...
/* Yes there is, call ast_taken
* pretending that the user thread took an AST exception here,
#if DEBUG
/* debug assert - make sure that we're not returning to kernel */
- lwz r3,savesrr1(r30)
+ lwz r3,savesrr1+4(r30)
andi. r3,r3,MASK(MSR_PR)
- bne+ 0f /* returning to user level, check */
+ bne++ scrnotkern ; returning to user level, check
- lis r0,hi16(Choke) ; Choke code
- ori r0,r0,lo16(Choke) ; and the rest
- li r3,failContext ; Bad state code
- sc ; System ABEND
+ lis r0,hi16(Choke) ; Choke code
+ ori r0,r0,lo16(Choke) ; and the rest
+ li r3,failContext ; Bad state code
+ sc ; System ABEND
-
-0:
+scrnotkern:
#endif /* DEBUG */
- li r3, AST_ALL
- li r4, 1
- bl EXT(ast_taken)
-
- b .L_thread_syscall_ret_check_ast
+ lis r3,hi16(AST_ALL) ; Set ast flags
+ li r4,1 ; Set interrupt allowed
+ ori r3,r3,lo16(AST_ALL)
+ bl EXT(ast_taken) ; Process the pending ast
+ b .L_thread_syscall_ret_check_ast ; Go see if there was another...
+.L_mach_invalid_ret:
+/*
+ * need to figure out why we got an KERN_INVALID_ARG
+ * if it was due to a non-existent system call
+ * then we want to throw an exception... otherwise
+ * we want to pass the error code back to the caller
+ */
+ lwz r0,saver0+4(r30) ; reload the original syscall number
+ neg r28,r0 ; Make this positive
+ mr r4,r28 ; save a copy
+ slwi r27,r4,4 ; multiply by 16
+ slwi r4,r4,2 ; and another 4
+ lis r28,hi16(EXT(mach_trap_table)) ; Get address of table
+ add r27,r27,r4 ; for a total of 20x (5 words/entry)
+ ori r28,r28,lo16(EXT(mach_trap_table)) ; Get address of table
+ add r28,r27,r28 ; Point right to the syscall table entry
+ lwz r27,MACH_TRAP_FUNCTION(r28) ; Pick up the function address
+ lis r28,hi16(EXT(kern_invalid)) ; Get high half of invalid syscall function
+ ori r28,r28,lo16(EXT(kern_invalid)) ; Get low half of invalid syscall function
+ cmpw cr0,r27,r28 ; Check if this is an invalid system call
+ beq-- .L_call_server_syscall_exception ; We have a bad system call
+ b .L_mach_invalid_arg ; a system call returned KERN_INVALID_ARG
+
+
/* thread_exception_return returns to here, almost all
* registers intact. It expects a full context restore
* of what it hasn't restored itself (ie. what we use).
* r30 points to the current pcb
*/
+ .align 5
+
.L_syscall_no_ast:
.L_thread_syscall_return:
- mr r3,r30 ; Get savearea to the correct register for common exit
- lwz r8,THREAD_TOP_ACT(r31) /* Now find the current activation */
-
- lwz r11,SAVflags(r30) /* Get the flags */
- lwz r5,THREAD_KERNEL_STACK(r31) /* Get the base pointer to the stack */
- rlwinm r11,r11,0,15,13 /* Clear the syscall flag */
- lwz r4,SAVprev(r30) ; Get the previous save area
- stw r11,SAVflags(r30) /* Stick back the flags */
- addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE /* Reset to empty */
- stw r4,ACT_MACT_PCB(r8) ; Save previous save area
- stw r5,ACT_MACT_KSP(r8) /* Save the empty stack pointer */
-
- b chkfac ; Go end it all...
-
-
-.L_syscall_copyin_recover:
+ mr r3,r30 ; Get savearea to the correct register for common exit
- /* This is the catcher for any data faults in the copyin
- * of arguments from the user's stack.
- * r30 still holds a pointer to the PCB
- *
- * call syscall_error(EXC_BAD_ACCESS, EXC_PPC_VM_PROT_READ, sp, ssp),
- *
- * we already had a frame so we can do this
- */
-
- li r3,EXC_BAD_ACCESS
- li r4,EXC_PPC_VM_PROT_READ
- lwz r5,saver1(r30)
- mr r6,r30
-
- bl EXT(syscall_error)
- b .L_syscall_return
+ lwz r11,SAVflags(r30) ; Get the flags
+ lwz r5,THREAD_KERNEL_STACK(r31) ; Get the base pointer to the stack
+ lwz r4,SAVprev+4(r30) ; Get the previous save area
+ rlwinm r11,r11,0,15,13 ; Clear the syscall flag
+ mfsprg r8,1 ; Now find the current activation
+ addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE ; Reset to empty
+ stw r11,SAVflags(r30) ; Stick back the flags
+ stw r5,ACT_MACT_KSP(r8) ; Save the empty stack pointer
+ stw r4,ACT_MACT_PCB(r8) ; Save previous save area
+ b chkfac ; Go end it all...
-
/*
* thread_exception_return()
*
LEXT(thread_exception_return) ; Directly return to user mode
.L_thread_exc_ret_check_ast:
-
- mfmsr r3 /* Get the MSR */
- rlwinm r3,r3,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Clear EE */
- mtmsr r3 /* Disable interrupts */
+ lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable
+ mfmsr r3 ; Get the MSR
+ ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE
+ andc r3,r3,r10 ; Turn off VEC, FP, and EE
+ mtmsr r3 ; Disable interrupts
/* Check to see if there's an outstanding AST */
/* We don't bother establishing a call frame even though CHECK_AST
frame, given that we're not going to return.
*/
- mfsprg r10,0 /* Get the per_processor block */
- lwz r4,PP_NEED_AST(r10)
- lwz r4,0(r4)
+ mfsprg r10,1 ; Get the current activation
+ lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block
+ lwz r4,PP_PENDING_AST(r10)
cmpi cr0,r4, 0
- beq cr0,.L_exc_ret_no_ast
+ beq+ cr0,.L_exc_ret_no_ast
- /* Yes there is, call ast_taken
- * pretending that the user thread took an AST exception here,
- * ast_taken will save all state and bring us back here
- */
+/* Yes there is, call ast_taken
+ * pretending that the user thread took an AST exception here,
+ * ast_taken will save all state and bring us back here
+ */
-
- li r3,AST_ALL
+ lis r3,hi16(AST_ALL)
li r4,1
+ ori r3,r3,lo16(AST_ALL)
bl EXT(ast_taken)
- b .L_thread_exc_ret_check_ast /* check for a second AST (rare)*/
+ b .L_thread_exc_ret_check_ast ; check for a second AST (rare)
/* arriving here, interrupts should be disabled */
/* Get the active thread's PCB pointer to restore regs
*/
.L_exc_ret_no_ast:
- lwz r31,PP_CPU_DATA(r10)
- lwz r31,CPU_ACTIVE_THREAD(r31)
- lwz r30,THREAD_TOP_ACT(r31)
+ mfsprg r30,1 ; Get the currrent activation
+ mr r31,r30
+
lwz r30,ACT_MACT_PCB(r30)
- mr. r30,r30 ; Is there any context yet?
- beq- makeDummyCtx ; No, hack one up...
+ mr. r30,r30 ; Is there any context yet?
+ beq- makeDummyCtx ; No, hack one up...
#if DEBUG
/*
* debug assert - make sure that we're not returning to kernel
* get the active thread's PCB pointer and thus pointer to user state
*/
- lwz r3,savesrr1(r30)
+ lwz r3,savesrr1+4(r30)
andi. r3,r3,MASK(MSR_PR)
- bne+ ret_user2 ; We are ok...
+ bne+ ret_user2 ; We are ok...
- lis r0,hi16(Choke) ; Choke code
- ori r0,r0,lo16(Choke) ; and the rest
- li r3,failContext ; Bad state code
- sc ; System ABEND
+ lis r0,hi16(Choke) ; Choke code
+ ori r0,r0,lo16(Choke) ; and the rest
+ li r3,failContext ; Bad state code
+ sc ; System ABEND
ret_user2:
#endif /* DEBUG */
-/* If the MSR_SYSCALL_MASK isn't set, then we came from a trap,
+/* If the system call flag isn't set, then we came from a trap,
* so warp into the return_from_trap (thread_return) routine,
* which takes PCB pointer in R3, not in r30!
*/
- lwz r0,SAVflags(r30)
- mr r3,r30 /* Copy pcb pointer into r3 in case */
- andis. r0,r0,SAVsyscall>>16 /* Are we returning from a syscall? */
- beq- cr0,thread_return /* Nope, must be a thread return... */
- b .L_thread_syscall_return
+ lwz r0,SAVflags(r30) ; Grab the savearea flags
+ andis. r0,r0,SAVsyscall>>16 ; Are we returning from a syscall?
+ mr r3,r30 ; Copy pcb pointer into r3 in case we need it
+ beq-- cr0,thread_return ; Nope, must be a thread return...
+ b .L_thread_syscall_return ; Join up with the system call return...
;
; This is where we handle someone trying who did a thread_create followed
; by a thread_resume with no intervening thread_set_state. Just make an
; empty context, initialize it to trash and let em execute at 0...
+;
+
+ .align 5
makeDummyCtx:
- bl EXT(save_get) ; Get a save_area
- li r0,0 ; Get a 0
- addi r2,r3,savefp0 ; Point past what we are clearing
- mr r4,r3 ; Save the start
-
-cleardummy: stw r0,0(r4) ; Clear stuff
- addi r4,r4,4 ; Next word
- cmplw r4,r2 ; Still some more?
- blt+ cleardummy ; Yeah...
+ bl EXT(save_get) ; Get a save_area
+ li r4,SAVgeneral ; Get the general context type
+ li r0,0 ; Get a 0
+ stb r4,SAVflags+2(r3) ; Set type
+ addi r2,r3,savefpscr+4 ; Point past what we are clearing
+ mr r4,r3 ; Save the start
+
+cleardummy: stw r0,0(r4) ; Clear stuff
+ addi r4,r4,4 ; Next word
+ cmplw r4,r2 ; Still some more?
+ blt+ cleardummy ; Yeah...
lis r2,hi16(MSR_EXPORT_MASK_SET) ; Set the high part of the user MSR
ori r2,r2,lo16(MSR_EXPORT_MASK_SET) ; And the low part
- stw r2,savesrr1(r3) ; Set the default user MSR
+ stw r2,savesrr1+4(r3) ; Set the default user MSR
- b thread_return ; Go let em try to execute, hah!
+ b thread_return ; Go let em try to execute, hah!
/*
* ihandler(type)
.align 5
.globl EXT(ihandler)
-LEXT(ihandler) /* Interrupt handler */
+LEXT(ihandler) ; Interrupt handler */
/*
* get the value of istackptr, if it's zero then we're already on the
- * interrupt stack, otherwise it points to a saved_state structure
- * at the top of the interrupt stack.
+ * interrupt stack.
*/
- lwz r10,savesrr1(r4) /* Get SRR1 */
- mfsprg r25,0 /* Get the per_proc block */
- li r14,0 /* Zero this for now */
- rlwinm. r13,r10,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on?
- lwz r16,PP_CPU_DATA(r25) /* Assume we need this */
- crmove cr1_eq,cr0_eq ; Remember vector enablement
- lwz r1,PP_ISTACKPTR(r25) /* Get the interrupt stack */
- rlwinm. r10,r10,0,MSR_FP_BIT,MSR_FP_BIT ; Was floating point on?
- li r13,0 /* Zero this for now */
- lwz r16,CPU_ACTIVE_THREAD(r16) /* Get the thread pointer */
-
- beq+ cr1,ivecoff ; Vector off, do not save vrsave...
- lwz r7,savevrsave(r4) ; Get the VRSAVE register
- stw r7,liveVRS(r25) ; Set the live value
-
-ivecoff: li r0,0 /* Get a constant 0 */
- cmplwi cr1,r16,0 /* Are we still booting? */
- beq+ ifpoff ; Skip if floating point is off...
- lwz r9,savexfpscr(r4) ; Grab the just saved FPSCR
- stw r9,liveFPSCR(r25) ; Make it the live copy
-
-ifpoff: mr. r1,r1 /* Is it active? */
- beq- cr1,ihboot1 /* We're still coming up... */
- lwz r13,THREAD_TOP_ACT(r16) /* Pick up the active thread */
- lwz r14,ACT_MACT_PCB(r13) /* Now point to the PCB */
-
-ihboot1: lwz r9,saver1(r4) /* Pick up the 'rupt time stack */
- stw r14,SAVprev(r4) /* Queue the new save area in the front */
- stw r13,SAVact(r4) /* Point the savearea at its activation */
- beq- cr1,ihboot4 /* We're still coming up... */
- stw r4,ACT_MACT_PCB(r13) /* Point to our savearea */
-
-ihboot4: bne .L_istackfree /* Nope... */
+ lwz r10,savesrr1+4(r4) ; Get SRR1
+ lwz r7,savevrsave(r4) ; Get the VRSAVE register
+ mfsprg r13,1 ; Get the current activation
+ lwz r25,ACT_PER_PROC(r13) ; Get the per_proc block
+ li r14,0 ; Zero this for now
+ rlwinm. r16,r10,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on?
+ lwz r1,PP_ISTACKPTR(r25) ; Get the interrupt stack
+ li r16,0 ; Zero this for now
+
+ beq+ ivecoff ; Vector off, do not save vrsave...
+ stw r7,liveVRS(r25) ; Set the live value
+
+ivecoff: li r0,0 ; Get a constant 0
+ rlwinm r5,r10,0,MSR_PR_BIT,MSR_PR_BIT ; Are we trapping from supervisor state?
+ mr. r1,r1 ; Is it active?
+ cmplwi cr2,r5,0 ; cr2_eq == 1 if yes
+ mr r16,r13
+ lwz r14,ACT_MACT_PCB(r13) ; Now point to the PCB
+ lwz r9,saver1+4(r4) ; Pick up the rupt time stack
+ stw r14,SAVprev+4(r4) ; Queue the new save area in the front
+ stw r13,SAVact(r4) ; Point the savearea at its activation
+ stw r4,ACT_MACT_PCB(r13) ; Point to our savearea
+ beq cr2,ifromk
+ stw r4,ACT_MACT_UPCB(r13) ; Store user savearea
+
+ifromk: bne .L_istackfree ; Nope...
/* We're already on the interrupt stack, get back the old
* stack pointer and make room for a frame
*/
- lwz r10,PP_INTSTACK_TOP_SS(r25) ; Get the top of the interrupt stack
- addi r5,r9,INTSTACK_SIZE-FM_SIZE ; Shift stack for bounds check
- subi r1,r9,FM_REDZONE ; Back up beyond the red zone
- sub r5,r5,r10 ; Get displacement into stack
- cmplwi r5,INTSTACK_SIZE-FM_SIZE ; Is the stack actually invalid?
- blt+ ihsetback ; The stack is ok...
+ lwz r10,PP_INTSTACK_TOP_SS(r25) ; Get the top of the interrupt stack
+ addi r5,r9,INTSTACK_SIZE-FM_SIZE ; Shift stack for bounds check
+ subi r1,r9,FM_REDZONE ; Back up beyond the red zone
+ sub r5,r5,r10 ; Get displacement into stack
+ cmplwi r5,INTSTACK_SIZE-FM_SIZE ; Is the stack actually invalid?
+ blt+ ihsetback ; The stack is ok...
- lwz r5,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top
+ lwz r5,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top
subi r5,r5,KERNEL_STACK_SIZE-FM_SIZE ; Adjust to start of stack
- sub r5,r1,r5 ; Get displacement into debug stack
+ sub r5,r1,r5 ; Get displacement into debug stack
cmplwi cr2,r5,KERNEL_STACK_SIZE-FM_SIZE ; Check if we are on debug stack
- blt+ ihsetback ; Yeah, that is ok too...
+ blt+ cr2,ihsetback ; Yeah, that is ok too...
+
+ lis r0,hi16(Choke) ; Choke code
+ ori r0,r0,lo16(Choke) ; and the rest
+ li r3,failStack ; Bad stack code
+ sc ; System ABEND
- lis r0,hi16(Choke) ; Choke code
- ori r0,r0,lo16(Choke) ; and the rest
- li r3,failStack ; Bad stack code
- sc ; System ABEND
+intUnalignedStk:
+ lis r0,hi16(Choke) ; Choke code
+ ori r0,r0,lo16(Choke) ; and the rest
+ li r3,failUnalignedStk ; Unaligned stack code
+ sc ; System ABEND
.align 5
.L_istackfree:
- lwz r10,SAVflags(r4)
- stw r0,PP_ISTACKPTR(r25) /* Mark the stack in use */
- oris r10,r10,HIGH_ADDR(SAVrststk) /* Indicate we reset stack when we return from this one */
- stw r10,SAVflags(r4) /* Stick it back */
+ rlwinm. r0,r1,0,28,31 ; Check if stack is aligned (and get 0)
+ lwz r10,SAVflags(r4) ; Get savearea flags
+ bne-- intUnalignedStk ; Stack is unaligned...
+ stw r0,PP_ISTACKPTR(r25) ; Mark the stack in use
+ oris r10,r10,hi16(SAVrststk) ; Indicate we reset stack when we return from this one
+ stw r10,SAVflags(r4) ; Stick it back
- /*
- * To summarize, when we reach here, the state has been saved and
- * the stack is marked as busy. We now generate a small
- * stack frame with backpointers to follow the calling
- * conventions. We set up the backpointers to the trapped
- * routine allowing us to backtrace.
- */
+/*
+ * To summarize, when we reach here, the state has been saved and
+ * the stack is marked as busy. We now generate a small
+ * stack frame with backpointers to follow the calling
+ * conventions. We set up the backpointers to the trapped
+ * routine allowing us to backtrace.
+ */
-ihsetback: subi r1,r1,FM_SIZE /* Make a new frame */
- stw r9,FM_BACKPTR(r1) /* point back to previous stackptr */
+ihsetback: subi r1,r1,FM_SIZE ; Make a new frame
+ stw r9,FM_BACKPTR(r1) ; Point back to previous stackptr
#if VERIFYSAVE
- bl versave ; (TEST/DEBUG)
+ beq- cr1,ihbootnover ; (TEST/DEBUG)
+ bl versave ; (TEST/DEBUG)
+ihbootnover: ; (TEST/DEBUG)
#endif
#if DEBUG
* which links back to the trapped routine. The second is
* that which the C routine below will need
*/
- lwz r5,savesrr0(r4) /* Get interrupt address */
- stw r5,FM_LR_SAVE(r1) /* save old instr ptr as LR value */
- stwu r1,-FM_SIZE(r1) /* Make another new frame for C routine */
+ lwz r5,savesrr0+4(r4) ; Get interrupt address
+ stw r5,FM_LR_SAVE(r1) ; save old instr ptr as LR value
+ stwu r1,-FM_SIZE(r1) ; Make another new frame for C routine
#endif /* DEBUG */
- lwz r5,savedsisr(r4) /* Get the DSISR */
- lwz r6,savedar(r4) /* Get the DAR */
+ mr r31,r3
+ mr r30,r4
+
+ lwz r3,SAVtime+4(r4)
+ addi r4,r13,SYSTEM_TIMER
+ bl EXT(timer_event)
+
+ mr r3,r31
+ mr r4,r30
+ lwz r5,savedsisr(r30) ; Get the DSISR
+ lwz r6,savedar+4(r30) ; Get the DAR
+#if FPFLOOD
+ stfd f31,emfp31(r25) ; (TEST/DEBUG)
+#endif
+
bl EXT(interrupt)
*/
.globl EXT(ihandler_ret)
-LEXT(ihandler_ret) /* Marks our return point from debugger entry */
+LEXT(ihandler_ret) ; Marks our return point from debugger entry
- mfmsr r0 /* Get our MSR */
- rlwinm r0,r0,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Flip off the interrupt enabled bit */
- mtmsr r0 /* Make sure interrupts are disabled */
- mfsprg r10,0 /* Get the per_proc block */
+ lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable
+ mfmsr r0 ; Get our MSR
+ ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE
+ andc r0,r0,r10 ; Turn off VEC, FP, and EE
+ mtmsr r0 ; Make sure interrupts are disabled
+ mfsprg r8,1 ; Get the current activation
+ lwz r10,ACT_PER_PROC(r8) ; Get the per_proc block
- lwz r8,PP_CPU_DATA(r10) /* Get the CPU data area */
- lwz r7,SAVflags(r3) /* Pick up the flags */
- lwz r8,CPU_ACTIVE_THREAD(r8) /* and the active thread */
- lwz r9,SAVprev(r3) /* Get previous save area */
- cmplwi cr1,r8,0 /* Are we still initializing? */
- lwz r12,savesrr1(r3) /* Get the MSR we will load on return */
- beq- cr1,ihboot2 /* Skip if we are still in init... */
- lwz r8,THREAD_TOP_ACT(r8) /* Pick up the active thread */
-
-ihboot2: andis. r11,r7,HIGH_ADDR(SAVrststk) /* Is this the first on the stack? */
- beq- cr1,ihboot3 /* Skip if we are still in init... */
- stw r9,ACT_MACT_PCB(r8) /* Point to previous context savearea */
-
-ihboot3: mr r4,r3 /* Move the savearea pointer */
- beq .L_no_int_ast2 /* Get going if not the top o' stack... */
+ lwz r7,SAVflags(r3) ; Pick up the flags
+ lwz r9,SAVprev+4(r3) ; Get previous save area
+ cmplwi cr1,r8,0 ; Are we still initializing?
+ lwz r12,savesrr1+4(r3) ; Get the MSR we will load on return
+ andis. r11,r7,hi16(SAVrststk) ; Is this the first on the stack?
+ stw r9,ACT_MACT_PCB(r8) ; Point to previous context savearea
+ mr r4,r3 ; Move the savearea pointer
+ beq .L_no_int_ast2 ; Get going if not the top-o-stack...
/* We're the last frame on the stack. Restore istackptr to empty state.
* returning to user mode
* returning to a kloaded server
*/
- lwz r9,PP_INTSTACK_TOP_SS(r10) /* Get the empty stack value */
- lwz r5,PP_CPU_DATA(r10) /* Get cpu_data ptr */
- andc r7,r7,r11 /* Remove the stack reset bit in case we pass this one */
- stw r9,PP_ISTACKPTR(r10) /* Save that saved state ptr */
- lwz r3,CPU_PREEMPTION_LEVEL(r5) /* Get preemption level */
- stw r7,SAVflags(r4) /* Save the flags */
- cmplwi r3, 0 /* Check for preemption */
- bne .L_no_int_ast /* Don't preempt if level is not zero */
- andi. r6,r12,MASK(MSR_PR) /* privilege mode */
- lwz r11,PP_NEED_AST(r10) /* Get the AST request address */
- lwz r11,0(r11) /* Get the request */
- beq- .L_kernel_int_ast /* In kernel space, AST_URGENT check */
- li r3,T_AST /* Assume the worst */
- mr. r11,r11 /* Are there any pending? */
- beq .L_no_int_ast /* Nope... */
+ lwz r9,PP_INTSTACK_TOP_SS(r10) ; Get the empty stack value
+ andc r7,r7,r11 ; Remove the stack reset bit in case we pass this one
+ stw r9,PP_ISTACKPTR(r10) ; Save that saved state ptr
+ lwz r3,ACT_PREEMPT_CNT(r8) ; Get preemption level
+ stw r7,SAVflags(r4) ; Save the flags
+ cmplwi r3, 0 ; Check for preemption
+ bne .L_no_int_ast ; Do not preempt if level is not zero
+ andi. r6,r12,MASK(MSR_PR) ; privilege mode
+ lwz r11,PP_PENDING_AST(r10) ; Get the pending AST mask
+ beq- .L_kernel_int_ast ; In kernel space, AST_URGENT check
+ li r3,T_AST ; Assume the worst
+ mr. r11,r11 ; Are there any pending?
+ beq .L_no_int_ast ; Nope...
b .L_call_thandler
.L_kernel_int_ast:
- andi. r11,r11,AST_URGENT /* AST_URGENT */
- li r3,T_PREEMPT /* Assume the worst */
- beq .L_no_int_ast /* Nope... */
-
-.L_call_thandler:
+ andi. r11,r11,AST_URGENT ; Do we have AST_URGENT?
+ li r3,T_PREEMPT ; Assume the worst
+ beq .L_no_int_ast ; Nope...
/*
* There is a pending AST. Massage things to make it look like
* trap instead of performing an rfi.
*/
- stw r3,saveexception(r4) /* Set the exception code to T_AST/T_PREEMPT */
- b EXT(thandler) /* hyperspace into AST trap */
+.L_call_thandler:
+ stw r3,saveexception(r4) ; Set the exception code to T_AST/T_PREEMPT
+ b EXT(thandler) ; We need to preempt so treat like a trap...
.L_no_int_ast:
- mr r3,r4 ; Get into the right register for common code
+ mr r3,r4 ; Get into the right register for common code
+
.L_no_int_ast2:
- rlwinm r7,r7,0,15,13 /* Clear the syscall bit */
- li r4,0 ; Assume for a moment that we are in init
- stw r7,SAVflags(r3) /* Set the flags */
- beq- cr1,chkfac ; Jump away if we are in init...
- lwz r4,ACT_MACT_PCB(r8) ; Get the new level marker
+ rlwinm r7,r7,0,15,13 ; Clear the syscall flag
+ li r4,0 ; Assume for a moment that we are in init
+ stw r7,SAVflags(r3) ; Set the flags with cleared syscall flag
+ beq-- cr1,chkfac ; Jump away if we are in init...
+
+ lwz r4,ACT_MACT_PCB(r8) ; Get the new level marker
;
; R8 = pointer to activation
; R10 = per_proc block
;
-chkfac: mr. r8,r8 ; Are we still in boot?
- beq- chkenax ; Yeah, skip it all...
-
- lwz r20,ACT_MACT_FPUlvl(r8) ; Get the FPU level
- lwz r12,savesrr1(r3) ; Get the current MSR
- cmplw cr1,r20,r3 ; Are we returning from the active level?
- lwz r23,PP_FPU_THREAD(r10) ; Get floating point owner
- rlwinm r12,r12,0,MSR_FP_BIT+1,MSR_FP_BIT-1 ; Turn off floating point for now
- cmplw cr2,r23,r8 ; Are we the facility owner?
- lhz r26,PP_CPU_NUMBER(r10) ; Get the current CPU number
- cror cr0_eq,cr1_eq,cr2_eq ; Check if returning from active or we own facility
- bne- cr0,chkvecnr ; Nothing to do if not returning from active or not us...
-
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3301 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
+; Note that barring unforseen crashes, there is no escape from this point
+; on. We WILL call exception_exit and launch this context. No worries
+; about preemption or interruptions here.
+;
+; Note that we will set up R26 with whatever context we will be launching,
+; so it will indicate the current, or the deferred it it is set and we
+; are going to user state. CR2_eq will be set to indicate deferred.
+;
- li r22,ACT_MACT_FPUcpu ; Point to the CPU indication/lock word
-
-cfSpin2: lwarx r27,r22,r8 ; Get and reserve the last used CPU
- mr. r27,r27 ; Is it changing now?
- oris r0,r27,hi16(fvChk) ; Set the "changing" flag
- blt- cfSpin2 ; Spin if changing
- stwcx. r0,r22,r8 ; Lock it up
- bne- cfSpin2 ; Someone is messing right now
-
- isync ; Make sure we see everything
-
- cmplw r4,r20 ; Are we going to be in the right level?
- beq- cr1,chkfpfree ; Leaving active level, can not possibly enable...
- cmplw cr1,r27,r26 ; Are we on the right CPU?
- li r0,0 ; Get a constant 0
- beq+ cr1,chkfpnlvl ; Right CPU...
-
- stw r0,PP_FPU_THREAD(r10) ; Show facility unowned so we do not get back here
- b chkvec ; Go check out the vector facility...
-
-chkfpnlvl: bne- chkvec ; Different level, can not enable...
- lwz r24,ACT_MACT_FPU(r8) ; Get the floating point save area
- ori r12,r12,lo16(MASK(MSR_FP)) ; Enable facility
- mr. r24,r24 ; Does the savearea exist?
- li r0,1 ; Get set to invalidate
- beq- chkvec ; Nothing to invalidate...
- lwz r25,SAVlvlfp(r24) ; Get the level of top savearea
- cmplw r4,r25 ; Is the top one ours?
- bne+ chkvec ; Not ours...
- stw r0,SAVlvlfp(r24) ; Invalidate the first one
- b chkvec ; Go check out the vector facility...
-
-chkfpfree: li r0,0 ; Clear a register
- lwz r24,ACT_MACT_FPU(r8) ; Get the floating point save area
-
- bne- cr2,chkfpnfr ; Not our facility, do not clear...
- stw r0,PP_FPU_THREAD(r10) ; Clear floating point owner
-chkfpnfr:
+chkfac: lwz r29,savesrr1+4(r3) ; Get the current MSR
+ mr. r28,r8 ; Are we still in boot?
+ mr r31,r10 ; Move per_proc address
+ mr r30,r4 ; Preserve new level
+ mr r27,r3 ; Save the old level
+ beq-- chkenax ; Yeah, skip it all...
+
+ rlwinm. r0,r29,0,MSR_PR_BIT,MSR_PR_BIT ; Are we going into user state?
+
+ lwz r20,curctx(r28) ; Get our current context
+ lwz r26,deferctx(r28) ; Get any deferred context switch
+ li r0,1 ; Get set to hold off quickfret
+ rlwinm r29,r29,0,MSR_FP_BIT+1,MSR_FP_BIT-1 ; Turn off floating point for now
+ lwz r21,FPUlevel(r20) ; Get the facility level
+ cmplwi cr2,r26,0 ; Are we going into a deferred context later?
+ rlwinm r29,r29,0,MSR_VEC_BIT+1,MSR_VEC_BIT-1 ; Turn off vector for now
+ crnor cr2_eq,cr0_eq,cr2_eq ; Set cr2_eq if going to user state and there is deferred
+ lhz r19,PP_CPU_NUMBER(r31) ; Get our CPU number
+ cmplw r27,r21 ; Are we returning from the active level?
+ stw r0,holdQFret(r31) ; Make sure we hold off releasing quickfret
+ bne++ fpuchkena ; Nope...
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3302 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
+;
+; First clean up any live context we are returning from
+;
- mr. r24,r24 ; Do we even have a savearea?
- beq+ chkvec ; Nope...
+ lwz r22,FPUcpu(r20) ; Get CPU this context was last dispatched on
-#if FPVECDBG
- rlwinm. r0,r24,0,0,15 ; (TEST/DEBUG)
- bne+ notbadxxx1 ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
-notbadxxx1: ; (TEST/DEBUG)
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3303 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
-
- lwz r25,SAVlvlfp(r24) ; Get the level of top savearea
- cmplwi r25,1 ; Is the top area invalid?
- cmplw cr1,r25,r3 ; Is it for the returned from context?
- beq fptoss ; It is invalid...
- bne cr1,chkvec ; Not for the returned context...
+ stw r19,FPUcpu(r20) ; Claim context for us
-fptoss: lwz r25,SAVprefp(r24) ; Get previous savearea
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3304 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- mr r5,r25 ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
- mr. r25,r25 ; Is there one?
- stw r25,ACT_MACT_FPU(r8) ; Set the new pointer
- beq fptoplvl ; Nope, we are at the top...
-#if FPVECDBG
- rlwinm. r0,r25,0,0,15 ; (TEST/DEBUG)
- bne+ notbadxxx2 ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
-notbadxxx2: ; (TEST/DEBUG)
-#endif
- lwz r25,SAVlvlfp(r25) ; Get the new level
-
-fptoplvl: lwz r19,SAVflags(r24) ; Get the savearea flags
-#if FPVECDBG
- rlwinm. r0,r19,0,1,1 ; (TEST/DEBUG)
- bne+ donotdie3 ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
-donotdie3: ; (TEST/DEBUG)
+ eieio ; Make sure this gets out before owner clear
+
+#if ppeSize != 16
+#error per_proc_entry is not 16bytes in size
#endif
+
+ lis r23,hi16(EXT(PerProcTable)) ; Set base PerProcTable
+ slwi r22,r22,4 ; Find offset to the owner per_proc_entry
+ ori r23,r23,lo16(EXT(PerProcTable)) ; Set base PerProcTable
+ li r24,FPUowner ; Displacement to float owner
+ add r22,r23,r22 ; Point to the owner per_proc_entry
+ lwz r22,ppe_vaddr(r22) ; Point to the owner per_proc
+
+fpuinvothr: lwarx r23,r24,r22 ; Get the owner
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3305 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
- rlwinm r22,r24,0,0,19 ; Round down to the base savearea block
- rlwinm r19,r19,0,2,0 ; Remove the floating point in use flag
- stw r25,ACT_MACT_FPUlvl(r8) ; Set the new top level
- andis. r0,r19,hi16(SAVinuse) ; Still in use?
- stw r19,SAVflags(r24) ; Set the savearea flags
- bne- invlivefp ; Go invalidate live FP
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3306 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
-#if FPVECDBG
- rlwinm. r0,r24,0,0,15 ; (TEST/DEBUG)
- bne+ notbadxxx3 ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
-notbadxxx3: ; (TEST/DEBUG)
-#endif
- lwz r23,SACvrswap(r22) ; Get the conversion from virtual to real
- lwz r20,PP_QUICKFRET(r10) ; Get the old quick fret head
- xor r23,r24,r23 ; Convert to physical
- stw r20,SAVqfret(r24) ; Back chain the quick release queue
- stw r23,PP_QUICKFRET(r10) ; Anchor it
-
-invlivefp: lis r20,hi16(EXT(real_ncpus)) ; Get number of CPUs
- lis r23,hi16(EXT(per_proc_info)) ; Set base per_proc
- ori r20,r20,lo16(EXT(real_ncpus)) ; Other half of number of CPUs
- li r25,PP_FPU_THREAD ; Point to the FP owner address
- lwz r20,0(r20) ; Get number of processors active
- ori r23,r23,lo16(EXT(per_proc_info)) ; Set base per_proc
- li r2,0 ; Get something clear
-
-invlivefl: cmplw r23,r10 ; We can skip our processor
- addi r20,r20,-1 ; Count remaining processors
- beq invlivefn ; Skip ourselves...
-
-invlivefa: lwarx r0,r25,r23 ; Get FP owner for this processor
- cmplw r0,r8 ; Do we own it?
- bne invlivefn ; Nope...
- stwcx. r2,r25,r23 ; Show not live
- bne- invlivefa ; Someone else did this, try again...
-
-invlivefn: mr. r20,r20 ; Have we finished?
- addi r23,r23,ppSize ; Bump to next
- bgt invlivefl ; Make sure we do all processors...
+ sub r0,r23,r20 ; Subtract one from the other
+ sub r21,r20,r23 ; Subtract the other from the one
+ or r21,r21,r0 ; Combine them
+ srawi r21,r21,31 ; Get a 0 if equal or -1 of not
+ and r23,r23,r21 ; Make 0 if same, unchanged if not
+ stwcx. r23,r24,r22 ; Try to invalidate it
+ bne-- fpuinvothr ; Try again if there was a collision...
+ isync
;
-; Check out vector stuff (and translate savearea to physical for exit)
+; Now if there is a savearea associated with the popped context, release it.
+; Either way, pop the level to the top stacked context.
;
-chkvec: sync ; Make sure all is saved
- stw r27,ACT_MACT_FPUcpu(r8) ; Set the active CPU and release
+
+ lwz r22,FPUsave(r20) ; Get pointer to the first savearea
+ li r21,0 ; Assume we popped all the way out
+ mr. r22,r22 ; Is there anything there?
+ beq++ fpusetlvl ; No, see if we need to enable...
-chkvecnr: lwz r20,ACT_MACT_VMXlvl(r8) ; Get the vector level
- lwz r23,PP_VMX_THREAD(r10) ; Get vector owner
- cmplw cr1,r20,r3 ; Are we returning from the active level?
- cmplw cr2,r23,r8 ; Are we the facility owner?
- rlwinm r12,r12,0,MSR_VEC_BIT+1,MSR_VEC_BIT-1 ; Turn off vector for now
- cror cr0_eq,cr1_eq,cr2_eq ; Check if returning from active or we own facility
- bne- cr0,setenanr ; Not our facility, nothing to do here...
+ lwz r21,SAVlevel(r22) ; Get the level of that savearea
+ cmplw r21,r27 ; Is this the saved copy of the live stuff?
+ bne fpusetlvl ; No, leave as is...
+
+ lwz r24,SAVprev+4(r22) ; Pick up the previous area
+ li r21,0 ; Assume we popped all the way out
+ mr. r24,r24 ; Any more context stacked?
+ beq-- fpuonlyone ; Nope...
+ lwz r21,SAVlevel(r24) ; Get the level associated with save
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3401 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
+fpuonlyone: stw r24,FPUsave(r20) ; Dequeue this savearea
- li r22,ACT_MACT_VMXcpu ; Point to the CPU indication/lock word
-
-cvSpin2: lwarx r27,r22,r8 ; Get and reserve the last used CPU
- mr. r27,r27 ; Is it changing now?
- oris r0,r27,hi16(fvChk) ; Set the "changing" flag
- blt- cvSpin2 ; Spin if changing
- stwcx. r0,r22,r8 ; Lock it up
- bne- cvSpin2 ; Someone is messing right now
-
- isync ; Make sure we see everything
-
- cmplw r4,r20 ; Are we going to be in the right level?
- beq- cr1,chkvecfree ; Leaving active level, can not possibly enable...
- cmplw cr1,r27,r26 ; Are we on the right CPU?
- li r0,0 ; Get a constant 0
- beq+ cr1,chkvecnlvl ; Right CPU...
-
- stw r0,PP_VMX_THREAD(r10) ; Show facility unowned so we do not get back here
- b setena ; Go actually exit...
-
-chkvecnlvl: bne- setena ; Different level, can not enable...
- lwz r24,ACT_MACT_VMX(r8) ; Get the vector save area
- oris r12,r12,hi16(MASK(MSR_VEC)) ; Enable facility
- mr. r24,r24 ; Does the savearea exist?
- li r0,1 ; Get set to invalidate
- beq- setena ; Nothing to invalidate...
- lwz r25,SAVlvlvec(r24) ; Get the level of top savearea
- cmplw r4,r25 ; Is the top one ours?
- bne+ setena ; Not ours...
- stw r0,SAVlvlvec(r24) ; Invalidate the first one
- b setena ; Actually exit...
-
-chkvecfree: li r0,0 ; Clear a register
- lwz r24,ACT_MACT_VMX(r8) ; Get the vector save area
-
- bne- cr2,chkvecnfr ; Not our facility, do not clear...
- stw r0,PP_VMX_THREAD(r10) ; Clear vector owner
-chkvecnfr:
+ rlwinm r3,r22,0,0,19 ; Find main savearea header
+
+ lwz r8,quickfret(r31) ; Get the first in quickfret list (top)
+ lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom)
+ lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top)
+ lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom)
+ stw r8,SAVprev(r22) ; Link the old in (top)
+ stw r9,SAVprev+4(r22) ; Link the old in (bottom)
+ xor r3,r22,r3 ; Convert to physical
+ stw r2,quickfret(r31) ; Set the first in quickfret list (top)
+ stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom)
#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3402 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
+ lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
+ li r2,0x3301 ; (TEST/DEBUG)
+ oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
+ sc ; (TEST/DEBUG)
+#endif
- mr. r24,r24 ; Do we even have a savearea?
- beq+ setena ; Nope...
+fpusetlvl: stw r21,FPUlevel(r20) ; Save the level
+
+;
+; Here we check if we are at the right level
+; We need to check the level we are entering, not the one we are exiting.
+; Therefore, we will use the defer level if it is non-zero and we are
+; going into user state.
+;
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3403 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
- lwz r25,SAVlvlvec(r24) ; Get the level
- cmplwi r25,1 ; Is the top area invalid?
- cmplw cr1,r25,r3 ; Is it for the returned from context?
- beq vectoss ; It is invalid...
- bne cr1,setena ; Not for the returned context...
+fpuchkena: bt-- cr2_eq,fpuhasdfrd ; Skip if deferred, R26 already set up...
+ mr r26,r20 ; Use the non-deferred value
-vectoss: lwz r25,SAVprevec(r24) ; Get previous savearea
+fpuhasdfrd:
+#if 0
+ rlwinm. r0,r29,0,MSR_PR_BIT,MSR_PR_BIT ; (TEST/DEBUG) Going into user state?
+ beq fpunusrstt ; (TEST/DEBUG) Nope...
+ lwz r23,FPUlevel(r26) ; (TEST/DEBUG) Get the level ID
+ lwz r24,FPUsave(r26) ; (TEST/DEBUG) Get the first savearea
+ mr. r23,r23 ; (TEST/DEBUG) Should be level 0
+ beq++ fpulvl0 ; (TEST/DEBUG) Yes...
+
+ lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code
+ ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest
+ sc ; (TEST/DEBUG) System ABEND
+
+fpulvl0: mr. r24,r24 ; (TEST/DEBUG) Any context?
+ beq fpunusrstt ; (TEST/DEBUG) No...
+ lwz r23,SAVlevel(r24) ; (TEST/DEBUG) Get level of context
+ lwz r21,SAVprev+4(r24) ; (TEST/DEBUG) Get previous pointer
+ mr. r23,r23 ; (TEST/DEBUG) Is this our user context?
+ beq++ fpulvl0b ; (TEST/DEBUG) Yes...
+
+ lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code
+ ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest
+ sc ; (TEST/DEBUG) System ABEND
+
+fpulvl0b: mr. r21,r21 ; (TEST/DEBUG) Is there a forward chain?
+ beq++ fpunusrstt ; (TEST/DEBUG) Nope...
+
+ lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code
+ ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest
+ sc ; (TEST/DEBUG) System ABEND
+
+fpunusrstt: ; (TEST/DEBUG)
+#endif
+
+ lwz r21,FPUowner(r31) ; Get the ID of the live context
+ lwz r23,FPUlevel(r26) ; Get the level ID
+ lwz r24,FPUcpu(r26) ; Get the CPU that the context was last dispatched on
+ cmplw cr3,r26,r21 ; Do we have the live context?
+ cmplw r30,r23 ; Are we about to launch the live level?
+ bne-- cr3,chkvec ; No, can not possibly enable...
+ cmplw cr1,r19,r24 ; Was facility used on this processor last?
+ bne-- chkvec ; No, not live...
+ bne-- cr1,chkvec ; No, wrong cpu, have to enable later....
+
+ lwz r24,FPUsave(r26) ; Get the first savearea
+ mr. r24,r24 ; Any savearea?
+ beq++ fpuena ; Nope...
+ lwz r25,SAVlevel(r24) ; Get the level of savearea
+ lwz r0,SAVprev+4(r24) ; Get the previous
+
+ cmplw r30,r25 ; Is savearea for the level we are launching?
+ bne++ fpuena ; No, just go enable...
+
+ stw r0,FPUsave(r26) ; Pop the chain
+
+ rlwinm r3,r24,0,0,19 ; Find main savearea header
+
+ lwz r8,quickfret(r31) ; Get the first in quickfret list (top)
+ lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom)
+ lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top)
+ lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom)
+ stw r8,SAVprev(r24) ; Link the old in (top)
+ stw r9,SAVprev+4(r24) ; Link the old in (bottom)
+ xor r3,r24,r3 ; Convert to physical
+ stw r2,quickfret(r31) ; Set the first in quickfret list (top)
+ stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom)
+
#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3504 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- mr r5,r25 ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
- mr. r25,r25 ; Is there one?
- stw r25,ACT_MACT_VMX(r8) ; Set the new pointer
- beq vectoplvl ; Nope, we are at the top...
- lwz r25,SAVlvlvec(r25) ; Get the new level
+ lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
+ li r2,0x3302 ; (TEST/DEBUG)
+ oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
+ sc ; (TEST/DEBUG)
+#endif
-vectoplvl: lwz r19,SAVflags(r24) ; Get the savearea flags
+fpuena: ori r29,r29,lo16(MASK(MSR_FP)) ; Enable facility
+
+chkvec:
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3405 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
- rlwinm r22,r24,0,0,19 ; Round down to the base savearea block
- rlwinm r19,r19,0,3,1 ; Remove the vector in use flag
- stw r25,ACT_MACT_VMXlvl(r8) ; Set the new top level
- andis. r0,r19,hi16(SAVinuse) ; Still in use?
- stw r19,SAVflags(r24) ; Set the savearea flags
- bne- invliveve ; Go invalidate live vec...
-#if FPVECDBG
- lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
- li r2,0x3406 ; (TEST/DEBUG)
- oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
- sc ; (TEST/DEBUG)
-#endif
- lwz r23,SACvrswap(r22) ; Get the conversion from virtual to real
- lwz r20,PP_QUICKFRET(r10) ; Get the old quick fret head
- xor r23,r24,r23 ; Convert to physical
- stw r20,SAVqfret(r24) ; Back chain the quick release queue
- stw r23,PP_QUICKFRET(r10) ; Anchor it
-
-invliveve: lis r20,hi16(EXT(real_ncpus)) ; Get number of CPUs
- lis r23,hi16(EXT(per_proc_info)) ; Set base per_proc
- ori r20,r20,lo16(EXT(real_ncpus)) ; Other half of number of CPUs
- li r25,PP_VMX_THREAD ; Point to the vector owner address
- lwz r20,0(r20) ; Get number of processors active
- ori r23,r23,lo16(EXT(per_proc_info)) ; Set base per_proc
- li r2,0 ; Get something clear
-
-invlivevl: cmplw r23,r10 ; We can skip our processor
- addi r20,r20,-1 ; Count remaining processors
- beq invlivevn ; Skip ourselves...
-
-invliveva: lwarx r0,r25,r23 ; Get vector owner for this processor
- cmplw r0,r8 ; Do we own it?
- bne invlivevn ; Nope...
- stwcx. r2,r25,r23 ; Show not live
- bne- invliveva ; Someone else did this, try again...
+ lwz r21,VMXlevel(r20) ; Get the facility level
-invlivevn: mr. r20,r20 ; Have we finished?
- addi r23,r23,ppSize ; Bump to next
- bgt invlivevl ; Make sure we do all processors...
+ cmplw r27,r21 ; Are we returning from the active level?
+ bne+ vmxchkena ; Nope...
+
-setena: sync ; Make sure all is saved
- stw r27,ACT_MACT_VMXcpu(r8) ; Set the active CPU and release
+;
+; First clean up any live context we are returning from
+;
-setenanr: rlwinm r20,r12,(((31-vectorCngbit)+(MSR_VEC_BIT+1))&31),vectorCngbit,vectorCngbit ; Set flag if we enabled vector
- rlwimi. r20,r12,(((31-floatCngbit)+(MSR_FP_BIT+1))&31),floatCngbit,floatCngbit ; Set flag if we enabled floats
- beq setenaa ; Neither float nor vector turned on....
+ lwz r22,VMXcpu(r20) ; Get CPU this context was last dispatched on
+
+ stw r19,VMXcpu(r20) ; Claim context for us
+
+ eieio ; Make sure this gets out before owner clear
- lwz r5,ACT_MACT_SPF(r8) ; Get activation copy
- lwz r6,spcFlags(r10) ; Get per_proc copy
- or r5,r5,r20 ; Set vector/float changed bits in activation
- or r6,r6,r20 ; Set vector/float changed bits in per_proc
- stw r5,ACT_MACT_SPF(r8) ; Set activation copy
- stw r6,spcFlags(r10) ; Set per_proc copy
+ lis r23,hi16(EXT(PerProcTable)) ; Set base PerProcTable
+ slwi r22,r22,4 ; Find offset to the owner per_proc_entry
+ ori r23,r23,lo16(EXT(PerProcTable)) ; Set base PerProcTable
+ li r24,VMXowner ; Displacement to float owner
+ add r22,r23,r22 ; Point to the owner per_proc_entry
+ lwz r22,ppe_vaddr(r22) ; Point to the owner per_proc
+
+vmxinvothr: lwarx r23,r24,r22 ; Get the owner
-setenaa: stw r12,savesrr1(r3) ; Turn facility on or off
-
- mfdec r24 ; Get decrementer
- lwz r22,qactTimer(r8) ; Get high order quick activation timer
- mr. r24,r24 ; See if it has popped already...
- lwz r23,qactTimer+4(r8) ; Get low order qact timer
- ble- chkenax ; We have popped or are just about to...
-
-segtb: mftbu r20 ; Get the upper time base
- mftb r21 ; Get the low
- mftbu r19 ; Get upper again
- or. r0,r22,r23 ; Any time set?
- cmplw cr1,r20,r19 ; Did they change?
- beq+ chkenax ; No time set....
- bne- cr1,segtb ; Timebase ticked, get them again...
-
- subfc r6,r21,r23 ; Subtract current from qact time
- li r0,0 ; Make a 0
- subfe r5,r20,r22 ; Finish subtract
- subfze r0,r0 ; Get a 0 if qact was bigger than current, -1 otherwise
- andc. r12,r5,r0 ; Set 0 if qact has passed
- andc r13,r6,r0 ; Set 0 if qact has passed
- bne chkenax ; If high order is non-zero, this is too big for a decrementer
- cmplw r13,r24 ; Is this earlier than the decrementer? (logical compare takes care of high bit on)
- bge+ chkenax ; No, do not reset decrementer...
-
- mtdec r13 ; Set our value
-
-chkenax: lwz r6,SAVflags(r3) ; Pick up the flags of the old savearea
+ sub r0,r23,r20 ; Subtract one from the other
+ sub r21,r20,r23 ; Subtract the other from the one
+ or r21,r21,r0 ; Combine them
+ srawi r21,r21,31 ; Get a 0 if equal or -1 of not
+ and r23,r23,r21 ; Make 0 if same, unchanged if not
+ stwcx. r23,r24,r22 ; Try to invalidate it
+ bne-- vmxinvothr ; Try again if there was a collision...
+
+ isync
+
+;
+; Now if there is a savearea associated with the popped context, release it.
+; Either way, pop the level to the top stacked context.
+;
+
+ lwz r22,VMXsave(r20) ; Get pointer to the first savearea
+ li r21,0 ; Assume we popped all the way out
+ mr. r22,r22 ; Is there anything there?
+ beq++ vmxsetlvl ; No, see if we need to enable...
+
+ lwz r21,SAVlevel(r22) ; Get the level of that savearea
+ cmplw r21,r27 ; Is this the saved copy of the live stuff?
+ bne vmxsetlvl ; No, leave as is...
+
+ lwz r24,SAVprev+4(r22) ; Pick up the previous area
+ li r21,0 ; Assume we popped all the way out
+ mr. r24,r24 ; Any more context?
+ beq-- vmxonlyone ; Nope...
+ lwz r21,SAVlevel(r24) ; Get the level associated with save
+
+vmxonlyone: stw r24,VMXsave(r20) ; Dequeue this savearea
+
+ rlwinm r3,r22,0,0,19 ; Find main savearea header
+
+ lwz r8,quickfret(r31) ; Get the first in quickfret list (top)
+ lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom)
+ lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top)
+ lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom)
+ stw r8,SAVprev(r22) ; Link the old in (top)
+ stw r9,SAVprev+4(r22) ; Link the old in (bottom)
+ xor r3,r22,r3 ; Convert to physical
+ stw r2,quickfret(r31) ; Set the first in quickfret list (top)
+ stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom)
+
+#if FPVECDBG
+ lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
+ li r2,0x3401 ; (TEST/DEBUG)
+ oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
+ sc ; (TEST/DEBUG)
+#endif
+
+vmxsetlvl: stw r21,VMXlevel(r20) ; Save the level
+
+;
+; Here we check if we are at the right level
+;
+
+vmxchkena: lwz r21,VMXowner(r31) ; Get the ID of the live context
+ lwz r23,VMXlevel(r26) ; Get the level ID
+ cmplw r26,r21 ; Do we have the live context?
+ lwz r24,VMXcpu(r26) ; Get the CPU that the context was last dispatched on
+ bne-- setena ; No, can not possibly enable...
+ cmplw r30,r23 ; Are we about to launch the live level?
+ cmplw cr1,r19,r24 ; Was facility used on this processor last?
+ bne-- setena ; No, not live...
+ bne-- cr1,setena ; No, wrong cpu, have to enable later....
+
+ lwz r24,VMXsave(r26) ; Get the first savearea
+ mr. r24,r24 ; Any savearea?
+ beq++ vmxena ; Nope...
+ lwz r25,SAVlevel(r24) ; Get the level of savearea
+ lwz r0,SAVprev+4(r24) ; Get the previous
+ cmplw r30,r25 ; Is savearea for the level we are launching?
+ bne++ vmxena ; No, just go enable...
+
+ stw r0,VMXsave(r26) ; Pop the chain
+
+ rlwinm r3,r24,0,0,19 ; Find main savearea header
+
+ lwz r8,quickfret(r31) ; Get the first in quickfret list (top)
+ lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom)
+ lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top)
+ lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom)
+ stw r8,SAVprev(r24) ; Link the old in (top)
+ stw r9,SAVprev+4(r24) ; Link the old in (bottom)
+ xor r3,r24,r3 ; Convert to physical
+ stw r2,quickfret(r31) ; Set the first in quickfret list (top)
+ stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom)
+
+#if FPVECDBG
+ lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG)
+ li r2,0x3402 ; (TEST/DEBUG)
+ oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG)
+ sc ; (TEST/DEBUG)
+#endif
+
+vmxena: oris r29,r29,hi16(MASK(MSR_VEC)) ; Enable facility
+
+setena: lwz r18,umwSpace(r28) ; Get the space ID in case we are launching user
+ rlwinm. r0,r29,0,MSR_PR_BIT,MSR_PR_BIT ; Are we about to launch user state?
+ li r0,0 ; Get set to release quickfret holdoff
+ crmove cr7_eq,cr0_eq ; Remember if we are going to user state
+ rlwimi. r20,r29,(((31-floatCngbit)+(MSR_FP_BIT+1))&31),floatCngbit,floatCngbit ; Set flag if we enabled floats
+ lwz r19,deferctx(r28) ; Get any deferred facility context switch
+ rlwinm r20,r29,(((31-vectorCngbit)+(MSR_VEC_BIT+1))&31),vectorCngbit,vectorCngbit ; Set flag if we enabled vector
+ stw r29,savesrr1+4(r27) ; Turn facility on or off
+ stw r0,holdQFret(r31) ; Release quickfret
+ oris r18,r18,hi16(umwSwitchAway) ; Set the switch-away bit in case we go to user
+
+ beq setenaa ; Neither float nor vector turned on....
+
+ lwz r5,ACT_MACT_SPF(r28) ; Get activation copy
+ lwz r6,spcFlags(r31) ; Get per_proc copy
+ or r5,r5,r20 ; Set vector/float changed bits in activation
+ or r6,r6,r20 ; Set vector/float changed bits in per_proc
+ stw r5,ACT_MACT_SPF(r28) ; Set activation copy
+ stw r6,spcFlags(r31) ; Set per_proc copy
+
+setenaa: mfdec r24 ; Get decrementer
+ bf+ cr2_eq,nodefer ; No deferred to switch to...
+
+ li r20,0 ; Clear this
+ stw r26,curctx(r28) ; Make the facility context current
+ stw r20,deferctx(r28) ; Clear deferred context
+
+nodefer: lwz r22,qactTimer(r28) ; Get high order quick activation timer
+ mr. r24,r24 ; See if it has popped already...
+ lwz r23,qactTimer+4(r28) ; Get low order qact timer
+ ble- chkifuser ; We have popped or are just about to...
+
+segtb: mftbu r20 ; Get the upper time base
+ mftb r21 ; Get the low
+ mftbu r19 ; Get upper again
+ or. r0,r22,r23 ; Any time set?
+ cmplw cr1,r20,r19 ; Did they change?
+ beq++ chkifuser ; No time set....
+ bne-- cr1,segtb ; Timebase ticked, get them again...
+
+ subfc r6,r21,r23 ; Subtract current from qact time
+ li r0,0 ; Make a 0
+ subfe r5,r20,r22 ; Finish subtract
+ subfze r0,r0 ; Get a 0 if qact was bigger than current, -1 otherwise
+ andc. r12,r5,r0 ; Set 0 if qact has passed
+ andc r13,r6,r0 ; Set 0 if qact has passed
+ bne chkifuser ; If high order is non-zero, this is too big for a decrementer
+ cmplw r13,r24 ; Is this earlier than the decrementer? (logical compare takes care of high bit on)
+ bge++ chkifuser ; No, do not reset decrementer...
+
+ mtdec r13 ; Set our value
+
+chkifuser: addi r4,r28,SYSTEM_TIMER
+ mftb r3
+ beq-- cr7,chkifuser1 ; Skip this if we are going to kernel...
+ stw r18,umwSpace(r28) ; Half-invalidate to force MapUserAddressWindow to reload SRs
+ addi r4,r28,USER_TIMER
+
+chkifuser1: bl EXT(timer_event)
+
+chkenax:
-
#if DEBUG
- lwz r20,SAVact(r3) ; (TEST/DEBUG) Make sure our restore
- lwz r21,PP_CPU_DATA(r10) ; (TEST/DEBUG) context is associated
- lwz r21,CPU_ACTIVE_THREAD(r21) ; (TEST/DEBUG) with the current act.
- cmpwi r21,0 ; (TEST/DEBUG)
- beq- yeswereok ; (TEST/DEBUG)
- lwz r21,THREAD_TOP_ACT(r21) ; (TEST/DEBUG)
- cmplw r21,r20 ; (TEST/DEBUG)
- beq+ yeswereok ; (TEST/DEBUG)
-
- lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code
- ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest
- mr r21,r3 ; (TEST/DEBUG) Save the savearea address
- li r3,failContext ; (TEST/DEBUG) Bad state code
- sc ; (TEST/DEBUG) System ABEND
+ lwz r20,SAVact(r27) ; (TEST/DEBUG) Make sure our restore
+ mfsprg r21, 1 ; (TEST/DEBUG) with the current act.
+ cmpwi r21,0 ; (TEST/DEBUG)
+ beq-- yeswereok ; (TEST/DEBUG)
+ cmplw r21,r20 ; (TEST/DEBUG)
+ beq++ yeswereok ; (TEST/DEBUG)
+
+ lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code
+ ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest
+ mr r21,r27 ; (TEST/DEBUG) Save the savearea address
+ li r3,failContext ; (TEST/DEBUG) Bad state code
+ sc ; (TEST/DEBUG) System ABEND
yeswereok:
#endif
- rlwinm r5,r3,0,0,19 ; Round savearea down to page bndry
- rlwinm r6,r6,0,1,31 ; Mark savearea free
- lwz r5,SACvrswap(r5) ; Get the conversion from virtual to real
- stw r6,SAVflags(r3) ; Set savearea flags
- xor r3,r3,r5 ; Flip to physical address
- b EXT(exception_exit) ; We are all done now...
+ mr r3,r27 ; Pass savearea back
+ b EXT(exception_exit) ; We are all done now...
+;
+; Null PPC call - performance testing, does absolutely nothing
+;
+
+ .align 5
+
+ .globl EXT(ppcNull)
+
+LEXT(ppcNull)
+
+ li r3,-1 ; Make sure we test no asts
+ blr
+
+
+;
+; Instrumented null PPC call - performance testing, does absolutely nothing
+; Forces various timestamps to be returned.
+;
+
+ .align 5
+
+ .globl EXT(ppcNullinst)
+
+LEXT(ppcNullinst)
+
+ li r3,-1 ; Make sure we test no asts
+ blr
+
+
/*
* Here's where we handle the fastpath stuff
* We'll do what we can here because registers are already
.align 5
-fastpath: cmplwi cr3,r0,0x7FF1 ; Is it CthreadSetSelfNumber?
- bnelr- cr3 ; Not a fast path...
+fastpath: cmplwi cr3,r0,0x7FF5 ; Is this a null fastpath?
+ beq-- cr3,fastexutl ; Yes, bail fast...
+ cmplwi cr3,r0,0x7FF1 ; Is it CthreadSetSelfNumber?
+ bnelr-- cr3 ; Not a fast path...
/*
* void cthread_set_self(cproc_t p)
*
- * set's thread state "user_value"
+ * Set's thread state "user_value". In practice this is the thread-local-data-pointer (TLDP),
+ * though we do not interpret it. This call is mostly used by 32-bit tasks, but we save all 64 bits
+ * in case a 64-bit task wants to use this facility. They normally do not, because the 64-bit
+ * ABI reserves r13 for the TLDP.
*
* This op is invoked as follows:
* li r0, CthreadSetSelfNumber // load the fast-trap number
* sc // invoke fast-trap
* blr
- *
*/
CthreadSetSelfNumber:
-
- lwz r5,saver3(r4) /* Retrieve the self number */
- stw r5,CTHREAD_SELF(r13) /* Remember it */
- stw r5,UAW(r25) /* Prime the per_proc_info with it */
+ lwz r3,saver3+0(r4) /* get the TLDP passed in r3 */
+ lwz r5,saver3+4(r4) /* (all 64 bits, in case this is a 64-bit task) */
+ stw r3,CTHREAD_SELF+0(r13) /* Remember it in the activation... */
+ stw r5,CTHREAD_SELF+4(r13)
+ stw r3,UAW+0(r25) /* ...and in the per-proc */
+ stw r5,UAW+4(r25)
.globl EXT(fastexit)
EXT(fastexit):
- lwz r8,SAVflags(r4) /* Pick up the flags */
- rlwinm r9,r4,0,0,19 /* Round down to the base savearea block */
- rlwinm r8,r8,0,1,31 /* Clear the attached bit */
- lwz r9,SACvrswap(r9) /* Get the conversion from virtual to real */
- stw r8,SAVflags(r4) /* Set the flags */
- xor r3,r4,r9 /* Switch savearea to physical addressing */
- b EXT(exception_exit) /* Go back to the caller... */
+fastexutl: mr r3,r4 ; Pass back savearea
+ b EXT(exception_exit) ; Go back to the caller...
/*
checkassist:
lwz r0,saveexception(r4) ; Get the exception code
- lwz r23,savesrr1(r4) ; Get the interrupted MSR
+ lwz r23,savesrr1+4(r4) ; Get the interrupted MSR
lwz r26,ACT_MACT_BEDA(r13) ; Get Blue Box Descriptor Area
mtcrf 0x18,r23 ; Check what SRR1 says
lwz r24,ACT_MACT_BTS(r13) ; Get the table start
cmplwi r0,T_AST ; Check for T_AST trap
- lwz r27,savesrr0(r4) ; Get trapped address
+ lwz r27,savesrr0+4(r4) ; Get trapped address
crnand cr1_eq,SRR1_PRG_TRAP_BIT,MSR_PR_BIT ; We need both trap and user state
sub r24,r27,r24 ; See how far into it we are
cror cr0_eq,cr0_eq,cr1_eq ; Need to bail if AST or not trap or not user state
.globl EXT(retFromVM)
LEXT(retFromVM)
- mfsprg r10,0 ; Restore the per_proc info
+ mfsprg r10,1 ; Get the current activation
+ lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block
mr r8,r3 ; Get the activation
- lwz r4,SAVprev(r30) ; Pick up the previous savearea
+ lwz r4,SAVprev+4(r30) ; Pick up the previous savearea
mr r3,r30 ; Put savearea in proper register for common code
lwz r11,SAVflags(r30) ; Get the flags of the current savearea
rlwinm r11,r11,0,15,13 ; Clear the syscall flag
- lwz r1,ACT_THREAD(r8) ; and the active thread
+ mr r1,r8
stw r11,SAVflags(r3) ; Save back the flags (with reset stack cleared)
stw r4,ACT_MACT_PCB(r8) ; Point to the previous savearea (or 0 if none)
.align 5
.globl EXT(chandler)
-LEXT(chandler) /* Choke handler */
+LEXT(chandler) ; Choke handler
- lis r25,hi16(EXT(trcWork)) ; (TEST/DEBUG)
- li r31,0 ; (TEST/DEBUG)
- ori r25,r25,lo16(EXT(trcWork)) ; (TEST/DEBUG)
- stw r31,traceMask(r25) ; (TEST/DEBUG)
+ li r31,0 ; Get a 0
+ mfsprg r25,1 ; Get the current activation
+ lwz r25,ACT_PER_PROC(r25) ; Get the per_proc block
+ stw r31,traceMask(0) ; Force tracing off right now
- mfsprg r25,0 ; Get the per_proc
lwz r1,PP_DEBSTACKPTR(r25) ; Get debug stack pointer
cmpwi r1,-1 ; Are we already choking?
chokefirst: li r0,-1 ; Set choke value
mr. r1,r1 ; See if we are on debug stack yet
- lwz r10,saver1(r4) ;
+ lwz r10,saver1+4(r4) ;
stw r0,PP_DEBSTACKPTR(r25) ; Show we are choking
bne chokestart ; We are not on the debug stack yet...
;
versave:
-
#if 0
-;
-; Make sure that only the top FPU savearea is marked invalid
-;
-
- lis r28,hi16(EXT(default_pset)) ; (TEST/DEBUG)
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r28,r28,lo16(EXT(default_pset)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- li r20,0 ; (TEST/DEBUG)
- lwz r26,0(r27) ; (TEST/DEBUG)
- lwz r27,psthreadcnt(r28) ; (TEST/DEBUG)
- mr. r26,r26 ; (TEST/DEBUG)
- lwz r28,psthreads(r28) ; (TEST/DEBUG)
- bnelr- ; (TEST/DEBUG)
-
-fcknxtth: mr. r27,r27 ; (TEST/DEBUG)
+ lis r22,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
+ ori r22,r22,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
+ lwz r23,0(r22) ; (TEST/DEBUG)
+ mr. r23,r23 ; (TEST/DEBUG)
beqlr- ; (TEST/DEBUG)
-
- lwz r26,THREAD_TOP_ACT(r28) ; (TEST/DEBUG)
-
-fckact: mr. r26,r26 ; (TEST/DEBUG)
- bne+ fckact2 ; (TEST/DEBUG)
-
- lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Next in line
- subi r27,r27,1 ; (TEST/DEBUG)
- b fcknxtth ; (TEST/DEBUG)
-
-fckact2: lwz r20,ACT_MACT_FPU(r26) ; (TEST/DEBUG) Get FPU chain
- mr. r20,r20 ; (TEST/DEBUG) Are there any?
- beq+ fcknact ; (TEST/DEBUG) No...
-
-fckact3: lwz r20,SAVprefp(r20) ; (TEST/DEBUG) Get next in list
- mr. r20,r20 ; (TEST/DEBUG) Check next savearea
- beq+ fcknact ; (TEST/DEBUG) No...
-
- lwz r29,SAVlvlfp(r20) ; (TEST/DEBUG) Get the level
-
- cmplwi r29,1 ; (TEST/DEBUG) Is it invalid??
- bne+ fckact3 ; (TEST/DEBUG) Nope...
-
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- stw r27,0(r27) ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
-
-fcknact: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation
- b fckact ; (TEST/DEBUG)
+ mfsprg r20,1 ; Get the current activation
+ lwz r20,ACT_PER_PROC(r20) ; Get the per_proc block
+ lwz r21,pfAvailable(r20) ; (TEST/DEBUG)
+ mr. r21,r21 ; (TEST/DEBUG)
+ bnelr+ ; (TEST/DEBUG)
+
+ stw r22,0(r22) ; (TEST/DEBUG) Lock out more checks
+ BREAKPOINT_TRAP ; (TEST/DEBUG) Get into debugger
#endif
-#if 1
+#if 0
+ ;; This code is broken and migration will make the matter even worse
;
-; Make sure there are no circular links in the float chain
-; And that FP is marked busy in it.
-; And the only the top is marked invalid.
-; And that the owning PCB is correct.
+; Make sure that all savearea chains have the right type on them
;
lis r28,hi16(EXT(default_pset)) ; (TEST/DEBUG)
li r20,0 ; (TEST/DEBUG)
lwz r26,0(r27) ; (TEST/DEBUG)
lwz r27,psthreadcnt(r28) ; (TEST/DEBUG)
- mr. r26,r26 ; (TEST/DEBUG)
+ mr. r26,r26 ; (TEST/DEBUG) Have we locked the test out?
lwz r28,psthreads(r28) ; (TEST/DEBUG)
- bnelr- ; (TEST/DEBUG)
-
-fcknxtth: mr. r27,r27 ; (TEST/DEBUG)
- beqlr- ; (TEST/DEBUG)
-
- lwz r26,THREAD_TOP_ACT(r28) ; (TEST/DEBUG)
-
-fckact: mr. r26,r26 ; (TEST/DEBUG)
- bne+ fckact2 ; (TEST/DEBUG)
-
- lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Next in line
- subi r27,r27,1 ; (TEST/DEBUG)
- b fcknxtth ; (TEST/DEBUG)
-
-fckact2: lwz r20,ACT_MACT_FPU(r26) ; (TEST/DEBUG) Get FPU chain
- li r29,1 ; (TEST/DEBUG)
- li r22,0 ; (TEST/DEBUG)
-
-fckact3: mr. r20,r20 ; (TEST/DEBUG) Are there any?
- beq+ fckact5 ; (TEST/DEBUG) No...
+ mflr r31 ; (TEST/DEBUG) Save return
+ bnelr- ; (TEST/DEBUG) Test already triggered, skip...
+ b fckgo ; (TEST/DEBUG) Join up...
- addi r22,r22,1 ; (TEST/DEBUG) Count chain depth
-
- lwz r21,SAVflags(r20) ; (TEST/DEBUG) Get the flags
- rlwinm. r21,r21,0,1,1 ; (TEST/DEBUG) FP busy?
- bne+ fckact3a ; (TEST/DEBUG) Yeah...
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- stw r27,0(r27) ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG) Die
-
-fckact3a: cmplwi r22,1 ; (TEST/DEBUG) At first SA?
- beq+ fckact3b ; (TEST/DEBUG) Yeah, invalid is ok...
- lwz r21,SAVlvlfp(r20) ; (TEST/DEBUG) Get level
- cmplwi r21,1 ; (TEST/DEBUG) Is it invalid?
- bne+ fckact3b ; (TEST/DEBUG) Nope, it is ok...
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- stw r27,0(r27) ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG) Die
-
-fckact3b: lwz r21,SAVact(r20) ; (TEST/DEBUG) Get the owner
- cmplw r21,r26 ; (TEST/DEBUG) Correct activation?
- beq+ fckact3c ; (TEST/DEBUG) Yup...
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- stw r27,0(r27) ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG) Die
+fcknext: mr. r27,r27 ; (TEST/DEBUG) Any more threads?
+ bne+ fckxxx ; (TEST/DEBUG) Yes...
-fckact3c: ; (TEST/DEBUG)
- lbz r21,SAVflags+3(r20) ; (TEST/DEBUG) Pick up the test byte
- mr. r21,r21 ; (TEST/DEBUG) marked?
- beq+ fckact4 ; (TEST/DEBUG) No, good...
-
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- stw r27,0(r27) ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
+ mtlr r31 ; (TEST/DEBUG) Restore return
+ blr ; (TEST/DEBUG) Leave...
-fckact4: stb r29,SAVflags+3(r20) ; (TEST/DEBUG) Set the test byte
- lwz r20,SAVprefp(r20) ; (TEST/DEBUG) Next in list
- b fckact3 ; (TEST/DEBUG) Try it...
-
-fckact5: lwz r20,ACT_MACT_FPU(r26) ; (TEST/DEBUG) Get FPU chain
- li r29,0 ; (TEST/DEBUG)
+fckxxx: lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Get next thread
-fckact6: mr. r20,r20 ; (TEST/DEBUG) Are there any?
- beq+ fcknact ; (TEST/DEBUG) No...
+fckgo: subi r27,r27,1 ; (TEST/DEBUG) Decrement thread count
+ lwz r24,THREAD_TOP_ACT(r28) ; (TEST/DEBUG) Get activation for the thread
+ lwz r20,ACT_MACT_PCB(r24) ; (TEST/DEBUG) Get the normal context
+ li r21,SAVgeneral ; (TEST/DEBUG) Make sure this is all general context
+ bl versavetype ; (TEST/DEBUG) Check the chain
- stb r29,SAVflags+3(r20) ; (TEST/DEBUG) Clear the test byte
- lwz r20,SAVprefp(r20) ; (TEST/DEBUG) Next in list
- b fckact6 ; (TEST/DEBUG) Try it...
+ lwz r20,facctx+FPUsave(r24) ; (TEST/DEBUG) Get regular floating point
+ li r21,SAVfloat ; (TEST/DEBUG) Make sure this is all floating point
+ bl versavetype ; (TEST/DEBUG) Check the chain
-fcknact: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation
- b fckact ; (TEST/DEBUG)
-#endif
-
-
-#if 0
-;
-; Make sure in use count matches found savearea. This is
-; not always accurate. There is a variable "fuzz" factor in count.
-
- lis r28,hi16(EXT(default_pset)) ; (TEST/DEBUG)
- lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r28,r28,lo16(EXT(default_pset)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- li r20,0 ; (TEST/DEBUG)
- lwz r26,0(r27) ; (TEST/DEBUG)
- lwz r27,psthreadcnt(r28) ; (TEST/DEBUG)
- mr. r26,r26 ; (TEST/DEBUG)
- lwz r28,psthreads(r28) ; (TEST/DEBUG)
- bnelr- ; (TEST/DEBUG)
-
-cknxtth: mr. r27,r27 ; (TEST/DEBUG)
- beq- cktotal ; (TEST/DEBUG)
+ lwz r20,facctx+VMXsave(r24) ; (TEST/DEBUG) Get regular vector point
+ li r21,SAVvector ; (TEST/DEBUG) Make sure this is all vector
+ bl versavetype ; (TEST/DEBUG) Check the chain
- lwz r26,THREAD_TOP_ACT(r28) ; (TEST/DEBUG)
-
-ckact: mr. r26,r26 ; (TEST/DEBUG)
- bne+ ckact2 ; (TEST/DEBUG)
+ lwz r29,vmmControl(r24) ; (TEST/DEBUG) Get the virtual machine control blocks
+ mr. r29,r29 ; (TEST/DEBUG) Are there any?
+ beq+ fcknext ; (TEST/DEBUG) Nope, next thread...
- lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Next in line
- subi r27,r27,1 ; (TEST/DEBUG)
- b cknxtth ; (TEST/DEBUG)
+ li r22,kVmmMaxContextsPerThread ; (TEST/DEBUG) Get the number of control blocks
+ subi r29,r29,vmmCEntrySize ; (TEST/DEBUG) Get running start
-ckact2: lwz r29,ACT_MACT_PCB(r26) ; (TEST/DEBUG)
+fcknvmm: subi r22,r22,1 ; (TEST/DEBUG) Do all of them
+ mr. r22,r22 ; (TEST/DEBUG) Are we all done?
+ addi r29,r29,vmmCEntrySize ; (TEST/DEBUG) Get the next entry
+ blt- fcknext ; (TEST/DEBUG) Yes, check next thread...
-cknorm: mr. r29,r29 ; (TEST/DEBUG)
- beq- cknormd ; (TEST/DEBUG)
+ lwz r23,vmmFlags(r29) ; (TEST/DEBUG) Get entry flags
+ rlwinm. r23,r23,0,0,0 ; (TEST/DEBUG) Is this in use?
+ beq+ fcknvmm ; (TEST/DEBUG) Not in use...
- addi r20,r20,1 ; (TEST/DEBUG) Count normal savearea
+ lwz r20,vmmFacCtx+FPUsave(r29) ; (TEST/DEBUG) Get regular floating point
+ li r21,SAVfloat ; (TEST/DEBUG) Make sure this is all floating point
+ bl versavetype ; (TEST/DEBUG) Check the chain
- lwz r29,SAVprev(r29) ; (TEST/DEBUG)
- b cknorm ; (TEST/DEBUG)
-
-cknormd: lwz r29,ACT_MACT_FPU(r26) ; (TEST/DEBUG)
-
-ckfpu: mr. r29,r29 ; (TEST/DEBUG)
- beq- ckfpud ; (TEST/DEBUG)
+ lwz r20,vmmFacCtx+VMXsave(r29) ; (TEST/DEBUG) Get regular vector point
+ li r21,SAVvector ; (TEST/DEBUG) Make sure this is all vector
+ bl versavetype ; (TEST/DEBUG) Check the chain
+ b fcknvmm ; (TEST/DEBUG) Get then vmm block...
- lwz r21,SAVflags(r29) ; (TEST/DEBUG)
- rlwinm. r21,r21,0,0,0 ; (TEST/DEBUG) See if already counted
- bne- cknfpu ; (TEST/DEBUG)
+versavetype:
+ mr. r20,r20 ; (TEST/DEBUG) Chain done?
+ beqlr- ; (TEST/DEBUG) Yes...
- addi r20,r20,1 ; (TEST/DEBUG) Count fpu savearea
+ lwz r23,SAVflags(r20) ; (TEST/DEBUG) Get the flags
+ rlwinm r23,r23,24,24,31 ; (TEST/DEBUG) Position it
+ cmplw r23,r21 ; (TEST/DEBUG) Are we the correct type?
+ beq+ versvok ; (TEST/DEBUG) This one is ok...
-cknfpu: lwz r29,SAVprefp(r29) ; (TEST/DEBUG)
- b ckfpu ; (TEST/DEBUG)
+ lis r22,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
+ ori r22,r22,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
+ stw r22,0(r22) ; (TEST/DEBUG) Lock out more checks
+ BREAKPOINT_TRAP ; (TEST/DEBUG) Get into debugger
-ckfpud: lwz r29,ACT_MACT_VMX(r26) ; (TEST/DEBUG)
-
-ckvmx: mr. r29,r29 ; (TEST/DEBUG)
- beq- ckvmxd ; (TEST/DEBUG)
-
- lwz r21,SAVflags(r29) ; (TEST/DEBUG)
- rlwinm. r21,r21,0,0,1 ; (TEST/DEBUG) See if already counted
- bne- cknvmx ; (TEST/DEBUG)
-
- addi r20,r20,1 ; (TEST/DEBUG) Count vector savearea
-
-cknvmx: lwz r29,SAVprevec(r29) ; (TEST/DEBUG)
- b ckvmx ; (TEST/DEBUG)
-
-ckvmxd: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation
- b ckact ; (TEST/DEBUG)
+versvok: lwz r20,SAVprev+4(r20) ; (TEST/DEBUG) Get the previous one
+ b versavetype ; (TEST/DEBUG) Go check its type...
+#endif
-cktotal: lis r28,hi16(EXT(saveanchor)) ; (TEST/DEBUG)
- lis r27,hi16(EXT(real_ncpus)) ; (TEST/DEBUG)
- ori r28,r28,lo16(EXT(saveanchor)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(real_ncpus)) ; (TEST/DEBUG)
- lwz r21,SVinuse(r28) ; (TEST/DEBUG)
- lwz r27,0(r27) ; (TEST/DEBUG) Get the number of CPUs
- sub. r29,r21,r20 ; (TEST/DEBUG) Get number accounted for
- blt- badsave ; (TEST/DEBUG) Have too many in use...
- sub r26,r29,r27 ; (TEST/DEBUG) Should be 1 unaccounted for for each processor
- cmpwi r26,10 ; (TEST/DEBUG) Allow a 10 area slop factor
- bltlr+ ; (TEST/DEBUG)
-
-badsave: lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG)
- ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG)
- stw r27,0(r27) ; (TEST/DEBUG)
- BREAKPOINT_TRAP ; (TEST/DEBUG)
-#endif
#endif