X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/0b4e3aa066abc0728aacb4bbeb86f53f9737156e..0c530ab8987f0ae6a1a3d9284f40182b88852816:/osfmk/ppc/hw_exception.s?ds=sidebyside diff --git a/osfmk/ppc/hw_exception.s b/osfmk/ppc/hw_exception.s index 55b30c73e..4c66797de 100644 --- a/osfmk/ppc/hw_exception.s +++ b/osfmk/ppc/hw_exception.s @@ -1,5 +1,5 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2005 Apple Computer, Inc. All rights reserved. * * @APPLE_LICENSE_HEADER_START@ * @@ -36,6 +36,7 @@ #include #include #include +#include #include #include @@ -44,12 +45,14 @@ #include #include #include -#include +#include #define VERIFYSAVE 0 #define FPVECDBG 0 - +#define FPFLOOD 0 +#define INSTRUMENT 0 + /* * thandler(type) * @@ -62,12 +65,12 @@ /* * If pcb.ksp == 0 then the kernel stack is already busy, - * we save ppc_saved state below the current stack pointer, + * we make a stack frame * leaving enough space for the 'red zone' in case the * trapped thread was in the middle of saving state below * its stack pointer. * - * otherwise we save a ppc_saved_state in the pcb, and switch to + * otherwise we make a stack frame and * the kernel stack (setting pcb.ksp to 0) * * on return, we do the reverse, the last state is popped from the pcb @@ -87,177 +90,152 @@ .align 5 .globl EXT(thandler) -LEXT(thandler) /* Trap handler */ - -#if 0 -; -; NOTE: This trap will hang VPC running Windows98 (and probably others)... -; - lwz r25,savedar(r4) ; (TEST/DEBUG) - cmplwi r25,0x298 ; (TEST/DEBUG) - -deadloop: addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - addi r25,r25,1 ; (TEST/DEBUG) - beq- deadloop ; (TEST/DEBUG) -#endif +LEXT(thandler) ; Trap handler - mfsprg r25,0 /* Get the per_proc */ + mfsprg r13,1 ; Get the current activation + lwz r25,ACT_PER_PROC(r13) ; Get the per_proc block - lwz r1,PP_ISTACKPTR(r25) ; Get interrupt stack pointer + lwz r1,PP_ISTACKPTR(r25) ; Get interrupt stack pointer - lwz r6,PP_CPU_DATA(r25) /* Get point to cpu specific data */ - cmpwi cr0,r1,0 ; Are we on interrupt stack? - lwz r6,CPU_ACTIVE_THREAD(r6) /* Get the pointer to the currently active thread */ - beq- cr0,EXT(ihandler) ; If on interrupt stack, treat this as interrupt... - lwz r13,THREAD_TOP_ACT(r6) /* Point to the active activation */ - lwz r26,ACT_MACT_SPF(r13) ; Get special flags - lwz r8,ACT_MACT_PCB(r13) /* Get the last savearea used */ + cmpwi cr0,r1,0 ; Are we on interrupt stack? + mr r6,r13 + beq- cr0,EXT(ihandler) ; If on interrupt stack, treat this as interrupt... + lwz r26,ACT_MACT_SPF(r13) ; Get special flags + lwz r8,ACT_MACT_PCB(r13) ; Get the last savearea used rlwinm. r26,r26,0,bbThreadbit,bbThreadbit ; Do we have Blue Box Assist active? - lwz r1,ACT_MACT_KSP(r13) ; Get the top of kernel stack - bnel- checkassist /* See if we should assist this */ - stw r4,ACT_MACT_PCB(r13) /* Point to our savearea */ - stw r8,SAVprev(r4) /* Queue the new save area in the front */ + lwz r1,ACT_MACT_KSP(r13) ; Get the top of kernel stack + bnel- checkassist ; See if we should assist this + stw r4,ACT_MACT_PCB(r13) ; Point to our savearea + stw r8,SAVprev+4(r4) ; Queue the new save area in the front #if VERIFYSAVE - bl versave ; (TEST/DEBUG) + bl versave ; (TEST/DEBUG) #endif - lwz r9,THREAD_KERNEL_STACK(r6) ; Get our kernel stack start - cmpwi cr1,r1,0 ; Are we already on kernel stack? - stw r13,SAVact(r4) ; Mark the savearea as belonging to this activation - lwz r26,saver1(r4) ; Get the stack at interrupt time + lwz r9,THREAD_KERNEL_STACK(r6) ; Get our kernel stack start + cmpwi cr1,r1,0 ; Are we already on kernel stack? + stw r13,SAVact(r4) ; Mark the savearea as belonging to this activation + lwz r26,saver1+4(r4) ; Get the stack at interrupt time - bne+ cr1,.L_kstackfree ; We are not on kernel stack yet... + bne+ cr1,.L_kstackfree ; We are not on kernel stack yet... - subi r1,r26,FM_REDZONE ; Make a red zone on interrupt time kernel stack + subi r1,r26,FM_REDZONE ; Make a red zone on interrupt time kernel stack .L_kstackfree: - lwz r7,savesrr1(r4) /* Pick up the entry MSR */ - sub r9,r1,r9 ; Get displacment into the kernel stack - li r0,0 /* Make this 0 */ - cmplwi cr2,r9,KERNEL_STACK_SIZE ; Do we still have room on the stack? - beq cr1,.L_state_on_kstack /* using above test for pcb/stack */ + lwz r31,savesrr1+4(r4) ; Pick up the entry MSR + sub r9,r1,r9 ; Get displacment into the kernel stack + li r0,0 ; Make this 0 + rlwinm. r0,r9,0,28,31 ; Verify that we have a 16-byte aligned stack (and get a 0) + cmplwi cr2,r9,KERNEL_STACK_SIZE ; Do we still have room on the stack? + beq cr1,.L_state_on_kstack ; using above test for pcb/stack - stw r0,ACT_MACT_KSP(r13) /* Show that we have taken the stack */ + stw r0,ACT_MACT_KSP(r13) ; Show that we have taken the stack .L_state_on_kstack: - rlwinm. r6,r7,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on? - bgt- cr2,kernelStackBad ; Kernel stack is bogus... -kernelStackNotBad: ; Not really - beq+ tvecoff ; Vector off, do not save vrsave... - lwz r3,savevrsave(r4) ; Get the VRSAVE register - stw r3,liveVRS(r25) ; Set the live value - -tvecoff: rlwinm. r3,r7,0,MSR_FP_BIT,MSR_FP_BIT ; Was floating point on? - subi r1,r1,FM_SIZE /* Push a header onto the current stack */ - beq+ tfpoff /* Floating point was off... */ - lwz r3,savexfpscr(r4) ; Grab the just saved FPSCR - stw r3,liveFPSCR(r25) ; Make it the live copy - -tfpoff: stw r26,FM_BACKPTR(r1) ; Link back to the previous frame + lwz r9,savevrsave(r4) ; Get the VRSAVE register + bne-- kernelStackUnaligned ; Stack is unaligned... + rlwinm. r6,r31,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on? + subi r1,r1,FM_SIZE ; Push a header onto the current stack + bgt-- cr2,kernelStackBad ; Kernel stack is bogus... + +kernelStackNotBad: ; Vector was off + beq++ tvecoff ; Vector off, do not save vrsave... + stw r9,liveVRS(r25) ; Set the live value + +tvecoff: stw r26,FM_BACKPTR(r1) ; Link back to the previous frame #if DEBUG /* If debugging, we need two frames, the first being a dummy * which links back to the trapped routine. The second is * that which the C routine below will need */ - lwz r3,savesrr0(r4) /* Get the point of interruption */ - stw r3,FM_LR_SAVE(r1) /* save old instr ptr as LR value */ - stwu r1, -FM_SIZE(r1) /* and make new frame */ + lwz r3,savesrr0+4(r4) ; Get the point of interruption + stw r3,FM_LR_SAVE(r1) ; save old instr ptr as LR value + stwu r1, -FM_SIZE(r1) ; and make new frame #endif /* DEBUG */ + mr r30,r4 + lwz r3,SAVtime+4(r4) + addi r4,r13,SYSTEM_TIMER + bl EXT(timer_event) /* call trap handler proper, with - * ARG0 = type (not yet, holds pcb ptr) - * ARG1 = saved_state ptr (already there) - * ARG2 = dsisr (already there) - * ARG3 = dar (already there) + * ARG0 = type + * ARG1 = saved_state ptr + * ARG2 = dsisr + * ARG3 = dar */ - - lwz r3,saveexception(r4) /* Get the exception code */ - lwz r0,ACT_MACT_SPF(r13) ; Get the special flags + mr r4,r30 + lwz r3,saveexception(r30) ; Get the exception code + lwz r0,ACT_MACT_SPF(r13) ; Get the special flags - addi r5,r3,-T_DATA_ACCESS ; Adjust to start of range + addi r5,r3,-T_DATA_ACCESS ; Adjust to start of range rlwinm. r0,r0,0,runningVMbit,runningVMbit ; Are we in VM state? (cr0_eq == 0 if yes) - cmplwi cr2,r5,T_RUNMODE_TRACE-T_DATA_ACCESS ; Are we still in range? (cr_gt if not) + cmplwi cr2,r5,T_TRACE-T_DATA_ACCESS ; Are we still in range? (cr_gt if not) - lwz r5,savedsisr(r4) /* Get the saved DSISR */ + lwz r5,savedsisr(r4) ; Get the saved DSISR - crnor cr7_eq,cr0_eq,cr2_gt ; We should intercept if in VM and is a true trap (cr7_eq == 1 if yes) - rlwinm. r0,r7,0,MSR_PR_BIT,MSR_PR_BIT ; Are we trapping from supervisor state? (cr0_eq == 1 if yes) + crnor cr7_eq,cr0_eq,cr2_gt ; We should intercept if in VM and is a true trap (cr7_eq == 1 if yes) + rlwinm. r0,r31,0,MSR_PR_BIT,MSR_PR_BIT ; Are we trapping from supervisor state? (cr0_eq == 1 if yes) - cmpi cr2,r3,T_PREEMPT ; Is this a preemption? + cmpi cr2,r3,T_PREEMPT ; Is this a preemption? + + beq-- .L_check_VM + stw r4,ACT_MACT_UPCB(r13) ; Store user savearea +.L_check_VM: - crandc cr0_eq,cr7_eq,cr0_eq ; Do not intercept if we are in the kernel (cr0_eq == 1 if yes) + crandc cr0_eq,cr7_eq,cr0_eq ; Do not intercept if we are in the kernel (cr0_eq == 1 if yes) - lwz r6,savedar(r4) /* Get the DAR */ + lwz r6,savedar(r4) ; Get the DAR (top) + lwz r7,savedar+4(r4) ; Get the DAR (bottom) - beq- cr2, .L_call_trap /* Don't turn on interrupts for T_PREEMPT */ - beq- exitFromVM ; Any true trap but T_MACHINE_CHECK exits us from the VM... + beq- cr2,.L_call_trap ; Do not turn on interrupts for T_PREEMPT + beq- exitFromVM ; Any true trap but T_MACHINE_CHECK exits us from the VM... /* syscall exception might warp here if there's nothing left * to do except generate a trap */ .L_call_trap: -#if 0 - lis r0,HIGH_ADDR(CutTrace) /* (TEST/DEBUG) */ - oris r0,r0,LOW_ADDR(CutTrace) /* (TEST/DEBUG) */ - sc /* (TEST/DEBUG) */ -#endif +#if FPFLOOD + stfd f31,emfp31(r25) ; (TEST/DEBUG) +#endif + bl EXT(trap) + lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable + mfmsr r7 ; Get the MSR + ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE + andc r7,r7,r10 ; Turn off VEC, FP, and EE + mtmsr r7 ; Disable for interrupts + mfsprg r8,1 ; Get the current activation + lwz r10,ACT_PER_PROC(r8) ; Get the per_proc block /* - * Ok, return from C function - * * This is also the point where new threads come when they are created. * The new thread is setup to look like a thread that took an * interrupt and went immediatly into trap. - * */ thread_return: + lwz r11,SAVflags(r3) ; Get the flags of the current savearea + lwz r0,savesrr1+4(r3) ; Get the MSR we are going to + lwz r4,SAVprev+4(r3) ; Pick up the previous savearea + mfsprg r8,1 ; Get the current thread + rlwinm r11,r11,0,15,13 ; Clear the syscall flag + rlwinm. r0,r0,0,MSR_PR_BIT,MSR_PR_BIT ; Are we going to the user? + mr r1,r8 + stw r11,SAVflags(r3) ; Save back the flags (with reset stack cleared) + + lwz r5,THREAD_KERNEL_STACK(r1) ; Get the base pointer to the stack + stw r4,ACT_MACT_PCB(r8) ; Point to the previous savearea (or 0 if none) + addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE ; Reset to empty - mfmsr r7 /* Get the MSR */ - lwz r4,SAVprev(r3) /* Pick up the previous savearea */ - rlwinm r7,r7,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Clear the interrupt enable mask */ - lwz r11,SAVflags(r3) /* Get the flags of the current savearea */ - mtmsr r7 /* Disable for interrupts */ - - mfsprg r10,0 /* Restore the per_proc info */ - - lwz r8,savesrr1(r3) ; Get the MSR we are going to - lwz r1,PP_CPU_DATA(r10) /* Get the CPU data area */ - rlwinm r11,r11,0,15,13 /* Clear the syscall flag */ - lwz r1,CPU_ACTIVE_THREAD(r1) /* and the active thread */ - rlwinm. r8,r8,0,MSR_PR_BIT,MSR_PR_BIT ; Are we going to the user? - lwz r8,THREAD_TOP_ACT(r1) /* Now find the current activation */ - stw r11,SAVflags(r3) /* Save back the flags (with reset stack cleared) */ - -#if 0 - lis r0,HIGH_ADDR(CutTrace) /* (TEST/DEBUG) */ - oris r0,r0,LOW_ADDR(CutTrace) /* (TEST/DEBUG) */ - sc /* (TEST/DEBUG) */ -#endif - stw r4,ACT_MACT_PCB(r8) /* Point to the previous savearea (or 0 if none) */ - - beq- chkfac ; We are not leaving the kernel yet... + beq-- chkfac ; We are not leaving the kernel yet... - lwz r5,THREAD_KERNEL_STACK(r1) /* Get the base pointer to the stack */ - addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE /* Reset to empty */ - stw r5,ACT_MACT_KSP(r8) /* Save the empty stack pointer */ - b chkfac /* Go end it all... */ + stw r5,ACT_MACT_KSP(r8) ; Save the empty stack pointer + b chkfac ; Go end it all... ; @@ -267,16 +245,22 @@ thread_return: kernelStackBad: - lwz r3,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top + lwz r3,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top subi r3,r3,KERNEL_STACK_SIZE-FM_SIZE ; Adjust to start of stack - sub r3,r1,r3 ; Get displacement into debug stack + sub r3,r1,r3 ; Get displacement into debug stack cmplwi cr2,r3,KERNEL_STACK_SIZE-FM_SIZE ; Check if we are on debug stack - blt+ cr2,kernelStackNotBad ; Yeah, that is ok too... + blt+ cr2,kernelStackNotBad ; Yeah, that is ok too... + + lis r0,hi16(Choke) ; Choke code + ori r0,r0,lo16(Choke) ; and the rest + li r3,failStack ; Bad stack code + sc ; System ABEND - lis r0,hi16(Choke) ; Choke code - ori r0,r0,lo16(Choke) ; and the rest - li r3,failStack ; Bad stack code - sc ; System ABEND +kernelStackUnaligned: + lis r0,hi16(Choke) ; Choke code + ori r0,r0,lo16(Choke) ; and the rest + li r3,failUnalignedStk ; Unaligned stack code + sc ; System ABEND /* @@ -311,129 +295,138 @@ kernelStackBad: .align 5 .globl EXT(shandler) -LEXT(shandler) /* System call handler */ - - mfsprg r25,0 /* Get the per proc area */ - lwz r0,saver0(r4) /* Get the original syscall number */ - lwz r17,PP_ISTACKPTR(r25) ; Get interrupt stack pointer - rlwinm r15,r0,0,0,19 ; Clear the bottom of call number for fast check - lwz r16,PP_CPU_DATA(r25) /* Assume we need this */ - mr. r17,r17 ; Are we on interrupt stack? - lwz r7,savesrr1(r4) ; Get the SRR1 value - beq- EXT(ihandler) ; On interrupt stack, not allowed... +LEXT(shandler) ; System call handler + + lwz r7,savesrr1+4(r4) ; Get the SRR1 value + mfsprg r13,1 ; Get the current activation + lwz r25,ACT_PER_PROC(r13) ; Get the per_proc block + lwz r0,saver0+4(r4) ; Get the original syscall number + lwz r17,PP_ISTACKPTR(r25) ; Get interrupt stack pointer + rlwinm r15,r0,0,0,19 ; Clear the bottom of call number for fast check + mr. r17,r17 ; Are we on interrupt stack? + lwz r9,savevrsave(r4) ; Get the VRsave register + beq-- EXT(ihandler) ; On interrupt stack, not allowed... rlwinm. r6,r7,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on? - lwz r16,CPU_ACTIVE_THREAD(r16) /* Get the thread pointer */ - - beq+ svecoff ; Vector off, do not save vrsave... - lwz r6,savevrsave(r4) ; Get the VRSAVE register - stw r6,liveVRS(r25) ; Set the live value - -svecoff: rlwinm. r6,r7,0,MSR_FP_BIT,MSR_FP_BIT ; Was floating point on? - lwz r13,THREAD_TOP_ACT(r16) /* Pick up the active thread */ - beq+ sfpoff ; Skip if floating point is off... - lwz r9,savexfpscr(r4) ; Grab the just saved FPSCR - stw r9,liveFPSCR(r25) ; Make it the live copy + mr r16,r13 + beq++ svecoff ; Vector off, do not save vrsave... + stw r9,liveVRS(r25) ; Set the live value +; ; Check if SCs are being redirected for the BlueBox or to VMM +; -sfpoff: lwz r6,ACT_MACT_SPF(r13) ; Pick up activation special flags - mtcrf 0x41,r6 ; Check special flags - crmove cr6_eq,runningVMbit ; Remember if we are in VMM - bf+ bbNoMachSCbit,noassist ; Take branch if SCs are not redirected - lwz r26,ACT_MACT_BEDA(r13) ; Pick up the pointer to the blue box exception area - b EXT(atomic_switch_syscall) ; Go to the assist... - -noassist: cmplwi r15,0x7000 /* Do we have a fast path trap? */ - lwz r14,ACT_MACT_PCB(r13) /* Now point to the PCB */ - beql+ fastpath /* We think it's a fastpath... */ - - lwz r1,ACT_MACT_KSP(r13) /* Get the kernel stack pointer */ +svecoff: lwz r6,ACT_MACT_SPF(r13) ; Pick up activation special flags + mtcrf 0x40,r6 ; Check special flags + mtcrf 0x01,r6 ; Check special flags + crmove cr6_eq,runningVMbit ; Remember if we are in VMM + bne++ cr6,sVMchecked ; Not running VM + lwz r18,spcFlags(r25) ; Load per_proc special flags + rlwinm. r18,r18,0,FamVMmodebit,FamVMmodebit ; Is FamVMmodebit set? + beq sVMchecked ; Not in FAM + cmpwi r0,0x6004 ; Is it vmm_dispatch syscall: + bne sVMchecked + lwz r26,saver3+4(r4) ; Get the original syscall number + cmpwi cr6,r26,kvmmExitToHost ; vmm_exit_to_host request +sVMchecked: + bf++ bbNoMachSCbit,noassist ; Take branch if SCs are not redirected + lwz r26,ACT_MACT_BEDA(r13) ; Pick up the pointer to the blue box exception area + b EXT(atomic_switch_syscall) ; Go to the assist... + +noassist: cmplwi r15,0x7000 ; Do we have a fast path trap? + lwz r14,ACT_MACT_PCB(r13) ; Now point to the PCB + beql fastpath ; We think it is a fastpath... + + lwz r1,ACT_MACT_KSP(r13) ; Get the kernel stack pointer #if DEBUG - mr. r1,r1 /* Are we already on the kernel stack? */ - li r3,T_SYSTEM_CALL /* Yup, pretend we had an interrupt... */ - beq- EXT(ihandler) /* Bad boy, bad boy... What'cha gonna do when they come for you? */ + mr. r1,r1 ; Are we already on the kernel stack? + li r3,T_SYSTEM_CALL ; Yup, pretend we had an interrupt... + beq- EXT(ihandler) ; Bad boy, bad boy... What cha gonna do when they come for you? #endif /* DEBUG */ - stw r4,ACT_MACT_PCB(r13) /* Point to our savearea */ - li r0,0 /* Clear this out */ - stw r14,SAVprev(r4) /* Queue the new save area in the front */ - stw r13,SAVact(r4) /* Point the savearea at its activation */ + stw r4,ACT_MACT_PCB(r13) ; Point to our savearea + stw r4,ACT_MACT_UPCB(r13) ; Store user savearea + li r0,0 ; Clear this out + stw r14,SAVprev+4(r4) ; Queue the new save area in the front + stw r13,SAVact(r4) ; Point the savearea at its activation #if VERIFYSAVE - bl versave ; (TEST/DEBUG) + bl versave ; (TEST/DEBUG) #endif - mr r30,r4 /* Save pointer to the new context savearea */ - lwz r15,saver1(r4) /* Grab interrupt time stack */ - stw r0,ACT_MACT_KSP(r13) /* Mark stack as busy with 0 val */ - stw r15,FM_BACKPTR(r1) /* Link backwards */ + lwz r15,saver1+4(r4) ; Grab interrupt time stack + mr r30,r4 ; Save pointer to the new context savearea + stw r0,ACT_MACT_KSP(r13) ; Mark stack as busy with 0 val + stw r15,FM_BACKPTR(r1) ; Link stack frame backwards + + lwz r3,SAVtime+4(r30) + addi r4,r13,SYSTEM_TIMER + bl EXT(timer_event) #if DEBUG - /* If debugging, we need two frames, the first being a dummy - * which links back to the trapped routine. The second is - * that which the C routine below will need - */ - lwz r8,savesrr0(r30) /* Get the point of interruption */ - stw r8,FM_LR_SAVE(r1) /* save old instr ptr as LR value */ - stwu r1, -FM_SIZE(r1) /* and make new frame */ +/* If debugging, we need two frames, the first being a dummy + * which links back to the trapped routine. The second is + * that which the C routine below will need + */ + lwz r8,savesrr0+4(r30) ; Get the point of interruption + stw r8,FM_LR_SAVE(r1) ; Save old instr ptr as LR value + stwu r1, -FM_SIZE(r1) ; and make new frame #endif /* DEBUG */ - mfmsr r11 /* Get the MSR */ - lwz r15,SAVflags(r4) /* Get the savearea flags */ - ori r11,r11,lo16(MASK(MSR_EE)) /* Turn on interruption enabled bit */ - lwz r0,saver0(r30) ; Get R0 back - oris r15,r15,SAVsyscall >> 16 /* Mark that it this is a syscall */ - rlwinm r10,r0,0,0,19 ; Keep only the top part - stwu r1,-(FM_SIZE+ARG_SIZE)(r1) /* Make a stack frame */ - cmplwi r10,0x6000 ; Is it the special ppc-only guy? - stw r15,SAVflags(r30) /* Save syscall marker */ - beq- cr6,exitFromVM ; It is time to exit from alternate context... - - beq- ppcscall ; Call the ppc-only system call handler... - - mtmsr r11 /* Enable interruptions */ - - /* Call a function that can print out our syscall info */ - /* Note that we don't care about any volatiles yet */ mr r4,r30 - bl EXT(syscall_trace) - - lwz r0,saver0(r30) /* Get the system call selector */ - mr. r0,r0 /* What kind is it? */ - blt- .L_kernel_syscall /* -ve syscall - go to kernel */ - /* +ve syscall - go to server */ - cmpwi cr0,r0,0x7FFA - beq- .L_notify_interrupt_syscall - -#ifdef MACH_BSD - mr r3,r30 /* Get PCB/savearea */ - lwz r4,saver4(r30) /* Restore r4 */ - lwz r5,saver5(r30) /* Restore r5 */ - lwz r6,saver6(r30) /* Restore r6 */ - lwz r7,saver7(r30) /* Restore r7 */ - lwz r8,saver8(r30) /* Restore r8 */ - lwz r9,saver9(r30) /* Restore r9 */ - lwz r10,saver10(r30) /* Restore r10 */ - bl EXT(unix_syscall) /* Check out unix... */ + + lwz r15,SAVflags(r30) ; Get the savearea flags + lwz r0,saver0+4(r30) ; Get R0 back + mfmsr r11 ; Get the MSR + stwu r1,-(FM_SIZE+ARG_SIZE+MUNGE_ARGS_SIZE)(r1) ; Make a stack frame + ori r11,r11,lo16(MASK(MSR_EE)) ; Turn on interruption enabled bit + rlwinm r10,r0,0,0,19 ; Keep only the top part + oris r15,r15,SAVsyscall >> 16 ; Mark that it this is a syscall + cmplwi r10,0x6000 ; Is it the special ppc-only guy? + stw r15,SAVflags(r30) ; Save syscall marker + beq-- cr6,exitFromVM ; It is time to exit from alternate context... + + beq-- ppcscall ; Call the ppc-only system call handler... + + mr. r0,r0 ; What kind is it? + mtmsr r11 ; Enable interruptions + + blt-- .L_kernel_syscall ; System call number if negative, this is a mach call... + + lwz r8,ACT_TASK(r13) ; Get our task + cmpwi cr0,r0,0x7FFA ; Special blue box call? + beq-- .L_notify_interrupt_syscall ; Yeah, call it... + + lwz r7,TASK_SYSCALLS_UNIX(r8) ; Get the current count + mr r3,r30 ; Get PCB/savearea + mr r4,r13 ; current activation + addi r7,r7,1 ; Bump it + stw r7,TASK_SYSCALLS_UNIX(r8) ; Save it + +#if FPFLOOD + stfd f31,emfp31(r25) ; (TEST/DEBUG) #endif + bl EXT(unix_syscall) ; Check out unix... + .L_call_server_syscall_exception: - li r3,EXC_SYSCALL /* doexception(EXC_SYSCALL, num, 1) */ + li r3,EXC_SYSCALL ; doexception(EXC_SYSCALL, num, 1) .L_call_server_exception: - mr r4,r0 /* Set syscall selector */ + mr r4,r0 ; Set syscall selector li r5,1 - b EXT(doexception) /* Go away, never to return... */ - -/* The above, but with EXC_MACH_SYSCALL */ -.L_call_server_mach_syscall: - li r3,EXC_MACH_SYSCALL - b .L_call_server_exception /* Join the common above... */ + b EXT(doexception) ; Go away, never to return... .L_notify_interrupt_syscall: - lwz r3,saver3(r30) ; Get the new PC address to pass in + lwz r3,saver3+4(r30) ; Get the new PC address to pass in bl EXT(syscall_notify_interrupt) - b .L_syscall_return +/* + * Ok, return from C function, R3 = return value + * + * saved state is still in R30 and the active thread is in R16 . + */ + mr r31,r16 ; Move the current thread pointer + stw r3,saver3+4(r30) ; Stash the return code + b .L_thread_syscall_ret_check_ast ; ; Handle PPC-only system call interface @@ -441,7 +434,7 @@ noassist: cmplwi r15,0x7000 /* Do we have a fast path trap? */ ; and the savearea/pcb as the first parameter. ; It is up to the callee to enable interruptions if ; they should be. We are in a state here where -; both interrupts and preemption is ok, but because we could +; both interrupts and preemption are ok, but because we could ; be calling diagnostic code we will not enable. ; ; Also, the callee is responsible for finding any parameters @@ -455,14 +448,18 @@ noassist: cmplwi r15,0x7000 /* Do we have a fast path trap? */ ; ; Note: the last option is intended for special diagnostics calls that ; want the thread to return and execute before checking for preemption. +; +; NOTE: Both R16 (thread) and R30 (savearea) need to be preserved over this call!!!! ; -ppcscall: rlwinm r11,r0,2,18,29 ; Make an index into the table - lis r10,hi16(EXT(PPCcalls)) ; Get PPC-only system call table - cmplwi r11,PPCcallmax ; See if we are too big - ori r10,r10,lo16(EXT(PPCcalls)) ; Merge in low half + .align 5 + +ppcscall: rlwinm r11,r0,2,18,29 ; Make an index into the table + lis r10,hi16(EXT(PPCcalls)) ; Get PPC-only system call table + cmplwi r11,PPCcallmax ; See if we are too big + ori r10,r10,lo16(EXT(PPCcalls)) ; Merge in low half bgt- .L_call_server_syscall_exception ; Bogus call... - lwzx r11,r10,r11 ; Get function address + lwzx r11,r10,r11 ; Get function address ; ; Note: make sure we do not change the savearea in R30 to @@ -470,250 +467,186 @@ ppcscall: rlwinm r11,r0,2,18,29 ; Make an index into the table ; depend upon it being there. ; - mr r3,r30 ; Pass the savearea - mr r4,r13 ; Pass the activation - mr. r11,r11 ; See if there is a function here - mtlr r11 ; Set the function address + mr r3,r30 ; Pass the savearea + mr r4,r13 ; Pass the activation + mr. r11,r11 ; See if there is a function here + mtctr r11 ; Set the function address beq- .L_call_server_syscall_exception ; Disabled call... - blrl ; Call it - - +#if INSTRUMENT + mfspr r4,pmc1 ; Get stamp + stw r4,0x6100+(9*16)+0x0(0) ; Save it + mfspr r4,pmc2 ; Get stamp + stw r4,0x6100+(9*16)+0x4(0) ; Save it + mfspr r4,pmc3 ; Get stamp + stw r4,0x6100+(9*16)+0x8(0) ; Save it + mfspr r4,pmc4 ; Get stamp + stw r4,0x6100+(9*16)+0xC(0) ; Save it +#endif + bctrl ; Call it + .globl EXT(ppcscret) + LEXT(ppcscret) - mr. r3,r3 ; See what we should do - mr r31,r16 ; Restore the current thread pointer + mr. r3,r3 ; See what we should do + mr r31,r16 ; Restore the current thread pointer bgt+ .L_thread_syscall_ret_check_ast ; Take normal AST checking return.... - mfsprg r10,0 ; Get the per_proc - blt+ .L_thread_syscall_return ; Return, but no ASTs.... - lwz r0,saver0(r30) ; Restore the system call number + mfsprg r10,1 ; Get the current activation + lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block + blt+ .L_thread_syscall_return ; Return, but no ASTs.... + lwz r0,saver0+4(r30) ; Restore the system call number b .L_call_server_syscall_exception ; Go to common exit... + +/* + * we get here for mach system calls + * when kdebug tracing is enabled + */ + +ksystrace: + mr r4,r30 ; Pass in saved state + bl EXT(syscall_trace) + + cmplw r31,r29 ; Is this syscall in the table? + add r31,r27,r28 ; Point right to the syscall table entry + + bge- .L_call_server_syscall_exception ; The syscall number is invalid + + lwz r0,savesrr1(r30) ; Get the saved srr1 + rlwinm. r0,r0,0,MSR_SF_BIT,MSR_SF_BIT ; Test for 64 bit caller + lwz r0,MACH_TRAP_ARG_MUNGE32(r31) ; Pick up the 32 bit munge function address + beq-- .L_ksystrace_munge + lwz r0,MACH_TRAP_ARG_MUNGE64(r31) ; Pick up the 64 bit munge function address + +.L_ksystrace_munge: + cmplwi r0,0 ; do we have a munger to call? + mtctr r0 ; Set the function call address + addi r3,r30,saver3 ; Pointer to args from save area + addi r4,r1,FM_ARG0+ARG_SIZE ; Pointer for munged args + beq-- .L_ksystrace_trapcall ; just make the trap call + bctrl ; Call the munge function + +.L_ksystrace_trapcall: + lwz r0,MACH_TRAP_FUNCTION(r31) ; Pick up the function address + mtctr r0 ; Set the function call address + addi r3,r1,FM_ARG0+ARG_SIZE ; Pointer to munged args + bctrl + + mr r4,r30 ; Pass in the savearea + bl EXT(syscall_trace_end) ; Trace the exit of the system call + b .L_mach_return + + + /* Once here, we know that the syscall was -ve * we should still have r1=ksp, * r16 = pointer to current thread, * r13 = pointer to top activation, * r0 = syscall number * r30 = pointer to saved state (in pcb) - */ -.L_kernel_syscall: - neg r31, r0 /* Make number +ve and put in r31*/ - - /* If out of range, call server with syscall exception */ - addis r29, 0, HIGH_CADDR(EXT(mach_trap_count)) - addi r29, r29, LOW_ADDR(EXT(mach_trap_count)) - lwz r29, 0(r29) - - cmp cr0, r31, r29 - bge- cr0, .L_call_server_syscall_exception - - addis r29, 0, HIGH_CADDR(EXT(mach_trap_table)) - addi r29, r29, LOW_ADDR(EXT(mach_trap_table)) - - /* multiply the trap number to get offset into table */ - slwi r31, r31, MACH_TRAP_OFFSET_POW2 - - /* r31 now holds offset into table of our trap entry, - * add on the table base, and it then holds pointer to entry - */ - add r31, r31, r29 - - /* If the function is kern_invalid, prepare to send an exception. - This is messy, but parallels the x86. We need it for task_by_pid, - at least. */ - lis r29, HIGH_CADDR(EXT(kern_invalid)) - addi r29, r29, LOW_ADDR(EXT(kern_invalid)) - lwz r0, MACH_TRAP_FUNCTION(r31) - cmp cr0, r0, r29 - beq- .L_call_server_syscall_exception - - /* get arg count. If argc > 8 then not all args were in regs, - * so we must perform copyin. - */ - lwz r29, MACH_TRAP_ARGC(r31) - cmpwi cr0, r29, 8 - ble+ .L_syscall_got_args - -/* argc > 8 - perform a copyin */ -/* if the syscall came from kernel space, we can just copy */ - - lwz r0,savesrr1(r30) /* Pick up exception time MSR */ - andi. r0,r0,MASK(MSR_PR) /* Check the priv bit */ - bne+ .L_syscall_arg_copyin /* We're not priviliged... */ - -/* we came from a privilaged task, just do a copy */ -/* get user's stack pointer */ - - lwz r28,saver1(r30) /* Get the stack pointer */ - - subi r29,r29,8 /* Get the number of arguments to copy */ - - addi r28,r28,COPYIN_ARG0_OFFSET-4 /* Point to source - 4 */ - addi r27,r1,FM_ARG0-4 /* Point to sink - 4 */ - -.L_syscall_copy_word_loop: - addic. r29,r29,-1 /* Count down the number of arguments left */ - lwz r0,4(r28) /* Pick up the argument from the stack */ - addi r28,r28,4 /* Point to the next source */ - stw r0,4(r27) /* Store the argument */ - addi r27,r27,4 /* Point to the next sink */ - bne+ .L_syscall_copy_word_loop /* Move all arguments... */ - b .L_syscall_got_args /* Go call it now... */ - - -/* we came from a user task, pay the price of a real copyin */ -/* set recovery point */ - -.L_syscall_arg_copyin: - lwz r8,ACT_VMMAP(r13) ; Get the vm_map for this activation - lis r28,hi16(.L_syscall_copyin_recover) - lwz r8,VMMAP_PMAP(r8) ; Get the pmap - ori r28,r28,lo16(.L_syscall_copyin_recover) - addi r8,r8,PMAP_SEGS ; Point to the pmap SR slots - stw r28,THREAD_RECOVER(r16) /* R16 still holds thread ptr */ - -/* We can manipulate the COPYIN segment register quite easily - * here, but we've also got to make sure we don't go over a - * segment boundary - hence some mess. - * Registers from 12-29 are free for our use. - */ - - - lwz r28,saver1(r30) /* Get the stack pointer */ - subi r29,r29,8 /* Get the number of arguments to copy */ - addi r28,r28,COPYIN_ARG0_OFFSET /* Set source in user land */ - -/* set up SR_COPYIN to allow us to copy, we may need to loop - * around if we change segments. We know that this previously - * pointed to user space, so the sid doesn't need setting. */ - rlwinm r7,r28,6,26,29 ; Get index to the segment slot + .align 5 -.L_syscall_copyin_seg_loop: - - - lwzx r10,r8,r7 ; Get the source SR value - rlwinm r26,r28,0,4,31 ; Clear the segment number from source address - mtsr SR_COPYIN,r10 ; Set the copyin SR - isync - - oris r26,r26,(SR_COPYIN_NUM << (28-16)) ; Insert the copyin segment number into source address - -/* Make r27 point to address-4 of where we will store copied args */ - addi r27,r1,FM_ARG0-4 - -.L_syscall_copyin_word_loop: - - lwz r0,0(r26) /* MAY CAUSE PAGE FAULT! */ - subi r29,r29,1 ; Decrement count - addi r26,r26,4 ; Bump input - stw r0,4(r27) ; Save the copied in word - mr. r29,r29 ; Are they all moved? - addi r27,r27,4 ; Bump output - beq+ .L_syscall_copyin_done ; Escape if we are done... - - rlwinm. r0,r26,0,4,29 ; Did we just step into a new segment? - addi r28,r28,4 ; Bump up user state address also - bne+ .L_syscall_copyin_word_loop ; We are still on the same segment... - - addi r7,r7,4 ; Bump to next slot - b .L_syscall_copyin_seg_loop /* On new segment! remap */ - -/* Don't bother restoring SR_COPYIN, we can leave it trashed */ -/* clear thread recovery as we're done touching user data */ - -.L_syscall_copyin_done: - li r0,0 - stw r0,THREAD_RECOVER(r16) /* R16 still holds thread ptr */ - -.L_syscall_got_args: - lwz r8,ACT_TASK(r13) /* Get our task */ - lis r10,hi16(EXT(c_syscalls_mach)) /* Get top half of counter address */ - lwz r7,TASK_SYSCALLS_MACH(r8) ; Get the current count - lwz r3,saver3(r30) /* Restore r3 */ - addi r7,r7,1 ; Bump it - ori r10,r10,lo16(EXT(c_syscalls_mach)) /* Get low half of counter address */ - stw r7,TASK_SYSCALLS_MACH(r8) ; Save it - lwz r4,saver4(r30) /* Restore r4 */ - lwz r9,0(r10) /* Get counter */ - lwz r5,saver5(r30) /* Restore r5 */ - lwz r6,saver6(r30) /* Restore r6 */ - addi r9,r9,1 /* Add 1 */ - lwz r7,saver7(r30) /* Restore r7 */ - lwz r8,saver8(r30) /* Restore r8 */ - stw r9,0(r10) /* Save it back */ - lwz r9,saver9(r30) /* Restore r9 */ - lwz r10,saver10(r30) /* Restore r10 */ - - lwz r0,MACH_TRAP_FUNCTION(r31) - -/* calling this function, all the callee-saved registers are - * still valid except for r30 and r31 which are in the PCB - * r30 holds pointer to saved state (ie. pcb) - * r31 is scrap - */ - mtctr r0 - bctrl /* perform the actual syscall */ +.L_kernel_syscall: +; +; Call a function that can print out our syscall info +; Note that we don t care about any volatiles yet +; + lwz r10,ACT_TASK(r13) ; Get our task + lwz r0,saver0+4(r30) + lis r8,hi16(EXT(kdebug_enable)) ; Get top of kdebug_enable + lis r28,hi16(EXT(mach_trap_table)) ; Get address of table + ori r8,r8,lo16(EXT(kdebug_enable)) ; Get bottom of kdebug_enable + lwz r8,0(r8) ; Get kdebug_enable + + lwz r7,TASK_SYSCALLS_MACH(r10) ; Get the current count + neg r31,r0 ; Make this positive + mr r3,r31 ; save it + slwi r27,r3,4 ; multiply by 16 + slwi r3,r3,2 ; and the original by 4 + ori r28,r28,lo16(EXT(mach_trap_table)) ; Get address of table + add r27,r27,r3 ; for a total of 20x (5 words/entry) + addi r7,r7,1 ; Bump TASK_SYSCALLS_MACH count + cmplwi r8,0 ; Is kdebug_enable non-zero + stw r7,TASK_SYSCALLS_MACH(r10) ; Save count + bne-- ksystrace ; yes, tracing enabled + + cmplwi r31,MACH_TRAP_TABLE_COUNT ; Is this syscall in the table? + add r31,r27,r28 ; Point right to the syscall table entry + + bge-- .L_call_server_syscall_exception ; The syscall number is invalid + + lwz r0,savesrr1(r30) ; Get the saved srr1 + rlwinm. r0,r0,0,MSR_SF_BIT,MSR_SF_BIT ; Test for 64 bit caller + lwz r0,MACH_TRAP_ARG_MUNGE32(r31) ; Pick up the 32 bit munge function address + beq-- .L_kernel_syscall_munge + lwz r0,MACH_TRAP_ARG_MUNGE64(r31) ; Pick up the 64 bit munge function address + +.L_kernel_syscall_munge: + cmplwi r0,0 ; test for null munger + mtctr r0 ; Set the function call address + addi r3,r30,saver3 ; Pointer to args from save area + addi r4,r1,FM_ARG0+ARG_SIZE ; Pointer for munged args + beq-- .L_kernel_syscall_trapcall ; null munger - skip to trap call + bctrl ; Call the munge function + +.L_kernel_syscall_trapcall: + lwz r0,MACH_TRAP_FUNCTION(r31) ; Pick up the function address + mtctr r0 ; Set the function call address + addi r3,r1,FM_ARG0+ARG_SIZE ; Pointer to munged args + +#if FPFLOOD + stfd f31,emfp31(r25) ; (TEST/DEBUG) +#endif -/* 'standard' syscall returns here - INTERRUPTS ARE STILL ON */ + bctrl -/* r3 contains value that we're going to return to the user - */ /* - * Ok, return from C function, ARG0 = return value + * Ok, return from C function, R3 = return value * * get the active thread's PCB pointer and thus pointer to user state - * saved state is still in R30 and the active thread is in R16 . + * saved state is still in R30 and the active thread is in R16 */ -/* Store return value into saved state structure, since - * we need to pick up the value from here later - the - * syscall may perform a thread_set_syscall_return +.L_mach_return: + srawi r0,r3,31 ; properly extend the return code + cmpi cr0,r3,KERN_INVALID_ARGUMENT ; deal with invalid system calls + mr r31,r16 ; Move the current thread pointer + stw r0, saver3(r30) ; stash the high part of the return code + stw r3,saver3+4(r30) ; Stash the low part of the return code + beq-- cr0,.L_mach_invalid_ret ; otherwise fall through into the normal return path +.L_mach_invalid_arg: + + +/* 'standard' syscall returns here - INTERRUPTS ARE STILL ON + * the syscall may perform a thread_set_syscall_return * followed by a thread_exception_return, ending up * at thread_syscall_return below, with SS_R3 having * been set up already - */ - -/* When we are here, r16 should point to the current thread, + * + * When we are here, r31 should point to the current thread, * r30 should point to the current pcb + * r3 contains value that we're going to return to the user + * which has already been stored back into the save area */ - -/* save off return value, we must load it - * back anyway for thread_exception_return - * TODO NMGS put in register? - */ -.L_syscall_return: - mr r31,r16 /* Move the current thread pointer */ - stw r3,saver3(r30) /* Stash the return code */ - - /* Call a function that records the end of */ - /* the mach system call */ - mr r4,r30 - bl EXT(syscall_trace_end) - -#if 0 - lis r0,HIGH_ADDR(CutTrace) /* (TEST/DEBUG) */ - mr r4,r31 /* (TEST/DEBUG) */ - oris r0,r0,LOW_ADDR(CutTrace) /* (TEST/DEBUG) */ - mr r5,r30 /* (TEST/DEBUG) */ - sc /* (TEST/DEBUG) */ -#endif - + .L_thread_syscall_ret_check_ast: - mfmsr r12 /* Get the current MSR */ - rlwinm r12,r12,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Turn off interruptions enable bit */ - mtmsr r12 /* Turn interruptions off */ + lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable + mfmsr r12 ; Get the current MSR + ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE + andc r12,r12,r10 ; Turn off VEC, FP, and EE + mtmsr r12 ; Turn interruptions off - mfsprg r10,0 /* Get the per_processor block */ + mfsprg r10,1 ; Get the current activation + lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block /* Check to see if there's an outstanding AST */ - lwz r4,PP_NEED_AST(r10) - lwz r4,0(r4) - cmpi cr0,r4, 0 - beq cr0,.L_syscall_no_ast + lwz r4,PP_PENDING_AST(r10) + cmpi cr0,r4, 0 ; Any pending asts? + beq++ cr0,.L_syscall_no_ast ; Nope... /* Yes there is, call ast_taken * pretending that the user thread took an AST exception here, @@ -722,25 +655,48 @@ LEXT(ppcscret) #if DEBUG /* debug assert - make sure that we're not returning to kernel */ - lwz r3,savesrr1(r30) + lwz r3,savesrr1+4(r30) andi. r3,r3,MASK(MSR_PR) - bne+ 0f /* returning to user level, check */ + bne++ scrnotkern ; returning to user level, check - lis r0,hi16(Choke) ; Choke code - ori r0,r0,lo16(Choke) ; and the rest - li r3,failContext ; Bad state code - sc ; System ABEND - + lis r0,hi16(Choke) ; Choke code + ori r0,r0,lo16(Choke) ; and the rest + li r3,failContext ; Bad state code + sc ; System ABEND -0: +scrnotkern: #endif /* DEBUG */ - li r3, AST_ALL - li r4, 1 - bl EXT(ast_taken) - - b .L_thread_syscall_ret_check_ast + lis r3,hi16(AST_ALL) ; Set ast flags + li r4,1 ; Set interrupt allowed + ori r3,r3,lo16(AST_ALL) + bl EXT(ast_taken) ; Process the pending ast + b .L_thread_syscall_ret_check_ast ; Go see if there was another... +.L_mach_invalid_ret: +/* + * need to figure out why we got an KERN_INVALID_ARG + * if it was due to a non-existent system call + * then we want to throw an exception... otherwise + * we want to pass the error code back to the caller + */ + lwz r0,saver0+4(r30) ; reload the original syscall number + neg r28,r0 ; Make this positive + mr r4,r28 ; save a copy + slwi r27,r4,4 ; multiply by 16 + slwi r4,r4,2 ; and another 4 + lis r28,hi16(EXT(mach_trap_table)) ; Get address of table + add r27,r27,r4 ; for a total of 20x (5 words/entry) + ori r28,r28,lo16(EXT(mach_trap_table)) ; Get address of table + add r28,r27,r28 ; Point right to the syscall table entry + lwz r27,MACH_TRAP_FUNCTION(r28) ; Pick up the function address + lis r28,hi16(EXT(kern_invalid)) ; Get high half of invalid syscall function + ori r28,r28,lo16(EXT(kern_invalid)) ; Get low half of invalid syscall function + cmpw cr0,r27,r28 ; Check if this is an invalid system call + beq-- .L_call_server_syscall_exception ; We have a bad system call + b .L_mach_invalid_arg ; a system call returned KERN_INVALID_ARG + + /* thread_exception_return returns to here, almost all * registers intact. It expects a full context restore * of what it hasn't restored itself (ie. what we use). @@ -750,44 +706,24 @@ LEXT(ppcscret) * r30 points to the current pcb */ + .align 5 + .L_syscall_no_ast: .L_thread_syscall_return: - mr r3,r30 ; Get savearea to the correct register for common exit - lwz r8,THREAD_TOP_ACT(r31) /* Now find the current activation */ - - lwz r11,SAVflags(r30) /* Get the flags */ - lwz r5,THREAD_KERNEL_STACK(r31) /* Get the base pointer to the stack */ - rlwinm r11,r11,0,15,13 /* Clear the syscall flag */ - lwz r4,SAVprev(r30) ; Get the previous save area - stw r11,SAVflags(r30) /* Stick back the flags */ - addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE /* Reset to empty */ - stw r4,ACT_MACT_PCB(r8) ; Save previous save area - stw r5,ACT_MACT_KSP(r8) /* Save the empty stack pointer */ - - b chkfac ; Go end it all... - + mr r3,r30 ; Get savearea to the correct register for common exit -.L_syscall_copyin_recover: + lwz r11,SAVflags(r30) ; Get the flags + lwz r5,THREAD_KERNEL_STACK(r31) ; Get the base pointer to the stack + lwz r4,SAVprev+4(r30) ; Get the previous save area + rlwinm r11,r11,0,15,13 ; Clear the syscall flag + mfsprg r8,1 ; Now find the current activation + addi r5,r5,KERNEL_STACK_SIZE-FM_SIZE ; Reset to empty + stw r11,SAVflags(r30) ; Stick back the flags + stw r5,ACT_MACT_KSP(r8) ; Save the empty stack pointer + stw r4,ACT_MACT_PCB(r8) ; Save previous save area + b chkfac ; Go end it all... - /* This is the catcher for any data faults in the copyin - * of arguments from the user's stack. - * r30 still holds a pointer to the PCB - * - * call syscall_error(EXC_BAD_ACCESS, EXC_PPC_VM_PROT_READ, sp, ssp), - * - * we already had a frame so we can do this - */ - - li r3,EXC_BAD_ACCESS - li r4,EXC_PPC_VM_PROT_READ - lwz r5,saver1(r30) - mr r6,r30 - - bl EXT(syscall_error) - b .L_syscall_return - - /* * thread_exception_return() * @@ -802,10 +738,11 @@ LEXT(thread_bootstrap_return) ; NOTE: THIS IS GOING AWAY IN A FEW DAYS.... LEXT(thread_exception_return) ; Directly return to user mode .L_thread_exc_ret_check_ast: - - mfmsr r3 /* Get the MSR */ - rlwinm r3,r3,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Clear EE */ - mtmsr r3 /* Disable interrupts */ + lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable + mfmsr r3 ; Get the MSR + ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE + andc r3,r3,r10 ; Turn off VEC, FP, and EE + mtmsr r3 ; Disable interrupts /* Check to see if there's an outstanding AST */ /* We don't bother establishing a call frame even though CHECK_AST @@ -813,84 +750,89 @@ LEXT(thread_exception_return) ; Directly return to user mode frame, given that we're not going to return. */ - mfsprg r10,0 /* Get the per_processor block */ - lwz r4,PP_NEED_AST(r10) - lwz r4,0(r4) + mfsprg r10,1 ; Get the current activation + lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block + lwz r4,PP_PENDING_AST(r10) cmpi cr0,r4, 0 - beq cr0,.L_exc_ret_no_ast + beq+ cr0,.L_exc_ret_no_ast - /* Yes there is, call ast_taken - * pretending that the user thread took an AST exception here, - * ast_taken will save all state and bring us back here - */ +/* Yes there is, call ast_taken + * pretending that the user thread took an AST exception here, + * ast_taken will save all state and bring us back here + */ - - li r3,AST_ALL + lis r3,hi16(AST_ALL) li r4,1 + ori r3,r3,lo16(AST_ALL) bl EXT(ast_taken) - b .L_thread_exc_ret_check_ast /* check for a second AST (rare)*/ + b .L_thread_exc_ret_check_ast ; check for a second AST (rare) /* arriving here, interrupts should be disabled */ /* Get the active thread's PCB pointer to restore regs */ .L_exc_ret_no_ast: - lwz r31,PP_CPU_DATA(r10) - lwz r31,CPU_ACTIVE_THREAD(r31) - lwz r30,THREAD_TOP_ACT(r31) + mfsprg r30,1 ; Get the currrent activation + mr r31,r30 + lwz r30,ACT_MACT_PCB(r30) - mr. r30,r30 ; Is there any context yet? - beq- makeDummyCtx ; No, hack one up... + mr. r30,r30 ; Is there any context yet? + beq- makeDummyCtx ; No, hack one up... #if DEBUG /* * debug assert - make sure that we're not returning to kernel * get the active thread's PCB pointer and thus pointer to user state */ - lwz r3,savesrr1(r30) + lwz r3,savesrr1+4(r30) andi. r3,r3,MASK(MSR_PR) - bne+ ret_user2 ; We are ok... + bne+ ret_user2 ; We are ok... - lis r0,hi16(Choke) ; Choke code - ori r0,r0,lo16(Choke) ; and the rest - li r3,failContext ; Bad state code - sc ; System ABEND + lis r0,hi16(Choke) ; Choke code + ori r0,r0,lo16(Choke) ; and the rest + li r3,failContext ; Bad state code + sc ; System ABEND ret_user2: #endif /* DEBUG */ -/* If the MSR_SYSCALL_MASK isn't set, then we came from a trap, +/* If the system call flag isn't set, then we came from a trap, * so warp into the return_from_trap (thread_return) routine, * which takes PCB pointer in R3, not in r30! */ - lwz r0,SAVflags(r30) - mr r3,r30 /* Copy pcb pointer into r3 in case */ - andis. r0,r0,SAVsyscall>>16 /* Are we returning from a syscall? */ - beq- cr0,thread_return /* Nope, must be a thread return... */ - b .L_thread_syscall_return + lwz r0,SAVflags(r30) ; Grab the savearea flags + andis. r0,r0,SAVsyscall>>16 ; Are we returning from a syscall? + mr r3,r30 ; Copy pcb pointer into r3 in case we need it + beq-- cr0,thread_return ; Nope, must be a thread return... + b .L_thread_syscall_return ; Join up with the system call return... ; ; This is where we handle someone trying who did a thread_create followed ; by a thread_resume with no intervening thread_set_state. Just make an ; empty context, initialize it to trash and let em execute at 0... +; + + .align 5 makeDummyCtx: - bl EXT(save_get) ; Get a save_area - li r0,0 ; Get a 0 - addi r2,r3,savefp0 ; Point past what we are clearing - mr r4,r3 ; Save the start - -cleardummy: stw r0,0(r4) ; Clear stuff - addi r4,r4,4 ; Next word - cmplw r4,r2 ; Still some more? - blt+ cleardummy ; Yeah... + bl EXT(save_get) ; Get a save_area + li r4,SAVgeneral ; Get the general context type + li r0,0 ; Get a 0 + stb r4,SAVflags+2(r3) ; Set type + addi r2,r3,savefpscr+4 ; Point past what we are clearing + mr r4,r3 ; Save the start + +cleardummy: stw r0,0(r4) ; Clear stuff + addi r4,r4,4 ; Next word + cmplw r4,r2 ; Still some more? + blt+ cleardummy ; Yeah... lis r2,hi16(MSR_EXPORT_MASK_SET) ; Set the high part of the user MSR ori r2,r2,lo16(MSR_EXPORT_MASK_SET) ; And the low part - stw r2,savesrr1(r3) ; Set the default user MSR + stw r2,savesrr1+4(r3) ; Set the default user MSR - b thread_return ; Go let em try to execute, hah! + b thread_return ; Go let em try to execute, hah! /* * ihandler(type) @@ -905,91 +847,93 @@ cleardummy: stw r0,0(r4) ; Clear stuff .align 5 .globl EXT(ihandler) -LEXT(ihandler) /* Interrupt handler */ +LEXT(ihandler) ; Interrupt handler */ /* * get the value of istackptr, if it's zero then we're already on the - * interrupt stack, otherwise it points to a saved_state structure - * at the top of the interrupt stack. + * interrupt stack. */ - lwz r10,savesrr1(r4) /* Get SRR1 */ - mfsprg r25,0 /* Get the per_proc block */ - li r14,0 /* Zero this for now */ - rlwinm. r13,r10,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on? - lwz r16,PP_CPU_DATA(r25) /* Assume we need this */ - crmove cr1_eq,cr0_eq ; Remember vector enablement - lwz r1,PP_ISTACKPTR(r25) /* Get the interrupt stack */ - rlwinm. r10,r10,0,MSR_FP_BIT,MSR_FP_BIT ; Was floating point on? - li r13,0 /* Zero this for now */ - lwz r16,CPU_ACTIVE_THREAD(r16) /* Get the thread pointer */ - - beq+ cr1,ivecoff ; Vector off, do not save vrsave... - lwz r7,savevrsave(r4) ; Get the VRSAVE register - stw r7,liveVRS(r25) ; Set the live value - -ivecoff: li r0,0 /* Get a constant 0 */ - cmplwi cr1,r16,0 /* Are we still booting? */ - beq+ ifpoff ; Skip if floating point is off... - lwz r9,savexfpscr(r4) ; Grab the just saved FPSCR - stw r9,liveFPSCR(r25) ; Make it the live copy - -ifpoff: mr. r1,r1 /* Is it active? */ - beq- cr1,ihboot1 /* We're still coming up... */ - lwz r13,THREAD_TOP_ACT(r16) /* Pick up the active thread */ - lwz r14,ACT_MACT_PCB(r13) /* Now point to the PCB */ - -ihboot1: lwz r9,saver1(r4) /* Pick up the 'rupt time stack */ - stw r14,SAVprev(r4) /* Queue the new save area in the front */ - stw r13,SAVact(r4) /* Point the savearea at its activation */ - beq- cr1,ihboot4 /* We're still coming up... */ - stw r4,ACT_MACT_PCB(r13) /* Point to our savearea */ - -ihboot4: bne .L_istackfree /* Nope... */ + lwz r10,savesrr1+4(r4) ; Get SRR1 + lwz r7,savevrsave(r4) ; Get the VRSAVE register + mfsprg r13,1 ; Get the current activation + lwz r25,ACT_PER_PROC(r13) ; Get the per_proc block + li r14,0 ; Zero this for now + rlwinm. r16,r10,0,MSR_VEC_BIT,MSR_VEC_BIT ; Was vector on? + lwz r1,PP_ISTACKPTR(r25) ; Get the interrupt stack + li r16,0 ; Zero this for now + + beq+ ivecoff ; Vector off, do not save vrsave... + stw r7,liveVRS(r25) ; Set the live value + +ivecoff: li r0,0 ; Get a constant 0 + rlwinm r5,r10,0,MSR_PR_BIT,MSR_PR_BIT ; Are we trapping from supervisor state? + mr. r1,r1 ; Is it active? + cmplwi cr2,r5,0 ; cr2_eq == 1 if yes + mr r16,r13 + lwz r14,ACT_MACT_PCB(r13) ; Now point to the PCB + lwz r9,saver1+4(r4) ; Pick up the rupt time stack + stw r14,SAVprev+4(r4) ; Queue the new save area in the front + stw r13,SAVact(r4) ; Point the savearea at its activation + stw r4,ACT_MACT_PCB(r13) ; Point to our savearea + beq cr2,ifromk + stw r4,ACT_MACT_UPCB(r13) ; Store user savearea + +ifromk: bne .L_istackfree ; Nope... /* We're already on the interrupt stack, get back the old * stack pointer and make room for a frame */ - lwz r10,PP_INTSTACK_TOP_SS(r25) ; Get the top of the interrupt stack - addi r5,r9,INTSTACK_SIZE-FM_SIZE ; Shift stack for bounds check - subi r1,r9,FM_REDZONE ; Back up beyond the red zone - sub r5,r5,r10 ; Get displacement into stack - cmplwi r5,INTSTACK_SIZE-FM_SIZE ; Is the stack actually invalid? - blt+ ihsetback ; The stack is ok... + lwz r10,PP_INTSTACK_TOP_SS(r25) ; Get the top of the interrupt stack + addi r5,r9,INTSTACK_SIZE-FM_SIZE ; Shift stack for bounds check + subi r1,r9,FM_REDZONE ; Back up beyond the red zone + sub r5,r5,r10 ; Get displacement into stack + cmplwi r5,INTSTACK_SIZE-FM_SIZE ; Is the stack actually invalid? + blt+ ihsetback ; The stack is ok... - lwz r5,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top + lwz r5,PP_DEBSTACK_TOP_SS(r25) ; Pick up debug stack top subi r5,r5,KERNEL_STACK_SIZE-FM_SIZE ; Adjust to start of stack - sub r5,r1,r5 ; Get displacement into debug stack + sub r5,r1,r5 ; Get displacement into debug stack cmplwi cr2,r5,KERNEL_STACK_SIZE-FM_SIZE ; Check if we are on debug stack - blt+ ihsetback ; Yeah, that is ok too... + blt+ cr2,ihsetback ; Yeah, that is ok too... - lis r0,hi16(Choke) ; Choke code - ori r0,r0,lo16(Choke) ; and the rest - li r3,failStack ; Bad stack code - sc ; System ABEND + lis r0,hi16(Choke) ; Choke code + ori r0,r0,lo16(Choke) ; and the rest + li r3,failStack ; Bad stack code + sc ; System ABEND + +intUnalignedStk: + lis r0,hi16(Choke) ; Choke code + ori r0,r0,lo16(Choke) ; and the rest + li r3,failUnalignedStk ; Unaligned stack code + sc ; System ABEND .align 5 .L_istackfree: - lwz r10,SAVflags(r4) - stw r0,PP_ISTACKPTR(r25) /* Mark the stack in use */ - oris r10,r10,HIGH_ADDR(SAVrststk) /* Indicate we reset stack when we return from this one */ - stw r10,SAVflags(r4) /* Stick it back */ + rlwinm. r0,r1,0,28,31 ; Check if stack is aligned (and get 0) + lwz r10,SAVflags(r4) ; Get savearea flags + bne-- intUnalignedStk ; Stack is unaligned... + stw r0,PP_ISTACKPTR(r25) ; Mark the stack in use + oris r10,r10,hi16(SAVrststk) ; Indicate we reset stack when we return from this one + stw r10,SAVflags(r4) ; Stick it back - /* - * To summarize, when we reach here, the state has been saved and - * the stack is marked as busy. We now generate a small - * stack frame with backpointers to follow the calling - * conventions. We set up the backpointers to the trapped - * routine allowing us to backtrace. - */ +/* + * To summarize, when we reach here, the state has been saved and + * the stack is marked as busy. We now generate a small + * stack frame with backpointers to follow the calling + * conventions. We set up the backpointers to the trapped + * routine allowing us to backtrace. + */ -ihsetback: subi r1,r1,FM_SIZE /* Make a new frame */ - stw r9,FM_BACKPTR(r1) /* point back to previous stackptr */ +ihsetback: subi r1,r1,FM_SIZE ; Make a new frame + stw r9,FM_BACKPTR(r1) ; Point back to previous stackptr #if VERIFYSAVE - bl versave ; (TEST/DEBUG) + beq- cr1,ihbootnover ; (TEST/DEBUG) + bl versave ; (TEST/DEBUG) +ihbootnover: ; (TEST/DEBUG) #endif #if DEBUG @@ -997,14 +941,27 @@ ihsetback: subi r1,r1,FM_SIZE /* Make a new frame */ * which links back to the trapped routine. The second is * that which the C routine below will need */ - lwz r5,savesrr0(r4) /* Get interrupt address */ - stw r5,FM_LR_SAVE(r1) /* save old instr ptr as LR value */ - stwu r1,-FM_SIZE(r1) /* Make another new frame for C routine */ + lwz r5,savesrr0+4(r4) ; Get interrupt address + stw r5,FM_LR_SAVE(r1) ; save old instr ptr as LR value + stwu r1,-FM_SIZE(r1) ; Make another new frame for C routine #endif /* DEBUG */ - lwz r5,savedsisr(r4) /* Get the DSISR */ - lwz r6,savedar(r4) /* Get the DAR */ + mr r31,r3 + mr r30,r4 + + lwz r3,SAVtime+4(r4) + addi r4,r13,SYSTEM_TIMER + bl EXT(timer_event) + + mr r3,r31 + mr r4,r30 + lwz r5,savedsisr(r30) ; Get the DSISR + lwz r6,savedar+4(r30) ; Get the DAR +#if FPFLOOD + stfd f31,emfp31(r25) ; (TEST/DEBUG) +#endif + bl EXT(interrupt) @@ -1014,28 +971,24 @@ ihsetback: subi r1,r1,FM_SIZE /* Make a new frame */ */ .globl EXT(ihandler_ret) -LEXT(ihandler_ret) /* Marks our return point from debugger entry */ +LEXT(ihandler_ret) ; Marks our return point from debugger entry - mfmsr r0 /* Get our MSR */ - rlwinm r0,r0,0,MSR_EE_BIT+1,MSR_EE_BIT-1 /* Flip off the interrupt enabled bit */ - mtmsr r0 /* Make sure interrupts are disabled */ - mfsprg r10,0 /* Get the per_proc block */ + lis r10,hi16(MASK(MSR_VEC)) ; Get the vector enable + mfmsr r0 ; Get our MSR + ori r10,r10,lo16(MASK(MSR_FP)|MASK(MSR_EE)) ; Add in FP and EE + andc r0,r0,r10 ; Turn off VEC, FP, and EE + mtmsr r0 ; Make sure interrupts are disabled + mfsprg r8,1 ; Get the current activation + lwz r10,ACT_PER_PROC(r8) ; Get the per_proc block - lwz r8,PP_CPU_DATA(r10) /* Get the CPU data area */ - lwz r7,SAVflags(r3) /* Pick up the flags */ - lwz r8,CPU_ACTIVE_THREAD(r8) /* and the active thread */ - lwz r9,SAVprev(r3) /* Get previous save area */ - cmplwi cr1,r8,0 /* Are we still initializing? */ - lwz r12,savesrr1(r3) /* Get the MSR we will load on return */ - beq- cr1,ihboot2 /* Skip if we are still in init... */ - lwz r8,THREAD_TOP_ACT(r8) /* Pick up the active thread */ - -ihboot2: andis. r11,r7,HIGH_ADDR(SAVrststk) /* Is this the first on the stack? */ - beq- cr1,ihboot3 /* Skip if we are still in init... */ - stw r9,ACT_MACT_PCB(r8) /* Point to previous context savearea */ - -ihboot3: mr r4,r3 /* Move the savearea pointer */ - beq .L_no_int_ast2 /* Get going if not the top o' stack... */ + lwz r7,SAVflags(r3) ; Pick up the flags + lwz r9,SAVprev+4(r3) ; Get previous save area + cmplwi cr1,r8,0 ; Are we still initializing? + lwz r12,savesrr1+4(r3) ; Get the MSR we will load on return + andis. r11,r7,hi16(SAVrststk) ; Is this the first on the stack? + stw r9,ACT_MACT_PCB(r8) ; Point to previous context savearea + mr r4,r3 ; Move the savearea pointer + beq .L_no_int_ast2 ; Get going if not the top-o-stack... /* We're the last frame on the stack. Restore istackptr to empty state. @@ -1044,29 +997,25 @@ ihboot3: mr r4,r3 /* Move the savearea pointer */ * returning to user mode * returning to a kloaded server */ - lwz r9,PP_INTSTACK_TOP_SS(r10) /* Get the empty stack value */ - lwz r5,PP_CPU_DATA(r10) /* Get cpu_data ptr */ - andc r7,r7,r11 /* Remove the stack reset bit in case we pass this one */ - stw r9,PP_ISTACKPTR(r10) /* Save that saved state ptr */ - lwz r3,CPU_PREEMPTION_LEVEL(r5) /* Get preemption level */ - stw r7,SAVflags(r4) /* Save the flags */ - cmplwi r3, 0 /* Check for preemption */ - bne .L_no_int_ast /* Don't preempt if level is not zero */ - andi. r6,r12,MASK(MSR_PR) /* privilege mode */ - lwz r11,PP_NEED_AST(r10) /* Get the AST request address */ - lwz r11,0(r11) /* Get the request */ - beq- .L_kernel_int_ast /* In kernel space, AST_URGENT check */ - li r3,T_AST /* Assume the worst */ - mr. r11,r11 /* Are there any pending? */ - beq .L_no_int_ast /* Nope... */ + lwz r9,PP_INTSTACK_TOP_SS(r10) ; Get the empty stack value + andc r7,r7,r11 ; Remove the stack reset bit in case we pass this one + stw r9,PP_ISTACKPTR(r10) ; Save that saved state ptr + lwz r3,ACT_PREEMPT_CNT(r8) ; Get preemption level + stw r7,SAVflags(r4) ; Save the flags + cmplwi r3, 0 ; Check for preemption + bne .L_no_int_ast ; Do not preempt if level is not zero + andi. r6,r12,MASK(MSR_PR) ; privilege mode + lwz r11,PP_PENDING_AST(r10) ; Get the pending AST mask + beq- .L_kernel_int_ast ; In kernel space, AST_URGENT check + li r3,T_AST ; Assume the worst + mr. r11,r11 ; Are there any pending? + beq .L_no_int_ast ; Nope... b .L_call_thandler .L_kernel_int_ast: - andi. r11,r11,AST_URGENT /* AST_URGENT */ - li r3,T_PREEMPT /* Assume the worst */ - beq .L_no_int_ast /* Nope... */ - -.L_call_thandler: + andi. r11,r11,AST_URGENT ; Do we have AST_URGENT? + li r3,T_PREEMPT ; Assume the worst + beq .L_no_int_ast ; Nope... /* * There is a pending AST. Massage things to make it look like @@ -1076,17 +1025,20 @@ ihboot3: mr r4,r3 /* Move the savearea pointer */ * trap instead of performing an rfi. */ - stw r3,saveexception(r4) /* Set the exception code to T_AST/T_PREEMPT */ - b EXT(thandler) /* hyperspace into AST trap */ +.L_call_thandler: + stw r3,saveexception(r4) ; Set the exception code to T_AST/T_PREEMPT + b EXT(thandler) ; We need to preempt so treat like a trap... .L_no_int_ast: - mr r3,r4 ; Get into the right register for common code + mr r3,r4 ; Get into the right register for common code + .L_no_int_ast2: - rlwinm r7,r7,0,15,13 /* Clear the syscall bit */ - li r4,0 ; Assume for a moment that we are in init - stw r7,SAVflags(r3) /* Set the flags */ - beq- cr1,chkfac ; Jump away if we are in init... - lwz r4,ACT_MACT_PCB(r8) ; Get the new level marker + rlwinm r7,r7,0,15,13 ; Clear the syscall flag + li r4,0 ; Assume for a moment that we are in init + stw r7,SAVflags(r3) ; Set the flags with cleared syscall flag + beq-- cr1,chkfac ; Jump away if we are in init... + + lwz r4,ACT_MACT_PCB(r8) ; Get the new level marker ; @@ -1101,386 +1053,435 @@ ihboot3: mr r4,r3 /* Move the savearea pointer */ ; R8 = pointer to activation ; R10 = per_proc block ; -chkfac: mr. r8,r8 ; Are we still in boot? - beq- chkenax ; Yeah, skip it all... - - lwz r20,ACT_MACT_FPUlvl(r8) ; Get the FPU level - lwz r12,savesrr1(r3) ; Get the current MSR - cmplw cr1,r20,r3 ; Are we returning from the active level? - lwz r23,PP_FPU_THREAD(r10) ; Get floating point owner - rlwinm r12,r12,0,MSR_FP_BIT+1,MSR_FP_BIT-1 ; Turn off floating point for now - cmplw cr2,r23,r8 ; Are we the facility owner? - lhz r26,PP_CPU_NUMBER(r10) ; Get the current CPU number - cror cr0_eq,cr1_eq,cr2_eq ; Check if returning from active or we own facility - bne- cr0,chkvecnr ; Nothing to do if not returning from active or not us... - -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3301 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif +; Note that barring unforseen crashes, there is no escape from this point +; on. We WILL call exception_exit and launch this context. No worries +; about preemption or interruptions here. +; +; Note that we will set up R26 with whatever context we will be launching, +; so it will indicate the current, or the deferred it it is set and we +; are going to user state. CR2_eq will be set to indicate deferred. +; - li r22,ACT_MACT_FPUcpu ; Point to the CPU indication/lock word - -cfSpin2: lwarx r27,r22,r8 ; Get and reserve the last used CPU - mr. r27,r27 ; Is it changing now? - oris r0,r27,hi16(fvChk) ; Set the "changing" flag - blt- cfSpin2 ; Spin if changing - stwcx. r0,r22,r8 ; Lock it up - bne- cfSpin2 ; Someone is messing right now - - isync ; Make sure we see everything - - cmplw r4,r20 ; Are we going to be in the right level? - beq- cr1,chkfpfree ; Leaving active level, can not possibly enable... - cmplw cr1,r27,r26 ; Are we on the right CPU? - li r0,0 ; Get a constant 0 - beq+ cr1,chkfpnlvl ; Right CPU... - - stw r0,PP_FPU_THREAD(r10) ; Show facility unowned so we do not get back here - b chkvec ; Go check out the vector facility... - -chkfpnlvl: bne- chkvec ; Different level, can not enable... - lwz r24,ACT_MACT_FPU(r8) ; Get the floating point save area - ori r12,r12,lo16(MASK(MSR_FP)) ; Enable facility - mr. r24,r24 ; Does the savearea exist? - li r0,1 ; Get set to invalidate - beq- chkvec ; Nothing to invalidate... - lwz r25,SAVlvlfp(r24) ; Get the level of top savearea - cmplw r4,r25 ; Is the top one ours? - bne+ chkvec ; Not ours... - stw r0,SAVlvlfp(r24) ; Invalidate the first one - b chkvec ; Go check out the vector facility... - -chkfpfree: li r0,0 ; Clear a register - lwz r24,ACT_MACT_FPU(r8) ; Get the floating point save area - - bne- cr2,chkfpnfr ; Not our facility, do not clear... - stw r0,PP_FPU_THREAD(r10) ; Clear floating point owner -chkfpnfr: +chkfac: lwz r29,savesrr1+4(r3) ; Get the current MSR + mr. r28,r8 ; Are we still in boot? + mr r31,r10 ; Move per_proc address + mr r30,r4 ; Preserve new level + mr r27,r3 ; Save the old level + beq-- chkenax ; Yeah, skip it all... + + rlwinm. r0,r29,0,MSR_PR_BIT,MSR_PR_BIT ; Are we going into user state? + + lwz r20,curctx(r28) ; Get our current context + lwz r26,deferctx(r28) ; Get any deferred context switch + li r0,1 ; Get set to hold off quickfret + rlwinm r29,r29,0,MSR_FP_BIT+1,MSR_FP_BIT-1 ; Turn off floating point for now + lwz r21,FPUlevel(r20) ; Get the facility level + cmplwi cr2,r26,0 ; Are we going into a deferred context later? + rlwinm r29,r29,0,MSR_VEC_BIT+1,MSR_VEC_BIT-1 ; Turn off vector for now + crnor cr2_eq,cr0_eq,cr2_eq ; Set cr2_eq if going to user state and there is deferred + lhz r19,PP_CPU_NUMBER(r31) ; Get our CPU number + cmplw r27,r21 ; Are we returning from the active level? + stw r0,holdQFret(r31) ; Make sure we hold off releasing quickfret + bne++ fpuchkena ; Nope... -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3302 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif +; +; First clean up any live context we are returning from +; - mr. r24,r24 ; Do we even have a savearea? - beq+ chkvec ; Nope... + lwz r22,FPUcpu(r20) ; Get CPU this context was last dispatched on -#if FPVECDBG - rlwinm. r0,r24,0,0,15 ; (TEST/DEBUG) - bne+ notbadxxx1 ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) -notbadxxx1: ; (TEST/DEBUG) - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3303 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - - lwz r25,SAVlvlfp(r24) ; Get the level of top savearea - cmplwi r25,1 ; Is the top area invalid? - cmplw cr1,r25,r3 ; Is it for the returned from context? - beq fptoss ; It is invalid... - bne cr1,chkvec ; Not for the returned context... + stw r19,FPUcpu(r20) ; Claim context for us -fptoss: lwz r25,SAVprefp(r24) ; Get previous savearea -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3304 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - mr r5,r25 ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - mr. r25,r25 ; Is there one? - stw r25,ACT_MACT_FPU(r8) ; Set the new pointer - beq fptoplvl ; Nope, we are at the top... -#if FPVECDBG - rlwinm. r0,r25,0,0,15 ; (TEST/DEBUG) - bne+ notbadxxx2 ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) -notbadxxx2: ; (TEST/DEBUG) -#endif - lwz r25,SAVlvlfp(r25) ; Get the new level - -fptoplvl: lwz r19,SAVflags(r24) ; Get the savearea flags -#if FPVECDBG - rlwinm. r0,r19,0,1,1 ; (TEST/DEBUG) - bne+ donotdie3 ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) -donotdie3: ; (TEST/DEBUG) + eieio ; Make sure this gets out before owner clear + +#if ppeSize != 16 +#error per_proc_entry is not 16bytes in size #endif + + lis r23,hi16(EXT(PerProcTable)) ; Set base PerProcTable + slwi r22,r22,4 ; Find offset to the owner per_proc_entry + ori r23,r23,lo16(EXT(PerProcTable)) ; Set base PerProcTable + li r24,FPUowner ; Displacement to float owner + add r22,r23,r22 ; Point to the owner per_proc_entry + lwz r22,ppe_vaddr(r22) ; Point to the owner per_proc + +fpuinvothr: lwarx r23,r24,r22 ; Get the owner -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3305 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - rlwinm r22,r24,0,0,19 ; Round down to the base savearea block - rlwinm r19,r19,0,2,0 ; Remove the floating point in use flag - stw r25,ACT_MACT_FPUlvl(r8) ; Set the new top level - andis. r0,r19,hi16(SAVinuse) ; Still in use? - stw r19,SAVflags(r24) ; Set the savearea flags - bne- invlivefp ; Go invalidate live FP -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3306 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif -#if FPVECDBG - rlwinm. r0,r24,0,0,15 ; (TEST/DEBUG) - bne+ notbadxxx3 ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) -notbadxxx3: ; (TEST/DEBUG) -#endif - lwz r23,SACvrswap(r22) ; Get the conversion from virtual to real - lwz r20,PP_QUICKFRET(r10) ; Get the old quick fret head - xor r23,r24,r23 ; Convert to physical - stw r20,SAVqfret(r24) ; Back chain the quick release queue - stw r23,PP_QUICKFRET(r10) ; Anchor it - -invlivefp: lis r20,hi16(EXT(real_ncpus)) ; Get number of CPUs - lis r23,hi16(EXT(per_proc_info)) ; Set base per_proc - ori r20,r20,lo16(EXT(real_ncpus)) ; Other half of number of CPUs - li r25,PP_FPU_THREAD ; Point to the FP owner address - lwz r20,0(r20) ; Get number of processors active - ori r23,r23,lo16(EXT(per_proc_info)) ; Set base per_proc - li r2,0 ; Get something clear - -invlivefl: cmplw r23,r10 ; We can skip our processor - addi r20,r20,-1 ; Count remaining processors - beq invlivefn ; Skip ourselves... - -invlivefa: lwarx r0,r25,r23 ; Get FP owner for this processor - cmplw r0,r8 ; Do we own it? - bne invlivefn ; Nope... - stwcx. r2,r25,r23 ; Show not live - bne- invlivefa ; Someone else did this, try again... - -invlivefn: mr. r20,r20 ; Have we finished? - addi r23,r23,ppSize ; Bump to next - bgt invlivefl ; Make sure we do all processors... + sub r0,r23,r20 ; Subtract one from the other + sub r21,r20,r23 ; Subtract the other from the one + or r21,r21,r0 ; Combine them + srawi r21,r21,31 ; Get a 0 if equal or -1 of not + and r23,r23,r21 ; Make 0 if same, unchanged if not + stwcx. r23,r24,r22 ; Try to invalidate it + bne-- fpuinvothr ; Try again if there was a collision... + isync ; -; Check out vector stuff (and translate savearea to physical for exit) +; Now if there is a savearea associated with the popped context, release it. +; Either way, pop the level to the top stacked context. ; -chkvec: sync ; Make sure all is saved - stw r27,ACT_MACT_FPUcpu(r8) ; Set the active CPU and release + + lwz r22,FPUsave(r20) ; Get pointer to the first savearea + li r21,0 ; Assume we popped all the way out + mr. r22,r22 ; Is there anything there? + beq++ fpusetlvl ; No, see if we need to enable... + + lwz r21,SAVlevel(r22) ; Get the level of that savearea + cmplw r21,r27 ; Is this the saved copy of the live stuff? + bne fpusetlvl ; No, leave as is... -chkvecnr: lwz r20,ACT_MACT_VMXlvl(r8) ; Get the vector level - lwz r23,PP_VMX_THREAD(r10) ; Get vector owner - cmplw cr1,r20,r3 ; Are we returning from the active level? - cmplw cr2,r23,r8 ; Are we the facility owner? - rlwinm r12,r12,0,MSR_VEC_BIT+1,MSR_VEC_BIT-1 ; Turn off vector for now - cror cr0_eq,cr1_eq,cr2_eq ; Check if returning from active or we own facility - bne- cr0,setenanr ; Not our facility, nothing to do here... + lwz r24,SAVprev+4(r22) ; Pick up the previous area + li r21,0 ; Assume we popped all the way out + mr. r24,r24 ; Any more context stacked? + beq-- fpuonlyone ; Nope... + lwz r21,SAVlevel(r24) ; Get the level associated with save -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3401 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif +fpuonlyone: stw r24,FPUsave(r20) ; Dequeue this savearea - li r22,ACT_MACT_VMXcpu ; Point to the CPU indication/lock word - -cvSpin2: lwarx r27,r22,r8 ; Get and reserve the last used CPU - mr. r27,r27 ; Is it changing now? - oris r0,r27,hi16(fvChk) ; Set the "changing" flag - blt- cvSpin2 ; Spin if changing - stwcx. r0,r22,r8 ; Lock it up - bne- cvSpin2 ; Someone is messing right now - - isync ; Make sure we see everything - - cmplw r4,r20 ; Are we going to be in the right level? - beq- cr1,chkvecfree ; Leaving active level, can not possibly enable... - cmplw cr1,r27,r26 ; Are we on the right CPU? - li r0,0 ; Get a constant 0 - beq+ cr1,chkvecnlvl ; Right CPU... - - stw r0,PP_VMX_THREAD(r10) ; Show facility unowned so we do not get back here - b setena ; Go actually exit... - -chkvecnlvl: bne- setena ; Different level, can not enable... - lwz r24,ACT_MACT_VMX(r8) ; Get the vector save area - oris r12,r12,hi16(MASK(MSR_VEC)) ; Enable facility - mr. r24,r24 ; Does the savearea exist? - li r0,1 ; Get set to invalidate - beq- setena ; Nothing to invalidate... - lwz r25,SAVlvlvec(r24) ; Get the level of top savearea - cmplw r4,r25 ; Is the top one ours? - bne+ setena ; Not ours... - stw r0,SAVlvlvec(r24) ; Invalidate the first one - b setena ; Actually exit... - -chkvecfree: li r0,0 ; Clear a register - lwz r24,ACT_MACT_VMX(r8) ; Get the vector save area - - bne- cr2,chkvecnfr ; Not our facility, do not clear... - stw r0,PP_VMX_THREAD(r10) ; Clear vector owner -chkvecnfr: + rlwinm r3,r22,0,0,19 ; Find main savearea header + + lwz r8,quickfret(r31) ; Get the first in quickfret list (top) + lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom) + lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top) + lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom) + stw r8,SAVprev(r22) ; Link the old in (top) + stw r9,SAVprev+4(r22) ; Link the old in (bottom) + xor r3,r22,r3 ; Convert to physical + stw r2,quickfret(r31) ; Set the first in quickfret list (top) + stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom) #if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3402 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif + lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) + li r2,0x3301 ; (TEST/DEBUG) + oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) + sc ; (TEST/DEBUG) +#endif - mr. r24,r24 ; Do we even have a savearea? - beq+ setena ; Nope... +fpusetlvl: stw r21,FPUlevel(r20) ; Save the level + +; +; Here we check if we are at the right level +; We need to check the level we are entering, not the one we are exiting. +; Therefore, we will use the defer level if it is non-zero and we are +; going into user state. +; -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3403 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - lwz r25,SAVlvlvec(r24) ; Get the level - cmplwi r25,1 ; Is the top area invalid? - cmplw cr1,r25,r3 ; Is it for the returned from context? - beq vectoss ; It is invalid... - bne cr1,setena ; Not for the returned context... +fpuchkena: bt-- cr2_eq,fpuhasdfrd ; Skip if deferred, R26 already set up... + mr r26,r20 ; Use the non-deferred value -vectoss: lwz r25,SAVprevec(r24) ; Get previous savearea +fpuhasdfrd: +#if 0 + rlwinm. r0,r29,0,MSR_PR_BIT,MSR_PR_BIT ; (TEST/DEBUG) Going into user state? + beq fpunusrstt ; (TEST/DEBUG) Nope... + lwz r23,FPUlevel(r26) ; (TEST/DEBUG) Get the level ID + lwz r24,FPUsave(r26) ; (TEST/DEBUG) Get the first savearea + mr. r23,r23 ; (TEST/DEBUG) Should be level 0 + beq++ fpulvl0 ; (TEST/DEBUG) Yes... + + lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code + ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest + sc ; (TEST/DEBUG) System ABEND + +fpulvl0: mr. r24,r24 ; (TEST/DEBUG) Any context? + beq fpunusrstt ; (TEST/DEBUG) No... + lwz r23,SAVlevel(r24) ; (TEST/DEBUG) Get level of context + lwz r21,SAVprev+4(r24) ; (TEST/DEBUG) Get previous pointer + mr. r23,r23 ; (TEST/DEBUG) Is this our user context? + beq++ fpulvl0b ; (TEST/DEBUG) Yes... + + lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code + ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest + sc ; (TEST/DEBUG) System ABEND + +fpulvl0b: mr. r21,r21 ; (TEST/DEBUG) Is there a forward chain? + beq++ fpunusrstt ; (TEST/DEBUG) Nope... + + lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code + ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest + sc ; (TEST/DEBUG) System ABEND + +fpunusrstt: ; (TEST/DEBUG) +#endif + + lwz r21,FPUowner(r31) ; Get the ID of the live context + lwz r23,FPUlevel(r26) ; Get the level ID + lwz r24,FPUcpu(r26) ; Get the CPU that the context was last dispatched on + cmplw cr3,r26,r21 ; Do we have the live context? + cmplw r30,r23 ; Are we about to launch the live level? + bne-- cr3,chkvec ; No, can not possibly enable... + cmplw cr1,r19,r24 ; Was facility used on this processor last? + bne-- chkvec ; No, not live... + bne-- cr1,chkvec ; No, wrong cpu, have to enable later.... + + lwz r24,FPUsave(r26) ; Get the first savearea + mr. r24,r24 ; Any savearea? + beq++ fpuena ; Nope... + lwz r25,SAVlevel(r24) ; Get the level of savearea + lwz r0,SAVprev+4(r24) ; Get the previous + + cmplw r30,r25 ; Is savearea for the level we are launching? + bne++ fpuena ; No, just go enable... + + stw r0,FPUsave(r26) ; Pop the chain + + rlwinm r3,r24,0,0,19 ; Find main savearea header + + lwz r8,quickfret(r31) ; Get the first in quickfret list (top) + lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom) + lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top) + lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom) + stw r8,SAVprev(r24) ; Link the old in (top) + stw r9,SAVprev+4(r24) ; Link the old in (bottom) + xor r3,r24,r3 ; Convert to physical + stw r2,quickfret(r31) ; Set the first in quickfret list (top) + stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom) + #if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3504 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - mr r5,r25 ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - mr. r25,r25 ; Is there one? - stw r25,ACT_MACT_VMX(r8) ; Set the new pointer - beq vectoplvl ; Nope, we are at the top... - lwz r25,SAVlvlvec(r25) ; Get the new level + lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) + li r2,0x3302 ; (TEST/DEBUG) + oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) + sc ; (TEST/DEBUG) +#endif -vectoplvl: lwz r19,SAVflags(r24) ; Get the savearea flags +fpuena: ori r29,r29,lo16(MASK(MSR_FP)) ; Enable facility + +chkvec: -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3405 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - rlwinm r22,r24,0,0,19 ; Round down to the base savearea block - rlwinm r19,r19,0,3,1 ; Remove the vector in use flag - stw r25,ACT_MACT_VMXlvl(r8) ; Set the new top level - andis. r0,r19,hi16(SAVinuse) ; Still in use? - stw r19,SAVflags(r24) ; Set the savearea flags - bne- invliveve ; Go invalidate live vec... -#if FPVECDBG - lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) - li r2,0x3406 ; (TEST/DEBUG) - oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) - sc ; (TEST/DEBUG) -#endif - lwz r23,SACvrswap(r22) ; Get the conversion from virtual to real - lwz r20,PP_QUICKFRET(r10) ; Get the old quick fret head - xor r23,r24,r23 ; Convert to physical - stw r20,SAVqfret(r24) ; Back chain the quick release queue - stw r23,PP_QUICKFRET(r10) ; Anchor it - -invliveve: lis r20,hi16(EXT(real_ncpus)) ; Get number of CPUs - lis r23,hi16(EXT(per_proc_info)) ; Set base per_proc - ori r20,r20,lo16(EXT(real_ncpus)) ; Other half of number of CPUs - li r25,PP_VMX_THREAD ; Point to the vector owner address - lwz r20,0(r20) ; Get number of processors active - ori r23,r23,lo16(EXT(per_proc_info)) ; Set base per_proc - li r2,0 ; Get something clear - -invlivevl: cmplw r23,r10 ; We can skip our processor - addi r20,r20,-1 ; Count remaining processors - beq invlivevn ; Skip ourselves... - -invliveva: lwarx r0,r25,r23 ; Get vector owner for this processor - cmplw r0,r8 ; Do we own it? - bne invlivevn ; Nope... - stwcx. r2,r25,r23 ; Show not live - bne- invliveva ; Someone else did this, try again... + lwz r21,VMXlevel(r20) ; Get the facility level -invlivevn: mr. r20,r20 ; Have we finished? - addi r23,r23,ppSize ; Bump to next - bgt invlivevl ; Make sure we do all processors... + cmplw r27,r21 ; Are we returning from the active level? + bne+ vmxchkena ; Nope... + -setena: sync ; Make sure all is saved - stw r27,ACT_MACT_VMXcpu(r8) ; Set the active CPU and release +; +; First clean up any live context we are returning from +; -setenanr: rlwinm r20,r12,(((31-vectorCngbit)+(MSR_VEC_BIT+1))&31),vectorCngbit,vectorCngbit ; Set flag if we enabled vector - rlwimi. r20,r12,(((31-floatCngbit)+(MSR_FP_BIT+1))&31),floatCngbit,floatCngbit ; Set flag if we enabled floats - beq setenaa ; Neither float nor vector turned on.... + lwz r22,VMXcpu(r20) ; Get CPU this context was last dispatched on + + stw r19,VMXcpu(r20) ; Claim context for us - lwz r5,ACT_MACT_SPF(r8) ; Get activation copy - lwz r6,spcFlags(r10) ; Get per_proc copy - or r5,r5,r20 ; Set vector/float changed bits in activation - or r6,r6,r20 ; Set vector/float changed bits in per_proc - stw r5,ACT_MACT_SPF(r8) ; Set activation copy - stw r6,spcFlags(r10) ; Set per_proc copy + eieio ; Make sure this gets out before owner clear + + lis r23,hi16(EXT(PerProcTable)) ; Set base PerProcTable + slwi r22,r22,4 ; Find offset to the owner per_proc_entry + ori r23,r23,lo16(EXT(PerProcTable)) ; Set base PerProcTable + li r24,VMXowner ; Displacement to float owner + add r22,r23,r22 ; Point to the owner per_proc_entry + lwz r22,ppe_vaddr(r22) ; Point to the owner per_proc + +vmxinvothr: lwarx r23,r24,r22 ; Get the owner -setenaa: stw r12,savesrr1(r3) ; Turn facility on or off - - mfdec r24 ; Get decrementer - lwz r22,qactTimer(r8) ; Get high order quick activation timer - mr. r24,r24 ; See if it has popped already... - lwz r23,qactTimer+4(r8) ; Get low order qact timer - ble- chkenax ; We have popped or are just about to... - -segtb: mftbu r20 ; Get the upper time base - mftb r21 ; Get the low - mftbu r19 ; Get upper again - or. r0,r22,r23 ; Any time set? - cmplw cr1,r20,r19 ; Did they change? - beq+ chkenax ; No time set.... - bne- cr1,segtb ; Timebase ticked, get them again... - - subfc r6,r21,r23 ; Subtract current from qact time - li r0,0 ; Make a 0 - subfe r5,r20,r22 ; Finish subtract - subfze r0,r0 ; Get a 0 if qact was bigger than current, -1 otherwise - andc. r12,r5,r0 ; Set 0 if qact has passed - andc r13,r6,r0 ; Set 0 if qact has passed - bne chkenax ; If high order is non-zero, this is too big for a decrementer - cmplw r13,r24 ; Is this earlier than the decrementer? (logical compare takes care of high bit on) - bge+ chkenax ; No, do not reset decrementer... - - mtdec r13 ; Set our value - -chkenax: lwz r6,SAVflags(r3) ; Pick up the flags of the old savearea + sub r0,r23,r20 ; Subtract one from the other + sub r21,r20,r23 ; Subtract the other from the one + or r21,r21,r0 ; Combine them + srawi r21,r21,31 ; Get a 0 if equal or -1 of not + and r23,r23,r21 ; Make 0 if same, unchanged if not + stwcx. r23,r24,r22 ; Try to invalidate it + bne-- vmxinvothr ; Try again if there was a collision... + + isync + +; +; Now if there is a savearea associated with the popped context, release it. +; Either way, pop the level to the top stacked context. +; + + lwz r22,VMXsave(r20) ; Get pointer to the first savearea + li r21,0 ; Assume we popped all the way out + mr. r22,r22 ; Is there anything there? + beq++ vmxsetlvl ; No, see if we need to enable... + + lwz r21,SAVlevel(r22) ; Get the level of that savearea + cmplw r21,r27 ; Is this the saved copy of the live stuff? + bne vmxsetlvl ; No, leave as is... + + lwz r24,SAVprev+4(r22) ; Pick up the previous area + li r21,0 ; Assume we popped all the way out + mr. r24,r24 ; Any more context? + beq-- vmxonlyone ; Nope... + lwz r21,SAVlevel(r24) ; Get the level associated with save + +vmxonlyone: stw r24,VMXsave(r20) ; Dequeue this savearea + + rlwinm r3,r22,0,0,19 ; Find main savearea header + + lwz r8,quickfret(r31) ; Get the first in quickfret list (top) + lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom) + lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top) + lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom) + stw r8,SAVprev(r22) ; Link the old in (top) + stw r9,SAVprev+4(r22) ; Link the old in (bottom) + xor r3,r22,r3 ; Convert to physical + stw r2,quickfret(r31) ; Set the first in quickfret list (top) + stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom) + +#if FPVECDBG + lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) + li r2,0x3401 ; (TEST/DEBUG) + oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) + sc ; (TEST/DEBUG) +#endif + +vmxsetlvl: stw r21,VMXlevel(r20) ; Save the level + +; +; Here we check if we are at the right level +; + +vmxchkena: lwz r21,VMXowner(r31) ; Get the ID of the live context + lwz r23,VMXlevel(r26) ; Get the level ID + cmplw r26,r21 ; Do we have the live context? + lwz r24,VMXcpu(r26) ; Get the CPU that the context was last dispatched on + bne-- setena ; No, can not possibly enable... + cmplw r30,r23 ; Are we about to launch the live level? + cmplw cr1,r19,r24 ; Was facility used on this processor last? + bne-- setena ; No, not live... + bne-- cr1,setena ; No, wrong cpu, have to enable later.... + + lwz r24,VMXsave(r26) ; Get the first savearea + mr. r24,r24 ; Any savearea? + beq++ vmxena ; Nope... + lwz r25,SAVlevel(r24) ; Get the level of savearea + lwz r0,SAVprev+4(r24) ; Get the previous + cmplw r30,r25 ; Is savearea for the level we are launching? + bne++ vmxena ; No, just go enable... + + stw r0,VMXsave(r26) ; Pop the chain + + rlwinm r3,r24,0,0,19 ; Find main savearea header + + lwz r8,quickfret(r31) ; Get the first in quickfret list (top) + lwz r9,quickfret+4(r31) ; Get the first in quickfret list (bottom) + lwz r2,SACvrswap(r3) ; Get the virtual to real conversion (top) + lwz r3,SACvrswap+4(r3) ; Get the virtual to real conversion (bottom) + stw r8,SAVprev(r24) ; Link the old in (top) + stw r9,SAVprev+4(r24) ; Link the old in (bottom) + xor r3,r24,r3 ; Convert to physical + stw r2,quickfret(r31) ; Set the first in quickfret list (top) + stw r3,quickfret+4(r31) ; Set the first in quickfret list (bottom) + +#if FPVECDBG + lis r0,HIGH_ADDR(CutTrace) ; (TEST/DEBUG) + li r2,0x3402 ; (TEST/DEBUG) + oris r0,r0,LOW_ADDR(CutTrace) ; (TEST/DEBUG) + sc ; (TEST/DEBUG) +#endif + +vmxena: oris r29,r29,hi16(MASK(MSR_VEC)) ; Enable facility + +setena: lwz r18,umwSpace(r28) ; Get the space ID in case we are launching user + rlwinm. r0,r29,0,MSR_PR_BIT,MSR_PR_BIT ; Are we about to launch user state? + li r0,0 ; Get set to release quickfret holdoff + crmove cr7_eq,cr0_eq ; Remember if we are going to user state + rlwimi. r20,r29,(((31-floatCngbit)+(MSR_FP_BIT+1))&31),floatCngbit,floatCngbit ; Set flag if we enabled floats + lwz r19,deferctx(r28) ; Get any deferred facility context switch + rlwinm r20,r29,(((31-vectorCngbit)+(MSR_VEC_BIT+1))&31),vectorCngbit,vectorCngbit ; Set flag if we enabled vector + stw r29,savesrr1+4(r27) ; Turn facility on or off + stw r0,holdQFret(r31) ; Release quickfret + oris r18,r18,hi16(umwSwitchAway) ; Set the switch-away bit in case we go to user + + beq setenaa ; Neither float nor vector turned on.... + + lwz r5,ACT_MACT_SPF(r28) ; Get activation copy + lwz r6,spcFlags(r31) ; Get per_proc copy + or r5,r5,r20 ; Set vector/float changed bits in activation + or r6,r6,r20 ; Set vector/float changed bits in per_proc + stw r5,ACT_MACT_SPF(r28) ; Set activation copy + stw r6,spcFlags(r31) ; Set per_proc copy + +setenaa: mfdec r24 ; Get decrementer + bf+ cr2_eq,nodefer ; No deferred to switch to... + + li r20,0 ; Clear this + stw r26,curctx(r28) ; Make the facility context current + stw r20,deferctx(r28) ; Clear deferred context + +nodefer: lwz r22,qactTimer(r28) ; Get high order quick activation timer + mr. r24,r24 ; See if it has popped already... + lwz r23,qactTimer+4(r28) ; Get low order qact timer + ble- chkifuser ; We have popped or are just about to... + +segtb: mftbu r20 ; Get the upper time base + mftb r21 ; Get the low + mftbu r19 ; Get upper again + or. r0,r22,r23 ; Any time set? + cmplw cr1,r20,r19 ; Did they change? + beq++ chkifuser ; No time set.... + bne-- cr1,segtb ; Timebase ticked, get them again... + + subfc r6,r21,r23 ; Subtract current from qact time + li r0,0 ; Make a 0 + subfe r5,r20,r22 ; Finish subtract + subfze r0,r0 ; Get a 0 if qact was bigger than current, -1 otherwise + andc. r12,r5,r0 ; Set 0 if qact has passed + andc r13,r6,r0 ; Set 0 if qact has passed + bne chkifuser ; If high order is non-zero, this is too big for a decrementer + cmplw r13,r24 ; Is this earlier than the decrementer? (logical compare takes care of high bit on) + bge++ chkifuser ; No, do not reset decrementer... + + mtdec r13 ; Set our value + +chkifuser: addi r4,r28,SYSTEM_TIMER + mftb r3 + beq-- cr7,chkifuser1 ; Skip this if we are going to kernel... + stw r18,umwSpace(r28) ; Half-invalidate to force MapUserAddressWindow to reload SRs + addi r4,r28,USER_TIMER + +chkifuser1: bl EXT(timer_event) + +chkenax: - #if DEBUG - lwz r20,SAVact(r3) ; (TEST/DEBUG) Make sure our restore - lwz r21,PP_CPU_DATA(r10) ; (TEST/DEBUG) context is associated - lwz r21,CPU_ACTIVE_THREAD(r21) ; (TEST/DEBUG) with the current act. - cmpwi r21,0 ; (TEST/DEBUG) - beq- yeswereok ; (TEST/DEBUG) - lwz r21,THREAD_TOP_ACT(r21) ; (TEST/DEBUG) - cmplw r21,r20 ; (TEST/DEBUG) - beq+ yeswereok ; (TEST/DEBUG) - - lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code - ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest - mr r21,r3 ; (TEST/DEBUG) Save the savearea address - li r3,failContext ; (TEST/DEBUG) Bad state code - sc ; (TEST/DEBUG) System ABEND + lwz r20,SAVact(r27) ; (TEST/DEBUG) Make sure our restore + mfsprg r21, 1 ; (TEST/DEBUG) with the current act. + cmpwi r21,0 ; (TEST/DEBUG) + beq-- yeswereok ; (TEST/DEBUG) + cmplw r21,r20 ; (TEST/DEBUG) + beq++ yeswereok ; (TEST/DEBUG) + + lis r0,hi16(Choke) ; (TEST/DEBUG) Choke code + ori r0,r0,lo16(Choke) ; (TEST/DEBUG) and the rest + mr r21,r27 ; (TEST/DEBUG) Save the savearea address + li r3,failContext ; (TEST/DEBUG) Bad state code + sc ; (TEST/DEBUG) System ABEND yeswereok: #endif - rlwinm r5,r3,0,0,19 ; Round savearea down to page bndry - rlwinm r6,r6,0,1,31 ; Mark savearea free - lwz r5,SACvrswap(r5) ; Get the conversion from virtual to real - stw r6,SAVflags(r3) ; Set savearea flags - xor r3,r3,r5 ; Flip to physical address - b EXT(exception_exit) ; We are all done now... + mr r3,r27 ; Pass savearea back + b EXT(exception_exit) ; We are all done now... + + + +; +; Null PPC call - performance testing, does absolutely nothing +; + + .align 5 + + .globl EXT(ppcNull) + +LEXT(ppcNull) + + li r3,-1 ; Make sure we test no asts + blr + + +; +; Instrumented null PPC call - performance testing, does absolutely nothing +; Forces various timestamps to be returned. +; + .align 5 + + .globl EXT(ppcNullinst) + +LEXT(ppcNullinst) + + li r3,-1 ; Make sure we test no asts + blr /* @@ -1501,37 +1502,38 @@ yeswereok: .align 5 -fastpath: cmplwi cr3,r0,0x7FF1 ; Is it CthreadSetSelfNumber? - bnelr- cr3 ; Not a fast path... +fastpath: cmplwi cr3,r0,0x7FF5 ; Is this a null fastpath? + beq-- cr3,fastexutl ; Yes, bail fast... + cmplwi cr3,r0,0x7FF1 ; Is it CthreadSetSelfNumber? + bnelr-- cr3 ; Not a fast path... /* * void cthread_set_self(cproc_t p) * - * set's thread state "user_value" + * Set's thread state "user_value". In practice this is the thread-local-data-pointer (TLDP), + * though we do not interpret it. This call is mostly used by 32-bit tasks, but we save all 64 bits + * in case a 64-bit task wants to use this facility. They normally do not, because the 64-bit + * ABI reserves r13 for the TLDP. * * This op is invoked as follows: * li r0, CthreadSetSelfNumber // load the fast-trap number * sc // invoke fast-trap * blr - * */ CthreadSetSelfNumber: - - lwz r5,saver3(r4) /* Retrieve the self number */ - stw r5,CTHREAD_SELF(r13) /* Remember it */ - stw r5,UAW(r25) /* Prime the per_proc_info with it */ + lwz r3,saver3+0(r4) /* get the TLDP passed in r3 */ + lwz r5,saver3+4(r4) /* (all 64 bits, in case this is a 64-bit task) */ + stw r3,CTHREAD_SELF+0(r13) /* Remember it in the activation... */ + stw r5,CTHREAD_SELF+4(r13) + stw r3,UAW+0(r25) /* ...and in the per-proc */ + stw r5,UAW+4(r25) .globl EXT(fastexit) EXT(fastexit): - lwz r8,SAVflags(r4) /* Pick up the flags */ - rlwinm r9,r4,0,0,19 /* Round down to the base savearea block */ - rlwinm r8,r8,0,1,31 /* Clear the attached bit */ - lwz r9,SACvrswap(r9) /* Get the conversion from virtual to real */ - stw r8,SAVflags(r4) /* Set the flags */ - xor r3,r4,r9 /* Switch savearea to physical addressing */ - b EXT(exception_exit) /* Go back to the caller... */ +fastexutl: mr r3,r4 ; Pass back savearea + b EXT(exception_exit) ; Go back to the caller... /* @@ -1545,12 +1547,12 @@ EXT(fastexit): checkassist: lwz r0,saveexception(r4) ; Get the exception code - lwz r23,savesrr1(r4) ; Get the interrupted MSR + lwz r23,savesrr1+4(r4) ; Get the interrupted MSR lwz r26,ACT_MACT_BEDA(r13) ; Get Blue Box Descriptor Area mtcrf 0x18,r23 ; Check what SRR1 says lwz r24,ACT_MACT_BTS(r13) ; Get the table start cmplwi r0,T_AST ; Check for T_AST trap - lwz r27,savesrr0(r4) ; Get trapped address + lwz r27,savesrr0+4(r4) ; Get trapped address crnand cr1_eq,SRR1_PRG_TRAP_BIT,MSR_PR_BIT ; We need both trap and user state sub r24,r27,r24 ; See how far into it we are cror cr0_eq,cr0_eq,cr1_eq ; Need to bail if AST or not trap or not user state @@ -1578,13 +1580,14 @@ exitFromVM: mr r30,r4 ; Get the savearea .globl EXT(retFromVM) LEXT(retFromVM) - mfsprg r10,0 ; Restore the per_proc info + mfsprg r10,1 ; Get the current activation + lwz r10,ACT_PER_PROC(r10) ; Get the per_proc block mr r8,r3 ; Get the activation - lwz r4,SAVprev(r30) ; Pick up the previous savearea + lwz r4,SAVprev+4(r30) ; Pick up the previous savearea mr r3,r30 ; Put savearea in proper register for common code lwz r11,SAVflags(r30) ; Get the flags of the current savearea rlwinm r11,r11,0,15,13 ; Clear the syscall flag - lwz r1,ACT_THREAD(r8) ; and the active thread + mr r1,r8 stw r11,SAVflags(r3) ; Save back the flags (with reset stack cleared) stw r4,ACT_MACT_PCB(r8) ; Point to the previous savearea (or 0 if none) @@ -1614,15 +1617,14 @@ LEXT(retFromVM) .align 5 .globl EXT(chandler) -LEXT(chandler) /* Choke handler */ +LEXT(chandler) ; Choke handler - lis r25,hi16(EXT(trcWork)) ; (TEST/DEBUG) - li r31,0 ; (TEST/DEBUG) - ori r25,r25,lo16(EXT(trcWork)) ; (TEST/DEBUG) - stw r31,traceMask(r25) ; (TEST/DEBUG) + li r31,0 ; Get a 0 + mfsprg r25,1 ; Get the current activation + lwz r25,ACT_PER_PROC(r25) ; Get the per_proc block + stw r31,traceMask(0) ; Force tracing off right now - mfsprg r25,0 ; Get the per_proc lwz r1,PP_DEBSTACKPTR(r25) ; Get debug stack pointer cmpwi r1,-1 ; Are we already choking? @@ -1638,7 +1640,7 @@ chokespin: addi r31,r31,1 ; Spin and hope for an analyzer connection... chokefirst: li r0,-1 ; Set choke value mr. r1,r1 ; See if we are on debug stack yet - lwz r10,saver1(r4) ; + lwz r10,saver1+4(r4) ; stw r0,PP_DEBSTACKPTR(r25) ; Show we are choking bne chokestart ; We are not on the debug stack yet... @@ -1663,63 +1665,26 @@ chokestart: li r0,0 ; Get a zero ; versave: - #if 0 -; -; Make sure that only the top FPU savearea is marked invalid -; - - lis r28,hi16(EXT(default_pset)) ; (TEST/DEBUG) - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r28,r28,lo16(EXT(default_pset)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - li r20,0 ; (TEST/DEBUG) - lwz r26,0(r27) ; (TEST/DEBUG) - lwz r27,psthreadcnt(r28) ; (TEST/DEBUG) - mr. r26,r26 ; (TEST/DEBUG) - lwz r28,psthreads(r28) ; (TEST/DEBUG) - bnelr- ; (TEST/DEBUG) - -fcknxtth: mr. r27,r27 ; (TEST/DEBUG) + lis r22,hi16(EXT(DebugWork)) ; (TEST/DEBUG) + ori r22,r22,lo16(EXT(DebugWork)) ; (TEST/DEBUG) + lwz r23,0(r22) ; (TEST/DEBUG) + mr. r23,r23 ; (TEST/DEBUG) beqlr- ; (TEST/DEBUG) - - lwz r26,THREAD_TOP_ACT(r28) ; (TEST/DEBUG) - -fckact: mr. r26,r26 ; (TEST/DEBUG) - bne+ fckact2 ; (TEST/DEBUG) - - lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Next in line - subi r27,r27,1 ; (TEST/DEBUG) - b fcknxtth ; (TEST/DEBUG) - -fckact2: lwz r20,ACT_MACT_FPU(r26) ; (TEST/DEBUG) Get FPU chain - mr. r20,r20 ; (TEST/DEBUG) Are there any? - beq+ fcknact ; (TEST/DEBUG) No... - -fckact3: lwz r20,SAVprefp(r20) ; (TEST/DEBUG) Get next in list - mr. r20,r20 ; (TEST/DEBUG) Check next savearea - beq+ fcknact ; (TEST/DEBUG) No... - - lwz r29,SAVlvlfp(r20) ; (TEST/DEBUG) Get the level - - cmplwi r29,1 ; (TEST/DEBUG) Is it invalid?? - bne+ fckact3 ; (TEST/DEBUG) Nope... - - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - stw r27,0(r27) ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) - -fcknact: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation - b fckact ; (TEST/DEBUG) + mfsprg r20,1 ; Get the current activation + lwz r20,ACT_PER_PROC(r20) ; Get the per_proc block + lwz r21,pfAvailable(r20) ; (TEST/DEBUG) + mr. r21,r21 ; (TEST/DEBUG) + bnelr+ ; (TEST/DEBUG) + + stw r22,0(r22) ; (TEST/DEBUG) Lock out more checks + BREAKPOINT_TRAP ; (TEST/DEBUG) Get into debugger #endif -#if 1 +#if 0 + ;; This code is broken and migration will make the matter even worse ; -; Make sure there are no circular links in the float chain -; And that FP is marked busy in it. -; And the only the top is marked invalid. -; And that the owning PCB is correct. +; Make sure that all savearea chains have the right type on them ; lis r28,hi16(EXT(default_pset)) ; (TEST/DEBUG) @@ -1729,171 +1694,76 @@ fcknact: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation li r20,0 ; (TEST/DEBUG) lwz r26,0(r27) ; (TEST/DEBUG) lwz r27,psthreadcnt(r28) ; (TEST/DEBUG) - mr. r26,r26 ; (TEST/DEBUG) + mr. r26,r26 ; (TEST/DEBUG) Have we locked the test out? lwz r28,psthreads(r28) ; (TEST/DEBUG) - bnelr- ; (TEST/DEBUG) - -fcknxtth: mr. r27,r27 ; (TEST/DEBUG) - beqlr- ; (TEST/DEBUG) + mflr r31 ; (TEST/DEBUG) Save return + bnelr- ; (TEST/DEBUG) Test already triggered, skip... + b fckgo ; (TEST/DEBUG) Join up... - lwz r26,THREAD_TOP_ACT(r28) ; (TEST/DEBUG) +fcknext: mr. r27,r27 ; (TEST/DEBUG) Any more threads? + bne+ fckxxx ; (TEST/DEBUG) Yes... -fckact: mr. r26,r26 ; (TEST/DEBUG) - bne+ fckact2 ; (TEST/DEBUG) + mtlr r31 ; (TEST/DEBUG) Restore return + blr ; (TEST/DEBUG) Leave... - lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Next in line - subi r27,r27,1 ; (TEST/DEBUG) - b fcknxtth ; (TEST/DEBUG) - -fckact2: lwz r20,ACT_MACT_FPU(r26) ; (TEST/DEBUG) Get FPU chain - li r29,1 ; (TEST/DEBUG) - li r22,0 ; (TEST/DEBUG) +fckxxx: lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Get next thread -fckact3: mr. r20,r20 ; (TEST/DEBUG) Are there any? - beq+ fckact5 ; (TEST/DEBUG) No... +fckgo: subi r27,r27,1 ; (TEST/DEBUG) Decrement thread count + lwz r24,THREAD_TOP_ACT(r28) ; (TEST/DEBUG) Get activation for the thread + lwz r20,ACT_MACT_PCB(r24) ; (TEST/DEBUG) Get the normal context + li r21,SAVgeneral ; (TEST/DEBUG) Make sure this is all general context + bl versavetype ; (TEST/DEBUG) Check the chain - addi r22,r22,1 ; (TEST/DEBUG) Count chain depth + lwz r20,facctx+FPUsave(r24) ; (TEST/DEBUG) Get regular floating point + li r21,SAVfloat ; (TEST/DEBUG) Make sure this is all floating point + bl versavetype ; (TEST/DEBUG) Check the chain - lwz r21,SAVflags(r20) ; (TEST/DEBUG) Get the flags - rlwinm. r21,r21,0,1,1 ; (TEST/DEBUG) FP busy? - bne+ fckact3a ; (TEST/DEBUG) Yeah... - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - stw r27,0(r27) ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) Die - -fckact3a: cmplwi r22,1 ; (TEST/DEBUG) At first SA? - beq+ fckact3b ; (TEST/DEBUG) Yeah, invalid is ok... - lwz r21,SAVlvlfp(r20) ; (TEST/DEBUG) Get level - cmplwi r21,1 ; (TEST/DEBUG) Is it invalid? - bne+ fckact3b ; (TEST/DEBUG) Nope, it is ok... - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - stw r27,0(r27) ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) Die + lwz r20,facctx+VMXsave(r24) ; (TEST/DEBUG) Get regular vector point + li r21,SAVvector ; (TEST/DEBUG) Make sure this is all vector + bl versavetype ; (TEST/DEBUG) Check the chain -fckact3b: lwz r21,SAVact(r20) ; (TEST/DEBUG) Get the owner - cmplw r21,r26 ; (TEST/DEBUG) Correct activation? - beq+ fckact3c ; (TEST/DEBUG) Yup... - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - stw r27,0(r27) ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) Die - -fckact3c: ; (TEST/DEBUG) - lbz r21,SAVflags+3(r20) ; (TEST/DEBUG) Pick up the test byte - mr. r21,r21 ; (TEST/DEBUG) marked? - beq+ fckact4 ; (TEST/DEBUG) No, good... - - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - stw r27,0(r27) ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) - -fckact4: stb r29,SAVflags+3(r20) ; (TEST/DEBUG) Set the test byte - lwz r20,SAVprefp(r20) ; (TEST/DEBUG) Next in list - b fckact3 ; (TEST/DEBUG) Try it... - -fckact5: lwz r20,ACT_MACT_FPU(r26) ; (TEST/DEBUG) Get FPU chain - li r29,0 ; (TEST/DEBUG) - -fckact6: mr. r20,r20 ; (TEST/DEBUG) Are there any? - beq+ fcknact ; (TEST/DEBUG) No... - - stb r29,SAVflags+3(r20) ; (TEST/DEBUG) Clear the test byte - lwz r20,SAVprefp(r20) ; (TEST/DEBUG) Next in list - b fckact6 ; (TEST/DEBUG) Try it... - -fcknact: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation - b fckact ; (TEST/DEBUG) -#endif - - -#if 0 -; -; Make sure in use count matches found savearea. This is -; not always accurate. There is a variable "fuzz" factor in count. - - lis r28,hi16(EXT(default_pset)) ; (TEST/DEBUG) - lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r28,r28,lo16(EXT(default_pset)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - li r20,0 ; (TEST/DEBUG) - lwz r26,0(r27) ; (TEST/DEBUG) - lwz r27,psthreadcnt(r28) ; (TEST/DEBUG) - mr. r26,r26 ; (TEST/DEBUG) - lwz r28,psthreads(r28) ; (TEST/DEBUG) - bnelr- ; (TEST/DEBUG) + lwz r29,vmmControl(r24) ; (TEST/DEBUG) Get the virtual machine control blocks + mr. r29,r29 ; (TEST/DEBUG) Are there any? + beq+ fcknext ; (TEST/DEBUG) Nope, next thread... -cknxtth: mr. r27,r27 ; (TEST/DEBUG) - beq- cktotal ; (TEST/DEBUG) + li r22,kVmmMaxContextsPerThread ; (TEST/DEBUG) Get the number of control blocks + subi r29,r29,vmmCEntrySize ; (TEST/DEBUG) Get running start - lwz r26,THREAD_TOP_ACT(r28) ; (TEST/DEBUG) - -ckact: mr. r26,r26 ; (TEST/DEBUG) - bne+ ckact2 ; (TEST/DEBUG) - - lwz r28,THREAD_PSTHRN(r28) ; (TEST/DEBUG) Next in line - subi r27,r27,1 ; (TEST/DEBUG) - b cknxtth ; (TEST/DEBUG) - -ckact2: lwz r29,ACT_MACT_PCB(r26) ; (TEST/DEBUG) +fcknvmm: subi r22,r22,1 ; (TEST/DEBUG) Do all of them + mr. r22,r22 ; (TEST/DEBUG) Are we all done? + addi r29,r29,vmmCEntrySize ; (TEST/DEBUG) Get the next entry + blt- fcknext ; (TEST/DEBUG) Yes, check next thread... -cknorm: mr. r29,r29 ; (TEST/DEBUG) - beq- cknormd ; (TEST/DEBUG) + lwz r23,vmmFlags(r29) ; (TEST/DEBUG) Get entry flags + rlwinm. r23,r23,0,0,0 ; (TEST/DEBUG) Is this in use? + beq+ fcknvmm ; (TEST/DEBUG) Not in use... - addi r20,r20,1 ; (TEST/DEBUG) Count normal savearea + lwz r20,vmmFacCtx+FPUsave(r29) ; (TEST/DEBUG) Get regular floating point + li r21,SAVfloat ; (TEST/DEBUG) Make sure this is all floating point + bl versavetype ; (TEST/DEBUG) Check the chain - lwz r29,SAVprev(r29) ; (TEST/DEBUG) - b cknorm ; (TEST/DEBUG) - -cknormd: lwz r29,ACT_MACT_FPU(r26) ; (TEST/DEBUG) - -ckfpu: mr. r29,r29 ; (TEST/DEBUG) - beq- ckfpud ; (TEST/DEBUG) + lwz r20,vmmFacCtx+VMXsave(r29) ; (TEST/DEBUG) Get regular vector point + li r21,SAVvector ; (TEST/DEBUG) Make sure this is all vector + bl versavetype ; (TEST/DEBUG) Check the chain + b fcknvmm ; (TEST/DEBUG) Get then vmm block... - lwz r21,SAVflags(r29) ; (TEST/DEBUG) - rlwinm. r21,r21,0,0,0 ; (TEST/DEBUG) See if already counted - bne- cknfpu ; (TEST/DEBUG) +versavetype: + mr. r20,r20 ; (TEST/DEBUG) Chain done? + beqlr- ; (TEST/DEBUG) Yes... - addi r20,r20,1 ; (TEST/DEBUG) Count fpu savearea + lwz r23,SAVflags(r20) ; (TEST/DEBUG) Get the flags + rlwinm r23,r23,24,24,31 ; (TEST/DEBUG) Position it + cmplw r23,r21 ; (TEST/DEBUG) Are we the correct type? + beq+ versvok ; (TEST/DEBUG) This one is ok... -cknfpu: lwz r29,SAVprefp(r29) ; (TEST/DEBUG) - b ckfpu ; (TEST/DEBUG) + lis r22,hi16(EXT(DebugWork)) ; (TEST/DEBUG) + ori r22,r22,lo16(EXT(DebugWork)) ; (TEST/DEBUG) + stw r22,0(r22) ; (TEST/DEBUG) Lock out more checks + BREAKPOINT_TRAP ; (TEST/DEBUG) Get into debugger -ckfpud: lwz r29,ACT_MACT_VMX(r26) ; (TEST/DEBUG) - -ckvmx: mr. r29,r29 ; (TEST/DEBUG) - beq- ckvmxd ; (TEST/DEBUG) - - lwz r21,SAVflags(r29) ; (TEST/DEBUG) - rlwinm. r21,r21,0,0,1 ; (TEST/DEBUG) See if already counted - bne- cknvmx ; (TEST/DEBUG) - - addi r20,r20,1 ; (TEST/DEBUG) Count vector savearea - -cknvmx: lwz r29,SAVprevec(r29) ; (TEST/DEBUG) - b ckvmx ; (TEST/DEBUG) - -ckvmxd: lwz r26,ACT_LOWER(r26) ; (TEST/DEBUG) Next activation - b ckact ; (TEST/DEBUG) +versvok: lwz r20,SAVprev+4(r20) ; (TEST/DEBUG) Get the previous one + b versavetype ; (TEST/DEBUG) Go check its type... +#endif -cktotal: lis r28,hi16(EXT(saveanchor)) ; (TEST/DEBUG) - lis r27,hi16(EXT(real_ncpus)) ; (TEST/DEBUG) - ori r28,r28,lo16(EXT(saveanchor)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(real_ncpus)) ; (TEST/DEBUG) - lwz r21,SVinuse(r28) ; (TEST/DEBUG) - lwz r27,0(r27) ; (TEST/DEBUG) Get the number of CPUs - sub. r29,r21,r20 ; (TEST/DEBUG) Get number accounted for - blt- badsave ; (TEST/DEBUG) Have too many in use... - sub r26,r29,r27 ; (TEST/DEBUG) Should be 1 unaccounted for for each processor - cmpwi r26,10 ; (TEST/DEBUG) Allow a 10 area slop factor - bltlr+ ; (TEST/DEBUG) - -badsave: lis r27,hi16(EXT(DebugWork)) ; (TEST/DEBUG) - ori r27,r27,lo16(EXT(DebugWork)) ; (TEST/DEBUG) - stw r27,0(r27) ; (TEST/DEBUG) - BREAKPOINT_TRAP ; (TEST/DEBUG) -#endif #endif