/*
- * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
*
* @APPLE_LICENSE_HEADER_START@
*
- * Copyright (c) 1999-2003 Apple Computer, Inc. All Rights Reserved.
- *
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
*/
#include <cputypes.h>
-#include <cpus.h>
#include <platforms.h>
#include <kern/ast.h>
#include <kern/queue.h>
#include <kern/sched_prim.h>
#include <kern/thread.h>
-#include <kern/thread_act.h>
-#include <kern/thread_swap.h>
#include <kern/processor.h>
#include <kern/spl.h>
+#include <kern/wait_queue.h>
#include <mach/policy.h>
-volatile ast_t need_ast[NCPUS];
+#ifdef __ppc__
+#include <ppc/trap.h> // for CHUD AST hook
+#endif
void
ast_init(void)
{
-#ifndef MACHINE_AST
- register int i;
-
- for (i=0; i<NCPUS; i++) {
- need_ast[i] = AST_NONE;
- }
-#endif /* MACHINE_AST */
}
+/*
+ * Called at splsched.
+ */
void
ast_taken(
- ast_t reasons,
- boolean_t enable
+ ast_t reasons,
+ boolean_t enable
)
{
- register int mycpu;
- register processor_t myprocessor;
- register thread_t self = current_thread();
- boolean_t preempt_trap = (reasons == AST_PREEMPT);
-
- disable_preemption();
- mycpu = cpu_number();
- reasons &= need_ast[mycpu];
- need_ast[mycpu] &= ~reasons;
- enable_preemption();
+ boolean_t preempt_trap = (reasons == AST_PREEMPTION);
+ ast_t *myast = ast_pending();
+ thread_t thread = current_thread();
+#ifdef __ppc__
/*
- * No ast for an idle thread
+ * CHUD hook - all threads including idle processor threads
*/
- if (self->state & TH_IDLE)
- goto enable_and_return;
-
- /*
- * Check for urgent preemption
- */
- if ((reasons & AST_URGENT) && wait_queue_assert_possible(self)) {
- if (reasons & AST_BLOCK) {
- counter(c_ast_taken_block++);
- thread_block_reason((void (*)(void))0, AST_BLOCK);
+ if(perfASTHook) {
+ if(*myast & AST_PPC_CHUD_ALL) {
+ perfASTHook(0, NULL, 0, 0);
+
+ if(*myast == AST_NONE) {
+ return; // nothing left to do
+ }
}
-
- reasons &= ~AST_PREEMPT;
- if (reasons == 0)
- goto enable_and_return;
+ } else {
+ *myast &= ~AST_PPC_CHUD_ALL;
}
+#endif
- if (preempt_trap)
- goto enable_and_return;
-
- ml_set_interrupts_enabled(enable);
+ reasons &= *myast;
+ *myast &= ~reasons;
-#ifdef MACH_BSD
/*
- * Check for BSD hook
+ * Handle ASTs for all threads
+ * except idle processor threads.
*/
- if (reasons & AST_BSD) {
- extern void bsd_ast(thread_act_t act);
- thread_act_t act = self->top_act;
+ if (!(thread->state & TH_IDLE)) {
+ /*
+ * Check for urgent preemption.
+ */
+ if ( (reasons & AST_URGENT) &&
+ wait_queue_assert_possible(thread) ) {
+ if (reasons & AST_PREEMPT) {
+ counter(c_ast_taken_block++);
+ thread_block_reason(THREAD_CONTINUE_NULL, NULL,
+ AST_PREEMPT | AST_URGENT);
+ }
+
+ reasons &= ~AST_PREEMPTION;
+ }
- thread_ast_clear(act, AST_BSD);
- bsd_ast(act);
- }
-#endif
+ /*
+ * The kernel preempt traps
+ * skip all other ASTs.
+ */
+ if (!preempt_trap) {
+ ml_set_interrupts_enabled(enable);
- /*
- * migration APC hook
- */
- if (reasons & AST_APC) {
- act_execute_returnhandlers();
- }
+#ifdef MACH_BSD
+ /*
+ * Handle BSD hook.
+ */
+ if (reasons & AST_BSD) {
+ thread_ast_clear(thread, AST_BSD);
+ bsd_ast(thread);
+ }
+#endif
- /*
- * Check for normal preemption
- */
- reasons &= AST_BLOCK;
- if (reasons == 0) {
- disable_preemption();
- myprocessor = current_processor();
- if (csw_needed(self, myprocessor))
- reasons = AST_BLOCK;
- enable_preemption();
- }
- if ( (reasons & AST_BLOCK) &&
- wait_queue_assert_possible(self) ) {
- counter(c_ast_taken_block++);
- thread_block_reason(thread_exception_return, AST_BLOCK);
+ /*
+ * Thread APC hook.
+ */
+ if (reasons & AST_APC)
+ act_execute_returnhandlers();
+
+ ml_set_interrupts_enabled(FALSE);
+
+ /*
+ * Check for preemption.
+ */
+ if (reasons & AST_PREEMPT) {
+ processor_t myprocessor = current_processor();
+
+ if (csw_needed(thread, myprocessor))
+ reasons = AST_PREEMPT;
+ else
+ reasons = AST_NONE;
+ }
+ if ( (reasons & AST_PREEMPT) &&
+ wait_queue_assert_possible(thread) ) {
+ counter(c_ast_taken_block++);
+ thread_block_reason((thread_continue_t)thread_exception_return, NULL, AST_PREEMPT);
+ }
+ }
}
- goto just_return;
-
-enable_and_return:
- ml_set_interrupts_enabled(enable);
-
-just_return:
- return;
+ ml_set_interrupts_enabled(enable);
}
/*
ast_check(
processor_t processor)
{
- register thread_t self = processor->cpu_data->active_thread;
+ register thread_t thread = processor->active_thread;
- processor->current_pri = self->sched_pri;
- if (processor->state == PROCESSOR_RUNNING) {
+ processor->current_pri = thread->sched_pri;
+ if ( processor->state == PROCESSOR_RUNNING ||
+ processor->state == PROCESSOR_SHUTDOWN ) {
register ast_t preempt;
-processor_running:
/*
* Propagate thread ast to processor.
*/
- ast_propagate(self->top_act->ast);
+ ast_propagate(thread->ast);
/*
* Context switch check.
*/
- if ((preempt = csw_check(self, processor)) != AST_NONE)
+ if ((preempt = csw_check(thread, processor)) != AST_NONE)
ast_on(preempt);
}
- else
- if ( processor->state == PROCESSOR_DISPATCHING ||
- processor->state == PROCESSOR_IDLE ) {
- return;
- }
- else
- if (processor->state == PROCESSOR_SHUTDOWN)
- goto processor_running;
- else
- if (processor->state == PROCESSOR_ASSIGN)
- ast_on(AST_BLOCK);
}