/*
- * Copyright (c) 2002,2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2007 Apple Inc. All rights reserved.
*
- * @APPLE_LICENSE_HEADER_START@
- *
- * Copyright (c) 1999-2003 Apple Computer, Inc. All Rights Reserved.
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
- * compliance with the License. Please obtain a copy of the License at
- * http://www.opensource.apple.com/apsl/ and read it before using this
- * file.
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
+ *
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
*
* The Original Code and all software distributed under the License are
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* Please see the License for the specific language governing rights and
* limitations under the License.
*
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
/*
* @OSF_COPYRIGHT@
#include <mach/mach_types.h>
#include <mach/machine/vm_types.h>
-#include <sys/appleapiopts.h>
-
-#if !defined(MACH_KERNEL_PRIVATE)
+#ifdef KERNEL_PRIVATE
-/*
- * Declare empty structure definitions for export to other
- * kernel components. This lets us still provide some level
- * of type checking, without exposing our internal data
- * structures.
- */
+#ifndef MACH_KERNEL_PRIVATE
struct zone ;
-struct wait_queue { unsigned int opaque[2]; uintptr_t opaquep[2]; } ;
-
-#endif /* MACH_KERNEL_PRIVATE */
+#ifndef __LP64__
+struct wait_queue { unsigned int opaque[2]; uintptr_t opaquep[2]; } ;
+#else
+struct wait_queue { unsigned char opaque[32]; };
+#endif
+#endif /* MACH_KERNEL_PRIVATE */
typedef struct zone *zone_t;
#define ZONE_NULL ((zone_t) 0)
typedef vm_offset_t ipc_kobject_t;
#define IKO_NULL ((ipc_kobject_t) 0)
+#endif /* KERNEL_PRIVATE */
+
typedef void *event_t; /* wait event */
#define NO_EVENT ((event_t) 0)
typedef uint64_t event64_t; /* 64 bit wait event */
#define NO_EVENT64 ((event64_t) 0)
+#define CAST_EVENT64_T(a_ptr) ((event64_t)((uintptr_t)(a_ptr)))
/*
* Possible wait_result_t values.
#define THREAD_TIMED_OUT 1 /* timeout expired */
#define THREAD_INTERRUPTED 2 /* aborted/interrupted */
#define THREAD_RESTART 3 /* restart operation entirely */
+#define THREAD_NOT_WAITING 10 /* thread didn't need to wait */
-
-typedef void (*thread_continue_t)(void); /* where to resume it */
+typedef void (*thread_continue_t)(void *, wait_result_t);
#define THREAD_CONTINUE_NULL ((thread_continue_t) 0)
/*
* Interruptible flag for waits.
+ *
+ * THREAD_UNINT: Uninterruptible wait
+ * Wait will only end when someone explicitly wakes up the thread, or if the
+ * wait timeout expires.
+ *
+ * Use this state if the system as a whole cannot recover from a thread being
+ * interrupted out of the wait.
+ *
+ * THREAD_INTERRUPTIBLE:
+ * Wait will end if someone explicitly wakes up the thread, the wait timeout
+ * expires, or the current thread is being terminated.
+ *
+ * This value can be used when your operation may not be cleanly restartable
+ * for the current process or thread (i.e. the loss of state would be only visible
+ * to the current client). Since the thread is exiting anyways, you're willing
+ * to cut the operation short. The system as a whole must be able to cleanly
+ * deal with the interruption (i.e. remain in a consistent and recoverable state).
+ *
+ * THREAD_ABORTSAFE:
+ * Wait will end if someone explicitly wakes up the thread, the wait timeout
+ * expires, the current thread is being terminated, if any signal arrives for
+ * the task, or thread_abort_safely() is called on the thread.
+ *
+ * Using this value means that you are willing to be interrupted in the face
+ * of any user signal, and safely rewind the thread back to the user/kernel
+ * boundary. Many syscalls will try to restart the operation they were performing
+ * after the signal has been handled.
+ *
+ * You must provide this value for any unbounded wait - otherwise you will
+ * pend user signals forever.
+ *
+ * THREAD_WAIT_NOREPORT:
+ * The scheduler has a callback (sched_call) that some subsystems use to
+ * decide whether more threads should be thrown at a given problem by trying
+ * to maintain a good level of concurrency.
+ *
+ * When the wait will not be helped by adding more threads (e.g. lock
+ * contention), using this flag as an argument to assert_wait* (or any of its
+ * wrappers) will prevent the next wait/block to cause thread creation.
+ *
+ * This comes in two flavors: THREAD_WAIT_NOREPORT_KERNEL, and
+ * THREAD_WAIT_NOREPORT_USER to prevent reporting about the wait for kernel
+ * and user threads respectively.
+ *
+ * Thread interrupt mask:
+ *
+ * The current maximum interruptible state for the thread, as set by
+ * thread_interrupt_level(), will limit the conditions that will cause a wake.
+ * This is useful for code that can't be interrupted to set before calling code
+ * that doesn't know that.
+ *
+ * Thread termination vs safe abort:
+ *
+ * Termination abort: thread_abort(), thread_terminate()
+ *
+ * A termination abort is sticky. Once a thread is marked for termination, every
+ * THREAD_INTERRUPTIBLE wait will return immediately with THREAD_INTERRUPTED
+ * until the thread successfully exits.
+ *
+ * Safe abort: thread_abort_safely()
+ *
+ * A safe abort is not sticky. The current wait, (or the next wait if the thread
+ * is not currently waiting) will be interrupted, but then the abort condition is cleared.
+ * The next wait will sleep as normal. Safe aborts only have a single effect.
+ *
+ * The path back to the user/kernel boundary must not make any further unbounded
+ * wait calls. The waiter should detect the THREAD_INTERRUPTED return code
+ * from an ABORTSAFE wait and return an error code that causes its caller
+ * to understand that the current operation has been interrupted, and its
+ * caller should return a similar error code, and so on until the
+ * user/kernel boundary is reached. For Mach, the error code is usually KERN_ABORTED,
+ * for BSD it is EINTR.
+ *
+ * Debuggers rely on the safe abort mechanism - a signaled thread must return to
+ * the AST at the user/kernel boundary for the debugger to finish attaching.
+ *
+ * No wait/block will ever disappear a thread out from under the waiter. The block
+ * call will always either return or call the passed in continuation.
*/
typedef int wait_interrupt_t;
-#define THREAD_UNINT 0 /* not interruptible */
-#define THREAD_INTERRUPTIBLE 1 /* may not be restartable */
-#define THREAD_ABORTSAFE 2 /* abortable safely */
+#define THREAD_UNINT 0x00000000 /* not interruptible */
+#define THREAD_INTERRUPTIBLE 0x00000001 /* may not be restartable */
+#define THREAD_ABORTSAFE 0x00000002 /* abortable safely */
+#define THREAD_WAIT_NOREPORT_KERNEL 0x80000000
+#define THREAD_WAIT_NOREPORT_USER 0x40000000
+#define THREAD_WAIT_NOREPORT (THREAD_WAIT_NOREPORT_KERNEL | THREAD_WAIT_NOREPORT_USER)
-#ifdef __APPLE_API_PRIVATE
+typedef int wait_timeout_urgency_t;
+#define TIMEOUT_URGENCY_SYS_NORMAL 0x00 /* use default leeway thresholds for system */
+#define TIMEOUT_URGENCY_SYS_CRITICAL 0x01 /* use critical leeway thresholds for system */
+#define TIMEOUT_URGENCY_SYS_BACKGROUND 0x02 /* use background leeway thresholds for system */
-#ifdef MACH_KERNEL_PRIVATE
+#define TIMEOUT_URGENCY_USER_MASK 0x10 /* mask to identify user timeout urgency classes */
+#define TIMEOUT_URGENCY_USER_NORMAL 0x10 /* use default leeway thresholds for user */
+#define TIMEOUT_URGENCY_USER_CRITICAL 0x11 /* use critical leeway thresholds for user */
+#define TIMEOUT_URGENCY_USER_BACKGROUND 0x12 /* use background leeway thresholds for user */
-#include <kern/misc_protos.h>
-typedef struct clock *clock_t;
+#define TIMEOUT_URGENCY_MASK 0x13 /* mask to identify timeout urgency */
-#endif /* MACH_KERNEL_PRIVATE */
+#define TIMEOUT_URGENCY_LEEWAY 0x20 /* don't ignore provided leeway value */
-#ifdef __APPLE_API_EVOLVING
+#define TIMEOUT_URGENCY_FIRST_AVAIL 0x40 /* first available bit outside of urgency mask/leeway */
+#define TIMEOUT_URGENCY_RATELIMITED 0x80
-#ifndef MACH_KERNEL_PRIVATE
-struct wait_queue_set ;
-struct wait_queue_link ;
-#endif
+/*
+ * Timeout and deadline tokens for waits.
+ * The following tokens define common values for leeway and deadline parameters.
+ */
+#define TIMEOUT_NO_LEEWAY (0ULL)
+#define TIMEOUT_WAIT_FOREVER (0ULL)
-typedef struct wait_queue_set *wait_queue_set_t;
-#define WAIT_QUEUE_SET_NULL ((wait_queue_set_t)0)
-#define SIZEOF_WAITQUEUE_SET wait_queue_set_size()
+#ifdef KERNEL_PRIVATE
-typedef struct wait_queue_link *wait_queue_link_t;
-#define WAIT_QUEUE_LINK_NULL ((wait_queue_link_t)0)
-#define SIZEOF_WAITQUEUE_LINK wait_queue_link_size()
+/*
+ * n.b. this is defined in thread_call.h, but in the TIMEOUT_URGENCY flags space:
+ * #define THREAD_CALL_CONTINUOUS 0x100
+ */
+
+#ifdef MACH_KERNEL_PRIVATE
+
+#include <kern/misc_protos.h>
+typedef struct clock *clock_t;
typedef struct mig_object *mig_object_t;
#define MIG_OBJECT_NULL ((mig_object_t) 0)
typedef struct mig_notify *mig_notify_t;
#define MIG_NOTIFY_NULL ((mig_notify_t) 0)
-typedef boolean_t (*thread_roust_t)(thread_t, wait_result_t);
-#define THREAD_ROUST_NULL ((thread_roust_t) 0)
+typedef struct pset_node *pset_node_t;
+#define PSET_NODE_NULL ((pset_node_t) 0)
+
+typedef struct affinity_set *affinity_set_t;
+#define AFFINITY_SET_NULL ((affinity_set_t) 0)
+
+typedef struct run_queue *run_queue_t;
+#define RUN_QUEUE_NULL ((run_queue_t) 0)
-#endif /* __APPLE_API_EVOLVING */
+typedef struct grrr_run_queue *grrr_run_queue_t;
+#define GRRR_RUN_QUEUE_NULL ((grrr_run_queue_t) 0)
-#ifdef __APPLE_API_UNSTABLE
+typedef struct grrr_group *grrr_group_t;
+#define GRRR_GROUP_NULL ((grrr_group_t) 0)
-/* legacy definitions - going away */
-typedef struct thread *thread_shuttle_t;
-#define THREAD_SHUTTLE_NULL ((thread_shuttle_t)0)
-struct wait_queue_sub ;
-typedef struct wait_queue_sub *wait_queue_sub_t;
-#define WAIT_QUEUE_SUB_NULL ((wait_queue_sub_t)0)
-#define SIZEOF_WAITQUEUE_SUB wait_queue_set_size()
+#if defined(CONFIG_SCHED_MULTIQ)
+typedef struct sched_group *sched_group_t;
+#define SCHED_GROUP_NULL ((sched_group_t) 0)
+#endif /* defined(CONFIG_SCHED_MULTIQ) */
-#endif /* __APPLE_API_UNSTABLE */
+#else /* MACH_KERNEL_PRIVATE */
-#endif /* __APPLE_API_PRIVATE */
+struct wait_queue_set ;
+struct _wait_queue_link ;
+
+#endif /* MACH_KERNEL_PRIVATE */
+
+typedef struct wait_queue_set *wait_queue_set_t;
+#define WAIT_QUEUE_SET_NULL ((wait_queue_set_t)0)
+#define SIZEOF_WAITQUEUE_SET wait_queue_set_size()
+
+typedef struct _wait_queue_link *wait_queue_link_t;
+#define WAIT_QUEUE_LINK_NULL ((wait_queue_link_t)0)
+#define SIZEOF_WAITQUEUE_LINK wait_queue_link_size()
+
+typedef struct perfcontrol_state *perfcontrol_state_t;
+#define PERFCONTROL_STATE_NULL ((perfcontrol_state_t)0)
+
+/*
+ * Enum to define the event which caused the CLPC callout
+ */
+typedef enum perfcontrol_event {
+ /*
+ * Thread State Update Events
+ * Used to indicate events that update properties for
+ * a given thread. These events are passed as part of the
+ * sched_perfcontrol_state_update_t callout
+ */
+ QUANTUM_EXPIRY = 1,
+ THREAD_GROUP_UPDATE = 2,
+ PERFCONTROL_ATTR_UPDATE = 3,
+ /*
+ * Context Switch Events
+ * Used to indicate events that switch from one thread
+ * to the other. These events are passed as part of the
+ * sched_perfcontrol_csw_t callout.
+ */
+ CONTEXT_SWITCH = 10,
+ IDLE = 11
+} perfcontrol_event;
+
+/*
+ * Flags for the sched_perfcontrol_csw_t & sched_perfcontrol_state_update_t
+ * callouts.
+ * Currently defined flags are:
+ * PERFCONTROL_CALLOUT_WAKE_UNSAFE - Flag to indicate its unsafe to
+ * do a wakeup as part of this callout. If this is set, it
+ * indicates that the scheduler holds a spinlock which might be needed
+ * in the wakeup path. In that case CLPC should do a thread_call
+ * instead of a direct wakeup to run their workloop thread.
+ */
+#define PERFCONTROL_CALLOUT_WAKE_UNSAFE 0x1
+
+/*
+ * Enum to define the perfcontrol class for thread.
+ * thread_get_perfcontrol_class() takes the thread's
+ * priority, QoS, urgency etc. into consideration and
+ * produces a value in this enum.
+ */
+typedef enum perfcontrol_class {
+ /* Idle thread */
+ PERFCONTROL_CLASS_IDLE = 1,
+ /* Kernel thread */
+ PERFCONTROL_CLASS_KERNEL = 2,
+ /* Realtime Thread */
+ PERFCONTROL_CLASS_REALTIME = 3,
+ /* Background Thread */
+ PERFCONTROL_CLASS_BACKGROUND = 4,
+ /* Utility Thread */
+ PERFCONTROL_CLASS_UTILITY = 5,
+ /* Non-UI Thread (Default/Legacy) */
+ PERFCONTROL_CLASS_NONUI = 6,
+ /* UI Thread (UI/IN) */
+ PERFCONTROL_CLASS_UI = 7,
+ /* Above UI Thread */
+ PERFCONTROL_CLASS_ABOVEUI = 8,
+} perfcontrol_class_t;
+
+#endif /* KERNEL_PRIVATE */
#endif /* _KERN_KERN_TYPES_H_ */